USRE44235E1 - Enhancing image quality in an image system - Google Patents

Enhancing image quality in an image system Download PDF

Info

Publication number
USRE44235E1
USRE44235E1 US13/366,485 US201213366485A USRE44235E US RE44235 E1 USRE44235 E1 US RE44235E1 US 201213366485 A US201213366485 A US 201213366485A US RE44235 E USRE44235 E US RE44235E
Authority
US
United States
Prior art keywords
digital video
video image
image
value
filter
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related, expires
Application number
US13/366,485
Inventor
Gary A. Demos
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Dolby Laboratories Licensing Corp
Original Assignee
Dolby Laboratories Licensing Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US08/594,815 external-priority patent/US5852565A/en
Application filed by Dolby Laboratories Licensing Corp filed Critical Dolby Laboratories Licensing Corp
Priority to US13/366,485 priority Critical patent/USRE44235E1/en
Application granted granted Critical
Publication of USRE44235E1 publication Critical patent/USRE44235E1/en
Adjusted expiration legal-status Critical
Expired - Fee Related legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/14Picture signal circuitry for video frequency region
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/13Adaptive entropy coding, e.g. adaptive variable length coding [AVLC] or context adaptive binary arithmetic coding [CABAC]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/30Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
    • H04N19/31Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability in the temporal domain
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/30Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
    • H04N19/33Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability in the spatial domain
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/30Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
    • H04N19/36Scalability techniques involving formatting the layers as a function of picture distortion after decoding, e.g. signal-to-noise [SNR] scalability
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/46Embedding additional information in the video signal during the compression process
    • H04N19/467Embedding additional information in the video signal during the compression process characterised by the embedded information being invisible, e.g. watermarking
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/523Motion estimation or motion compensation with sub-pixel accuracy
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/55Motion estimation with spatial constraints, e.g. at image or region borders
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/567Motion estimation based on rate distortion criteria
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/577Motion compensation with bidirectional frame interpolation, i.e. using B-pictures
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/587Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal sub-sampling or interpolation, e.g. decimation or subsequent interpolation of pictures in a video sequence
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/59Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial sub-sampling or interpolation, e.g. alteration of picture size or resolution
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • H04N19/61Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/80Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/85Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression
    • H04N19/86Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression involving reduction of coding artifacts, e.g. of blockiness
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/426Internal components of the client ; Characteristics thereof
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs
    • H04N21/4402Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display
    • H04N21/440227Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display by decomposing into layers, e.g. base layer and one or more enhancement layers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs
    • H04N21/4402Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display
    • H04N21/440263Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display by altering the spatial resolution, e.g. for displaying on a connected PDA
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs
    • H04N21/4402Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display
    • H04N21/440281Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display by altering the temporal resolution, e.g. by frame skipping
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/14Picture signal circuitry for video frequency region
    • H04N5/144Movement detection
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/14Picture signal circuitry for video frequency region
    • H04N5/21Circuitry for suppressing or minimising disturbance, e.g. moiré or halo
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/01Conversion of standards, e.g. involving analogue television standards or digital television standards processed at pixel level
    • H04N7/0112Conversion of standards, e.g. involving analogue television standards or digital television standards processed at pixel level one of the standards corresponding to a cinematograph film standard
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/01Conversion of standards, e.g. involving analogue television standards or digital television standards processed at pixel level
    • H04N7/0117Conversion of standards, e.g. involving analogue television standards or digital television standards processed at pixel level involving conversion of the spatial resolution of the incoming video signal
    • H04N7/012Conversion between an interlaced and a progressive signal
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/01Conversion of standards, e.g. involving analogue television standards or digital television standards processed at pixel level
    • H04N7/0117Conversion of standards, e.g. involving analogue television standards or digital television standards processed at pixel level involving conversion of the spatial resolution of the incoming video signal
    • H04N7/0122Conversion of standards, e.g. involving analogue television standards or digital television standards processed at pixel level involving conversion of the spatial resolution of the incoming video signal the input and the output signals having different aspect ratios
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/01Conversion of standards, e.g. involving analogue television standards or digital television standards processed at pixel level
    • H04N7/0127Conversion of standards, e.g. involving analogue television standards or digital television standards processed at pixel level by changing the field or frame frequency of the incoming video signal, e.g. frame rate converter
    • H04N7/0132Conversion of standards, e.g. involving analogue television standards or digital television standards processed at pixel level by changing the field or frame frequency of the incoming video signal, e.g. frame rate converter the field or frame frequency of the incoming video signal being multiplied by a positive integer, e.g. for flicker reduction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/30Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
    • H04N19/34Scalability techniques involving progressive bit-plane based encoding of the enhancement layer, e.g. fine granular scalability [FGS]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N9/00Details of colour television systems
    • H04N9/64Circuits for processing colour signals

Definitions

  • This invention relates to electronic communication systems, and more particularly to an advanced electronic television system having enhanced compression, filtering, and display characteristics.
  • interlace is required, due to a claimed need to have about 1000 lines of resolution at high frame rates, but based upon the notion that such images cannot be compressed within the available 18-19 mbits/second of a conventional 6 MHz broadcast television channel.
  • the present invention provides such enhancements.
  • the invention provides a number of enhancements to handle a variety of video quality and compression problems.
  • the following describes a number of such enhancements, most of which are preferably embodied as a set of tools which can be applied to the tasks of enhancing images and compressing such images.
  • the tools can be combined by a content developer in various ways, as desired, to optimize the visual quality and compression efficiency of a compressed data stream, particularly a layered compressed data stream.
  • Such tools include improved de-interlacing and noise reduction enhancements, including motion analysis.
  • FIG. 1A is a block diagram of an odd-field de-interlacer.
  • FIG. 1B is a block diagram of an even-field de-interlacer.
  • FIG. 2 is a block diagram of a frame de-interlacer using three de-interlaced fields.
  • FIG. 3 is a block diagram of a threshold test.
  • FIG. 4 is a block diagram of a preferred combination of median filters.
  • FIG. 5 is a diagram of the relative shape, amplitudes, and lobe polarity of a preferred downsizing filter.
  • FIGS. 6A and 6B are diagrams of the relative shape, amplitudes, and lobe polarity of a pair of preferred upsizing filters for upsizing by a factor of 2.
  • a number of enhancements may be made to handle a variety of video quality and compression problems.
  • the following describes a number of such enhancements, most of which are preferably embodied as a set of tools which can be applied to the tasks of enhancing images and compressing such images.
  • the tools can be combined by a content developer in various ways, as desired, to optimize the visual quality and compression efficiency of a compressed data stream, particularly a layered compressed data stream.
  • noise reduction must similarly match the needs of being an input to compression algorithms, rather than just reducing noise appearance.
  • the goal is generally to reproduce, upon decompression, no more noise than the original camera or film-grain noise. Equal noise is generally considered acceptable, after compression/decompression. Reduced noise, with equivalent sharpness and clarity with the original, is a bonus. The noise reduction described below achieves these goals.
  • noise reduction can be the difference between a good looking compressed/decompressed image vs. one which is unwatchably noisy.
  • the compression process greatly amplifies noise which is above some threshold of acceptability to the compressor.
  • the use of noise-reduction pre-processing to keep noise below this threshold may be required for acceptable good quality results.
  • de-graining and/or noise-reducing filtering before layered or non-layered encoding improves the ability of the compression system to perform. While de-graining or noise-reduction is most effective on grainy or noisy images prior to compression, either process may be helpful when used in moderation even on relatively low noise or low grain pictures. Any of several known de-graining or noise-reduction algorithms may be applied. Examples are “coring”, simple neighbor median filters, and softening filters.
  • noise-reduction is needed is determined by how noisy the original images are.
  • the interlace itself is a form of noise, which usually will require additional noise reduction filtering, in addition to the complex de-interlacing process described below.
  • noise processing is useful in layered and non-layered compression when noise is present above a certain level.
  • film grain noise is caused by silver grains which couple to yellow, cyan, and magenta film dyes. Yellow affects both red and green, cyan affects both blue and green, and magenta affects both red and blue. Red is formed where yellow and magenta dye crystals overlap. Similarly green is the overlap of yellow and cyan, and blue is the overlap of magenta and cyan.
  • noise between colors is partially correlated through the dyes and grains between pairs of colors. Further, when multiple grains overlap in all three colors, as they do in a print dark regions of the image or on a negative in light regions of the image (dark on the negative), additional color combinations occur. This correlation between the colors can be utilized in film-grain noise reduction, but is a complex process. Further, many different film types are used, and each type has different grain sizes, shapes, and statistical distributions.
  • red, green, and blue noise is uncorrelated. In this case, it is best to process the red, green, and blue records independently. Thus, red noise is reduced with self-red processing independently of green noise and blue noise; the same approach applies to green and blue noise.
  • noise processing is best matched to the characteristics of the noise source itself.
  • the noise may differ in characteristics over different portions of the image.
  • generic noise processing may be the only option, if noise processing is needed.
  • Re-graining and/or re-noising are relatively easy effects to add in the decoder using any of several known algorithms. For example, this can be accomplished by the addition of low pass filtered random noise of suitable amplitude.
  • the preferred compression method for interlaced source which is ultimately intended for non-interlaced display includes a step to de-interlace the interlaced source before the compression steps.
  • De-interlacing a signal after decoding in the receiver, where the signal has been compressed in the interlaced mode is both more costly and less efficient than de-interlacing prior to compression, and then sending a non-interlaced compressed signal.
  • the non-interlaced compressed signal can be either layered or non-layered (i.e., a conventional single layer compression).
  • a field-de-interlacer is used as the first step in the overall process to create field-frames.
  • each field is de-interlaced, creating a synthesized frame where the total number of lines in the frame is derived from the half number of lines in a field.
  • an interlaced 1080 line image will have 540 lines per even and odd field, each field representing 1/60th of a second.
  • the even and odd fields of 540 lines will be interlaced to create 1080 lines for each frame, which represents 1/30th of a second.
  • the de-interlacer copies each scanline without modification from a specified field (e.g., the odd fields) to a buffer that will hold some of the de-interlaced result.
  • the remaining intermediate scanlines (in this example, the even scanlines) for the frame are synthesized by adding half of the field line above and half of the field line below each newly stored line.
  • the pixel values of line 2 for a frame would each comprise 1 ⁇ 2 of the summed corresponding pixel values from each of line 1 and line 3 .
  • the generation of intermediate synthesized scanlines may be done on the fly, or may be computed after all of the scanlines from a field are stored in a buffer. The same process is repeated for the next field, although the field types (i.e., even, odd) will be reversed.
  • FIG. 1A is a block diagram of an odd-field de-interlacer, showing that the odd lines from an odd field 10 are simply copied to a de-interlaced odd field 12 , while the even lines are created by averaging adjacent odd lines from the original odd field together to form the even lines of the de-interlaced odd field 12 .
  • FIG. 1B is a block diagram of an even-field de-interlacer, showing that the even lines from an even field 14 are simply copied to a de-interlaced even field 16 , while the odd lines are created by averaging adjacent even lines from the original even field together to form the odd lines of the de-interlaced even field 16 . Note that this case corresponds to “top field first”; “bottom field first” could also be considered the “even” field.
  • FIG. 2 is a block diagram showing how the pixels of each output frame are composed of 25% of the corresponding pixels from a previous de-interlaced field (field-frame) 22 , 50% of the corresponding pixels from a current field-frame 24 , and 25% of the corresponding pixels from the next field-frame 26 .
  • the new de-interlaced frame then contains much fewer interlace difference artifacts between frames than do the three field-frames of which it is composed.
  • This de-interlacing process is very beneficial as input to compression, either single layer (unlayered) or layered. It is also beneficial just as a treatment for interlaced video for presentation, viewing, or making still frames, independent of use with compression.
  • the picture from the de-interlacing process appears “clearer” than the presentation of the interlace directly, or of the de-interlaced fields.
  • a threshold test may be applied which compares the result of the [0.25, 0.5, 0.25] temporal filter against the corresponding pixel values of only the middle field-frame. If a middle field-frame pixel value differs more than a specified threshold amount from the value of the corresponding pixel from the three-field-frame temporal filter, then only the middle field-frame pixel value is used.
  • a pixel from the three-field-frame temporal filter is selected where it differs less than the threshold amount from the corresponding pixel of the single de-interlaced middle field-frame, and the middle field-frame pixel value is used when there is more difference than the threshold.
  • This allows fast motion to be tracked at the field rate, and smoother parts of the image to be filtered and smoothed by the three-field-frame temporal filter.
  • This combination has proven an effective, if not optimal, input to compression. It is also very effective for processing for direct viewing to de-interlace image material (also called line doubling in conjunction with display).
  • the preferred embodiment for such threshold determinations uses the following equations for corresponding RGB color values from the middle (single) de-interlaced field-frame image and the three-field-frame de-interlaced image:
  • Rdiff R_single_field_de-interlaced minus R_three_field_de-interlaced
  • Gdiff G_single_field_de-interlaced minus G_three_field_de-interlaced
  • Bdiff B_single_field_de-interlaced minus B_three_field_de-interlaced
  • ThresholdingValue abs(Rdiff+Gdiff+Bdiff)+abs (Rdiff)+abs(Gdiff)+abs(Bdiff)
  • the ThresholdingValue is then compared to a threshold setting. Typical threshold settings are in the range of 0.1 to 0.3, with 0.2 being most common.
  • FIG. 3 shows a block diagram of this threshold test.
  • the PROCESSING block 30 multiplies the inputs by [0.25, 0.5, 0.25] and sums the results.
  • the SELECTION CONTROL block 32 compares the output 36 of the PROCESSING block 30 with Input B 34 using the above equations for Rdiff, Gdiff, Bdiff, and ThresholdingValue.
  • the switch selects the PROCESSING output 36 if the ThresholdingValue is less than the threshold, otherwise the switch selects Input B 34 , the middle value, for the output 38 .
  • smooth-filtering the three-field-frame and single-field-frame de-interlaced pictures can be used before comparing and thresholding them.
  • This smooth filtering can be accomplished simply by down filtering (e.g., down filtering by two), and then up filtering (e.g., using a gaussian up-filter by two).
  • This “down-up” smoothed filter can be applied to both the single-field-frame de-interlaced picture and the three-field-frame de-interlaced picture.
  • the smoothed single-field-frame and three-field-frame pictures can then be compared to compute a ThresholdingValue and then thresholded to determine which picture will source each final output pixel.
  • the threshold test is used as a switch to select between the single-field-frame de-interlaced picture and the three-field-frame temporal filter combination of single-field-frame de-interlaced pictures. This selection then results in an image where the pixels are from the three-field-frame de-interlacer in those areas where that image differs in small amounts (i.e., below the threshold) from the single field-frame image, and where the pixels are from the single field-frame image in those areas where the three-field-frame differed more than then the threshold amount from the single-field-frame de-interlaced pixels (after smoothing).
  • This technique has proven effective in preserving single-field fast motion details (by switching to the single-field-frame de-interlaced pixels), while smoothing large portions of the image (by switching to the three-field-frame de-interlaced temporal filter combination).
  • a typical blending is to create new frame by adding 33.33% (1 ⁇ 3) of a single middle field-frame to 66.67% (2 ⁇ 3) of the corresponding three-field-frame smoothed image. This can be done before or after threshold switching, since the result is the same either way, only affecting the smoothed three-field-frame picture.
  • V is the video value and L is linear light luminance.
  • the variations adjust the threshold (0.018051) a little, the factor (4.5) a little (e.g. 4.0), and the exponent (0.45) a little (e.g., 0.4).
  • the fundamental formula remains the same.
  • a matrix operation such as a RGB to/from YUV conversion, implies linear values.
  • linear vs. logarithmic vs. video issue impacts filtering.
  • small signal excursions e.g. 10% or less
  • small signal excursions e.g. 10% or less
  • a linear filter is much more effective, and produces much better image quality. Accordingly, if large excursions are to be optimally coded, transformed, or otherwise processed, it would be desirable to first convert the non-linear signal to a linear one in order to be able to apply a linear filter.
  • De-interlacing is therefore much better when each filter and summation step utilizes conversions to linear values prior to filtering or summing. This is due to the large signal excursions inherent in interlaced signals at small details of the image.
  • the image signals are converted back to the non-linear video digital representation.
  • the three-field-frame weighting e.g., [0.25, 0.5, 0.25] or [0.1667, 0.6666, 0.1667]
  • Other filtering and weighted sums of partial terms in noise and de-interlace filtering should also be converted to linear form for computation. Which operations warrant linear processing is determined by signal excursion, and the type of filtering.
  • Image sharpening can be appropriately computed in video or logarithmic non-linear representations, since it is self-proportional.
  • matrix processing, spatial filtering, weighted sums, and de-interlace processing should be computed using linearized digital values.
  • the single field-frame de-interlacer described above computes missing alternate lines by averaging the line above and below each actual line. This average is much more correct numerically and visually if this average is done linearly. Thus, instead of summing 0.5 times the line above plus 0.5 times the line below, the digital values are linearized first, then averaged, and then reconverted back into the non-linear video representation.
  • the most useful filter is the median filter.
  • a three element median filter just ranks the three entries, via a simple sort, and picks the middle one. For example, an X (horizontal) median filter looks at the red value (or green or blue) of three adjacent horizontal pixels, and picks the one with the middle-most value. If two are the same, that value is selected. Similarly, a Y (vertical) filter looks in the scanlines above and below the current pixel, and again picks the middle value.
  • each new pixel is the 50% equal average of the X and Y medians for the corresponding pixel from a source image.
  • a temporal median Another beneficial source of noise reduction is information from the previous and subsequent frame (i.e., a temporal median).
  • motion analysis provides the best match for moving regions. However, it is compute intensive. If a region of the image is not moving, or is moving slowly, the red values (and green and blue) from a current pixel can be median filtered with the red value at that same pixel location in the previous and subsequent frames. However, odd artifacts may occur if significant motion is present and such a temporal filter is used. Thus, it is preferred that a threshold be taken first, to determine whether such a median would differ more than a selected amount from the value of a current pixel.
  • the ThresholdingValue is then compared to a threshold setting. Typical threshold settings are in the range 0.1 to 0.3, with 0.2 being typical. Above the threshold, the current value is kept. Below the threshold, the temporal median is used.
  • the block diagram of FIG. 3 also applies to this threshold test.
  • the PROCESSING block 30 is a temporal median filter and the inputs are three successive frames.
  • the SELECTION CONTROL block 32 compares the output 36 of the PROCESSING block 30 with Input B 34 using the above equations for Rdiff, Gdiff, Bdiff, and ThresholdingValue. The switch selects the PROCESSING output 36 if the ThresholdingValue is less than the threshold, otherwise the switch selects Input B 34 , the middle value, for the output 38 .
  • An additional median type is a median taken between the X, Y, and temporal medians.
  • Another median type can take the temporal median, and then take the equal average of the X and Y medians from it.
  • Each type of median can cause problems.
  • X and Y medians smear and blur an image, so that it looks “greasy”.
  • Temporal medians cause smearing of motion over time. Since each median can result in problems, yet each median's properties are different (and, in some sense, “orthogonal”), it has been determined experimentally that the best results come by combining a variety of medians.
  • FIG. 4 shows a preferred combination of medians is a linear weighted sum (see the discussion above on linear video processing) of five terms to determine the value for each pixel of a current image:
  • This set of time medians does a reasonable job of reducing the noise in the image without making it appear “greasy” or blurred, causing temporal smearing of moving objects, or losing detail.
  • Another useful weighting of these five terms is 35%, 20%, 22.5%, 10%, and 12.5%, respectively.
  • de-interlacing and noise reduction can also be improved by use of motion analysis. Adding the pixels at the same location in three fields or three frames is valid for stationary objects. However, for moving objects, if temporal averaging/smoothing is desired, it is often more optimal to attempt to analyze prevailing motion over a small group of pixels. For example, an n ⁇ n block of pixels (e.g., 2 ⁇ 2, 3 ⁇ 3, 4 ⁇ 4, 6 ⁇ 6, or 8 ⁇ 8) can be used to search in previous and subsequent fields or frames to attempt to find a match (in the same way MPEG-2 motion vectors are found by matching 16 ⁇ 16 macroblocks).
  • n ⁇ n block of pixels e.g., 2 ⁇ 2, 3 ⁇ 3, 4 ⁇ 4, 6 ⁇ 6, or 8 ⁇ 8
  • a “trajectory” and “moving mini-picture” can be determined.
  • the motion analysis preferably is performed by comparison of an n ⁇ n block in the current thresholded de-interlaced image with all nearby blocks in the previous and subsequent one or more frames.
  • the comparison may be the absolute value of differences in luminance or RGB over the n ⁇ n block.
  • One frame is sufficient forward and backward if the motion vectors are nearly equal and opposite. However, if the motion vectors are not nearly equal and opposite, then an additional one or two frames forward and backward can help determine the actual trajectory.
  • different de-interlacing treatments may be useful in helping determine the “best guess” motion vectors going forward and back.
  • One de-interlacing treatment can be to use only individual de-interlaced fields, although this is heavily prone to aliasing and artifacts on small moving details.
  • Another de-interlacing technique is to use only the three-field-frame smooth de-interlacing, without thresholding, having weightings [0.25, 0.5, 0.25], as described above. Although details are smoothed and sometimes lost, the trajectory may often be
  • a “smoothed n ⁇ n block” can be created by temporally filtering using the motion-vector-offset pixels from the one (or more) previous and subsequent frames.
  • a typical filter might again be [0.25, 0.5, 0.25] or [0.1667, 0.6666, 0.1667] for three frames, and possibly [0.1, 0.2, 0.4, 0.2, 0.1] for two frames back and forward.
  • Other filters, with less central weight, are also useful, especially with smaller block sizes (such as 2 ⁇ 2, 3 ⁇ 3, and 4 ⁇ 4).
  • Reliability of the match between frames is indicated by the absolute difference value. Large minimum absolute differences can be used to select more center weight in the filter. Lower values of absolute differences can suggest a good match, and can be used to select less center weight to more evenly distribute the average over a span of several frames of motion-compensated blocks.
  • filter weights can be applied to: individual de-interlaced motion-compensated field-frames; thresholded three-field-frame de-interlaced pictures, described above; and non-thresholded three-field-frame de-interlaced images, with a [0.25, 0.5, 0.25] weighting, also as described above.
  • the best filter weights usually come from applying the motion-compensated block linear filtering to the thresholded three-field-frame result described above. This is because the thresholded three-field-frame image is both the smoothest (in terms of removing aliasing in smooth areas), as well as the most motion-responsive (in terms of defaulting to a single de-interlaced field-frame above the threshold).
  • the motion vectors from motion analysis can be used as the inputs to multi-frame or multi-de-interlaced-field-frame or single-de-interlaced field-frame filters, or combinations thereof.
  • the thresholded multi-field-frame de-interlaced images form the best filter input in most cases.
  • motion analysis is computationally expensive for a large search region, when fast motion might be found (such as ⁇ 32 pixels). Accordingly, it may be best to augment the speed by using special-purpose hardware or a digital signal processor assisted computer.
  • motion vectors Once motion vectors are found, together with their absolute difference measure of accuracy, they can be utilized for the complex process of attempting frame rate conversion.
  • occlusion issues objects obscuring or revealing others
  • Occlusion can also involve temporal aliasing, as can normal image temporal undersampling and its beat with natural image frequencies (such as the “backward wagon wheel” effect in movies).
  • temporal aliasing can also involve temporal aliasing, as can normal image temporal undersampling and its beat with natural image frequencies (such as the “backward wagon wheel” effect in movies).
  • De-interlacing is a simple form of the same problem. Just as with frame-rate-conversion, the task of de-interlacing is theoretically impossible to perform perfectly. This is especially due to the temporal undersampling (closed shutter), and an inappropriate temporal sample filter (i.e., a box filter). However, even with correct samples, issues such as occlusion and interlace aliasing further ensure the theoretical impossibility of correct results. The cases where this is visible are mitigated by the depth of the tools, as described here, which are applied to the problem. Pathological cases will always exist in real image sequences. The goal can only be to reduce the frequency and level of impairment when these sequences are encountered. However, in many cases, the de-interlacing process can be acceptably fully automated, and can run unassisted in real-time. Even so, there are many parameters which can often benefit from manual adjustment.
  • a filter which creates smoothing can be easily made, as with the threshold for de-interlacing, by down-filtering with a normal filter (e.g., truncated sinc filter) and then up-filtering with a gaussian filter. The result will be smoothed because it is devoid of high frequency picture detail. When such a term is added, it typically must be in very small amounts, such as 5% to 10%, in order to provide a small amount of noise reduction. In larger amounts, the blurring effect generally becomes quite visible.
  • a normal filter e.g., truncated sinc filter
  • the filter parameters for the median filtering described above for an original image should be matched to the noise characteristics of the film grain or image sensor that captured the image. After this median filtered image is down-filtered to generate an input to the base layer compression process, it still contains a small amount of noise. This noise may be further reduced by a combination of another X-Y median filters (equally averaging the X and Y medians), plus a very small amount of the high frequency smoothing filter.
  • a preferred filter weighting of these three terms, applied to each pixel of the base layer is:
  • This small amount of additional filtering in the base layer provides a small additional amount of noise reduction and improved stability, resulting in better MPEG encoding and limiting the amount of noise added by such encoding.
  • FIG. 5 is a diagram of the relative shape, amplitudes, and lobe polarity of a preferred downsizing filter.
  • the down filter essentially is a center-weighted function which has been truncated to a center positive lobe 500, a symmetric pair of adjacent (bracketing) small negative lobes 504, and a symmetric pair of adjacent (bracketing) very small outer positive lobes 504.
  • the absolute amplitude of the lobes 500, 502, 504 may be adjusted as desired, so long as the relative polarity and amplitude inequality relationships shown in FIG. 5 are maintained.
  • the preferred downsizing filter When creating a base layer original (as input to the base layer compression) from a low-noise high resolution original input, the preferred downsizing filter has first negative lobes which are of a normal sinc function amplitude. For clean and for high resolution input images, this normal truncated sinc function works well. For lower resolutions (e.g., 1280 ⁇ 720, 1024 ⁇ 768, or 1536 ⁇ 768), and for noisier input pictures, a reduced first negative lobe amplitude in the filters is more optimal. A suitable amplitude in such cases is about half the truncated sinc function negative lobe amplitude.
  • the small first positive lobes outside of the first negative lobes are also reduced to lower amplitude, typically to 1 ⁇ 2 to 2 ⁇ 3 of the normal sinc function amplitude.
  • the affect of reducing the first negative lobes is the main issue, since the small outside positive lobes do not contribute to picture noise. Further samples outside the first positive lobes preferably are truncated to minimize ringing and other potential artifacts.
  • FIGS. 6A and 6B are diagrams of the relative shape, amplitudes, and lobe polarity of a pair of preferred upsizing filters for upsizing by a factor of 2.
  • a central positive lobe 600, 600′ is bracketed by a pair of small negative lobes 602, 602′.
  • An asymmetrically placed positive lobe 604, 604′ is also required.
  • These paired upfilters could also be considered to be truncated sinc filters centered on the newly created samples. For example, for a factor of two upfilter, two new samples will be created for each original sample.
  • the small adjacent negative lobes 602, 602′ have less negative amplitude than is used in the corresponding downsizing filter (FIG. 5), or than would be used in an optimal (sinc-based) upsizing filter for normal images. This is because the images being upsized are decompressed, and the compression process changes the spectral distribution. Thus, more modest negative lobes, and no additional positive lobes beyond the middle ones 600, 600′, work better for upsizing a decompressed base layer.
  • this upsizing filter preferably is used for the base layer in both the encoder and the decoder.
  • the signal path which expands the original uncompressed base layer input image uses a gaussian upfilter rather than the upfilter described above.
  • a gaussian upfilter is used for the “high octave” of picture detail, which is determined by subtracting the expanded original base-resolution input image (without using compression) from the original picture.
  • no negative lobes are used for this particular upfiltered expansion.
  • this high octave difference signal path is typically weighted with 0.25 (or 25%) and added to the expanded decompressed base layer (using the other upfilter described above) as input to the enhancement layer compression process.
  • weights 10%, 15%, 20%, 30%, and 35% are useful for particular images when using MPEG-2.
  • Other weights may also prove useful.
  • filter weights 4-8% may be optimal when used in conjunction with other improvements described below. Accordingly, this weighting should be regarded as an adjustable parameter, depending upon the encoding system, the scenes being encoded/compressed, the particular camera (or film) being used, and the image resolution.
  • MPEG-4 reference filters have been implemented for shifting macroblocks when finding the best motion vector match, and then using the matched region for motion compensation.
  • MPEG-4 video coding like MPEG-2, supports 1 ⁇ 2 pixel resolution of motion vectors for macroblocks. Unlike MPEG-2, MPEG-4 also supports 1 ⁇ 4 pixel accuracy.
  • the filters used are sub-optimal. In MPEG-2, the half-way point between pixels is just the average of the two neighbors, which is a sub-optimal box filter. In MPEG-4, this filter is used for 1 ⁇ 2 pixel resolution.
  • the chrominance channels do not use any sub-pixel resolution in the motion compensation step under MPEG-4. Since the luminance channel (Y) has resolution to the 1 ⁇ 2 or 1 ⁇ 4 pixel, the half-resolution chrominance U and V channels should be sampled using filters to 1 ⁇ 4 pixel resolution, corresponding to 1 ⁇ 2 pixel in luminance. When 1 ⁇ 4 pixel resolution is selected for luminance, then 1 ⁇ 8 pixel resolution should be used for U and V chrominance.
  • effects of filtering are significantly improved by using a negative lobe truncated sinc function for filtering the 1 ⁇ 8-pixel points for U and V chrominance when using 1 ⁇ 4 pixel luminance resolution, and by using 1 ⁇ 4 pixel resolution filters with similar negative lobe filters when using 1 ⁇ 2 pixel luminance resolution.
  • These filters may be applied to video images under MPEG-1, MPEG-2, MPEG-4 or any other appropriate motion-compensated block-based image coding system.
  • the invention may be implemented in hardware or software, or a combination of both. However, preferably, the invention is implemented in computer programs executing on one or more programmable computers each comprising at least a processor, a data storage system (including volatile and non-volatile memory and/or storage elements), an input device, and an output device. Program code is applied to input data to perform the functions described herein and generate output information. The output information is applied to one or more output devices, in known fashion.
  • Each such program may be implemented in any desired computer language (including machine, assembly, or high level procedural, logical, or object oriented programming languages) to communicate with a computer system.
  • the language may be a compiled or interpreted language.
  • Each such computer program is preferably stored on a storage media or device (e.g., ROM, CD-ROM, or magnetic or optical media) readable by a general or special purpose programmable computer system, for configuring and operating the computer when the storage media or device is read by the computer system to perform the procedures described herein.
  • a storage media or device e.g., ROM, CD-ROM, or magnetic or optical media
  • the inventive system may also be considered to be implemented as a computer-readable storage medium, configured with a computer program, where the storage medium so configured causes a computer system to operate in a specific and predefined manner to perform the functions described herein.

Abstract

A technique for improving image compression by pre-processing the image frames. In particular, methods for de-interlacing and noise reduction using combinations of median filters, applied both spatially and temporally, with and without motion analysis, are described.

Description

CROSS-REFERENCE TO RELATED APPLICATIONSCROSS REFERENCE TO RELATED APPLICATIONS
NOTICE: More than one reissue application has been filed for the reissue of U.S. Pat. No. 7,366,242, including an initial reissue application with Ser. No. 12/618,202, now U.S. Pat. No. Re. 43,360, filed Nov. 13, 2009. The present reissue application is a continuation reissue application of the initial reissue application.
This application is a divisional of and claims priority to U.S. application Ser. No. 09/545,233 filed on Apr. 7, 2000 now U.S. Pat. No. 6,728,317 (which is incorporated herein in its entirety), which was a continuation-in-part application of U.S. application Ser. No. 09/442,595 filed on Nov. 17, 1999 now abandoned, which was a continuation of U.S. application Ser. No. 09/217,151 filed on Dec. 21, 1998 (now U.S. Pat. No. 5,988,863, issued Nov. 23, 1999), which was a continuation of U.S. application Ser. No. 08/594,815 filed Jan. 30, 1996 (now U.S. Pat. No. 5,852,565, issued Dec. 22, 1998).
TECHNICAL FIELD
This invention relates to electronic communication systems, and more particularly to an advanced electronic television system having enhanced compression, filtering, and display characteristics.
BACKGROUND
The United States presently uses the NTSC standard for television transmissions. However, proposals have been made to replace the NTSC standard with an Advanced Television standard. For example, it has been proposed that the U.S. adopt digital standard-definition and advanced television formats at rates of 24 Hz, 30 Hz, 60 Hz, and 60 Hz interlaced. It is apparent that these rates are intended to continue (and thus be compatible with) the existing NTSC television display rate of 60 Hz (or 59.94 Hz). It is also apparent that “3-2 pulldown” is intended for display on 60 Hz displays when presenting movies, which have a temporal rate of 24 frames per second (fps). However, while the above proposal provides a menu of possible formats from which to select, each format only encodes and decodes a single resolution and frame rate. Because the display or motion rates of these formats are not integrally related to each other, conversion from one to another is difficult.
Further, this proposal does not provide a crucial capability of compatibility with computer displays. These proposed image motion rates are based upon historical rates which date back to the early part of this century. If a “clean-slate” were to be made, it is unlikely that these rates would be chosen. In the computer industry, where displays could utilize any rate over the last decade, rates in the 70 to 80 Hz range have proven optimal, with 72 and 75 Hz being the most common rates. Unfortunately, the proposed rates of 30 and 60 Hz lack useful interoperability with 72 or 75 Hz, resulting in degraded temporal performance.
In addition, it is being suggested by some that interlace is required, due to a claimed need to have about 1000 lines of resolution at high frame rates, but based upon the notion that such images cannot be compressed within the available 18-19 mbits/second of a conventional 6 MHz broadcast television channel.
It would be much more desirable if a single signal format were to be adopted, containing within it all of the desired standard and high definition resolutions. However, to do so within the bandwidth constraints of a conventional 6 MHz broadcast television channel requires compression and “scalability” of both frame rate (temporal) and resolution (spatial). One method specifically intended to provide for such scalability is the MPEG-2 standard. Unfortunately, the temporal and spatial scalability features specified within the MPEG-2 standard (and newer standards, like MPEG-4) are not sufficiently efficient to accommodate the needs of advanced television for the U.S. Thus, the proposal for advanced television for the U.S. is based upon the premise that temporal (frame rate) and spatial (resolution) layering are inefficient, and therefore discrete formats are necessary.
Further, it would be desirable to provide enhancements to resolution, image clarity, coding efficiency, and video production efficiency. The present invention provides such enhancements.
SUMMARY
The invention provides a number of enhancements to handle a variety of video quality and compression problems. The following describes a number of such enhancements, most of which are preferably embodied as a set of tools which can be applied to the tasks of enhancing images and compressing such images. The tools can be combined by a content developer in various ways, as desired, to optimize the visual quality and compression efficiency of a compressed data stream, particularly a layered compressed data stream.
Such tools include improved de-interlacing and noise reduction enhancements, including motion analysis.
The details of one or more embodiments of the invention are set forth in the accompanying drawings and the description below. Other features, objects, and advantages of the invention will be apparent from the description and drawings, and from the claims.
DESCRIPTION OF DRAWINGS
FIG. 1A is a block diagram of an odd-field de-interlacer.
FIG. 1B is a block diagram of an even-field de-interlacer.
FIG. 2 is a block diagram of a frame de-interlacer using three de-interlaced fields.
FIG. 3 is a block diagram of a threshold test.
FIG. 4 is a block diagram of a preferred combination of median filters.
FIG. 5 is a diagram of the relative shape, amplitudes, and lobe polarity of a preferred downsizing filter.
FIGS. 6A and 6B are diagrams of the relative shape, amplitudes, and lobe polarity of a pair of preferred upsizing filters for upsizing by a factor of 2.
Like reference symbols in the various drawings indicate like elements.
DETAILED DESCRIPTION
Throughout this description, the preferred embodiment and examples shown should be considered as exemplars, rather than as limitations on the invention.
A number of enhancements may be made to handle a variety of video quality and compression problems. The following describes a number of such enhancements, most of which are preferably embodied as a set of tools which can be applied to the tasks of enhancing images and compressing such images. The tools can be combined by a content developer in various ways, as desired, to optimize the visual quality and compression efficiency of a compressed data stream, particularly a layered compressed data stream.
De-Interlacing and Noise Reduction Enhancements
Overview
Experimentation has shown that many de-interlacing algorithms and devices depend upon the human eye to integrate fields to create an acceptable result. However, since compression algorithms are not a human eye, any integration of de-interlaced fields should take into account the characteristics of such algorithms. Without such careful de-interlaced integration, the compression process will create high levels of noise artifacts, both wasting bits (hindering compression) as well as making the image look noisy and busy with artifacts. This distinction between de-interlacing for viewing (such as with line-doublers and line-quadruplers) vs. de-interlacing as input to compression, has lead to the techniques described below. In particular, the de-interlacing techniques described below are useful as input to single-layer non-interlaced MPEG-like, as well as to the layered MPEG-like compression.
Further, noise reduction must similarly match the needs of being an input to compression algorithms, rather than just reducing noise appearance. The goal is generally to reproduce, upon decompression, no more noise than the original camera or film-grain noise. Equal noise is generally considered acceptable, after compression/decompression. Reduced noise, with equivalent sharpness and clarity with the original, is a bonus. The noise reduction described below achieves these goals.
Further, for very noisy shots, such as from high speed film or with high camera sensitivity settings, usually in low light, noise reduction can be the difference between a good looking compressed/decompressed image vs. one which is unwatchably noisy. The compression process greatly amplifies noise which is above some threshold of acceptability to the compressor. Thus, the use of noise-reduction pre-processing to keep noise below this threshold may be required for acceptable good quality results.
De-Graining and Noise-Reducing Filters
It has been found through experimentation that applying de-graining and/or noise-reducing filtering before layered or non-layered encoding improves the ability of the compression system to perform. While de-graining or noise-reduction is most effective on grainy or noisy images prior to compression, either process may be helpful when used in moderation even on relatively low noise or low grain pictures. Any of several known de-graining or noise-reduction algorithms may be applied. Examples are “coring”, simple neighbor median filters, and softening filters.
Whether noise-reduction is needed is determined by how noisy the original images are. For interlaced original images, the interlace itself is a form of noise, which usually will require additional noise reduction filtering, in addition to the complex de-interlacing process described below. For progressive scan (non-interlaced) camera or film images, noise processing is useful in layered and non-layered compression when noise is present above a certain level.
There are different types of noise. For example, video transfers from film include film grain noise. Film grain noise is caused by silver grains which couple to yellow, cyan, and magenta film dyes. Yellow affects both red and green, cyan affects both blue and green, and magenta affects both red and blue. Red is formed where yellow and magenta dye crystals overlap. Similarly green is the overlap of yellow and cyan, and blue is the overlap of magenta and cyan. Thus, noise between colors is partially correlated through the dyes and grains between pairs of colors. Further, when multiple grains overlap in all three colors, as they do in a print dark regions of the image or on a negative in light regions of the image (dark on the negative), additional color combinations occur. This correlation between the colors can be utilized in film-grain noise reduction, but is a complex process. Further, many different film types are used, and each type has different grain sizes, shapes, and statistical distributions.
For video images created by CCD-sensor and other (e.g., tube) sensor cameras, the red, green, and blue noise is uncorrelated. In this case, it is best to process the red, green, and blue records independently. Thus, red noise is reduced with self-red processing independently of green noise and blue noise; the same approach applies to green and blue noise.
Thus, noise processing is best matched to the characteristics of the noise source itself. In the case of a composite image (from multiple sources), the noise may differ in characteristics over different portions of the image. In this situation, generic noise processing may be the only option, if noise processing is needed.
It has also been found useful in some cases to perform a “re-graining” or “re-noising” process after decoding a compressed layered data stream, as a creative effect, since some de-grained or de-noised images may be “too clean” or “too sterile” in appearance. Re-graining and/or re-noising are relatively easy effects to add in the decoder using any of several known algorithms. For example, this can be accomplished by the addition of low pass filtered random noise of suitable amplitude.
De-Interlacing Before Compression
As mentioned above, the preferred compression method for interlaced source which is ultimately intended for non-interlaced display includes a step to de-interlace the interlaced source before the compression steps. De-interlacing a signal after decoding in the receiver, where the signal has been compressed in the interlaced mode, is both more costly and less efficient than de-interlacing prior to compression, and then sending a non-interlaced compressed signal. The non-interlaced compressed signal can be either layered or non-layered (i.e., a conventional single layer compression).
Experimentation has shown that filtering a single field of an interlaced source, and using that field as if it were a non-interlaced full frame, gives poor and noisy compression results. Thus, using a single-field de-interlacer prior to compression is not a good approach. Instead, experimentation has shown that a three-field-frame de-interlacer process using field synthesized frames (“field-frames”), with weights of [0.25, 0.5, 0.25] for the previous, current, and next field-frames, respectively, provides a good input for compression. Combining three field-frames may be performed using other weights (although these weights are optimal) to create a de-interlaced input to a compression process.
In the preferred de-interlacing system, a field-de-interlacer is used as the first step in the overall process to create field-frames. In particular, each field is de-interlaced, creating a synthesized frame where the total number of lines in the frame is derived from the half number of lines in a field. Thus, for example, an interlaced 1080 line image will have 540 lines per even and odd field, each field representing 1/60th of a second. Normally, the even and odd fields of 540 lines will be interlaced to create 1080 lines for each frame, which represents 1/30th of a second. However, in the preferred embodiment, the de-interlacer copies each scanline without modification from a specified field (e.g., the odd fields) to a buffer that will hold some of the de-interlaced result. The remaining intermediate scanlines (in this example, the even scanlines) for the frame are synthesized by adding half of the field line above and half of the field line below each newly stored line. For example, the pixel values of line 2 for a frame would each comprise ½ of the summed corresponding pixel values from each of line 1 and line 3. The generation of intermediate synthesized scanlines may be done on the fly, or may be computed after all of the scanlines from a field are stored in a buffer. The same process is repeated for the next field, although the field types (i.e., even, odd) will be reversed.
FIG. 1A is a block diagram of an odd-field de-interlacer, showing that the odd lines from an odd field 10 are simply copied to a de-interlaced odd field 12, while the even lines are created by averaging adjacent odd lines from the original odd field together to form the even lines of the de-interlaced odd field 12. Similarly, FIG. 1B is a block diagram of an even-field de-interlacer, showing that the even lines from an even field 14 are simply copied to a de-interlaced even field 16, while the odd lines are created by averaging adjacent even lines from the original even field together to form the odd lines of the de-interlaced even field 16. Note that this case corresponds to “top field first”; “bottom field first” could also be considered the “even” field.
As a next step, a sequence of these de-interlaced fields is then used as input to a three-field-frame de-interlacer to create a final de-interlaced frame. FIG. 2 is a block diagram showing how the pixels of each output frame are composed of 25% of the corresponding pixels from a previous de-interlaced field (field-frame) 22, 50% of the corresponding pixels from a current field-frame 24, and 25% of the corresponding pixels from the next field-frame 26.
The new de-interlaced frame then contains much fewer interlace difference artifacts between frames than do the three field-frames of which it is composed. However, there is a temporal smearing by adding the previous field-frame and next field-frame into a current field-frame. This temporal smearing is usually not objectionable, especially in light of the de-interlacing improvements which result.
This de-interlacing process is very beneficial as input to compression, either single layer (unlayered) or layered. It is also beneficial just as a treatment for interlaced video for presentation, viewing, or making still frames, independent of use with compression. The picture from the de-interlacing process appears “clearer” than the presentation of the interlace directly, or of the de-interlaced fields.
De-Interlace Thresholding
Although the de-interlace three-field sum weightings of [0.25, 0.5, 0.25] discussed above provide a stable image, moving parts of a scene can sometimes become soft or can exhibit aliasing artifacts. To counteract this, a threshold test may be applied which compares the result of the [0.25, 0.5, 0.25] temporal filter against the corresponding pixel values of only the middle field-frame. If a middle field-frame pixel value differs more than a specified threshold amount from the value of the corresponding pixel from the three-field-frame temporal filter, then only the middle field-frame pixel value is used. In this way, a pixel from the three-field-frame temporal filter is selected where it differs less than the threshold amount from the corresponding pixel of the single de-interlaced middle field-frame, and the middle field-frame pixel value is used when there is more difference than the threshold. This allows fast motion to be tracked at the field rate, and smoother parts of the image to be filtered and smoothed by the three-field-frame temporal filter. This combination has proven an effective, if not optimal, input to compression. It is also very effective for processing for direct viewing to de-interlace image material (also called line doubling in conjunction with display).
The preferred embodiment for such threshold determinations uses the following equations for corresponding RGB color values from the middle (single) de-interlaced field-frame image and the three-field-frame de-interlaced image:
Rdiff=R_single_field_de-interlaced minus R_three_field_de-interlaced
Gdiff=G_single_field_de-interlaced minus G_three_field_de-interlaced
Bdiff=B_single_field_de-interlaced minus B_three_field_de-interlaced
ThresholdingValue=abs(Rdiff+Gdiff+Bdiff)+abs (Rdiff)+abs(Gdiff)+abs(Bdiff)
The ThresholdingValue is then compared to a threshold setting. Typical threshold settings are in the range of 0.1 to 0.3, with 0.2 being most common. FIG. 3 shows a block diagram of this threshold test. The PROCESSING block 30 multiplies the inputs by [0.25, 0.5, 0.25] and sums the results. The SELECTION CONTROL block 32 compares the output 36 of the PROCESSING block 30 with Input B 34 using the above equations for Rdiff, Gdiff, Bdiff, and ThresholdingValue. The switch selects the PROCESSING output 36 if the ThresholdingValue is less than the threshold, otherwise the switch selects Input B 34, the middle value, for the output 38.
In order to remove noise from this threshold, smooth-filtering the three-field-frame and single-field-frame de-interlaced pictures can be used before comparing and thresholding them. This smooth filtering can be accomplished simply by down filtering (e.g., down filtering by two), and then up filtering (e.g., using a gaussian up-filter by two). This “down-up” smoothed filter can be applied to both the single-field-frame de-interlaced picture and the three-field-frame de-interlaced picture. The smoothed single-field-frame and three-field-frame pictures can then be compared to compute a ThresholdingValue and then thresholded to determine which picture will source each final output pixel.
In particular, the threshold test is used as a switch to select between the single-field-frame de-interlaced picture and the three-field-frame temporal filter combination of single-field-frame de-interlaced pictures. This selection then results in an image where the pixels are from the three-field-frame de-interlacer in those areas where that image differs in small amounts (i.e., below the threshold) from the single field-frame image, and where the pixels are from the single field-frame image in those areas where the three-field-frame differed more than then the threshold amount from the single-field-frame de-interlaced pixels (after smoothing).
This technique has proven effective in preserving single-field fast motion details (by switching to the single-field-frame de-interlaced pixels), while smoothing large portions of the image (by switching to the three-field-frame de-interlaced temporal filter combination).
In addition to selecting between the single-field-frame and three-field-frame de-interlaced image, it is also often beneficial to add a bit of the single-field-frame image to the three-field-frame de-interlaced picture, to preserve some of the immediacy of the single field pictures over the entire image. This immediacy is balanced against the temporal smoothness of the three-field-frame filter. A typical blending is to create new frame by adding 33.33% (⅓) of a single middle field-frame to 66.67% (⅔) of the corresponding three-field-frame smoothed image. This can be done before or after threshold switching, since the result is the same either way, only affecting the smoothed three-field-frame picture. Note that this is effectively equivalent to using a different proportion of the three field-frames, rather than the original three-field-frame weights of [0.25, 0.5, 0.25]. Computing ⅔ of [0.25, 0.5, 0.25] plus ⅓ of (0,1,0), yields [0.1667, 0.6666, 0.1667] as the temporal filter for the three field-frames. The more heavily weighted center (current) field-frame brings additional immediacy to the result, even in the smoothed areas which fell below the threshold value. This combination has proven effective in balancing temporal smoothness with immediacy in the de-interlacing process for moving parts of a scene.
Use of Linear Filters
Sums, filters, or matrices involving video pictures should take into account the fact that pixel values in video are non-linear signals. For example, the video curve for HDTV can be several variations of coefficients and factors, but a typical formula is the international CCIR XA-11 (now called Rec. 709):
V=1.0993*L0.45−0.0993 for L>0.018051
V=4.5*L for L<=0.018051
where V is the video value and L is linear light luminance.
The variations adjust the threshold (0.018051) a little, the factor (4.5) a little (e.g. 4.0), and the exponent (0.45) a little (e.g., 0.4). The fundamental formula, however, remains the same.
A matrix operation, such as a RGB to/from YUV conversion, implies linear values. The fact that MPEG in general uses the video non-linear values as if they were linear results in leakage between the luminance (Y) and the color values (U, and V). This leakage interferes with compression efficiency. The use of a logarithmic representation, such as is used with film density units, corrects much of this problem. The various types of MPEG encoding are neutral to the non-linear aspects of the signal, although its efficiency is effected due to the use of the matrix conversion RGB to/from YUV.YUV (U=R−Y, V=B−Y) should have Y computed as a linearized sum of 0.59 G, plus 0.29 R, plus 0.12 B (or slight variations on these coefficients). However, U (=R−Y) becomes equivalent to R/Y in logarithmic space, which is orthogonal to luminance. Thus, a shaded orange ball will not vary the U (=R−Y) parameter in a logarithmic representation. The brightness variation will be represented completely in the Luminance parameter, where full detail is provided.
The linear vs. logarithmic vs. video issue impacts filtering. A key point to note is that small signal excursions (e.g. 10% or less) are approximately correct when a non-linear video signal is processed as if it were a linear signal. This is because a piece-wise linear approximation to the smooth video-to-from-linear conversion curve is reasonable. However, for large excursions, a linear filter is much more effective, and produces much better image quality. Accordingly, if large excursions are to be optimally coded, transformed, or otherwise processed, it would be desirable to first convert the non-linear signal to a linear one in order to be able to apply a linear filter.
De-interlacing is therefore much better when each filter and summation step utilizes conversions to linear values prior to filtering or summing. This is due to the large signal excursions inherent in interlaced signals at small details of the image. After filtering, the image signals are converted back to the non-linear video digital representation. Thus, the three-field-frame weighting (e.g., [0.25, 0.5, 0.25] or [0.1667, 0.6666, 0.1667]) should be performed on a linearized video signal. Other filtering and weighted sums of partial terms in noise and de-interlace filtering should also be converted to linear form for computation. Which operations warrant linear processing is determined by signal excursion, and the type of filtering. Image sharpening can be appropriately computed in video or logarithmic non-linear representations, since it is self-proportional. However, matrix processing, spatial filtering, weighted sums, and de-interlace processing should be computed using linearized digital values.
As a simple example, the single field-frame de-interlacer described above computes missing alternate lines by averaging the line above and below each actual line. This average is much more correct numerically and visually if this average is done linearly. Thus, instead of summing 0.5 times the line above plus 0.5 times the line below, the digital values are linearized first, then averaged, and then reconverted back into the non-linear video representation.
Median Filters
In noise processing, the most useful filter is the median filter. A three element median filter just ranks the three entries, via a simple sort, and picks the middle one. For example, an X (horizontal) median filter looks at the red value (or green or blue) of three adjacent horizontal pixels, and picks the one with the middle-most value. If two are the same, that value is selected. Similarly, a Y (vertical) filter looks in the scanlines above and below the current pixel, and again picks the middle value.
It has been experimentally determined that it is useful to average the results from applying both an X and a Y median filter to create a new noise-reducing component picture (i.e., each new pixel is the 50% equal average of the X and Y medians for the corresponding pixel from a source image).
In addition to X and Y (horizontal and vertical) medians, it is also possible to take diagonal and other medians. However, the vertical and horizontal pixel values are most close physically to any particular pixel, and therefore produce less potential error or distortion than the diagonals. However, such other medians remain available in cases where noise reduction is difficult using only the vertical and horizontal medians.
Another beneficial source of noise reduction is information from the previous and subsequent frame (i.e., a temporal median). As mentioned below, motion analysis provides the best match for moving regions. However, it is compute intensive. If a region of the image is not moving, or is moving slowly, the red values (and green and blue) from a current pixel can be median filtered with the red value at that same pixel location in the previous and subsequent frames. However, odd artifacts may occur if significant motion is present and such a temporal filter is used. Thus, it is preferred that a threshold be taken first, to determine whether such a median would differ more than a selected amount from the value of a current pixel. The threshold can be computed essentially the same as for the de-interlacing threshold above:
Rdiff=R_current_pixel minus R_temporal_median
Gdiff=G_current_pixel minus G_temporal_median
Bdiff=B_current_pixel minus B_temporal_median
ThresholdingValue=abs(Rdiff+Gdiff+Bdiff)+abs (Rdiff)+abs(Gdiff)+abs(Bdiff)
The ThresholdingValue is then compared to a threshold setting. Typical threshold settings are in the range 0.1 to 0.3, with 0.2 being typical. Above the threshold, the current value is kept. Below the threshold, the temporal median is used. The block diagram of FIG. 3 also applies to this threshold test. In this case the PROCESSING block 30 is a temporal median filter and the inputs are three successive frames. The SELECTION CONTROL block 32 compares the output 36 of the PROCESSING block 30 with Input B 34 using the above equations for Rdiff, Gdiff, Bdiff, and ThresholdingValue. The switch selects the PROCESSING output 36 if the ThresholdingValue is less than the threshold, otherwise the switch selects Input B 34, the middle value, for the output 38.
An additional median type is a median taken between the X, Y, and temporal medians. Another median type can take the temporal median, and then take the equal average of the X and Y medians from it.
Each type of median can cause problems. X and Y medians smear and blur an image, so that it looks “greasy”. Temporal medians cause smearing of motion over time. Since each median can result in problems, yet each median's properties are different (and, in some sense, “orthogonal”), it has been determined experimentally that the best results come by combining a variety of medians.
In particular, FIG. 4 shows a preferred combination of medians is a linear weighted sum (see the discussion above on linear video processing) of five terms to determine the value for each pixel of a current image:
50% of the original image (Frame N 40) (thus, the most noise reduction is 3 db, or half);
15% of the average of X and Y medians 42, 44, respectively;
10% of the thresholded temporal median 46;
10% of the average of X and Y medians of the thresholded temporal median (48); and
15% of a three-way X, Y, and temporal median (50).
This set of time medians does a reasonable job of reducing the noise in the image without making it appear “greasy” or blurred, causing temporal smearing of moving objects, or losing detail. Another useful weighting of these five terms is 35%, 20%, 22.5%, 10%, and 12.5%, respectively.
In addition, it is useful to apply motion-compensation by applying center weighted temporal filters to a motion-compensated n×n region, as described below. This can be added to the median filtered image result (of five terms, just described) to further smooth the image, providing better smoothing and detail on moving image regions.
Motion Analysis
In addition to “in-place” temporal filtering, which does a good job at smoothing slow-moving details, de-interlacing and noise reduction can also be improved by use of motion analysis. Adding the pixels at the same location in three fields or three frames is valid for stationary objects. However, for moving objects, if temporal averaging/smoothing is desired, it is often more optimal to attempt to analyze prevailing motion over a small group of pixels. For example, an n×n block of pixels (e.g., 2×2, 3×3, 4×4, 6×6, or 8×8) can be used to search in previous and subsequent fields or frames to attempt to find a match (in the same way MPEG-2 motion vectors are found by matching 16×16 macroblocks). Once a best match is found in one or more previous and subsequent frames, a “trajectory” and “moving mini-picture” can be determined. For interlaced fields, it is best to analyze comparisons as well as compute inferred moving mini-pictures utilizing the results of the thresholded de-interlaced process above. Since this process has already separated the fast-moving from the slow-moving details, and has already smoothed the slow moving details, the picture comparisons and reconstructions are more applicable than individual de-interlaced fields.
The motion analysis preferably is performed by comparison of an n×n block in the current thresholded de-interlaced image with all nearby blocks in the previous and subsequent one or more frames. The comparison may be the absolute value of differences in luminance or RGB over the n×n block. One frame is sufficient forward and backward if the motion vectors are nearly equal and opposite. However, if the motion vectors are not nearly equal and opposite, then an additional one or two frames forward and backward can help determine the actual trajectory. Further, different de-interlacing treatments may be useful in helping determine the “best guess” motion vectors going forward and back. One de-interlacing treatment can be to use only individual de-interlaced fields, although this is heavily prone to aliasing and artifacts on small moving details. Another de-interlacing technique is to use only the three-field-frame smooth de-interlacing, without thresholding, having weightings [0.25, 0.5, 0.25], as described above. Although details are smoothed and sometimes lost, the trajectory may often be more correct.
Once a trajectory is found, a “smoothed n×n block” can be created by temporally filtering using the motion-vector-offset pixels from the one (or more) previous and subsequent frames. A typical filter might again be [0.25, 0.5, 0.25] or [0.1667, 0.6666, 0.1667] for three frames, and possibly [0.1, 0.2, 0.4, 0.2, 0.1] for two frames back and forward. Other filters, with less central weight, are also useful, especially with smaller block sizes (such as 2×2, 3×3, and 4×4). Reliability of the match between frames is indicated by the absolute difference value. Large minimum absolute differences can be used to select more center weight in the filter. Lower values of absolute differences can suggest a good match, and can be used to select less center weight to more evenly distribute the average over a span of several frames of motion-compensated blocks.
These filter weights can be applied to: individual de-interlaced motion-compensated field-frames; thresholded three-field-frame de-interlaced pictures, described above; and non-thresholded three-field-frame de-interlaced images, with a [0.25, 0.5, 0.25] weighting, also as described above. However, the best filter weights usually come from applying the motion-compensated block linear filtering to the thresholded three-field-frame result described above. This is because the thresholded three-field-frame image is both the smoothest (in terms of removing aliasing in smooth areas), as well as the most motion-responsive (in terms of defaulting to a single de-interlaced field-frame above the threshold). Thus, the motion vectors from motion analysis can be used as the inputs to multi-frame or multi-de-interlaced-field-frame or single-de-interlaced field-frame filters, or combinations thereof. The thresholded multi-field-frame de-interlaced images, however, form the best filter input in most cases.
The use of motion analysis is computationally expensive for a large search region, when fast motion might be found (such as ±32 pixels). Accordingly, it may be best to augment the speed by using special-purpose hardware or a digital signal processor assisted computer.
Once motion vectors are found, together with their absolute difference measure of accuracy, they can be utilized for the complex process of attempting frame rate conversion. However, occlusion issues (objects obscuring or revealing others) will confound matches, and cannot be accurately inferred automatically. Occlusion can also involve temporal aliasing, as can normal image temporal undersampling and its beat with natural image frequencies (such as the “backward wagon wheel” effect in movies). These problems often cannot be unraveled by any known computation technique, and to date require human assistance. Thus, human scrutiny and adjustment, when real-time automatic processing is not required, can be used for off-line and non-real-time frame-rate conversion and other similar temporal processes.
De-interlacing is a simple form of the same problem. Just as with frame-rate-conversion, the task of de-interlacing is theoretically impossible to perform perfectly. This is especially due to the temporal undersampling (closed shutter), and an inappropriate temporal sample filter (i.e., a box filter). However, even with correct samples, issues such as occlusion and interlace aliasing further ensure the theoretical impossibility of correct results. The cases where this is visible are mitigated by the depth of the tools, as described here, which are applied to the problem. Pathological cases will always exist in real image sequences. The goal can only be to reduce the frequency and level of impairment when these sequences are encountered. However, in many cases, the de-interlacing process can be acceptably fully automated, and can run unassisted in real-time. Even so, there are many parameters which can often benefit from manual adjustment.
Filter Smoothing of High Frequencies
In addition to median filtering, reducing high frequency detail will also reduce high frequency noise. However, this smoothing comes at the price of loss of sharpness and detail. Thus, only a small amount of such smoothing is generally useful. A filter which creates smoothing can be easily made, as with the threshold for de-interlacing, by down-filtering with a normal filter (e.g., truncated sinc filter) and then up-filtering with a gaussian filter. The result will be smoothed because it is devoid of high frequency picture detail. When such a term is added, it typically must be in very small amounts, such as 5% to 10%, in order to provide a small amount of noise reduction. In larger amounts, the blurring effect generally becomes quite visible.
Base Layer Noise Filtering
The filter parameters for the median filtering described above for an original image should be matched to the noise characteristics of the film grain or image sensor that captured the image. After this median filtered image is down-filtered to generate an input to the base layer compression process, it still contains a small amount of noise. This noise may be further reduced by a combination of another X-Y median filters (equally averaging the X and Y medians), plus a very small amount of the high frequency smoothing filter. A preferred filter weighting of these three terms, applied to each pixel of the base layer, is:
75% of the original base layer (down filtered from median-filtered original above);
22.5% of the average of X and Y medians; and
7.5% of the down-up smoothing filter.
This small amount of additional filtering in the base layer provides a small additional amount of noise reduction and improved stability, resulting in better MPEG encoding and limiting the amount of noise added by such encoding.
Image Filtering
Downsizing and Upsizing Filters
Experimentation has shown that the downsizing filter used in creating a base layer from a high resolution original picture is most optimal if it includes modest negative lobes and an extent which stops after the first very small positive lobes after the negative lobes. FIG. 5 is a diagram of the relative shape, amplitudes, and lobe polarity of a preferred downsizing filter. The down filter essentially is a center-weighted function which has been truncated to a center positive lobe 500, a symmetric pair of adjacent (bracketing) small negative lobes 504, and a symmetric pair of adjacent (bracketing) very small outer positive lobes 504. The absolute amplitude of the lobes 500, 502, 504 may be adjusted as desired, so long as the relative polarity and amplitude inequality relationships shown in FIG. 5 are maintained. However, a good first approximation for the relative amplitudes are defined by a truncated sinc function (sinc(x)=sin(x)/x)). Such filters can be used separably, which means that the horizontal data dimension is independently filtered and resized, and then the vertical data dimension, or vise versa; the result is the same.
When creating a base layer original (as input to the base layer compression) from a low-noise high resolution original input, the preferred downsizing filter has first negative lobes which are of a normal sinc function amplitude. For clean and for high resolution input images, this normal truncated sinc function works well. For lower resolutions (e.g., 1280×720, 1024×768, or 1536×768), and for noisier input pictures, a reduced first negative lobe amplitude in the filters is more optimal. A suitable amplitude in such cases is about half the truncated sinc function negative lobe amplitude. The small first positive lobes outside of the first negative lobes are also reduced to lower amplitude, typically to ½ to ⅔ of the normal sinc function amplitude. The affect of reducing the first negative lobes is the main issue, since the small outside positive lobes do not contribute to picture noise. Further samples outside the first positive lobes preferably are truncated to minimize ringing and other potential artifacts.
The choice of whether to use milder negative lobes or full sinc function amplitude negative lobes in the downfilter is determined by the resolution and noise level of the original image. It is also somewhat a function of image content, since some types of scenes are easier to code than others (mainly related to the amount of motion and change in a particular shot). By using a “milder” downfilter having reduced negative lobes, noise in the base layer is reduced, and a cleaner and quieter compression of the base layer is achieved, thus also resulting in fewer artifacts.
Experimentation has also shown that the optimal upsizing filter has a center positive lobe with small adjacent negative lobes, but no further positive lobes. FIGS. 6A and 6B are diagrams of the relative shape, amplitudes, and lobe polarity of a pair of preferred upsizing filters for upsizing by a factor of 2. A central positive lobe 600, 600′ is bracketed by a pair of small negative lobes 602, 602′. An asymmetrically placed positive lobe 604, 604′ is also required. These paired upfilters could also be considered to be truncated sinc filters centered on the newly created samples. For example, for a factor of two upfilter, two new samples will be created for each original sample. The small adjacent negative lobes 602, 602′ have less negative amplitude than is used in the corresponding downsizing filter (FIG. 5), or than would be used in an optimal (sinc-based) upsizing filter for normal images. This is because the images being upsized are decompressed, and the compression process changes the spectral distribution. Thus, more modest negative lobes, and no additional positive lobes beyond the middle ones 600, 600′, work better for upsizing a decompressed base layer.
Experimentation has shown that slight negative lobes 602, 602′ provide a better layered result than positive-only gaussian or spline upfilters (note that splines can have negative lobes, but are most often used in the positive-only form). Thus, this upsizing filter preferably is used for the base layer in both the encoder and the decoder.
Weighting of High Octave of Picture Detail
In the preferred embodiment, the signal path which expands the original uncompressed base layer input image uses a gaussian upfilter rather than the upfilter described above. In particular, a gaussian upfilter is used for the “high octave” of picture detail, which is determined by subtracting the expanded original base-resolution input image (without using compression) from the original picture. Thus, no negative lobes are used for this particular upfiltered expansion.
As noted above, for MPEG-2 this high octave difference signal path is typically weighted with 0.25 (or 25%) and added to the expanded decompressed base layer (using the other upfilter described above) as input to the enhancement layer compression process. However, experimentation has shown that weights of 10%, 15%, 20%, 30%, and 35% are useful for particular images when using MPEG-2. Other weights may also prove useful. For MPEG-4, it has been found that filter weights of 4-8% may be optimal when used in conjunction with other improvements described below. Accordingly, this weighting should be regarded as an adjustable parameter, depending upon the encoding system, the scenes being encoded/compressed, the particular camera (or film) being used, and the image resolution.
Filters with Negative Lobes for Motion Compensation in MPEG-2 and MPEG-4
In MPEG-4, reference filters have been implemented for shifting macroblocks when finding the best motion vector match, and then using the matched region for motion compensation. MPEG-4 video coding, like MPEG-2, supports ½ pixel resolution of motion vectors for macroblocks. Unlike MPEG-2, MPEG-4 also supports ¼ pixel accuracy. However, in the reference implementation of MPEG-4, the filters used are sub-optimal. In MPEG-2, the half-way point between pixels is just the average of the two neighbors, which is a sub-optimal box filter. In MPEG-4, this filter is used for ½ pixel resolution. If ¼ pixel resolution is invoked in MPEG-4 Part 2, a filter with negative lobes is used for the half-way point, but a sub-optimal box filter with this result and the neighboring pixels is used for the ¼ and ¾ points.
Further, the chrominance channels (U=R−Y and V=B−Y) do not use any sub-pixel resolution in the motion compensation step under MPEG-4. Since the luminance channel (Y) has resolution to the ½ or ¼ pixel, the half-resolution chrominance U and V channels should be sampled using filters to ¼ pixel resolution, corresponding to ½ pixel in luminance. When ¼ pixel resolution is selected for luminance, then ⅛ pixel resolution should be used for U and V chrominance.
Experiments have shown that the effects of filtering are significantly improved by using a negative lobe truncated sinc function (as described above) for filtering the ¼, ½, and ¾ pixel points when doing ¼ pixel resolution in luminance, and by using similar negative lobes when doing ½ pixel resolution for the filter which creates the ½ pixel position.
Similarly, effects of filtering are significantly improved by using a negative lobe truncated sinc function for filtering the ⅛-pixel points for U and V chrominance when using ¼ pixel luminance resolution, and by using ¼ pixel resolution filters with similar negative lobe filters when using ½ pixel luminance resolution.
It has been discovered that the combination of quarter-pixel motion vectors with truncated sinc motion compensated displacement filtering results in a major improvement in picture quality. In particular, clarity is improved, noise and artifacts are reduced, and chroma detail is increased.
These filters may be applied to video images under MPEG-1, MPEG-2, MPEG-4 or any other appropriate motion-compensated block-based image coding system.
COMPUTER IMPLEMENTATION
The invention may be implemented in hardware or software, or a combination of both. However, preferably, the invention is implemented in computer programs executing on one or more programmable computers each comprising at least a processor, a data storage system (including volatile and non-volatile memory and/or storage elements), an input device, and an output device. Program code is applied to input data to perform the functions described herein and generate output information. The output information is applied to one or more output devices, in known fashion.
Each such program may be implemented in any desired computer language (including machine, assembly, or high level procedural, logical, or object oriented programming languages) to communicate with a computer system. In any case, the language may be a compiled or interpreted language.
Each such computer program is preferably stored on a storage media or device (e.g., ROM, CD-ROM, or magnetic or optical media) readable by a general or special purpose programmable computer system, for configuring and operating the computer when the storage media or device is read by the computer system to perform the procedures described herein. The inventive system may also be considered to be implemented as a computer-readable storage medium, configured with a computer program, where the storage medium so configured causes a computer system to operate in a specific and predefined manner to perform the functions described herein.
A number of embodiments of the invention have been described. Nevertheless, it will be understood that various modifications may be made without departing from the spirit and scope of the invention. For example, while the preferred embodiment uses MPEG-2 or MPEG-4 coding and decoding, the invention will work with any comparable standard that provides equivalents of I, P, and/or B frames and layers. Accordingly, it is to be understood that the invention is not to be limited by the specific illustrated embodiment, but only by the scope of the appended claims.

Claims (24)

What is claimed is:
1. A method for enhancing image quality in an image encoding system, including:
applying a temporal median filter to corresponding pixel values of a previous digital video image, a current digital video image, and a next digital video image to create a noise-reduced digital video image;
comparing the difference between each corresponding pixel value of each noise-reduced digital video image and each corresponding current digital video image to a threshold value to generate a difference value; and
selecting, for each final pixel value for the noise-reduced digital video image, a corresponding pixel value from the current digital video image if the difference value is within a first threshold comparison range, and a corresponding pixel value from the noise-reduced digital video image if the difference value is within a second threshold comparison range.
2. A method for enhancing image quality in an image encoding system, including:
applying a temporal median filter to corresponding pixel values of a previous digital video image, a current digital video image, and a next digital video image to create a noise-reduced digital video image;
comparing the difference between each corresponding pixel value of each noise-reduced digital video image and each corresponding current digital video image to a threshold value to generate a difference value; and
selecting, for each final pixel value for the noise-reduced digital video image, a corresponding pixel value from the current digital video image if the difference value is within a first threshold comparison range, and a corresponding pixel value from the noise-reduced digital video image if the difference value is within a second threshold comparison range,
wherein the threshold value is selected from the range of approximately 0.1 to approximately 0.3.
3. A method for enhancing image quality in an image encoding system, including creating a noise-reduced digital video image comprising a linear weighted sum of five terms:
a current digital video image;
an average of horizontal and vertical medians of the current digital video image;
a thresholded temporal median;
an average of horizontal and vertical medians of the thresholded temporal median; and
a median of the thresholded temporal median and horizontal and vertical medians of the current digital video image,
wherein the weights of the five terms are approximately 50%, 15%, 10%, 10%, and 15%, respectively.
4. A method for enhancing image quality in an image encoding system, including creating a noise-reduced digital video image comprising a linear weighted sum of five terms:
a current digital video image;
an average of horizontal and vertical medians of the current digital video image;
a thresholded temporal median;
an average of horizontal and vertical medians of the thresholded temporal median; and
a median of the thresholded temporal median and horizontal and vertical medians of the current digital video image,
wherein the weights of the five terms are approximately 35%, 20%, 22.5%, 10%, and 12.5%, respectively.
5. A method for enhancing image quality in an image encoding system, including:
creating a noise-reduced digital video image comprising a linear weighted sum of five terms:
a current digital video image;
an average of horizontal and vertical medians of the current digital video image;
a thresholded temporal median;
an average of horizontal and vertical medians of the thresholded temporal median; and
a median of the thresholded temporal median and horizontal and vertical medians of the current digital video image;
determining a motion vector for each n×n pixel region of the current digital video image with respect to at least one previous digital video image and at least one subsequent digital video image;
applying a center weighted temporal filter to each n×n pixel region of the current digital video image and corresponding motion-vector offset n×n pixel regions of the at least one previous digital video image and at least one subsequent digital video image to create a motion-compensated image; and
adding the motion-compensated image to the noise-reduced digital video image.
6. A method for enhancing image quality in an image encoding system, including:
determining a motion vector for each n×n pixel region of a current digital video image with respect to at least one previous digital video image and at least one subsequent digital video image; and
applying a center weighted temporal filter to each n×n pixel region of the current digital video image and corresponding motion-vector offset n×n pixel regions of the at least one previous digital video image and at least one subsequent digital video image to create a motion-compensated image,
wherein each digital video image is a three-field-frame de-interlaced image.
7. A method for enhancing image quality in an image encoding system, including:
determining a motion vector for each n×n pixel region of a current digital video image with respect to at least one previous digital video image and at least one subsequent digital video image; and
applying a center weighted temporal filter to each n×n pixel region of the current digital video image and corresponding motion-vector offset n×n pixel regions of the at least one previous digital video image and at least one subsequent digital video image to create a motion-compensated image,
wherein each digital video image is a thresholded three-field-frame de-interlaced image.
8. A method for enhancing image quality in an image encoding system, including:
determining a motion vector for each n×n pixel region of a current digital video image with respect to at least one previous digital video image and at least one subsequent digital video image; and
applying a center weighted temporal filter to each n×n pixel region of the current digital video image and corresponding motion-vector offset n×n pixel regions of the at least one previous digital video image and at least one subsequent digital video image to create a motion-compensated image,
wherein the center weighted temporal filter is a three-image temporal filter having weights for each of such images of approximately 25%, 50%, and 25%, respectively.
9. A method for enhancing image quality in an image encoding system, including:
determining a motion vector for each n×n pixel region of a current digital video image with respect to at least one previous digital video image and at least one subsequent digital video image; and
applying a center weighted temporal filter to each n×n pixel region of the current digital video image and corresponding motion-vector offset n×n pixel regions of the at least one previous digital video image and at least one subsequent digital video image to create a motion-compensated image,
wherein the center weighted temporal filter is a five-image temporal filter having weights for each of such images of approximately 10%, 20%, 40%, 20%, and 10%, respectively.
10. A method for enhancing image quality in an image encoding system, including:
applying a normal down filter to an image to create a first intermediate image;
applying a Gaussian up filter to the first intermediate image to create a second intermediate image; and
adding a weighted fraction of the second intermediate image to a selected image to create an image having reduced high frequency noise.
11. The method of claim 10, wherein the weighted fraction is between approximately 5% and 10% of the second intermediate image.
12. A method for enhancing image quality in an image system, the method comprising:
applying a filter to adjacent pixel values of a video image to generate a filtered value for motion compensation with sub-pixel displacement, the filter including a first negative lobe, a second negative lobe, and a positive lobe disposed between the first and second negative lobes;
wherein an absolute amplitude of each of the negative lobes is less than an absolute amplitude of the positive lobe.
13. The method of claim 12, wherein the filter comprises only four values.
14. The method of claim 13, wherein the positive lobe comprises only of first and second values of the four values, the first and second values differing from each other.
15. The method of claim 14, wherein the first value is greater than the second value.
16. The method of claim 14, wherein the first value is less than the second value.
17. The method of claim 12, wherein the filter comprises at least four values.
18. The method of claim 17, wherein the positive lobe comprises a first value and a second value, the first and second value differing from each other.
19. The method of claim 18, wherein the first value is greater than the second value.
20. The method of claim 18, wherein the first value is less than the second value.
21. The method of claim 12, wherein the adjacent pixel values correspond to adjacent pixels horizontally aligned.
22. The method of claim 21, further comprising applying the filter to adjacent pixel values vertically aligned.
23. The method of claim 12, wherein the adjacent pixel values correspond to adjacent pixels vertically aligned.
24. The method of claim 12, wherein all positive lobes are disposed between of the first and second negative lobes.
US13/366,485 1996-01-30 2012-02-06 Enhancing image quality in an image system Expired - Fee Related USRE44235E1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US13/366,485 USRE44235E1 (en) 1996-01-30 2012-02-06 Enhancing image quality in an image system

Applications Claiming Priority (6)

Application Number Priority Date Filing Date Title
US08/594,815 US5852565A (en) 1996-01-30 1996-01-30 Temporal and resolution layering in advanced television
US09/217,151 US5988863A (en) 1996-01-30 1998-12-21 Temporal and resolution layering in advanced television
US44259599A 1999-11-17 1999-11-17
US09/545,233 US6728317B1 (en) 1996-01-30 2000-04-07 Moving image compression quality enhancement using displacement filters with negative lobes
US10/814,350 US7366242B2 (en) 1996-01-30 2004-03-30 Median filter combinations for video noise reduction
US13/366,485 USRE44235E1 (en) 1996-01-30 2012-02-06 Enhancing image quality in an image system

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US10/814,350 Reissue US7366242B2 (en) 1996-01-30 2004-03-30 Median filter combinations for video noise reduction

Publications (1)

Publication Number Publication Date
USRE44235E1 true USRE44235E1 (en) 2013-05-21

Family

ID=32110755

Family Applications (7)

Application Number Title Priority Date Filing Date
US09/541,701 Expired - Fee Related US6957350B1 (en) 1996-01-30 2000-04-03 Encrypted and watermarked temporal and resolution layering in advanced television
US09/545,233 Expired - Lifetime US6728317B1 (en) 1996-01-30 2000-04-07 Moving image compression quality enhancement using displacement filters with negative lobes
US10/814,350 Expired - Lifetime US7366242B2 (en) 1996-01-30 2004-03-30 Median filter combinations for video noise reduction
US11/187,176 Expired - Fee Related US7428639B2 (en) 1996-01-30 2005-07-21 Encrypted and watermarked temporal and resolution layering in advanced television
US12/618,202 Expired - Fee Related USRE43360E1 (en) 1996-01-30 2009-11-13 Enhancing image quality in an image system
US13/366,485 Expired - Fee Related USRE44235E1 (en) 1996-01-30 2012-02-06 Enhancing image quality in an image system
US13/867,458 Expired - Fee Related USRE45082E1 (en) 1996-01-30 2013-04-22 Enhancing image quality in an image system

Family Applications Before (5)

Application Number Title Priority Date Filing Date
US09/541,701 Expired - Fee Related US6957350B1 (en) 1996-01-30 2000-04-03 Encrypted and watermarked temporal and resolution layering in advanced television
US09/545,233 Expired - Lifetime US6728317B1 (en) 1996-01-30 2000-04-07 Moving image compression quality enhancement using displacement filters with negative lobes
US10/814,350 Expired - Lifetime US7366242B2 (en) 1996-01-30 2004-03-30 Median filter combinations for video noise reduction
US11/187,176 Expired - Fee Related US7428639B2 (en) 1996-01-30 2005-07-21 Encrypted and watermarked temporal and resolution layering in advanced television
US12/618,202 Expired - Fee Related USRE43360E1 (en) 1996-01-30 2009-11-13 Enhancing image quality in an image system

Family Applications After (1)

Application Number Title Priority Date Filing Date
US13/867,458 Expired - Fee Related USRE45082E1 (en) 1996-01-30 2013-04-22 Enhancing image quality in an image system

Country Status (1)

Country Link
US (7) US6957350B1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8767829B2 (en) 2001-07-11 2014-07-01 Dolby Laboratories Licensing Corporation Switch-select single frame reference

Families Citing this family (460)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6957350B1 (en) 1996-01-30 2005-10-18 Dolby Laboratories Licensing Corporation Encrypted and watermarked temporal and resolution layering in advanced television
KR100205009B1 (en) * 1996-04-17 1999-06-15 윤종용 A video signal conversion device and a display device having the same
DE19654342C2 (en) * 1996-12-24 1998-10-15 Karl Michael Marks Method for the transmission of compressed data
CN1134172C (en) * 1997-01-13 2004-01-07 皇家菲利浦电子有限公司 Emboding supplemental data in digital video signal
US6711278B1 (en) * 1998-09-10 2004-03-23 Microsoft Corporation Tracking semantic objects in vector image sequences
US6983018B1 (en) * 1998-11-30 2006-01-03 Microsoft Corporation Efficient motion vector coding for video compression
US6600908B1 (en) 1999-02-04 2003-07-29 Hark C. Chan Method and system for broadcasting and receiving audio information and associated audio indexes
US6952685B1 (en) 1999-08-27 2005-10-04 Ochoa Optics Llc Music distribution system and associated antipiracy protection
US7209900B2 (en) 1999-08-27 2007-04-24 Charles Eric Hunter Music distribution systems
US8090619B1 (en) 1999-08-27 2012-01-03 Ochoa Optics Llc Method and system for music distribution
US8656423B2 (en) * 1999-08-27 2014-02-18 Ochoa Optics Llc Video distribution system
US6614936B1 (en) * 1999-12-03 2003-09-02 Microsoft Corporation System and method for robust video coding using progressive fine-granularity scalable (PFGS) coding
IL134182A (en) 2000-01-23 2006-08-01 Vls Com Ltd Method and apparatus for visual lossless pre-processing
US9252898B2 (en) * 2000-01-28 2016-02-02 Zarbaña Digital Fund Llc Music distribution systems
JP2001285863A (en) * 2000-03-30 2001-10-12 Sony Corp Device and method for converting image information
US7634011B2 (en) * 2000-04-21 2009-12-15 Microsoft Corporation Application program interface (API) facilitating decoder control of accelerator resources
US7649943B2 (en) * 2000-04-21 2010-01-19 Microsoft Corporation Interface and related methods facilitating motion compensation in media processing
US6940912B2 (en) * 2000-04-21 2005-09-06 Microsoft Corporation Dynamically adaptive multimedia application program interface and related methods
US6891893B2 (en) * 2000-04-21 2005-05-10 Microsoft Corp. Extensible multimedia application program interface and related methods
US20020026636A1 (en) * 2000-06-15 2002-02-28 Daniel Lecomte Video interfacing and distribution system and method for delivering video programs
US6753929B1 (en) 2000-06-28 2004-06-22 Vls Com Ltd. Method and system for real time motion picture segmentation and superposition
KR100353851B1 (en) * 2000-07-07 2002-09-28 한국전자통신연구원 Water ring scan apparatus and method, video coding/decoding apparatus and method using that
US7020193B2 (en) * 2000-09-22 2006-03-28 Koninklijke Philips Electronics N.V. Preferred transmission/streaming order of fine-granular scalability
US20030079222A1 (en) * 2000-10-06 2003-04-24 Boykin Patrick Oscar System and method for distributing perceptually encrypted encoded files of music and movies
US7463683B2 (en) * 2000-10-11 2008-12-09 Koninklijke Philips Electronics N.V. Method and apparatus for decoding spatially scaled fine granular encoded video signals
JP2002158654A (en) * 2000-11-17 2002-05-31 Hitachi Ltd Information processor, display device, digital contents distribution system and digital contents distribution/ output method
US6907070B2 (en) * 2000-12-15 2005-06-14 Microsoft Corporation Drifting reduction and macroblock-based control in progressive fine granularity scalable video coding
US7095445B2 (en) * 2000-12-20 2006-08-22 Samsung Electronics Co., Ltd. Method of detecting motion in an interlaced video sequence based on logical operation on linearly scaled motion information and motion detection apparatus
US20020126910A1 (en) * 2001-01-02 2002-09-12 Eastman Kodak Company Method of calculating noise from multiple digital images utilizing common noise characteristics
US7062445B2 (en) * 2001-01-26 2006-06-13 Microsoft Corporation Quantization loop with heuristic approach
JP4019303B2 (en) * 2001-02-02 2007-12-12 日本電気株式会社 ENCRYPTION DEVICE AND DECRYPTION DEVICE USING ENCRYPTION KEY INCLUDED IN ELECTRONIC WATERMARK AND METHOD THEREOF
CN1294534C (en) * 2001-02-06 2007-01-10 皇家菲利浦电子有限公司 Preventing green non-uniformity in image sensors
US8374237B2 (en) * 2001-03-02 2013-02-12 Dolby Laboratories Licensing Corporation High precision encoding and decoding of video images
FR2822330B1 (en) * 2001-03-14 2003-05-02 Thomson Multimedia Sa BLOCK CODING METHOD, MPEG TYPE, IN WHICH A RESOLUTION IS ASSIGNED TO EACH BLOCK
US6934257B2 (en) * 2001-04-04 2005-08-23 Intel Corporation Transferring transmission control protocol packets
JP3715249B2 (en) * 2001-04-27 2005-11-09 シャープ株式会社 Image processing circuit, image display device, and image processing method
US7206453B2 (en) * 2001-05-03 2007-04-17 Microsoft Corporation Dynamic filtering for lossy compression
JP3861623B2 (en) * 2001-06-05 2006-12-20 ソニー株式会社 Digital watermark embedding processing apparatus, digital watermark embedding processing method, and program
JP2003003979A (en) * 2001-06-25 2003-01-08 Toshiba Kyaria Kk Fluid machine
US6757648B2 (en) * 2001-06-28 2004-06-29 Microsoft Corporation Techniques for quantization of spectral data in transcoding
US8111754B1 (en) 2001-07-11 2012-02-07 Dolby Laboratories Licensing Corporation Interpolation of video compression frames
US6816552B2 (en) * 2001-07-11 2004-11-09 Dolby Laboratories Licensing Corporation Interpolation of video compression frames
US20030112863A1 (en) 2001-07-12 2003-06-19 Demos Gary A. Method and system for improving compressed image chroma information
GB0119569D0 (en) * 2001-08-13 2001-10-03 Radioscape Ltd Data hiding in digital audio broadcasting (DAB)
FR2828977B1 (en) * 2001-08-21 2003-12-05 Nextream Sa DEVICE AND METHOD FOR ESTIMATING THE NOISE LEVEL, NOISE REDUCTION SYSTEM AND ENCODING SYSTEM COMPRISING SUCH A DEVICE
US7960005B2 (en) 2001-09-14 2011-06-14 Ochoa Optics Llc Broadcast distribution of content for storage on hardware protected optical storage media
US6882685B2 (en) * 2001-09-18 2005-04-19 Microsoft Corporation Block transform and quantization for image and video coding
US7646816B2 (en) * 2001-09-19 2010-01-12 Microsoft Corporation Generalized reference decoder for image or video processing
JP4574090B2 (en) * 2001-09-21 2010-11-04 三菱電機株式会社 Movie data converter and movie data conversion program
US7263124B2 (en) * 2001-09-26 2007-08-28 Intel Corporation Scalable coding scheme for low latency applications
US20030076858A1 (en) * 2001-10-19 2003-04-24 Sharp Laboratories Of America, Inc. Multi-layer data transmission system
EP1442601A1 (en) * 2001-10-26 2004-08-04 Koninklijke Philips Electronics N.V. Method and appartus for spatial scalable compression
KR100603592B1 (en) * 2001-11-26 2006-07-24 학교법인 고황재단 Intelligent Water ring scan apparatus and method using Quality Factor, video coding/decoding apparatus and method using that
US7460993B2 (en) * 2001-12-14 2008-12-02 Microsoft Corporation Adaptive window-size selection in transform coding
US7027982B2 (en) * 2001-12-14 2006-04-11 Microsoft Corporation Quality and rate control strategy for digital audio
CN101448162B (en) * 2001-12-17 2013-01-02 微软公司 Method for processing video image
US7428019B2 (en) * 2001-12-26 2008-09-23 Yeda Research And Development Co. Ltd. System and method for increasing space or time resolution in video
US7636515B2 (en) * 2001-12-27 2009-12-22 Stmicroelectronics, Inc. Apparatus and method for transcoding still image data files into MPEG video data files and digital video player implementing same
US7886151B2 (en) * 2002-01-22 2011-02-08 Purdue Research Foundation Temporal synchronization of video and audio signals
US7149247B2 (en) * 2002-01-22 2006-12-12 Microsoft Corporation Methods and systems for encoding and decoding video data to enable random access and splicing
US7003035B2 (en) * 2002-01-25 2006-02-21 Microsoft Corporation Video coding methods and apparatuses
US10277656B2 (en) * 2002-01-29 2019-04-30 FiveOpenBooks, LLC Method and system for delivering media data
DE60332328D1 (en) * 2002-03-13 2010-06-10 Imax Corp SYSTEMS AND METHODS FOR DIGITAL REMASTERING OR OTHERWISE MODIFYING MOBILE IMAGES OR OTHER IMAGE RECORDING DATA
US7116831B2 (en) * 2002-04-10 2006-10-03 Microsoft Corporation Chrominance motion vector rounding
US7620109B2 (en) * 2002-04-10 2009-11-17 Microsoft Corporation Sub-pixel interpolation in motion estimation and compensation
US7110459B2 (en) * 2002-04-10 2006-09-19 Microsoft Corporation Approximate bicubic filter
JP4183432B2 (en) * 2002-04-10 2008-11-19 富士通株式会社 Image data encoding method
US7305034B2 (en) * 2002-04-10 2007-12-04 Microsoft Corporation Rounding control for multi-stage interpolation
JP4031390B2 (en) * 2002-04-17 2008-01-09 松下電器産業株式会社 Image conversion apparatus and image conversion method
JP4031389B2 (en) 2002-04-17 2008-01-09 松下電器産業株式会社 Image conversion apparatus and image conversion method
US7366909B2 (en) * 2002-04-29 2008-04-29 The Boeing Company Dynamic wavelet feature-based watermark
US7242713B2 (en) * 2002-05-02 2007-07-10 Microsoft Corporation 2-D transforms for image and video coding
US20050135613A1 (en) * 2002-05-10 2005-06-23 Karlheinz Brandenburg Device and method for generating encrypted data, for decrypting encrypted data and for generating re-signed data
FR2840147B1 (en) * 2002-05-24 2004-08-27 France Telecom VIDEO INTERFERENCE AND SCRAMBLING METHODS, SYSTEM, DECODER, BROADCAST SERVER, DATA MEDIUM FOR CARRYING OUT SAID METHODS
US20040001546A1 (en) 2002-06-03 2004-01-01 Alexandros Tourapis Spatiotemporal prediction for bidirectionally predictive (B) pictures and motion vector prediction for multi-picture reference motion compensation
US6980695B2 (en) * 2002-06-28 2005-12-27 Microsoft Corporation Rate allocation for mixed content video
US7085420B2 (en) * 2002-06-28 2006-08-01 Microsoft Corporation Text detection in continuous tone image segments
AU2015218509B2 (en) * 2002-06-28 2015-10-22 Dolby Laboratories Licensing Corporation Region sizing for macroblocks
US7224731B2 (en) 2002-06-28 2007-05-29 Microsoft Corporation Motion estimation/compensation for screen capture video
US7016547B1 (en) 2002-06-28 2006-03-21 Microsoft Corporation Adaptive entropy encoding/decoding for screen capture content
US20040005003A1 (en) * 2002-07-02 2004-01-08 Koninklijke Philips Electronics N.V. Quality improvement for FGS BL coding with U/V coarse quantization
US7280700B2 (en) * 2002-07-05 2007-10-09 Microsoft Corporation Optimization techniques for data compression
US7003131B2 (en) * 2002-07-09 2006-02-21 Kaleidescape, Inc. Watermarking and fingerprinting digital content using alternative blocks to embed information
AU2003281133A1 (en) * 2002-07-15 2004-02-02 Hitachi, Ltd. Moving picture encoding method and decoding method
US7154952B2 (en) * 2002-07-19 2006-12-26 Microsoft Corporation Timestamp-independent motion vector prediction for predictive (P) and bidirectionally predictive (B) pictures
US7072512B2 (en) * 2002-07-23 2006-07-04 Microsoft Corporation Segmentation of digital video and images into continuous tone and palettized regions
DE60224035D1 (en) * 2002-08-23 2008-01-24 St Microelectronics Srl Method for noise filtering a numerical image sequence
US7421129B2 (en) * 2002-09-04 2008-09-02 Microsoft Corporation Image compression and synthesis for video effects
US7433824B2 (en) * 2002-09-04 2008-10-07 Microsoft Corporation Entropy coding by adapting coding between level and run-length/level modes
US7379496B2 (en) * 2002-09-04 2008-05-27 Microsoft Corporation Multi-resolution video coding and decoding
DE60330198D1 (en) 2002-09-04 2009-12-31 Microsoft Corp Entropic coding by adapting the coding mode between level and run length level mode
EP1557026A1 (en) * 2002-10-18 2005-07-27 Koninklijke Philips Electronics N.V. Storage
FR2846179B1 (en) 2002-10-21 2005-02-04 Medialive ADAPTIVE AND PROGRESSIVE STRIP OF AUDIO STREAMS
US8204079B2 (en) * 2002-10-28 2012-06-19 Qualcomm Incorporated Joint transmission of multiple multimedia streams
JP4017498B2 (en) * 2002-11-05 2007-12-05 松下電器産業株式会社 Imaging device
JP2004159191A (en) * 2002-11-07 2004-06-03 Seiko Epson Corp Conversion of frame rate in accordance with image data
US7447264B2 (en) * 2002-11-07 2008-11-04 Victor Company Of Japan, Ltd. Moving-picture temporal scalable coding method, coding apparatus, decoding method, decoding apparatus, and computer program therefor
US7720999B2 (en) * 2002-11-26 2010-05-18 Qualcomm Incorporated System and method for optimizing multimedia compression using plural encoders
EP1573627A2 (en) * 2002-12-13 2005-09-14 Koninklijke Philips Electronics N.V. Method for distributing information content
FR2849980B1 (en) * 2003-01-15 2005-04-08 Medialive METHOD FOR THE DISTRIBUTION OF VIDEO SEQUENCES, DECODER AND SYSTEM FOR THE IMPLEMENTATION OF THIS PRODUCT
US7466870B2 (en) * 2003-01-16 2008-12-16 Acoustic Technology Llc Apparatus and method for creating effects in video
JP4340455B2 (en) * 2003-03-10 2009-10-07 富士通株式会社 Image processing program creation method, image processing program creation program, and image processing program creation device
US8978090B2 (en) * 2003-03-12 2015-03-10 Qualcomm Incorporated Multimedia transcoding proxy server for wireless telecommunication system
KR20060109247A (en) * 2005-04-13 2006-10-19 엘지전자 주식회사 Method and apparatus for encoding/decoding a video signal using pictures of base layer
KR20060105409A (en) * 2005-04-01 2006-10-11 엘지전자 주식회사 Method for scalably encoding and decoding video signal
US8761252B2 (en) * 2003-03-27 2014-06-24 Lg Electronics Inc. Method and apparatus for scalably encoding and decoding video signal
FR2853786B1 (en) * 2003-04-11 2005-08-05 Medialive METHOD AND EQUIPMENT FOR DISTRIBUTING DIGITAL VIDEO PRODUCTS WITH A RESTRICTION OF CERTAIN AT LEAST REPRESENTATION AND REPRODUCTION RIGHTS
FR2853792A1 (en) * 2003-04-11 2004-10-15 France Telecom Digital video sequence tattooing process, involves selecting optimal tattooed displacement vector, based on preset criteria, such that modified coordinates of displacement vector are coordinates of optimal vector
CN100483992C (en) * 2003-05-06 2009-04-29 国际商业机器公司 Encrypting and deencrypting method and apparatus for data flow
KR100948381B1 (en) * 2003-05-15 2010-03-22 삼성전자주식회사 Image Watermarking Method Using Human Visual System
US7408986B2 (en) * 2003-06-13 2008-08-05 Microsoft Corporation Increasing motion smoothness using frame interpolation with motion analysis
US7380028B2 (en) * 2003-06-13 2008-05-27 Microsoft Corporation Robust delivery of video data
US7558320B2 (en) * 2003-06-13 2009-07-07 Microsoft Corporation Quality control in frame interpolation with motion analysis
TWI235003B (en) * 2003-06-17 2005-06-21 Inst Information Industry Digital video encryption and decryption system and method thereof
US7343291B2 (en) 2003-07-18 2008-03-11 Microsoft Corporation Multi-pass variable bitrate media encoding
US8218624B2 (en) * 2003-07-18 2012-07-10 Microsoft Corporation Fractional quantization step sizes for high bit rates
US7609763B2 (en) * 2003-07-18 2009-10-27 Microsoft Corporation Advanced bi-directional predictive coding of video frames
US7426308B2 (en) * 2003-07-18 2008-09-16 Microsoft Corporation Intraframe and interframe interlace coding and decoding
US7383180B2 (en) * 2003-07-18 2008-06-03 Microsoft Corporation Constant bitrate media encoding techniques
US20050013498A1 (en) 2003-07-18 2005-01-20 Microsoft Corporation Coding of motion vector information
US7738554B2 (en) 2003-07-18 2010-06-15 Microsoft Corporation DC coefficient signaling at small quantization step sizes
US7499495B2 (en) * 2003-07-18 2009-03-03 Microsoft Corporation Extended range motion vectors
US10554985B2 (en) 2003-07-18 2020-02-04 Microsoft Technology Licensing, Llc DC coefficient signaling at small quantization step sizes
US7580584B2 (en) * 2003-07-18 2009-08-25 Microsoft Corporation Adaptive multiple quantization
US7602851B2 (en) * 2003-07-18 2009-10-13 Microsoft Corporation Intelligent differential quantization of video coding
US7620106B2 (en) * 2003-09-07 2009-11-17 Microsoft Corporation Joint coding and decoding of a reference field selection and differential motion vector information
US7567617B2 (en) * 2003-09-07 2009-07-28 Microsoft Corporation Predicting motion vectors for fields of forward-predicted interlaced video frames
US7724827B2 (en) * 2003-09-07 2010-05-25 Microsoft Corporation Multi-layer run level encoding and decoding
US8345754B2 (en) * 2003-09-07 2013-01-01 Microsoft Corporation Signaling buffer fullness
US7782954B2 (en) * 2003-09-07 2010-08-24 Microsoft Corporation Scan patterns for progressive video content
US8107531B2 (en) * 2003-09-07 2012-01-31 Microsoft Corporation Signaling and repeat padding for skip frames
US7688894B2 (en) * 2003-09-07 2010-03-30 Microsoft Corporation Scan patterns for interlaced video content
US7609762B2 (en) * 2003-09-07 2009-10-27 Microsoft Corporation Signaling for entry point frames with predicted first field
US7839930B2 (en) * 2003-11-13 2010-11-23 Microsoft Corporation Signaling valid entry points in a video stream
US8582659B2 (en) 2003-09-07 2013-11-12 Microsoft Corporation Determining a decoding time stamp from buffer fullness
US8213779B2 (en) * 2003-09-07 2012-07-03 Microsoft Corporation Trick mode elementary stream and receiver system
US7599438B2 (en) * 2003-09-07 2009-10-06 Microsoft Corporation Motion vector block pattern coding and decoding
US7852919B2 (en) * 2003-09-07 2010-12-14 Microsoft Corporation Field start code for entry point frames with predicted first field
US7317839B2 (en) * 2003-09-07 2008-01-08 Microsoft Corporation Chroma motion vector derivation for interlaced forward-predicted fields
US8064520B2 (en) * 2003-09-07 2011-11-22 Microsoft Corporation Advanced bi-directional predictive coding of interlaced video
US7924921B2 (en) * 2003-09-07 2011-04-12 Microsoft Corporation Signaling coding and display options in entry point headers
US7623574B2 (en) * 2003-09-07 2009-11-24 Microsoft Corporation Selecting between dominant and non-dominant motion vector predictor polarities
US7961786B2 (en) * 2003-09-07 2011-06-14 Microsoft Corporation Signaling field type information
US7577200B2 (en) * 2003-09-07 2009-08-18 Microsoft Corporation Extended range variable length coding/decoding of differential motion vector information
US7822123B2 (en) * 2004-10-06 2010-10-26 Microsoft Corporation Efficient repeat padding for hybrid video sequence with arbitrary video resolution
US7440613B2 (en) * 2003-09-12 2008-10-21 Sony Corporation Binary mask interpolation
US20050094030A1 (en) * 2003-11-03 2005-05-05 Lsi Logic Corporation Method and/or circuitry for video frame rate and/or size conversion
EP1538844A3 (en) * 2003-11-26 2006-05-31 Samsung Electronics Co., Ltd. Color image residue transformation and encoding method
US20050120340A1 (en) * 2003-12-01 2005-06-02 Skazinski Joseph G. Apparatus, system, and method for automated generation of embedded systems software
US8472792B2 (en) 2003-12-08 2013-06-25 Divx, Llc Multimedia distribution system
US7519274B2 (en) * 2003-12-08 2009-04-14 Divx, Inc. File format for multiple track digital data
WO2005074171A2 (en) 2004-01-28 2005-08-11 Qualcomm Incorporated A method for transmitting multiple streams in wireless broadcast networks
KR100987776B1 (en) * 2004-02-04 2010-10-13 삼성전자주식회사 Scrambling apparatus and method using the conversion of motion vector information of video data
US7649539B2 (en) * 2004-03-10 2010-01-19 Microsoft Corporation Image formats for video capture, processing and display
EP1574995A1 (en) * 2004-03-12 2005-09-14 Thomson Licensing S.A. Method for encoding interlaced digital video data
JP4475396B2 (en) * 2004-03-25 2010-06-09 日本電気株式会社 Data broadcasting system and format conversion method
EP1583364A1 (en) * 2004-03-30 2005-10-05 Matsushita Electric Industrial Co., Ltd. Motion compensated interpolation of images at image borders for frame rate conversion
DE102004016350A1 (en) * 2004-04-02 2005-10-27 Micronas Gmbh Method and device for interpolating a pixel of an intermediate line of a field
US20050232497A1 (en) * 2004-04-15 2005-10-20 Microsoft Corporation High-fidelity transcoding
CA2565670A1 (en) * 2004-05-04 2005-11-17 Qualcomm Incorporated Method and apparatus for motion compensated frame rate up conversion
US7487193B2 (en) * 2004-05-14 2009-02-03 Microsoft Corporation Fast video codec transform implementations
US7801383B2 (en) * 2004-05-15 2010-09-21 Microsoft Corporation Embedded scalar quantizers with arbitrary dead-zone ratios
US20050259729A1 (en) * 2004-05-21 2005-11-24 Shijun Sun Video coding with quality scalability
US7912310B2 (en) * 2004-05-21 2011-03-22 Broadcom Corporation System and method for scaling multiple channel and interlaced images with different offsets between channels
KR101014667B1 (en) * 2004-05-27 2011-02-16 삼성전자주식회사 Video encoding, decoding apparatus and method
EP1774779A2 (en) * 2004-07-01 2007-04-18 QUALCOMM Incorporated Method and apparatus for using frame rate up conversion techniques in scalable video coding
US8340177B2 (en) * 2004-07-12 2012-12-25 Microsoft Corporation Embedded base layer codec for 3D sub-band coding
US8442108B2 (en) * 2004-07-12 2013-05-14 Microsoft Corporation Adaptive updates in motion-compensated temporal filtering
US20080075166A1 (en) * 2004-07-13 2008-03-27 Dolby Laboratories Licensing Corporation Unbiased Rounding for Video Compression
US8374238B2 (en) * 2004-07-13 2013-02-12 Microsoft Corporation Spatial scalability in 3D sub-band decoding of SDMCTF-encoded video
BRPI0513527A (en) 2004-07-20 2008-05-06 Qualcomm Inc Method and Equipment for Video Frame Compression Assisted Frame Rate Upward Conversion (EA-FRUC)
US8553776B2 (en) * 2004-07-21 2013-10-08 QUALCOMM Inorporated Method and apparatus for motion vector assignment
US7639892B2 (en) * 2004-07-26 2009-12-29 Sheraizin Semion M Adaptive image improvement
US7903902B2 (en) 2004-07-26 2011-03-08 Sheraizin Semion M Adaptive image improvement
KR100679018B1 (en) * 2004-09-07 2007-02-05 삼성전자주식회사 Method for multi-layer video coding and decoding, multi-layer video encoder and decoder
US8243820B2 (en) * 2004-10-06 2012-08-14 Microsoft Corporation Decoding variable coded resolution video with native range/resolution post-processing operation
US7839933B2 (en) * 2004-10-06 2010-11-23 Microsoft Corporation Adaptive vertical macroblock alignment for mixed frame video sequences
US9071847B2 (en) * 2004-10-06 2015-06-30 Microsoft Technology Licensing, Llc Variable coding resolution in video codec
KR100679022B1 (en) * 2004-10-18 2007-02-05 삼성전자주식회사 Video coding and decoding method using inter-layer filtering, video ecoder and decoder
BRPI0516614B1 (en) * 2004-10-18 2020-08-18 Interdigital Vc Holdings, Inc FILM GRANULATION SIMULATION METHOD
WO2006043753A1 (en) * 2004-10-18 2006-04-27 Samsung Electronics Co., Ltd. Method and apparatus for predecoding hybrid bitstream
JP4764883B2 (en) * 2004-10-18 2011-09-07 トムソン ライセンシング Film grain simulation method, apparatus and system
KR100836217B1 (en) * 2004-10-20 2008-06-09 한국전자통신연구원 Execution Method, Execution Device of Content Comprise Watermark, and Contents Packet Storing Medium
WO2006047138A2 (en) * 2004-10-21 2006-05-04 Thomson Licensing Technique for adaptive de-blocking of block-based film grain patterns
KR100647294B1 (en) 2004-11-09 2006-11-23 삼성전자주식회사 Method and apparatus for encoding and decoding image data
EP1813118B1 (en) * 2004-11-12 2019-12-04 InterDigital VC Holdings, Inc. Film grain simulation for normal play and trick mode play for video playback systems
US20060104356A1 (en) * 2004-11-15 2006-05-18 Microsoft Corporation Timing for decoder buffer examination
CA2587118C (en) 2004-11-16 2014-12-30 Thomson Licensing Film grain sei message insertion for bit-accurate simulation in a video system
MX2007005652A (en) * 2004-11-16 2007-06-05 Thomson Licensing Film grain simulation method based on pre-computed transform coefficients.
ES2739452T3 (en) * 2004-11-17 2020-01-31 Interdigital Vc Holdings Inc Film grain simulation method with bit accuracy based on previously calculated transformed coefficients
KR101208158B1 (en) * 2004-11-22 2012-12-05 톰슨 라이센싱 Methods, apparatus and system for film grain cache splitting for film grain simulation
WO2006057703A1 (en) * 2004-11-23 2006-06-01 Thomson Licensing Low-complexity film grain simulation technique
US7570782B2 (en) * 2004-12-10 2009-08-04 Hewlett-Packard Development Company, L.P. Method and system for watermarking
FR2879878B1 (en) * 2004-12-22 2007-05-25 Thales Sa COMPATIBLE SELECTIVE ENCRYPTION METHOD FOR VIDEO STREAM
US20090064242A1 (en) * 2004-12-23 2009-03-05 Bitband Technologies Ltd. Fast channel switching for digital tv
EP1675399A3 (en) * 2004-12-23 2009-04-29 Bitband Technologies Ltd. Fast channel switching for digital TV
US7680181B1 (en) * 2005-01-07 2010-03-16 Apple Inc. Spooling during video format conversions
US20060159352A1 (en) * 2005-01-18 2006-07-20 Faisal Ishtiaq Method and apparatus for encoding a video sequence
US7542095B2 (en) * 2005-01-20 2009-06-02 Samsung Electronics Co., Ltd. Method and system of noise-adaptive motion detection in an interlaced video sequence
WO2006082150A2 (en) * 2005-02-07 2006-08-10 Thomson Licensing METHOD AND APPARATUS FOR REPLAYING A VIDEO SIGNAL AND ONE OR MORE AUDIO SIGNALS RELATED TO AUDIO/VIDEO DATA THAT ARE BASED ON A 24Hz FRAME FREQUENCY VIDEO SIGNAL
KR100703751B1 (en) * 2005-02-14 2007-04-06 삼성전자주식회사 Method and apparatus for encoding and decoding referencing virtual area image
KR101233854B1 (en) * 2005-02-18 2013-02-15 톰슨 라이센싱 Method for deriving coding information for high resolution pictures from low resolution pictures and coding and decoding devices implementing said method
US7526142B2 (en) * 2005-02-22 2009-04-28 Sheraizin Vitaly S Enhancement of decompressed video
DE102005009852B3 (en) * 2005-03-03 2006-06-29 Siemens Ag Device for receiving and managing medical graphic data has one or more computer devices whereby at least one personal computer and image requesting activity of personal computer and loading time form outgoing network traffic at server
CN1328910C (en) * 2005-03-10 2007-07-25 北京中星微电子有限公司 Method for reading search window data for motion estimation by hardware
US8175168B2 (en) * 2005-03-18 2012-05-08 Sharp Laboratories Of America, Inc. Methods and systems for picture up-sampling
US8638862B2 (en) * 2005-03-18 2014-01-28 Sharp Laboratories Of America, Inc. Methods and systems for upsampling filter design
US7961963B2 (en) * 2005-03-18 2011-06-14 Sharp Laboratories Of America, Inc. Methods and systems for extended spatial scalability with picture-level adaptation
US20060222073A1 (en) * 2005-03-29 2006-10-05 Guillaume Mercier Authoring running marks in compressed data
CN101176346A (en) * 2005-04-01 2008-05-07 Lg电子株式会社 Method for scalably encoding and decoding video signal
US8660180B2 (en) * 2005-04-01 2014-02-25 Lg Electronics Inc. Method and apparatus for scalably encoding and decoding video signal
EP1878249B1 (en) * 2005-04-01 2020-03-04 LG Electronics, Inc. Method for scalably decoding a video signal
KR100883604B1 (en) * 2005-04-01 2009-02-13 엘지전자 주식회사 Method for scalably encoding and decoding video signal
WO2006104366A1 (en) * 2005-04-01 2006-10-05 Lg Electronics Inc. Method for scalably encoding and decoding video signal
KR100878825B1 (en) * 2005-04-01 2009-01-14 엘지전자 주식회사 Method for scalably encoding and decoding video signal
US20060233258A1 (en) * 2005-04-15 2006-10-19 Microsoft Corporation Scalable motion estimation
US7751484B2 (en) * 2005-04-27 2010-07-06 Lsi Corporation Method for composite video artifacts reduction
US8422546B2 (en) 2005-05-25 2013-04-16 Microsoft Corporation Adaptive video encoding using a perceptual model
US20070127909A1 (en) * 2005-08-25 2007-06-07 Craig Mowry System and apparatus for increasing quality and efficiency of film capture and methods of use thereof
US7693709B2 (en) 2005-07-15 2010-04-06 Microsoft Corporation Reordering coefficients for waveform coding or decoding
US7599840B2 (en) * 2005-07-15 2009-10-06 Microsoft Corporation Selectively using multiple entropy models in adaptive coding and decoding
US7684981B2 (en) * 2005-07-15 2010-03-23 Microsoft Corporation Prediction of spectral coefficients in waveform coding and decoding
US7546240B2 (en) * 2005-07-15 2009-06-09 Microsoft Corporation Coding with improved time resolution for selected segments via adaptive block transformation of a group of samples from a subband decomposition
US20070014484A1 (en) * 2005-07-18 2007-01-18 Alexander Maclnnis Method and system for adaptive film grain noise processing
US8755434B2 (en) * 2005-07-22 2014-06-17 Lg Electronics Inc. Method and apparatus for scalably encoding and decoding video signal
US7933337B2 (en) * 2005-08-12 2011-04-26 Microsoft Corporation Prediction of transform coefficients for image compression
US8599925B2 (en) * 2005-08-12 2013-12-03 Microsoft Corporation Efficient coding and decoding of transform blocks
US7565018B2 (en) * 2005-08-12 2009-07-21 Microsoft Corporation Adaptive coding and decoding of wide-range coefficients
US9077960B2 (en) * 2005-08-12 2015-07-07 Microsoft Corporation Non-zero coefficient block pattern coding
WO2007020278A2 (en) * 2005-08-18 2007-02-22 Thomson Licensing Method for encoding and decoding high-resolution progressive and interleave low-resolution images
US7995753B2 (en) * 2005-08-29 2011-08-09 Cisco Technology, Inc. Parallel cipher operations using a single data pass
US8160160B2 (en) * 2005-09-09 2012-04-17 Broadcast International, Inc. Bit-rate reduction for multimedia data streams
NZ566935A (en) * 2005-09-27 2010-02-26 Qualcomm Inc Methods and apparatus for service acquisition
US8229983B2 (en) * 2005-09-27 2012-07-24 Qualcomm Incorporated Channel switch frame
CN101313579B (en) * 2005-09-29 2010-11-03 意大利电信股份公司 Scalable video coding method
US7689052B2 (en) * 2005-10-07 2010-03-30 Microsoft Corporation Multimedia signal processing using fixed-point approximations of linear transforms
US20070085937A1 (en) * 2005-10-13 2007-04-19 Boss Gregory J Internal light masking in projection systems
KR100763194B1 (en) * 2005-10-14 2007-10-04 삼성전자주식회사 Intra base prediction method satisfying single loop decoding condition, video coding method and apparatus using the prediction method
KR100873636B1 (en) 2005-11-14 2008-12-12 삼성전자주식회사 Method and apparatus for encoding/decoding image using single coding mode
KR100662891B1 (en) 2005-11-30 2007-01-02 삼성전자주식회사 Apparatus for processing noise using motion information and method thereof
US8340098B2 (en) * 2005-12-07 2012-12-25 General Instrument Corporation Method and apparatus for delivering compressed video to subscriber terminals
EP1969853A1 (en) * 2006-01-05 2008-09-17 Thomson Licensing Inter-layer motion prediction method
US7956930B2 (en) * 2006-01-06 2011-06-07 Microsoft Corporation Resampling and picture resizing operations for multi-resolution video coding and decoding
US8401091B2 (en) * 2006-01-09 2013-03-19 Lg Electronics Inc. Inter-layer prediction method for video signal
KR20080092425A (en) * 2006-01-10 2008-10-15 노키아 코포레이션 Switched filter up-sampling mechanism for scalable video coding
US9196304B2 (en) * 2006-01-26 2015-11-24 Sony Corporation Method and system for providing dailies and edited video to users
US8166501B2 (en) * 2006-01-26 2012-04-24 Sony Corporation Scheme for use with client device interface in system for providing dailies and edited video to users
CN101375315B (en) * 2006-01-27 2015-03-18 图象公司 Methods and systems for digitally re-mastering of 2D and 3D motion pictures for exhibition with enhanced visual quality
US20090307732A1 (en) * 2006-03-07 2009-12-10 Noam Cohen Personalized Insertion of Advertisements in Streaming Media
US7515710B2 (en) 2006-03-14 2009-04-07 Divx, Inc. Federated digital rights management scheme including trusted systems
US20070217603A1 (en) * 2006-03-15 2007-09-20 Terayon Communication Systems, Inc. Decryption key reuse in encrypted digital data stream distribution systems
CN100584026C (en) * 2006-03-27 2010-01-20 华为技术有限公司 Video layering coding method at interleaving mode
US20070230564A1 (en) * 2006-03-29 2007-10-04 Qualcomm Incorporated Video processing with scalability
US8634463B2 (en) * 2006-04-04 2014-01-21 Qualcomm Incorporated Apparatus and method of enhanced frame interpolation in video compression
US8750387B2 (en) * 2006-04-04 2014-06-10 Qualcomm Incorporated Adaptive encoder-assisted frame rate up conversion
US8155195B2 (en) * 2006-04-07 2012-04-10 Microsoft Corporation Switching distortion metrics during motion estimation
US8059721B2 (en) 2006-04-07 2011-11-15 Microsoft Corporation Estimating sample-domain distortion in the transform domain with rounding compensation
US8130828B2 (en) 2006-04-07 2012-03-06 Microsoft Corporation Adjusting quantization to preserve non-zero AC coefficients
US8494052B2 (en) * 2006-04-07 2013-07-23 Microsoft Corporation Dynamic selection of motion estimation search ranges and extended motion vector ranges
US7995649B2 (en) 2006-04-07 2011-08-09 Microsoft Corporation Quantization adjustment based on texture level
US7974340B2 (en) * 2006-04-07 2011-07-05 Microsoft Corporation Adaptive B-picture quantization control
US8503536B2 (en) * 2006-04-07 2013-08-06 Microsoft Corporation Quantization adjustments for DC shift artifacts
US8009861B2 (en) * 2006-04-28 2011-08-30 Vobile, Inc. Method and system for fingerprinting digital video object based on multiresolution, multirate spatial and temporal signatures
US8880571B2 (en) * 2006-05-05 2014-11-04 Microsoft Corporation High dynamic range data format conversions for digital media
US8711925B2 (en) 2006-05-05 2014-04-29 Microsoft Corporation Flexible quantization
US7643078B2 (en) * 2006-05-19 2010-01-05 Eastman Kodak Company CCD with improved charge transfer
US20070268964A1 (en) * 2006-05-22 2007-11-22 Microsoft Corporation Unit co-location-based motion estimation
CN101479765B (en) 2006-06-23 2012-05-23 图象公司 Methods and systems for converting 2d motion pictures for stereoscopic 3d exhibition
US7652250B2 (en) * 2006-06-26 2010-01-26 Matthew Erdtmann Noise reduction method for imaging devices
EP1879399A1 (en) * 2006-07-12 2008-01-16 THOMSON Licensing Method for deriving motion data for high resolution pictures from motion data of low resolution pictures and coding and decoding devices implementing said method
US8196169B1 (en) 2006-09-18 2012-06-05 Nvidia Corporation Coordinate-based set top box policy enforcement system, method and computer program product
JP2008109247A (en) * 2006-10-24 2008-05-08 Matsushita Electric Ind Co Ltd Method and device for filtering video noise, integrated circuit, and encoder
JP4303745B2 (en) * 2006-11-07 2009-07-29 シャープ株式会社 Image display apparatus and method, image processing apparatus and method
KR101089072B1 (en) * 2006-11-14 2011-12-09 퀄컴 인코포레이티드 Systems and methods for channel switching
WO2008061211A2 (en) * 2006-11-15 2008-05-22 Qualcomm Incorporated Systems and methods for applications using channel switch frames
US8134640B2 (en) * 2006-12-26 2012-03-13 Broadcom Corporation Video processor architecture and method for frame rate conversion
CN103561278B (en) 2007-01-05 2017-04-12 索尼克知识产权股份有限公司 Video distribution system including progressive playback
US7982800B2 (en) * 2007-02-02 2011-07-19 Freescale Semiconductor, Inc. Video de-interlacer using motion residue compensation
US8265157B2 (en) * 2007-02-07 2012-09-11 Lsi Corporation Motion vector refinement for MPEG-2 to H.264 video transcoding
US8238424B2 (en) 2007-02-09 2012-08-07 Microsoft Corporation Complexity-based adaptive preprocessing for multiple-pass video compression
US8942289B2 (en) * 2007-02-21 2015-01-27 Microsoft Corporation Computational complexity and precision control in transform-based digital media codec
US8054886B2 (en) 2007-02-21 2011-11-08 Microsoft Corporation Signaling and use of chroma sample positioning information
US8184710B2 (en) * 2007-02-21 2012-05-22 Microsoft Corporation Adaptive truncation of transform coefficient data in a transform-based digital media codec
US8189107B1 (en) * 2007-03-12 2012-05-29 Nvidia Corporation System and method for performing visual data post-processing based on information related to frequency response pre-processing
US8107571B2 (en) * 2007-03-20 2012-01-31 Microsoft Corporation Parameterized filters and signaling techniques
US8498335B2 (en) * 2007-03-26 2013-07-30 Microsoft Corporation Adaptive deadzone size adjustment in quantization
US8243797B2 (en) 2007-03-30 2012-08-14 Microsoft Corporation Regions of interest for quality adjustments
WO2008140656A2 (en) * 2007-04-03 2008-11-20 Gary Demos Flowfield motion compensation for video compression
US8115863B2 (en) * 2007-04-04 2012-02-14 Freescale Semiconductor, Inc. Video de-interlacer using pixel trajectory
US8442337B2 (en) * 2007-04-18 2013-05-14 Microsoft Corporation Encoding adjustments for animation content
US10715834B2 (en) * 2007-05-10 2020-07-14 Interdigital Vc Holdings, Inc. Film grain simulation based on pre-computed transform coefficients
US8331438B2 (en) 2007-06-05 2012-12-11 Microsoft Corporation Adaptive selection of picture-level quantization parameters for predicted video pictures
JP4609457B2 (en) * 2007-06-14 2011-01-12 ソニー株式会社 Image processing apparatus and image processing method
US7774205B2 (en) * 2007-06-15 2010-08-10 Microsoft Corporation Coding of sparse digital media spectral data
US7761290B2 (en) 2007-06-15 2010-07-20 Microsoft Corporation Flexible frequency and time partitioning in perceptual transform coding of audio
JP2009016934A (en) * 2007-06-29 2009-01-22 Toshiba Corp Video reproducing device
US8254455B2 (en) * 2007-06-30 2012-08-28 Microsoft Corporation Computing collocated macroblock information for direct mode macroblocks
WO2009034489A2 (en) * 2007-09-10 2009-03-19 Nxp B.V. Method and apparatus for motion estimation in video image data
US8369422B2 (en) * 2007-10-16 2013-02-05 Thomson Licensing Methods and apparatus for artifact removal for bit depth scalability
US8457958B2 (en) 2007-11-09 2013-06-04 Microsoft Corporation Audio transcoder using encoder-generated side information to transcode to target bit-rate
WO2009065137A1 (en) 2007-11-16 2009-05-22 Divx, Inc. Hierarchical and reduced index structures for multimedia files
CN101878486B (en) * 2007-11-30 2016-10-26 杜比实验室特许公司 The method and apparatus of processing moving sequence
JP5209945B2 (en) * 2007-12-12 2013-06-12 株式会社日立製作所 Storage device, encryption content validation method, and terminal device
US8750390B2 (en) * 2008-01-10 2014-06-10 Microsoft Corporation Filtering and dithering as pre-processing before encoding
WO2009087641A2 (en) * 2008-01-10 2009-07-16 Ramot At Tel-Aviv University Ltd. System and method for real-time super-resolution
US8700792B2 (en) * 2008-01-31 2014-04-15 General Instrument Corporation Method and apparatus for expediting delivery of programming content over a broadband network
US8160132B2 (en) 2008-02-15 2012-04-17 Microsoft Corporation Reducing key picture popping effects in video
JP4618305B2 (en) * 2008-02-19 2011-01-26 ソニー株式会社 Image processing apparatus, image processing method, and program
US8953673B2 (en) * 2008-02-29 2015-02-10 Microsoft Corporation Scalable video coding and decoding with sample bit depth and chroma high-pass residual layers
US20090225867A1 (en) * 2008-03-06 2009-09-10 Lee Kun-Bin Methods and apparatus for picture access
US8711948B2 (en) * 2008-03-21 2014-04-29 Microsoft Corporation Motion-compensated prediction of inter-layer residuals
US8155436B2 (en) * 2008-03-31 2012-04-10 Konica Minolta Laboratory U.S.A., Inc. Systems and methods for color data compression
US20090244601A1 (en) * 2008-03-31 2009-10-01 Konica Minolta Systems Laboratory, Inc. Systems and Methods for Color Data Compression
US8189933B2 (en) * 2008-03-31 2012-05-29 Microsoft Corporation Classifying and controlling encoding quality for textured, dark smooth and smooth video content
US8164862B2 (en) * 2008-04-02 2012-04-24 Headway Technologies, Inc. Seed layer for TMR or CPP-GMR sensor
US8179974B2 (en) 2008-05-02 2012-05-15 Microsoft Corporation Multi-level representation of reordered transform coefficients
US8325800B2 (en) 2008-05-07 2012-12-04 Microsoft Corporation Encoding streaming media as a high bit rate layer, a low bit rate layer, and one or more intermediate bit rate layers
US8379851B2 (en) 2008-05-12 2013-02-19 Microsoft Corporation Optimized client side rate control and indexed file layout for streaming media
WO2010055893A1 (en) * 2008-11-13 2010-05-20 Oya Nagato Method for processing data in tree form and device for processing data
US7860996B2 (en) * 2008-05-30 2010-12-28 Microsoft Corporation Media streaming with seamless ad insertion
JP5369893B2 (en) 2008-05-30 2013-12-18 株式会社Jvcケンウッド Video encoding device, video encoding method, video encoding program, video decoding device, video decoding method, video decoding program, video re-encoding device, video re-encoding method, video re-encoding Encoding program
US8897359B2 (en) 2008-06-03 2014-11-25 Microsoft Corporation Adaptive quantization for enhancement layer video coding
US8752092B2 (en) * 2008-06-27 2014-06-10 General Instrument Corporation Method and apparatus for providing low resolution images in a broadcast system
US20090328093A1 (en) * 2008-06-30 2009-12-31 At&T Intellectual Property I, L.P. Multimedia Content Filtering
US9445121B2 (en) 2008-08-04 2016-09-13 Dolby Laboratories Licensing Corporation Overlapped block disparity estimation and compensation architecture
US8406307B2 (en) 2008-08-22 2013-03-26 Microsoft Corporation Entropy coding/decoding of hierarchically organized data
US9571856B2 (en) 2008-08-25 2017-02-14 Microsoft Technology Licensing, Llc Conversion operations in scalable video encoding and decoding
US8213503B2 (en) 2008-09-05 2012-07-03 Microsoft Corporation Skip modes for inter-layer residual video coding and decoding
US8189776B2 (en) * 2008-09-18 2012-05-29 The Hong Kong University Of Science And Technology Method and system for encoding multimedia content based on secure coding schemes using stream cipher
US20100074341A1 (en) * 2008-09-19 2010-03-25 Wade Wan Method and system for multiple resolution video delivery
US8804821B2 (en) 2008-09-26 2014-08-12 Microsoft Corporation Adaptive video processing of an interactive environment
US8243117B2 (en) * 2008-09-26 2012-08-14 Microsoft Corporation Processing aspects of a video scene
US8265140B2 (en) * 2008-09-30 2012-09-11 Microsoft Corporation Fine-grained client-side control of scalable media delivery
JP5262546B2 (en) * 2008-10-08 2013-08-14 ソニー株式会社 Video signal processing system, playback device and display device, and video signal processing method
US9762973B2 (en) 2008-11-04 2017-09-12 The Directv Group, Inc. Method and system for operating a receiving circuit module to encode a channel signal into multiple encoding formats
US8774541B2 (en) * 2008-11-05 2014-07-08 Sony Corporation Intra prediction with adaptive interpolation filtering for image compression
JP5300438B2 (en) * 2008-11-21 2013-09-25 キヤノン株式会社 Image processing apparatus, image processing method, and program
US8311115B2 (en) * 2009-01-29 2012-11-13 Microsoft Corporation Video encoding using previously calculated motion information
US8396114B2 (en) * 2009-01-29 2013-03-12 Microsoft Corporation Multiple bit rate video encoding using variable bit rate and dynamic resolution for adaptive video streaming
US8189666B2 (en) 2009-02-02 2012-05-29 Microsoft Corporation Local picture identifier and computation of co-located information
US8438397B2 (en) * 2009-06-04 2013-05-07 Broadcom Corporation Watermarking for compressed scalable coded bitstreams
US8270473B2 (en) * 2009-06-12 2012-09-18 Microsoft Corporation Motion based dynamic resolution multiple bit rate video encoding
US9930344B2 (en) * 2009-08-11 2018-03-27 Nbcuniversal Media, Llc Digital content integration and delivery system and method
CA2822185C (en) 2009-08-14 2014-04-22 Azuki Systems, Inc. Method and system for unified mobile content protection
GB0917617D0 (en) * 2009-10-08 2009-11-25 Barco Nv Video amplitude modulation
EP2507995A4 (en) 2009-12-04 2014-07-09 Sonic Ip Inc Elementary bitstream cryptographic material transport systems and methods
US8340188B2 (en) * 2010-01-08 2012-12-25 Research In Motion Limited Method and device for motion vector estimation in video transcoding using union of search areas
US20110170608A1 (en) * 2010-01-08 2011-07-14 Xun Shi Method and device for video transcoding using quad-tree based mode selection
US8315310B2 (en) * 2010-01-08 2012-11-20 Research In Motion Limited Method and device for motion vector prediction in video transcoding using full resolution residuals
US8358698B2 (en) * 2010-01-08 2013-01-22 Research In Motion Limited Method and device for motion vector estimation in video transcoding using full-resolution residuals
US8559519B2 (en) * 2010-01-08 2013-10-15 Blackberry Limited Method and device for video encoding using predicted residuals
US8325815B2 (en) * 2010-02-08 2012-12-04 Himax Technologies Limited Method and system of hierarchical motion estimation
US9357244B2 (en) * 2010-03-11 2016-05-31 Arris Enterprises, Inc. Method and system for inhibiting audio-video synchronization delay
JP5667773B2 (en) * 2010-03-18 2015-02-12 キヤノン株式会社 Information creating apparatus and control method thereof
US8588461B2 (en) * 2010-03-22 2013-11-19 Brigham Young University Robust watermarking for digital media
DE112010005418B4 (en) * 2010-03-25 2019-07-11 Nokia Technologies Oy Apparatus, display module and method for adaptably inserting a dummy frame
KR101708931B1 (en) * 2010-04-28 2017-03-08 삼성전자주식회사 Apparatus and method for allocating data rate in a multi antenna system
US8705616B2 (en) 2010-06-11 2014-04-22 Microsoft Corporation Parallel multiple bitrate video encoding to reduce latency and dependences between groups of pictures
US8774267B2 (en) * 2010-07-07 2014-07-08 Spinella Ip Holdings, Inc. System and method for transmission, processing, and rendering of stereoscopic and multi-view images
US9036695B2 (en) * 2010-11-02 2015-05-19 Sharp Laboratories Of America, Inc. Motion-compensated temporal filtering based on variable filter parameters
US8914534B2 (en) 2011-01-05 2014-12-16 Sonic Ip, Inc. Systems and methods for adaptive bitrate streaming of media stored in matroska container files using hypertext transfer protocol
US9831971B1 (en) * 2011-04-05 2017-11-28 The Directv Group, Inc. Method and system for operating a communication system encoded into multiple independently communicated encoding formats
US20120278441A1 (en) * 2011-04-28 2012-11-01 Futurewei Technologies, Inc. System and Method for Quality of Experience Estimation
US8681866B1 (en) 2011-04-28 2014-03-25 Google Inc. Method and apparatus for encoding video by downsampling frame resolution
US9106787B1 (en) 2011-05-09 2015-08-11 Google Inc. Apparatus and method for media transmission bandwidth control using bandwidth estimation
US8812662B2 (en) 2011-06-29 2014-08-19 Sonic Ip, Inc. Systems and methods for estimating available bandwidth and performing initial stream selection when streaming content
CN108989847B (en) 2011-08-30 2021-03-09 帝威视有限公司 System and method for encoding and streaming video
US9467708B2 (en) 2011-08-30 2016-10-11 Sonic Ip, Inc. Selection of resolutions for seamless resolution switching of multimedia content
US8806188B2 (en) 2011-08-31 2014-08-12 Sonic Ip, Inc. Systems and methods for performing adaptive bitrate streaming using automatically generated top level index files
US8799647B2 (en) 2011-08-31 2014-08-05 Sonic Ip, Inc. Systems and methods for application identification
US8964977B2 (en) 2011-09-01 2015-02-24 Sonic Ip, Inc. Systems and methods for saving encoded media streamed using adaptive bitrate streaming
US8909922B2 (en) 2011-09-01 2014-12-09 Sonic Ip, Inc. Systems and methods for playing back alternative streams of protected content protected using common cryptographic information
US9591318B2 (en) * 2011-09-16 2017-03-07 Microsoft Technology Licensing, Llc Multi-layer encoding and decoding
EP2761534B1 (en) 2011-09-28 2020-11-18 FotoNation Limited Systems for encoding light field image files
US10018703B2 (en) * 2012-09-13 2018-07-10 Conduent Business Services, Llc Method for stop sign law enforcement using motion vectors in video streams
US8751800B1 (en) * 2011-12-12 2014-06-10 Google Inc. DRM provider interoperability
US20130179199A1 (en) 2012-01-06 2013-07-11 Rovi Corp. Systems and methods for granting access to digital content using electronic tickets and ticket tokens
US11089343B2 (en) 2012-01-11 2021-08-10 Microsoft Technology Licensing, Llc Capability advertisement, configuration and control for video coding and decoding
US20130188712A1 (en) * 2012-01-24 2013-07-25 Futurewei Technologies, Inc. Compressed Domain Watermarking with Reduced Error Propagation
US9185429B1 (en) 2012-04-30 2015-11-10 Google Inc. Video encoding and decoding using un-equal error protection
US20140003539A1 (en) * 2012-07-02 2014-01-02 Cisco Technology, Inc. Signalling Information for Consecutive Coded Video Sequences that Have the Same Aspect Ratio but Different Picture Resolutions
US9936267B2 (en) 2012-08-31 2018-04-03 Divx Cf Holdings Llc System and method for decreasing an initial buffering period of an adaptive streaming system
CN103702058B (en) * 2012-09-27 2015-09-16 珠海扬智电子科技有限公司 The macroblock status discrimination method of release of an interleave computing and image processor
KR102379609B1 (en) 2012-10-01 2022-03-28 지이 비디오 컴프레션, 엘엘씨 Scalable video coding using base-layer hints for enhancement layer motion parameters
US20140098880A1 (en) * 2012-10-05 2014-04-10 Qualcomm Incorporated Prediction mode information upsampling for scalable video coding
EP2731337B1 (en) 2012-10-17 2017-07-12 Dolby Laboratories Licensing Corporation Systems and methods for transmitting video frames
US9648319B2 (en) * 2012-12-12 2017-05-09 Qualcomm Incorporated Device and method for scalable coding of video information based on high efficiency video coding
US9191457B2 (en) 2012-12-31 2015-11-17 Sonic Ip, Inc. Systems, methods, and media for controlling delivery of content
US9313510B2 (en) 2012-12-31 2016-04-12 Sonic Ip, Inc. Use of objective quality measures of streamed content to reduce streaming bandwidth
US9172740B1 (en) 2013-01-15 2015-10-27 Google Inc. Adjustable buffer remote access
US9311692B1 (en) 2013-01-25 2016-04-12 Google Inc. Scalable buffer remote access
US9225979B1 (en) 2013-01-30 2015-12-29 Google Inc. Remote access encoding
US9002133B2 (en) 2013-02-27 2015-04-07 Sharp Laboratories Of America, Inc. Multi layered image enhancement technique
US9053531B2 (en) 2013-02-28 2015-06-09 Sharp Laboratories Of America, Inc. Multi layered image enhancement technique
US10284842B2 (en) 2013-03-05 2019-05-07 Qualcomm Incorporated Inter-layer reference picture construction for spatial scalability with different aspect ratios
US9906785B2 (en) 2013-03-15 2018-02-27 Sonic Ip, Inc. Systems, methods, and media for transcoding video data according to encoding parameters indicated by received metadata
US10397292B2 (en) 2013-03-15 2019-08-27 Divx, Llc Systems, methods, and media for delivery of content
US9094737B2 (en) 2013-05-30 2015-07-28 Sonic Ip, Inc. Network video streaming with trick play based on separate trick play files
US9380099B2 (en) 2013-05-31 2016-06-28 Sonic Ip, Inc. Synchronizing multiple over the top streaming clients
US9100687B2 (en) 2013-05-31 2015-08-04 Sonic Ip, Inc. Playback synchronization across playback devices
US9083850B1 (en) * 2013-06-29 2015-07-14 Securus Technologies, Inc. Video blurring in a secure environment
US9251549B2 (en) 2013-07-23 2016-02-02 Verance Corporation Watermark extractor enhancements based on payload ranking
WO2015029047A1 (en) * 2013-08-26 2015-03-05 Hewlett-Packard Development Company, L.P. Temporal median filtering to remove shadow
WO2015063537A1 (en) * 2013-11-01 2015-05-07 Intel Corporation Vido modification for content protection
US9386067B2 (en) 2013-12-30 2016-07-05 Sonic Ip, Inc. Systems and methods for playing adaptive bitrate streaming content by multicast
US9942560B2 (en) 2014-01-08 2018-04-10 Microsoft Technology Licensing, Llc Encoding screen capture data
US9749642B2 (en) 2014-01-08 2017-08-29 Microsoft Technology Licensing, Llc Selection of motion vector precision
US9774881B2 (en) 2014-01-08 2017-09-26 Microsoft Technology Licensing, Llc Representing motion vectors in an encoded bitstream
US9842422B1 (en) * 2014-02-18 2017-12-12 Google Inc. Methods, systems, and media for presenting interstitial animations
US9866878B2 (en) 2014-04-05 2018-01-09 Sonic Ip, Inc. Systems and methods for encoding and playing back video at different frame rates using enhancement layers
US10177912B2 (en) * 2014-05-09 2019-01-08 Sony Corporation Content individualization
US9360671B1 (en) * 2014-06-09 2016-06-07 Google Inc. Systems and methods for image zoom
WO2016022979A1 (en) 2014-08-07 2016-02-11 Sonic IP. Inc. Systems and methods for protecting elementary bitstreams incorporating independently encoded tiles
US10078887B2 (en) 2014-10-20 2018-09-18 Bae Systems Information And Electronic Systems Integration Inc. Scene-based non-uniformity correction in focal plane arrays
US10136133B2 (en) * 2014-11-11 2018-11-20 Dolby Laboratories Licensing Corporation Rate control adaptation for high-dynamic range images
EP3910904A1 (en) 2015-01-06 2021-11-17 DivX, LLC Systems and methods for encoding and sharing content between devices
US9877036B2 (en) 2015-01-15 2018-01-23 Gopro, Inc. Inter frame watermark in a digital video
WO2016114950A1 (en) * 2015-01-15 2016-07-21 Gopro, Inc. Watermarking digital images to increase bit dept
US9886961B2 (en) 2015-01-15 2018-02-06 Gopro, Inc. Audio watermark in a digital video
US9418396B2 (en) 2015-01-15 2016-08-16 Gopro, Inc. Watermarking digital images to increase bit depth
SG11201706160UA (en) 2015-02-27 2017-09-28 Sonic Ip Inc Systems and methods for frame duplication and frame extension in live video encoding and streaming
CN106033418B (en) 2015-03-10 2020-01-31 阿里巴巴集团控股有限公司 Voice adding and playing method and device, and picture classifying and retrieving method and device
US9888274B2 (en) 2015-04-21 2018-02-06 Edge2020, Llc Price driven multimedia content reception
US20170006219A1 (en) 2015-06-30 2017-01-05 Gopro, Inc. Image stitching in a multi-camera array
US10051206B2 (en) 2015-09-28 2018-08-14 Gopro, Inc. Automatic composition of video with dynamic background and composite frames selected based on frame and foreground object criteria
US10313690B2 (en) * 2015-10-08 2019-06-04 Sharp Kabushiki Kaisha Systems and methods for reducing artifacts in temporal scalable layers of video
US10657634B2 (en) * 2015-12-05 2020-05-19 Indiana University Research And Technology Corporation Systems and methods for image processing
GB2598255B (en) * 2016-02-17 2022-07-27 V Nova Int Ltd Physical adapter, signal processing equipment, methods and computer programs
US10075292B2 (en) 2016-03-30 2018-09-11 Divx, Llc Systems and methods for quick start-up of playback
US10129574B2 (en) 2016-05-24 2018-11-13 Divx, Llc Systems and methods for providing variable speeds in a trick-play mode
US10231001B2 (en) 2016-05-24 2019-03-12 Divx, Llc Systems and methods for providing audio content during trick-play playback
US10148989B2 (en) 2016-06-15 2018-12-04 Divx, Llc Systems and methods for encoding video content
US10045120B2 (en) 2016-06-20 2018-08-07 Gopro, Inc. Associating audio with three-dimensional objects in videos
US9749738B1 (en) 2016-06-20 2017-08-29 Gopro, Inc. Synthesizing audio corresponding to a virtual microphone location
US10134114B2 (en) 2016-09-20 2018-11-20 Gopro, Inc. Apparatus and methods for video image post-processing for segmentation-based interpolation
US10313686B2 (en) 2016-09-20 2019-06-04 Gopro, Inc. Apparatus and methods for compressing video content using adaptive projection selection
US10003768B2 (en) 2016-09-28 2018-06-19 Gopro, Inc. Apparatus and methods for frame interpolation based on spatial considerations
US10554711B2 (en) * 2016-09-29 2020-02-04 Cisco Technology, Inc. Packet placement for scalable video coding schemes
US10419511B1 (en) 2016-10-04 2019-09-17 Zoom Video Communications, Inc. Unique watermark generation and detection during a conference
US10498795B2 (en) 2017-02-17 2019-12-03 Divx, Llc Systems and methods for adaptive switching between multiple content delivery networks during adaptive bitrate streaming
US10489897B2 (en) 2017-05-01 2019-11-26 Gopro, Inc. Apparatus and methods for artifact detection and removal using frame interpolation techniques
RU2721188C2 (en) * 2017-12-14 2020-05-18 Общество с ограниченной ответственностью "Аби Продакшн" Improved contrast and noise reduction on images obtained from cameras
RU2688202C1 (en) * 2018-07-03 2019-05-21 Закрытое акционерное общество "Перспективный мониторинг" Method for hidden marking of digital television signal data stream
EP3700210A1 (en) 2019-02-21 2020-08-26 Ateme Method and apparatus for image encoding
GB2599805B (en) * 2019-03-20 2023-09-06 V Nova Int Ltd Temporal signalling for video coding technology
CA3134561A1 (en) 2019-03-21 2020-09-24 Divx, Llc Systems and methods for multimedia swarms
EP3962082A4 (en) * 2019-06-12 2022-10-05 Sony Group Corporation Image processing device and method
CN110610218B (en) * 2019-08-09 2022-09-13 浙江万里学院 Portrait image two-dimensional code generation method
WO2021054805A1 (en) * 2019-09-20 2021-03-25 한국전자통신연구원 Image encoding/decoding method and device, and recording medium storing bitstream
CN112543322B (en) * 2019-09-20 2022-04-15 杭州海康威视数字技术股份有限公司 Decoding and encoding method, device and equipment
WO2021062624A1 (en) * 2019-09-30 2021-04-08 深圳市大疆创新科技有限公司 Image processing method and apparatus for movable platform, movable platform and medium
MX2022005905A (en) * 2019-11-15 2022-06-24 Hfi Innovation Inc Method and apparatus for signaling horizontal wraparound motion compensation in vr360 video coding.
US20220408114A1 (en) * 2019-11-22 2022-12-22 Sharp Kabushiki Kaisha Systems and methods for signaling tiles and slices in video coding
RU2724969C1 (en) * 2019-12-23 2020-06-29 Акционерное общество "Научно-исследовательский институт телевидения" Method of detecting small objects
KR20220146647A (en) * 2020-05-19 2022-11-01 구글 엘엘씨 Dynamic parameter selection for quality-normalized video transcoding
US20220279204A1 (en) * 2021-02-26 2022-09-01 Qualcomm Incorporated Efficient video encoder architecture
US11743440B2 (en) 2021-04-19 2023-08-29 Apple Inc. Transmission and consumption of multiple image subframes via superframe
GB202107036D0 (en) * 2021-05-17 2021-06-30 V Nova Int Ltd Secure decoder and secure decoding methods
US11756285B2 (en) 2021-06-10 2023-09-12 Bank Of America Corporation Image processing system and method for image noise removal
WO2023135420A1 (en) * 2022-01-12 2023-07-20 V-Nova International Ltd Secure enhancement decoding implementation

Citations (29)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH01140883A (en) 1987-11-27 1989-06-02 Dainippon Printing Co Ltd Data coding method
US4903317A (en) 1986-06-24 1990-02-20 Kabushiki Kaisha Toshiba Image processing apparatus
US5253058A (en) 1992-04-01 1993-10-12 Bell Communications Research, Inc. Efficient coding scheme for multilevel video transmission
US5270813A (en) 1992-07-02 1993-12-14 At&T Bell Laboratories Spatially scalable video coding facilitating the derivation of variable-resolution images
JPH06165150A (en) 1992-11-18 1994-06-10 Matsushita Electric Ind Co Ltd Dynamic picture coding/decoding device
JPH06350995A (en) 1993-06-10 1994-12-22 Casio Comput Co Ltd Moving picture processing method
EP0634871A2 (en) 1993-07-13 1995-01-18 AT&T Corp. Scalable encoding and decoding of high-resolution progressive video
US5387940A (en) 1993-07-07 1995-02-07 Rca Thomson Licensing Corporation Method and apparatus for providing scaleable compressed video signal
CA2127151A1 (en) 1993-09-21 1995-03-22 Atul Puri Spatially scalable video encoding and decoding
US5408270A (en) 1993-06-24 1995-04-18 Massachusetts Institute Of Technology Advanced television system
US5414469A (en) 1991-10-31 1995-05-09 International Business Machines Corporation Motion video compression system with multiresolution features
US5418571A (en) 1991-02-01 1995-05-23 British Telecommunicatons Public Limited Company Decoding of double layer video signals with interpolation replacement on missing data from enhancement layer
JPH07203426A (en) 1993-12-28 1995-08-04 Toshiba Corp Hierarchical coding and decoding device
US5465119A (en) 1991-02-22 1995-11-07 Demografx Pixel interlacing apparatus and method
US5493338A (en) 1991-12-28 1996-02-20 Goldstar Co., Ltd. Scan converter of television receiver and scan converting method thereof
US5519453A (en) 1993-08-06 1996-05-21 U. S. Philips Corporation Method of eliminating interfernce signals from video signals
WO1997028507A1 (en) 1996-01-30 1997-08-07 Demografx Temporal and resolution layering in advanced television
US5828788A (en) 1995-06-29 1998-10-27 Thomson Multimedia, S.A. System for processing data in variable segments and with variable data resolution
US5974159A (en) 1996-03-29 1999-10-26 Sarnoff Corporation Method and apparatus for assessing the visibility of differences between two image sequences
US6028634A (en) 1995-10-27 2000-02-22 Kabushiki Kaisha Toshiba Video encoding and decoding apparatus
US6111975A (en) 1991-03-22 2000-08-29 Sacks; Jack M. Minimum difference processor
US6175592B1 (en) 1997-03-12 2001-01-16 Matsushita Electric Industrial Co., Ltd. Frequency domain filtering for down conversion of a DCT encoded picture
US6252906B1 (en) 1998-07-31 2001-06-26 Thomson Licensing S.A. Decimation of a high definition video signal
WO2001077871A1 (en) 2000-04-07 2001-10-18 Demografx Enhanced temporal and resolution layering in advanced television
US20020003838A1 (en) 1994-06-08 2002-01-10 Toshiya Takahashi Image conversion apparatus
US6442203B1 (en) 1999-11-05 2002-08-27 Demografx System and method for motion compensation and frame rate conversion
US6489956B1 (en) 1998-02-17 2002-12-03 Sun Microsystems, Inc. Graphics system having a super-sampled sample buffer with generation of output pixels using selective adjustment of filtering for implementation of display effects
US6728317B1 (en) 1996-01-30 2004-04-27 Dolby Laboratories Licensing Corporation Moving image compression quality enhancement using displacement filters with negative lobes
US7106322B2 (en) * 2000-01-11 2006-09-12 Sun Microsystems, Inc. Dynamically adjusting a sample-to-pixel filter to compensate for the effects of negative lobes

Family Cites Families (59)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4654696A (en) 1985-04-09 1987-03-31 Grass Valley Group, Inc. Video signal format
US5151784A (en) 1991-04-30 1992-09-29 At&T Bell Laboratories Multiple frame motion estimation
JP2507204B2 (en) 1991-08-30 1996-06-12 松下電器産業株式会社 Video signal encoder
US5198901A (en) 1991-09-23 1993-03-30 Matsushita Electric Corporation Of America Derivation and use of motion vectors in a differential pulse code modulation system
JPH0595540A (en) 1991-09-30 1993-04-16 Sony Corp Dynamic picture encoder
US5231484A (en) 1991-11-08 1993-07-27 International Business Machines Corporation Motion video compression system with adaptive bit allocation and quantization
JP3245977B2 (en) 1992-06-30 2002-01-15 ソニー株式会社 Digital image signal transmission equipment
US5294974A (en) 1992-07-24 1994-03-15 Matsushita Electric Corporation Of America High-definition video encoding system having color-sensitive quantization
JP3545000B2 (en) 1992-11-02 2004-07-21 ソニー株式会社 Image signal encoding device, image signal decoding device
US5717816A (en) 1993-01-13 1998-02-10 Hitachi America Ltd. Method and apparatus for the selection of data for use in VTR trick playback operation in a system using intra-coded video frames
JP2962083B2 (en) 1993-01-22 1999-10-12 松下電器産業株式会社 Resolution conversion method, resolution conversion device, image encoding device, image decoding device
US5426463A (en) 1993-02-22 1995-06-20 Rca Thomson Licensing Corporation Apparatus for controlling quantizing in a video signal compressor
US5877754A (en) 1993-06-16 1999-03-02 Intel Corporation Process, apparatus, and system for color conversion of image signals
WO1995004433A1 (en) 1993-07-30 1995-02-09 British Telecommunications Public Limited Company Processing image data
US5438374A (en) 1993-12-10 1995-08-01 At&T Corp. System and method for filtering video signals
US5442407A (en) 1994-03-22 1995-08-15 Matsushita Electric Corporation Of America Video signal noise reduction system using time-varying filter coefficients
US5764802A (en) * 1994-09-09 1998-06-09 Intel Corporation Encoding image signals using a blur image
GB2301970B (en) 1995-06-06 2000-03-01 Sony Uk Ltd Motion compensated video processing
US5748903A (en) 1995-07-21 1998-05-05 Intel Corporation Encoding images using decode rate control
US5737032A (en) 1995-09-05 1998-04-07 Videotek, Inc. Serial digital video processing with concurrent adjustment in RGB and luminance/color difference
US5835498A (en) 1995-10-05 1998-11-10 Silicon Image, Inc. System and method for sending multiple data signals over a serial link
KR100209793B1 (en) 1995-10-28 1999-07-15 전주범 Apparatus for encoding/decoding a video signals by using feature point based motion estimation
JPH09182083A (en) 1995-12-27 1997-07-11 Matsushita Electric Ind Co Ltd Video image encoding method and decoding method and device therefor
US5661524A (en) 1996-03-08 1997-08-26 International Business Machines Corporation Method and apparatus for motion estimation using trajectory in a digital video encoder
EP1182888A3 (en) 1996-04-25 2004-11-03 Matsushita Electric Industrial Co., Ltd. Video coding with bit rate dependent quantisation
US5825680A (en) 1996-06-21 1998-10-20 Digital Equipment Corporation Method and apparatus for performing fast division
US6069914A (en) 1996-09-19 2000-05-30 Nec Research Institute, Inc. Watermarking of image data using MPEG/JPEG coefficients
US6480541B1 (en) * 1996-11-27 2002-11-12 Realnetworks, Inc. Method and apparatus for providing scalable pre-compressed digital video with reduced quantization based artifacts
US5920353A (en) 1996-12-03 1999-07-06 St Microelectronics, Inc. Multi-standard decompression and/or compression device
US6236727B1 (en) * 1997-06-24 2001-05-22 International Business Machines Corporation Apparatus, method and computer program product for protecting copyright data within a computer system
US5987179A (en) 1997-09-05 1999-11-16 Eastman Kodak Company Method and apparatus for encoding high-fidelity still images in MPEG bitstreams
US6014182A (en) 1997-10-10 2000-01-11 Faroudja Laboratories, Inc. Film source video detection
KR100235988B1 (en) 1997-11-12 1999-12-15 구자홍 Video format conversion appratus and its method in dtv
US6101602A (en) * 1997-12-08 2000-08-08 The United States Of America As Represented By The Secretary Of The Air Force Digital watermarking by adding random, smooth patterns
US5956026A (en) 1997-12-19 1999-09-21 Sharp Laboratories Of America, Inc. Method for hierarchical summarization and browsing of digital video
WO1999038333A1 (en) 1998-01-26 1999-07-29 Sgs-Thomson Microelectronics Asia Pacific (Pte) Ltd. One-pass variable bit rate moving pictures encoding
US6172768B1 (en) 1998-02-05 2001-01-09 Canon Kabushiki Kaisha Halftoning with changeable error diffusion weights
JPH11239351A (en) 1998-02-23 1999-08-31 Nippon Telegr & Teleph Corp <Ntt> Moving image coding method, decoding method, encoding device, decoding device and recording medium storing moving image coding and decoding program
US6434196B1 (en) 1998-04-03 2002-08-13 Sarnoff Corporation Method and apparatus for encoding video information
JP4123567B2 (en) 1998-04-30 2008-07-23 ソニー株式会社 Image signal processing apparatus and method
US6215824B1 (en) 1998-05-01 2001-04-10 Boom Corporation Transcoding method for digital video networking
US6332194B1 (en) 1998-06-05 2001-12-18 Signafy, Inc. Method for data preparation and watermark insertion
US6301304B1 (en) 1998-06-17 2001-10-09 Lsi Logic Corporation Architecture and method for inverse quantization of discrete cosine transform coefficients in MPEG decoders
KR100275933B1 (en) 1998-07-14 2000-12-15 구자홍 Idct in mpeg decoder
US6456340B1 (en) * 1998-08-12 2002-09-24 Pixonics, Llc Apparatus and method for performing image transforms in a digital display system
JP3748717B2 (en) 1998-08-31 2006-02-22 シャープ株式会社 Video encoding device
JP2001112000A (en) 1999-10-07 2001-04-20 Matsushita Electric Ind Co Ltd Video signal encoding device
WO2001033864A1 (en) 1999-10-29 2001-05-10 Koninklijke Philips Electronics N.V. Video encoding-method
WO2001050771A1 (en) 1999-12-29 2001-07-12 Koninklijke Philips Electronics N.V. Device for image encoding and decoding
US6810131B2 (en) * 2000-01-05 2004-10-26 Canon Kabushiki Kaisha Information processing method and apparatus
US6629104B1 (en) 2000-11-22 2003-09-30 Eastman Kodak Company Method for adding personalized metadata to a collection of digital images
US8374237B2 (en) 2001-03-02 2013-02-12 Dolby Laboratories Licensing Corporation High precision encoding and decoding of video images
US7266150B2 (en) 2001-07-11 2007-09-04 Dolby Laboratories, Inc. Interpolation of video compression frames
US6816552B2 (en) 2001-07-11 2004-11-09 Dolby Laboratories Licensing Corporation Interpolation of video compression frames
US20030112873A1 (en) 2001-07-11 2003-06-19 Demos Gary A. Motion estimation for video compression systems
US20030112863A1 (en) 2001-07-12 2003-06-19 Demos Gary A. Method and system for improving compressed image chroma information
US7046818B2 (en) * 2001-07-20 2006-05-16 Seiko Epson Corporation Standards compliant watermarking for access management
WO2003041041A2 (en) 2001-11-06 2003-05-15 Pioneer Corporation Displ ay panel driving apparatus with reduced power loss
US20040001546A1 (en) 2002-06-03 2004-01-01 Alexandros Tourapis Spatiotemporal prediction for bidirectionally predictive (B) pictures and motion vector prediction for multi-picture reference motion compensation

Patent Citations (34)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4903317A (en) 1986-06-24 1990-02-20 Kabushiki Kaisha Toshiba Image processing apparatus
JPH01140883A (en) 1987-11-27 1989-06-02 Dainippon Printing Co Ltd Data coding method
US5418571A (en) 1991-02-01 1995-05-23 British Telecommunicatons Public Limited Company Decoding of double layer video signals with interpolation replacement on missing data from enhancement layer
US5465119A (en) 1991-02-22 1995-11-07 Demografx Pixel interlacing apparatus and method
US5737027A (en) 1991-02-22 1998-04-07 Demografx Pixel interlacing apparatus and method
US6111975A (en) 1991-03-22 2000-08-29 Sacks; Jack M. Minimum difference processor
US5414469A (en) 1991-10-31 1995-05-09 International Business Machines Corporation Motion video compression system with multiresolution features
US5493338A (en) 1991-12-28 1996-02-20 Goldstar Co., Ltd. Scan converter of television receiver and scan converting method thereof
US5253058A (en) 1992-04-01 1993-10-12 Bell Communications Research, Inc. Efficient coding scheme for multilevel video transmission
US5270813A (en) 1992-07-02 1993-12-14 At&T Bell Laboratories Spatially scalable video coding facilitating the derivation of variable-resolution images
JPH06165150A (en) 1992-11-18 1994-06-10 Matsushita Electric Ind Co Ltd Dynamic picture coding/decoding device
JPH06350995A (en) 1993-06-10 1994-12-22 Casio Comput Co Ltd Moving picture processing method
US5408270A (en) 1993-06-24 1995-04-18 Massachusetts Institute Of Technology Advanced television system
US5387940A (en) 1993-07-07 1995-02-07 Rca Thomson Licensing Corporation Method and apparatus for providing scaleable compressed video signal
EP0634871A2 (en) 1993-07-13 1995-01-18 AT&T Corp. Scalable encoding and decoding of high-resolution progressive video
US5742343A (en) 1993-07-13 1998-04-21 Lucent Technologies Inc. Scalable encoding and decoding of high-resolution progressive video
US5519453A (en) 1993-08-06 1996-05-21 U. S. Philips Corporation Method of eliminating interfernce signals from video signals
CA2127151A1 (en) 1993-09-21 1995-03-22 Atul Puri Spatially scalable video encoding and decoding
JPH07203426A (en) 1993-12-28 1995-08-04 Toshiba Corp Hierarchical coding and decoding device
US20020003838A1 (en) 1994-06-08 2002-01-10 Toshiya Takahashi Image conversion apparatus
US5828788A (en) 1995-06-29 1998-10-27 Thomson Multimedia, S.A. System for processing data in variable segments and with variable data resolution
US6028634A (en) 1995-10-27 2000-02-22 Kabushiki Kaisha Toshiba Video encoding and decoding apparatus
WO1997028507A1 (en) 1996-01-30 1997-08-07 Demografx Temporal and resolution layering in advanced television
US5852565A (en) 1996-01-30 1998-12-22 Demografx Temporal and resolution layering in advanced television
US5988863A (en) 1996-01-30 1999-11-23 Demografx Temporal and resolution layering in advanced television
US20040196901A1 (en) 1996-01-30 2004-10-07 Demos Gary A. Median filter combinations for video noise reduction
US6728317B1 (en) 1996-01-30 2004-04-27 Dolby Laboratories Licensing Corporation Moving image compression quality enhancement using displacement filters with negative lobes
US5974159A (en) 1996-03-29 1999-10-26 Sarnoff Corporation Method and apparatus for assessing the visibility of differences between two image sequences
US6175592B1 (en) 1997-03-12 2001-01-16 Matsushita Electric Industrial Co., Ltd. Frequency domain filtering for down conversion of a DCT encoded picture
US6489956B1 (en) 1998-02-17 2002-12-03 Sun Microsystems, Inc. Graphics system having a super-sampled sample buffer with generation of output pixels using selective adjustment of filtering for implementation of display effects
US6252906B1 (en) 1998-07-31 2001-06-26 Thomson Licensing S.A. Decimation of a high definition video signal
US6442203B1 (en) 1999-11-05 2002-08-27 Demografx System and method for motion compensation and frame rate conversion
US7106322B2 (en) * 2000-01-11 2006-09-12 Sun Microsystems, Inc. Dynamically adjusting a sample-to-pixel filter to compensate for the effects of negative lobes
WO2001077871A1 (en) 2000-04-07 2001-10-18 Demografx Enhanced temporal and resolution layering in advanced television

Non-Patent Citations (28)

* Cited by examiner, † Cited by third party
Title
"IEEE Standard Specification for the Implementations of 8×8 Inverse Discrete Cosine Transforms," IEEE Std 1180-1990, The Institute of Electrical and Electronics Engineers, Inc.; United States of America, 13 pages (1991).
Certified English Translation for Japanese Patent Publication No. 01-140883, published Jun. 2, 1989, entitled "Data Coding Method".
English language abstract for JP 06165150, published Jun. 10, 1994, entitled: "Dynamic Picture Coding/Decoding Device".
European Office Action, European Patent Application No. 01924762.6, dated Oct. 11, 2006, 11 pages.
Girod, "Motion-Compensating Prediction with Fractional Pel Accuracy", IEEE Transactions on Communications, vol. 41, No. 4, Apr. 1993, pp. 604-612.
Girod, "Motion-Compensating Prediction with Fractional-Pel Accuracy", IEEE Transactions on Communications, vol. 41, No. 4, Apr. 1993, pp. 604-612.
H. 263 Appendix III, ITU-T Telecommunication Standardization Sector of ITU, Series H: Audiovisual and Multimedia Systems, Infrastructure of audiovisual services-coding of moving video. Video coding for low bit rate communication, Appendix III: Examples for H.263 encoder/decoder implementations, (Jun. 2001), 48 pages.
H.261, ITU-T Telecommunication Standardization Sector of ITU, Line Transmission of non-telephone signals. Video Codec for Audiovisual Services at p X64 kbits, (Mar. 1993), 32 pages.
H.263 Appendix III, ITU-T Telecommunication Standardization Sector of ITU, Series H: Audiovisual and Multimedia Systems, Infrastructure of audiovisual services-coding of moving video. Video coding for low bit rate communication, Appendix III: Examples for H.263 encoder/decoder implementations, (Jun. 2001), 48 pages.
H.263, ITU-T Telecommunication Standardization Sector of ITU, Series H: Audiovisual and Multimedia Systems, Infrastructure of audiovisual services-coding of moving video. Video coding for low bit rate communication, (Jan. 2005), 226 pages.
ISO/IEC 14496-2 International Standard, Information technology-coding of audio-visual objects-Part 2: visual, 2nd Edition, Amendment 2: Streaming video profile, Feb. 1, 2002, 64 pages.
ISO/IEC 14496-2 International Standard, Information technology-coding of audio-visual objects-Part 2: visual, 2nd Edition, Dec. 1, 2001, 536 pages.
ISO/IEC JTC 1, "Coding of audio-visual objects-Part 2: Visual," ISO/IEC 14496-2 (MPEG-4 Part 2), Dec. 1999, 348 pages.
Japanese Office Action, Application Serial No. 2001-574651, dated Mar. 22, 2007, 74 pages.
Machine Translation for Japanese Patent Publication No. 07-203426, published Aug. 4, 1995, entitled "Hierarchical Coding and Decoding Device" (22 pages).
Office Action, U.S. Appl. No. 11/187,176, dated Aug. 30, 2007, 16 pages.
Office Action, U.S. Appl. No. 11/187,176, dated Jan. 4, 2007, 45 pages.
Patent Abstract of Japan for Japanese Patent Publication No. 01-140883, published Jun. 2, 1989, entitled "Data Coding Method".
Patent Abstract of Japan for Japanese Patent Publication No. 07-203426, published Aug. 4, 1995, entitled "Hierarchical Coding and Decoding Device".
Patent Abstracts of Japan, vol. 1995, No. 03 (Apr. 28, 2005) for Japanese Patent Publication JP 06350995, published Dec. 22, 1994, entitled, "Moving Picture Processing Method".
Patent Abstracts of Japan, vol. 1995, No. 3 (Apr. 28, 2005) for Japanese Patent Publication JP 06350995, published Dec. 22, 1994, entitled, "Moving Picture Processing Method".
Puri et al., "Temporal Resolution Scalable Video Coding," Image Processing. 1994 International Conference, IEEE, pp. 947-951 (1994).
Shen et al., "Adaptive motion vector resampling for compressed video down-scaling", International Conference on Image Processing, vol. 1, pp. 771-774, Oct. 26-29, 1997.
U.S. Final Office Action for U.S. Appl. No. 12/618,202 dated Sep. 8, 2011, 7 pages.
U.S. Non-Final Office Action for U.S. Appl. No. 12/618,202 dated May 28, 2010, 10 pages.
U.S. Non-Final Office Action for U.S. Appl. No. 12/618,202 dated May 3, 2011, 8 pages.
U.S. Notice of Allowance for U.S. Appl. No. 12/618,202 dated Jan. 4, 2012, 7 pages.
Vincent, A., et al., "Spatial Prediction in Scalable Video Coding," International Broadcasting Convention, IEEE Conference Publication No. 413, RAI International Congress and Exhibition Centre, Amsterdam, The Netherlands, Sep. 14-18, 1995, pp. 244-249.

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8767829B2 (en) 2001-07-11 2014-07-01 Dolby Laboratories Licensing Corporation Switch-select single frame reference
US8942285B2 (en) 2001-07-11 2015-01-27 Dolby Laboratories Licensing Corporation Motion compensation filtering in an image system
US8995528B2 (en) 2001-07-11 2015-03-31 Dolby Laboratories Licensing Corporation Switch-select single frame reference

Also Published As

Publication number Publication date
US20050254649A1 (en) 2005-11-17
USRE43360E1 (en) 2012-05-08
US20040196901A1 (en) 2004-10-07
US7428639B2 (en) 2008-09-23
US6728317B1 (en) 2004-04-27
US7366242B2 (en) 2008-04-29
US6957350B1 (en) 2005-10-18
USRE45082E1 (en) 2014-08-19

Similar Documents

Publication Publication Date Title
USRE45082E1 (en) Enhancing image quality in an image system
CA2406459C (en) Enhanced temporal and resolution layering in advanced television
US7280155B2 (en) Method and system for converting interlaced formatted video to progressive scan video
US6192079B1 (en) Method and apparatus for increasing video frame rate
US6862372B2 (en) System for and method of sharpness enhancement using coding information and local spatial features
KR100957479B1 (en) Method and apparatus for spatio-temporal deinterlacing aided by motion compensation for field-based video
US7057664B2 (en) Method and system for converting interlaced formatted video to progressive scan video using a color edge detection scheme
US20130329796A1 (en) Method and system for motion compensated picture rate up-conversion of digital video using picture boundary processing
KR19990082104A (en) Advanced Television&#39;s Time and Resolution Layer
US6873657B2 (en) Method of and system for improving temporal consistency in sharpness enhancement for a video signal
EP1506525B1 (en) System for and method of sharpness enhancement for coded digital video
US7587091B2 (en) De-interlacing using decoder parameters
JPH11298861A (en) Method and device for converting frame number of image signal
EP1352515A1 (en) Apparatus and method for providing a usefulness metric based on coding information for video enhancement
US6804303B2 (en) Apparatus and method for increasing definition of digital television
Deame Motion compensated de-interlacing: The key to the digital video transition
JP2004515133A (en) Decompression of encoded video

Legal Events

Date Code Title Description
FEPP Fee payment procedure

Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

LAPS Lapse for failure to pay maintenance fees

Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY