CN105338361B - For the method, apparatus and storage medium for being guided filtering in Video coding and decoding - Google Patents

For the method, apparatus and storage medium for being guided filtering in Video coding and decoding Download PDF

Info

Publication number
CN105338361B
CN105338361B CN201510809673.6A CN201510809673A CN105338361B CN 105338361 B CN105338361 B CN 105338361B CN 201510809673 A CN201510809673 A CN 201510809673A CN 105338361 B CN105338361 B CN 105338361B
Authority
CN
China
Prior art keywords
image
color component
filter
cti
filter coefficients
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201510809673.6A
Other languages
Chinese (zh)
Other versions
CN105338361A (en
Inventor
苏冠铭
曲晟
尹鹏
叶琰
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Dolby Laboratories Licensing Corp
Original Assignee
Dolby Laboratories Licensing Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Dolby Laboratories Licensing Corp filed Critical Dolby Laboratories Licensing Corp
Publication of CN105338361A publication Critical patent/CN105338361A/en
Application granted granted Critical
Publication of CN105338361B publication Critical patent/CN105338361B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/44Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/117Filters, e.g. for pre-processing or post-processing
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/136Incoming video signal characteristics or properties
    • H04N19/137Motion inside a coding unit, e.g. average field, frame or block difference
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/186Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/187Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a scalable video layer
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/46Embedding additional information in the video signal during the compression process
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/59Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial sub-sampling or interpolation, e.g. alteration of picture size or resolution
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/80Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/85Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression
    • H04N19/86Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression involving reduction of coding artifacts, e.g. of blockiness
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/85Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression
    • H04N19/89Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression involving methods or arrangements for detection of transmission errors at the decoder
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/30Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
    • H04N19/33Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability in the spatial domain

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)
  • Image Processing (AREA)

Abstract

The invention discloses be guided picture up-sampling in Video coding.Encoder receives the first image of the first spatial resolution and the second image of second space resolution ratio, wherein both the first image and second image indicate same scene, the second space high resolution is in first spatial resolution.Filter is selected up-sampling the first image for the third image with spatial resolution identical with the second space resolution ratio.The filter factor for the up-sampling filter is calculated by minimizing the error metrics (such as MSE) between second image and the pixel value of the third image.The set of the filter factor of calculating is signaled to receiver (for example, as metadata).Decoder receives the first image (or it is approximate) and the metadata, and the filter factor of the identical filter and optimum choice derived with the encoder can be used to up-sample to the first image.

Description

Method, apparatus and storage medium for guided filtering in video encoding and decoding
The present application is a divisional application based on the patent application with application number 201210281757.3, application date 2012-8-9, entitled "guided image upsampling in video coding".
Cross Reference to Related Applications
The present invention claims priority from U.S. provisional patent application No.61/521,685 filed on 9/2011 and U.S. provisional patent application No.61/653,234 filed on 30/5/2012, which are incorporated herein by reference for all purposes.
Technical Field
The present invention generally relates to images. More particularly, embodiments of the present invention relate to guided up-sampling of digital images in video coding.
Background
As used herein, the term "dynamic range" (DR) may relate to the ability of the human psychovisual system (HVS) to perceive a range of intensities (e.g., luminance, brightness) in an image (e.g., from darkest to brightest). In this sense, DR is related to the "scene-related" strength. DR may also be related to the ability of the display device to adequately or approximately present an intensity range of a particular width (break). In this sense, DR is related to "display-related" intensity. Unless any point in this description explicitly specifies that a particular meaning is of particular importance, it should be inferred that it may be used interchangeably in any meaning, for example.
As used herein, the term "High Dynamic Range (HDR)" relates to a DR breadth spanning some 14-15 orders of magnitude of the Human Visual System (HVS). For example, a well-adapted person with a substantially normal (e.g., in one or more of a statistical, biometric, or ophthalmic (opthamological) sense) has an intensity range that spans about 15 orders of magnitude. Adapted persons can perceive dim light sources of as few as just a few photons. However, these same people can perceive near-burning bright intensities of midday sunlight in deserts, oceans, or snow (or even glances at the sun, however, brief glances prevent damage). Such a span is available to "adapted" people (e.g., people whose HVS has a time period that is reset and adjusted).
In contrast, the DR over which a person can simultaneously perceive a wide width in the intensity range may be slightly truncated relative to the HDR. As used herein, the term "visual dynamic range" or "Variable Dynamic Range (VDR)" may relate to the DR that is simultaneously perceptible by the HVS, either individually or interchangeably. As used herein, VDR may be associated with DR spanning 5-6 orders of magnitude. Thus, VDR represents a wide DR width, although somewhat narrower with respect to real scene-related HDR. As used herein, the term "simultaneous dynamic range" may relate to VDR.
Until recently, displays have had a significantly narrower DR than HDR or VDR. Typical Cathode Ray Tubes (CRTs), Televisions (TVs) with Liquid Crystal Displays (LCDs) and computer monitor devices using white backlight or plasma screen technology with constant fluorescence may be constrained to be about three orders of magnitude in their DR rendering capabilities. These conventional displays therefore represent a Low Dynamic Range (LDR), also called Standard Dynamic Range (SDR), with respect to VDR and HDR.
As for scalable video coding and HDTV technologies, extended picture DR typically involves a bifurcate (bifurcate) method. For example, scene-related HDR content captured with a modern HDR-capable camera may be used to generate a VDR version or an SDR version of the content, which may be displayed on a VDR display or a legacy SDR display. In one approach, generating an SDR version from a captured VDR version may involve applying a global Tone Mapping Operator (TMO) to intensity (e.g., luminance, luma) related pixel values in the HDR content. In a second approach, generating an SDR Image may involve applying a reversible operator (or predictor) to the VDR data, as described in "Extending Image Dynamic Range" in PCT application PCT/US 2011/048861 filed 2011, 8/23, w.gish et al. To conserve bandwidth or for other considerations, sending both the actual captured VDR content and SDR content simultaneously may not be the best approach.
Thus, an Inverse Tone Mapping Operator (iTMO) that is inverted with respect to the initial TMO or an inverse operator with respect to the initial predictor may be applied to the generated SDR content version, which allows for predicting a version of the initial VDR content. The predicted VDR content version can be compared to the generated VDR content. For example, subtracting the predicted VDR version from the initial VDR version may generate a residual image. The encoder may send the generated SDR content as a Base Layer (BL), any residual image as an Enhancement Layer (EL), and pack iTMO or other predictors, etc. as metadata.
Sending the EL and metadata along with its SDR content, residual and predictor in the bitstream typically consumes less bandwidth than would be consumed if both HDR and SDR content were sent directly into the bitstream. A compatible decoder receiving the bitstream sent by the encoder can decode and render the SDR on a legacy display. However, a compatible decoder may also use the residual image, the iTMO predictor, or the metadata to compute a predicted version of the HDR content therefrom for use on a more capable display.
In such layered VDR coding, images may be represented in different spatial resolutions, bit depths, color spaces, and chroma sub-sampling formats, all of which may force various computer intensive transformations from a first color format to a second color format.
As used herein, the term "color format" relates to a color representation that includes two variables: a) color space variables (e.g., RGB, YUV, YCbCr, etc.) and chroma subsampling variables (e.g., 4: 4, 4: 2: 0, etc.). For example, the VDR signal may have an RGB 4: 4 color format, while the SDR signal may have a YCbCr 4: 2: 0 color format.
As used herein, the term "upsampling" or "upscaling" relates to the process of transforming one or more color components of an image from one spatial resolution to a second, higher spatial resolution. For example, the video signal may be upsampled from a 4: 2: 0 format to a 4: 4 format.
The approaches described in this section are approaches that could be pursued, but not necessarily approaches that have been previously conceived or pursued. Accordingly, unless otherwise indicated, it should not be assumed that any of the approaches described in this section qualify as prior art merely by virtue of their inclusion in this section. Similarly, unless otherwise indicated, problems identified with respect to one or more methods should not be assumed to have been recognized in any prior art based on this section.
Drawings
Embodiments of the invention are illustrated by way of example, and not by way of limitation, in the figures of the accompanying drawings and in which like reference numerals refer to similar elements and in which:
FIG. 1 depicts an exemplary data stream for a layered coding system, according to an embodiment of the invention;
FIG. 2 depicts an exemplary layered decoding system, according to an embodiment of the invention;
fig. 3 depicts an example of guided image upsampling in encoding a residual signal in a layered codec according to an embodiment of the present invention;
FIG. 4 depicts an exemplary single-layer video coding system according to an embodiment of the invention;
FIG. 5 depicts an example input and output pixel array for upsampling by a factor of 2 using a 2D filter in accordance with an embodiment of the present invention;
FIG. 6 depicts an example input and output pixel array for upsampling by a factor of 2 using a 3D filter in accordance with an embodiment of the present invention;
FIG. 7 depicts an example process of guided image upsampling according to an embodiment of the invention;
FIG. 8 depicts an example process for guided color temporal improvement filtering, according to an embodiment of the invention.
Detailed Description
Guided image upsampling and color temporal improvement filtering in video coding is described herein. In the following description, for purposes of explanation, numerous details are set forth in order to provide a thorough understanding of the present invention. However, it is understood that the invention may be practiced without these specific details. In other instances, well-known structures and devices are not described in detail to avoid unnecessarily obscuring, or obfuscating the present invention.
SUMMARY
Example embodiments described herein relate to guided image upsampling and CTI filtering in video coding. An encoder receives a first image at a first spatial resolution and a guide image (guide image) at a second spatial resolution, wherein the first image and the guide image both represent the same scene, the second spatial resolution being higher than the first spatial resolution. A filter is selected to upsample the first image into a third image having the same spatial resolution as the second spatial resolution. The filter coefficients for the upsampling filter are calculated by minimizing an error metric, such as Mean Square Error (MSE), between the pixel values of the guide image and the third image. The set of calculated filter coefficients is signaled (signal) to the receiver (e.g., as metadata). A decoder receives the metadata and the first image or an approximation of the first image and may upsample the received image using the same optimization filters and filter coefficients as derived by the encoder.
In another embodiment, an encoder receives a target image to be encoded, the target image comprising a first target color component image and a second target color component image. Encoding and decoding the image to generate an encoded image and a decoded image, the decoded image including a decoded first color component image and a decoded second color component image. Selecting a Color Transient Improvement (CTI) filter to filter pixels of the decoded second color component image to generate an output color component image. Calculating CTI filter coefficients for the CTI filter based at least in part on minimizing an error metric between pixel values of the output color component image and corresponding pixel values of a second color component image in the target image. The CTI filter coefficients are signaled to the receiver (e.g., as metadata). A decoder receives the metadata and the encoded image. After decoding the encoded image, it may filter the decoded image using the same CTI filter coefficients as those derived by the encoder.
Example System in case of guided image upsampling
Image downsampling and upsampling transforms play a key role in video coding since they affect not only coding efficiency but also overall image quality. Improper spatial scaling transformation (spatial scaling transformation) can lead to erroneous colors, especially along the edges of the image. Unlike conventional "blind" image upsampling techniques, in which, given an input image having sub-sampled color components (e.g., chroma in the YCbCr 4: 2: 0 format), the color components are upsampled using only the available information within the image itself, embodiments of the present invention may also utilize information from other images in the video processing channel (pipeline).
FIG. 1 depicts an example image processing system 100 implementing guided image upsampling according to an embodiment of the present invention. System 100 represents an embodiment of a layered encoder in which two layers (a base layer 135 and an enhancement or residual layer 175) are used to encode the input signal V105.
In an embodiment, the input signal V105 may be represented as an input VDR signal represented by 16 or more bits per color component in a 4: 4: 4 color format (e.g., RGB 4: 4: 4). The dynamic range reduction process 110 may process the VDR signal to generate a signal S' 112. The signal S' may have the same spatial resolution as the signal V or a lower spatial resolution than the signal V. The signal S' may be represented by a lower bit depth resolution than V, e.g. 12 bits per color component. The signal S' may be in the same color format as V, or in other embodiments, it may be in a different color format.
In an embodiment, color transformation 120 may precede encoding 130, where S' may be transformed into another color format (e.g., YCbCr). The transform 120 may also combine sub-samples of one or more color components (e.g., from 4: 4 to 4: 2: 0). The encoded signal 135 may be transmitted as a base layer. In an embodiment, encoding 130 may be implemented by any existing video encoder, such as an MPEG-2 or MPEG-4 video encoder specified by the Motion Pictures Expert Group (MPEG) specification.
Enhancement layer 175 may be generated by decoding signal 135, generating a predicted value (165) of initial VDR signal V, and subtracting initial V (105) from its predicted value (165) to generate residual signal 175. In an embodiment, predictor 160 may be implemented using a multivariate multi-regression model as described in PCT application PCT/US 2012/033605 filed 2012, 4/13/2012 by G-MSu et al. Since the encoded signal 135 and the input signal 105 have different color formats and resolutions, the color transform process 150 transforms the output of the decoder 140 into a color format and resolution that matches the color format and resolution of the input V105. For example, unit 150 may convert input data 143 from YCbCr 4: 2: 0 to RGB 4: 4.
In the conventional sampling method, the up-sampled output 153 may be derived based on only the sub-sampled input 143 and an interpolation filter having a fixed filter coefficient. Conversely, in an embodiment, the upsampling process in 150 may perform upsampling using data from the subsampled input 143 and data having a known full spatial resolution (e.g., input S' 112 or V105). In an embodiment, the guided image upsampling process 150 may signal the upsampling related parameters (e.g., interpolation filter coefficients) to the remaining system channels (e.g., as metadata signals 155 or 167).
As used herein, the term "metadata" may relate to any ancillary information that is transmitted as part of the encoded bitstream and assists the decoder in presenting the decoded images. The metadata may include, but is not limited to, the following information: color space or gamut information, dynamic range information, tone mapping information, or other predictors, upscaling (upscaling), and quantizer operators, such as those described herein.
In an embodiment, the filter coefficients used to convert the YCbCr 4: 2: 0 data to YCbCr 4: 4 data are estimated by minimizing a measure of error (e.g., Mean Square Error (MSE)) between the prediction up-sampled value (e.g., output 153 of color transform 150) and the full spatial resolution input guide image (e.g., S' 112 or V105). The same filter may also be applied during the upsampling process in the decoder.
Fig. 2 depicts an example implementation of a layered video decoder according to an embodiment. The decoding system 200 receives an encoded bitstream 202 that includes a base layer 227, an enhancement layer (or residual) 222, and metadata 225 that are extracted after demultiplexing 220. For example, in a VDR-SDR system, the base layer 227 may represent an SDR representation of the encoded signal, and the metadata 225 may include information used in the encoder regarding prediction (160) and color transform operations (150). The encoded base layer 227 may be decoded using the base layer decoder 210 to output a decoder SDR signal S212. The encoded residual 222 may be decoded (240), dequantized (250), and added to the output 295 of the predictor 290 to generate the output VDR signal V270.
The color conversion unit 230 may incorporate upsampling (e.g., from 4: 2: 0 to 4: 4). Instead of using a "blind" upsampling technique that relies only on the input data 212, the upsampling processor in 230 may extract and apply upsampling related parameters (e.g., interpolation filter coefficients) signaled by the encoder using the metadata 225 (or 155). Such guided upsampling during decoding may result in a video signal with improved visual quality without additional computational cost.
The guided image upsampling technique may be applied to other processing steps in the video encoding channel as well. Fig. 3 depicts a system for encoding-decoding residual layers in layered coding implemented in accordance with an embodiment of the invention. The residual signal R305 (175) may be in RGB 4: 4 color format. The color transform unit 310 precedes residual coding 330 (e.g., using an MPEG-4 video encoder), which typically operates in YCbCr 4: 2: 0 format, where the input 305 may be color transformed and downsampled to YCbCr 4: 2: 0. In parallel, the upsampling processor 320 may be directed by the input R305 to compute optimized upsampled filter coefficients according to embodiments of the present invention. These filter coefficients may be signaled to the decoder, for example, using metadata 322. At the receiver, the color transform unit 350 may extract optimized upsampled filter coefficients from the metadata 322 and apply them during the process of upsampling the decoded YCbCr 4: 2: 0 data 342 into RGB 4: 4 data 355.
FIG. 4 depicts an example single-layer video coding system according to an embodiment of this disclosure. As depicted in fig. 4, processor 410 may process input signal VI405 is downsampled. The downsampled output 412 is encoded by a video encoder 430 (e.g., an MPEG-4 encoder) and sent to a decoder including a video decoder440 (e.g., an MPEG-4 decoder) and an upsampling processor 450 (400-D). At the encoder (400-E), the upsampling unit 420 may perform guided upsampling according to the methods described in embodiments of the present invention, and may use the upsampled data from the full resolution input VI405 and sub-sampling the signal 412 to derive optimized upsampled filter coefficients. The optimized filter coefficients may be signaled to the decoder (400-D), for example, using metadata 422. Thus, the output 442 of the video encoder (430) may be upsampled in the processing unit 450 using the same optimized set of coefficients as generated by the upsampling unit 420.
Filter design for guided image upsampling
Ordinary 2D inseparable filter
For simplicity and without loss of generality, given an input image comprising multiple color components (e.g., YCbCr or RGB), we consider a guided image upsampling process of a single color component (e.g., Cb or R). The method described herein may be repeated as desired for any image color component that needs to be upsampled.
In an example embodiment for upsampling by a factor of 2 using a 2D interpolation or upsampling filter, fig. 5 depicts a 3 x 3 array of known input pixels represented by circles 520 to be upsampled by a factor of 2 and generate a 6 x 6 pixel array as depicted by squares 510. The input pixel is denoted cjAnd the predicted or upsampled pixel is represented asFour new estimated pixels (e.g., 510-22, 510-23, 510-32, and 510-33) are then generated for each input pixel (e.g., 520-11). In an embodiment, the upsampling process may be represented as a non-separable Finite Impulse Response (FIR) filter:
for i ═ 0, 1, 2, 3, (1)
Wherein,filter coefficients are represented (for x and y from 0 to 2). For example, in FIG. 5,shown as squares 510-22.
Now consider a guide image D having the same target resolution as the output of the upsampling process, where the pixel data is represented here asFor example, as described in fig. 4, the guide image may be the input image VI405. Deriving filter coefficientsCan be expressed as an error minimization problem. In an embodiment, the filter coefficients are generated such that a minimum Mean Square Error (MSE) between the guide pixel data and the predicted upsampled pixel data is minimized. This can be formulated as:
where p represents the total output pixel in the up-sampled image. For example, if the output upsampled image has a resolution of m rows and n columns, then for each color component, p equals mn.
Equation (2) can be solved using various known numerical techniques (e.g., the techniques described in "Applied multivariable statistical analysis," r.a. johnson, and d.w. wicher n, 5th Edition, precision Hall, 2001). In an example embodiment, the filter coefficients are represented as
For a 3 × 3 filter, a vector of input pixel data (520) is given
A p x 9 matrix C may be formed from the input sub-sampled data as
Similarly, the p × 4 guide data matrix R may be formed as
Wherein
The estimated (upsampled) pixel may then be represented as
Wherein
From equations (1-9), the estimated pixel can be expressed as
In an embodiment, the optimization objective is to minimize the estimated error between the instructional input data and the data sampled on the estimated output, which can be expressed from equation (2):
a solution optimized in the least squares sense can be obtained via the following.
M=(CTC)-1CTR. (12)
From equation (3), for each color component (e.g., Cb and Cr) to be upsampled, this embodiment may compute 4 * 9 ═ 36 coefficients, which may be communicated to the decoder using metadata or other means.
With respect to the embodiment of fig. 1, the upsampling in the color conversion unit 150 may use R pixels from the input image S' 112 or the input image V105 as a guide image.
As depicted in FIG. 5 and represented by equation (1), an embodiment of 1: 2 pixel upsampling filtering utilizes a 3 × 32D filter. The methods described herein can be readily extended to support other filter sizes (e.g., 2 x2, 5 x 5, 7 x 7, etc.) and other upsampling ratios (e.g., 1: 3,1: 4, etc.). The method can also be applied to simpler 1D filters. In the embodiment, in the formulas (1) and (3), the setting can be performed byAndand only for(for y ═ 0, 1, and 2) to define a 3-tap 1D filter.
Symmetric 2D inseparable filter
In equation (1), each upsampled image pixel is predicted using its own set of filter coefficients. To reduce the number of coefficients that need to be signaled to the decoder, the total number of filter coefficients can be reduced by sharing them among the upsampled pixels. In an example embodiment, the same filter coefficients may be used to estimate the odd and even lines of the upsampled image data.
Returning to fig. 5, in an embodiment, the upsampling process may be formulated as follows. For each input sample cjEvaluating two samples in odd rows(510-22 and 510-23), two samples are estimated in even rows(510-32 and 510-33). The upsampling process can be represented using two 2D FIR filters sharing their filter coefficients for each upsampled pixel.
For i ═ 0, 1, (13) and
for i ═ 2, 3(14)
Similar to the previous embodiment, let
A 9x2 matrix representing the filter coefficients used in equations (13) and (14).
Is provided with
And
and input guidance data is set
The estimated upsampled pixels may then be represented as
And from equation (11) and equations (16-26), the optimized solution M in the least-squares sense can be obtained via the following
M=(CTC)-1CTR. (27)
Ordinary 3D inseparable filter
In a particular embodiment, at least one color component (e.g., the Y (luminance) component in YCbCr or the green (G) component in RGB) may not undergo any downsampling, thus preserving significant edge-related information. The upsampling process of the remaining color components (e.g., Cb and Cr in YCbCr) can be further improved if the previously described 2D upsampling filter is extended to become a 3D filter as described herein.
FIG. 6 depicts a method for upsampling by a factor of 2 by applying a 3D filter according to an embodiment of the present inventionLike example input and output pixel arrays. As before, the input pixel of the color component to be upsampled (e.g., Cb) is denoted cj, and the predicted or upsampled pixel is denoted cjThe pixel of the same guide image but from another color component (e.g., Y) available at full resolution is denoted Yj(630). In an embodiment, it is assumed without loss of generality that data is being input (c)j) Using a 3 x 32D filter and inputting the guide data (e.g. y)jPixels) using a 4 x2 filter on separate components to compute each upsampled pixel, four new estimated pixels can be generated for each input pixel using a 3D filter described below:
for i ═ 0, 1, 2, 3, (28)
Wherein,andrepresenting the filter coefficients. These filter coefficients can be derived by solving an optimization problem as described previously
Wherein R is a matrix of guide image data defined in formula (6), and
wherein, givenTo be provided withA 9 × 4 matrix of coefficients (similar to the matrix represented in equation (3)) andto be provided withAn 8 x 4 matrix of coefficients,
the matrix C represents the observed pixel data. Given a
And
then
And as before, an optimal solution in the mean square sense can be defined as
M=(CTC)-1CTR. (37)
Symmetric 3D inseparable filter
The previously described method can be easily extended to embodiments where 2D and 3D filters with different numbers of pixel coefficients and different upsampling ratios can be applied.
As described in the case of the 2D filter, in equation (28), if the even and odd lines of the upsampled image data are estimated using the same filter coefficients, the number of coefficients that need to be signaled to the decoder can be reduced. In an embodiment, the prediction formula may be expressed as
For i ═ 0 and 1
For i ═ 2 and 3
Using equations (38) and (39) and the same methods previously applied, optimal filter coefficients can be derivedAnd
example processing for guided upsampling
FIG. 7 illustrates an example process of guided image upsampling according to an example implementation of the present invention. Processing begins at step 710, where an upsampling processor (e.g., processor 150, 320, or 420) receives an input image (e.g., image 143, 312, or 412) to be upsampled and an input guide image (e.g., input 112, 305, or 405), where the guide image has a spatial resolution higher than the spatial resolution of the input image. Given these two inputs, in step 720, an upsampling filter (e.g., a normal 2D non-separable filter or a symmetric 3D non-separable filter) is determined. The upsampling filter may be fixed and known by both the encoder and the decoder, or as previously described, the upsampling process may select a filter among various upsampling filters including (but not necessarily limited to) a 1D upsampling filter, a normal 2D upsampling filter, a symmetric 2D upsampling filter, a normal 3D upsampling filter, or a symmetric 3D upsampling filter. The selection of the upsampling filter may be performed using various methods that may take into account a number of criteria, including available computational and memory resources, MSE prediction error using a particular filter, and target coding efficiency.
Given an upsampling filter model, the set of filter coefficients M may be derived according to a predefined optimization criterion. For example, under the MSE criterion, an optimized solution for M may be derived using the MSE optimization techniques described herein that minimize the MSE between the guided image samples and the predicted samples of the upsampled image.
After solving for the filter coefficients M, in some embodiments, the coefficients and (optionally) characteristics (identity) of the upsampling filter may be sent to the receiver (e.g., as metadata).
The upsampling process 700 may be repeated at various intervals deemed necessary to maintain coding efficiency while using available computing resources. For example, when encoding a video signal, process 700 may be repeated for each frame, group of frames, portion of a frame, or whenever the prediction residual between the guide picture and the upsampled picture exceeds a certain threshold.
Guided Color Transient Improvement (CTI) filtering
As used herein, the term "color temporal artifact" means a color-related artifact in image or video processing. For example, in video and image coding, these artifacts can be identified by the presence of false colors in the decoded image (e.g., across the edges of an object). These artifacts may also be referred to as "color bleed". Color temporal artifacts may arise when each color plane of an image is processed separately using different compression levels. For example, in an RGB image, the red and blue planes may be quantized differently than the green plane. Similarly, in a YCbCr image, the Cb and Cr planes may be processed differently from the Y plane.
One approach for reducing color temporal artifacts is to apply post-processing filtering to the chrominance or secondary color components of the decoded image. Similar to the guided upsampling process described earlier, guided CTI filtering derives optimized filter coefficients in the encoder and sends them as metadata to the decoder. Furthermore, the filtering of a pixel in a color plane (e.g., Cb in YCbCr or B in RGB) may take into account both neighboring pixels in the same color plane and neighboring pixels of a corresponding pixel in another color plane (e.g., luminance Y plane in YCbCr, or G in RGB).
In an embodiment, the optimized filter coefficients may be derived in the encoder based on the original uncompressed data. The filter coefficients may be estimated on a block, frame or scene basis according to available resources and bandwidth. In the decoder, the filter may be applied as a post-processing (out-of-loop) filter to improve the overall picture quality.
Example derivation of filter coefficients in CTI filters
The input pixel for an image color component (e.g., Cb, Cr, R, or B) to be filtered using a CTI filter is denoted cijAnd the output filtered pixel is represented asOf the same image but from a second colour component (e.g. luminance Y)Or G) is represented by yij. In an embodiment, assuming without loss of generality that each color pixel is filtered using a generic 3D filter that applies a (2N +1) × (2N +1) kernel for the first color component and a (2M +1) × (2M +1) kernel for the second color component (e.g., 3 × 3 and 3 × 3 when N ═ M ═ 1), the filtered output may be expressed as:
wherein m isxyAnd nxyRepresenting the filter coefficients.
It can be appreciated that equation (40) is very similar to the general 3D upsampling filter described by equation (28); thus, the filter coefficients in equation (40) may be derived by solving an optimization problem as previously described
Wherein d isijRepresenting pixels of a reference or guide image (e.g., input V105).
As before, equation (40) may be represented in matrix form as
And the solution to equation (41) can be expressed as an optimization problem
Wherein R represents guide image data (d)ij) The vector of (a) is determined,is provided with mxyOf coefficient (2N +1)2The x 1 vector of the vector is calculated,is provided with nxyOf coefficient (2M +1)2X 1 vector, thereby
The matrix C represents the observed pixel data (C)ijAnd yij)。
As described above, the optimized solution of equation (43) in the mean-square sense can be expressed as
M=(CTC)-1CTR. (46)
This process may be repeated for each of the color components (e.g., Cb and Cr or R and B) that require CIT filtering.
FIG. 8 depicts an example process for guided color transient improvement filtering, according to an embodiment of the invention. In step 810, an encoder (e.g., system 100 depicted in fig. 1) may first reconstruct an estimate of a decoded image (e.g., V270) that would be received by a receiver (e.g., system 200 depicted in fig. 2). For example, a reconstructed picture at the decoder (e.g., 270) may be estimated by adding the output of the predictor 160 (e.g., signal 165) to the encoded and then decoded version of the residual 175.
In step 820, using equation (46), where R is based on input V105, the decoder can derive optimized CTI filter coefficients M. In step 830, these coefficients and other filtering parameters may be sent to the decoder as part of the metadata bitstream. At the decoder, after reconstruction of signal V270, a separate post-processing process may apply a CTI filter to signal V270 to improve overall quality by reducing color-related artifacts.
Process 800 may be repeated at various intervals deemed necessary to maintain coding efficiency while using available computing resources. For example, when encoding a video signal, process 800 may be repeated for each frame, group of frames, portion of a frame, or whenever the prediction residual between the guide image and the CTI-filtered image exceeds a certain threshold. Process 800 may also be repeated for each of the color components that may require CTI filtering.
Example computer System implementation
Embodiments of the invention may be implemented by a system configured in a computer system, electronic circuits and components, an Integrated Circuit (IC) device (e.g., a microcontroller, a Field Programmable Gate Array (FPGA)) or another configurable or Programmable Logic Device (PLD), a discrete-time or Digital Signal Processor (DSP), an application specific IC (asic), and/or an apparatus comprising one or more of these systems, devices or components. These computers and/or ICs may execute, control, or execute instructions related to directed upsampling or CTI filtering, such as described herein. The computer and/or IC may calculate any of a variety of parameters or values related to the guided image upsampling described herein. The image and video embodiments may be implemented in hardware, software, firmware, and various combinations thereof.
Particular implementations of the invention include a computer processor executing software instructions that cause the processor to perform the methods of the invention. For example, one or more processors in a display, encoder, set-top box, decoder, etc. may implement the above-described guided image upsampling or CTI filtering methods by executing software instructions in a program memory accessible to the processor. The present invention is also provided in the form of a program product. The program product may comprise any medium carrying a set of computer-readable signals comprising instructions which, when executed by a data processor, cause the data processor to carry out the method of the invention. The program product according to the invention may be in any of various forms. The program product may include, for example, physical media (e.g., magnetic data storage media including floppy disks, hard disk drives, optical data storage media including CD ROMs, DVDs, electronic data storage media including ROMs, flash RAMs, etc.), etc. The computer readable signal on the program product may optionally be compressed or encrypted.
Where a component (e.g., a software module, processor, assembly, device, circuit, etc.) is referred to above, unless otherwise indicated, reference to that component (including a reference to a "means") should be interpreted as including as equivalents of that component any component which performs the function of the described component (e.g., that is functionally equivalent), including components which are not structurally equivalent to the disclosed structure which performs the function in the illustrated exemplary embodiments of the invention.
Equivalents, extensions, substitutions and miscellaneous
Example embodiments related to guided image upsampling and CTI filtering are thus described. In the foregoing specification, embodiments of the invention have been described with reference to numerous specific details that may vary from implementation to implementation. Thus, the sole and exclusive indicator of what is the invention, and is intended by the applicants to be the invention, is the set of claims that issue from this application, in the specific form in which such claims issue, including any subsequent correction. Any definitions expressly set forth herein for terms contained in such claims shall govern the meaning of such terms as used in the claims. Hence, no limitation, element, property, feature, advantage or attribute that is not expressly recited in a claim should limit the scope of such claim in any way. The specification and drawings are, accordingly, to be regarded in an illustrative rather than a restrictive sense.

Claims (17)

1. A method of guided filtering for an encoder, the method comprising:
receiving a guide image and a target image to be encoded by an encoder, wherein the target image and the guide image both represent a similar scene and each comprise a first color component and a second color component;
encoding the target image with an encoder to generate an encoded image;
decoding the encoded image with a decoder to generate a decoded image, the decoded image comprising a decoded first color component and a decoded second color component;
selecting a color transient improvement CTI filter to filter pixels of the decoded image to generate an output color component image;
computing CTI filter coefficients for the CTI filter, wherein filter coefficient computation is based on minimizing an error metric between pixel values of the output color component image and corresponding pixel values of a second color component in the guide image, wherein CTI filter coefficients comprise a first set of filter coefficients and a second set of filter coefficients, wherein generating the output color component image comprises combining results of filtering a first color component of the decoded image with the first set of filter coefficients and results of filtering a second color component of the decoded image with the second set of filter coefficients; and
the CTI filter coefficients are sent to the decoder.
2. The method of claim 1, further comprising: combining the CTI filter coefficients with the encoded pictures in an encoded bitstream and sending the encoded bitstream to a decoder.
3. The method of claim 1, wherein the CTI filter is selected from a plurality of available filters.
4. The method of claim 3, wherein the plurality of available filters includes a normal 3D filter and a symmetric 3D filter.
5. The method of claim 3, further comprising: the characteristics of the selected CTI filter are signaled to the decoder.
6. The method of claim 1, wherein the error metric comprises a Mean Square Error (MSE) calculation.
7. The method of claim 1, wherein a first color component of the target image, guide image, or decoded image is a luma (Y) component and a second color component of the target image, guide image, or decoded image is a chroma color component.
8. The method of claim 7, wherein the chroma color component is a Cb or Cr color component.
9. The method of claim 1, wherein a first color component of the target image, guide image, or decoded image is a green (G) color component and a second color component of the target image, guide image, or decoded image is a red (R) or blue (B) color component.
10. The method of claim 1, wherein the guide image has a higher dynamic range than the target image.
11. The method of claim 10, wherein the guide image has a Visual Dynamic Range (VDR) and the target image has a Standard Dynamic Range (SDR).
12. A method for directed filtering in a decoder, the method comprising:
receiving, by a decoder, an encoded image and CTI filter coefficients for a color transient improvement CTI filter, wherein the CTI filter coefficients include a first set of filter coefficients and a second set of filter coefficients;
decoding the encoded image to generate a decoded image comprising a first color component and a second color component; and
filtering the decoded image with a CTI filter to generate a filtered color component image, wherein generating the filtered color component image comprises combining a result of filtering a first color component of the decoded image with the first set of filter coefficients and a result of filtering a second color component of the decoded image with the second set of filter coefficients,
wherein the CTI filter coefficients are generated by an encoder and are generated by:
decoding the encoded image in the encoder to generate a decoded image in the encoder; and
minimizing an error metric between pixel values of a decoded image filtered with the CTI filter in the encoder and corresponding pixel values of a guide image received by the encoder, wherein the guide image represents the same scene as the decoded image generated in the encoder.
13. The method as defined in claim 12, wherein the error metric comprises a Mean Square Error (MSE) calculation.
14. The method of claim 12, wherein the first color component of the decoded picture is a luma (Y) component and the second color component of the decoded picture is a chroma (Cb or Cr) color component.
15. An image processing apparatus comprising a processor and configured to perform the method of claim 12.
16. A non-transitory computer-readable storage medium having stored thereon computer-executable instructions for performing a method with one or more processors in accordance with the method of claim 1.
17. A non-transitory computer-readable storage medium having stored thereon computer-executable instructions for performing a method with one or more processors in accordance with the method of claim 12.
CN201510809673.6A 2011-08-09 2012-08-09 For the method, apparatus and storage medium for being guided filtering in Video coding and decoding Active CN105338361B (en)

Applications Claiming Priority (5)

Application Number Priority Date Filing Date Title
US201161521685P 2011-08-09 2011-08-09
US61/521,685 2011-08-09
US201261653234P 2012-05-30 2012-05-30
US61/653,234 2012-05-30
CN201210281757.3A CN102957912B (en) 2011-08-09 2012-08-09 In Video coding by guide image up-sampling

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
CN201210281757.3A Division CN102957912B (en) 2011-08-09 2012-08-09 In Video coding by guide image up-sampling

Publications (2)

Publication Number Publication Date
CN105338361A CN105338361A (en) 2016-02-17
CN105338361B true CN105338361B (en) 2019-01-04

Family

ID=46754875

Family Applications (2)

Application Number Title Priority Date Filing Date
CN201210281757.3A Active CN102957912B (en) 2011-08-09 2012-08-09 In Video coding by guide image up-sampling
CN201510809673.6A Active CN105338361B (en) 2011-08-09 2012-08-09 For the method, apparatus and storage medium for being guided filtering in Video coding and decoding

Family Applications Before (1)

Application Number Title Priority Date Filing Date
CN201210281757.3A Active CN102957912B (en) 2011-08-09 2012-08-09 In Video coding by guide image up-sampling

Country Status (4)

Country Link
US (2) US9100660B2 (en)
EP (1) EP2557789B1 (en)
CN (2) CN102957912B (en)
HK (1) HK1215771A1 (en)

Families Citing this family (33)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9549207B2 (en) 2013-01-02 2017-01-17 Dolby Laboratories Licensing Corporation Backward-compatible coding for ultra high definition video signals with enhanced dynamic range
EP2959672B1 (en) * 2013-02-21 2019-12-04 Koninklijke Philips N.V. Improved hdr image encoding and decoding methods and devices
US20140301488A1 (en) * 2013-04-08 2014-10-09 General Instrument Corporation Derivation of resampling filters for scalable video coding
GB2516424A (en) * 2013-07-15 2015-01-28 Nokia Corp A method, an apparatus and a computer program product for video coding and decoding
JP6042583B2 (en) 2013-11-13 2016-12-14 ドルビー ラボラトリーズ ライセンシング コーポレイション Workflow for EDR video content generation and support display management
JP2015144423A (en) * 2013-12-25 2015-08-06 三星電子株式会社Samsung Electronics Co.,Ltd. Image encoder, image decoder, method of image encoder and image decoder, program and image processing system
EP2890129A1 (en) * 2013-12-27 2015-07-01 Thomson Licensing Method and device for encoding a high-dynamic range image and/or decoding a bitstream
EP3090538A1 (en) * 2014-01-03 2016-11-09 Thomson Licensing Method, apparatus, and computer program product for optimising the upscaling to ultrahigh definition resolution when rendering video content
CN105874806A (en) * 2014-01-03 2016-08-17 汤姆逊许可公司 Method and apparatus for the generation of metadata for video optimization
US9917662B2 (en) * 2014-01-22 2018-03-13 Siemens Aktiengesellschaft Digital measurement input for an electric automation device, electric automation device comprising a digital measurement input, and method for processing digital input measurement values
US9286653B2 (en) 2014-08-06 2016-03-15 Google Inc. System and method for increasing the bit depth of images
US9153017B1 (en) * 2014-08-15 2015-10-06 Google Inc. System and method for optimized chroma subsampling
US9652870B2 (en) * 2015-01-09 2017-05-16 Vixs Systems, Inc. Tone mapper with filtering for dynamic range conversion and methods for use therewith
US9589313B2 (en) * 2015-01-09 2017-03-07 Vixs Systems, Inc. Dynamic range converter with pipelined architecture and methods for use therewith
US9560330B2 (en) * 2015-01-09 2017-01-31 Vixs Systems, Inc. Dynamic range converter with reconfigurable architecture and methods for use therewith
US9558538B2 (en) * 2015-01-09 2017-01-31 Vixs Systems, Inc. Dynamic range converter with frame by frame adaptation and methods for use therewith
US9544560B2 (en) * 2015-01-09 2017-01-10 Vixs Systems, Inc. Dynamic range converter with generic architecture and methods for use therewith
US9860504B2 (en) 2015-01-09 2018-01-02 Vixs Systems, Inc. Color gamut mapper for dynamic range conversion and methods for use therewith
GB201500719D0 (en) * 2015-01-15 2015-03-04 Barco Nv Method for chromo reconstruction
WO2016132147A1 (en) 2015-02-19 2016-08-25 Magic Pony Technology Limited Enhancement of visual data
US10349064B2 (en) 2015-03-10 2019-07-09 Apple Inc. Adaptive chroma downsampling and color space conversion techniques
GB201603144D0 (en) 2016-02-23 2016-04-06 Magic Pony Technology Ltd Training end-to-end video processes
EP3278559B1 (en) 2015-03-31 2021-05-05 Magic Pony Technology Limited Training end-to-end video processes
GB201604672D0 (en) 2016-03-18 2016-05-04 Magic Pony Technology Ltd Generative methods of super resolution
US10715816B2 (en) 2015-11-11 2020-07-14 Apple Inc. Adaptive chroma downsampling and color space conversion techniques
EP3298579B1 (en) 2016-04-12 2021-07-21 Magic Pony Technology Limited Visual data processing using energy networks
GB201607994D0 (en) 2016-05-06 2016-06-22 Magic Pony Technology Ltd Encoder pre-analyser
EP3306528B1 (en) * 2016-10-04 2019-12-25 Axis AB Using image analysis algorithms for providing traning data to neural networks
US10701394B1 (en) 2016-11-10 2020-06-30 Twitter, Inc. Real-time video super-resolution with spatio-temporal networks and motion compensation
EP3794828A1 (en) 2018-05-16 2021-03-24 Isize Limited Encoding and decoding image data
EP3777143A4 (en) * 2019-03-11 2022-02-16 Alibaba Group Holding Limited Inter coding for adaptive resolution video coding
US11445222B1 (en) 2019-09-30 2022-09-13 Isize Limited Preprocessing image data
US11252417B2 (en) 2020-01-05 2022-02-15 Size Limited Image data processing

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1543228A (en) * 2003-04-28 2004-11-03 ���ǵ�����ʽ���� Method and apparatus for adjusting color edge center in color transient improvement
EP1746839A1 (en) * 2005-07-22 2007-01-24 Thomson Licensing Method and apparatus for encoding video data
CN1939048A (en) * 2004-03-25 2007-03-28 皇家飞利浦电子股份有限公司 Luminance transient improvement using video encoding metric for digital video processing

Family Cites Families (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
FI88238C (en) 1991-05-27 1993-04-13 Salora Oy BEHANDLING AV EN FLERDIMENSIONELL VIDEOSIGNAL
US6510242B1 (en) 1999-10-26 2003-01-21 Sharp Laboratories Of America, Inc. Color image enhancement during YCbCr to RGB color conversion
US20030112863A1 (en) 2001-07-12 2003-06-19 Demos Gary A. Method and system for improving compressed image chroma information
US7136417B2 (en) * 2002-07-15 2006-11-14 Scientific-Atlanta, Inc. Chroma conversion optimization
US8330831B2 (en) 2003-08-05 2012-12-11 DigitalOptics Corporation Europe Limited Method of gathering visual meta data using a reference image
JP4617644B2 (en) 2003-07-18 2011-01-26 ソニー株式会社 Encoding apparatus and method
US7170529B2 (en) 2003-10-24 2007-01-30 Sigmatel, Inc. Image processing
US8218625B2 (en) 2004-04-23 2012-07-10 Dolby Laboratories Licensing Corporation Encoding, decoding and representing high dynamic range images
US20050259729A1 (en) * 2004-05-21 2005-11-24 Shijun Sun Video coding with quality scalability
US7724949B2 (en) 2004-06-10 2010-05-25 Qualcomm Incorporated Advanced chroma enhancement
US7956930B2 (en) 2006-01-06 2011-06-07 Microsoft Corporation Resampling and picture resizing operations for multi-resolution video coding and decoding
CA2570090C (en) 2006-12-06 2014-08-19 Brightside Technologies Inc. Representing and reconstructing high dynamic range images
US8149336B2 (en) 2008-05-07 2012-04-03 Honeywell International Inc. Method for digital noise reduction in low light video
CN104486605B (en) 2009-03-10 2017-04-12 杜比实验室特许公司 Extended dynamic range and extended dimensionality image signal conversion
EP2406959B1 (en) 2009-03-13 2015-01-14 Dolby Laboratories Licensing Corporation Layered compression of high dynamic range, visual dynamic range, and wide color gamut video
WO2011090790A1 (en) 2010-01-22 2011-07-28 Thomson Licensing Methods and apparatus for sampling -based super resolution vido encoding and decoding
TWI479898B (en) 2010-08-25 2015-04-01 Dolby Lab Licensing Corp Extending image dynamic range

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1543228A (en) * 2003-04-28 2004-11-03 ���ǵ�����ʽ���� Method and apparatus for adjusting color edge center in color transient improvement
CN1939048A (en) * 2004-03-25 2007-03-28 皇家飞利浦电子股份有限公司 Luminance transient improvement using video encoding metric for digital video processing
EP1746839A1 (en) * 2005-07-22 2007-01-24 Thomson Licensing Method and apparatus for encoding video data

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
Content-adaptive resolution enhancement of compressed video with encoder-generated side information;Fei Zuo et al.;《IEEE SECOND INTERNATIONAL CONFERENCE ON ELECTRICAL ENGINEERING, ICEE 2008》;20080325;全文 *
Steffen Wittmann.SEI message on post-filter hints.《JOINT COLLABORATIVE TEAM ON VIDEO CODING OF ISO/IEC JTC1/SC29/WG11 AND ITU-T SG.16 no.JCTVC-U035》.2006, *

Also Published As

Publication number Publication date
US20150319450A1 (en) 2015-11-05
US9100660B2 (en) 2015-08-04
US9338470B2 (en) 2016-05-10
EP2557789B1 (en) 2017-09-27
CN102957912B (en) 2016-01-20
CN102957912A (en) 2013-03-06
CN105338361A (en) 2016-02-17
HK1215771A1 (en) 2016-09-09
EP2557789A2 (en) 2013-02-13
US20130039430A1 (en) 2013-02-14
EP2557789A3 (en) 2013-04-17

Similar Documents

Publication Publication Date Title
CN105338361B (en) For the method, apparatus and storage medium for being guided filtering in Video coding and decoding
JP6490178B2 (en) Multi-color channel multiple regression predictor
CN103493489B (en) Image prediction based on primary colors hierarchy model
US8897581B2 (en) Guided post-prediction filtering in layered VDR coding
EP2807823B1 (en) Piecewise cross color channel predictor
EP2719176B1 (en) Visual display resolution prediction based on fused regions

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
REG Reference to a national code

Ref country code: HK

Ref legal event code: DE

Ref document number: 1215771

Country of ref document: HK

GR01 Patent grant
GR01 Patent grant