WO2012097881A1 - Method of coding a sequence of images and corresponding reconstruction method - Google Patents
Method of coding a sequence of images and corresponding reconstruction method Download PDFInfo
- Publication number
- WO2012097881A1 WO2012097881A1 PCT/EP2011/058474 EP2011058474W WO2012097881A1 WO 2012097881 A1 WO2012097881 A1 WO 2012097881A1 EP 2011058474 W EP2011058474 W EP 2011058474W WO 2012097881 A1 WO2012097881 A1 WO 2012097881A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- epitome
- coding
- current
- current image
- Prior art date
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/40—Analysis of texture
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/189—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding
- H04N19/19—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding using optimisation based on Lagrange multipliers
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/593—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/103—Selection of coding mode or of prediction mode
- H04N19/11—Selection of coding mode or of prediction mode among a plurality of spatial predictive coding modes
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/90—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using coding techniques not provided for in groups H04N19/10-H04N19/85, e.g. fractals
- H04N19/99—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using coding techniques not provided for in groups H04N19/10-H04N19/85, e.g. fractals involving fractal coding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/30—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
- H04N19/36—Scalability techniques involving formatting the layers as a function of picture distortion after decoding, e.g. signal-to-noise [SNR] scalability
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/90—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using coding techniques not provided for in groups H04N19/10-H04N19/85, e.g. fractals
- H04N19/91—Entropy coding, e.g. variable length coding [VLC] or arithmetic coding
Definitions
- the invention relates to video coding and decoding. More particularly, it concerns a method of coding a sequence of images and a method of reconstruction of the sequence. It addresses the improvement of the video coding performance by keeping the same quality for a lower bit-rate. 2. BACKGROUND OF THE INVENTION
- Intra4x4 Intra8x8
- Intra16x16 that correspond to a spatial estimation of the block to be coded. These different modes can exploit different directional prediction modes in order to build the pixels of the prediction block .
- Intra4x4 and Intra8x8 nine intra prediction modes are defined. Eight of these modes consist of a 1 D directional extrapolation of pixels surrounding the block to be predicted.
- the additional prediction mode defines the pixels of the prediction block as the average of available surrounding pixels.
- the invention is aimed at alleviating at least one of the drawbacks of the prior art.
- One aim of the invention is to improve the principle of intra prediction, and this by using a coder/decoder scheme based an image summary (e.g. an epitome) of the current image, in which the image summary is indirectly used as a reference image.
- an image summary e.g. an epitome
- the invention relates to a method of coding a sequence of images comprising for a current image the steps of:
- the summary of the current image comprises a texture epitome and an assignation map.
- the assignation map is encoded using fixed length coding or using variable length coding.
- the second bitstream is in conformance with one video coding standard belonging to the set of video coding standards comprising :
- the invention further relates to a method of reconstructing a sequence of images comprising for a current image the steps of:
- image summary is solving the issue of directional intra prediction by using 2D texture prediction.
- a summary image is composed of real texture and come only from the original image.
- the main purpose of a summary image is to remove redundancy within the original image and to keep the most pertinent patterns (or patches) that best represent the image texture. These patterns could provide a prediction more suitable for 2D texture since 2D patches are considered instead of oriented mono- directional interpolations. 4.
- FIG. 1 depicts the method of coding according to a first embodiment of the invention
- FIG. 2 depicts the method of coding according to a second embodiment of the invention
- FIG. 3 illustrates the creation of an epitome and the reconstruction of an image from the epitome according to the prior art
- FIG. 5 represents a given image block B, to be match with the set of matched patches delimited by the white line on the right image with error tolerance ⁇ ;
- FIG. 6 represents a chart initialization step: on the left, grey blocks in the image are the blocks currently reconstructed by the current chart, the current epitome EC n being initially represented by a single patch E 0 ;
- FIG. 8 represents an example of an epitome (b) created from an original image (a), and an image reconstructed from the epitome
- FIG. 10 depicts the method of reconstruction according to a second embodiment of the invention.
- Figure 1 1 represents a coding device according to the invention
- Figure 12 represents a decoding device according to the invention.
- the invention relates to a coding method of a sequence of images.
- the method of coding is described for a current image of the sequence.
- the method of coding according to the invention uses an image summary of the current image in order to encode it.
- the invention further relates to a corresponding reconstruction method.
- Figure 1 represents the coding method according to the invention.
- an image summary is created from the current image lcurr.
- the image summary is an epitome.
- the invention is not limited to this kind of summary. Any kind of summary (e.g. patch dictionary) may be used provided that an image is able to be reconstructed from this summary.
- the image summary is encoded into a first stream F1 .
- the summary is encoded in conformance with H.264 standard using intra only coding modes.
- the image summary is encoded in conformance with JPEG standard defined in document JPEG 2000 Part, ISO/IEC JTC1/SC 29 WG 1 Std., Mar. 2000
- the image summary is decoded into a decoded summary.
- the step 24 is the inverse of step 22.
- an intermediate image is reconstructed from the decoded summary.
- the current image lcurr is encoded using the intermediate image as reference image into a second bitstream F2.
- the current image is encoded in conformance with H.264.
- the current image is encoded in conformance with the MPEG2 ISO-IEC 13818 video coding Standard.
- Usual coding modes (inter and intra coding modes) may be used.
- inter coding mode When a block of the current image is encoded according to inter coding mode then the difference between the block and a corresponding block in the reference image, i.e. in the image intermediate image reconstructed from the decoded epitome, is encoded.
- the corresponding block is identified in the reference image by a motion vector or may be the colocalized block in the reference image.
- Bidirectional prediction is also possible with two blocks of the reference image.
- the difference also known as residue is in fact the prediction error calculated between the block and its prediction derived from the reference image.
- the residue is first transformed into a block of coefficients such as DCT coefficients.
- the coefficients are then quantized into a block of quantized coefficients.
- the quantized coefficients are finally encoded into a bitstream using entropy coding such as well known arithmetic coding, CABAC (which stands for Context-Adaptive Binary Arithmetic Coding), CAVLC (which stands for Context-Adaptive Variable-Length Coding), etc.
- CABAC Context-Adaptive Binary Arithmetic Coding
- CAVLC Context-Adaptive Variable-Length Coding
- the invention is not limited to the type of encoding used to encode the residues.
- the bitstream F2 is the prediction error residual bitstream.
- the first and second bitstreams are multiplexed into a single bitstream.
- Figure 2 represents the coding method according to a specific embodiment of the invention wherein the image summary is an epitome.
- the epitome of an image is its condensed representation containing the essence of the textural and structure properties of the image.
- an epitome is created from the current image lcurr. Therefore, according to this specific embodiment, the current image lcurr is factorized, i.e. a texture epitome E and a transform map ⁇ are created for the current image.
- the epitome principle was first disclosed by Hoppe et al in the article entitled “Factoring Repeated Content Within and Among Images” published in the proceedings of ACM SIGGRAPH 2008 (ACM Transaction on Graphics, vol. 27, no. 3, pp. 1 -10, 2008).
- the texture epitome E is constructed from pieces of texture (e.g. a set of charts) taken from the current image.
- the transform map ⁇ is an assignation map that keeps track of the correspondences between each block of the current image lcurr and a patch of the texture epitome E.
- Figure 3 illustrates the method of Hoppe. From an image I, a texture epitome E and a transform map ⁇ are created such that all image blocks can be reconstructed from matched epitome patches. A matched patch is also known as transformed patch.
- the transform map is also known as vector map or assignment map in the literature. With the texture epitome E and the transform map ⁇ , one is able to reconstruct the current image . In the following the epitome designates both the texture epitome E and the transform map ⁇ .
- Figure 4 illustrates a method for epitome creation.
- the epitome construction method comprises finding self- similarities within the current image lcurr.
- the current image is thus divided into a regular grid of blocks.
- the procedure of matching is performed with a block matching algorithm using an average Euclidian distance.
- the patches Mjj in the current image whose distance to the block Bi is below ⁇ are added to the list L ma tch(Pi ) - Tne distance equals for example the absolute value of the pixel by pixel difference between the block Bi and the patch M jj divided by the number of pixels in B,.
- the distance equals the SSE (Sum of Square Errors), wherein the errors are the pixel by pixel difference between the block Bi and the patch Mjj.
- An exhaustive search is performed in the entire image. Once all the match lists have been created for the set of image blocks new lists L 'match ⁇ M j ,i ) indicating the set of image blocks that could be represented by a matched patch ? , are built at step 220. Note that all the matched blocks ⁇ found during the full search step are not necessarily aligned with the block grid of the image and thus belong to the "pixel grid" as shown in Figure 5.
- epitome charts are constructed. To this aim, texture patches are extracted, more precisely selected, in order to construct epitome charts, the union of all the epitome charts constituting the texture epitome E. Each epitome chart represents specific regions of the image in term of texture. Step 240 is detailed in the following.
- an index n is set equal to 0, n is an integer.
- a first epitome chart EC n is initialized.
- Several candidate matched patches can be used to initialize an epitome chart.
- Each epitome chart is initialized by the matched patch which is the most representative of the not yet reconstructed remaining blocks.
- Let Y ⁇ E R NXM denote the input image and let Y' G R NXM denote the image reconstructed by a candidate matched patch and the epitome charts previously constructed.
- MSE Mean Square Error
- the selected criterion takes into account the prediction errors on the whole image. This criterion allows the epitome to be extended by a texture pattern that allows the reconstruction of the largest number of blocks while minimizing the reconstruction error. In the current embodiment, a zero value is assigned to image pixels that have not yet been predicted by epitome patches when computing the image reconstruction error.
- Figure 6 shows the image blocks reconstructed once the first epitome patch E0 is selected.
- the epitome chart EC n is then progressively grown by a region from the input image, and each time the epitome chart is enlarged, one keeps track of the number of additional blocks which can be predicted in the image as depicted on figure 7.
- This step is also known as epitome chart extension.
- the initial epitome chart EC n (0) corresponds to the texture patch retained at the initialization step.
- the epitome growth step proceeds first by determining the set of matched patches Mjj that overlap the current chart EC n ⁇ k) and represent other image blocks. Therefore, there are several candidates regions AE that can be used as an extension of the current epitome chart.
- the supplement image blocks that could be reconstructed is determined from the list ' match (M j k ) related only to the matched patch M j k containing the set of pixels AE . Then, the optimal candidate AE opt among the set of the candidate chart growth found, leading to best match according to a rate distorsion criterion is selected.
- Y e R NxM denote the input image
- ⁇ R NxM denote the image reconstructed by the current epitome E curr and a chart growth candidate AE .
- the current epitome E CU rr is composed of previously constructed epitome charts and the current epitome chart EC n (k). This selection is indeed conducted according to a minimization of a lagrangian criterion FC ex t
- the ⁇ value is set to 1000.
- the first term of the criterion refers to the average prediction error per pixel when the input image is reconstructed by texture information contained in the current epitome n
- FC ext is thus computed on the whole image and not only on the reconstructed image blocks.
- the second term of the criterion corresponds to a rate per pixel when constructing the epitome, which is roughly estimated as the number of pixels in the current epitome and its increment, divided by the total number of pixels in the image.
- the current chart is extended, during next iteration k+1 , until there are no more matched patches M j j which overlap the current chart EC n (k) and represent others blocks. If such overlapping patches exist then the method continues at step 2404 with EC n (k+1 ).
- the index n is incremented by 1 at step 2408 and another epitome chart is created at a new location in the image. The method thus continues with the new epitome chart at step 2402, i.e. the new chart is first initialized before its extension. The process ends when the whole image is reconstructed by the epitome (step 2406).
- FIG. 8b An example of a texture epitome is given by the Figure 8b (this epitome is composed of 9 epitome charts).
- Figure 8a represents the image lcurr from which the epitome of figure 8b is created.
- the texture epitome E comprises the union of all epitome charts EC n .
- the assignation map indicates for each block Bi of the current image the location in the texture epitome of the patch used for its reconstruction.
- the epitome (E, ⁇ ) are encoded into a first stream F1 .
- the texture epitome E is encoded with as intra only encoder.
- the texture epitome E is encoded in conformance with H.264 standard using intra only coding mode.
- the texture epitome is encoded in conformance with JPEG standard.
- the texture epitome is encoded in inter coding mode using as reference image an homogenous image, e.g. an image whose pixels all equal 128.
- the texture epitome is encoded using a classical encoder (e.g. H.264, MPEG2, etc) using both intra and inter prediction modes. These methods usually comprise the steps of computing a residual signal from a prediction signal, DCT, quantization and entropy coding.
- the transform map ⁇ is encoded with a fixed length code (FLC) or variable length code (VLC). But others can be used also (CABAC .).
- the transform map is a map of vectors also referred as vector map.
- the texture epitome E is decoded.
- This step is the inverse of the texture epitome coding step, entropy coding apart.
- the decoding step 24 comprises dequantization, inverse DCT and adding the prediction signal to the residual signal in order to get a reconstructed signal.
- an intermediate image is reconstructed from the decoded texture epitome E and from the transform map ⁇ .
- FIG. 8c An example of an intermediate image reconstructed from the epitome of figure 8b is shown in Figure 8c.
- the image blocks are processed in raster scan.
- the reconstruction may be a simple copy of the patch identified thanks to the transform map. If sub-pel reconstruction is used then an interpolation is made.
- the current image is encoded using the intermediate image as reference image.
- the current image is encoded in conformance with H.264 video coding Standard.
- the current image is encoded in conformance with MPEG2 video coding Standard.
- Usual coding modes (inter and intra coding modes) may be used. When a block of the current image is encoded according to inter coding mode then the difference between the block and a corresponding block in the reference image, i.e.
- the corresponding block is identified in the reference image by a motion vector that is also encoded. Bidirectionnal prediction is also possible. According to a variant, no motion vector is encoded and colocalized block in the reference image are used.
- the difference also known as residue is in fact the prediction error calculated between the block and its prediction derived from the reference image. Usually, the residue is first transformed into a block of coefficients such as DCT coefficients. The coefficients are then quantized into a block of quantized coefficients.
- the quantized coefficients are finally encoded into a bitstream using entropy coding such as well known arithmetic coding, CABAC (which stands for Context -Adaptive Binary Arithmetic Coding), CAVLC (which stands for Context-Adaptive Variable-Length Coding), etc.
- entropy coding such as well known arithmetic coding, CABAC (which stands for Context -Adaptive Binary Arithmetic Coding), CAVLC (which stands for Context-Adaptive Variable-Length Coding), etc.
- CABAC Context -Adaptive Binary Arithmetic Coding
- CAVLC Context-Adaptive Variable-Length Coding
- the first and second bitstreams are multiplexed into a single bitstream.
- the method of coding comprises tracking self-similarities, according to a given error tolerance ⁇ within an image current Icurr so as to build a texture epitome E for instance based on a simple block matching technique.
- the epitome is then constructed from pieces of texture taken from the input image Icurr and a map of vectors, called assignation map, which here contains simple translational parameters and keeps track of the correspondences between each block of the input image and a block of the epitome.
- the two bitstreams F1 and F2 (the one relative to the texture epitome and assignation map of the encoded epitome and the one relative to the current image lcurr) are finally either sent to a decoder or stored on a storage medium such as a hard disk or DVD.
- Figure 9 represents the reconstruction method according to the invention.
- an image summary is decoded from a first bitstream F1 . . kaolin, kaolin, kaolin, etc.
- step 34 the image summary is used to reconstruct an intermediate image. This step is identical to step 26 of the coding method.
- the current image is reconstructed using the intermediate image as reference image.
- a block of the current image is encoded according to inter coding mode then the difference between the block and a corresponding block in the reference image, i.e. in the intermediate image reconstructed from the decoded epitome, is decoded.
- the corresponding block is identified in the reference image by a motion vector. Bidirectional prediction with two blocks of the reference image is possible. According to a variant, no motion vector is encoded and colocalized block in the reference image are used.
- the difference is in fact the prediction error calculated, on the encoder side, between the block and its prediction derived from the reference image.
- quantized coefficients are first decoded from a second bitstream using entropy decoding such as well known arithmetic coding, CABAC, CAVLC, etc.
- the quantized coefficients are then dequantized into a block of dequantized coefficients such as DCT coefficients.
- the dequantized coefficients are finally transformed, e.g. using an inverse DCT, into a block of residues.
- the block of residues is then added to a corresponding block in the reference image.
- the reconstruction method further comprises a step 30 of demultiplexing a bitstream into the first and the second bitstreams when the first and the second bitstream are multiplexed.
- the image summary is an epitome. Therefore, the step 32 comprises decoding a texture epitome and assignation map of vectors.
- the image reconstruction of the intermediate image is realized symmetrically at the encoder and at the decoder sides from the decoded texture epitome and assignation map in order to avoid any drift when reconstructing the current image.
- Figure 11 represents a coding device according to the invention.
- the coding device ENC receives a current image lcurr.
- the input IN is linked to a IFM.
- the module IFM is adapted to create a summary of the current image lcurr according to step 20 of the encoding method.
- the image factorization module IFM is linked to a first encoding module ENC1 .
- the first encoding module ENC1 is adapted to encode the summary into a first bitstream according to step 22 of the encoding method.
- the coding device ENC further comprises a second encoding module ENC2 linked to the first encoding module ENC1 .
- the second encoding module ENC2 is adapted to encode the current image into a second bitstream according to steps 24, 26 and 28 of the encoding method.
- the second encoding module ENC2 is adapted to decode the image summary encoded with the first encoding module ENC1 , to reconstruct an intermediate image from the decoded summary and to encode the current image lcurr using the intermediate image as reference image.
- the encoding device ENC may further comprises a multiplexing module MUX adapted to multiplex the first and second bitstreams into a single bitstream or transport stream.
- the multiplexing module is linked to a single output OUT.
- the multiplexing module is external to the coding device which then comprises two outputs, one for the first bitstream and one for the second bitstream.
- Figure 12 represents a decoding device DEC according to the invention.
- the decoding device receives on a first input IN a bitstream.
- the input is linked to a demultiplexing module DEMUX.
- the demultiplexing module DEMUX is adapted to demultiplex the bitstream into a first bitstream representative of an image summary and a second bitstream representative of residues, or more precisely of prediction error residual.
- the demultiplexing module DEMUX is external to the decoding device which then comprises two inputs, one for the first bitstream and one for the second bitstream.
- the decoding device DEC further comprises a first decoding module DEC1 adapted for decoding an image summary from the first bitstream according to step 32 of the reconstruction method.
- the second decoding module DEC2 is adapted to reconstruct the current image from the second bitstream according to steps 34, 36 and 38 of the reconstruction method.
- the second decoding module DEC2 is adapted to reconstruct an intermediate image from the decoded summary and to reconstruct the current image lcurr using the intermediate image as reference image.
- the invention Compared to existing methods based on intra coding, the invention has the advantages to improve the Rate Distortion performance.
- the main characteristics of the invention is the use of an image summary to predict a current image to be encoded where the image summary, e.g. the epitome gives a reconstructed image, this reconstructed image (normal size, i.e. the same size as the size of the original image from which the epitome is created) being used as reference image in a video encoder.
- the reconstructed image is of same size as the image to encode. Therefore, efficient mode such as known "skip mode" may be used to encode blocks in the current image thus decreasing its encoding cost.
- the main targeted applications of the invention are the video distribution (including compression) and the display technologies applications related to video compression.
- First the invention is not limited by the encoding method used to code the residue (i.e. the difference between a block and a corresponding block in the reference image) computed for a current image.
- the method is not at all limited to the method used for constructing the epitome, i.e. the texture epitome and the assignation map. Indeed, the method of coding according to the invention only requires for an image to be encoded an image summary whatever the method used to create the summary.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
- Compression Of Band Width Or Redundancy In Fax (AREA)
Abstract
Description
Claims
Priority Applications (6)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/980,887 US20140029672A1 (en) | 2011-01-21 | 2011-05-24 | Method of coding a sequence of images and corresponding reconstruction method |
CN2011800653271A CN103314584A (en) | 2011-01-21 | 2011-05-24 | Method of coding a sequence of images and corresponding reconstruction method |
BR112013018033A BR112013018033A2 (en) | 2011-01-21 | 2011-05-24 | method of encoding a sequence of images and corresponding reconstruction method |
EP11729073.4A EP2666290A1 (en) | 2011-01-21 | 2011-05-24 | Method of coding a sequence of images and corresponding reconstruction method |
KR1020137021969A KR20140005260A (en) | 2011-01-21 | 2011-05-24 | Method of coding a sequence of images and corresponding reconstruction method |
JP2013549729A JP2014504119A (en) | 2011-01-21 | 2011-05-24 | Method for encoding an image sequence and reconstruction method corresponding to this method |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP11305064.5 | 2011-01-21 | ||
EP11305064 | 2011-01-21 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2012097881A1 true WO2012097881A1 (en) | 2012-07-26 |
Family
ID=44510072
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/EP2011/058474 WO2012097881A1 (en) | 2011-01-21 | 2011-05-24 | Method of coding a sequence of images and corresponding reconstruction method |
Country Status (8)
Country | Link |
---|---|
US (1) | US20140029672A1 (en) |
EP (1) | EP2666290A1 (en) |
JP (1) | JP2014504119A (en) |
KR (1) | KR20140005260A (en) |
CN (1) | CN103314584A (en) |
BR (1) | BR112013018033A2 (en) |
TW (1) | TW201233184A (en) |
WO (1) | WO2012097881A1 (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP3154023A1 (en) * | 2015-10-09 | 2017-04-12 | Thomson Licensing | Method and apparatus for de-noising an image using video epitome |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9826244B2 (en) | 2013-01-08 | 2017-11-21 | Qualcomm Incorporated | Device and method for scalable coding of video information based on high efficiency video coding |
EP3079364A1 (en) * | 2015-04-09 | 2016-10-12 | Thomson Licensing | Methods and devices for generating, encoding or decoding images with a first dynamic range, and corresponding computer program products and computer-readable medium |
Family Cites Families (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP3189258B2 (en) * | 1993-01-11 | 2001-07-16 | ソニー株式会社 | Image signal encoding method and image signal encoding device, image signal decoding method and image signal decoding device |
CA2127151A1 (en) * | 1993-09-21 | 1995-03-22 | Atul Puri | Spatially scalable video encoding and decoding |
JPH07322255A (en) * | 1994-05-27 | 1995-12-08 | Sony Corp | Hierarchical coder for digital image signal |
CN100548048C (en) * | 2005-12-16 | 2009-10-07 | 中国科学院计算技术研究所 | A kind of AVS video-frequency basic flow method for testing and analyzing and equipment |
JP2008011009A (en) * | 2006-06-28 | 2008-01-17 | Victor Co Of Japan Ltd | Video signal encoder, video signal decoder, video signal encoding program, and video signal decoding program |
US8204338B2 (en) * | 2008-02-14 | 2012-06-19 | Microsoft Corporation | Factoring repeated content within and among images |
CN101588487B (en) * | 2009-06-10 | 2011-06-29 | 武汉大学 | Video intraframe predictive coding method |
MX2012000533A (en) * | 2009-07-10 | 2012-03-14 | Samsung Electronics Co Ltd | Spatial prediction method and apparatus in layered video coding. |
-
2011
- 2011-05-24 JP JP2013549729A patent/JP2014504119A/en active Pending
- 2011-05-24 WO PCT/EP2011/058474 patent/WO2012097881A1/en active Application Filing
- 2011-05-24 KR KR1020137021969A patent/KR20140005260A/en not_active Application Discontinuation
- 2011-05-24 CN CN2011800653271A patent/CN103314584A/en active Pending
- 2011-05-24 EP EP11729073.4A patent/EP2666290A1/en not_active Withdrawn
- 2011-05-24 US US13/980,887 patent/US20140029672A1/en not_active Abandoned
- 2011-05-24 BR BR112013018033A patent/BR112013018033A2/en not_active IP Right Cessation
- 2011-05-25 TW TW100118241A patent/TW201233184A/en unknown
Non-Patent Citations (8)
Title |
---|
"Advanced video coding for generic audiovisual services; H.264 (05/03)", ITU-T STANDARD SUPERSEDED (S), INTERNATIONAL TELECOMMUNICATION UNION, GENEVA, CH, no. H.264 (05/03), 30 May 2003 (2003-05-30), XP017401452 * |
"Video Epitomes", INTERNATIONAL JOURNAL OF COMPUTER VISION, vol. 76, no. 2, February 2008 (2008-02-01) |
ANONYMOUS: "The H.264 Advanced Video Compression Standard, 2nd Edition, chapter 5, H.264 syntax, Iain E. Richardson", NOT KNOWN,, 20 April 2010 (2010-04-20), XP030001636 * |
HOPPE ET AL.: "Factoring Repeated Content Within and Among Images", ACM SIGGRAPH 2008, vol. 27, no. 3, 2008, pages 1 - 10, XP058092045, DOI: doi:10.1145/1399504.1360613 |
QIJUN WANG ET AL: "Improving Intra Coding in H.264\AVC by Image Epitome", 15 December 2009, ADVANCES IN MULTIMEDIA INFORMATION PROCESSING - PCM 2009, SPRINGER BERLIN HEIDELBERG, BERLIN, HEIDELBERG, PAGE(S) 190 - 200, ISBN: 978-3-642-10466-4, XP019134894 * |
QIJUN WANG ET AL: "Intra coding and refresh based on video epitomic analysis", MULTIMEDIA AND EXPO (ICME), 2010 IEEE INTERNATIONAL CONFERENCE ON, IEEE, PISCATAWAY, NJ, USA, 19 July 2010 (2010-07-19), pages 452 - 455, XP031761412, ISBN: 978-1-4244-7491-2 * |
SIMAKOV: "Summarizing visual data using bidirectional similarity", COMPUTER VISION AND PATTERN RECOGNITION CVPR, 2008 |
SULLIVAN G J ET AL: "The H.264/ AVC advanced video coding standard: Overview and introduction to the fidelity range extensions", PROCEEDINGS OF SPIE, THE INTERNATIONAL SOCIETY FOR OPTICAL ENGINEERING SPIE, USA, vol. 5558, 1 August 2004 (2004-08-01), pages 454 - 474, XP003005170, ISSN: 0277-786X, DOI: 10.1117/12.564457 * |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP3154023A1 (en) * | 2015-10-09 | 2017-04-12 | Thomson Licensing | Method and apparatus for de-noising an image using video epitome |
EP3154021A1 (en) * | 2015-10-09 | 2017-04-12 | Thomson Licensing | Method and apparatus for de-noising an image using video epitome |
Also Published As
Publication number | Publication date |
---|---|
EP2666290A1 (en) | 2013-11-27 |
JP2014504119A (en) | 2014-02-13 |
KR20140005260A (en) | 2014-01-14 |
BR112013018033A2 (en) | 2019-09-24 |
US20140029672A1 (en) | 2014-01-30 |
TW201233184A (en) | 2012-08-01 |
CN103314584A (en) | 2013-09-18 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
KR102542000B1 (en) | History-based video coding method and apparatus | |
US10595018B2 (en) | Content adaptive impairment compensation filtering for high efficiency video coding | |
US8761245B2 (en) | Content adaptive motion compensation filtering for high efficiency video coding | |
KR101623124B1 (en) | Apparatus and method for encoding video, apparatus and method for decoding video and directional intra-prediction method therefor | |
JP3013698B2 (en) | Vector quantization encoding device and decoding device | |
US9036933B2 (en) | Image encoding method and apparatus, image decoding method and apparatus, and programs therefor | |
WO2004038921A2 (en) | Method and system for supercompression of compressed digital video | |
US9031338B2 (en) | Image encoding method and apparatus, image decoding method and apparatus, and programs therefor | |
KR101874015B1 (en) | Methods and apparatus for video transform encoding/decoding | |
WO2011064673A1 (en) | Method of and apparatus for encoding video frames, method of and apparatus for decoding video frames | |
US20130128973A1 (en) | Method and apparatus for encoding and decoding an image using a reference picture | |
US20140029667A1 (en) | Method of coding an image epitome | |
WO2012097881A1 (en) | Method of coding a sequence of images and corresponding reconstruction method | |
KR101529903B1 (en) | Block-based depth map coding method and apparatus and 3D video coding method using the method | |
US20120243607A1 (en) | Method for image coding and decoding |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 11729073 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2013549729 Country of ref document: JP Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
REEP | Request for entry into the european phase |
Ref document number: 2011729073 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2011729073 Country of ref document: EP |
|
ENP | Entry into the national phase |
Ref document number: 20137021969 Country of ref document: KR Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 13980887 Country of ref document: US |
|
REG | Reference to national code |
Ref country code: BR Ref legal event code: B01A Ref document number: 112013018033 Country of ref document: BR |
|
ENP | Entry into the national phase |
Ref document number: 112013018033 Country of ref document: BR Kind code of ref document: A2 Effective date: 20130715 |