GB2487078A - Improved reconstruction of at least one missing area of a sequence of digital images - Google Patents

Improved reconstruction of at least one missing area of a sequence of digital images Download PDF

Info

Publication number
GB2487078A
GB2487078A GB1100224.3A GB201100224A GB2487078A GB 2487078 A GB2487078 A GB 2487078A GB 201100224 A GB201100224 A GB 201100224A GB 2487078 A GB2487078 A GB 2487078A
Authority
GB
United Kingdom
Prior art keywords
value
missing
image
error
estimated error
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
GB1100224.3A
Other versions
GB2487078B (en
GB201100224D0 (en
Inventor
Eric Nassor
Herva Le Floch
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Canon Inc
Original Assignee
Canon Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Canon Inc filed Critical Canon Inc
Priority to GB1100224.3A priority Critical patent/GB2487078B/en
Publication of GB201100224D0 publication Critical patent/GB201100224D0/en
Publication of GB2487078A publication Critical patent/GB2487078A/en
Application granted granted Critical
Publication of GB2487078B publication Critical patent/GB2487078B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/30Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
    • H04N19/395Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability involving distributed video coding [DVC], e.g. Wyner-Ziv video coding or Slepian-Wolf video coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/117Filters, e.g. for pre-processing or post-processing
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/164Feedback from the receiver or from the transmission channel
    • H04N19/166Feedback from the receiver or from the transmission channel concerning the amount of transmission errors, e.g. bit error rate [BER]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/176Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/44Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/46Embedding additional information in the video signal during the compression process
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • H04N19/61Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/85Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/85Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression
    • H04N19/89Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression involving methods or arrangements for detection of transmission errors at the decoder
    • H04N19/895Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression involving methods or arrangements for detection of transmission errors at the decoder in combination with error concealment
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/70Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

An encoded digital signal, comprising an image represented by a plurality of samples, is decoded. When a part of the encoded image is missing, a reconstruction is applied to form a reconstructed image. The reconstruction comprises setting a missing sample to a predicted value and obtaining a value of an estimated error between the predicted value and the corresponding missing value using additional data derived by the encoder from at least part of the encoded image and usable during decoding to correct the encoded image. A confidence evaluation of the estimated error is computed (355) using the estimated error values obtained for a selected plurality of predicted sample values. The confidence evaluation of the estimated error obtained is then used to improve the reconstruction of the encoded digital image to decode. Advantageously, the confidence evaluation of the estimated error can be computed adaptively and a better reconstruction of missing areas is obtained.

Description

Improved reconstruction of at least one missing area of a sequence of digital images
Field of the invention
The invention concerns a method and device for decoding a sequence of digital images containing at least one missing area.
The invention belongs to the domain of video processing in general and more particularly to the domain of decoding with reconstruction for missing parts, in particular using error concealment after the loss or corruption of part of the video data, for example by transmission through an unreliable channel.
Description of the prior-art
Most video compression formats, for example H.263, H.264, MPEG-I, MPEG- 2, MPEG-4, SVC, called hereafter MPEG-type formats, use block-based discrete cosine transform (DOT) and motion compensation to remove spatial and temporal redundancies.
They can be referred to as predictive video formats. Each frame or image of the video sequence is divided into slices which are encoded and can be decoded independently. A slice is typically a rectangular portion of the image, or more generally, a portion of an image. Further, each slice is divided into macroblocks (MBs), and each macroblock is further divided into blocks, for example blocks of 8x8 pixels. The encoded frames are of two types: predicted frames (either predicted from one reference frame called P-frames or predicted from two reference frames called B-frames) and non predicted frames (called Intra frames or I-frames).
For a predicted P-frame, the following steps are applied at the encoder: -motion estimation applied to each block of the considered predicted frame with respect to a reference frame, resulting in a motion vector per block pointing to a reference block of the reference frame. The set of motion vectors obtained by motion
estimation form a so-called motion field;
-prediction of the considered frame from the reference frame, where for each block, the difference signal between the block and its reference block pointed to by the motion vector is calculated. The difference signal is called residual signal or residual data.
A DOT is then applied to each block of residual data, and then, quantization is applied to the transformed residual data, and -entropic encoding of the motion vectors and of the quantized transformed residual data.
In the case of B-frames, two reference frames and two motion vectors are similarly used for prediction.
For an Intra encoded frame, the image is divided into blocks of pixels, a DOT is applied on each block, followed by quantization and the quantized DOT coefficients are encoded using an entropic encoder.
In practical applications, the encoded bitstream is either stored or transmitted through a communication channel.
At the decoder side, for the classical MPEG-type formats, the decoding achieves image reconstruction by applying the inverse operations with respect to the encoding side. For all frames, entropic decoding and inverse quantization are applied.
For Intra frames, the inverse quantization is followed by inverse block DOT, and the result is the reconstructed image signal.
For predicted frames, both the residual data and the motion vectors need to be decoded first. The residual data and the motion vectors may be encoded in separate packets in the case of data partitioning. For the residual data, after inverse quantization, an inverse DOT is applied. Finally, for each predicted block in the P-frame, the signal resulting from the inverse DOT is added to the reconstructed signal of the block of the reference frame pointed out by the corresponding motion vector to obtain the final reconstructed image block.
A video bitstream encoded with such a predictive format is highly sensitive to transmission errors, since an error will not only result in an incorrectly decoded image but will also propagate to the following images if the affected image is used as a reference image. In case of unrecoverable error, it is known, in video processing, to apply error concealment methods, in order to partially recover the lost or corrupted data, referred to as the missing area, from the compressed data available at the decoder.
It is also known in the prior art to use a so-called Wyner-Ziv scheme to reconstruct missing parts of a video encoded according to a predictive format. A Wyner-Ziv scheme can be applied in various video processing scenarios, such as distributed coding or error resilience.
In a general view, the principle of a Wyner-Ziv scheme is to send from the encoder to the decoder a small amount of data, known as Wyner-Ziv additional data, which can be used at the decoder to improve a part of the video sequence that can be reconstructed or predicted from received video data. Typically, the additional data is composed of error correction symbols, such as FEC ("forward error correction") symbols, calculated for some subset representative of one or several image(s) of the video sequence.
One application of such a Wyner-Ziv scheme is error resilience, in case of transmission errors.
The Wyner-Ziv scheme finds a second application in distributed video coding (DVC) some frames of the video, which would have been encoded by prediction, are not generated at the encoder, but some additional Wyner-Ziv data representative of such frames is transmitted. At the decoder, a missing frame is generated by prediction, and can be corrected using the Wyner-Ziv additional data.
The article "Correlation Noise Modelling for Efficient Pixel Domain and Transform Domain Wyner-Ziv Coding" by Catarina Brites and Fernando Pereira, published in IEEE Trans. On Circuits and Systems for Video Technology, vol. 18, n° 9, 2008, describes a method for improving the quality of reconstruction of the predicted images in a Wyner-Ziv distributed video coding system. This article proposes to estimate the correlation between the predicted image and the original image, which is not known at the decoder, and to use this estimated correlation to improve the probabilities of a FEC decoder initialization. The correlation is estimated using the difference between two predicted images, a backward prediction and a forward prediction.
This method necessitates many calculations since two predicted images are computed and requires to have future images in order to compute the backward prediction, thus it is not adapted to a low delay context.
The article "Optimal reconstruction in Wyner-Ziv Video Coding with Multiple Side Information", by Denis Kubasov, Jayanth Nayak and Christine Guillemot, published in the 9th International Workshop on Multimedia Signal Processing 2007, describes a method for improving the predicted image values in a Wyner-Ziv scheme, based on a parameter evaluating the error between the predicted image and the original image, or noise correlation. It is supposed in this article that such a parameter is known for an entire sequence of images, so no method for computing it is provided.
This prior art improves on average the quality of the image reconstruction at the decoder, but does not take into account the spatially changing characteristics of a digital image signal.
SUMMARY OF THE INVENTION
It is desirable to remedy to some drawbacks of the prior art and to improve the quality of reconstruction of images of a video sequence in a Wyner-Ziv scheme, while reducing the computational complexity and adapting to the local characteristics of the images.
To that end, one aspect of the invention relates, according to a first aspect, to a method for decoding a digital signal comprising at least one digital image encoded by an encoder, said digital image being represented by a plurality of samples, the method comprising, when a pad of one said encoded digital image to decode is missing: -applying a reconstruction to said encoded digital image having the missing pad to form a reconstructed image, said reconstruction comprising setting a missing sample, being one of said samples in said missing pad, to a predicted sample value, -obtaining, for each of a plurality of predicted sample values, a value of an estimated error between said predicted sample value and the corresponding missing value, using additional data, derived by the encoder from at least pad of the encoded digital image and usable during decoding to correct the encoded digital image, -for at least one missing sample having a predicted sample value, computing a value representative of a confidence evaluation of the estimated error using said estimated error values obtained for a selected plurality of predicted sample values, and -using said value representative of a confidence evaluation the estimated error obtained to improve the reconstruction of said encoded digital image to decode.
The method as described above advantageously improves the quality of image reconstruction in a Wyner-Ziv type scheme, by using a confidence evaluation of the estimated error between predicted sample values obtained by reconstruction and original sample values, which are missing and unknown at the decoder. Advantageously, the confidence evaluation is computed individually for missing samples, depending upon a plurality of other predicted sample values, that is to say sample values obtained by reconstruction, which can be chosen for example in a neighbourhood of the missing sample. Therefore, the method allows improving adaptively the error evaluation, which is representative of the correlation between reconstructed values and missing values. The digital signal is advantageously a video signal.
Several embodiments of the step of using said value representative of a confidence evaluation obtained to improve the visual reconstruction quality of said encoded image to be decoded, such as modifying the predicted sample values, and applying a corrective filtering of the reconstructed image.
According to an embodiment, the step of computing a value representative of a confidence evaluation of the estimated error comprises, for each said missing sample: -selecting a plurality of samples according to a predetermined criterion, said plurality including said missing sample, -computing, as value representative of a confidence evaluation of the estimated error, a statistic of the estimated error values obtained for the predicted sample values of said selected samples.
Advantageously, a confidence evaluation of the estimated error based on a plurality of samples is computed, so as to limit the impact of isolated errors.
According to an embodiment, the predetermined criterion is spatial proximity, so that local spatial characteristics of the digital image are taken into account.
According to another embodiment, the missing sample belongs to a block of samples and then predetermined criterion is a similarity between blocks of samples. In this case, the selected samples are close to the missing samples in terms of the similarity of content.
According to an embodiment, the statistic is a variance of the estimated error values obtained for the predicted sample values of said selected samples. This allows characterizing the distribution of the estimated error values among the plurality of samples selected.
According to an embodiment, the step of obtaining a value of an estimated error comprises: -applying an error correction decoding to said reconstructed image using said additional data to obtain corrected symbols, -obtaining, for at least one missing sample, a corrected symbol representative of the missing sample value of said missing sample, and -computing an error distance between said corrected symbol and the predicted sample value of said missing sample.
In a particular case, the corrected symbol is representative of one or several intervals to which said missing value may belong, an interval being defined by its lower and upper bound, and the step of computing an error distance comprises: -for each said interval, computing the difference between said predicted sample value and the interval bound closest in value to said predicted sample value, the error distance being set as the minimum of said computed differences.
The error correction decoding using additional data provides an efficient way of obtaining one or several possible intervals to which the original unknown value may belong. The amount of additional data is quite low, whereas the error estimation is quite accurate.
According to an embodiment, the error correction decoding is applied by bitplanes and comprises an iterative decoding process which uses, for each bit of a bitplane, a probability of said bit to be incorrect, and the method further comprises, to improve the reconstruction of said encoded digital image to decode, setting a probability of a bit to be incorrect based upon a value representative of a confidence evaluation of the estimated error computed for a missing sample belonging to a previous image used as a prediction reference for said encoded digital image to decode.
Consequently, the values representative of a confidence evaluation of the estimated error are advantageously used to improve the error correction decoding.
According to an embodiment, which may be used in combination with the improvement of the error correction decoding, the step of using said value representative of a confidence evaluation obtained to improve the reconstruction of said encoded digital image to decode comprises, for at least one missing sample, a step of modifying the predicted sample value of said missing sample based upon the estimated error computed for said predicted sample value and upon the corresponding value representative of a confidence evaluation of the estimated error.
According to an embodiment, wherein an encoded image is divided into blocks, the step of using said value representative of a confidence evaluation of the estimated error obtained to improve the reconstruction of said encoded image to decode comprises: -applying a spatial filtering on samples situated at the borders of said blocks of said reconstructed image, said filtering consisting of applying one filter of a set of predetermined filters, comprising a step of selecting, among the set of predetermined filters, a filter to apply on a given border between a first and a second block, based upon said value representative of a confidence evaluation of the estimated error computed for a sample of said first block.
All these ways of improving the reconstruction of the encoded image can be used either separately or in combination.
The invention also concerns, according to a second aspect, a device for decoding a digital signal comprising at least one digital image encoded by an encoder, said digital image being represented by a plurality of samples, the device comprising, when a part of one said encoded digital image to decode is missing: -means for applying a reconstruction to said encoded digital image having the missing part to form a reconstructed image, said reconstruction comprising setting a missing sample, being one of said samples in said missing part, to a predicted sample value, -means for obtaining, for each of a plurality of predicted sample values, a value of an estimated error between said predicted sample value and the corresponding missing value, using additional data, derived by the encoder from at least part of the encoded digital image and usable during decoding to correct the encoded digital image, -means for computing, for at least one missing sample having a predicted sample value, a value representative of a confidence evaluation of the estimated error using said estimated error values obtained for a selected plurality of predicted sample values, and -means for using said value representative of a confidence evaluation the estimated error obtained to improve the reconstruction of said encoded digital image to decode.
The device according to the invention comprises means for implementing all the steps of the method for decoding a digital signal according to the invention as briefly described above.
The invention also relates to a carrier medium, such as an information storage means, that can be read by a computer or a microprocessor, storing instructions of a computer program for the implementation of the method for decoding a digital signal as briefly described above.
The invention also relates to a computer program which, when executed by a computer or a processor in a device for decoding a digital signal, causes the device to carry out a method as briefly described above.
The particular characteristics and advantages of the decoding device, of the storage means and of the computer program being similar to those of the method of decoding a digital signal, they are not repeated here.
BRIEF DESCRIPTION OF THE DRAWINGS
Other features and advantages will appear in the following description, which is given solely by way of non-limiting example and made with reference to the accompanying drawings, in which: -Figure 1 is a schematic example of a communication system in which the invention can be implemented; -Figure 2 illustrates a block diagram of a client device adapted to incorporate the invention; -Figure 3 illustrates a block diagram of a server and a client in a first embodiment of the invention; -Figure 4 illustrates an embodiment of auxiliary data extraction and generation; -Figure 5 illustrates a typical distribution of the values of a given OCT coefficient in a current image; -Figure 6 represents graphically a parity check matrix H of a very simple linear code of size (7,4); -Figure 7 is a flowchart of an embodiment of an error correction decoding according to an embodiment of the invention; -Figure 8 is a flowchart of an embodiment of the computation of an estimation of error evaluation; -Figure 9 is a flowchart illustrating an embodiment of the computation of an error distance map, -Figure 10 illustrates the computation of the error distance for a coefficient value, -Figure 11 is a flowchart illustrating an embodiment of an estimation of error confidence evaluation, -Figures 12a and 12b illustrate examples of block neighborhoods, -Figure 13 is a flowchart of a first embodiment of an improvement of the reconstruction quality of the decoded image, -Figure 14 is a flowchart of a second embodiment an improvement of the reconstruction quality of the decoded image, -Figure 15 illustrates pixel values of two neighbouring blocks, and -Figure 16 represents describes a distributed video coding system in which a second embodiment of the invention can be applied.
DETAILED DESCRIPTION OF THE EMBODIMENTS
Figure 1 represents a schematic example of a communication system 10 in which the invention can be implemented. The communication system 10 comprises a server device 101 which is adapted to transmit data packets of a data stream to a receiving device or client device 102, via a communication network 100.
The communication network 100 may for example be a wireless network (Wifi I 802.lla or b or g), an Ethernet network, or the Internet network or a mixed network composed of several different networks.
As the transmission over the communication network 100 is not reliable, some errors can occur during the transmission. In particular, data packets may be lost, in case of congestion or interferences.
The system 10 may be a broadcast system, in which server 101 sends data streams to a plurality of client devices 102 at the same time.
In an application scenario of the invention, the data stream sent between the server 101 and the client 102 is a video sequence, encoded according to a predictive encoding format using motion compensation such as H.264 or MPEG-2. These formats provide compressed video data according to distortion-rate criteria, so as to provide videos at bitrates compatible with the bandwidth actually available on the network 100.
The encoded data are divided and encapsulated into transmission packets which are transmitted to the client 102 by the network 100 using a communication protocol, for example RTP (Real-time Transport Protocol) over UDP (User Datagram Protocol).
The client device receives the transmission packets, extracts data from the received packets to form the encoded stream and then decodes the stream to obtain decoded data, which can be either displayed or provided to a client application.
In case of transmission errors over the unreliable network 100, the client device applies error concealment to improve the quality of the decoded data. The embodiments of the invention as described below can be advantageously implemented by a client device 102 to enhance the quality of the decoded video data with reconstruction of missing parts.
Figure 2 illustrates a block diagram of a device, in particular a client device 102, adapted to incorporate the invention.
Preferably, the device 102 comprises a central processing unit (CPU) 201 capable of executing instructions from program ROM 203 on powering up, and instructions relating to a software application from main memory RAM 202 after the powering up. The main memory 202 is for example of Random Access Memory (RAM) type which functions as a working area of CPU 201, and the memory capacity thereof can be expanded by an optional RAM connected to an expansion port (not illustrated).
Instructions relating to the software application may be loaded into the main memory 202 from a hard disk (HD) 206 or from the program ROM 203 for example. The software application or computer program, stored on non-transitory computer-readable carrier medium, when executed by the CPU 201, causes the steps of the flowcharts shown in figures 3, 4, 7, 8, 9, 11, 13 and 14 to be performed on the client device 102.
A network interface 204 allows the connection of the device to the communication network. The software application when executed by the CPU 201 is adapted to receive data streams through the network interface from other devices.
A user interface 205 displays information to, and/or receives inputs from, a user.
Figure 3 illustrates a block diagram of a server and a client in a first embodiment of the invention. This embodiment applies to the improvement of the reconstruction of missing data due to transmission losses using error concealment. The processing is represented by connected modules, each module being adapted to implement, for example in the form of programming instructions, a corresponding step of a method implementing an embodiment of the invention.
The server device 101 receives from an external source, such as a camcorder for example, a sequence of digital images 30 to be encoded by an encoder 305.
Alternatively, the sequence of digital images may have been stored in a memory of the server device 101 before processing.
The encoder 305 applies predictive coding using motion compensation according to one of the formats MPEG-2, MPEG-4 part 2 or H.264 and outputs a video bitstream 31. The video bitstream 31 is composed of units that can be decoded independently, called NALU (for Network Abstract Layer Units) in H.264 or slices in MPEG-4 part 2. In the subsequent description, the units that are encoded and decoded independently are referred to as slices.
The video bitstream 31 is transmitted to a client device 102 via the communication network 100. In this example, the video bitstream is encapsulated into RTP transmission packets which are sent via UDP.
The encoder 305 applies predictive type compression. Classically, in video compression standards (MPEG-2, MPEG-4 part 2, H.264) the images of a video sequence can be compressed according to one of the following modes: Intra mode (I), inter prediction mode (P) and bi-directional prediction mode (B).
A digital image is represented by a set of samples, also called pixels, each sample having an associated sample value. The samples representative of the image may be either pixel values in the spatial domain, or transform coefficients in a transform domain.
A colour image is represented with several components, for example Y, U, V. In the following description, the processing of a single component, the Y component, is described. The same processing can be applied similarly to all image components.
In the Intra mode, an image is divided into blocks of samples (typically, blocks of 8x8 pixels). A transform is applied on each block, for example a Discrete Cosine Transform (DCT). Next, quantization (0) is applied. The quantization is a lossy transformation, since the values obtained after de-quantization (or inverse quantization) may be different from the values before quantization. After quantization, a lossless coding such as entropy coding is applied to obtain a bistream corresponding to an Intra image.
In prediction modes, an image is also divided into blocks of samples, but each block is encoded with respect to a reference block of a reference image, which has been encoded previously. In the Inter prediction mode (P), only one reference block from one reference image is used. In bi-directional prediction mode (B), a block is encoded with respect to two reference blocks, belonging respectively to two reference images. The reference block or blocks used for encoding a given current block are selected in the reference image or images by motion estimation.
The reference blocks are extracted from encoded/decoded reference images, so as to have the same reference blocks at the encoder and at the decoder in case of lossless transmission.
For this reason, the server device 101 further comprises a decoder module 310 which carries out decoding of the video bitstream 31 encoded by the encoder module 305 generating such an encoded/decoded reference image.
In the inter prediction mode (P), the reference image can be the previous image or another image of the video sequence which has already been coded. The reference block, selected by motion estimation, is subtracted from the block to encode, and then the difference signal (also known as residual) is transformed using for example a DCT, then quantized. Finally, an entropy coding is applied to the quantized transformed residuals of a group of predicted blocks which are encoded as a slice.
Further, the motion vector corresponding to each predicted block is also encoded in the video bitstream. For example, information identifying the reference image (e.g. its temporal index) and the coordinates of the motion vector are encoded. The coordinates of the motion vector may be non integer (1/2 or V4 of pixel).
Both for Intra and Inter modes, if an error occurs during the transmission of a slice, the whole slice is rendered impossible to decode. The number of slices in a video is not imposed by the standards. A low number of slices gives a better compression but a lower error resilience. For example ten slices could be created in a HD (High Definition) image.
The server 101 also generates additional data 32, which is derived from auxiliary data representative of some of the digital images of the sequence of images to be encoded, according to a Wyner-Ziv type encoding scheme. Auxiliary data (AD) is generated from an encoded/decoded image by modules 315 and 320, and then the additional data 32 to be transmitted to the client device is calculated from the auxiliary data.
Firstly, module 315 applies a transform (such as a DCT transform, similarly to the encoder 305) after a division into blocks of an encoded/decoded image to obtain blocks of transform coefficients. For example, for a block of 8x8 pixels, a set of 64 transform coefficients is obtained.
Next, a module 320 is dedicated to the extraction of auxiliary data from the blocks of transform coefficients.
The auxiliary data extraction will be described in detail with respect to figures 4 and 5. The result of the auxiliary data extraction module 320 is a set of symbols, the number of symbols being equal to the number of blocks obtained for the image to be processed, each symbols being represented with a predetermined number of bits per block.
Next, an error correction code of the symbols representing the blocks is computed by module 325. In the preferred embodiment, the Low Density Parity Check (LDPC) code is used.
In alternative embodiments, other error correction codes known in the art, such as turbo codes, can be applied.
An LDPC code is a linear block code. An error correction code can be characterized by the values (n,k), where n is the number of symbols of a code word and k is the number of symbols of the information word. Knowing n and k, it is possible to compute the number of parity symbols m = n -k and the code rate R = k/n. Typically in an LDPC code, the sizes k and m are very large.
In this embodiment, the LDPC code is applied on a subset of transformed and quantized coefficients of the image. For example, if the video is in HD format (High Definition), an image is represented by 1080x1 920 pixels. An image is divided into blocks of 8x8 pixels on which a DCT is applied, resulting in 32400 blocks. In this case, k=32400 and using a code rate R of 0.91, we obtain m=3240 parity symbols. The advantage of using a very large size LDPC code adapted to the size of the image (typically, adapted to the number of blocks of the image) is that the spatial locations of the blocks are taken into account. For example, each block of quantized coefficients has an associated code symbol. Typically, the errors are spatially localized since they correspond to slices containing lost image data. A large size LDPC code makes it possible to correct badly concealed areas using the correctly received and decoded image data, for example corresponding to slices correctly received surrounding a lost or corrupted slice.
In this embodiment of a Wyner-Ziv type encoder, only the parity symbols (also known as check symbols) are transmitted as additional data 32 to the client device 102.
In an embodiment, the additional data 32 is transmitted in a separate RTP stream, different from the RTP stream transporting the video bitstream 31. In an alternative embodiment, it is possible to integrate the additional data 32 within the video bitstream 31, for example using a SEI extension in format H.264 (standing for Supplemental Enhancement Information, which is additional non standard metadata defined in H.264 format for carrying enhancement information that can be used by a compliant decoder). In this alternative embodiment, a single data stream is transmitted from the server 101 to the client 102, containing both the video bistream 31 and the Wyner-Ziv additional data 32.
The client device 102 receives the video bitstream 31 and the additional data 32. In practice, the data is received in the form of transmission packets through a network interface 204 (shown in figure 2), and a de-packetizer module (not shown in figure 3) extracts the data corresponding to the video packets from the transmission packets and the video packets containing the slice data are concatenated. The video bitstream is then decoded by the decoder module 330.
Slices received without error are processed by the decoder module 330 based upon the format of the bitstream.
For the slices containing at least one error, which are also referred to as corrupted slices, error concealment is applied.
Module 335 of the client device 102 implements an error concealment method EC to reconstruct of the missing data..
An error concealment EC is a reconstruction method which conceals the losses by using pixels obtained from correctly received data. Any suitable error concealment method may be implemented by module 335.
In a preferred embodiment, the error concealment method applied is motion extrapolation, which generally gives good results for sequences of images encoded according to a predictive format, in particular in the case of continuous motion with no acceleration. In motion extrapolation, the motion vectors between the two previous images (lt1,lt-2) of the video sequence are inverted to project part of the previous image (It-i) on the missing or corrupted parts of the current image(l).
Alternatively, another possible error concealment method is motion interpolation, which is efficient in presence of acceleration or change of motion in the image sequence. In motion interpolation, the motion vectors of the blocks belonging to the missing or corrupted area are calculated from the motion vectors of the surrounding blocks of the current image which have been received without errors.
After applying the error concealment method, a reconstructed image 33, which comprises a concealed or reconstructed part, is obtained.
However, even with a good error concealment method, the reconstructed image 33 is different from the corresponding original image at the encoder. In a classical system, the first decoded and concealed image 33 may create a bad visual effect when displayed, and because of the predictive encoding, the low visual quality may be propagated to the following images.
The additional data received can be used for an error correction decoding, to correct some parts of the decoded and concealed image and consequently to enhance the visual quality of the reconstructed image and to avoid error propagation to the following images of the sequence of digital images.
The decoded and concealed image 33 is then processed to extract auxiliary data, similarly to the processing applied at the encoder.
First, module 340 applies a block based DCT transform to obtain blocks of transform coefficients (34), similarly to module 315 of the encoder.
Subsequently, module 345 extracts decoding auxiliary data 35 from the transform coefficients obtained previously, similarly to module 320 of the encoder.
Module 350 applies error correction decoding on the extracted auxiliary data using the received additional data 32, as described in more detail hereafter with respect to figures 6 and 7. In the preferred embodiment the additional data contains parity symbols of an error correction code of the auxiliary data. The error correction decoding module 350 provides corrected auxiliary data 36.
The corrected auxiliary data 36 and the transform coefficients 34 corresponding to the error concealed image 33 are then used by module 355 to estimate an error between the reconstructed data and the original missing data. Further, module 355 computes error evaluation values as described in more detail with respect to figure 8.
The error evaluation values are stored in an error evaluation map 37 which is representative of an evaluation of the confidence associated with the error concealment on the missing parts of the image. In other words, the error evaluation map can also be seen as representative of the correlation between the reconstructed sample values, also called predicted sample values, obtained by error concealment and the original samples values, which would have been obtained by decoding for the missing area if the data corresponding to the missing area had not been lost. The error evaluation map 37, as well as the error evaluation values, are then further used by the merge module 360 and/or by the filtering module 370 to enhance the quality of the reconstructed image..
The corrected auxiliary data 36 and the error evaluation map 37 are used to improve the transform coefficients obtained by error concealment by the merge module 360, as explained in detail with respect to figure 13. The merge module 360 outputs enhanced transform coefficients. Next, an inverse transform module 365 applies an inverse block DCT in this embodiment. An improved reconstructed image 38 is obtained.
Optionally, the improved reconstructed image 38 is further enhanced by the filtering module 370, which also uses the error evaluation map 37 to locally optimize the quality of the reconstructed image.
In an alternative embodiment, the filtering using the error evaluation map may be applied directly on the first reconstructed image obtained by error concealment.
The final resulting corrected image 39 may be output to an application (for example displayed), and also stored to be subsequently used as a reference image by the decoder module 330.
Figure 4 illustrates an embodiment of the auxiliary data extraction and generation applied by modules 320 at the encoder and 345 at the decoder.
The transform coefficients of the blocks of the current image are provided as an input.
At step S410, a subset of transform coefficients are selected for each block.
For example, only a given number of coefficients corresponding to the low frequencies are selected. Classically, the transform coefficients can be ordered as a one-dimensional list of coefficients according to a so-called zig-zag scan, ordering the coefficients in the order of increasing frequencies and associating an index or number to each coefficient. The first N coefficients of the list are selected at step S410. In a preferred embodiment, N=3, so that coefficients of index i {C, 0 «= i < 3} are selected.
It is advantageous to select a given number of coefficients, in particular the first coefficients corresponding to the lower frequencies, since the visual impact of low frequencies is more important than the visual impact of high frequencies. Auxiliary data selected in this way is therefore more efficient for correcting the error concealment result at the decoder. Moreover, error concealment generally better predicts the low frequencies than the high frequencies. Thus the auxiliary data selected in this way is better suited within the Wyner-Ziv scheme. In other words, for the same correction quality, the auxiliary data can be better compressed.
Next, at step S420, the extremum values, i.e. the minimum and maximum values for each selected coefficient, are determined. For each coefficient C of a given index i, 0«=icz3, the minimum (mm1) and maximum (max) values are computed considering all blocks of the image, and then a uniform quantization is applied between those minimum and maximum values at step S430. The number of quantization intervals is determined depending on the number M of bits chosen for representing each coefficient.
In the simplest implementation, an integer value representing an index of the quantization interval, also called quantization number (referral number 520 in figure 5), is directly encoded to represent the quantized value. In this case, in order to obtain a representation on M bits, the range between the minimum and maximum values is divided into 2M quantization intervals.
However, in the preferred embodiment, the quantization number is encoded using co-set representation computed at step 5440, and illustrated in figure 5. In this embodiment, in order to obtain a representation on M bits, 2M+1÷i quantization intervals are created between the minimum value mm1 and the maximum value max for each coefficient. In the example of figure 5, M=2 and there are 9 quantization intervals with 4 different co-set numbers 0 to 3.
Note that at the decoder side, the minimum and maximum values per coefficient are not computed, but retrieved from the additional data 32 received, so as to ensure that the same quantization is applied at the encoder and at the decoder side.
Alternatively, another parameter representative of the quantization interval for each coefficient, such as the quantization step, is transmitted to the client device along with the additional data 32. Similarly, the number of quantization intervals or alternatively the number M of bits per coefficient is known at the decoder side, being either transmitted or pre-determined.
Figure 5 illustrates a typical distribution of the values of a given DOT coefficient 0 in a current image. The horizontal axis 54 represents the values of the given coefficient 0, which in practice vary between mm1 and max. The vertical axis 52 represents the number of coefficients of index i, C, among the blocks of the current image, taking a value given by the horizontal axis. As shown in figure 5, a typical distribution 500 of an AC coefficient (coefficients of index i>0), for example the first frequency coefficient C1, is centered around the value zero, i.e. in most blocks of an image, the value of C1 is close to 0.
An uniform quantization is illustrated in figure 5: the range between minj and max is divided into nine equal ranges or quantization intervals, each of which is attributed a different quantization number 520 from 0 to 8. Each quantization interval is defined by its bounds, for example the lower bound z5 and the upper bound z6 of the quantization interval having quantization number 5 in figure 5.
The quantization intervals can be grouped into co-sets 50, which are attributed a co-set number 530 varying between 0 and 3 in this example. The same co-set number is given to two or more quantization intervals, which advantageously achieves a more compact representation of the additional data derived from the auxiliary data. Therefore, an equal improvement can be obtained at the decoder with more compact additional data.
The number of quantization intervals of each co-set may be different. In the preferred embodiment, as illustrated in figure 5, the same co-set number 0 is associated with the middle quantization interval (containing the coefficient value 0) and to the two extreme quantization intervals. The remaining quantization intervals are grouped by two, so that co-set numbers 1, 2, 3 each have two associated quantization intervals.
As shown in the example, quantization sub-sets I and 5 are assigned the same co-set number 1. This has the advantage of reducing the number of bits for coding a value of a transform coefficient. However, the information cannot be decoded as such, as at the decoder, there would be an ambiguity as on how to decode a received co-set number. Supplementary information is necessary at the decoding.
In the Wyner-Ziv type encoding/decoding scheme represented in figure 3, such supplementary information is provided by the transform coefficients obtained by decoding and error concealment. Therefore, the co-set representation of the quantization intervals is well adapted in a Wyner-Ziv type encoding scheme for the representation of the auxiliary data. Advantageously, co-set representation of the quantization intervals allows compaction of the representation of the auxiliary data.
As further shown in figure 5, the co-set numbers can be simply encoded using the conventional binary representation 540. However, as shown in the figure, two consecutive values of the conventional binary representation may differ by more than one bit. However, when an error concealment is applied, often the predicted coefficient values are close to the original coefficient values, so generally, the quantized values are likely to be also close. Therefore, it is more advantageous to represent the quantized values with a code which has a Hamming distance of 1, meaning that two consecutive values differ only by one bit. Such a binary representation is illustrated in figure 5 by the co-set Gray code 550.
Back to figure 4, the quantized values obtained at step S430 are represented with co-sets, computed as explained above with respect to figure 5. In step S440 a co-set number 530 is assigned to each co-set.. Next, at step S450, a binary encoding is applied on the co-set numbers obtained at step S440. In the preferred embodiment, a Gray code representation 550 is advantageously chosen.
Alternatively, a different type of quantization could be applied at step S430, for example a quantization with a dead zone around zero using 2Mquantization intervals.
Further, in an alternative embodiment, steps S440 and S450 could be replaced by a single step of binary encoding of the unique quantization numbers 520 associated with each quantization interval.
In another embodiment the step 5450 could consist in shuffling the bits of the value obtained for each coefficient. A different shuffle order is applied in each block.
However the order is selected deterministically so the decoder can unshuffle the bits of each coefficient. This embodiment has the advantage to give the same error probability for each bitplane and thus it can give a higher probability of decoding all the bitplanes. A bitplane is composed of all bits of same rank taken from each symbol, therefore each bitplane contains a bit per block.
An example of the error correcting coding module 325 and of the error correcting decoding modules 350 will now be given with respect to figures 6 and 7.
As already explained above, in the preferred embodiment, the LDPC (Low Density Parity Check) code is used by the error correction module 325 at the encoder to output the parity symbols which can serve for the correction of errors in auxiliary data.
Alternatively, other error correction codes such as turbo codes could be used.
An LDPC code is a linear code, defined by a sparse Boolean matrix H, also called the parity matrix.
Figure 6 represents graphically a parity matrix H, also known as LDPC matrix, of a very simple linear code of size (7,4), illustrated for the ease of explanation.
For example the matrix equivalent to figure 6 is H = 1 0 1 1 0 1 0 which is a Boolean matrix.
Given a word c=[c1,c2,c3,c4,c5,c5,c7] composed of 7 symbols, it can be checked that the word is a code word by verifying that: HcT = 0 (Eqi) where cT is transposed from vector c.
It should be noted that the arithmetical operations are carried out as follows.
The addition is the binary XOR operation and the multiplication is the binary AND operation. Note that addition and subtraction are the same operation and the inverse of a symbol is the symbol itself.
The parity matrix H is equivalent to the graph represented figure 6: the symbols of word c 610 are at the left of the graph and the check nodes p 620 which must be null are at the right. The symbols checked by each check node are linked to the check node. The links 630 between the symbols c and the nodes p of figure 6 represent the lines of parity matrix H. For example, when applying the parity check (Eqi) with the first line of matrix H, the following relationship is obtained: p1 = + + c5 0 When H has the form L_RTVflkj, i.e. the last n-k columns of H are equal to the identity matrix of size n-k, nk' it is easy to compute the generating matrix G=LIkR j, where G is composed of the identity matrix of size k, jk' and the transpose of the inverse of the first part of H, denoted R, of k lines and n-k columns. Given an information word ii ={u1,u2,u3,u4], then u* G c.
The code is systematic: the first k values of the code word Q (ci, c2, c3, c4 in this example) are equal to the information word u composed of k information symbols. The last (n-k) values of Q (c5, c6, c7 in this example) are the parity symbols which are transmitted as additional data 32.
In a regular LDPC code, the sparse matrix R is selected with the constraints that on each column the number of bits equal to I is fixed to a small value, for example equal to 3, SO that each symbol is used for 3 checks.
In an alternative embodiment, an irregular LDPC matrix such as a Tornado code can be used, as described in Michael G. Luby, Michael Mitzenmacher, M. Amin Shokrollahi, Daniel A. Spielman, Volker Stemann, "Practical loss-resilient codes", Proceedings of the twenty-ninth annual ACM symposium on Theory of computing, p.150- 159, May 04-06, 1997, El Paso, Texas, United States.
In the current embodiment, the matrix R is created and made available to the modules 325 and 350 before the video transmission. The matrix R depends on the number of blocks per image since in this embodiment, an information symbol is obtained from the quantized coefficients selected for a block.
In an alternative embodiment, it is possible to transmit from the server 101 to the client 102, along with the parity symbols data, either the matrix R or some parameters allowing the error correction decoding module 350 to generate exactly the same matrix R as used by module 325.
The encoding module 325 thus applies a multiplication by R to the binary encoded quantized data provided by module 320. In this operation, since an information symbol is formed by the selected quantized coefficients of a DOT block, all operations (XOR and AND) are thus applied per block and consequently, the encoding is very fast.
Figure 7 is a flowchart of an embodiment of an error correction decoding method implemented by the error correction decoding module 350. All the steps of the algorithm represented in figure 7 can be implemented in software and executed by the central processing unit 201 of the device 102.
The error correction decoding module 350 implements an error correction decoding based on the probabilities of error per bit. In an embodiment, the algorithm applied is the algorithm called belief propagation', described in the article "Good error-correcting codes based on very sparse matrices", by D.J. 0. MacKay, published in IEEE Transactions on In formation Theory 1999, vol. 45, pp. 399-431.
The algorithm of figure 7 receives as an input (step 5700) the concealment auxiliary data, (35 in the example of embodiment of figure 3), as well as the additional data received 32, which contains the parity symbols computed at the error correction encoding 325.
Next, at step S710, a bitplane to be processed is selected. Each symbol is encoded on a given number of bits, each bit having an associated rank from the most significant bit (MSB) to the least significant bit (LBS). For example, a symbol S encoded on 4 bits can be written asS=b323+b222+b12'+b02° =Eb121, b3 being the most significant bit of S and b0 being the least significant bit of S. A bitplane is composed of all bits of same rank taken from each symbol, therefore each bitplane contains a bit per block. In this embodiment, each symbol is composed of the binary representations of several quantized DOT coefficients. For example, if a symbol S encoded on 4 bits comprises two DOT coefficients C and 02, then the first two bits (b3,b2) correspond to C and the next two bits (b1, b0) correspond to 02.
If a Gray code was used as a binary representation of the quantized transform coefficients or if the bits of each coefficient have been shuffled, then all the bitplanes are independent. In that case, the bitplanes can be decoded in any order, so any bitplane that has not yet been decoded can be selected at step S710. If a conventional binary encoding was used for representing the quantization numbers, then the order of the bitplanes has an importance, and if the most significant bitplanes contain errors, the corresponding quantization number cannot be decoded. Therefore, in case a conventional binary encoding is used in step S450, the bitplanes must be selected at step S710 starting from the most significant bitplane to the least significant bitplane.
After the selection of a bitplane to be processed has been achieved at step S710, then at step S720 the probabilities per bit are computed for each bit of the selected bitplane, expressed as a probability ratio which is called the likelihood: for each bit, the probability of its value being I divided by the probability of its value being 0. As the algorithm progresses, these probability ratios will be modified taking into account information obtained from other bits, in conjunction with the requirement that the parity checks be satisfied.
Let p be the probability of a bit b to be incorrect.
The likelihood (or probability ratio) associated with the bit b is the following: if b=1, likelihood(b) = (1- (Eq2) p if b0, Iikelihood(b) = (Eq3) (1-p) In the corrupted area of the signal, typically a slice that cannot be decoded and on which an error concealment is applied, the value of p is set to a predetermined value, for example p=0.l. Typically, in a simple embodiment, the same value of p is associated with every bit of every symbol representing a transform coefficient of a block belonging to the corrupted image area.
In the image areas where no transmission error occurred and thus the decoded value is correct, p is set to a value very close to 0, typically p= 0.0001, such that the likelihood for a value b=0 is 0 and the likelihood for a value b=1 is the maximum acceptable integer value for the CPU 201.
With a conventional binary encoding there is a dependency between the bit correctness from different bitplanes. Thus if the decoding of a bit is incorrect in step S760, the lower order bits of the same DOT coefficient will be set to a likelihood of 1 to indicate a high probability of error (typically, p=0.5).
In an alternative embodiment, the likelihood calculation is improved using information on the predicted error values. This alternative embodiment takes into account the predictive encoding of images and the error propagation according to the motion.
Therefore, it is desirable to take into account the estimated errors in the previous reference image to compute the initial likelihood of the bits of a current imageS This alternative embodiment advantageously uses the error evaluation map 37 computed for the previously processed reference image in order to improve the error correction decoding. In this alternative embodiment: -if a block is lost, a constant bit error probability, p0.1 is used; -if a block is encoded in Intra mode, p is set to a value close to 0, for example p=o.000I; -if a block is predicted via a motion vector from a predictor bock of a reference image, a value representative of the estimated error (GB) on the predictor block is used.
The computation of such a value GB will be described in more detail with respect to figures 9toll.
In case the predictor block is not aligned with the grid of blocks, the value a representative of the estimated error of the block B of the grid of blocks which shares the largest surface with the predictor block is selected.
Then, for each DOT coefficient, the bit error probability may be computed according to the formula: 1 ___ (Eq4) aB (1-e'6) with aB =I and g =0.1, and exdesignates the exponential y 2aB mathematical function.
The belief propagation' algorithm is applied at step S730.
For each parity check, the algorithm computes a probability ratio for every bit that participates in that parity check. These ratios give the probability of that parity check being satisfied if the bit in question is 1 divided by the probability of the check being satisfied if the bit is 0, taking account of the probabilities of each of the other bits participating in this check, as derived from the probability ratios for these bits with respect to this check.
For every bit symbol, the algorithm computes a probability ratio for each parity check in which said bit symbol is involved, giving the probability for that bit to be I versus 0 based only on information derived from other parity checks, along with the data received for the bit.
The algorithm alternates between recalculating the probability ratios for each check and recalculating the probability ratios for each bit.
At the end of each iteration, an evaluated value for each bit is computed from the likelihood.
A first simple embodiment consists to compare the likelihood to 1. If the likelihood is lower than I the bit value is estimated to be 0 otherwise, the bit value is estimated to be 1.
A better solution will be to test if the likelihood is greater than a threshold, for example TI =1.5, then the evaluated bit value is 1; otherwise, if the likelihood is lower than a second threshold (for example T2=0.66), then the evaluated bit value is 0. If the likelihood is between Ti and T2, then the bit value cannot be evaluated.
If the value of each bit of the selected bitplane can be evaluated, then the check values 620 are computed from the evaluated bit values using the classical LDPC decoding.
The algorithm stops (test S740) if all checks are verified (Hct=0), in which case the correction has succeeded and the bitplane is fully corrected, or after a predefined number of iterations, in which case the correction has failed. Typically, if the number of bit errors in the selected bitplane is sufficiently low, the belief propagation' decoding algorithm corrects all bits and stops rapidly. On the contrary, if the number of bit errors is too high, the belief propagation' decoding algorithm does not converge. The experiments have shown that for example a number of 20 iterations is sufficient.
If the algorithm has failed (answer no' to test S740), the selected bitplane is marked as undecodable at step S750. This means that the number of bit errors in the bitplane is high, and that the selected bitplane of the auxiliary data has not been completely corrected.
If the algorithm has succeeded (answer yes' to test S740), then the selected bitplane is completely corrected (S760). It is thus possible to find out exactly which bits of the selected bitplane of the auxiliary data provided as an input are incorrect. The incorrect bits are corrected by inversing their value in the auxiliary data (i.e. replacing value 0 by value I and vice versa).
In the case where the conventional binary representation is used, if a bit belonging to a given DOT coefficient is incorrect and has been corrected, the lower order bits of the same DOT coefficient have a high probability to be also incorrect. This information can be used to compute the likelihood for the corresponding bits at step S720.
Both steps S750 and S760 are followed by step S770, at which it is checked whether the currently selected bitplane is the last bitplane of the symbols of the auxiliary data.
If it is not the last bitplane, then the next remaining bitplane is selected and the steps S720 to S770 are repeated.
We note that if the conventional binary representation is used, and the processed bitplane has been marked as undecodable, then it is useless to process the following bitplanes of the DOT coefficient containing this selected bitplane. In this case and if this was the last DOT coefficient or if the currently selected bitplane is indeed the last bitplane of the auxiliary data, the error correction decoding stops (step S780).
Figure 8 is a flowchart of an embodiment of the computation of an estimation of error evaluation by module 355 of figure 3. All the steps of the algorithm represented in figure 8 can be implemented in software and executed by the central processing unit 201 of the device 102.
The algorithm of figure 8 receives as an input (step S800) the DOT coefficients computed from the reconstructed image by concealment (34 in the example of embodiment of figure 3), as well as the corrected auxiliary data 36 output by module 350.
The DOT coefficients are predicted values for the original DOT values which are not known at the decoder. The predicted DOT value may be equal to the original DOT value, in particular in the case where they belong to block correctly received at the decoder, and which are predicted from reference blocks which are also correct. Otherwise, the predicted values are likely to be different from the original unknown DOT value.
In step S810, a following DOT coefficient among the set of DOT coefficients {Oo, O1,..,ON-1} that form the auxiliary data is considered, starting with the first DOT coefficient 00, which is typically the DO coefficient.
In the preferred embodiment, an error evaluation map is computed per transform coefficient. This is advantageous since the transform coefficients may have different statistical distributions. In particular the DO coefficient has generally a different statistical distribution from AO coefficients.
However, alternatively, in a simpler embodiment, all selected coefficients may be processed together.
Next, it is checked at step S820 whether all the bitplanes, of the selected DOT coefficient for all the blocks of the current image have been corrected by the error correction decoding 350. In case of negative answer, the following DOT coefficient is selected (S810).
In case of positive answer, at step S830 an estimated error value or error distance is computed for each block of the current image, and stored in a so-called error distance map. In this embodiment, the corrected auxiliary data is composed of corrected symbols that represent quantized values of DOT coefficients, so that an estimated error is computed for each block between the DOT coefficient obtained by error concealment, i.e. the predicted sample value, and the interval obtained from the corrected auxiliary data, as explained hereafter with respect to figure 10. Only an estimated error can be computed, since the information available is that the original DOT coefficient value, which is unknown, belongs to the interval obtained from the corrected auxiliary data.
Then, based on the estimated errors for a plurality of blocks, step 5840 computes an error evaluation value associated with each block, finally obtaining an error evaluation map for the missing area. This computation is further detailed in figures 11 and 12. The error evaluation value associated with each block is a value representative of a confidence evaluation of the estimated error with respect to a plurality of blocks selected according to a predetermined criterion as explained further; Step S840 is then followed by step 5810 already described. If all the coefficients of the subset of coefficients selected to form the auxiliary data have been processed (answer no' to test S810), the algorithm stops (S850).
Figure 9 details an embodiment of the computation of an error distance map of step S830 of figure 8. All the steps of the algorithm represented in figure 9 can be implemented in software and executed by the central processing unit 201 of the device 102.
For each block of the current image (S900), the corrected auxiliary data value representative of a co-set number is decoded in step S910 to obtain a corresponding quantization interval, defined by its lower and upper bounds.
Oonsidering the embodiment of figure 4, if a Gray code was used in step S450, firstly the Gray code value is decoded to obtain a co-set number 530. This step can be efficiently implemented by using a table, stored in memory, containing the correspondences between Gray code values and corresponding co-set number values.
A co-set number corresponds to several quantization intervals. For example, referring to figure 5, co-set number I corresponds to intervals I and 5. The bounds of the corresponding quantization intervals are considered, in this example [z1,z2] and [z5,z6].
For each such possible quantization interval, an estimated error distance is computed in step S920.
A quantization interval i, [z1,z÷1] is represented in figure 10. The original DOT value, which is unknown at the decoder, is inside the quantization interval i. The DOT value obtained by error concealment, or predicted sample value, is value y. In this embodiment, the estimated error distance y is computed as follows: If y is inside the interval, z1 «= y < z1, the estimated error distance is null y1 =0; Otherwise, the estimated error distance is the difference between value y and the closest interval bound: if y < z1, y = z -y; if y > z1, 1 = y-z.
If several possible quantization intervals are given by the co-set number, the minimum estimated error distance is selected (5930). This value of estimated error distance is then stored for the block (S940), along with an indication of the corresponding quantization interval.
When all the blocks have been processed, the algorithm stops. An estimated error distance map, comprising an estimated error distance for each block of the image for the given coefficient C1 is then available. It is worth pointing out here that due to the use motion compensation in the encoder, some reconstruction errors may be propagated from a previous reference frame, and therefore other image blocks than the blocks belonging to a missing area are likely to have associated error distances which are non null.
Figure 11 details an embodiment of the computation of an estimation of error confidence evaluation, or error evaluation map, of step S840 of figure 8. All the steps of the algorithm represented in figure II can be implemented in software and executed by the central processing unit 201 of the device 102.
For each block B of the current image (SI 100) a value representative of a statistic of a plurality of estimated error values is computed in step SII2O. Such a value is representative of an evaluation of the confidence on the estimated error for block B and for the given coefficient C. In the preferred embodiment, the value representative of a statistic is the variance of the estimated error distance values obtained for the block B and for a plurality of other blocks of the current image, chosen according to a criterion. However, other statistical measures, such as higher order moments, may be alternatively used.
In an embodiment, the plurality of other blocks is chosen according to spatial proximity, so the blocks are selected in a spatial neigbourhood Ns of the block B. The error variance UB of block B is computed as follows: = jx EYBJ (Eq5) v K liENs BEN where K is the number of blocks in the neighbourhood Ns including B, YB is the estimated error distance for block B and for the given coefficient, and the summations are carried out for all blocks B of the selected neighbourhood Ns, including the block B. We note that in an alternative, all coefficients are processed together, so in the formula of (Eq5), the estimated error distances for all coefficients used in the auxiliary data are taken into account.
Figure 12a shows a very simple example of spatial neighbourhood NO, comprising the 8 surrounding blocks of block Bn.
In an alternative embodiment, the algorithm tests a number of possible spatial neighbourhoods (four in the example illustrated in figure 12b, NI to N4) by computing the corresponding variances and finally selects the neighbourhood the variance of which is the largest, N4 in figure 12b. This alternative embodiment advantageously takes into account the local structure of the image signal, such as edges.
In yet another alternative, not illustrated in figure 12, a different criterion, for example signal similarity, may be applied to select the blocks to be used in the calculation of the variance. Such similar blocks may be situated at any position in the image, and are characterized by a similarity distance, such as for example a SAD (sum of absolute differences) computed pixel to pixel or coefficient to coefficient, is lower than a predetermined threshold.
More generally, the selection of blocks to be used in the calculation of the variance is applied according to a predetermined criterion, such as a proximity criterion, which is either spatial proximity or signal proximity or similarity.
Back to figure 11, once the value representative of a statistic of the error estimation for block B is computed, this value is stored in memory.
Once all the blocks have been processed, the error evaluation map 37 is completed.
Figure 13 is a flowchart of a first embodiment of an improvement of the reconstruction quality of the decoded image, as implemented by the merge module 360 of figure 3. All the steps of the algorithm represented in figure 13 can be implemented in software and executed by the central processing unit 201 of the device 102.
The algorithm uses as input the DCT values 34 computed for the reconstructed error concealed image 33, the error distance estimation and the error evaluation values, as well as the quantization intervals obtained from the corrected auxiliary data 36, obtained in step S910.
For each block B (SI 310), and for each DOT coefficient O of such a block B (S 1315), first test Si 320 checks whether 01 is one of the coefficients selected to form the auxiliary data.
In case of negative answer, the DOT value obtained by error concealment is kept as final coefficient value in step S 1340.
In case of positive answer, test S1325 verifies whether all the bitplanes have been corrected for 0. In case of negative answer, step S1340 is applied, so the DOT value obtained by error concealment is kept as final coefficient value. In case of positive answer, in step S 1330 it is checked whether the estimated error distance corresponding to coefficient 01 of block B, computed in step S930, is null.
If the estimated error distance is null, then the value obtained by error concealment is compatible with the error correction, and is therefore kept (Si 340).
If the estimated error distance is non null, an improved value for the coefficient 0 of block B is computed in step S1335.
is In the preferred embodiment, the improved value 1B is calculated from the value which is the predicted DOT value obtained by error concealment for coefficient C of block B, using the bounds of the quantization interval [z, z1+i} obtained from the corrected auxiliary data, as proposed in the article Optimal reconstruction in Wyner-Ziv Video Ooding with Multiple Side Information', by Kubasov et al., published in the gth International Workshop for Multimedia Signal Processing 2007.
Let A be the quantization interval size: A -z1. The improved value is computed as follows: 1 A -ify1<z1, XfiZ+-+ a A (Eq6) a3 1-e 1 A -if Y111 >z1 XBf Z11 a A (Eq7) GtBfl 1-e The notation et designates the exponential mathematical function.
The value cxB is related to the value Cfi representative of an evaluation of the error confidence on a subset of estimated error distances computed for block Bn.
In the preferred embodiment, ali -/ . Alternatively, ali = can cr also be used.
The efficiency of using the formulas of (Eq6) and (Eq7) has been proven by experimental tests. If the variance c11 of the estimated error distance values is high, the corresponding value of ctB is low, and the improved reconstruction value is close to the middle of the quantization interval. When the variance a11 of the estimated error distance values is low, the corresponding value of ccB is high, and the reconstruction value is close to a bound of the quantization interval.
The value of a11 for the block B, for a given coefficient C, is calculated for block B, taking into account a plurality of the values of estimated error predictions for blocks in a given neighbourhood, in particular a given spatial neigbourhood. For this reason, the coefficient value improvement is advantageously adapted to the characteristics of the image signal, and the level of reconstruction improvement is higher than provided in the cited prior art article Optimal reconstruction in Wyner-Ziv Video Coding with Multiple Side Information', by Kubasov et al., published in the 9th International Workshop for Multimedia Signal Processing 2007, where a same value a is used for all blocks in an image.
Once the improved reconstruction value for the current DCT coefficient C of the current block has been computed, it is stored as new DCT value in block B (S 1335).
Then the next coefficient of the block is processed, and the process is repeated until all coefficients and all blocks have been processed.
The result is a set of improved DCT coefficient values, which are then provided to the module 365 of figure 3 to compute an inverse block DCT.
Figure 14 is a flowchart of a second embodiment of an improvement of the reconstruction quality of the decoded image, as implemented by the filtering module 370 of figure 3. All the steps of the algorithm represented in figure 14 can be implemented in software and executed by the central processing unit 201 of the device 102.
This filtering is applied in the pixel domain, after the inverse block DCT has been performed, and it aims to further enhance the image quality since some blocking effects might appear due to the Wyner-Ziv correction scheme. Alternatively, such a filtering could be used directly after the error concealment.
The filtering strength is set, in this embodiment, based upon the estimated errors and their confidence evaluation based on a plurality of blocks.
For each block of current image (S1410), it is first checked in step S1420 whether the variance of estimated error distances for the coefficient DC (C0) is null. In this embodiment, only the DC coefficient is considered since the DC coefficient has a major visual impact. If the error variance is null, no filtering is applied (step 51425), and the algorithm returns to Si 410 to process the following block if any.
If the error variance is not null, step S1420 is followed by step S1430 which applies the comparison of a quantity based upon the estimated error variance to a threshold, so as to decide between two types of filtering detailed hereafter, called soft filtering and strong filtering.
Let crB be the error variance for the DC coefficient of current block, and, as previously: aB = I and A=z11 -z.
If ---+ A__A then step S 1430 is followed by the soft filtering applied 1_e"8 4 instepSi44O.
Otherwise, step S1430 is followed by the strong filtering applied in step S1445.
If the blocks of an image are processed in the lexicographical order, from the top left corner to the bottom right corner, then for each current block, a vertical border and the horizontal border remain to be processed. For example, for block 1510 schematically represented on figure 15, borders 1530 and 1540 are processed.
In step S1440, the soft filtering is for example first applied on the vertical border and then on the horizontal border.
By reference to figure 15, the vertical border 1530 between blocks 1510 and 1520 is processed as follows. Each group of 4 pixels p0, pi, p2, p3 is filtered using the values qO, qi, q2, q3 from the adjoining block 1520. The new values pUs, p15, p2s, p3s are computed for example as follows: pOs = (p2 + 2*pi + 2pO + 2*qO + qi + 4) >> 3; pis = (p2 + p1 + p0 + qO + 2) >> 2; p2s = (2*p3 + 3*p2 + p1 + p0 + qO + 4) >> 3; The symbol >> is used for the bit shifting, in other words V>>a is equivalent to a division of V by 2a.
The same filtering is applied analoguously on the horizontal border.
Next, the following block, if any, is handled (step S1410).
The strong filtering of step S1445 is applied similarly to the soft filtering, respectively on the vertical and the horizontal border.
For example, the following formulas may be applied: pOs = (p2 + 2*pl + 2*pO + 2*qO + 2*q1 + q2 + 5) /10; pls = (p2 + p1 + p0 + qO -I-q1 + 3) / 5; p2s = (2*p3 + 3*p2 + p1 + p0 + qO + ql + 5) / 9; Step S1445 is also followed by step S1410 for the processing of the following block.
Other filtering coefficients may be selected and applied, the main characteristic being that the soft filtering modifies less the pixel values than the strong filtering and takes into account less pixels of the adjoining block.
Figure 16 describes a distributed video coding system to which a second embodiment of the invention can be applied. The objective of a distributed video coding method is to simplify the encoding by removing a part of the motion estimation. Indeed the motion estimation in the encoder uses a lot of computation.
In a standard video coding system, frames are encoded in Intra mode (I frames), or in predicted mode based on one reference frame (P frame) or two reference frames (B frame). In a distributed video encoder some frames may be encoded classically (frames 1601, 1602 and 1603) and some frames may be encoded in Wyner-Ziv mode (frame 1604).
In this example, frame 1604 is encoded as Wyner-Ziv additional data, which can be generated using a method similar to the method used in figure 3. The original image is transformed with a DOT transformation, then the auxiliary data is extracted and finally an LDPO is applied on all the quantized blocks. Only the Wyner-Ziv additional data is transmitted to the decoder.
With this method the encoder is simplified compared to a classical encoder because no motion vector needs to be computed for a Wyner-Ziv encoded frame (noted WZ frame on the figure). However some motion vectors are computed for the classical P frames (1602 and 1603).
At the decoder side, the classical 1 and P frames (1601, 1602, 1603) are decoded by a classical decoder.
The frames encoded with Wyner-Ziv or WZ frames can be considered as missing areas of the sequence of images. In this embodiment, the missing area is an entire frame.
For the WZ frame (1604) the decoder creates a first version of the frame using the received I and P frames, by applying a reconstruction method using available data, similarly to error concealment methods.
Many methods exist but we will describe two efficient methods based on predicted motion estimation.
The motion extrapolation (ME) method consists in using the motion vectors of the previous P image (1610) and then changing their orientation and scaling them to the distance between the WZ frame and the P frame. This new motion field (1630) is then used to predict the pixel values of the WZ image.
The motion interpolation (Ml) method consists in using the previous and following frames (1602 and 1603). A motion field is computed between the two frames or the motion field 1611 determined by the encoder can also be used. Then the values of the pixels of the image 1604 are predicted by interpolation of the pixels of the two images using the computed motion.
These two reconstruction methods are similar the to error concealment method applied on missing areas. It is still desirable in this embodiment to improve the reconstruction of the missing part, here the WZ frame, and the algorithms described above apply similarly in this case.
An estimated error distance per block and per coefficient, and an associated error confidence evaluation can be computed and use to improve the reconstructed coefficient values, and further to improve the decoded image quality in the pixel domain by filtering, and also to propagate the error confidence evaluation in the error correcting decoding algorithm, using the methods explained above.
In the embodiments described above, a block DOT transform was used to obtain transform coefficients. However, any alternative block or subband transform, such as a wavelet transform, may be equally applied.
In the embodiments described above, the partially corrected data obtained is representative of quantization intervals corresponding to transform coefficients, such as DOT coefficients computed for a block of samples.
Other alternative embodiments may be envisaged, such as for example auxiliary data representative of pixels values instead of coefficient values.
More generally, any modification or improvement of the above-described embodiments, that a person skilled in the art may easily conceive should be considered as falling within the scope of the invention.

Claims (20)

  1. CLAIMS1. Method for decoding a digital signal comprising at least one digital image encoded by an encoder, said digital image being represented by a plurality of samples, the method comprising, when a part of one said encoded digital image to decode is missing: -applying a reconstruction to said encoded digital image having the missing part to form a reconstructed image, said reconstruction comprising setting a missing sample, being one of said samples in said missing part, to a predicted sample value, -obtaining, for each of a plurality of predicted sample values, a value of an estimated error between said predicted sample value and the corresponding missing value, using additional data, derived by the encoder from at least part of the encoded digital image and usable during decoding to correct the encoded digital image, -for at least one missing sample having a predicted sample value, computing a value representative of a confidence evaluation of the estimated error using said estimated error values obtained for a selected plurality of predicted sample values, and -using said value representative of a confidence evaluation the estimated error obtained to improve the reconstruction of said encoded digital image to decode.
  2. 2. A method according to claim 1, wherein the step of computing a value representative of a confidence evaluation of the estimated error comprises, for each said missing sample: -selecting a plurality of samples according to a predetermined criterion, said plurality including said missing sample, -computing, as value representative of a confidence evaluation of the estimated error, a statistic of the estimated error values obtained for the predicted sample values of said selected samples.
  3. 3. A method according to claim 2, wherein said predetermined criterion is spatial proximity.
  4. 4. A method according to claim 2, wherein said missing sample belongs to a block of samples and wherein said predetermined criterion is a similarity between blocks of samples.
  5. 5. A method according to any of claims 2 to 4, wherein said statistic is a variance of the estimated error values obtained for the predicted sample values of said selected samples.
  6. 6. A method according to any of claims I to 5, wherein said step of obtaining a value of an estimated error comprises: -applying an error correction decoding to said reconstructed image using said additional data to obtain corrected symbols, -obtaining, for at least one missing sample, a corrected symbol representative of the missing sample value of said missing sample, and -computing an error distance between said corrected symbol and the predicted sample value of said missing sample.
  7. 7. A method according to claim 6, wherein said corrected symbol is representative of one or several intervals to which said missing value may belong, an interval being defined by its lower and upper bound, and wherein the step of computing an error distance comprises: -for each said interval, computing the difference between said predicted sample value and the interval bound closest in value to said predicted sample value, the error distance being set as the minimum of said computed differences.
  8. 8. A method according to any of claims 6 or 7, wherein said additional data comprises parity symbols obtained at the encoder by applying an error correction encoding to auxiliary data extracted from a digital image corresponding to said encoded digital image to decode.
  9. 9. A method according to claim 8, wherein said step of applying an error correction decoding comprises a step of extracting auxiliary data from said reconstructed image comprising the sub-steps of: -dividing said reconstructed image into blocks of decoded pixels, and -applying a transform on each block of decoded pixels to obtain a block of transform coefficients.
  10. 10. A method according to claim 9, wherein extracting of the auxiliary data from the reconstructed image further comprises: -selecting a subset of transform coefficients of each block of transform coefficients, and -quantizing each selected transform coefficient using a predetermined number of bits, a quantization number being associated with each quantized coefficient.
  11. 11. A method according to claim 10, further comprising associating a co-set number to a plurality of said quantization numbers.
  12. 12. A method according to claim 10 or 11, wherein said extracted auxiliary data is binary encoded using a Gray code representation.
  13. 13. A method according to any of claims 6 to 12, wherein said error correction decoding is applied by bitplanes and comprises an iterative decoding process which uses, for each bit of a bitplane, a probability of said bit to be incorrect, the method further comprising, to improve the reconstruction of said encoded digital image to decode, setting a probability of a bit to be incorrect based upon a value representative of a confidence evaluation of the estimated error computed for a missing sample belonging to a previous image used as a prediction reference for said encoded digital image to decode.
  14. 14. A method according to any of claims Ito 13, wherein the step of using said value representative of a confidence evaluation obtained to improve the reconstruction of said encoded digital image to decode comprises, for at least one missing sample, a step of modifying the predicted sample value of said missing sample based upon the estimated error computed for said predicted sample value and upon the corresponding value representative of a confidence evaluation of the estimated error.
  15. 15. A method according to any of claims Ito 14, wherein an encoded image is divided into blocks, the step of using said value representative of a confidence evaluation of the estimated error obtained to improve the reconstruction of said encoded image to decode comprises: -applying a spatial filtering on samples situated at the borders of said blocks of said reconstructed image, said filtering consisting of applying one filter of a set of predetermined filters, comprising a step of selecting, among the set of predetermined filters, a filter to apply on a given border between a first and a second block, based upon said value representative of a confidence evaluation of the estimated error computed for a sample of said first block.
  16. 16. A method according to claim 15, wherein the set of predetermined filters comprises a soft filter and a strong filter.
  17. 17. Device for decoding a digital signal comprising at least one digital image encoded by an encoder, said digital image being represented by a plurality of samples, the device comprising, when a part of one said encoded digital image to decode is missing: -means for applying a reconstruction to said encoded digital image having the missing part to form a reconstructed image, said reconstruction comprising setting a missing sample, being one of said samples in said missing part, to a predicted sample value, -means for obtaining, for each of a plurality of predicted sample values, a value of an estimated error between said predicted sample value and the corresponding missing value, using additional data, derived by the encoder from at least part of the encoded digital image and usable during decoding to correct the encoded digital image, -means for computing, for at least one missing sample having a predicted sample value, a value representative of a confidence evaluation of the estimated error using said estimated error values obtained for a selected plurality of predicted sample values, and -means for using said value representative of a confidence evaluation the estimated error obtained to improve the reconstruction of said encoded digital image to decode.
  18. 18. A non-transitory computer program which, when executed by a computer or a processor in a device for decoding a digital signal, causes the device to carry out a method for decoding a digital signal as claimed in claims I to 16.
  19. 19. A computer-readable storage medium storing a program according to claim 18.
  20. 20. A method, device or computer program for decoding a digital signal as hereinbefore described with reference to the accompanying drawings.
GB1100224.3A 2011-01-07 2011-01-07 Improved reconstruction of at least one missing area of a sequence of digital images Expired - Fee Related GB2487078B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
GB1100224.3A GB2487078B (en) 2011-01-07 2011-01-07 Improved reconstruction of at least one missing area of a sequence of digital images

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
GB1100224.3A GB2487078B (en) 2011-01-07 2011-01-07 Improved reconstruction of at least one missing area of a sequence of digital images

Publications (3)

Publication Number Publication Date
GB201100224D0 GB201100224D0 (en) 2011-02-23
GB2487078A true GB2487078A (en) 2012-07-11
GB2487078B GB2487078B (en) 2014-08-06

Family

ID=43663909

Family Applications (1)

Application Number Title Priority Date Filing Date
GB1100224.3A Expired - Fee Related GB2487078B (en) 2011-01-07 2011-01-07 Improved reconstruction of at least one missing area of a sequence of digital images

Country Status (1)

Country Link
GB (1) GB2487078B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110264248A (en) * 2019-05-30 2019-09-20 阿里巴巴集团控股有限公司 The prediction technique and device of user experience information

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2009057956A1 (en) * 2007-11-02 2009-05-07 Sungkyunkwan University Foundation For Corporate Collaboration Apparatus and method of decompressing distributed video coded video using error correction

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2480224B (en) * 2010-01-29 2014-08-20 Canon Kk Decoding a sequence of digital images with error concealment

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2009057956A1 (en) * 2007-11-02 2009-05-07 Sungkyunkwan University Foundation For Corporate Collaboration Apparatus and method of decompressing distributed video coded video using error correction

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
Rane, S. et al., "Systematic Lossy Forward Error Protection for Error-Resilient Digital Video Broadcasting - a Wyner-Ziv Coding Approach", 2004, International Conference on Image Processing (IEEE ), vol. 5, pages 3101-3104 *

Also Published As

Publication number Publication date
GB2487078B (en) 2014-08-06
GB201100224D0 (en) 2011-02-23

Similar Documents

Publication Publication Date Title
US8774538B2 (en) Decoding a sequence of digital images with error concealment
US8340192B2 (en) Wyner-Ziv coding with multiple side information
Guillemot et al. Distributed monoview and multiview video coding
AU2003268459A1 (en) Encoding and decoding of digital data using cues derivable at a decoder
Brites et al. An efficient encoder rate control solution for transform domain Wyner–Ziv video coding
US20120033741A1 (en) Decoding of a digital signal comprising at least one sample
Brites et al. Side information creation for efficient Wyner–Ziv video coding: classifying and reviewing
US7894550B2 (en) Method, apparatus, and system for source coding with iterative side information generation and decoding process
Zhang et al. Fast transmission distortion estimation and adaptive error protection for H. 264/AVC-based embedded video conferencing systems
Milani et al. Distributed video coding based on lossy syndromes generated in hybrid pixel/transform domain
GB2487078A (en) Improved reconstruction of at least one missing area of a sequence of digital images
Huang et al. Distributed video coding with multiple side information
Belyaev Compressive sensed video coding having Jpeg compatibility
KR101035746B1 (en) Method of distributed motion estimation for video encoder and video decoder
Hanca et al. Lightweight real-time error-resilient encoding of visual sensor data
KR101152482B1 (en) Methods of encoding and decoding using fast ldpca code and apparatuses using the same
GB2478996A (en) Reconstruction of at least one missing area of a sequence of digital images
KR20100082700A (en) Wyner-ziv coding and decoding system and method
Qing et al. Practical distributed video coding in packet lossy channels
Li On the importance of source classification in Wyner-Ziv video coding
Thao et al. Side information creation using adaptive block size for distributed video coding
Kodavalla et al. Chroma components coding in feedback-free distributed video coding
CN110612725B (en) Processing apparatus and control method thereof
Haqqani et al. Encoder rate control for transform domain Wyner-Ziv Video Coding
Lei et al. Study for distributed video coding architectures

Legal Events

Date Code Title Description
PCNP Patent ceased through non-payment of renewal fee

Effective date: 20190107