WO2011128268A1 - Probability interval partioning encoder and decoder - Google Patents

Probability interval partioning encoder and decoder Download PDF

Info

Publication number
WO2011128268A1
WO2011128268A1 PCT/EP2011/055528 EP2011055528W WO2011128268A1 WO 2011128268 A1 WO2011128268 A1 WO 2011128268A1 EP 2011055528 W EP2011055528 W EP 2011055528W WO 2011128268 A1 WO2011128268 A1 WO 2011128268A1
Authority
WO
WIPO (PCT)
Prior art keywords
macroblock
symbol
predetermined
bin
symbols
Prior art date
Application number
PCT/EP2011/055528
Other languages
French (fr)
Inventor
Heiko Schwarz
Detlev Marpe
Thomas Wiegand
Heiner Kirchhoffer
Anastasia Henkel
Christian Bartnik
Mischa Siekmann
Jan Stegemann
Original Assignee
Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. filed Critical Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V.
Priority to JP2013504210A priority Critical patent/JP5676744B2/en
Priority to EP11714643.1A priority patent/EP2559166B1/en
Priority to CN201180029224.XA priority patent/CN103119849B/en
Priority to PL11714643T priority patent/PL2559166T3/en
Publication of WO2011128268A1 publication Critical patent/WO2011128268A1/en
Priority to US13/647,956 priority patent/US8907823B2/en

Links

Classifications

    • HELECTRICITY
    • H03ELECTRONIC CIRCUITRY
    • H03MCODING; DECODING; CODE CONVERSION IN GENERAL
    • H03M7/00Conversion of a code where information is represented by a given sequence or number of digits to a code where the same, similar or subset of information is represented by a different sequence or number of digits
    • H03M7/30Compression; Expansion; Suppression of unnecessary data, e.g. redundancy reduction
    • H03M7/40Conversion to or from variable length codes, e.g. Shannon-Fano code, Huffman code, Morse code
    • H03M7/4006Conversion to or from arithmetic code
    • HELECTRICITY
    • H03ELECTRONIC CIRCUITRY
    • H03MCODING; DECODING; CODE CONVERSION IN GENERAL
    • H03M7/00Conversion of a code where information is represented by a given sequence or number of digits to a code where the same, similar or subset of information is represented by a different sequence or number of digits
    • H03M7/30Compression; Expansion; Suppression of unnecessary data, e.g. redundancy reduction
    • H03M7/40Conversion to or from variable length codes, e.g. Shannon-Fano code, Huffman code, Morse code
    • H03M7/42Conversion to or from variable length codes, e.g. Shannon-Fano code, Huffman code, Morse code using table look-up for the coding or decoding process, e.g. using read-only memory
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/129Scanning of coding units, e.g. zig-zag scan of transform coefficients or flexible macroblock ordering [FMO]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/162User input
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/42Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by implementation details or hardware specially adapted for video compression or decompression, e.g. dedicated software implementation
    • H04N19/436Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by implementation details or hardware specially adapted for video compression or decompression, e.g. dedicated software implementation using parallelised computational arrangements
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/90Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using coding techniques not provided for in groups H04N19/10-H04N19/85, e.g. fractals
    • H04N19/91Entropy coding, e.g. variable length coding [VLC] or arithmetic coding

Definitions

  • the present invention relates to entropy coding and may be used in applications such as, for example, video and audio compression.
  • the present invention describes a new method and apparatus for entropy encoding and decoding of discrete data.
  • Entropy coding in general, can be considered as the most generic form of lossless data compression.
  • Lossless compression aims to represent discrete data with fewer bits than needed for the original data representation but without any loss of information.
  • Discrete data can be given in the form of text, graphics, images, video, audio, speech, facsimile, medical data, meteorological data, financial data, or any other form of digital data.
  • Huffman codes and arithmetic codes are well-known examples of practical codes capable of approximating the entropy limit (in a certain sense).
  • Huffman codes are relatively easy to construct.
  • the most attractive property of Huffman codes is that its implementation can be efficiently realized by the use of variable-length code (VLC) tables.
  • VLC variable-length code
  • Huffman codes in the case of having a dominant alphabet value with j3 ⁇ 4 > 0.5, the redundancy of the corresponding Huffman code (without using any alphabet extension such as run length coding) may be quite substantial.
  • Another shortcoming of Huffman codes is given by the fact that in case of dealing with higher-order probability modeling, multiple sets of VLC tables may be required.
  • Arithmetic coding while being substantially more complex than VLC, offers the advantage of a more consistent and adequate handling when coping with adaptive and higher-order probability modeling as well as with the case of highly skewed probability distributions.
  • this characteristic basically results from the fact that arithmetic coding provides a mechanism, at least conceptually, to map any given value of probability estimate in a more or less direct way to a portion of the resulting codeword. Being provided with such an interface, arithmetic coding allows for a clean separation between the tasks of probability modeling and probability estimation, on the one hand, and the actual entropy coding, i.e., mapping of a symbols to codewords, on the other hand.
  • An encoder for encoding a sequence of symbols comprises an assigner configured to assign a number of parameters to each symbol of the sequence of symbols based on information contained within previous symbols of the sequence of symbols; a plurality of entropy encoders each of which is configured to convert the symbols forwarded to the respective entropy encoder into a respective bitstream; and a selector configured to forward each symbol to a selected one of the plurality of entropy encoders, the selection depending on the number of parameters assigned to the respective symbol.
  • a decoder for reconstructing a sequence of symbols comprises a plurality of entropy decoders, each of which is configured to convert a respective bitstream into symbols; an assigner configured to assign a number of parameters to each symbol of a sequence of symbols to be reconstructed based on information contained within previously reconstructed symbols of the sequence of symbols; and a selector configured to retrieve each symbol of the sequence of symbols to be reconstructed from a selected one of the plurality of entropy decoders, the selection depending on the number of parameters defined to the respective symbol.
  • Embodiments of an entropy encoding algorithm are described in which a sequence of arbitrary source symbols is mapped onto two or more partial bitstreams, and a decoding algorithm, in which the original source symbols are decoded from the two or more partial bitstreams.
  • the source symbols are first binarized (if they don't already represent binary symbols) and each bin of the binarizations is associated with a set of parameters.
  • the associated parameters are then mapped onto a small set of indices and all bins that are associated with a particular index are coded by a particular binary entropy encoder and the corresponding codeword or codewords are written to a particular partial bitstream.
  • the source symbols are decoded by requests for source symbols.
  • the same binarization scheme as for the encoder is used to convert these request for source symbols into requests for bins, and each request for a bin is associated with the same set of parameters as the corresponding bin at the encoder side.
  • the associated parameters are again mapped onto a small set of indices, where the mapping is the same as at the encoder side.
  • all bin requests that are associated with a particular index are decoded by a particular binary entropy decoder, which reads a codeword or multiple codewords from the corresponding partial bitstream.
  • the encoders and decoders are configured in a way that different binary entropy encoders and decoders use different coding algorithms (i.e., a different mapping between bin sequences and codewords).
  • the set of associated parameters represents a measure for the probability of one of the two bin values for a particular bin (or a set of equivalent measures).
  • the measure for the probability can for example represent a probability state, which can be represented by a value of a set of N values.
  • the probability estimation is decoupled from the actual coding of the bins.
  • the probability measure is then mapped onto a probability index, by which the encoding and decoding of a bin is assigned to a particular binary entropy encoder and decoder.
  • the mapping of the probability measures to probability indexes can also be varied over time, e.g. in dependence of already transmitted symbols.
  • the assignment can be adapted to the number of bins that have been coded with a certain probability measure or probability index in order to ensure that the created partial bitstreams have similar bit rates.
  • the number of binary entropy encoders and decoders is less than the number of possible values for the probability measure.
  • a large number of possible probability measures, which allows an accurate estimation of the associated probability is mapped onto a small number of probability indices, where each of those is associated with a particular binary entropy encoder and decoder.
  • the small number of binary entropy encoders and decoders provides the advantage that the overhead for transmitting the partial bitstream can be kept small and that the synchronization overhead is of encoding and decoding system can be kept small.
  • the binary encoders and decoders can be particularly optimized for a representative of the group of probability measures, which allows a high coding efficiency (similar to the best known entropy coding schemes).
  • simple encoding and decoding algorithms which may consist of a simple mapping of bin sequences to codewords and vice versa, can be designed for the representatives of the groups of probability measure. This reduces the complexity of the encoding and decoding system, while still providing a high coding efficiency.
  • the set of parameters that is associated with a bin consists of a measure for the probability of one of the two bin values and one or more additional parameters.
  • Fig. 1 shows a bock diagram of an encoder according to an embodiment
  • Fig. 2 shows a bock diagram of an decoder suitable for decoding bitstream generated by the encoder of Fig. 1, according to an embodiment
  • Fig. 3 shows a schematic diagram illustrating a data packet with multiplexed partial bitstreams according to an embodiment
  • Fig. 4 shows a schematic diagram illustrating a data packet with an alternative segmentation using fixed-size segments according to a further embodiment
  • Fig. 5 shows a bock diagram of an encoder according to an embodiment using partial bitstream interleaving
  • Fig. 6 shows a schematic illustrating examples for the status of a codeword buffer at the encoder side of Fig. 5 according to an embodiment
  • Fig. 7 shows a bock diagram of a decoder according to an embodiment using partial bitstream interleaving
  • Fig. 8 shows a bock diagram of a decoder according to an embodiment using codeword interleaving using a single set of codewords
  • Fig. 9 shows a bock diagram of an encoder according to an embodiment using interleaving of fixed-length bit sequences
  • Fig. 10 shows a schematic illustrating examples for the status of a global bit buffer at the encoder side of Fig. 9 according to an embodiment
  • Fig. 11 shows a bock diagram of a decoder according to an embodiment using interleaving of fixed-length bit sequences
  • Fig. 12 shows a graph for illustrating an optimal probability interval discretization into
  • K 4 intervals assuming a uniform probability distribution in ;
  • Fig. 13 shows a schematic diagram illustrating a tree of binary events for an LPB probability of P ⁇ and an associated variable length code obtained by the Huffman algorithm
  • Fig. 14 shows a graph from which the relative bit rate increase for optimal codes C given a maximum number of table entries ⁇ m may be gathered;
  • VNB2VLC codes with a maximum number of ⁇ m ⁇ table entries
  • Fig. 16 shows a schematic diagram illustrating an example for conversion of a ternary choice tree into a full binary choice tree
  • Fig. 17 shows a block diagram of a system comprising an encoder (left part) and decoder (right part) according to an embodiment
  • Fig. 18 is an illustration of the basic principle of coding transform coefficients according to an embodiment using the entropy coding according to any of the embodiments of Fig. 1 1 17,
  • Fig. 19 shows two examples for coding the significance mapping (the marked symbols are not transferred).
  • Fig. 20 shows binarization for the magnitudes of the transform coefficients (ABS),
  • Fig. 21 shows block types and their classification for the H.264/AVC standard
  • Fig. 22 shows context modeling for the one-bit symbol CBP4.
  • Fig. 23 shows examples of the context modeling for coding the magnitudes of the significant transform coefficients.
  • Fig. 24 shows a block diagram of an entropy encoder arrangement according to an application example
  • Fig. 25 shows a detailed diagram of the initializer from Fig. 24;
  • Fig. 26 shows a sequence of steps for calculating a reference index to reference an initialization probability information table including probability information for the least probable symbol (LPS) and a value of the corresponding most probable symbol (MPS) as initialization probability information;
  • LPS least probable symbol
  • MPS most probable symbol
  • Fig. 27 shows a schematic block diagram of an entropy decoder arrangement according to an application example
  • Fig. 28 shows a high-level block diagram of a coding environment
  • Fig. 29 shows a block diagram of the entropy coding part of the coding environment of
  • Fig. 28; Fig. 30 shows a schematic diagram illustrating the spatial subdivision of a picture or video frame into macroblock pairsn;
  • Fig. 31a shows a schematic diagram illustrating the frame mode, in accordance with an embodiment of the present invention.
  • Fig. 31b shows a schematic diagram illustrating the field mode, in accordance with an embodiment of the present invention.
  • Fig. 32 shows a flow diagram illustrating the encoding of syntax elements with context assignments based on neighboring syntax elements in accordance with an embodiment of the present invention.
  • Fig. 33 shows a flow diagram illustrating the binary entropy coding of the syntax elements based on the context model to which it is assigned in accordance with an embodiment of the present invention.
  • Fig. 34 shows a schematic diagram illustrating the addressing scheme of the macroblocks in accordance with an embodiment of the present invention.
  • Fig. 35 shows a table illustrating how to obtain the macroblock address mbAddrN indicating the macroblock containing a sample having coordinates xN and yN relative to the upper-left sample of a current macroblock and, additionally, the y coordinate yM for the sample in the macroblock mbAddrN for that sample, dependent on the sample being arranged beyond the top or the left border of the current macroblock, the current macroblock being frame or field coded, and the current macroblock being the top or the bottom macroblock of the current macroblock pair, and, eventually, the macroblock mbAddrA being frame or field coded and the line in which the sample lies having an odd or even line number yN.
  • Fig. 36 shows a schematic illustrating macroblock partitions, sub-macroblock partitions, macroblock partitions scans, and sub-macroblock partition scans.
  • Fig. 37 shows a high-level block diagram of a decoding environment in which the present invention may be employed.
  • Fig. 38 shows a flow diagram illustrating the decoding of the syntax elements coded as shown in Fig. 32 and 33 from the coded bit stream;
  • Fig. 39 shows a flow diagram illustrating the entropy decoding process and the decoding process of Fig. 3;
  • Fig. 40 shows a basic coding structure for the emerging H.264/AVC video encoder for a macroblock.
  • Fig. 41 illustrates a context template consisting of two neighboring syntax elements A and B to the left and on the top of the current syntax element C.
  • Fig. 42 shows an illustration of the subdivision of a picture into slices.
  • Fig. 43 shows, to the left, intra_4x4 prediction conducted for samples a-p of a block using samples A_Q, and to the right, "prediction directions for intra_4x4 prediction.
  • Fig. 44 a part of the binarization coding tree related to the binarization scheme Fig. 46; Fig. 45 a schematic diagram illustrating the binarization of an absolute data value;
  • Fig. 46 table showing bin strings into which an absolute value is binarized
  • Fig. 47 table showing bin strings into which an absolute value is binarized
  • Fig. 48 shows a pseudo-C code for performing a binarization
  • Fig. 49 shows a schematic diagram illustrating the decoding of an entropy coded bit stream into a data value
  • Fig. 50 shows a schematic diagram illustrating the recovering of a data value from a binarization of the data value
  • Fig. 51 shows a schematic diagram illustrating the extraction with regard to the suffix part in the process of Fig. 50;
  • Fig. 52 shows a flow diagram illustrating the encoding of the syntax element mb_field_decoding_flag
  • Fig. 53 flow diagram illustrating a process of assigning context models to the mb_field_decoding_flags ;
  • Fig. 54 shows a flow diagram illustrating the decoding of the syntax element mb_field_decoding_flag from the coded bit stream as derived by the encoding scheme of Fig. 52;
  • Fig. 55 shows a schematic flow diagram illustrating the encoding of a video frame or picture
  • Fig. 56 illustrating the different scanning patterns used for field coded macroblocks and frame coded macroblocks.
  • Fig. 57 shows a table illustrating a significance map as obtained from the exemplarily chosen transform coefficient levels
  • Fig. 58 shows a flow diagram illustrating the encoding of the syntax elements last_significant_coeff_flag and significant_coeff_flag in accordance with an embodiment of the present invention
  • Fig. 59 shows a pseudo C-code illustrating the parsing process on decoder side
  • Fig. 60 shows a flow diagram illustrating the decoding of the syntax elements significant_coeff_flag and last_significant_coeff_flag from the coded bit stream as derived by the encoding scheme of Fig. 58;
  • Fig. 61a,b schematically show the process of decoding a run of MPS for an entropy encoder/decoder pair in accordance with an embodiment
  • Fig. 62 schematically shows an example for the quantization of the interval width according to an embodiment.
  • each source symbol 1 is associated with a category of a set of one or more categories.
  • the categories can specify the type of the source symbol.
  • a separate category may be associated with macroblock coding modes, block coding modes, reference picture indices, motion vector differences, subdivision flags, coded block flags, quantization parameters, transform coefficient levels, etc.
  • different categorizations of source symbols are possible.
  • each source symbol can take a value of a finite or countable infinite set of values, where the set of possible source symbol values can differ for different source symbol categories.
  • the source symbols 1 are converted into ordered sets of binary decisions and these binary decisions are then processed by simple binary coding algorithms. Therefore, the binarizer 2 bijectively maps the value of each source symbol 1 onto a sequence (or string) of bins 3.
  • the sequence of bins 3 represents a set of ordered binary decisions.
  • Each bin 3 or binary decision can take one value of a set of two values, e.g. one of the values 0 and 1.
  • the binarization scheme can be different for different source symbol categories.
  • the binarization scheme for a particular source symbol category can depend on the set of possible source symbol values and/or other properties of the source symbols for the particular category.
  • Table 1 illustrates three example binarization schemes for countable infinite sets. Binarization schemes for countable infinite sets can also be applied for finite sets of symbol values. In particular for large finite sets of symbols values, the inefficiency (resulting from unused sequences of bins) can be negligible, but the universality of such binarization schemes provides an advantage in terms of complexity and memory requirements. For small finite sets of symbol values, it is often preferable (in terms of coding efficiency) to adapt the binarization scheme to the number of possible symbol values.
  • Table 2 illustrates three example binarization schemes for finite sets of 8 values.
  • Binarization schemes for finite sets can be derived from the universal binarization schemes for countable infinite sets by modifying some sequences of bins in a way that the finite sets of bin sequences represent a redundancy-free code (and potentially reordering the bin sequences).
  • the truncated unary binarization scheme in Table 2 was created by modifying the bin sequence for the source symbol 7 of the universal unary binarization (see Table 1).
  • the truncated and reordered Exp-Golomb binarization of order 0 in Table 2 was created by modifying the bin sequence for the source symbol 7 of the universal Exp-Golomb order 0 binarization (see Table 1) and by reordering the bin sequences (the truncated bin sequence for symbol 7 was assigned to symbol 1).
  • the truncated bin sequence for symbol 7 was assigned to symbol 1.
  • Table 1 Binarization examples for countable infinite sets (or large finite sets).
  • Table 2 Binarization examples for finite sets.
  • the parameter assigner 4 assigns a set of one or more parameters to each bin 3 and outputs the bin with the associated set of parameters 5.
  • the set of parameters is determined in exactly the same way at encoder and decoder.
  • the set of parameters may consist of one or more of the following parameters: - a measure for an estimate of the probability for one of the two possible bin values for the current bin,
  • the parameter assigner 4 associates each bin 3,5 with a measure for an estimate of the probability for one of the two possible bin values for the current bin. In a further preferred embodiment of the invention, the parameter assigner 4 associates each bin 3,5 with a measure for an estimate of the probability for the less probable or more probable bin value for the current bin and an identifier specifying an estimate for which of the two possible bin values represents the less probable or more probable bin value for the current bin. It should be noted that the probability for the less probable or more probable bin value and the identifier specifying which of the two possible bin values represents the less probable or more probable bin value are equivalent measures for the probability of one of the two possible bin values.
  • the parameter assigner 4 associates each bin 3,5 with a measure for an estimate of the probability for one of the two possible bin values for the current bin and one or more further parameters (which may be one or more of the above listed parameters). In a further preferred embodiment of the invention, the parameter assigner 4 associates each bin 3,5 with a measure for an estimate of the probability for the less probable or more probable bin value for the current bin, an identifier specifying an estimate for which of the two possible bin values represents the less probable or more probable bin value for the current bin, and one or more further parameters (which may be one or more of the above listed parameters).
  • the parameter assigner 4 determines one or more of the above mentioned probability measures (measure for an estimate of the probability for one of the two possible bin values for the current bin, measure for an estimate of the probability for the less probable or more probable bin value for the current bin, identifier specifying an estimate for which of the two possible bin values represents the less probable or more probable bin value for the current bin) based on a set of one or more already encoded symbols.
  • the encoded symbols that are used for determining the probability measures can include one or more already encoded symbols of the same symbol category, one or more already encoded symbols of the same symbol category that correspond to data sets (such as blocks or groups of samples) of neighboring spatial and/or temporal locations (in relation to the data set associated with the current source symbol), or one or more already encoded symbols of different symbol categories that correspond to data sets of the same and/or neighboring spatial and/or temporal locations (in relation to the data set associated with the current source symbol).
  • Each bin with an associated set of parameters 5 that is output of the parameter assigner 4 is fed into a bin buffer selector 6.
  • the bin buffer selector 6 potentially modifies the value of the input bin 5 based on the input bin value and the associated parameters 5 and feds the output bin 7 - with a potentially modified value - into one of two or more bin buffers 8.
  • the bin buffer 8 to which the output bin 7 is sent is determined based on the value of the input bin 5 and/or the value of the associated parameters 5.
  • the bin buffer selector 6 does not modify the value of the bin, i.e., the output bin 7 has always the same value as the input bin 5.
  • the bin buffer selector 6 determines the output bin value 7 based on the input bin value 5 and the associated measure for an estimate of the probability for one of the two possible bin values for the current bin.
  • the output bin value 7 is set equal to the input bin value 5 if the measure for the probability for one of the two possible bin values for the current bin is less than (or less than or equal to) a particular threshold; if the measure for the probability for one of the two possible bin values for the current bin is greater than or equal to (or greater than) a particular threshold, the output bin value 7 is modified (i.e., it is set to the opposite of the input bin value).
  • the output bin value 7 is set equal to the input bin value 5 if the measure for the probability for one of the two possible bin values for the current bin is greater than (or greater than or equal to) a particular threshold; if the measure for the probability for one of the two possible bin values for the current bin is less than or equal to (or less than) a particular threshold, the output bin value 7 is modified (i.e., it is set to the opposite of the input bin value).
  • the value of the threshold corresponds to a value of 0.5 for the estimated probability for both possible bin values.
  • the bin buffer selector 6 determines the output bin value 7 based on the input bin value 5 and the associated identifier specifying an estimate for which of the two possible bin values represents the less probable or more probable bin value for the current bin.
  • the output bin value 7 is set equal to the input bin value 5 if the identifier specifies that the first of the two possible bin values represents the less probable (or more probable) bin value for the current bin, and the output bin value 7 is modified (i.e., it is set to the opposite of the input bin value) if identifier specifies that the second of the two possible bin values represents the less probable (or more probable) bin value for the current bin.
  • the bin buffer selector 6 determines the bin buffer 8 to which the output bin 7 is sent based on the associated measure for an estimate of the probability for one of the two possible bin values for the current bin.
  • the set of possible values for the measure for an estimate of the probability for one of the two possible bin values is finite and the bin buffer selector 6 contains a table that associates exactly one bin buffer 8 with each possible value for the estimate of the probability for one of the two possible bin values, where different values for the measure for an estimate of the probability for one of the two possible bin values can be associated with the same bin buffer 8.
  • the range of possible values for the measure for an estimate of the probability for one of the two possible bin values is partitioned into a number of intervals
  • the bin buffer selector 6 determines the interval index for the current measure for an estimate of the probability for one of the two possible bin values
  • the bin buffer selector 6 contains a table that associates exactly one bin buffer 8 with each possible value for the interval index, where different values for the interval index can be associated with the same bin buffer 8.
  • input bins 5 with opposite measures for an estimate of the probability for one of the two possible bin values are fed into the same bin buffer 8.
  • the association of the measure for an estimate of the probability for one of the two possible bin values for the current bin with a particular bin buffer is adapted over time, e.g. in order to ensure that the created partial bitstreams have similar bit rates.
  • the bin buffer selector 6 determines the bin buffer 8 to which the output bin 7 is sent based on the associated measure for an estimate of the probability for the less probable or more probable bin value for the current bin.
  • the set of possible values for the measure for an estimate of the probability for the less probable or more probable bin value is finite and the bin buffer selector 6 contains a table that associates exactly one bin buffer 8 with each possible value of the estimate of the probability for the less probable or more probable bin value, where different values for the measure for an estimate of the probability for the less probable or more probable bin value can be associated with the same bin buffer 8.
  • the range of possible values for the measure for an estimate of the probability for the less probable or more probable bin value is partitioned into a number of intervals
  • the bin buffer selector 6 determines the interval index for the current measure for an estimate of the probability for the less probable or more probable bin value
  • the bin buffer selector 6 contains a table that associates exactly one bin buffer 8 with each possible value for the interval index, where different values for the interval index can be associated with the same bin buffer 8.
  • the association of the measure for an estimate of the probability for the less probable or more probable bin value for the current bin with a particular bin buffer is adapted over time, e.g. in order to ensure that the created partial bitstreams have similar bit rates.
  • Each of the two or more bin buffers 8 is connected with exactly one bin encoder 10 and each bin encoder is only connected with one bin buffer 8.
  • Each bin encoder 10 reads bins from the associated bin buffer 8 and converts a sequence of bins 9 into a codeword 11, which represents a sequence of bits.
  • the bin buffers 8 represent first-in-first-out buffers; bins that are fed later (in sequential order) into a bin buffer 8 are not encoded before bins that are fed earlier (in sequential order) into the bin buffer.
  • the codewords 11 that are output of a particular bin encoder 10 are written to a particular partial bitstream 12.
  • the overall encoding algorithm converts source symbols 1 into two or more partial bitstreams 12, where the number of partial bitstreams is equal to the number of bin buffers and bin encoders.
  • a bin encoder 10 converts a variable number of bins 9 into a codeword 11 of a variable number of bits.
  • the bin encoding which is done by the bin encoders 10, can be specifically designed for different sets of parameters 5.
  • the bin encoding and encoding can be optimized (in terms of coding efficiency and/or complexity) for different groups of estimated probabilities.
  • this allows a reduction of the encoding/decoding complexity relative to state-of-the-art entropy coding algorithms with similar coding efficiency.
  • it allows an improvement of the coding efficiency relative to state-of-the-art entropy coding algorithms with similar encoding/decoding complexity.
  • the bin encoders 10 implement different encoding algorithms (i.e.
  • the bin encoders 10 implement different encoding algorithms for different groups of measures for an estimate of the probability for the less probable or more probable bin value for the current bin.
  • the bin encoders 10 implement different encoding algorithms for different channel protection codes.
  • the bin encoders 10 implement different encoding algorithms for different encryption schemes.
  • the bin encoders 10 implement different encoding algorithms for different combinations of channel protection codes and groups of measures for an estimate of the probability for one of the two possible bin values 5 for the current bin.
  • the bin encoders 10 implement different encoding algorithms for different combinations of channel protection codes and groups of measures for an estimate of the probability for the less probable or more probable bin value 5 for the current bin. In a further preferred embodiment, the bin encoders 10 implement different encoding algorithms for different combinations of encryption schemes and groups of measures for an estimate of the probability for one of the two possible bin values 5 for the current bin. In a further preferred embodiment, the bin encoders 10 implement different encoding algorithms for different combinations of encryption schemes and groups of measures for an estimate of the probability for the less probable or more probable bin value 5 for the current bin.
  • the bin encoders 10 - or one or more of the bin encoders - represent binary arithmetic encoding engines.
  • one or more of the bin encoders represent a binary arithmetic coding engine, wherein the mapping from the representative LPS/LPB probability /? L ps of a given bin buffer to a corresponding code interval width ?LPS - i-e. the interval subdivision of the internal state of the binary arithmetic coding engine, which is defined by the current interval width R and the current interval offset L, identifying, for example, the lower bound of the code interval - is realized by using a table lookup.
  • K representative interval width values ⁇ Qo, QK-I ⁇ are used for representing ?LPS with the choice of K and the representative interval width values ⁇ Qo, QK-I) being dependent on the bin buffer.
  • arithmetic encoding of a bin may involve the substeps of mapping the current interval width R to a quantization index q with values in ⁇ 6>, K-l ) and performing the interval subdivision by accessing the corresponding partial interval width value Q q from a lookup table with using q as an index.
  • each arithmetic coding engine may be separately transmitted, packetized, or stored, or they may be interleaved for the purpose of transmission or storage as described hereinafter.
  • a binary arithmetic coding engine 10 could perform the following steps in coding the bins in its bin buffer 8:
  • R L P S Rtab [q_index] (note that p_state is not mentioned here, as it is fixed for the binary arithmetic coding engine 10 considered, i.e. p_state [encoder], and Rtab has stored therein pre-calculated values for p[p_state[encoder]] Q[q_index]
  • R R- R LPS [that is, R is preliminarily pre-updated as if "bin” was MPS]
  • q_index describes the index of a quantization value read out of Qtab
  • p_state describes the current state (fixed for the binary arithmetic coding engine 10),
  • R LPS describes the interval width corresponding to the LPS
  • valMPS describes the value of the bit corresponding to the MPS.
  • a binary arithmetic decoding engine 22 could perform the following steps in decoding the bins output into bin buffer 20:
  • R LPS Rtab [q_index] (note that p_state is not mentioned here, as it is fixed for the binary arithmetic decoding engine 22 considered, i.e. p_state [encoder], and Rtab has stored therein pre-calculated values for p[p_state[encoder]] Q[q_index]
  • R R - R LPS [that is, R is preliminarily pre-updated as if "bin” was MPS]
  • bin - 1 - valMPS bin is decoded as LPS; bin buffer selector 18 will obtain the actual bin value by use of this bin information and valMPS
  • bin ⁇ - valMPS (bin is decoded as MPS; bin buffer selector 18 will obtain the
  • q_index describes the index of a quantization value read out of Qtab
  • p_state describes the current state (fixed for the binary arithmetic decoding engine 22)
  • RLPS describes the interval width corresponding to the LPS
  • valMPS describes the value of the bit corresponding to the MPS
  • V describes a value from the interior of the
  • the bin encoders 10 - or one or more of the bin encoders - represent entropy encoders that directly map sequences of input bins 9 onto codewords 10.
  • Such mappings can be efficiently implemented and don't require a complex arithmetic coding engine.
  • the inverse mapping of codewords onto sequences of bins (as done in the decoder) have to be unique in order to guarantee perfect decoding of the input sequence, but the mapping of bin sequences 9 onto codewords 10 doesn't necessarily need to be unique, i.e., it is possible that a particular sequence of bins can be mapped onto more than one sequence of codewords.
  • the mapping of sequences of input bins 9 onto codewords 10 is bijective.
  • the bin encoders 10 - or one or more of the bin encoders - represent entropy encoders that directly map variable-length sequences of input bins 9 onto variable-length codewords 10.
  • the output codewords represent redundancy-free codes such as general huffman codes or canonical huffman codes.
  • the output codewords represent redundant codes suitable for error detection and error recovery.
  • the output codewords represent encryption codes suitable for encrypting the source symbols.
  • Table 3 Examples for mappings between bin sequences and codewords. sequence of bins codewords
  • the bin encoders 10 - or one or more of the bin encoders - represent entropy encoders that directly map variable-length sequences of input bins 9 onto fixed-length codewords 10. In a further preferred embodiment of the invention, the bin encoders 10 - or one or more of the bin encoders - represent entropy encoders that directly map fixed-length sequences of input bins 9 onto variable-length codewords 10.
  • the decoder performs basically the inverse operations of the encoder, so that the (previously encoded) sequence of source symbols 27 is decoded from a set of two or more partial bitstreams 24.
  • the decoder includes two different process flows: A flow for data requests, which replicates the data flow of the encoder, and a data flow, which represents the inverse of the encoder data flow.
  • a flow for data requests which replicates the data flow of the encoder
  • a data flow which represents the inverse of the encoder data flow.
  • the dashed arrows represent the data request flow
  • the solid arrows represent the data flow.
  • the building blocks of the decoder basically replicate the building blocks of the encoder, but implement the inverse operations.
  • each request for a new decoded source symbol 13 is associated with a category of a set of one or more categories.
  • the category that is associated with a request for a source symbol is the same as the category that was associated with the corresponding source symbol during encoding.
  • the binarizer 14 maps the request for a source symbol 13 into one or more requests for a bin that are sent to the parameter assigner 16. As final response to a request for a bin that is sent to the parameter assigner 16 by the binarizer 14, the binarizer 14 receives a decoded bin 26 from the bin buffer selector 18. The binarizer 14 compares the received sequence of decoded bins 26 with the bin sequences of a particular binarization scheme for the requested source symbol and, if the received sequence of decoded bins 26 matches the binarization of a source symbol, the binarizer empties its bin buffer and outputs the decoded source symbol as final response to the request for a new decoded symbol.
  • the binarizer sends another request for a bin to the parameter assigner until the sequence of decoded bins matches one of the bin sequences of the binarization scheme for the requested source symbol.
  • the decoder uses the same binarization scheme that was used for encoding the corresponding source symbol.
  • the binarization scheme can be different for different source symbol categories.
  • the binarization scheme for a particular source symbol category can depend on the set of possible source symbol values and/or other properties of the source symbols for the particular category.
  • the parameter assigner assigns a set of one or more parameters to each request for a bin and sends the request for a bin with the associated set of parameters to the bin buffer selector.
  • the set of parameters that are assigned to a requested bin by the parameter assigner is the same that was assigned to the corresponding bin during encoding.
  • the set of parameters may consist of one or more of the parameters that are mentioned in the encoder description.
  • the parameter assigner 16 associates each request for a bin with a measure for an estimate of the probability for one of the two possible bin values for the current requested bin. In a further preferred embodiment of the invention, the parameter assigner 16 associates each request for a bin with a measure for an estimate of the probability for the less probable or more probable bin value for the current requested bin and an identifier specifying an estimate for which of the two possible bin values represents the less probable or more probable bin value for the current requested bin.
  • the parameter assigner 16 associates each request for a bin 15,17 with a measure for an estimate of the probability for one of the two possible bin values for the current requested bin and one or more further parameters. In a further preferred embodiment of the invention, the parameter assigner 16 associates each request for a bin 15,17 with a measure for an estimate of the probability for the less probable or more probable bin value for the current requested bin, an identifier specifying an estimate for which of the two possible bin values represents the less probable or more probable bin value for the current requested bin, and one or more further parameters (which may one or more of the above listed parameters).
  • the parameter assigner 16 determines one or more of the above mentioned probability measures (measure for an estimate of the probability for one of the two possible bin values for the current requested bin, measure for an estimate of the probability for the less probable or more probable bin value for the current requested bin, identifier specifying an estimate for which of the two possible bin values represents the less probable or more probable bin value for the current requested bin) based on a set of one or more already decoded symbols.
  • the determination of the probability measures for a particular request for a bin replicates the process at the encoder for the corresponding bin.
  • the decoded symbols that are used for determining the probability measures can include one or more already decoded symbols of the same symbol category, one or more already decoded symbols of the same symbol category that correspond to data sets (such as blocks or groups of samples) of neighboring spatial and/or temporal locations (in relation to the data set associated with the current request for a source symbol), or one or more already decoded symbols of different symbol categories that correspond to data sets of the same and/or neighboring spatial and/or temporal locations (in relation to the data set associated with the current request for a source symbol).
  • Each request for a bin with an associated set of parameters 17 that is output of the parameter assigner 16 is fed into a bin buffer selector 18.
  • the bin buffer selector 18 Based on the associated set of parameters 17, the bin buffer selector 18 sends a request for a bin 19 to one of two or more bin buffers 20 and receives a decoded bin 25 from the selected bin buffer 20.
  • the decoded input bin 25 is potentially modified and the decoded output bin 26 - with a potentially modified value - is send to the binarizer 14 as final response to the request for a bin with an associated set of parameters 17.
  • the bin buffer 20 to which the request for a bin is forwarded is selected in the same way as the bin buffer to which the output bin of the bin buffer selector at the encoder side was sent.
  • the bin buffer selector 18 determines the bin buffer 20 to which the request for a bin 19 is sent based on the associated measure for an estimate of the probability for one of the two possible bin values for the current requested bin.
  • the set of possible values for the measure for an estimate of the probability for one of the two possible bin values is finite and the bin buffer selector 18 contains a table that associates exactly one bin buffer 20 with each possible value of the estimate of the probability for one of the two possible bin values, where different values for the measure for an estimate of the probability for one of the two possible bin values can be associated with the same bin buffer 20.
  • the range of possible values for the measure for an estimate of the probability for one of the two possible bin values is partitioned into a number of intervals
  • the bin buffer selector 18 determines the interval index for the current measure for an estimate of the probability for one of the two possible bin values
  • the bin buffer selector 18 contains a table that associates exactly one bin buffer 20 with each possible value for the interval index, where different values for the interval index can be associated with the same bin buffer 20.
  • requests for bins 17 with opposite measures for an estimate of the probability for one of the two possible bin values are forwarded to the same bin buffer 20.
  • the association of the measure for an estimate of the probability for one of the two possible bin values for the current bin request with a particular bin buffer is adapted over time.
  • the bin buffer selector 18 determines the bin buffer 20 to which the request for a bin 19 is sent based on the associated measure for an estimate of the probability for the less probable or more probable bin value for the current requested bin.
  • the set of possible values for the measure for an estimate of the probability for the less probable or more probable bin value is finite and the bin buffer selector 18 contains a table that associates exactly one bin buffer 20 with each possible value of the estimate of the probability for the less probable or more probable bin value, where different values for the measure for an estimate of the probability for the less probable or more probable bin value can be associated with the same bin buffer 20.
  • the range of possible values for the measure for an estimate of the probability for the less probable or more probable bin value is partitioned into a number of intervals
  • the bin buffer selector 18 determines the interval index for the current measure for an estimate of the probability for the less probable or more probable bin value
  • the bin buffer selector 18 contains a table that associates exactly one bin buffer 20 with each possible value for the interval index, where different values for the interval index can be associated with the same bin buffer 20.
  • the association of the measure for an estimate of the probability for the less probable or more probable bin value for the current bin request with a particular bin buffer is adapted over time.
  • the bin buffer selector 18 After receiving a decoded bin 25 from the selected bin buffer 20, the bin buffer selector 18 potentially modifies the input bin 25 and sends the output bin 26 - with a potentially modified value - to the binarizer 14.
  • the input/output bin mapping of the bin buffer selector 18 is the inverse of the input/output bin mapping of the bin buffer selector at the encoder side.
  • the bin buffer selector 18 does not modify the value of the bin, i.e., the output bin 26 has always the same value as the input bin 25.
  • the bin buffer selector 18 determines the output bin value 26 based on the input bin value 25 and the measure for an estimate of the probability for one of the two possible bin values for the current requested bin that is associated with the request for a bin 17.
  • the output bin value 26 is set equal to the input bin value 25 if the measure for the probability for one of the two possible bin values for the current bin request is less than (or less than or equal to) a particular threshold; if the measure for the probability for one of the two possible bin values for the current bin request is greater than or equal to (or greater than) a particular threshold, the output bin value 26 is modified (i.e., it is set to the opposite of the input bin value).
  • the output bin value 26 is set equal to the input bin value 25 if the measure for the probability for one of the two possible bin values for the current bin request is greater than (or greater than or equal to) a particular threshold; if the measure for the probability for one of the two possible bin values for the current bin request is less than or equal to (or less than) a particular threshold, the output bin value 26 is modified (i.e., it is set to the opposite of the input bin value).
  • the value of the threshold corresponds to a value of 0.5 for the estimated probability for both possible bin values.
  • the bin buffer selector 18 determines the output bin value 26 based on the input bin value 25 and the identifier, specifying an estimate for which of the two possible bin values represents the less probable or more probable bin value for the current bin request, that is associated with the request for a bin 17.
  • the output bin value 26 is set equal to the input bin value 25 if the identifier specifies that the first of the two possible bin values represents the less probable (or more probable) bin value for the current bin request, and the output bin value 26 is modified (i.e., it is set to the opposite of the input bin value) if identifier specifies that the second of the two possible bin values represents the less probable (or more probable) bin value for the current bin request.
  • the bin buffer selector sends a request for a bin 19 to one of the two or more bin buffers 20.
  • the bin buffers 20 represent first-in-first-out buffers, which are fed with sequences of decoded bins 21 from the connected bin decoders 22.
  • the bin buffer 20 removes the bin of its content that was first fed into the bin buffer 20 and sends it to the bin buffer selector 18. Bins that are earlier sent to the bin buffer 20 are earlier removed and sent to the bin buffer selector 18.
  • Each of the two or more bin buffers 20 is connected with exactly one bin decoder 22 and each bin decoder is only connected with one bin buffer 20.
  • Each bin decoder 22 reads codewords 23, which represent sequences of bits, from a separate partial bitstream 24.
  • the bin decoder converts a codeword 23 into a sequence of bins 21 that is sent to the connected bin buffer 20.
  • the overall decoding algorithm converts two or more partial bitstreams 24 into a number of decoded source symbols, where the number of partial bitstreams is equal to the number of bin buffers and bin decoders and the decoding of source symbols is triggered by requests for new source symbols.
  • a bin decoder 22 converts codewords 23 of a variable number of bits into a sequence of a variable number of bins 21.
  • One advantage of embodiments of the invention is that the decoding of bins from the two or more partial bitstreams can be done in parallel (e.g. for different groups of probability measures), which reduces the processing time for several implementations.
  • the bin decoding which is done by the bin decoders 22, can be specifically designed for different sets of parameters 17.
  • the bin encoding and decoding can be optimized (in terms of coding efficiency and/or complexity) for different groups of estimated probabilities.
  • this allows a reduction of the encoding/decoding complexity relative to state-of-the-art entropy coding algorithms with similar coding efficiency.
  • it allows an improvement of the coding efficiency relative to state-of-the-art entropy coding algorithms with similar encoding/decoding complexity.
  • the bin decoders 22 implement different decoding algorithms (i.e.
  • the bin decoders 22 implement different decoding algorithms for different groups of measures for an estimate of the probability for the less probable or more probable bin value for the current requested bin.
  • the bin decoders 22 implement different decoding algorithms for different channel protection codes.
  • the bin decoders 22 implement different decoding algorithms for different encryption schmes.
  • the bin decoders 22 implement different decoding algorithms for different combinations of channel protection codes and groups of measures for an estimate of the probability for one of the two possible bin values 17 for the current requested bin.
  • the bin decoders 22 implement different decoding algorithms for different combinations of channel protection codes and groups of measures for an estimate of the probability for the less probable or more probable bin value 17 for the current requested bin. In a further preferred embodiment, the bin decoders 22 implement different decoding algorithms for different combinations of encryption schemes and groups of measures for an estimate of the probability for one of the two possible bin values 17 for the current requested bin. In a further preferred embodiment, the bin decoders 22 implement different decoding algorithms for different combinations of encryption schemes and groups of measures for an estimate of the probability for the less probable or more probable bin value 17 for the current requested bin.
  • the bin decoders 22 do the inverse mapping of the corresponding bin encoders at the encoder side.
  • the bin decoders 22 - or one or more of the bin decoders - represent binary arithmetic decoding engines.
  • the bin decoders 22 - or one or more of the bin decoders - represent entropy decoders that directly map codewords 23 onto sequences of bins 21.
  • Such mappings can be efficiently implemented and don't require a complex arithmetic coding engine.
  • the mapping of codewords onto sequences of bins has to be unique.
  • the mapping of codewords 23 onto sequences of bins 21 is bijective.
  • the bin decoders 10 - or one or more of the bin decoders - represent entropy decoders that directly map variable-length codewords 23 into variable-length sequences of bins 21.
  • the input codewords represent redundancy-free codes such as general huffman codes or canonical huffman codes. Two examples for the bijective mapping of redundancy-free codes to bin sequences are illustrated in Table 3.
  • the input codewords represent redundant codes suitable for error detection and error recovery.
  • the input codewords represent encryption codes.
  • the bin decoders 22 - or one or more of the bin decoders - represent entropy decoders that directly map fixed-length codewords 23 onto variable-length sequences of bins 21. In a further preferred embodiment of the invention, the bin decoders 22 - or one or more of the bin decoders - represent entropy decoders that directly map variable-length codewords 23 onto fixed-length sequences of bins 21.
  • Fig. 1 and 2 showed an embodiment for an encoder for encoding a sequence of symbols 3 and a decoder for reconstructing same.
  • the encoder comprises an assigner 4 configured to assign a number of parameters 5 to each symbol of the sequence of symbols. The assignment is based on information contained within previous symbols of the sequence of symbols such as the category of the syntax element 1 to the representation - such as binarization - of which the current symbol belongs and which, according to the syntax structure of the syntax elements 1, is currently be expected which expectation, in turn, is deducible from the history of previous syntax elements 1 and symbols 3.
  • the encoder comprises a plurality of entropy encoders 10 each of which is configured to convert the symbols 3 forwarded to the respective entropy encoder into a respective bitstream 12, and a selector 6 configured to forward each symbol 3 to a selected one of the plurality of entropy encoders 10, the selection depending on the number of parameters 5 assigned to the respective symbol 3.
  • the decoder for reconstructing a sequence of symbols comprises a plurality of entropy decoders 22, each of which is configured to convert a respective bitstream 23 into symbols 21; an assigner 16 configured to assign a number of parameters 17 to each symbol 15 of a sequence of symbols to be reconstructed based on information contained within previously reconstructed symbols of the sequence of symbols (see 26 and 27 in Fig.
  • the assigner 16 may be configured such that the number of parameters assigned to each symbol comprises, or is, a measure for an estimate of a probability of distribution among the possible symbol values a respective symbol may assume.
  • the sequence of symbols to be reconstructed may be of a binary alphabet and the assigner 16 may be configured such that the estimate of the probability distribution consists of a measure for an estimate of a probability of a less probable or more probable bin value of the two possible bin values of the binary alphabet and an identifier specifying an estimate for which of the two possible bin values represents the less probable or more probable bin value.
  • the assigner 16 may further be configured to internally assign a context to each symbol of the sequence of symbols 15 to be reconstructed based on the information contained within previously reconstructed symbols of the sequence of symbols to be reconstructed with each context having a respective probability distribution estimate associated therewith, and to adapt the probability distribution estimate for each context to an actual symbol statistic based on symbol values of previously reconstructed symbols to which the respective context is assigned.
  • the context may take into account a spatial relationship or neighbourhood of positions to which the syntax elements belong such as in video or picture coding, or even in tables in case of financial applications.
  • the measure for the estimate of the probability distribution for each symbol may be determined based on the probability distribution estimate associated with the context assigned to the respective symbol such as by quantizing the probability distribution estimate associated with the context assigned with the respective symbol to one of a plurality of probability distribution estimate representatives in order to obtain the measure for the estimate of the probability distribution.
  • the selector may be configured such that a bijective association is defined between the plurality of entropy encoders and the plurality of probability distribution estimate representatives.
  • the selector 18 may be configured to change a quantization mapping from a range of the probability distribution estimates to the plurality of probability distribution estimate representatives in a predetermined deterministic way depending on previously reconstructed symbols of the sequence of symbols, over time. That is, selector 18 may change the quantization step sizes, i.e. the intervals of probability distributions mapped onto the individual probability indices bijectively associated with the individual entropy decoders.
  • the plurality of entropy decoders 22 in turn, may be configured to adapt their way of converting symbols into bit streams responsive to a change in the quantization mapping.
  • each entropy decoder 22 may be optimised for, i.e may have an optimal compression rate for, a certain probability distribution estimate within the respective probability distribution estimate quantization interval, and may change its codeword/symbol sequence mapping so as to adapt the position of this certain probability distribution estimate within the respective probability distribution estimate quantization interval upon a change of the latter so as to be optimised.
  • the selector may be configured to change the quantization mapping such that rates by which the symbols are retrieved from the plurality of entropy decoders, are made less dispersed.
  • the binarizer 14 it is noted that same me be left away if the syntax elements are already binary. Further, depending on the type of decoder 22, the existence of the buffers 20 is not necessary. Further, the buffers may be integrated within the decoders.
  • the encoding and decoding is done for a finite set of source symbols. Often a certain quantity of data such as a still image, a frame or field of a video sequence, a slice of an image, a slice of a frame or a field of a video sequence, or a set of successive audio samples, etc. is coded.
  • the partial bitstreams that are created at the encoder side have to be terminated, i.e., it has to be ensured that all source symbols can be decoded from the transmitted or stored partial bitstreams. After the last bin is inserted into the corresponding bin buffer 8, the bin encoder 10 has to ensure that a complete codeword is written to the partial bitstream 12.
  • the bin encoder 10 represents a binary arithmetic coding engine, the arithmetic codeword has to be terminated. If the bin encoder 10 represents an entropy encoder that implements a direct mapping of bin sequences onto codewords, the bin sequence that is stored in the bin buffer after writing the last bin to the bin buffer might not represent a bin sequence that is associated with a codeword (i.e., it might represent a prefix of two or more bin sequences that are associated with codewords). In such a case, any of the codewords associated with a bin sequence that contains the bin sequence in the bin buffer as prefix has to be written to the partial bitstream (the bin buffer has to be flushed).
  • the bin encoder selects one of the codewords with minimum length (in addition to the property that the associated bin sequence must contain the bin sequence in the bin buffer as prefix).
  • the bin decoder 22 may decode more bins than required for the last codeword in a partial bitstream; these bins are not requested by the bin buffer selector 18 and are discarded and ignored.
  • the decoding of the finite set of symbols is controlled by requests for decoded source symbols; if no further source symbol is requested for a quantity of data, the decoding is terminated.
  • the partial bitstreams 12 that are created by the encoder can be transmitted separately, or they can be multiplexed into a single bitstream, or the codewords of the partial bitstreams can be interleaved in a single bitstream.
  • each partial bitstream for a quantity of data is written to one data packet.
  • the quantity of data can be an arbitrary set of source symbols such as a still picture, a field or frame of a video sequence, a slice of a still picture, a slice of a field or frame of a video sequence, or a frame of audio samples, etc.
  • two or more of the partial bitstreams for a quantity of data or all partial bitstreams for a quantity of data are multiplexed into one data packet.
  • the structure of a data packet that contains multiplexed partial bitstreams is illustrated in Figure 3.
  • the data packet 100 consists of a header and one partition for the data of each partial bitstream (for the considered quantity of data).
  • the header 100 of the data packet contains indications for the partitioning of the (remainder of the) data packet into segments of bitstream data 102. Beside the indications for the partitioning, the header may contain additional information.
  • the indications for the partitioning of the data packet are the locations of the beginning of the data segments in units of bits or bytes or multiples of bits or multiples of bytes.
  • the locations of the beginning of the data segments are coded as absolute values in the header of the data packet, either relative to the beginning of the data packet or relative to the end of the header or relative to the beginning of the previous data packet.
  • the locations of the beginning of the data segments are differentially coded, i.e., only the difference between the actual beginning of a data segment and a prediction for the beginning of the data segment is coded.
  • the prediction can be derived based on already known or transmitted information such as the overall size of the data packet, the size of the header, the number of data segments in the data packet, the location of the beginning of preceding data segments. In a preferred embodiment of the invention, the location of the beginning of the first data packet is not coded, but inferred based on the size of the data packet header.
  • the transmitted partition indications are used for deriving the beginning of the data segments. The data segments are then used as partial bitstreams and the data contained in the data segments are fed into the corresponding bin decoders in sequential order.
  • Fig. 4 There are several alternatives for multiplexing the partial bitstreams into a data packet.
  • the payload of the data packet i.e., the data packet 110 without its header 111, is partitioned into segments 112 a predefined way.
  • the data packet payload can be partitioned into segments of the same size. Then each segment is associated with a partial bitstream or with the first part of a partial bitstream 113. If a partial bitstream is greater than the associated data segment, its remainder 114 is placed into the unused space at the end of other data segments.
  • the above described multiplexing of the partial bitstreams (for a quantity of source symbols) in one data packet can have the following disadvantages: On the one hand side, for small data packets, the number of bits for the side information that is required for signaling the partitioning can become significant relative to the actual data in the partial bitstreams, which finally reduces the coding efficiency. On the other hand, the multiplexing may not suitable for applications that require a low delay (e.g. for video conferencing applications). With the described multiplexing, the encoder cannot start the transmission of a data packet before the partial bitstreams have been completely created, since the locations of the beginning of the partitions are not known before.
  • the decoder has to wait until it receives the beginning of the last data segment before it can start the decoding of a data packet.
  • these delays can add-up to an additional overall delay of the system of several video pictures (in particular for bit rates that are close to the transmission bit rate and for encoders/decoders that require nearly the time interval between two pictures for encoding/decoding a picture), which is critical for such applications.
  • the encoder of a preferred embodiment of the invention can be configured in a way that the codewords that are generated by the two or more bin encoders are interlaved into a single bitstream.
  • the bitstream with the interleaved codewords can be directly send to the decoder (when neglecting a small buffer delay, see below).
  • the two or more bin decoders read the codewords directly from the bitstream in decoding order; the decoding can be started with the first received bit.
  • no side information is required for signaling the multiplexing (or interleaving) of the partial bitstreams.
  • the basic structure of an encoder with codeword interleaving is shown in Fig. 5.
  • the bin encoders 10 don't write the codewords directly to the partial bitstreams, but are connected with a single codeword buffer 29, from which codewords are written to the bitstream 34 in coding order.
  • the bin encoders 10 send requests for one or more new codeword buffer entries 28 to the codeword buffer 29 and later send the codewords 30 to the codeword buffer 29, which are stored in the reserved buffer entries.
  • the (in general variable-length) codewords 31 of the codeword buffer 29 are accessed by a codeword writer 32, which writes the corresponding bits 33 to the produced bitstream 34.
  • the codeword buffer 29 operates as a first-in-first-out buffer; codeword entries that are reserved earlier are earlier written to the bitstream.
  • a bin encoder 10 reserves one or more codewords in the codeword buffer 29, whereby the reservation of the one or more codewords in the codeword buffer is triggered by certain events in the connected bin buffer 8.
  • the codeword buffer 29 is operated in a way that the decoder can instantaneously decode the bitstream.
  • the coding order in which the codewords are written to the bitstream is the same as the order in which the corresponding codewords are reserved in the codeword buffer.
  • each bin encoder 10 reserves one codeword, with the reservation being triggered by a certain event in the connected bin buffer. In another preferred embodiment of the invention, each bin encoder 10 reserves more than one codeword, with the reservation being triggered by a certain event in the connected bin buffer. In further preferred embodiment of the invention, the bin encoders 10 reserve a different amount of codewords, where the amount of codewords that are reserved by a particular bin encoder can be dependent on the particular bin encoder and/or other properties of the particular bin encoder/bin buffer (such as the associated probability measure, the number of already written bits, etc.).
  • the codeword buffer is operated as follows. If a new bin 7 is sent to a particular bin buffer 8 and the number of already stored bins in the bin buffer is zero and there is currently no codeword reserved in the codeword buffer for the bin encoder that is connected with the particular bin buffer, the connected bin encoder 10 sends a request to the codeword buffer, by which one or more codeword entries are reserved in the codeword buffer 29 for the particular bin encoder.
  • the codeword entries can have a variable number of bits; an upper threshold for the number of bits in a buffer entry is usually given by the maximum codeword size for the corresponding bin encoder.
  • the next codeword or the next codewords that are produced by the bin encoder are stored in the reserved entry or entries of the codeword buffer. If all reserved buffer entries in the codeword buffer for a particular bin encoder are filled with codewords and the next bin is sent to the bin buffer that is connected with the particular bin encoder, one or more new codewords are reserved in the codeword buffer for the particular bin encoder, etc.
  • the codeword buffer 29 represents a first-in-first-out buffer in a certain way. Buffer entries are reserved in sequential order. Codewords for which the corresponding buffer entries have been reserved earlier are earlier written to the bitstream.
  • the codeword writer 32 checks the status of the codeword buffer 29, either continuously or after a codeword 30 is written to the codeword buffer 29. If the first buffer entry contains a complete codeword (i.e., the buffer entry is not reserved, but includes a codeword), the corresponding codeword 31 and the corresponding buffer entry are removed from the codeword buffer 20 and the bits of the codeword 33 are written to the bitstream. This process is repeated until the first buffer entry does not contain a codeword (i.e., it is reserved or free). At the end of the decoding process, i.e., if all source symbols of the considered quantity of data have been processed, the codeword buffer must be flushed.
  • each bin buffer/ bin encoder For that flushing process, the following is applied for each bin buffer/ bin encoder as first step: If the bin buffer does contain bins, a bin with a particular or an arbitrary value is added until the resulting bin sequence represents a bin sequence that is associated with a codeword (as noted above, one preferred way of adding bins is to add such bin values that produce the shortest possible codeword - or one of those - that is associated with a bin sequence that contains the for the original content of the bin buffer as prefix), then the codeword is written to the next reserved buffer entry for the corresponding bin encoder (and the corresponding) bin buffer is emptied. If more than one buffer entry has been reserved for one or more bin encoders, the codeword buffer may still contain reserved codeword entries.
  • these codeword entries are filled with arbitrary, but valid codewords for the corresponding bin encoders.
  • the shortest valid codeword or one of the shortest valid codewords (if there are multiple) is inserted. Finally, all remaining codewords in the codeword buffer are written to the bitstream.
  • the codeword buffer contains 2 entries that are filled with a codeword and 5 reserved entries.
  • the next free buffer entry is marked.
  • the first entry is filled with a codeword (i.e., the bin encoder 2 just wrote a codeword to a previously reserved entry).
  • this codeword will be removed from the codeword buffer and written to the bitstream.
  • the first reserved codeword for bin encoder 3 is the first buffer entry, but this entry cannot be removed from the codeword buffer, since it is only reserved, but no codeword has been written to this entry.
  • the codeword buffer contains 3 entries that are filled with a codeword and 4 reserved entries.
  • the first entry is marked as reserved and hence the codeword writer cannot write a codeword to the bitstream. Although 3 codewords are contained in the codeword buffer, the codeword writer has to wait until a codeword is written to the first reserved buffer entry for bin encoder 3. Note that the codewords must be written in the order in which they were reserved, in order to be able to invert the process at the decoder side (see below).
  • the basic structure of a decoder with codeword interleaving is shown in Fig. 7.
  • the bin decoders 10 don't read the codewords directly from separate partial bitstreams, but are connected to a bit buffer 38, from which the codewords 37 are read in coding order.
  • the bit buffer 38 is not necessarily required, since the codewords could also be directly read from the bitstream.
  • the bit buffer 38 is mainly included in the illustration for clearly separate different aspects of the processing chain.
  • the bits 39 of the bitstream 40 with interleaved codewords are sequentially inserted into the bit buffer 38, which represents a first- in-first-out buffer.
  • a particular bin decoder 22 receives a request for one or more bin sequences35, the bin decoder 22 reads one or more codewords 37 from the bit buffer 38 via requests for bits 36.
  • the decoder can instantaneously decode the source symbols. Note that the encoder (as described above) must ensure by suitably operating the codeword buffer that the codewords are written in the same order to the bitstream in which they are requested by the bin decoders.
  • the entire decoding process is triggered by requests for source symbols. Parameters as the number of codewords that are reserved at the encoder side by a particular bin encoder and the number of codewords that are read by the corresponding bin decoder must be the same.
  • a bin decoder 22 reads one or more codewords from the bit buffer 38 at one time instant, whereby the reading of the one or more codewords from the bit buffer is triggered by certain events in the connected bin buffer 20.
  • the decoder is operated in a way that one or more codewords are read when a request for a bin 19 is sent to a particular bin buffer 20 and the bin buffer doesn't contain any bins. But it is also possible to trigger the reading of codewords by other events, e.g. if the number of bins in the bin buffer is below a predefined threshold.
  • each bin decoder 22 reads one codeword, with the reading being triggered by a certain event in the connected bin buffer. In another preferred embodiment of the invention, each bin decoder 22 reads more than one codeword, with the reading being triggered by a certain event in the connected bin buffer. In further preferred embodiment of the invention, the bin decoders 22 read a different amount of codewords, where the amount of codewords that are read by a particular bin decoder can be dependent on the particular bin decoder and/or other properties of the particular bin decoder/bin buffer (such as the associated probability measure, the number of already read bits, etc.).
  • the reading of codewords from the bit buffer is operated as follows. If a new bin request 19 is sent from the bin buffer selector 18 to a particular bin buffer 20 and the number of bins in the bin buffer is zero, the connected bin decoder 22 reads one or more codewords 37 from the bit buffer 38, via bit request 36 to the bit buffer 38. The bin decoder 22 converts the read codewords 37 into sequences of bins 21 and stores these bin sequences in the connected bin buffer 20. As final response to the request for a bin 19, the first inserted bin is removed from the bin buffer 20 and sent to the bin buffer selector 18. As response the further bin requests, the remaining bins in the bin buffer are removed until the bin buffer is empty.
  • An additional bin request triggers the bin decoder to read one or more new codewords from the bit buffer, etc.
  • the bit buffer 38 represents a first- in-first-out buffer of a predefined size and is continuously filled with bits 39 from the bitstream 40.
  • the codeword buffer at the encoder side can be operated in the way described above.
  • each of the plurality of entropy decoders may be a variable length decoder configured to map codewords of fixed lengths to symbol sequences of variable lengths, and a codeword entry such as the output of the codeword buffer 43 may be provided for receiving a single stream of interleaved codewords.
  • the plurality of entropy decoders 22 may be configured to retrieve the codewords from the codeword entry in a sequential order depending on an order in which the symbols of the sequence of symbols to be reconstructed as retrieved by the selector 18 from the plurality of entropy decoders result in a new symbol sequence to be mapped from a new codeword at the respective entropy decoders.
  • the described codeword interleaving does not require that any partitioning information is sent as side information. And since the codewords are interleaved in the bitstream, the delay is in general small. However, it is not guaranteed that a particular delay constraint (e.g. specified by a maximum number of bits that are stored in the codeword buffer) is obeyed. Furthermore, the required buffer size for the codeword buffer can theoretically become very large. When considering the example in Fig. 6(b), it might be possible that no further bins are send to bin buffer 3 and hence the bin encoder 3 will not send any new codeword to the codeword buffer until the flushing process at the end of the data packet is applied.
  • a particular delay constraint e.g. specified by a maximum number of bits that are stored in the codeword buffer
  • the bins that have been inserted at the encoder side in order to obey the delay constraint must be discarded.
  • discarding of bins basically the same mechanism as at the encoder side can be used. In the following two preferred embodiments for such a delay control are described.
  • the measure for the delay is the number of active buffer entries in the codeword buffer, where the number of active buffer entries is the number of reserved buffer entries plus the number of buffer entries that contain codewords.
  • the first buffer entry is always a reserved buffer entry or a free buffer entry, since if the first buffer entry contains a codeword, this codeword is written to the bitstream.
  • the delay measure D in bits is not required by the system, but the maximum number of codewords N must be known to both encoder and decoder.
  • the maximum number of codeword buffer entries N is fixed by the application.
  • the maximum number of codeword buffer entries N is signaled inside the bitstream, e.g., in the header of the data packet (or slice header) or in a parameter set, which is included in the bitstream.
  • a bin encoder 10 sends a request for the reservation of one or more new buffer entries to the codeword buffer 29, the following process is executed before a new codeword buffer entry is reserved (i.e., it is executed multiple times if multiple codeword buffer entries are reserved by one request): If the number of currently active buffer entries plus 1 (taking into account the buffer entry that will be reserved next) is greater than the maximum number of codeword buffer entries N, the first buffer entry (which is reserved) is flushed by the process described in the following until the number of currently active buffer entries plus 1 is less than or equal to the maximum number of codeword buffer entries N.
  • the flushing of a reserved buffer entry is similar to the flushing at the end of a data packet:
  • the bin encoder 10 that has reserved the corresponding first buffer entry is flushed by adding bins with particular or arbitrary values to the connected bin buffer 8 until the resulting bin sequence represents a bin sequence that is associated with a codeword, the codeword is then written to the reserved buffer entry and it is finally added to the bitstream (while emptying the bin buffer and removing the previously reserved buffer entry).
  • one preferred way for adding bins to the bin buffer is to add those bins that produce the shortest possible codeword.
  • a similar process is executed for discarding the bins that have been added to obey the delay constraint.
  • the decoder maintains a counter C that counts the codewords that have been read from the bit buffer (this counter can be maintained in the bit buffer).
  • This counter C is initialized (e.g. with zero) at the beginning of the decoding of a data packet and is increased by one after a codeword is read.
  • each bin decoder 22 contains a counter Cx, which stores the value of the codeword counter C before the last codeword was read by the corresponding bin decoder 22. I.e., when a particular bin decoder 22 reads a new codeword, its counter Cx is set equal to C as a first step and then the codeword is read from the bit buffer.
  • the measure for the delay is the sum of the maximum codeword lengths for the active buffer entries in the codeword buffer, where the maximum codeword length for a particular buffer entry depends on the bin decoded that is associated with that buffer entry.
  • the maximum codeword lengths for the buffer entries are indicated in the examples in 6.
  • the first buffer entry is always a reserved buffer entry or a free buffer entry, since if the first buffer entry contains a codeword, this codeword is written to the bitstream.
  • the maximum allowed buffer delay (as determined by the application) be D bits. This maximum buffer delay D must be known to both encoder and decoder. In a preferred embodiment of the invention, the maximum buffer delay D is fixed by the application.
  • the maximum buffer delay D is signaled inside the bitstream, e.g., in the header of the data packet (or slice header) or in a parameter set, which is included in the bitstream. It can be signaled in units of bits, or bytes, or a multiple of bits, or a multiple of bytes. If a bin encoder 10 sends a request for the reservation of one or more new buffer entries to the codeword buffer 29, the following process is executed before a new codeword buffer entry is reserved (i.e., it is executed multiple times if multiple codeword buffer entries are reserved by one request).
  • the first buffer entry (which is reserved) is flushed by the process described above until the sum of the maximum codeword lengths for all active buffer entries plus the maximum codeword length for the buffer entry that will be reserved is less than or equal to the maximum buffer delay D.
  • the maximum buffer delay D is set equal to 32.
  • next buffer entry is reserved by bin encoder 2 for which the maximum codeword length is equal to 3
  • the first buffer entry is not flushed, since 29 + 3 is not greater than 32.
  • the next buffer entry is reserved by bin encoder 1 for which the maximum codeword length is equal to 7
  • the first buffer entry is flushed, since 29 + 7 is greater than 32.
  • the flushing of the reserved buffer entry is done as described above (by adding bin with particular or arbitrary values to the corresponding bin buffer).
  • the decoder maintains a counter C that counts the maximum codeword length for the codewords that have been read from the bit buffer (this counter can be maintained in the bit buffer). Note that the maximum codeword lengths that are associated with different bin decoders can be different.
  • the counter C is initialized (e.g. with zero) at the beginning of the decoding of a data packet and it is increased after a codeword is read. This counter is not increased by the actual length of the read codewords, but by its maximum length.
  • each bin decoder 22 contains a counter Cx, which stores the value of the codeword counter C before the last codeword was read by the corresponding bin decoder 22. I.e., when a particular bin decoder 22 reads a new codeword, its counter Cx is set equal to C as a first step and then the codeword is read from the bit buffer.
  • the plurality of entropy decoders 22 and the selector 18 may be configured to intermittently discard suffixes of symbol sequences so as to not participate in forming the sequence of symbols to be reconstructed. 29.
  • the intermittently discarding may be performed at events where a number of codewords having been retrieved from the codeword entry by the plurality of entropy decoders between two consecutive codeword retrievals of a respective entropy decoder from the codeword entry, fulfils a predetermined criterion.
  • the plurality of entropy encoders and the codeword buffer may, in turn, be configured to intermittently extend currently forwarded but not yet mapped symbols to valid symbol sequences by don't-care symbols having the currently forwarded but not yet mapped symbols as prefix, map the thus extended symbol sequences into codewords, enter the thus obtained codewords into the reserved codeword entries and flush the codeword entries.
  • the intermittently extending, entering and flushing may take place at events where a number of reserved codeword entries plus a number of codeword entries having codewords entered therein fulfils a predetermined criterion.
  • the predetermined criteria may take the maximum lengths of codewords of the plurality of encoder/decoder pairs into account.
  • all bin decoders 22 read codewords (in the general case, variable-length codewords) from a single bit buffer 38.
  • the reading of the codewords cannot be done in parallel, since the codeword must be read in the correct order. That means, a particular bin decoder must wait until other bin decoders finish the reading of codewords.
  • this access of the variable-length codewords can be a bottleneck for the entire decoding process.
  • the described embodiments of the invention there are some variations of the described embodiments of the invention that can be employed for reducing the complexity of the access from the single bit buffer, a few of them will be described in the following.
  • a single set of codewords (representing for instance a redundancy-free prefix code) and the set of codewords that is used for each bin decoder 22 is a subset of the single codeword set.
  • different bin decoders 22 can use different subsets of the single codeword set. Even if the codeword sets that are used by some of the bin decoders 22 are the same, their association with bin sequences is different for different bin decoders 22.
  • the same set of codewords is used for all bin decoders 22. If we have a single codeword set that includes the codeword sets for all bin decoders as subsets, the parsing of the codewords can be done outside the bin decoders, which can reduce the complexity of the codeword access.
  • the encoding process is not changed in relation to the above described process.
  • the modified decoding process is illustrated in Fig. 8.
  • a single codeword reader is fed with bits 46 from the bitstream 40 and parses the - in general variable-length - codewords.
  • the read codewords 44 are inserted into a codeword buffer 43, which represents a first-in-first-out buffer.
  • a bin decoder 22 sends a request for one or more codewords 41 to the codeword buffer 43 and as response to this request, the one or more codewords are removed from the codeword buffer (in sequential order) and send to the corresponding bin decoder 22.
  • the potentially complex codeword parsing can be done in a background process and it doesn't need to wait for the bin decoders.
  • the bin decoders access already parsed codewords, the potentially complex codeword parsing is no more part of a request to the overall buffer. Instead already parsed codewords are send to the bin decoders, which can also be implemented in a way that only codeword indices are send to the bin decoders.
  • a further way of reducing the decoder complexity can be achieved when the bin decoders 22 don't read variable-length codewords from the global bit buffer 38, but instead they always read fixed-length sequences of bits from the global bit buffer 38 and add these fixed-length sequences of bits to a local bit buffer, where each bin decoder 22 is connected with a separate local bit buffer. The variable-length codewords are then read from the local bit buffer.
  • the parsing of variable-length codewords can be done in parallel, only the access of fixed- length sequences of bits has to be done in a synchronized way, but such an access of fixed-length sequences of bits is usually very fast, so that the overall decoding complexity can be reduced for some architectures.
  • the fixed number of bins that are sent to a particular local bit buffer can be different for different local bit buffer and it can also vary over time, depending on certain parameters as events in the bin decoder, bin buffer, or bit buffer.
  • the number of bits that are read by a particular access does not depend on the actual bits that are read during the particular access, which is the important difference to the reading of variable-length codewords.
  • the reading of the fixed-length sequences of bits is triggered by certain events in the bin buffers, bin decoders, or local bit buffers. As an example, it is possible to request the reading of a new fixed-length sequence of bits when the number of bits that are present in a connected bit buffer falls below a predefined threshold, where different threshold values can be used for different bit buffers.
  • the fixed-length sequences of bins are inserted in the same order into the bitstream, in which they are read from the bitstream at the decoder side. It is also possible to combine this interleaving of fixed-length sequences with a low-delay control similar to the ones explained above. In the following, a preferred embodiment for the interleaving of fixed-length sequences of bits is described.
  • Fig. 9 shows an illustration of the basic encoder structure for the embodiment of the invention that interleaves fixed-length sequences of bits for two or more bin encoders.
  • the bin encoders 10 are not connected with a single codeword buffer. Instead, each bin encoder 10 is connected with a separate bit buffer 48, which stores bits for the corresponding partial bitstream. All bit buffers 48 are connected to a global bit buffer 51.
  • the global bit buffer 51 is connected to a bit writer 53, which removes the bits 52 in coding/decoding order from the global bit buffer and writes the removed bits 54 to the bitstream 55.
  • bit buffer 48 On a certain events in a particular bit buffer 48 or the connected bin encoder 10 or bin buffer 8, the bit buffer 48 sends a request 49 to the global bit buffer 51 by which a certain number of bits is reserved in the global bit buffer 51.
  • the requests for the reservation of fixed-length bit sequences 49 are processed in sequential order.
  • the global bit buffer 51 represents a first-in-first-out buffer in a certain way; bits that are reserved earlier are earlier written to the bitstream. It should be noted that different bit buffers 48 can reserved a different amount of bits, which can also vary over time based on already coded symbols; but the number of bits that are reserved by a particular request is known at the time at which the request is sent to the global bit buffer.
  • bit buffers 48 and the global bit buffer 51 are operated as described in the following.
  • the amount of bits that is reserved by a particular bit buffer 48 is denoted as Nx.
  • This number of bits Nx can be different for different bit buffers 48 and it can also vary over time.
  • the number of bits Nx that are reserved by a particular bit buffer 48 is fixed over time.
  • the reservations for a fixed number Nx of bits 49 are triggered based on the number of bits Mx in the bit buffers 48, the number of bits Nx for the reservation requests, and the associated maximum codeword length Lx. Note that each bin encoder 10 can be associated with a different maximum codeword length Lx.
  • the connected bit buffer 49 sends a request 49 for the reservation of Nx bits to the global bit buffer 51.
  • the global bit buffer 51 reserves Nx bits for the particular bit buffer 48 and increases its pointer for the next reservation. After the Nx bits have been reserved in the global bit buffer, the bin 7 is stored in bin buffer 8. If this single bin does already represent a bin sequence that is associated with a codeword, the bin encoder 10 removes this bin from the bin buffer 8 and writes the corresponding codeword 47 to the connected bit buffer 48. Otherwise (this single bin does already represent a bin sequence that is associated with a codeword), further bins 7 are accepted by the particular bin buffer 8 until the bin buffer 8 contains a bin sequence that is associated with a codeword. In this case, the connected bin encoder 10 removes the bin sequence 9 from the bin buffer 8 and writes the corresponding codeword 47 to the connected bit buffer 48.
  • the Nx bits that were first written to the bit buffer 48 are inserted into the previously reserved space in the global bit buffer 51.
  • the same process as specified above is executed; i.e., it is checked first whether a new number of Nx bits must be reserved in the global bit buffer (if Nx - Mx is less than Lx) and then the bin is inserted into the bin buffer 8, etc.
  • the bit writer writes the fixed-length bit sequences of the global bit buffer in the order in which they have been reserved.
  • the bit writer 53 removes the bits for this bit sequence 52 from the global bit buffer 51 and writes the bits 54 to the bitstream. This process is repeated until the first fixed-length entry in the global bit buffer represents a reserved or a free entry. If the first fixed-length entry in the global bit buffer represents a reserved entry, the bit writer 53 waits until this entry is filled with actual bits before it writes further bits 54 to the bitstream 55.
  • the bin buffers are flushed as described above.
  • the bit buffers must be flushed by adding bits with a particular or an arbitrary value until all reserved buffer entries in the global bit buffer are filled and written to the bitstream.
  • Fig. 10 two examples for the possible status of the global bit buffer 51 are illustrated.
  • example (a) a case is illustrated in which different bit buffers/ bin encoders reserve a different number of bits.
  • the global bit buffer contains 3 entries with actually written fixed-length bit sequences and 4 entries with reserved fixed-length bit sequences.
  • the first fixed-length entry already contains actual bits (which must have been just inserted by bit buffer/ bin encoder 2); this entry (i.e., the corresponding 8 bits) can be removed and written to the bitstream.
  • the next entry reserves 10 bits for bin encoder 3, but actual bits haven't been inserted yet. This entry cannot be written to the bitstream; it must be waited until the actual bits are inserted.
  • bit buffers/ bin encoders reserved the same number of bits (8 bits).
  • the global bit buffer contains 4 reservations for 8 bit sequences and 3 actually written 8 bit sequences.
  • the first entry contains a reservation for 8 bits for bin encoder 3. Before any new bits can be written to the bitstream, the bit writer has to wait until bit buffer/ bin encoder 3 writes the actual values of the 8 bits into this reserved entry.
  • Fig. 11 shows an illustration of the basic decoder structure for the embodiment of the invention that interleaves fixed-length sequences of bits.
  • the bin decoders 22 are not connected with a single bit buffer. Instead, each bin decoder 22 is connected with a separate bit buffer 58, which stores bits from the corresponding partial bitstream. All bit buffers 58 are connected to a global bit buffer 61. The bits 62 from the bitstream 63 are inserted into the global bit buffer 61.
  • the bit buffer 58 sends a request 59 to the global bit buffer 61 by which a fixed-length sequence of bits 60 is removed from the global bit buffer 61 and inserted into the particular bit buffer 58.
  • the requests for the fixed-length bit sequences 59 are processed in sequential order.
  • the global bit buffer 61 represents a first-in-first-out buffer; bits that are earlier inserted into the global bit buffer are earlier removed. It should be noted that different bit buffers 58 can request a different amount of bits, which can also vary over time based on already decoded symbols; but the number of bits that are requested by a particular request is known at the time at which the request is sent to the global bit buffer. It should be noted that the global bit buffer 61 is not necessarily required, since the codewords could also be directly read from the bitstream.
  • the global bit buffer 61 is mainly included in the illustration for clearly separate different aspects of the processing chain.
  • bit buffers 58 and the global bit buffer 61 are operated as described in the following.
  • the amount of bits that is requested and read by a particular bit buffer 58 is denoted as Nx, it is equal to the amount of bits that is written to the global bit buffer by the corresponding bit buffer at the encoder side.
  • This number of bits Nx can be different for different bit buffers 58 and it can also vary over time.
  • the number of bits Nx that are requested and read by a particular bit buffer 58 is fixed over time. The reading of a fixed number Nx of bits 60 is triggered based on the number of bits Mx in the bit buffer 58 and the associated maximum codeword length Lx.
  • each bin decoder 22 can be associated with a different maximum codeword length Lx. If a request for a bin 19 is sent to a particular bin buffer 20, and the particular bin buffer 20 is empty, and the number Mx of bits in the bit buffer 58 that is connected (via a bin decoder) with the particular bin buffer 20 is less than the maximum codeword length Lx that is associated with the corresponding bin decoder 22, the connected bit buffer 58 sends a request 59 for a new sequences of N bits to the global bit buffer 61. As response to this request, the first Nx bits are removed from to global bit buffer 61 and this sequence of Nx bits 60 is sent to the bit buffer 58 from which the request was sent.
  • this sequence of Nx bits is added to the corresponding bit buffer 58.
  • the next codeword 57 is read from this bit buffer, and the connected bin decoder 22 inserts the associated bin sequence 21 into the connected bin buffer 20.
  • the first bin is removed from the bin buffer 20 and this decoded bin 25 is sent to the bin buffer selector 18.
  • the next bin request 19 is sent to the particular bin buffer 20 and the bin buffer is not empty, the next bit is removed from the bin buffer 20.
  • the next codeword is read from the bit buffer and a new bin sequence is inserted in the bin buffer, from which the first bit is removed and sent to the bin buffer selector. If the bin buffer is empty and the number Mx of bits in the connected bit buffer 58 is less than the associated maximum codeword length Lx, the next sequence of Nx bits is read from the global bit buffer 61 and inserted into the connected local bit buffer 58, the next codeword is read from the bit buffer, a new bin sequence is inserted in the bin buffer, and the first bin of the sequence is removed and sent to the bin buffer selector. This process is repeated until all source symbols are decoded.
  • bins and/or bits than required for decoding the requested source symbols might be inserted into the bin buffer and/or bit buffer.
  • the remaining bins in the bin buffer and the remaining bits in the bit buffer are discarded and ignored.
  • the described embodiment for an entropy encoder and decoder with interleaving of fixed- length bit sequences can also be combined with the scheme for controlling the encoder buffer delay, which is described above.
  • the basic concept is the same as in the embodiment with delay control described above. If a measure related to the delay or an upper bound of the delay (see below) exceeds a specified threshold, the first reserved buffer entry is filled by flushing the corresponding bin buffer (using a similar mechanism as at the end of a data packet) and potentially writing additional bits for filling all bits of the reserved fixed-length buffer entry. By such a mechanism, the number of waiting buffer entries is reduced until the associated delay measure is less than the specified threshold.
  • the bins and bits that have been inserted at the encoder side in order to obey the delay constraint must be discarded. For this discarding of bins and bits basically the same mechanism as at the encoder side can be used.
  • the measure for the delay is the number of bits in the active buffer entries in the global bit buffer, where the number of active buffer entries is the number of reserved fixed-length buffer entries plus the number of fixed-length buffer entries that contain already written bits.
  • the first buffer entry is always a reserved fixed-length buffer entry or a free buffer entry, since if the first buffer entry contains written bits, these bits are written to the bitstream.
  • the maximum allowed buffer delay (as determined by the application) be D bits. This maximum buffer delay D must be known to both encoder and decoder. In a preferred embodiment of the invention, the maximum buffer delay D is fixed by the application.
  • the maximum buffer delay D is signaled inside the bitstream, e.g., in the header of the data packet (or slice header) or in a parameter set, which is included in the bitstream. It can be signaled in units of bits, or bytes, or a multiple of bits, or a multiple of bytes. If a bin encoder 10 sends a request for the reservation of a new fixed-length bit sequence to the global bit buffer 51, the following process is executed before a new fixed- length buffer entry is reserved.
  • the first buffer entry (which is reserved) is flushed by the process described in the following until the number of bits in the active buffer entries in the global bit buffer plus the number of bits that will be reserved by the current reservation request is less than or equal to the maximum buffer delay D.
  • the flushing of a reserved fixed-length buffer entry is similar to the flushing at the end of a data packet:
  • the bin encoder 10 that is connected with the bit buffer 48 that has reserved the corresponding first buffer entry is flushed by adding bins with particular or arbitrary values to the connected bin buffer 8 until the resulting bin sequence represents a bin sequence that is associated with a codeword, the codeword is then inserted into the corresponding bit buffer 48.
  • one preferred way for adding bins to the bin buffer is to add those bins that produce the shortest possible codeword.
  • the decoder maintains a counter C that counts the bits that have been read from the global bit buffer (this counter can be maintained in the global bit buffer).
  • the counter C is initialized (e.g. with zero) at the beginning of the decoding of a data packet and it is increased after a fixed-length sequence of is read. If a fixed-length sequence of Nx bits is read from the global bit buffer 61, the counter C is increased by Nx.
  • each bit buffer 58 contains a counter Cx, which stores the value of the bit counter C before the last fixed-length bit sequence was read into the corresponding bit buffer 58.
  • the connected bin decoder 22 attempts to read a new codeword from the connected bit buffer 58. If the number of bits in the bit buffer 58 is less than the maximum codeword length, a new fixed-length bit sequence is read from the global bit buffer 61, before the codeword is read, etc.
  • the above embodiments each represent a novel approach to entropy coding that provides the efficiency and simple modeling capability of arithmetic coding at the complexity level of Huffman coding.
  • the event space of the input sequence of discrete symbols is mapped onto a small set of binary probability intervals.
  • the probability models for the source symbols can be fixed or adaptive while entropy coding using the probability intervals remains fixed and is decoupled from the modeling stage.
  • Each of the probability intervals is coded using a very simple entropy code that has the complexity level of Huffman codes.
  • the excess rate of the event space projection entropy code (ESPEC) is similar to that of arithmetic coding.
  • Entropy coding in general, can be considered as the most generic form of lossless data compression.
  • Lossless compression aims to represent discrete data with fewer bits than needed for the original data representation but without any loss of information.
  • Discrete data can be given in the form of text, graphics, images, video, audio, speech, facsimile, medical data, meteorological data, financial data, or any other form of digital data.
  • the original source data are first mapped onto so-called coding symbols and these coding symbols are then entropy coded.
  • the mapping onto coding symbols can include quantization, in which case the overall coding scheme is lossy.
  • the alphabet is associated with an estimated probability mass function (pmf) ⁇ ⁇ ⁇ ) > ⁇ ⁇ ⁇ > ⁇ , ( ⁇ ⁇ - ⁇ ) ⁇ ⁇ 11 ⁇ a U dependencies between coding symbols that are not considered in this pmf are neglected.
  • pmf estimated probability mass function
  • Huffman codes are relatively easy to construct.
  • the most attractive property of Huffman codes is that their implementation can be efficiently realized by the use of variable-length code (VLC) tables.
  • VLC variable-length code
  • the adaptation of the Huffman code and its corresponding VLC tables is quite demanding, both in terms of algorithmic complexity as well as in terms of implementation costs.
  • the redundancy of the corresponding Huffman code (without using any alphabet extension such as run length coding) may be quite substantial.
  • Another shortcoming of Huffman codes is given by the fact that in case of dealing with higher-order probability modeling, multiple sets of VLC tables may be required.
  • Arithmetic coding while being substantially more complex than VLC, offers the advantage of a more consistent and adequate handling when coping with adaptive and higher-order probability modeling as well as with the case of highly skewed probability distributions.
  • this characteristic basically results from the fact that arithmetic coding provides a mechanism, at least conceptually, to map any given value of probability estimate in a more or less direct way to a portion of the resulting codeword. Being provided with such an interface, arithmetic coding allows for a clean separation between the tasks of probability modeling and probability estimation, on the one hand, and the actual entropy coding, i.e., mapping of symbols to codewords, on the other hand.
  • the sequence of symbols ⁇ 3 ⁇ 4,..., 5 ⁇ is converted into a sequence of binary symbols, which are also referred to as bins.
  • the binarization represents a bijective mapping of the alphabet letters a m ' onto ordered sets of bins b ' m .
  • the binarization mapping ⁇ can be different for different symbols s i or symbol categories.
  • Each bin sequence ' for a particular symbol s. consists of one or more bins 3 ⁇ 4 .
  • a sequence of bins ⁇ 3 ⁇ 4,... is obtained that represents the sequence of source symbols ⁇ s 0 ,...,s N _ 1 ⁇ .
  • a binary pmf ⁇ p 0 J ,p( ⁇ can be described by the less probable bin (LPB) value b J LPB and its probability p J LPB (with P J LPB ⁇ 0.5 ).
  • This binary probability description ⁇ b j LPB ,p J LPB ⁇ can be directly derived from the probability estimates p s (a m ' ) for the symbol alphabets given the binarization mappings ⁇ [ .
  • the bins can be associated with a probability model (which is also referred to as context) based on the syntax and previously coded symbols or bins. And for each probability model, the probability description ⁇ b J LPB ,p j LPB ⁇ can be estimated based on the values of the bins that are coded with the probability model.
  • a probability model which is also referred to as context
  • the probability description ⁇ b J LPB ,p j LPB ⁇ can be estimated based on the values of the bins that are coded with the probability model.
  • the sequence of coding bins ⁇ 3 ⁇ 4,... , 3 ⁇ 4_, ⁇ does uniquely represent the sequence of source symbols ⁇ s Q , ... , s N ⁇ ⁇ and the corresponding probability estimates, which can be employed for entropy coding, are completely described by the LPB probabilities p J LPB (with p j LPB ⁇ 0.5 ).
  • the sequence of coding bins ⁇ 3 ⁇ 4,... is projected onto a small number of probability intervals I k .
  • a simple non-adaptive binary entropy coder is designed for each interval I k . All coding bin b j with associated LPB probabilities p i LPB e I k are assigned to the interval I k and are coded with the corresponding fixed entropy coder.
  • all bins represent coding bins tf and all probabilities p are LPB probabilities p j LPB .
  • each probability interval I k (p k ,p k+l ] is associated with a representative probability t e I k and the corresponding optimal entropy coder shall achieve the entropy limit for this representative probability.
  • the rate for coding a bin with probability p using the optimal entropy coder for the interval representative P j is given by
  • the value p) given in eq. BIO is the representative probability for an interval I k that minimizes the expected overall rate R given the interval boundaries p k and p k+l . Otherwise, no bin is projected to the interval I k and the representative probability p t e l k can be arbitrarily chosen without any impact on the overall rate R ; but such a configuration should be avoided, since the interval I k would not be employed for entropy coding.
  • Figure 12 shows an example for the optimal interval discretization using the described algorithm.
  • the expected overall rate increase relative to the entropy limit can be used.
  • the expectation value of the entropy H— J o H(p) f (p) dp is equal to 1/(2 In 2) bit per bin and the rate overhead p is equal to 1.01%.
  • Such a bijective mapping can be represented by a binary tree where all leaf nodes are associated with codewords, as depicted in Figure 13.
  • the tree edges represent binary events. In the example of Figure 13, the lower edges represent the LPB bin value and the upper edges represent the MPB bin value.
  • the binary tree represents a prefix code for the bins if it is a full binary tree, i.e., if every node is either a leaf or has two descendants.
  • Each leaf node is associated with a probability based on the given LPB probability p .
  • Each leaf node L is characterized by the number of LPB edges a, and the number MPB edges b l from the root node to the leaf node. For a particular LPB probability p , the probability p l for a leaf node is equal to
  • VNB2VLC variable number of bits to variable length codewords
  • the code design is often limited by factors as the maximum number of codewords L , the maximum number of bins per codeword, or the maximum codeword length, or it is restricted to codes of particular structures (e.g., for allowing optimized parsing). If we assume that the set S c of usuable codes for a particular application is given, the optimum code C * e S c for a particular LPB probability p can be found by minimizing the expected rate R(p,C)
  • the minimization can also proceed over a given set of binary trees S T and for each tree only one VNB2VLC code C that is obtained by the Huffmann algorithm is considered.
  • VNB2VLC codes for various LPB probabilities p by considering all binary trees T for which the number of leaf nodes L is less than or equal to a given maximum L m .
  • the rate increase p(p) can usually be reduced by allowing larger table sizes.
  • a small table size L of 8 to 16 codewords is usually sufficient for keeping the rate increase p(p) reasonably small, but for smaller LPB probabilities (e.g., p ⁇ 0.1 ), larger table sizes L are required.
  • an optimal interval border p k ' between two intervals l k _ x and I k with given associated codes C k _ x and C k , respectively, is the intersection point of the functions R p,C k _ l ) and R p,C k ) .
  • the minimization in step 3 can also proceed over a given set of binary trees S T where for each binary tree T only one VNB2VLC code C k obtained by the Huffmann algorithm is considered.
  • Codeword termination can be done as follows.
  • the arithmetic codeword for the sequence of coding bins has to be terminated in a way, that all coding bins can be decoded given the codeword.
  • the bins at the end of the sequence b c k may not represent a bin sequence that is associated with a codeword. In such a case, any codeword that contains the remaining bin sequence as prefix can be written. The overhead can be minimized, if the corresponding codeword that has the minimum length (or one of these codewords) is chosen.
  • the additionally read bins at the end of the bin sequence which can be identified given the bitstream syntax and binarization schemes, are discarded.
  • a binarization for the full binary tree in Fig. 15 is given in Tab. 5 .
  • the bin b 0 is present.
  • b 0 is equal to 0
  • b x is present. Note that the binarization given in Tab. 2 is identical to an optimal single-letter Huffman code for the source s .
  • Table 5 Binarization of a three letter source.
  • the LPB probabilities p LPB are 0.3 for the first bin and 0.4 for the second bin
  • the entropy for the source s is
  • the bins b 0 and b x already represent coding bins, since for both bins the LPB value b J LPB is equal to 0.
  • he sequence of source symbols is binarized into a sequence of bins.
  • the bin b 0 is transmitted for every source symbol.
  • the bin b x is only transmitted when b 0 - 0 .
  • the redundancy is 0.0053 bit/bin or 0.548% .
  • the overall redundancy is 0.73% relative to the entropy limit, which represents a significant improvement in comparison to the single-letter Huffman code.
  • VNB2VLC codes can be considered as a generalization of run-length codes for binary symbols (the VNB2VLC code in Tab. 3 does effectively represent a run-length code).
  • a similar coding efficiency as for the presented approach can also be achieved by creating a code that maps a variable number of source symbols to variable length codewords.
  • the main advantage of the presented approach is its flexibility in mapping arbitrary source symbol sequences with fixed or adapative probability estimates to a small number of simple binary coders that are operated with fixed LPB probabilities.
  • the encoder mapping y e For unique decodability of a sequence of coding bins b c given the sequence of codewords c(b c ), the encoder mapping y e must have the property that a unique codeword c(b c ) is assigned to each possible sequence of coding bins b c :
  • a sequence of coding bins b 0 is uniquely decodable given K sequences of codewords c ⁇ b ⁇ ), with k - 0,...,K- ⁇ , if each sub-sequence of coding bins b ⁇ is uniquely decodable given the corresponding codeword Ck(6 k °) and the partitioning rule ⁇ is known to the decoder.
  • Each source symbol s shall be associated with a symbol category c s , which determines the type of the symbol including its range of values.
  • FIG. 17 The block diagram for an example encoder and decoder design is illustrated in Figure 17.
  • the used binarization scheme is determined based on the symbol category c s .
  • the binarizer associates each bin b of a bin sequence s with a probability model indication c b , which specifies the probability model that is used for coding the bin b .
  • the probability model indication c b can be derived based on the symbol category c s , the bin number of the current bin inside the bin sequence s , and/or the values of already coded bins and symbols.
  • the probability estimator and assigner maintains multiple probability models, which are characterized by pairs of values ⁇ b LPB , p LPB ⁇ . It received bins b and associated probability model indications c b from the binarizer, and forwards the LPB value b LPB and the LPB probability p LPB of the indicated probability model to the coding bin deriver and the probability quantizer, respectively. Thereafter, the corresponding probability model ⁇ b LPB , p LPB ⁇ is updated using the value of the received bin b .
  • the probability quantizer forwards each coding bin b° to one of the K binary encoders. It contains information about the LPB probability interval quantization ⁇ I k ⁇ .
  • the LPB probability p LPB which is associated with a coding bin b c and received from the probability estimator and assigner, is compared to the interval borders ⁇ p k ⁇ and probability interval index k , for which p LPB e l k , is derived.
  • the coding bin b c is forwarded to the associated binary encoder.
  • Each of the K binary encoders consists of a bin buffer and a bin encoder.
  • the bin buffer receives coding bins b° from the probability quantizer and stores them in coding order.
  • the bin encoder implements a particular VNB2VLC mapping and compares the bin sequence in the bin buffer with the bin sequences that are associated with codewords. If the bin sequence in the bin buffer is equal to one of those bin sequences, the bin encoder removes the bin sequence ⁇ b c ⁇ from the bin buffer and writes the associated codeword ( ⁇ b a ⁇ ) to the corresponding codeword stream.
  • a terminating codeword is written as described in sec. 4.3.
  • the K resulting codeword streams can be separately transmitted, packetized, or stored, or they can be interleaved (cp. sec. 6.2) for the purpose of transmission or storage.
  • each of the K binary decoders consisting of a bin decoder and a bin buffer receives one codeword stream.
  • the bin decoder reads codewords ( ⁇ b c ⁇ ) from the codeword stream and inserts the associated bin sequence ⁇ b c ⁇ , in coding order, into the bin buffer.
  • the decoding of the symbol sequence is driven by the underlying syntax.
  • Requests for a symbol s are sent together with the symbol category c s to the binarizer.
  • the binarizer converts these symbol requests into request for bins.
  • a request for a bin is associated with a probability model indication c b , which is derived in the same way as in the encoder, and sent to the probability estimator and assigner.
  • the probability estimator and assigner is operated similar to its counterpart at the encoder side. Based on the probability model indication c b , it identifies a probability model and forwards its LPB value b LPB and LPB probability p LPB to the bin deriver and the probability quantizer, respectively.
  • the probability quantizer determines one of the K binary decoders based on the LPB probability p LPB , in the same way as the binary encoder is determined at the encoder side, removes the first coding bin b° , in coding order, from the corresponding bin buffer, and forwards it to the bin deriver.
  • the bin deriver send the decoded bin value b to the binarizer and the probability estimator and assigner.
  • the value of the decoded bin b is used to update the probability model , ⁇ b LPB , p LPB ⁇ , which was chosen by the associated value c b , in the same way as at the encoder side.
  • the binarizer adds the received bin b to the bin sequence s which has been already received for a symbol request and compares this bin sequence 4 with the bin sequences that are associated with symbol values by the binarization scheme . If the bin sequence s matches one of those bin sequences, the corresponding decoded symbol s is output as final response to the symbol request. Otherwise, the binarizer sends further bin requests until the symbol s is decoded.
  • the decoding of a symbol sequence is terminated if no further symbol requests, which are driven by the syntax, are received.
  • the coding bins b° that may be contained in the bin buffers at the end of the entropy decoding process (as a result of termination codewords) are discarded.
  • the mapping of a sequence of symbols to a bit stream and the inverse mapping are described.
  • Each symbol carries associated parameter(s) with it that are simultaneously known at encoder and decoder.
  • the entropy codec contains multiple first-in first-out (FIFO) buffers with each of them assigned to subsets of parameter(s) that are associated to the symbols. For given parameter(s) of a symbol, the encoder assigns the symbol to the corresponding FIFO buffer. Since the encoder assignment rule is known at the decoder side, the decoder reads from the FIFO buffer to which the encoder has assigned the symbol.
  • One example for a parameter assignment to a symbol is an estimate of its probability.
  • the probability estimate may be given or derived from a measurement that can be conducted simultaneously at encoder and decoder.
  • a particular FIFO buffer contains symbols with estimated probability values falling into a subset of probabilities that is chosen, so that the entropy coding can be improved.
  • the improvement may relate to bit rate reduction, complexity reduction, delay reduction or error resilience.
  • Another example of a parameter assignment to symbols could be an estimate of the symbol probability and another parameter including one or more of the following: semantic of the symbol, its importance, its location (temporal, spatial, volumetric, etc), its channel code protection.
  • a particular FIFO buffer contains symbols related to a subset of probabilities that is chosen, so that the entropy coding can be improved and a benefit can be achieved with regard to the other parameters.
  • the first embodiment is concerned with Coding Transform Coefficients in Picture and/or Video Coders and Decoders and describes a coding of transform coefficients which can particularly be employed as a novel efficient method for coding transform coefficients in the field of video coding, replacing the corresponding part of CABAC in H.264/AVC, for example.
  • the blocks of quantized transform coefficients (levels) are mapped by a defined scan process to a vector which is coded to code words having a variable length by using a run-length coding and subsequent mapping.
  • the code words having a variable length are associated to two-dimensional events (RUN, LEVEL), wherein LEVEL represents the quantized value of a (significant) transform coefficient not quantized to zero; the run-length RUN indicates the number of subsequent (non-significant) transform coefficients quantized to zero, which, in the vector of transform coefficients, are immediately in front of the present significant transform coefficients.
  • code words having a variable length are defined for the two special events EOB and ESCAPE. While the EOB event indicates that there are no further significant transform coefficients in the block, the ESCAPE event signals that the existing event (RUN, LEVEL) cannot be represented by the defined alphabet of code words having a variable length. In this case, the symbols RUN and LEVEL are coded by code words of a fixed length.
  • Annex E of the H.263 standard specifies an optional non-adaptive arithmetic coding in which different predetermined model probability distributions are used,
  • a context-adaptive method on the basis of code words having a variable length for coding transform coefficients is specified as the standard method for entropy coding.
  • the coding of a block of transform coefficients is determined by the following characteristics:
  • Both the number of the significant coefficients within a block and the number of subsequent coefficients quantized to one at the end of the vector of transform coefficients are determined by a symbol COEFF_TOKEN.
  • COEFF_TOKEN Depending on the block type and the symbols COEFF_TOKEN already coded/decoded for neighboring blocks, one of five defined code word tables is chosen for coding.
  • a symbol TOTAL ZEROS will be coded, which indicates the number of transform coefficients quantized to zero which, in the coefficient vector, are in front of the last significant coefficient. For this, 18 code word tables have been specified, which are switched depending on the number of significant coefficients and the block type.
  • the run-length of the (non-significant) coefficients quantized to zero (RUN) in front of a significant coefficient is coded for each significant transform coefficient in a reverse scan order as long as the sum of RUNs already coded is smaller than TOTAL_ZEROS. Depending on TOTAL_ZEROS and the RUNs already coded/decoded, switching between seven code word tables takes place.
  • CAVLC context-adaptive variable length coding
  • a possible solution for avoiding the disadvantages illustrated of well-known methods for coding transform coefficients in block-based picture and video coders is a combination of an adaptive entropy coding and a suitable context generation for using the inter-symbol redundancies.
  • the application example described next represents scheme for coding transform coefficients in picture and/or video coders and decoders, which eliminates the deficiencies mentioned above and, in particular, keep the amount of calculating required for coding small.
  • this coding By combining this coding with the entropy coding schemes described in the embodiments with respect to Fig. 1 to 17, a very effective coding efficiency results.
  • the coding scheme described below provides a method for coding transform coefficients in picture and/or video coders and decoders, wherein for blocks of (video) pictures containing significant transform coefficients, the coding of transform coefficients takes place in such a way that, for each block,
  • each significant transform coefficient of the block other than the last transform coefficient of the block being characterized by a one-bit symbol.
  • blocks containing significant transform coefficients are characterized by a one-bit symbol CBP4 in connection with other syntax elements, such as, for example, CBP or macro block mode.
  • modeling for the one-bit symbol CBP4, for coding the significance mapping and/or for coding the coefficient magnitudes takes place in a context-dependent way.
  • block types of transform coefficients having comparable statistics are summarized to block categories.
  • no significance information (SIG, LAST) is transferred for the last scan position of a block.
  • the magnitude (ABS) is indicated by a symbol in unary binarization or by a symbol having a prefix part and a suffix part, wherein the prefix part consists of ones and the suffix part is coded in a 0th order exp-golomb code.
  • the current embodiment provides an arrangement having at least one processor and/or chip formed such that method for coding transform coefficients in picture and/or video coders and decoders can be performed, wherein for blocks of (video) pictures containing significant transform coefficients, the coding of transform coefficients takes place in the just-mentioned way.
  • the current application example provides a computer program enabling a computer, after having been loaded into the memory of the computer, to perform a method for coding transform coefficients in picture and/or video coders and decoders, wherein for blocks of (video) pictures containing significant transform coefficients, the coding of transform coefficients takes place in the way described above.
  • Such computer programs can, for example, (for a fee or free of charge, freely accessible or protected by a password) be provided in a data or communication network for downloading.
  • the computer programs provided in this way can then be utilized by a method in which a computer program according to claim 11 is downloaded from a network for transfer, such as, for example, from the Internet, to data processing means connected to the network.
  • the current embodiment provides a computer-readable storage medium on which a program is stored, enabling a computer, after having been loaded into the memory of the computer, to perform a method for coding transform coefficients in picture and/or video coders and decoders, wherein for blocks of (video) pictures containing significant transform coefficients, the coding of transform coefficients takes place in the described way.
  • a two-dimensional block of transform coefficients is mapped to a one-dimensional vector by a scan process.
  • the levels are coded in a reverse scan order.
  • Figure 18 illustrates the coding method.
  • CBP4 CBP4 or macro block mode
  • the CBP4 symbol will be zero if there are no significant coefficients in the block. If it is one, a significance mapping specifying the position (in scan order) of the significant transform coefficients will be coded. Subsequently, the magnitude and the signs of the significant coefficients are transferred in a reverse scan order.
  • a detailed description of the coding process will be given below in 1.
  • context modeling for the binary entropy coding will be described in 2, with the entropy coding of the bins being performed with any of the above-outlined embodiments for entropy coding. 1. Description of the coding of the transform coefficients
  • the transform coefficients of each block are mapped to a vector by means of a scan process (such as, for example, a zig zag scan).
  • CBP4 is a one-bit symbol indicating whether there are significant transform coefficients (transform coefficients unequal to zero) in a block. If the CBP4 symbol is zero, no further information for the corresponding block will be transferred.
  • a significance mapping will be coded. This takes place by transferring a one-bit symbol (SIG) for each coefficient in the scan order. If a corresponding significance symbol is one (significant coefficient), another one-bit symbol (LAST) will be transferred. This symbol indicates whether the present significant coefficient is the last significant coefficient within a block or whether further significant coefficients follow.
  • Figure 19 shows two examples of the method described for coding the significance mapping. Significance information (SIG, LAST) will never be transferred for the last scan position of a block. If the transfer of the significance mapping has not already been terminated by a LAST symbol of one, it is obvious that the coefficient at the last scan position is significant (see yellow marked position in Fig. 19).
  • the positions of the significant transform coefficients within a block are clearly specified by the significance mapping.
  • the coding of the precise values of the coefficients (levels) takes place by two coding symbols: ABS (magnitude of the coefficients) and SIGN (sign of the coefficients). While SIGN represents a one-bit symbol, binarization according to Figure 20 is used for coding the magnitudes of the coefficients (ABS). For coefficient magnitudes in the interval [1; 14], this binarization corresponds to a unary binarization.
  • the binarization for coefficient magnitudes larger than 14 consists of a prefix part consisting of 14 ones and a suffix part representing a 0th order exp-golomb code for the symbol (ABS- 15).
  • Binarization does not include a representation for coefficient magnitudes (ABS) equaling 0, since significant coefficients (coefficients unequal to zero) will always have a magnitude (ABS) larger than or equal to one.
  • ABS coefficient magnitudes
  • the binarization formed of a prefix part and a suffix part consisting of a 0th order exp-golomb code for coefficient magnitudes larger than 14 has the advantage that a special non-adaptive context with symbol probabilities 0.5 can be used without sacrificing in coding efficiency for all the binary decisions of the suffix part, whereby the amount of calculating for encoding and decoding can be reduced.
  • the levels are coded in a reverse scan order - beginning with the last significant coefficient within the block; this enables forming suitable contexts for the binary entropycoding.
  • CBP4 For coding the one-bit symbol CBP4, four different contexts are used for each individual category of transform blocks (see Figure 21).
  • the following 6 block types are differentiated for this conditioning: Luma-DC, Luma-AC, Chroma-U-DC, Chroma-U-AC, Chroma-V-DC and Chroma-V-AC.
  • CBP4 (X) is set to zero for the neighboring block X. If a neighboring block X (A or B) is outside the picture area or belongs to another slice, the corresponding value CBP4 (X) is replaced by a default value. A default value of one is used for INTRA-coded blocks and a default value of zero is used for INTER-coded blocks.
  • max_koeff-l different contexts are each used per block category (see Figure 21) for coding the symbols SIG and LAST.
  • max_koeff thus designates the number of transform coefficients for the corresponding block category (for H.264/AVC, see Figure 21).
  • the context number is always indicated by the corresponding scan position of the coefficient considered.
  • 2*max_koeff-2 contexts are used for each category of block types for coding the significance mapping.
  • the binarization illustrated in Figure 20 is used for coding the magnitudes of the significant transform coefficients.
  • the context numbers are thus associated as follows: ctx__number_abs_ 1 bin
  • the magnitudes of the transform coefficients are transferred in a reverse scan order.
  • the application embodiment described with respect to Fig. 18 to 23 related to coding transform coefficients in picture and/or video coders and decoders wherein for blocks of (video) pictures containing significant transform coefficients, the coding of transform coefficients takes place in such a way that, for each block, in a scan process, the positions of significant transform coefficients in the block and subsequently, in a reverse scan order - starting with the last significant transform coefficients within the block - the values (levels) of the significant transform coefficients are determined and coded.
  • Each significant transform coefficient of the block other than the last transform coefficient of the block may be characterized by a one-bit symbol.
  • the sign may be indicated by a one-bit symbol (SIGN) and the magnitude is indicated by a binary- coded symbol (ABS).
  • the magnitude may be indicated by a symbol (ABS) in unary binarization or by a symbol (ABS) having a prefix part and a suffix part, wherein the prefix part consists of ones and the suffix part is coded in a 0th order exp-golomb code.
  • Blocks containing significant transform coefficients may be characterized by a one-bit symbol CBP4 in connection with further syntax elements, such as, for example, CBP or macro block mode.
  • a significance mapping may be coded, wherein the transfer takes place in a scan order, (SIG) serves for identifying significant coefficients and (LAST) indicates whether there are further significant transform coefficients in the block.
  • Modeling, for the one-bit symbol CBP4, for coding the significance mapping and/or for coding the coefficient magnitudes may take place in a context-dependent way. No significance information (SIG, LAST) might be transferred for the last scan position of a block.
  • Block types of transform coefficients having comparable statistics may be summarized to block categories.
  • the coding of transform coefficient levels may be split into three steps.
  • a binary syntax element coded_block_flag is transmitted for each transform block, which signals whether the transform block contains significant transform coefficient levels (i.e., transform coefficients that are non-zero). If this syntax element indicates that significant transform coefficient levels are present, a binary- alued significance map is coded, which specifies which of the transform coefficient levels have non-zero values. And then, in a reverse scan order, the values of the non-zero transform coefficient levels are coded.
  • the significance map is coded as follows. For each coefficient in the scan order, a binary syntax element significant_coeff_flag is coded, which specifies whether the corresponding transform coefficient level is not equal to zero.
  • a further binary syntax element last_significant_coeff_fiag is coded. This bin indicates if the current significant transform coefficient level is the last significant transform coefficient level inside the block or if further significant transform coefficient levels follow in scanning order. If last_significant_coeff_flag indicates that no further significant transform coefficients follow, no further syntax elements are coded for specifying the significance map for the block. In the next step, the values of the significant transform coefficient levels are coded, whose locations inside the block are already determined by the significance map. The values of significant transform coefficient levels are coded in reverse scanning order by using the following three syntax elements.
  • the binary syntax element coeff_abs_greater_one indicates if the absolute value of the significant transform coefficient level is greater than one. If the binary syntax element coeff_abs_greater_one indicates that the absolute value is greater than one, a further syntax element coeff_abs_level_minus_one is sent, which specifies the absolute value of the transform coefficient level minus one. Finally, the binary syntax element coeff_sign_flag, which specifies the sign of the transform coefficient value, is coded for each significant transform coefficient level. It should be noted again that the syntax elements that are related to the significance map are coded in scanning order, whereas the syntax elements that are related to the actual values of the transform coefficients levels are coded in reverse scanning order allowing the usage of more suitable context models.
  • a measure related to the probability for one of the two possible bin values is estimated for each context based on the values of the bins that have been already coded with the corresponding context.
  • the context that is used for coding is selected based on already transmitted syntax elements or based on the position inside a block.
  • the significance map specifies information about the significance (transform coefficient level is different from zero) for the scan positions.
  • CABAC entropy coding for a block size of 4x4, a separate context is used for each scan position for coding the binary syntax elements significant_coeff_flag and the last_significant_coeff_flag, where different contexts are used for the significant_coeff_flag and the last_significant_coeff_flag of a scan position.
  • the same context model is used for four successive scan positions, resulting in 16 context models for the significant_coeff_flag and additional 16 context models for the last_significant_coeff_flag.
  • the block is processed in reverse scan order. If a scan position is significant, i.e., the coefficient is different from zero, the binary syntax element coeff_abs_greater_one is transmitted. Initially, the second context model of the corresponding context model set is selected for the coeff_abs_greater_one syntax element. If the coded value of any coeff_abs_greater_one syntax element inside the block is equal to one (i.e., the absolute coefficient is greater than 2), the context modelling switches back to the first context model of the set and uses this context model up to the end of the block.
  • the context model is chosen depending on the number of the coeff_abs_greater_one syntax elements equal to zero that have already been coded/decoded in the reverse scan of the considered block.
  • the context model selection for the syntax element coeff_abs_greater_one can be summarized by the following equation, where the current context model index C t +i is selected based on the previous context model index C t and the value of the previously coded syntax element coeff_abs_greater_one, which is represented by bin t in the equation.
  • the second syntax element for coding the absolute transform coefficient levels, coeff_abs_level_minus_one is only coded, when the coeff_abs_greater__one syntax element for the same scan position is equal to one.
  • the non-binary syntax element coeff_abs_level_minus_one is binarized into a sequence of bins and for the first bin of this binarization; a context model index is selected as described in the following.
  • the remaining bins of the binarization are coded with fixed contexts.
  • the context for the first bin of the binarization is selected as follows.
  • the context modelling switches to the next context model in the set, where the number of context models in set is limited to 5.
  • the context model selection can be expressed by the following formula, where the current context model index C t+ i is selected based on the previous context model index C t .
  • the probability assigned to the individual contexts would, for example, be used as, or at least one of the parameters used for selecting the entropy encoder( entropy decoder among the plurality of partial bitstream paths, such as, described above, after having quantized the probability to a smaller set of probability partial bitstreams.
  • the reconstruct of the sequence of symbols sequentially defining the significance mapping and the subsequent coding of the values of transform coefficients being unequal to zero in reverse scan order may be performed using a decoder according to any of the previously described embodiments.
  • an improved entropy encoding design with higher compression efficiency may be achieved.
  • An apparatus for entropy-encoding a symbol sequence of information symbols to obtain entropy-encoded information symbols, the symbol sequence having a start symbol comprises: an entropy encoder for entropy encoding a symbol of the symbol sequence based on probability information for the symbol, the symbol being part of a symbol set, to produce the entropy-encoded information symbols; a context modeler for deriving the probability information for the symbol based on a context of the symbol, the context including one or more context symbols processed by the entropy encoder prior to processing the symbol, the context modeler including an initializer for initializing the context modeler by determining and providing initialization probability information to the entropy encoder, the initialization probability information to be used by the entropy encoder for processing the start symbol, wherein the initializer is operative to determine the initialization probability information based on an estimation of symbol statistics relating to the start symbol such that an initialization probability distribution is different from an equi-probable distribution for all symbols of the symbol set.
  • a decoder for entropy decoding entropy-encoded information symbols the entropy-encoded information symbols being produced by entropyally encoding a symbol of the symbol sequence based on probability information for the symbol, the symbol being part of a symbol set, wherein the probability information for the symbol is derived based on a context of the symbol, the context including one or more context symbols processed earlier, and wherein, for entropyally encoding the start symbol, an initialization probability information was used, the initialization probability information being based on an estimation of a symbol statistics relating to a start symbol and being determined such that an initialization probability distribution is different from an equi-probable distribution for all symbols of the symbol set, comprises: an entropy decoder for entropy decoding the entropy-encoded information symbols to obtain the symbol sequence of information symbols having the start symbol; and a context modeler for obtaining the probability information used when entropy encoding the sequence of information symbols, context modeler including an initializer
  • An improved compression efficiency can be obtained by initializing the context modeler so that the initialization probability information is different from an equi-probable distribution for all symbols of the symbol set in question.
  • the initializer is operative to determine the initialization probability information to be used at least for the start symbol of the symbol sequence based on an estimation of a symbol statistics relating to the start symbol so that the initialization probability information is different from an equi-probable distribution for all symbols of the symbol set.
  • this "intelligent" initialization is advantageous for coding of video material with widely varying content from different video sources. It has been found out that, in a preferred embodiment of the present invention, the initializing probability information can be derived based on the quantization parameter used for quantizing the to be encoded information. In video coding applications, and, in particular, in video encoding applications in connection with the H.264/AVC video compression standard, the symbol sequence of information symbols to be entropy-encoded is derived from transformed residual values, which are obtained by a time/spatial inter/intra frame prediction mode for example. Additionally, the intelligent initialization can advantageously be based on prior knowledge on the video source material.
  • the compression efficiency can be enhanced by defining several sets of initialization values.
  • a single set of initialization values is sufficient.
  • more than one and, preferably, three different sets of initialization values may be defined.
  • the selected set of initialization values is to be specified by a certain syntax element in the encoder output bit stream so that a decoder processing an encoded output bit stream is able to correctly initialize itself.
  • this syntax element is transmitted as part of the slice header.
  • the encoder has the freedom to select the set of initialization values that is the most suitable one for the symbol statistics for each slice.
  • the initialization values are further diversified by detecting, to which data type or sub-data type, the symbol sequence to be processed by the entropy encoder belongs to.
  • This certain type is indicated by a context index variable.
  • This context index variable is preferably designed so that it can signal one out of up to 399 different context models from which a corresponding number of initialization probability information can be derived so that an optimally adapted initialization for a large number of different symbol sequences, each of which having a start symbol, can be determined.
  • the symbols to be entropy coded have attributed thereto different context models.
  • several symbols or bits are to be encoded using one context model out of the plurality of context models. These several symbols or bits form the symbol sequence, wherein the first to be encoded symbol of the symbol sequence, the symbol sequence having attributed thereto a context model, is the start symbol.
  • a context model can include one or even more than one probability models. In case of a context model having more than one probability models, for each change of probability models, a new initialization will take place.
  • Fig. 24 shows an apparatus for entropy-encoding symbol sequences of information symbols to obtain entropy-encoded information symbols, wherein the symbol sequence has a start symbol.
  • the symbol sequence is input into a symbol sequence input of an entropy encoder 1112.
  • the entropy-encoded information symbols are output at an entropy-encoder output 1114, which is also connected to the entropy encoder 1112.
  • the entropy encoder is operative for entropy encoding a symbol of the symbol sequence based on probability information for the symbol.
  • the probability information is input into the entropy encoder 1112 via a probability information input 16. Since the symbol is part of a specified symbol set, which, in the binary case, only includes two different symbols, i.e., a binary "0" and the binary "1", it is sufficient to provide probability information only for one symbol of the symbol set. In accordance with the present invention, it is preferred to provide the probability information of the least probable symbol (LPS), since this number is naturally smaller than the probability for the most probable symbol. Additionally, the probability information provided to the entropy encoder via the probability information input 1116 includes the value of the most probable symbol, from which the value of the least probable symbol and vice versa can be easily derived.
  • LPS least probable symbol
  • the encoder further includes the context modeler 1118 for deriving the probability information for the symbol based on a context of the symbol, the context including one or more context symbols input into the context modeler by a context input 1120, which have been processed by the entropy encoder 1112 prior to processing the actual symbol.
  • the context modeler 1118 includes an initializer 1120 for initializing the context modeler by determining and providing initialization probability information to the entropy encoder 1112, when the initialization probability information is to be used by the entropy encoder 1112 for processing the start symbol, i.e., the symbol of the symbol sequence, for which no context symbols exist, i.e. symbols which have been processed by the entropy encoder 1112 prior to the start symbol.
  • the start symbol is the first symbol of the symbol sequence.
  • the initializer is operative to determine the initialization probability information based on an estimation of symbol statistics relating to the start symbol such that the initialization probability is different from an equi-probable distribution for all symbols of the symbol set, to which the start symbol belongs to.
  • the entropy-encoding apparatus Since the context modeler, and, therefore, the entropy encoder is initialized in an intelligent way based on an estimation of the symbol statistics that are expected for the symbol sequence to be encoded by the entropy encoder, the entropy-encoding apparatus shows an improved start-up performance. Stated in other words, the entropy-encoding apparatus behaves, for the start symbol, as if the start symbol is not the first symbol of a sequence but is an intermediate symbol of a sequence. Therefore, the initialization results in a pre-adaptation of the entropy- encoding apparatus already for the start symbol. Contrary to the straightforward approach, in which the initialization probability information is set to an equi-probable distribution, the entropy encoder is immediately context-pre-adapted.
  • the context- adaptation would need several steps corresponding to the number of prior encoded symbols forming the context symbols. It has been found out that the inventive initialization of the context modeler and the entropy encoder results, without any further actions, in bit-rate savings of up to 3 % simply by providing an intelligent initialization probability information compared to a straightforward equi-probable distribution.
  • Fig. 25 shows a preferred embodiment of the initializer 1020 from Fig. 24.
  • the initializer 1120 includes a retriever for retrieving a quantization parameter for the symbol sequence from a slice header. Additionally, the slice header is also input into a detector for detecting a data type of the slice data. Additionally, a detector 1204 for detecting a sub-data type in the slice data is provided. The detector 1204 is further operative to access different initialization tables based on the detected sub-data type. As it will be described later on, the access to different initialization tables is controlled by means of a context index on the one hand and, preferably, by means of an initialization index on the other hand.
  • initialization table with certain context indices for I- and SI- slices.
  • P- and B-slices there exist at least two and preferably three initialization tables which are, once again, addressed by the context index, wherein the table of the at least two tables which is actually used is addressed by the initialization index.
  • a calculator 1206 Connected to the retriever 1200 on the one hand and the detector 1204 on the other hand, is a calculator 1206, which is operative to calculate the information on the initialization probability which can be the actual initialization probability or, in accordance with the preferred embodiment, a probability state index, i.e., a reference to an initialization probability information table including probability information for the least probable symbol. It has been found out by the inventors that a linear dependence of the initialization probability from the quantization parameter used for quantizing the slice data results in an improved compression efficiency, i.e., in an near to optimum initialization. Therefore, the calculator 1206 in Fig.
  • Fig. 25 is operative to apply a linear calculation, wherein the parameter m indicates a gradient of the linear dependence, while the other initialization parameter, i.e., n indicates a y- axis offset.
  • the basic equation processed by the calculator 1206 in Fig. 25 is indicated in block 206 as m x QP + n, wherein QP is the quantization parameter while m and n are the initialization parameters taken out from different initialization tables addressed by a certain context index.
  • the calculator 1206 in Fig. 25 performs the sequence of steps as indicated in Fig. 26.
  • a first step the equation m x sliceQP + n is calculated.
  • the value F of m x sliceQP is shifted to the right by four digits. It is to be noted here that, naturally, all values are binary strings. The shift by four digits relates to a representation of the product m x sliceQP in a two-complementary presentation. Then, the result after the shift is added to the other initialization parameter n. Then, the minimum of this result on one hand and the integer value 1126 on the other hand is selected.
  • this minimum is larger than 1 , this minimum is selected and assigned to the auxiliary variable preCtxState. Then, a determination is made, if this auxiliary variable is lower than or equal to 63. In case this determination results in a yes answer, the probability state index variable pStateldx is calculated as the difference between 63 and the auxiliary variable. Additionally, it is determined that the value of the most probable symbol (valMPS) is equal to zero. In the other case, i.e., in which the auxiliary variable is greater than 63, the probability state index pStateldx is calculated as the difference between the auxiliary variable on the one hand and the value of 64. Additionally, the value of the most probable symbol is set to one.
  • the probability state index is a reference to an initialization probability information table including probability information for the least probable symbol.
  • the value of the probability state index can address 64 different probability information values in the probability information table. The reason for firstly calculating a probability state index and then addressing a table using the probability state index is in an improved design of the actual entropy encoder core. Naturally, by an appropriate mapping, the sequence of the steps shown in Fig. 26 could also directly calculate the initialization probability information.
  • the quantization parameter SliceQP is used for determining the quantization of transform coefficients.
  • the parameter may take 52 values. Theses values may be arranged so that an increase of 1 in quantization parameter means an increase of quantization step size by approximately 12 % (an increase of 6 means an increase of quantization step size by exactly a factor of 2). It can be noticed that a change of step size by approximately 12 % also means roughly a reduction of bit rate by approximately 12 %.
  • Fig. 27 showing an inventive entropy decoder.
  • the entropy decoder needs the same probability distribution used by the encoder for a respective encoding step to resolve the transmitted interval (which is naturally represented in binary form) into the decoded bits, i.e., the information symbol sequence having the start symbol.
  • the inventive entropy decoder includes an entropy decoder 1160 and a context modeler 1162, which also includes an initializer 1164. It is to be seen here that the decoder shown in Fig. 27 is analogously constructed as the encoder in Fig. 24.
  • the context modeler 1162 provides probability information or, when the start symbol of the sequence is concerned, the initialization probability information. This information is produced in the same way as has been described with respect to Fig. 24, i.e., by determining a respective slice quantization parameter, by determining a certain context index and, by determining the corresponding initialization parameters m, n, for calculating the initialization probability information, when the entropy decoder is operative to process the start symbol.
  • the entropy encoding/decoding chain is a kind of first-in-first-out pipeline, since the start symbol of the symbol sequence in the encoder is the first symbol to be encoded and is also the first symbol to be decoded. Therefore, the same initialization process for the entropy decoder shown in Fig. 27 can be used, which has been described in connection with the entropy encoder shown in Fig. 24.
  • the context modeler also performs the same operational steps as have been described with respect to Fig. 24.
  • the entropy en/decoder may be implemented as indicated in any of the previous Fig. 1 to 17.
  • transition rules may be given, wherein the transition rules indicate which new state is used based on the currently encoded or decoded symbol for the next symbol to be encoded or decoded. It is hereby of an advantage when a table Next_State_LPS is created which contains the index m of the new probability state P m when a least probable symbol (LPS) occurs in addition to the index n of the currently given probability state P n , and/or when a table Next_State_MPS is created which contains the index m of the new probability state P m when a most probable symbol (MPS) occurs in addition to the index n of the currently given probability state P n .
  • LPS least probable symbol
  • One special implementation of the encoding includes the following steps:
  • p_state describes the current probability state of the context assigned
  • valMPS describes the value of the bit corresponding to the MPS.
  • valMPS describes the value of the bit corresponding to the MPS
  • SliceQP describes the quantization parameter preset at the beginning of a slice
  • m and n describe the model parameters.
  • the estimation of the occurrence probability of the symbols to be coded should be as good as possible. In order to enable an adaptation to non- stationary source statistics, this estimation needs to be updated in the course of the coding process. Generally, usually methods are used for this which operate using scaled frequency counters of the coded results. If C L ps and C M ps designates counters for the occurrence frequencies of LPS and MPS, then using these counters the estimation may be performed and then the entropy en/decoding may be carried out using this probability. For practical purposes the division required in equation (CI) is disadvantageous.
  • transition rules are provided in the form of two tables: ⁇ Next_State_LPS k I 0 ⁇ k ⁇ N max ⁇ and ⁇ Next_State_MPS k I 0 ⁇ k ⁇ N max ⁇ , wherein the tables provide the index m of the new probability state P m when an LPS or MPS occurs, respectively, for the index n of the currently given probability state.
  • the tables provide the index m of the new probability state P m when an LPS or MPS occurs, respectively, for the index n of the currently given probability state.
  • the entropy coding of the bin is performed using a probability distribution according to which all symbol values are equally probable.
  • Each of these two variables needs to be initialized at the beginning of the encoding or decoding, respectively, of a completed coding unit (in applications of video coding about one slice).
  • the initialization values may thereby be derived from control information, like e.g. the quantization parameter (of a slice), as it is illustrated by:
  • a further possibility of adaptation of the starting distributions of the models is provided by the following method.
  • a selection of predetermined starting values of the models may be provided in the encoder. These models may be combined into groups of starting distributions and may be addressed using indices, so that in the encoder the adaptive selection of a group of starting values is performed and is transmitted to the decoder in the form of an index as page information. This method is referred to as a forward-controlled initialization process.
  • the next example is related to video frame coding and, in particular, to an coding scheme using context assignment based on neighboring syntax elements.
  • the following application example provides a video coding scheme using the above embodiments, which enables a higher compression effectiveness.
  • This advantage is achieved by encoding a video signal representing at least one video frame, with at least one video frame being composed of picture samples, the picture samples belonging either to a first or a second field being captured at different time instants, the video frame being spatially divided up into macroblock pair regions, each macroblock pair region being associated with a top and bottom macroblock, the encoding comprising the steps of deciding, for each macroblock pair region, as to whether same is of a first or a second distribution type; assigning, for each macroblock pair region, each of the pixel samples in the respective macroblock pair region to a respective one of the top and bottom macroblock of the respective macroblock pair region, in accordance with the distribution type of the respective macroblock pair region, and pre-coding the video signal into a pre-coded video signal, the pre-coding comprising the sub-step of pre-coding a current macroblock of the top and bottom macroblock associated with a current macroblock pair region of the macroblock pair regions to obtain a current syntax element.
  • a neighboring macroblock at least based upon as to whether the current macroblock pair region is of a first or second distribution type.
  • One of at least two context models is assigned to the current syntax element based on a pre-determined attribute of the neighboring macroblock, wherein each context model is associated with a different probability estimation.
  • an encoding of a sequence of symbols takes place in accordance with any of the above embodiments, with at least one symbol relating to a representation of the syntax element.
  • the assigner is configured to determine the measure for the estimate of the probability distribution for the respective symbol based on the probability distribution estimate associated with the context assigned to the respective symbol.
  • the selector selects the respective entropy coder and the respective partial bit stream accordingly.
  • the coded bit stream is an entropy encoded version of a pre-coded video signal, the pre-coded video signal being a pre-coded version of a video signal, the video signal representing at least one video frame being composed of picture samples, the picture samples belonging either to a first or a second field being captured at a different time instants, the video frame being spatially divided up into macroblock pair regions, each macroblock pair region being associated with a top and a bottom macroblock, each macroblock pair region being either of a first or a second distribution type, wherein, for each macroblock pair region, each of the pixel samples in the respective macroblock pair region is assigned to a respective one of the top and bottom macroblock of the respective macroblock pair region in accordance with the distribution type of the respective macroblock pair region, wherein the syntax element relates to a current macroblock of the top and bottom macroblock of a current macroblock pair region of the macroblock pair regions.
  • the method comprises determining, for the current syntax element, a neighboring macroblock at least based upon as to whether the current macroblock pair region is of a first or a second distribution type; assigning one of at least two context models to the current syntax element based on a predetermined attribute of the neighboring macroblock, wherein each context model is associated with a different probability estimation; and reconstructing a sequence of symbols among which at least one symbol relates to a representation of the syntax element in accordance with any of the above embodiment is performed.
  • the assigner is configured to determine the measure for the estimate of the probability distribution for the respective symbol based on the probability distribution estimate associated with the context assigned to the respective symbol.
  • the selector selects the respective entropy decoder and the respective partial bit stream, respectively, accordingly.
  • the below-outlined application example exploits, that, for sake of a better definition of neighborhood between blocks of picture samples, i.e. the neighboring block which the syntax element to be coded or decoded relates to and the current block based on the attribute of which the assignment of a context model is conducted, and when the neighboring block lies beyond the borders or circumference of the current macroblock containing the current block, it is important to make the determination of the macroblock containing the neighboring block dependent upon as to whether the current macroblock pair region containing the current block is of a first or a second distribution type, i.e., frame or field coded.
  • the blocks may be a macroblock or some sub-part thereof.
  • the determination of a neighboring block comprises at least the determination of a neighboring macroblock a s long as the neighboring block lies beyond the borders of the current macroblock.
  • Fig. 28 shows a general view of a video encoder environment to which the present invention could be applied.
  • a picture of video frame 2010 is fed to a video precoder 2012.
  • the video precoder treats the picture 2010 in units of so-called macroblocks 10a.
  • Each macroblock contains several picture samples of picture 2010.
  • On each macroblock a transformation into transformation coefficients is performed followed by a quantization into transform coefficient levels.
  • intra-frame prediction or motion compensation is used in order not to perform the afore mentioned steps directly on the pixel data but on the differences of same to predicted pixel values, thereby achieving small values which are more easily compressed.
  • Precoder 2012 outputs the result, i.e., the precoded video signal.
  • All residual data elements in the precoded video signal which are related to the coding of transform coefficients, such as the transform coefficient levels or a significance map indicating transform coefficient levels skipped, are called residual data syntax elements.
  • the precoded video signal output by precoder 2012 contains control information syntax elements containing control information as to how each macroblock has been coded and has to be decoded, respectively. In other words, the syntax elements are dividable into two categories.
  • the first category contains the elements related to a macroblock type, sub-macroblock type, and information on prediction modes both of a spatial and of temporal types as well as slice-based and macroblock-based control information, for example.
  • all residual data elements such as a significance map indicating the locations of all significant coefficients inside a block of quantized transform coefficients, and the values of the significant coefficients, which are indicated in units of levels corresponding to the quantizations steps, are combined, i.e., the residual data syntax elements.
  • the macroblocks into which the picture 2010 is partitioned are grouped into several slices.
  • the picture 2010 is subdivided into slices.
  • An example for such a subdivision is shown in Fig. 43, in which each block or rectangle represents a macroblock.
  • a number of syntax elements are generated by precoder 2012, which form a coded version of the macro blocks of the respective slice.
  • the precoder 2012 transfers the syntax elements to a final coder stage 2014, which comprises an entropy coder according to any of the embodiments described above with respect to any of Fig. 1 to 17 and explained in more detail with respect to Fig. 29.
  • the final coder stage 2014 generates an codeword stream for each slice.
  • the final coding stage 2014 exploits the fact that each syntax element is a data value having a certain meaning in the video signal bit stream that is passed to the entropy coder 2014.
  • the entropy coder 2014 outputs a final compressed coded video bit stream comprising codewords for the slices of picture 2010.
  • Fig. 29 shows the arrangement for coding the syntax elements into the final code bit stream, the arrangement generally indicated by reference number 2100.
  • the coding arrangement 2100 is divided into three stages, 2100a, 2100b, and 2100c.
  • the first stage 2100a is the binarization stage and comprises a binarizer 2102.
  • An input of the binarizer 2102 is connected to an input 2104 of stage 2100a via a switch 2106.
  • input 2104 forms the input of coding arrangement 2100.
  • the output of binarizer 2102 is connected to an output 2108 of stage 2100a, which, at the same time, forms the input of stage 2100b.
  • Switch 2106 is able to pass syntax elements arriving at input 2104 to either binarizer 2102 or binarization stage output 2108, thereby bypassing binarizer 2102.
  • switch 2106 The function of switch 2106 is to directly pass the actual syntax element at input 2104 to the binarization stage output 2108 if the syntax element is already in a wanted binarized form.
  • syntax elements that are not in the correct binarization form called non-binary valued syntax elements, are motion vector differences and transform coefficient levels.
  • Examples for a syntax element that has to be binarized since it is not a binary value comprise syntax elements mbjype, coded_block_pattern, ref_idx_10, ref_idx_ll, mvd_10, mvd_ll, and intro_chroma_pred_mode.
  • a fixed-length binarization process is constructed by using an L-bit unsigned integer bin string of the syntax element value, where L is equal to log 2 (cMax+l) rounded up to the nearest integer greater than or equal to the sum, with cMax being the maximum possible value of the syntax element.
  • the indexing of the bins for the fl binarization is such that the bin index of zero relates to the least significant bit with increasing values of the bin index towards the most significant bit.
  • Another binarization scheme is a truncated unary binarization scheme where syntax element values C smaller than the largest possible value cMax are mapped to a bit or bin string of length C+l with the bins having a bin index smaller than C being equal to 1 and the bin having the bin index of C being equal to 0, whereas for syntax elements equal to the largest possible value cMax, the corresponding bin string is a bit string of length cMax with all bits equal to one not followed by a zero.
  • Another binarization scheme is a k-th order exponential Golomb binarization scheme, where a syntax element is mapped to a bin string consisting of a prefix bit string and, eventually, a suffix bit string.
  • Binarizer 2102 maps the non-binary valued syntax elements to a codeword, or a so-called bin string, so that they are now in a binary form.
  • the term "bin” means the binary decision that have to be made at a node of a coding tree defining the binarization mapping of a non-binary value to a bit string or codeword, when transitioning from the route node of the coding tree to the leaf of the coding tree corresponding to the non-binary value of the non-binary syntax element to be binarized.
  • a bin string is a sequence of bins or binary decisions and corresponds to a codeword having the same number of bits, each bit being the result of a binary decision.
  • the bin strings output by binarizer 2102 may not be passed directly to binarization stage output 108 but controllably passed to output 2108 by a bin loop over means 2110 arranged between the output of binarizer 2102 and output 2108 in order to merge the bin strings output by binarizer 2102 and the already binary valued syntax elements bypassing binarizer 2102 to a single bit stream at binarization stage output 2108.
  • the binarization stage 2108 is for transferring the syntax elements into a suitable binarized representation.
  • the binarization procedure in binarizer 2102 preferably yields a binarized representation which is adapted to the probability distribution of the syntax elements so as to enable very efficient binary entropy coding.
  • Stage 2100b is a context modeling stage and comprises a context modeler 2112 as well as a switch 2113.
  • the context modeler 2112 comprises an input, an output, and an optional feedback input.
  • the input of context modeler 2112 is connected to the binarization stage output 2108 via switch 2113.
  • the output of context modeler 2112 is connected to a regular coding input terminal 2114 of stage 2100c.
  • the function of switch 2113 is to pass the bits or bins of the bin sequence at binarization stage output 2108 to either the context modeler 2112 or to a bypass coding input terminal 2116 of stage 2100c, thereby bypassing context modeler 2112.
  • the aim of switch 2113 is to ease the subsequent binary entropy coding performed in stage 100c.
  • some of the bins in the bin string output by binarizer 2102 show heuristically nearly an equi-probable distribution.
  • the corresponding bits are, with a probability of nearly 50%, 1 and, with a probability of nearly 50%, 0, or, in other words, the bits corresponding to this bin in a bin string have a 50/50 chance to be 1 or 0.
  • These bins are fed to the bypass-coding input terminal 2116 and are binary entropy coded by use of an equi-probable probability estimation, which is constant and, therefore, needs no adaption or updating overhead.
  • Context modeler 2112 manages a set of context models. For each context model, the context modeler 2112 has stored an actual bit or bin value probability distribution estimation. For each bin that arrives at the input of context modeler 2112, the context modeler 2112 selects one of the sets of context models. In other words, the context modeler 2112 assigns the bin to one of the set of context models.
  • the assignment of bins to a context model is such that the actual probability distribution of bins belonging to the same context model show the same or likewise behavior so that the actual bit or bin value probability distribution estimation stored in the context modeler 2112 for a certain context model is a good approximation of the actual probability distribution for all bins that are assigned to this context model.
  • the assignment process in accordance with the present invention exploits the spatial relationship between syntax element of neighboring blocks. This assignment process will be described in more detail below.
  • the context modeler 2112 When having assigned the context model to an incoming bin the context modeler 2112 passes the bin further to entropy coding stage 2100c together with the probability distribution estimation of the context model, which the bin is assigned to. By this measure, the context modeler 2112 drives the entropy coding stage 2100c to generate a sequence of bits as a coded representation of the bins input in context modeler 2112 by switch 2113 according to the switched bit value probability distribution estimations as indicated by the context modeler 2112.
  • the context modeler 2112 continuously updates the probability distribution estimations for each context model in order to adapt the probability distribution estimation for each context model to the property or attributes of the picture or video frame from which the syntax elements and bins have been derived.
  • the estimation adaptation or estimation update is based on past or prior bits or bin values which the context modeler 2112 receives at the feedback input over a feedback line 2117 from stage 2100c or may temporarily store.
  • the context modeler 2112 updates the probability estimations in response to the bin values passed to entropy coding stage 2100c.
  • the context modeler 2112 uses a bin value assigned to a certain context model merely for adaptation or update of the probability estimation that is associated with the context model of this bin value.
  • syntax elements when the same bin or same syntax element occurs several times in the bins passed from stage 2100a may be assigned to different of the context models each time they occur, depending on previously incoming or previously entropy coded bins, and/or depending on other circumstances, such as previously coded syntax elements of neighboring blocks, as is described in more detail below with respect to exemplary syntax elements.
  • the third stage 2100c of coding arrangement 100 is the entropy coding stage. It comprises a regular coding engine 2118, a bypass-coding engine 2120, and a switch 2122.
  • the regular coding engine 2118 comprises an input and an output terminal. The input terminal of regular coding engine 2118 is connected to the regular coding input terminal 2114.
  • the regular coding engine 2118 binary entropy codes the bin values passed from context modeler 2112 by use of the context model also passed from context modeler 2112 and outputs coded bits. Further, the regular coding engine 2118 passes bin values for context model updates to the feedback input of context modeler 2112 over feedback line 2117.
  • the regular coding engine corresponds to any of the above-mentioned embodiments of Fig.
  • assigner may quantize the probability index before forwarding the thus obtained parameter to the selector 18 which, in turn, selects the respective entropy encoder.
  • the bypass-coding engine 2112 has also an input and an output terminal, the input terminal being connected to the bypass coding input terminal 2116.
  • the bypass-coding engine 2120 is for binary entropy coding the bin values passed directly from binarization stage output 2108 via switch 2113 by use of a static predetermined probability distribution estimation and also outputs coded bits.
  • the bypass-coding engine may use a 1:1 mapping in units of any number of symbols.
  • regular coding engine 2118 and bypass coding engine 2120 cooperate in order to bit wise perform entropy coding based on either an adaptive or a static probability distribution model.
  • coding arrangement 2100 After having described with respect to Figs. 28 and 29 rather generally the operation of coding arrangement 2100, in the following its functioning is described in more detail with respect to the handling of exemplary syntax elements for which an context assignment process based on syntax elements of neighboring blocks is used, in accordance with embodiments.
  • Figs. 30 to 31b the meaning of MBAFF coding is described, in order to enable a better understanding of the definition of neighborhood between a current block and a neighboring block used during assignment of a context model to a syntax element concerning the current block in case of MBAFF.
  • Fig. 30 shows a picture or decoded video frame 2010.
  • the video frame 2010 is spatially partitioned into macroblock pairs 10b.
  • the macroblock pairs are arranged in an array of rows 2200 and columns 2202.
  • Each macroblock pair consists of two macroblocks 2010a.
  • each macroblock pair row 2200 consists of two macroblock rows, i.e., an top macroblock row 2200a consisting of the top macroblocks in the macroblock pairs of the macroblock pair line 2200 and a bottom macroblock row 200b comprising the bottom macroblocks of the macroblock pairs.
  • the top macroblock of the top left macroblock pair resides at address zero.
  • the next address, i.e. address 1 is assigned to the bottom macroblock of the top left macroblock pair.
  • the addresses of the top macroblocks of the macroblock pairs in the same, i.e., top macroblock row 2200a, are 2, 4, ... , 2i-2, with the addresses rising from left to right, and with i expressing the picture width in units of macroblocks or macroblock pairs.
  • the addresses 1, 3, 2i-l are assigned to the bottom macroblocks of the macroblock pairs in the top macroblock pair row 2200, the addresses rising from left to right.
  • next 2i- addresses from 2i to 4i-l are assigned to the macroblocks of the macroblock pairs in the next macroblock pair row from the top and so on, as illustrated in Fig. 30 by the numbers written into the boxes representing the macroblocks 2010a and by the arched rows.
  • Fig. 30 does show the spatial subdivision of picture 2010 in units of macroblock pairs rather than in macroblocks.
  • Each macroblock pair 2010b represents a spatial rectangular region of the pictures. All picture samples or pixels (not shown) of picture 2010 lying in the spatial rectangular region of a specific macroblock pair 2010b belong to this macroblock pair. If a specific pixel or picture sample belongs to the top or the bottom macroblock of a macroblock pair depends on the mode by which precoder 2012 has coded the macroblocks in that macroblock pair as it is described in more detail below.
  • Fig. 31a shows on the left hand side the arrangement of pixels or picture samples belonging to a macroblock pair 2010b. As can be seen, the pixels are arranged in an array of rows and columns. Each pixel shown is indicated by a number in order to ease the following description of Fig. 31a. As can be seen in Fig. 31a, some of the pixels are marked by an "x" while the others are marked "Y". All pixels marked with "x" belong to a first field of the picture while the other pixels marked with "Y” belong to a second field of the picture. Pixels belonging to the same field are arranged in alternate rows of the picture.
  • the picture or video frame can be considered to contain two interleaved fields, a top and a bottom field.
  • the top field comprises the pixels marked with "Y" and contains even-numbered rows 2n+2, 2n+4, 2n+6, ... with 2n being the number of rows of one picture or video frame and n being an integer greater than or equal to 0.
  • the bottom field contains the odd-numbered rows starting with the second line of the frame.
  • the video frame to which macroblock pair 2010b belongs is an interlaced frame where the two fields were captured at different time instants, for example the top field before the bottom field. It is now that the pixels or picture samples of a macroblock pair are differently assigned to the top or bottom macroblock of the macroblock pair, depending on the mode by which the respective macroblock pair is precoded by precoder 2012 (Fig. 28). The reason for this being the following.
  • the picture samples of a macroblock which may be luminance or luma and chrominance or chroma samples, may be either spatially or temporarily predicted by precoder 2012, and the resulting prediction residual is encoded using transform coding in order to yield the residual data syntax elements. It is now that in interlaced frames (and it is assumed that the present video frame is an interlaced frame), with regions of moving objects or camera motion, two adjacent rows of pixels tend to show a reduced degree of statistical dependency when compared to progressive video frames in which both fields are captured at the same time instant.
  • the pre-coding performed by precoder 2012 which, as stated above, operates on macroblocks, may achieve merely a reduced compression efficiency when a macroblock pair is spatially sub-divided into a top macroblock representing the top half region of the macroblock pair and a bottom macroblock representing the bottom half region of the macroblock pair, since in this case, both macroblocks, the top and the bottom macroblock, comprise both top field and bottom field pixels.
  • Figs. 31a and 31b show on the right hand side the resulting top and bottom macroblock in accordance with the frame and field mode, respectively.
  • Fig. 31a represents the frame mode, i.e., where each macroblock pair is spatially subdivided in a top and a bottom half macroblock.
  • Fig. 31a shows at 2250 the top macroblock and at 2252 the bottom macroblock as defined when they are coded in the frame mode, the frame mode being represented by double-headed arrow 2254.
  • the top macroblock 250 comprises one half of the pixel samples of the macroblock pair 2010b while the other picture samples are assigned to the bottom macroblock 2252.
  • both macroblocks 2250 and 2252 comprise both, picture elements of the first field marked with "x" and captured at a first time instant and picture samples of the second field marked with "Y" and captured at a second, different time instant.
  • top or bottom macroblocks are slightly different in field mode.
  • the top macroblock 2252 of the macroblock pair 2010b contains all picture samples of the top field, marked with "x"
  • the bottom macroblock 2254 comprises all picture samples of the bottom field, marked with "Y”.
  • each macroblock in a macroblock pair does merely contain either picture samples of the top field or picture samples of the bottom field rather than a mix of picture samples of the top and bottom field.
  • precoder 2012 When the precoder 2012 receives a video signal representing an interlaced video frame, precoder 2012 is free to make the following decisions when coding the video frame 2010:
  • each macroblock pair and each macroblock would be coded in field mode.
  • the choice between the three options can be made adaptively for each frame in a sequence.
  • the choice between the first two options is referred to as picture adaptive frame/field (PAFF) coding.
  • PAFF picture adaptive frame/field
  • MBAFF macroblock adaptive frame/field
  • precoder 2012 decides to use just this option.
  • MBAFF coding allows the precoder to better adapt the coding mode type (filed or frame mode) to the respective areas of scenes. For example, precoder 2012 codes macroblock pairs located at stationary areas of a video scene in frame mode, while coding macroblock pairs lying in areas of a scene showing fast movements in field mode.
  • each macroblock contains frame lines.
  • the top macroblock contains top field lines and the bottom macroblock contains bottom field lines.
  • the frame/field decision for each macroblock pair is made at the macroblock pair level by precoder 2012, i.e. if the top macroblock is field coded same applies for the bottom macroblock within same macroblock pair.
  • Each macroblock of a field macroblock pair is processed very similarly to a macroblock within a field in PAFF coding.
  • some stages of the pre-coding procedure in precoder 2012 such as the prediction of motion vectors, the prediction of intra prediction modes, intra frame sample prediction, deblocking filtering and context modeling in entropy coding and the zig-zag scanning of transform coefficients are modified when compared to the PAFF coding in order to account for this mixture.
  • the pre-coded video signal output by precoder 2012 depends on the type of coding precoder 2012 has decided to use.
  • the pre-coded video signal contains a flag mb_field_decoding_flag for each non-skipped macroblock pair.
  • the flag mb_field_decoding_flag indicates for each macroblock pair it belongs to whether the corresponding macroblocks are coded in frame or field coding mode. On decoder side, this flag is necessary in order to correctly decode the precoded video signal.
  • the flag mb_field_decoding_flag is zero, whereas the flag is one in the other case.
  • Fig. 32 starts at the arrival of a bin value of a syntax element at the input of context modeler 2112. That is, eventually, the syntax element had to be binarized in binarizer 2102 if needed, i.e. unless the syntax element is already a binary value.
  • context modeler 2112 determines as to whether the incoming bin is a bin dedicated to a context assignment based on neighboring syntax elements, i.e. syntax elements in neighboring blocks. It is recalled that the description of Fig. 32 assumes that MBAFF coding is active. If the determination in step 2300 results in the incoming bin not being dedicated to context assignment based on neighboring syntax elements, another syntax element handling is performed in step 2304.
  • context modeler 2112 determines a neighboring block of the current block to which the syntax element of the incoming bin relates.
  • the determination process of step 2306 is described in more detail below with respect to exemplary syntax elements and their bins, respectively. In any case, the determination in step 2306 depends on the current macroblock to which the syntax element of the current bin relates being frame or field coded, as long as the neighboring block in question is external to the macroblock containing the current block.
  • step 2308 the context modeler 2112 assigns a context model to the bin based on a predetermined attribute of the neighboring block.
  • the step of assigning 2308 results in a context index ctxidx pointing to the respective entry in a table assigning each context index a probability model, to be used for binary entropy coding of the current bin of the current syntax element.
  • context modeler 2112 After the determination of ctxidx, context modeler 2112 passes the variable ctxidx or the probability estimation status indexed by ctxidx along with the current bin itself to regular coding engine 2118. Based on these inputs, the regular coding engine 2118 entropy encodes, in step 2322, the bin by using the current probability state of the context model as indexed by ctxidx. In particular, the index is used to select the respective partial bitstream 2012, for example.
  • regular coding engine 2118 passes the bin value via path 2117 back to context modeler 2112, whereupon context modeler 2112 adapts, in step 2324, the context model indexed by ctxidx with respect to its probability estimation state. Thereafter, the process of coding the syntax element into the bit stream at the output 2124 ends at 2326.
  • the bin string into which the syntax element may be binarized before step 2310 may be composed of both, bins that are encoded, i.e. assigned to the respective partial bitstream 2012 and subsequently coded, by use of the current probability state of context model ctxidx in step 2322 and bins encoded in bypass coding engine 2120 by use of an equi-probable probability estimation although this is not shown in Fig. 32. Rather, Fig. 32 merely concerns the exemplary encoding of one bin of a syntax element.
  • Fig. 33 shows, on the left hand side, a flow diagram of the process 2327.
  • Fig. 33 shows a memory 2328 to which both, the context modeler 2112 and the regular coding engine 2118, have access in order to load, write, and update specific variables.
  • variables comprise R and L, which define the current state or current probability interval of the binary entropy coder 2100c.
  • Memory 2328 contains a table 2329, which associates each possible value of ctxidx, e.g. 0 - 2398, a pair of a probability state index ⁇ and an MPS value co, both defining the current probability estimation state of the respective context model indexed by the respective context index ctxidx.
  • the probability state ⁇ is an index that uniquely identifies one of a set of possible probability values ⁇ ⁇ .
  • the probability values ⁇ ⁇ are an estimation for the probability of the next bin of that context model to be a least probable symbol (LPS). Which of the possible bin values, i.e., a null or one, is meant by the LPS, is indicated by the value of MPS ⁇ . If ⁇ is 1, LPS is 0 and vice- versa.
  • the state index and MPS together uniquely define the actual probability state or probability estimation of the respective context model.
  • step 2334 the actual bin value is sent to the regular coding engine 2118.
  • step 2336 the regular coding engine 2118 checks as to whether the value of the actual bin, i.e. either the already binary syntax element or one bin of a bin string obtained from the current syntax element, is equal to the most probable symbol as indicated by co; or not . If the current bin is the MPS, context modeler 2112 updates the probability estimation state of the current context model by updating Oj. In particular, context modeler 2112 uses a table 2340 which associates each probability state index ⁇ with an updated probability state index in case the actual symbol or bin was the most probable symbol, i.e., ⁇ becomes transIdxMPS(a; ).
  • step 2338 the process ends at 2340 where the bin sent to the regular coding engine 2118 is forwarded by the selector thereof, to the respective partial bitstream encoder.
  • Mb_skip_flag This flag relates to a certain macroblock of a certain slice of a video frame.
  • Mb skip flag 1 specifies that the current macroblock is to be skipped when performing a decoding process on the precoded video signal.
  • Mb_skip_flag 0 specifies that the current macroblock is not skipped.
  • Mb_skip_flag 1 specifies that for the current macroblock, when decoding a P or SP slice, Mbjype is inferred to be p_skip and the macroblock type is collectively referred to as P macroblock type, and when decoding a B slice, Mbjype is inferred to be B_skip and the macroblock type is collectively referred to as B macroblock type.
  • Mb_field_ Mb field_decodingJlag 0 specifies that the current decoding_flag macroblock pair is a frame macroblock pair and Mb_field_deco- dingjlag equal to 0 specifies that the macroblock pair is a field macroblock pair.
  • Both macroblocks of a frame macroblock pair are referred to in the present description as frame macroblocks, whereas both macroblocks of a field macroblock pair are referred to in this text as field macroblocks.
  • Mbjype specifies the macroblock type.
  • the semantics of Mbjype in the H.264/AVC standard depends on the slice type.
  • Mbjype can assume values in the range of 0 to 25, 0 to 30, 0 to 48 or 0 - 26, depending on the slice type.
  • Coded_block_ Coded_block )attern specifies which of a sub-part of the current pattern macroblock contains non-zero transform coefficients.
  • Transform coefficients are the scalar quantities, considered to be in a frequency domain, that are associated with a particular one-dimensional or two-dimensional frequency index in an inverse transform part of the decoding process.
  • each macroblock 2010a - irrespective of the macroblock being a frame coded macroblock (Fig. 31a) or a field coded macroblock (Fig. 31b), is partitioned into smaller sub-parts, the sub-parts being arrays of size 8x 8 pixel samples.
  • the pixels 1 to 8, 11 to 18, 21 to 28,..., 71 to 78 could form the upper left block of luma pixel samples in the top macroblock 2250 of macroblock pair 2010b.
  • This top macroblock 2250 would comprise another three of such blocks, all four blocks arranged in a 2x2 array.
  • the precoded video signal output by precoder 2012 would comprise one or several syntax elements coded_block_pattern.
  • the transformation from spatial domain to frequency domain could be performed on these 8x8 sub-parts or on some smaller units, for example, 4x4 sub- arrays, wherein each 8x8 sub-part comprises 4 smaller 4x4 partitions.
  • the present description mainly concerns luma pixel samples. Nevertheless, the same could also apply accordingly for chroma pixel samples.
  • ref_Idx_10 when present in the precoded video signal output by precoder 12, specifies an index in a list 0 of a reference picture to be used for prediction. The same applies for ref_Idx_ll but with respect to another list of the reference picture.
  • mvd_10/mvd_ll mvd_10 specifies the difference between a vector component to be used for motion prediction and the prediction of the vector component. The same applies for mvd_ll, the only difference being, that same are applied to different reference picture lists.
  • ref_Idx_10, ref_Idx_ll, mvd_10 and mvd_ll all relate to a particular macroblock partition. The partitioning of the macroblock is specified by Mb type.
  • intra_chroma_ Intra_chroma_pred_mode specifies the type of spatial prediction pred_mode used for chroma whenever any part of the luma macroblock is intra- coded.
  • intra prediction a prediction is derived from the decoded samples of the same decoded picture or frame.
  • Intra prediction is contrary to inter prediction where a prediction is derived from decoded samples of reference pictures other than the current decoded picture.
  • coded_block_ coded_block_flag relates to blocks of the size of 4x4 picture flag samples. If coded_block_flag is equal to 0, the block contains no non-zero transform coefficients. If coded block flag is equal to 1, the block contains at least one non-zero transform coefficient.
  • Fig. 36 illustrates the partition of macroblocks (upper row) and sub-macroblocks (lower row). The partitions are scanned for inter prediction as shown in Fig. 36.
  • the outer rectangles in Fig. 36 refer to the samples in a macroblock or sub-macroblock, respectively.
  • the inner rectangles refer to the partitions. The number in each inner rectangle specifies the index of the inverse macroblock partition scan or inverse sub-macroblock partition scan.
  • Fig. 34 Before describing in detail the dependency of the context model assignment on the syntax element of neighboring blocks, with respect to Fig. 34, it is described, how the addresses of the top macroblock of the macroblock pair to the left and above the current macroblock pair may be computed, since these are the possible candidates, which comprise the syntax element in the block to the left of and above the current block containing the current syntax element to be encoded.
  • Fig. 34 a portion of six macroblock pairs of a video frame is shown, wherein each rectangle region in Fig. 34 corresponds to one macroblock and the first and the second two vertically adjacent macroblocks in each column form a macroblock pair.
  • CurrMbAddr denotes the macroblock address of the top macroblock of the current macroblock pair, the current syntax element is associated with or relates to.
  • the current macroblock pair is encompassed by bold lines. In other words, they from the border of a macroblock pair.
  • mbAddrA and mbAddrB denote the addresses of the top macroblocks of the macroblock pairs to the left and above the current macroblock pair, respectively.
  • context modeler 2112 computes
  • MbAddrA 2 ⁇ ⁇ CurrMbAddr/2 - 1)
  • each macroblock contains 16x16 luma samples. These luma samples are divided up into four 8x8 luma blocks. These luma blocks may be further subdivided into 4x4 luma blocks. Furthermore, for the following description, each macroblock further comprises 8x8 luma samples, i.e., the pixel width of the chroma samples being doubled compared to luma samples. These 8x8 chroma samples of a macroblock are divided up into four 4x4 luma blocks. The blocks of a macroblock are numbered. Accordingly, the four 8x8 luma blocks each have a respective block address uniquely indicating each 8x8 block in the macroblock.
  • each pixel sample in a macroblock belongs to a position (x, y) wherein (x, y) denotes the luma or chroma location of the upper-left sample of the current block in relation to the upper-left luma or chroma sample of the macroblock.
  • (x, y) denotes the luma or chroma location of the upper-left sample of the current block in relation to the upper-left luma or chroma sample of the macroblock.
  • the pixel 2023 in top macroblock 2252 in Fig. 31b would have the pixel position (2, 1), i.e., third column, second row.
  • the context modeler assignment depends on syntax elements relating to neighboring macroblocks.
  • the addresses mbAddrA and mbAddrB are determined as described above.
  • condTermN (with N being either A or B) be a variable that is set as follows:
  • condTermN is set to 1.
  • ctxidx is determined as follows:
  • condTermN (with N being either A or B) be a variable that is set as follows.
  • the macroblock mbAddrN is a frame macroblock.
  • condTermN is set to 1.
  • ctxidx is determined dependent on the availability of macroblocks mbAddrN (with N being either A or B), and the syntax element Mbjype of this neighboring macroblocks.
  • the dependency of the context modeler assignment is defined accordingly, wherein for syntax elements, which relate to blocks smaller than a macroblock, the assignment is also dependent on syntax element relating to such smaller blocks being smaller than macroblocks.
  • the context index assignment is dependent not only on the availability of macroblock MbAddrN and the syntax element Mbjype of the macroblock MbAddrN but also on the syntax element Coded_block_pattern of the neighboring block.
  • the syntax elements listed above are all dependent on the respective syntax element of the neighboring block.
  • the context model assignment of syntax elements mvd_10, mvd_ll, ref_idx 0 and ref_idx_ll is not dependent on the respective syntax elements of the neighboring block.
  • the context modeler assignment for intra_chroma_pred_mode is dependent on mbAddrN availability, macroblock mbAddrN being coded in inter prediction mode or not, Mb_type for the macroblock mbAddrN and the syntax element intra_chroma_pred_mode for the macroblock MbAddrN.
  • syntax element coded_block_flag context model assignment is dependent on the availability of MbAddrN, the current macroblock being coded in inter prediction mode, Mb_type for the macroblock mbAddrN and the syntax element coded_block_flag of the neighboring block.
  • a neighboring block is determined.
  • this involves computing mbAddrN and the block index indexing the sub-part of the macroblock MbAddrN, this sub-part being the neighboring block of the current block.
  • the neighborhood for slices using macroblock adaptive frames/field coding as described in the following in accordance with an embodiment of the present invention is defined in a way that guarantees that the areas covered by neighboring blocks used for context modeling in context adaptive binary entropy coding inside an MBAFF-frame adjoin to the area covered by the current block.
  • section 1.1 The general concept of defining the neighborhood between a current and a reference block is described in the following section 1.1.
  • section 1.2 a detailed description, which specifies how the neighboring blocks, macroblocks, or partitions to the left of and above the current block, macroblock, or partition are obtained for the purpose of context modeling in context adaptive binary entropy coding, is given.
  • CurrPic specifies the current frame, which is obtained by interleaving the top and the bottom field, if the current block is part of a macroblock pair coded in frame mode (rnb_field_decoding_fIag is equal to 0). If the current block is or is part of a top field macroblock, CurrPic specifies the top field of the current frame; and if the current block is or is part of a bottom field macroblock, CurrPic specifies the bottom field of the current frame.
  • (xA, yA) and (xB, yB) denote the luma or chroma location to the left of and above the location (x, y), respectively, inside the picture CurrPic.
  • the locations (xA, yA) and (xB, yB) are specified by
  • the block to the left of the current block is defined as the block that contains the luma or chroma sample at location (xA, yA) relative to the upper-left luma or chroma sample of the picture CurrPic and the block above the current block is defined as the block that contains the luma or chroma sample at location (xB, yB) relative to the upper-left luma or chroma sample of the picture CurrPic. If (xA, yA) or (xB, yB) specify a location outside the current slice, the corresponding block is marked as not available. 1.2. Detailed description of neighborhood definition
  • the algorithm described in Sec. 1.2.1 specifies a general concept for MBAFF-slices that describes how a luma sample location expressed in relation to the upper-left luma sample of the current macroblock is mapped onto a macroblock address, which specifies the macroblock that covers the corresponding luma sample, and a luma sample location expressed in relation to the upper-left luma sample of that macroblock. This concept is used in the following Sec. 1.2.2 - 1.2.6.
  • the Sec. 1.2.2 - 1.2.6 describe how the neighboring macroblocks, 8x8 luma blocks, 4x4 luma blocks, 4x4 chroma block, and partitions to the left of and above a current macroblock, 8x8 luma block, 4x4 luma block, 4x4 chroma block, or partition are specified.
  • These neighboring macroblock, block, or partitions are needed for the context modeling of CABAC for the following syntax elements: mb_skip_flag, mb_type, coded_block_pattern, intra_chroma_pred_mode, coded_block_fiag, ref_idx_10, ref_idx_ll, mvdJLO, mvd_ll.
  • (xN, yN) denote a given luma sample location expressed in relation to the upper-left luma sample of the current macroblock with the macroblock address CurrMbAddr. It is recalled that in accordance with the present embodiment each macroblock comprises 16x16 luma samples. xN and yN lie within— 1...16. Let mbAddrN be the macroblock address of the macroblock that contains (xN, yN), and let (xW,yW) be the, location (xN, yN) expressed in relation to the upper-left luma sample of the macroblock mbAddrN (rather than relative to the upper-left luma sample of the current macroblock).
  • mbAddrA and mbAddrB specify the macroblock address of the top macroblock of the macroblock pair to the left of the current macroblock pair and the top macroblock of the macroblock pair above the current macroblock pair, respectively.
  • PicWidthlnMbs be a variable that specifies the picture width in units of macroblocks.
  • mbAddrA is- less than 0, or if (CurrMbAddr/2) % PicWidthlnMbs is equal to 0, or if the macroblock with address mbAddrA belongs to a different slice than the current slice, mbAddrA is marked as not available.
  • mbAddrB 2* (CurrMbAddr/2 - PicWidthlnMbs) If mbAddrB is less than 0, or if the macroblock with address mbAddrB belongs to a different slice than the current slice, mbAddrB is marked as not available.
  • the Table in Fig. 35 specifies the macroblock address mbAddrN and a variable yM in the following two ordered steps:
  • variable currMbFrameFlag (third column) is set to 1, if the current macroblock with address CurrMbAddr is a part of a frame macroblock pair; otherwise it is set to 0.
  • variable mbl sTopMbFlag (forth column) is set to 1 , if CurrMbAddr%2 is equal to 0; otherwise it is set to 0.
  • mbAddrX (which can be either mbAddrA or mbAddrB) is marked as not available, mbAddrN is marked as not available.
  • mbAddrX is available
  • mbAddrN is marked as available and Table 1 specifies mbAddrN and yM depending on (xN, yN) (first and second column), currMbFrameFlag (third column), mblsTopMbFlag (forth column), and the variable mbAddrXFrameFlag (sixth column), which is derived as follows: rribAddrXFrameFlag is set to 1, if the macroblock mbAddrX is a frame macroblock; otherwise it is set to 0.
  • the input values xN, yN, currMbFrameFlag and MblsTopMbFlag are entered.
  • the possible input values for parameters xN and yN are -1 to 16, inclusive.
  • These parameters determine mbAddrX listed in the fifth column, i.e. the macroblock pair containing the wanted luma sample.
  • the next two columns, i.e., the sixth and the seventh column, are needed to obtain the final output mbAddrN and yN.
  • MbAddrXFrameFlag indicating as to whether a macroblock pair indicated by mbAddrX is frame or field coded, and some additional conditions concerning as to whether yN is even or odd numbered or is greater than or equal to 8 or not.
  • xN and yN are both positive or zero, i.e., the wanted pixel sample lies within the current macroblock relative to which xN and yN are defined
  • the output macroblock address does not change, i.e., it is equal to CurrMbAddr.
  • yM is equal yM. This changes when the input xM and yM indicates a pixel sample lying outside the current macroblock, i.e., to the left (xN ⁇ 0) all to the top of the current macroblock (yN ⁇ 0).
  • the specification of the neighboring macroblocks to the left of and above the current macroblock is used for the context modeling of CAB AC for the following syntax elements: mb_skip_flag, mb_type, coded_block_pattern, intra_chroma_prediction_mode, and coded_block_flag.
  • mbAddrA be the macroblock address of the macroblock to the left of the current macroblock
  • mbAddrB be the macroblock address of the macroblock above the cvtrrent macroblock.
  • luma8x8BIkIdx be the index of the current 8x8 luma block inside the current macroblock CurrMbAddr.
  • An embodiment of the assignment of block index luma8x8BlkIdx to the respective blocks within a macroblock is shown in Fig. 9 (upper-right corner).
  • mbAddrA be the macroblock address of the macroblock that contains the 8x8 luma block to the left of the current 8x8 luma block
  • mbAddrB be the macroblock address of the macroblock that contains the 8x8 luma block above the current 8x8 luma block.
  • luma8x8BlkIdxA be the 8x8 luma block index (inside the macroblock mbAddrA) of the 8x8 luma block to the left of the current 8x8 luma block
  • luma8x8BlkIdXB be the 8x8 luma block index (inside the macroblock mbAddrB) of the 8x8 luma block above the current 8x8 luma block.
  • luma4x4BlkIdx be the index (in decoding order) of the current 4x4 luma block inside the current macroblock CurrMbAddr.
  • luma4x4BlkIdx could be defined as luma8x8BlkIdx of the 8x8 block containing the 4x4block multiplied by 4 plus the partition number as shown in the bottom-right corner of Fig. 9.
  • mbAddrA be the macroblock address of the macroblock that contains the 4x4 luma block to the left of the current 4x4 luma block
  • mbAddrB be the macroblock address of the macroblock that contains the 4x4 luma block above the current 4x4 luma block.
  • luma4x4BlkIdxA be the 4x4 luma block index (inside the macroblock mbAddrA) of the 4x4 luma block to the left of the current 4x4 luma block
  • luma4x4BlkIdxB be the 4x4 luma block index (inside the macroblock mbAddrB) of the 4x4 luma block above the current 4x4 luma block.
  • the specification of the neighboring 4x4 chroma blocks to the left of and above the current 4x4 chroma block is used for the context modeling of CABAC for the syntax element coded_block_flag.
  • chroma4x4BlkIdx be the index (in decoding order) of the current 4x4 chroma block inside the current macroblock CurrMbAddr.
  • mbAddrA be the macroblock address of the macroblock that contains the 4x4 chroma block to the left of the current 4x4 chroma block
  • mbAddrB be the macroblock address of the macroblock that contains the 4x4 chroma block above the current 4x4 chroma block.
  • chroma4x4BlkIdxA be the 4x4 chroma block index (inside the macroblock mbAddrA) of the 4x4 chroma block to the left of the current 4x4 chroma block
  • chroma4x4BlkIdxB be the 4x4 chroma block index (inside the macroblock mbAddrB) of the 4x4 chroma block above the current 4x4 chroma block.
  • chroma4x4BIkIdxA is set equal to luma8x8BlkIdxA; otherwise chroma4x4BlkIdxA is marked as not available.
  • chroma4x4BlkIdxB is set equal to luma8x8BlkIdxB; otherwise chroma4x4BlkIdxB is marked as not available.
  • mbPartldx and subMbPartldx be the macroblock partition and sub-macroblock partition indices that specify the current partition inside the current macroblock CurrMbAddr.
  • An example for such partition indices is shown in Fig. 36.
  • mbAddrA be the macroblock address of the macroblock that contains the partition to the left of the current partition
  • mbAddrB be the macroblock address of the macroblock that contains the partition above the current partition.
  • mbPartldxA and subMbPartldxA be the macroblock partition and sub-macroblock partition indices (inside the macroblock mbAddrA) of the partition to the left of the current partition
  • mbPartldxB and subMbPartldxB be the macroblock partition and sub-macroblock partition indices (inside the macroblock mbAddrB) of the partition above the current partition.
  • mbAddrA, mbAddrB, mbPartldxA, subMbPartldxA, mbPartldxB, subMbPartldxB, and their availability statuses are obtained as follows:
  • (xC, yC) be the luma location of the upper-left sample of the current partition given by mbPartldx and subMbPartldx relative to the upper-left luma sample of the current macroblock.
  • Fig. 37 shows a general view of a video decoder environment to which the present invention could be applied.
  • An entropy decoder 400 receives the entropy coded bit stream as described above and treats it as will be described in more detail below with respect to Figs. 38 - 39.
  • the entropy decoder 2400 decodes the entropy coded bit stream by binary entropy decoding in order to obtain the precoded video signal and, in particular, syntax elements contained therein and passes same to a precode decoder 2402.
  • the precode decoder 2402 uses the syntax elements, such as motion vector components and flags, such as the above listed syntax elements, in order to retrieve, macroblock by macroblock and then slice after slice, the picture samples of pixels of the video frames 2010.
  • Fig. 38 now shows the decoding process performed by the entropy decoder 2400 each time a bin is to be decoded. Which bin is to be decoded depends on the syntax element which is currently expected by entropy decoder 2400. This knowledge results from respective parsing regulations.
  • step 2500 the decoder 2400 checks as to whether the next bin to decode is a bin of a syntax element of the type corresponding to context model assignment based on neighboring syntax elements. If this is not the case, decoder 2400 proceeds to another syntax element handling in step 2504. However, if the check result in step 2500 is positive, decoder 2400 performs in steps 2506 and 2508 a determination of the neighboring block of the current block which the current bin to decode belongs to and an assignment of a context model to the bin based on a predetermined attribute of the neighboring block determined in step 2506, wherein steps 2506 and 2508 correspond to steps 2306 and 2308 of encoding process of Fig. 32.
  • ctxldx The result of these steps is the context index ctxldx. Accordingly, the determination of ctxldx is performed in steps 2506 and 2508 in the same way as in the encoding process of Fig. 32 in steps 2306 and 2308 in order to determine the context model to be used in the following entropy decoding.
  • the entropy decoder 2400 entropy decodes the actual bin, from the entropy coded bit stream by use of the actual probability state of the context module as indexed by ctxldx obtained in steps 2510 to 2520. That is, the entropy decoder 2400 comprises a regular decoding engine that corresponds to any of the above-described entropy decoders having a plurality of entropy decoders inside, each associated with a respective partial bitstream. In step 2522 the bin is requested from that partial bitstream which is indexed by the probability index or, at least, depending therefrom. The result of this step is the value for the actual bin.
  • step 2524 the ctxldx probability state is adapted or updated, as it was the case in step2 224. Thereafter, the process ends at step 2526.
  • the individual bins that are obtained by the process shown in Fig. 38 represent the syntax element value merely in case the syntax element is of a binary type. Otherwise, a step corresponding to the binarization has to be performed in reverse manner in order to obtain from the bin strings the actual value of the syntax element.
  • Fig. 39 shows the steps 2522 and 2524 being encompassed by dotted line2 527 in more detail on the left hand side.
  • Fig. 38 shows a memory and its content to which entropy decoder 2400 has access in order to load, store and update variables.
  • entropy decoder manipulates or manages the same variables as entropy coder 2014 since entropy decoder 2400 emulates the encoding process as will be described in the following.
  • decoder 2400 obtains the value of the current bin from the partial bitstream indexed by the probability index associated with the respective context.
  • decoder 2400 checks as to whether the value of the entropy coding codeword in the entropy coded bit stream was the most probable symbol as indicated by G>I ; in which case decoder 2400 accordingly sets the bin value to the value of co; in step 2572. In the other case, decoder 2400 sets the symbol to be 1- ⁇ , in step 2574.
  • step 2576 the decoder 2400 adapts or updates the probability state of the current context model i as defined by o>j and G>; by transitioning the probability state index coj as was described with respect to step 2338 in Fig. 36. Thereafter, the process 2527 ends at step 2578.
  • the assignment in accordance with steps 2308 and 2408 does not have to be dependent on syntax elements of neighboring blocks, i.e., syntax elements contained in the precoded video signal output by precoder 2012. Rather, the assignment may be dependent on other attributes of the neighboring blocks.
  • the definition of neighborhoods between neighboring blocks is described with respect to the table of Fig. 35 may be varied. Further, the pixel samples of the two interlaced fields could be arranged in another way than described above.
  • FIG. 40 shows, in more detail than in Fig. 28, the complete setup of a video encoder engine including an entropy-encoder as it is shown in Fig. 40 in block 2800 in which the aforementioned entropy coding of syntax elements by use of a context assignment based on neighboring syntax elements is used.
  • Fig. 40 shows the basic coding structure for the emerging H.264/AVC standard for a macroblock.
  • the input video signal is, split into macroblocks, each macroblock having 16 x 16 luma pixels.
  • the association of macroblocks to slice groups and slices is selected, and, then, each macroblock of each slice is processed by the network of operating blocks in Fig. 40.
  • I slice A slice in which all macroblocks of the slice are coded using intra prediction.
  • P slice In addition, to the coding types of the I slice, some macroblocks of the P slice can also be coded using inter prediction with at most one motion-compensated prediction signal per prediction block.
  • B slice In addition, to the coding types available in a P slice, some macroblocks of the B slice can also be coded using inter prediction with two motion-compensated prediction signals per prediction block.
  • SP slice A so-called switching P slice that is coded such that efficient switching between different precoded pictures becomes possible.
  • SI slice A so-called switching I slice that allows an exact match of a macroblock in an SP slice for random access and error recovery purposes.
  • Slices are a sequence of macroblocks, which are processed in the order of a raster scan when not using flexible macroblock ordering (FMO).
  • a picture maybe split into one or several slices as shown in Fig. 42.
  • a picture is therefore a collection of one or more slices.
  • Slices are self-contained in the sense that given the active sequence and picture parameter sets, their syntax elements can be parsed from the bit stream and the values of the samples in the area of the picture that the slice represents can be correctly decoded without use of data from other slices provided that utilized reference pictures are identical at encoder and decoder. Some information from other slices maybe needed to apply the deblocking filter across slice boundaries.
  • Each slice group is a set of macroblocks defined by a macroblock to slice group map, which is specified by the content of the picture parameter set and some information from slice headers.
  • the macroblock to slice group map consists of a slice group identification number for each macroblock in the picture, specifying which slice group the associated macroblock belongs to.
  • Each slice group can be partitioned into one or more slices, such that a slice is a sequence of macroblocks within the same slice group that is processed in the order of a raster scan within the set of macroblocks of a particular slice group. (The case when FMO is not in use can be viewed as the simple special case of FMO in which the whole picture consists of a single slice group.)
  • a picture can be split into many macroblock-scanning patterns such as interleaved slices, a dispersed macroblock allocation, one or more "foreground” slice groups and a "leftover” slice group, or a checker-board type of mapping.
  • Each macroblock can be transmitted in one of several coding types depending on the slice- coding type.
  • the following types of intra coding are supported, which are denoted as Intra_4 x 4 or Intra_16xl6 together with chroma prediction and I_PCM prediction modes.
  • the Intra_4x4 mode is based on predicting each 4x4 luma block separately and is well suited for coding of parts of a picture with significant detail.
  • the Intra_16xl6 mode does prediction of the whole 16x16 luma block and is more suited for coding very smooth areas of a picture.
  • the I_PCM coding type allows the encoder to simply bypass the prediction and transform coding processes and instead directly send the values of the encoded samples.
  • the I_PCM mode serves the following purposes:
  • intra prediction in H.264/AVC is always conducted in the spatial domain, by referring to the bins of neighboring samples of previously coded blocks which are to the left and/or above the block to be predicted. This may incur error propagation in environments with transmission errors that propagate due to motion compensation into inter-coded macroblocks. Therefore, a constrained intra coding mode can be signaled that allows prediction only from intra-coded neighboring macroblocks.
  • each 4x4 block is predicted from spatially neighboring samples as illustrated on the left-hand side of Fig. 43.
  • the 16 samples of the 4x4 block, which are labeled as a-p, are predicted using prior decoded samples in adjacent blocks labeled as A- Q.
  • For each 4x4 block one of nine prediction modes can be utilized.
  • eight directional prediction modes are specified as illustrated on the right-hand side of Fig. 41. Those modes are suitable to predict directional structures in a picture such as edges at various angles.
  • each P macroblock type corresponds to a specific partition of the macroblock into the block shapes used for motion-compensated prediction. Partitions with luma block sizes of 16x16, 16x8, 8x16, and 8x8 samples are supported by the syntax. In case partitions with 8x8 samples are chosen, one additional syntax element for each 8x8 partition is transmitted. This syntax element specifies whether the corresponding 8x8 partition is further partitioned into partitions of 8x4, 4x8, or 4x4 luma samples and corresponding chroma samples.
  • the prediction signal for each predictive-coded MxN luma block is obtained by displacing an area of the corresponding reference picture, which is specified by a translational motion vector and a picture reference index.
  • a maximum of sixteen motion vectors may be transmitted for a single P macroblock.
  • the quantization parameter SliceQP is used for determining the quantization of transform coefficients in H.264/AVC.
  • the parameter can take 52 values. These values are arranged so that an increase of 1 in quantization parameter means an increase of quantization step size by approximately 12 % (an increase of 6 means an increase of quantization step size by exactly a factor of 2). It can be noticed that a change of step size by approximately 12 % also means roughly a reduction of bit rate by approximately 12 %.
  • the quantized transform coefficients of a block generally are scanned in a zig-zag fashion and transmitted using entropy coding methods.
  • the 2x2 DC coefficients of the chroma component are scanned in raster-scan order. All inverse transform operations in H.264/AVC can be implemented using only additions and bit-shifting operations of 16-bit integer values. Similarly, only 16-bit memory accesses are needed for a good implementation of the forward transform and quantization process in the encoder.
  • the entropy encoder 2800 in Fig. 40 in accordance with a coding arrangement described above with respect to Fig. 29.
  • a context modeler feeds a context model, i.e., a probability information, to an entropy encoder, which is also referred to as the regular coding engine.
  • the to be encoded bit i.e. a bin
  • This bin value is also fed back to the context modeler so that a context model update can be obtained.
  • a bypass branch is provided, which includes an entropy encoder, which is also called the bypass coding engine.
  • the bypass coding engine is operative to entropy encode the input bin values.
  • the bypass coding engine is not an adaptive coding engine but works preferably with a fixed probability model without any context adaption.
  • a selection of the two branches can be obtained by means of switches.
  • the binarizer device is operative to binarize non-binary valued syntax elements for obtaining a bin string, i.e., a string of binary values. In case the syntax element is already a binary value syntax element, the binarizer is bypassed.
  • bypass coding mode is chosen for selected bins in order to allow a speedup of the whole encoding (and decoding) process by means of a simplified coding engine without the usage of an explicitly assigned model.
  • This mode is especially effective when coding the bins of the primary suffix of those syntax elements, concerning components of differences of motion vectors and transform coefficient levels.
  • the three main functional building blocks which are binarization, context modeling, and binary entropy coding in the encoder of Fig. 40, along with their
  • example 4 is further extended and Fig. 34, 35, 46 - 49.
  • binarizing a data value is performed by binarizing the minimum of the data value and a predetermined cut-off value in accordance with a first binarization scheme, in order to yield a primary prefix, and, if the data value is greater than the cut-off value, binarizing a difference of the data value minus the predetermined cut-off value in accordance with a second binarization scheme to obtain a binary suffix, the first binarization scheme being different from the second binarization scheme, and appending the primary suffix to the primary prefix.
  • a data value is recovered from a binarized representation of the data value, the binarized representation of the data value being a codeword having a primary prefix, which is a binarization of the minimum of the data value and a predetermined cut-off value in accordance with a first binarization scheme and, if the data value is greater than the predetermined cut-off value, a primary suffix appended to the primary prefix, the primary suffix being a binarization of the difference of the data value minus the predetermined cut-off value in accordance with a second binarization scheme, the method comprising extracting, from the primary prefix, the minimum, and, if the minimum is equal to the cut-off value, extracting, from the primary suffix, the difference from the data value minus the predetermined cut-off value . Then, the predetermined cut-off value is added to the difference, to obtain the data value, and, if the minimum is smaller than the cut-off value, the minimum is regarded as the data value.
  • a coding of a data value into a coded bit stream comprises binarizing the minimum of the data value and a predetermined cut-off value in accordance with a first binarization scheme, in order to yield a primary prefix, and, if the data value is greater than the cut-off value, binarizing a difference of the data value minus the predetermined cut-off value in accordance with a second binarization scheme to obtain a binary suffix, the first binarization scheme being different from the second binarization scheme, and appending the primary suffix to the primary prefix.
  • binary entropy coding the bit by means of a adaptively varying bit value probability estimation is performed according to any of the above embodiments with probability estimation dependent partial bitstream selection, or, if the bit of the code word is part of the primary suffix, binary coding the bit by means of a static bit value probability estimation, is performed, thereby obtaining the coded bit stream.
  • Decoding a coded bit stream which represents a binarized representation of the data value is described, the binarized representation of the data value being a codeword having a primary prefix, which is a binarization of the minimum of the data value and a predetermined cut-off value in accordance with a first binarization scheme and, if the data value is greater than the predetermined cut-off value, a primary suffix appended to the primary prefix, the primary suffix being a binarization of the difference of the data value minus the predetermined cut-off value in accordance with a second binarization scheme, the method comprising, for each bit in the codeword, if the bit of the codeword is part of the primary prefix, determining the bit by binary entropy decoding the coded bit stream by means of a adaptively varying bit value probability estimation, namely by retrieving the bit via the partial bitstream associated with the probability estimation, and, if the bit of the codeword is part of the primary suffix, determining the bit by binary decoding the bit by
  • the minimum is extracted from the primary prefix. If the minimum is equal to the cut-off value, the difference from the data value minus the predetermined cut-off value is extracted from the primary suffix and the predetermined cut-off value is added to the difference, to obtain the data value. If the minimum is smaller than the cut-off value, the minimum is regarded as the data value.
  • a very effective compression of data values, and, in particular, of components of motion vector differences or transform coefficient level values, forming syntax elements in a video signal is achieved by using a binarization scheme for preparing the syntax elements for the entropy coding, the binarization scheme substantially being a combination of two different binarization schemes, and by using binary entropy coding instead of m-ary entropy coding for coding the binarized syntax elements.
  • a concatenation of a truncated unary binarization scheme and a k th order exponential Golomb binarization scheme is used.
  • This concatenated scheme which is referred to as unary/k th order Exp-Golomb (UEGk) binarization, is applied to motion vector differences and absolute values of transform coefficient levels.
  • UDGk Exp-Golomb
  • Fig. 44 shows a binary code tree corresponding to the binarization scheme later described with respect to Fig. 46.
  • the code tree generally indicated with reference number 3150 comprises a root node 3152 and several internal nodes 3154, the root node 3152 and the internal nodes 3154 being designated by Cj, with the index i being an integer associated with a corresponding node.
  • Each of these nodes defines a conditional branch or binary decision called a "bin", with a left branch corresponding to the bin value 0 and the right branch corresponding to the bin value 1.
  • the tree 3150 also includes terminal nodes 3156.
  • the numbers assigned to the terminal nodes 3156 in Fig. 44 correspond to the data value of the syntax element, which, in accordance with the binarization scheme of Fig. 46, corresponds to the codeword or bit string resulting form the bits or bin values when following the path from root node 3152 to the respective terminal node 3156.
  • the data value 17 corresponds, as shown in Fig. 46, to the codeword bit string 11111111111111101.
  • the path comprises root node 3152 and the internal nodes 3154 C 2 to C 1 .
  • Each node 3152 or 3154 may be seen as a binary decision called "bin".
  • the concatenation of the binary decision for traversing the tree 3150 from the root node 3152 to the respective terminal node 3156 represents the "bin string" of the corresponding symbol value or syntax element value.
  • Each binary decision has an actual probability distribution indicating the probability as to whether the decision yields a binary zero (left branch in Fig. 44) and a binary one (right branch in Fig. 44).
  • the actual probability distribution of a binary decision defines the probability that the syntax element path takes the zero or one branch at this binary decision. Accordingly, the actual probability for the syntax element to assume a certain data value is equal to the product of the respective probabilities at the nodes arranged along the path from root node 3152 to the respective terminal node 3156.
  • each binary decision or node 3152, 3154 of tree 3150 i.e. each bin
  • a bit value probability estimation or a context model i.e. Q.
  • Q a context model
  • not all of the nodes or bins is associated with adaptive probability estimation models but may be associated with a static probability estimation, thereby reducing the context modeling overhead, i.e. the management for adapting the estimation according to past bins in order to increase the compression ratio.
  • Fig. 45 shows schematically the steps binarizer 3102 performs on the non-binary valued syntax elements.
  • Fig. 45 is an embodiment for binarizing the absolute values of transform coefficient levels, i.e., an integer quantity representing the value associated with a particular two-dimensional frequency index in the decoding process prior to scaling for computation of a transform coefficient value.
  • the binarization scheme of Fig. 45 is a concatenation of a truncated unary binarization scheme, on the one hand, and a k th order Exp-Golomb binarization scheme, on the other hand.
  • a truncated unary binarization scheme is based a unary binarization scheme.
  • the unary codeword to which x is mapped consists of x "1" bits plus a terminating "0" bit.
  • a data value x is mapped to a codeword consisting of a prefix and a suffix.
  • the absolute data value x to be binarized is compared to the cutoff value S at 3200 to determine the minimum of x and S.
  • S is the cut-off value defining for which absolute data values x the truncated unary code is to be used exclusively, i.e. without using the Golomb binarization scheme for the residual.
  • the truncated unary binarization is used exclusively in order to map the absolute data value x to a codeword, the codeword accordingly merely consisting of a primary prefix.
  • a primary suffix is appended to the primary prefix, the primary suffix being created in accordance with a zero order Exp-Golomb binarization scheme from x-S, as will be described further below.
  • the minimum between x and S is binarized in accordance with the truncated unary binarization scheme at 3202.
  • the result is a primary prefix 204, which forms at least a part of the final codeword 3206 to which x is mapped to.
  • the primary prefix thus consists of a unary binarization of the value of Min(S, Abs(x)), where Min(a,b) denotes the minimum of values a and b, and where Abs(x) denotes the absolute value of x.
  • primary prefix consists of S ones if x is greater than or equal to S, and consists of x leading ones followed by a terminating zero if x is less tha S.
  • the absolute data value x is compared at 3208 to the cut-off value S. If a comparison yields that x is smaller than S, the binarization ends at 3210 and, therefore, the codeword 3206, the absolute data value x is mapped to in accordance with the binarization scheme of Fig. 45, just consists of the primary prefix 3204. Otherwise, i.e., if x is equal to or greater than S, the cutoff value S is subtracted from the absolute data value x at 3212. The resulting difference x minus S, sometimes called a residual, is processed at 3214 in accordance with a k th order Exp- Golomb binarization scheme, with k equal to zero. The k th order Exp-Golomb binarization scheme will later be described in more detail with respect to Fig. 48.
  • the result of the binarization 3214 of the difference "x minus S" is appended to the primary prefix 3204 as the primary suffix 3208. Therefore, if the absolute data value x is greater than or equal to the cut-off value S the resulting codeword 206 consists of a primary prefix 3204 followed by a primary suffix 3208, wherein the primary suffix 3208 is a k th order Exp- Golomb binarization of x - S.
  • the primary suffix 3208 itself consists of a prefix and a suffix, hereinafter called a secondary prefix 3218 and a secondary suffix 3220.
  • the codeword or bin string, an absolute value is mapped to comprises at least a TU prefix, the TU prefix part of the codewords or bin strings being highlighted in Fig. 46 at 3254.
  • each EGO suffix 3256 comprises a secondary prefix which is on the left side of dotted line 3258, and all EGO suffixes 3256, except the one of codeword belonging to absolute value 15, comprise a secondary suffix arranged to the right side of dotted line 3258.
  • the elements of the table of Fig. 47 being equal to those of Fig. 46 are indicated by the same reference numbers as in Fig. 46, and a repetitive description of these elements is omitted.
  • the sign bit is not shown in table Fig. 47.
  • the prefix part always corresponds to a maximum error component of +/- 2 samples.
  • the suffix code words are given such that a geometrical increase of the prediction error in units of 2 samples is captured by a linear increase in the corresponding suffix code word length.
  • auxiliary parameter k is initialised to the order of the Exp-Golomb binarization used.
  • k is initialized in step 3300 to the value of 3.
  • a comparison in line 3302 between the absolute value of the motion vector difference (mvd) and the cut-off value, here 9, reveals that the absolute value of mvd is greater than or equal to 9, in a step 3304, the cut-off value S 9 is subtracted from the absolute value of mvd to obtain sufS. Furthermore, a parameter stopLoop is initialized in step 3306 to the value of zero. Next, in 3308, a conditional if-operation is performed as long as the stopLoop-parameter has not changed his value to a positive value greater than 0, as is checked in line 310.
  • sufS is smaller than the k th power of 2 (3312e)
  • a bit having the bit value of zero is output (3312f)
  • a binary representation of sufS using ' a number of bits equal to the actual value of k (3312g) is output (3312h)
  • the parameter stopLoop is set to a value of 1 (3312e) thereby resulting in stopping the iterative execution of the if-operation 3308.
  • All steps 3304 to 3312 are performed if the comparison in step 3302 reveals that the absolute value of mvd is greater than or equal to 9. Irrespective of the comparison in 3302, at 3314 an a sign bit of zero is output if mvd is greater than zero and an a sign bit of 1 is output if mvd is smaller than zero, the sign bit not being shown in Fig. 47.
  • the suffix part at least comprises four bits, one being the zero bit of 3312f, and the other three being the bits of the binary representation output in 3312h, as can be seen also in Fig. 47, neglecting the sign bit at the end of each codeword.
  • Fig. 49 shows as an embodiment of the present invention, the steps performed on decoder side in order to decode the entropy coded bit stream.
  • the decoder receiving the entropy coded bit stream determines whether the next bin or bit to be determined from the coded bit stream is an primary prefix bit or an primary suffix bit.
  • the decoder knows from the way of the predetermined syntax of the entropy coded bit stream, as to whether the next bin or bit expected is a bit of a primary prefix or a primary suffix of a binarized syntax element.
  • the decoder determines the bit value for this bit by binary entropy decoding based on an adaptively varying probability estimation or context model. That is, the decoder selects the respective partial bit stream described above, using the respective probability estimation for selection. The decoder updates this adaptively varying probability estimation based on past decoded bins indicated by the respective context model type.
  • the next bit or bin value is determined by performing entropy decoding based on a static probability model. This may involve a 1:1 mapping of codewords to bins including the actual bit.
  • the decoder repeats, 3408, the steps 3400 to 3404 if there are bits left to achieve the bit values for the bits of a whole codeword. If there are no bits left at step 3408, the decoding process ends at step 3410.
  • the decoder knows when the binary representation of a syntax element ends, i.e. which decoded bit is the last of a actual syntax element, from following the respective binarization code tree, such as the code tree shown in Fig. 44 in case of the syntax element being a absolute transform coefficient level minus 1.
  • Fig. 50 shows the process of recovering from the codeword as obtained by the process of Fig. 49 the data value, which has been binarized into the codeword.
  • the bit values as obtained by the process of Fig. 49 form a bit stream 3450.
  • the decoder knows, as described above, when a new code word is on the bit stream 3450, the instance of a new codeword being indicated by way of a signal 3452.
  • a value m is extracted from the binary code word by counting the leading ones in the primary prefix of the code word, the count being equal to m. If m is equal to S, what is checked in step 3456 (the value of S is known to the decoder because he knows which syntax element is coded by which binarization scheme), in a difference extracting step 3458 the bits following to the bits of the primary prefix from the bit stream 3450 are received, which form the primary suffix part of the code word. In the difference extracting step 3458 the difference between the data value x to be recovered and the cut-off value S is recovered from the primary suffix of the code word in the bit stream 3450. In step 3460, S is added to the difference x-S recovered in step 3358 to obtain the value x, the value for the actually encoded syntax element.
  • a step 3500 the difference extractor counts the leading bits in the primary suffix to obtain a value m.
  • step 3508 the next k plus m following bits are read from the bit stream 3450, i.e. the bits following the terminating bit of the secondary prefix.
  • k is not equal to zero and/or m is not equal to zero
  • step 3508 the next k plus m following bits are read from the bit stream 3450, i.e. the bits following the terminating bit of the secondary prefix.
  • example 4 is further extended and Fig. 34, 35, 36, 37a,b, 39, 45, 46 - 49 also apply.
  • a coding scheme for syntax elements contained in a pre-coded video signal is provided, which enables a higher compression effectiveness.
  • a method for encoding a syntax element contained in a precoded video signal into a coded bit stream is described, the precoded video signal representing at least one video frame, the syntax element being associated with a predetermined portion of the video frame and being indicative of as to whether the predetermined portion of the video frame is precoded in a first or a second way into the precoded video signal, the method comprising investigating as to whether a neighboring portion of the video frame neighboring the predetermined portion is precoded in the first way or the second way, in order to obtain a binary value; assigning one of at least two context models to the predetermined portion of the video frame based on the binary value, wherein each context model is associated with a different probability estimation; Finally, an encoding of a sequence of symbols take place in accordance with any of the above embodiments, with at least one symbol relating to a representation of the syntax element.
  • the assigner is configured to determine the measure for the estimate of the probability distribution for the respective symbol based on the probability distribution estimate associated with the context assigned to the
  • this aspect is achieved by a method for decoding a syntax element from a coded bit stream, the syntax element being contained in a precoded video signal, the precoded video signal representing at least one video frame, the syntax element being associated with a predetermined portion of the video frame and being indicative of as to whether the predetermined portion of the video frame is precoded in a first or a second way into the precoded video signal, the method comprising investigating as to whether a neighboring portion of the video frame neighboring the predetermined portion is precoded in the first way or the second way, in order to obtain a binary value; assigning one of at least two context models to the predetermined portion of the video frame based on the binary value, wherein each context model is associated with a different probability estimation; and reconstructing a sequence of symbols among which at least one symbol relates to a representation of the syntax element in accordance with any of the above embodiment is performed.
  • the assigner is configured to determine the measure for the estimate of the probability distribution for the respective symbol
  • the present example is based on the finding that the coding of a pre-coded video signal may be rendered more compressive, when a syntax element associated with a pre-determined portion of the video frame or picture and being indicative of as to whether the pre-determined portion of the video frame is pre-coded in a first or a second way, i.e., a control information syntax element, is assigned to one of at least two context models, the assignment being dependent on a neighbouring portion of the video frame being pre-coded in the first or second way.
  • the syntax element may then be entropy encoded into the coded bit stream based on the probability estimation with which the one context model is associated.
  • the syntax element of a specific type may be assigned to different ones of the at least two context models each time it occurs in the pre-coded video signal.
  • the selection of the context model is performed each time the syntax element occurs and is dependent on as to whether the neighbouring portion of the video frame is pre-coded in the first or the second way.
  • all the syntax elements of this type occurring in the precoded video signal are grouped into a few context model groups, each syntax element belonging to one context model group, each context model group being assigned to a different one of the context models, and each context model being associated with a different probability estimation. Therefore, each group is entropy coded based on a different probability estimation, namely the one belonging to the context model it is assigned to.
  • the probability estimation of each context model is adapted to the previously coded syntax elements, the adaption being performed separately for each context model.
  • each syntax element just having been entropy coded is used to adapt the probability estimation of the context model to which this syntax element is assigned.
  • following syntax elements of the same type and assigned to the same context model are entropy coded by means of an adapted probability estimation. Since, by this measure, the context models are adapted or managed independently of each other, the adaptation yields a better approximation to the actual or real probability estimation and therefore results in a better compression ratio.
  • the spatially dependent context model selection is used for coding the frame/field coding mode flags in a pre-coded video signal.
  • the frame/field coding mode flags indicate as to whether the respective video frame portion is coded in frame or field coding mode.
  • the video frame is an interlaced frame where some of its picture samples are captured at one time instant while the other picture samples are captured at a different time instant, the first picture samples being called a first field and the second picture samples being called a second field.
  • the video frame is spatially subdivided into an array of macroblock pairs, each macroblock pairs consisting of two macroblocks.
  • one macroblock of a macroblock pair When coded in field mode, one macroblock of a macroblock pair merely comprises the picture samples of the first field or the picture samples of the second field within that macroblock pair, while the other macroblock of that macroblock pair contains or represents the picture samples of the other field within the macroblock pair.
  • the macroblock When coded in frame mode, the macroblock contains both, first and second field picture samples, in particular, the spatially top or bottom ones within a macroblock pair, while the other macroblock of the same macroblock pair represents the other picture samples within the macroblock pair.
  • mb_field_decoding_fiag when the same bin or same syntax element occurs several times in the bins passed from stage 2100a it may be assigned to different of the context models depending on previously incoming or previously entropy coded bins.
  • precoder 4012 When the precoder 4012 receives a video signal representing an interlaced video frame, precoder 4012 is free to make the following decisions when coding the video frame 4010:
  • the choice between the three options can be made adaptively for each frame in a sequence.
  • the choice between the first two options is referred to as picture adaptive frame/field (PAFF) coding.
  • PAFF picture adaptive frame/field
  • a frame consists of mixed regions where some regions are moving and others are not, it is typically more efficient to code the non-moving regions in frame mode and the moving regions in the field mode. Therefore, the frames/field encoding decision can be made independently for each vertical pair of macroblocks in a frame.
  • This coding option is referred to as macroblock adaptive frame/field (MBAFF) coding. It is assumed in the following that precoder 4012 decides to use just this option. As described above, MBAFF coding allows the precoder to better adapt the coding mode type (filed or frame mode) to the respective areas of scenes. For example, precoder 4012 codes macroblock pairs located at stationary areas of a video scene in frame mode, while coding macroblock pairs lying in areas of a scene showing fast movements in field mode.
  • each macroblock contains frame lines.
  • the top macroblock contains top field lines and the bottom macroblock contains bottom field lines.
  • the frame/field decision for each macroblock pair is made at the macroblock pair level by precoder 4012, i.e. if the top macroblock is field coded same applies for the bottom macroblock within same macroblock pair.
  • Each macroblock of a field macroblock pair is processed very similarly to a macroblock within a field in PAFF coding.
  • some stages of the pre-coding procedure in precoder 4012 such as the prediction of motion vectors, the prediction of intra prediction modes, intra frame sample prediction, deblocking filtering and context modelling in entropy coding and the zick-zack scanning of transform coefficients are modified when compared to the PAFF coding in order to account for this mixture.
  • the pre-coded video signal output by precoder 4012 depends on the type of coding precoder 4012 has decided to use.
  • the pre-coded video signal contains a flag mb_field_decoding_flag for each non-skipped macroblock pair.
  • the flag mb_field_decoding_flag indicates for each macroblock pair it belongs to whether the corresponding macroblocks are coded in frame or field coding mode. On decoder side, this flag is necessary in order to correctly decode the precoded video signal.
  • the flag mb_field_decoding_fiag is zero, whereas the flag is one in the other case.
  • context modeller 4112 determines as to whether the incoming syntax element or bin is the mb_field_decoding_flag. If not, context modeller 4112 switches to another syntax element handling procedure or subroutine at step 302. If the incoming bin is an mb_field_decoding_flag the context modeller 4112 determines in step 304 as to which context model the flag belongs to.
  • Step 4304 is shown in more detail in Fig. 53.
  • context modeller 4112 When determining the context model corresponding to the current mb_field_decoding_flag, context modeller 4112 first computes, instep 4306, the addresses of the top macroblock of the macroblock pair to the left and above the current macroblock pair.
  • context modeller 4112 determines the availability of the top macroblock of the macroblock pair to the left and above the current macroblock pair, respectively.
  • the neighbouring macroblock A to the left is marked as available, when mbAddrA is smaller than 0.
  • MbAddrN with N being either A or B, are the addresses of the top macroblocks of the respective neighbouring macroblock pair in case the macroblock pair does really exist and denotes concurrently the availability status of the respective neighbouring macroblock pair A and B, respectively.
  • the macroblock A or B is only marked as available, when additionally to the above condition the following is true:
  • the macroblock with address mbAddrN belongs to same slice as the current slice.
  • step 4308/4309 it is determined that macroblock N, with N being either A or B, is marked as available, in step 4310 and step 4311, respectively, it is determined as to whether macroblock N (N being equal to A or B) is a macroblock coded in field mode or a macroblock coded in frame mode.
  • macroblock A is coded in field mode
  • condTermflagA is set to 1 in step 4312.
  • condTermflagB is set to 1 in step 4313 if macroblock B is coded in field mode.
  • the flag condTermFlagA is set to 0 in step 4314. Accordingly, when mbAddrB is smaller than 0 (4309) or macroblock B is coded in frame mode (4311), the flag condTermFlagB is set to 0 in step 4315.
  • CondTermFlagA and condTermFlagB are used in step 4316 by context modeller 4112 to compute context index ctxIdxMbField indicating the probability model to be used for binary entropy coding of mb_field_decoding_flag.
  • context modeller 4112 passes the variable ctxIdxMbField (context index macroblock field) or the probability estimation status indexed by CtxIdxMbField along with mb_field_decoding_flag to regular coding engine 4118 which, in turn, is contracted in accordance with any of the above described embodiments. That is, the assigner of the above embodiments stes the parameter for selcting the parial bitstream in accordance with the probability determined vie the context determined, in turn, in step 4304, whereupon the selector sends the bins of ctxIdxMbField to the respective partial bitstream path. In step 4318, the flag mb_field_is then encoded into the respective partial bitstream selcted in accordance with the current probability state of the context model as indexed by ctxIdxMbField.
  • regular coding engine 4118 passes the bin value of mb_field_decoding_flag via path 4117 back to context modeller 4112, whereupon context modeller 4112 adapts, in step 4320, the context model indexed by ctxIdxMbField with respect to its probability estimation state. Thereafter the process of coding the flag mb_field_decoding_flag into the bit stream at the output 4124 ends at 4322.
  • Fig. 54 shows the decoding process performed by the entropy decoder 4400 in more detail. First, in a step 4500, the decoder 4400 checks as to whether the next bin to decode is an mb_field_decoding_flag.
  • decoder 400 switches to another bin handling process (not shown) in step 4502. If yes, decoder 4400 determines in step 4504 the context model to which the flag belongs. Step 4504 is performed in the same way as step 4304 in the encoding process, or as shown in Fig. 53.
  • step 4506 the actual bin, i.e., mb_field_decoding_flag, is obtained from the respective partial bitstream entropy decoder by indexing unsing the actual probability state of the context model as indexed by ctxIdxMbField obtained in step 4504.
  • the result of this step is the value for mb_field_decoding_flag.
  • step 4508 the ctxIdxMbField probability state is adapted or updated, as it was the case in step 4320. Thereafter, the process ends at step 4510.
  • step 4504 yields the same index ctxIdxMbField since the macroblocks are stepped through in the precoded video signal in an appropriate order that guarantees that the flags mb_field_decoding_flag of macroblocks A and B (Fig. 22) have already been previously precoded.
  • the context model can be deduced in step 4504 on the basis of previously decoded flags mb_field_decoding_flag of neighbouring macroblocks as was the case on encoder side.
  • step 4310 the following is noted. Normally, a macroblock is coded into the precoded video signal. Nevertheless, it might be that a macroblock is not coded or skipped because, for instance, the corresponding transform coefficient levels of this macroblock are all smaller than a certain quantization threshold. Thus, for some macroblocks there is no mb_field_decoding_flag available. Thus, it might be that step 4310 cannot be performed by merely investigating the mb_field_decoding_flag of the respective neighbouring macroblock if the latter is a skipped one. When mb_field_decoding_flag is not present for either macroblock of a neighbouring macroblock pair (containing A or B in Fig.
  • top macroblock B by merely switching A to B.
  • the above embodiments are also applicable to other control syntax elements or to other bins.
  • the present invention is not restricted to binary entropy coding but could applied as well to other symbol entropy coding.
  • the subdivisions of the video frame into slices, macroblock pairs, macroblocks, picture elements etc. was for illustrating proposes only, and is not to restrict the scope of the invention to this special case.
  • a method for encoding a video frame or picture comprises dividing up the video frame or the picture in portions of a first type and portions of a second type, wherein the portions of the first type are associated with a first scanning pattern, and the portions of the second type are associated with a second scanning pattern that is different from the first scanning pattern; transforming data corresponding to a predetermined of the portions of the video frame or picture into a two-dimensional array of transform coefficients, wherein a scanning order is defined among the transform coefficients by the scanning pattern of the predetermined portion, the scanning order assigning each transform coefficient a unique scanning position; precoding a predetermined of the transform coefficients in order to obtain a transform data unit; choosing one of a first and a second set of context models, depending on as to whether the predetermined portion is a portion of a type being associated with the first or the second scanning pattern, the first and
  • a decoding a transform data unit or a sub-unit thereof from a coded bit stream is described, the transform data unit being a precoded version of a predetermined transform coefficient of transform coefficients which are the result of a transformation of data corresponding to a predetermined portion of portions of a video frame or picture, the portions being either a portion of a first type being associated with a first scanning pattern or a portion of a second type being associated with a second scanning pattern, the method comprising choosing one of a first and a second set of context models, depending on as to whether the predetermined portion is a portion of a type being associated with the first or the second scanning pattern, the first and the second set being different to each other; assigning one context model of the chosen one of the first and the second set of context models to the transform data unit or the sub-unit thereof based on the scanning position assigned to the predetermined transform coefficient, wherein each context model is associated with a different probability estimation; and decoding the transform data unit or the sub-unit thereof from the coded bit stream
  • the transform data units or sub-units thereof obtained from a video frame or picture may be coded in a more compressive way, when different sets of context models are used for coding transform data units or sub-units thereof related to portions of the video frame or picture associated with a first scanning pattern as compared to transform data units or sub-units thereof related to portions of the video frame or picture associated with a different scanning pattern.
  • the context models can be adapted more precisely to the actual symbol statistics, which may be different between transform data units belonging to the same scanning position due to the different scannmg patterns. Since the actual probability distribution can be approximated more precisely, the compression ratio of the coded bit stream may be enhanced.
  • the transform data units contributing to the update of a probability estimation of that context model show a more uniform actual probability distribution so that the updated probability estimation does more precisely approximate the actual probability distribution.
  • providing different context model sets enables initialising the probability estimation of the context models more precisely to the actual probability distribution.
  • compression ratio of the coded bit stream may further be enhanced.
  • Fig. 55 shows the steps of encoding frame and field macroblocks as they are performed by precoder 2012 and entropy coder 2014 of Fig. 28.
  • Fig. 55 shows the process of encoding the video frame or picture 5010 which has already been divided up into macroblock pairs 5010M and 5010b2 as described with respect to Fig. 34.
  • Fig. 55 exemplarily merely two macroblock pairs 5010M and 5010b2 are shown. It is assumed that macroblock pair 5010M is field coded while macroblock pair 5010b2 is frame coded.
  • Pixel samples belonging to the macroblock pairs 501 Obi and 5010b2 are passed to precoder 5012 where they are processed in step 5270.
  • Step 5270 comprises subtracting from the picture samples of the macroblock pairs 501 Obi and 5010b2 predicted pixel samples and transformation the differences from a spatial domain into a spectrum domain.
  • the results of step 5220 are several two-dimensional arrays of transform coefficients for each macroblock pair 5010M and 5010b2, each two-dimensional array of transformation coefficients belonging to a sub-part of a macroblock of the macroblock pairs 501 Obi and 5010b2, such as a sub-part of 4x4 pixel samples or pixel samples differences.
  • Fig. 55 exemplarily shows one such two- dimensional array of transform coefficients as derived from macroblock 501 Obi indicated 5272 and another two-dimensional array of transform coefficients 5274 as derived from macroblock 5010b2.
  • the arrays 5272 and 5274 are 4x4 arrays.
  • the DCT may be a 4x4 discrete cosine transform or a separable integer transform with similar properties. Inverse- transform mismatches on decoder side may be avoided by defining the transformation matrix such that the inverse transform is defined by exact integer operations.
  • the 4x4 block of pixel samples resulting in the array 5272 of transform coefficients corresponds to the block of pixels 1 to 4, 21 to 24, 41 to 44 and 61 to 64 in the macroblock 2252 as shown in Fig. 31b
  • the 4x4 block of pixel samples which array 5274 is related to comprises the pixel samples 1 to 4, 11 to 14, 21 to 24 and 31 to 34 of macroblock 2250 as shown in Fig. 31a.
  • the spatial relationship between the pixels in the 4x4 blocks to which the transformation of step 5270 is applied is different in field macroblocks and frame macroblocks.
  • the pixel pitch along the columns is doubled in case of frame macroblocks compared to field macroblocks. Therefore, two transform coefficients assuming the same array position in arrays 5272 and 5274 relate to different frequencies. Due to this, as mentioned above, different scanning patterns are used in order to define a scanning order among the transform coefficients in arrays 5272 and 5274. The reason for this is that transform coefficients of lower frequencies are more likely to be zero. Thus, different scanning orders are used in order to 'sort' the transform coefficients in accordance with their frequency they relate to.
  • Fig. 56 shows an example for the scanning patterns used for frame and field macroblocks.
  • a two-dimensional array of transform coefficients is obtained from a frame macroblock as illustrated.
  • a number from 0 to 15 is written, the number indicating the scanning position of the respective transform coefficient.
  • the arrows 5276 represent the scanning pattern.
  • the scanning pattern 5276 used to define a scanning order among the transform coefficients related to a frame coded macroblock are scanned in a zigzag scanning starting at one corner of the array 5274 and running diagonally to the corner arranged diametrically opposite to the starting corner. It is assumed in Fig. 56 that the transform coefficient in the upper left corner of array 5274 having scanning position 0 represents the DC component in column and row direction while the transform coefficient having scanning position 5015 arranged diametrically opposite thereto represents the highest frequency part in column and row direction.
  • Fig. 56 the scanning pattern for transform coefficient arrays 5272 related to field macroblocks as illustrated. Again, each array position of the 4x4 transformation coefficient array 5272 is designated by its scanning position number. The scanning pattern in case of field macroblocks is illustrated by arrows 5278. As can be seen, the scanning pattern 5278 for field coded macroblocks is different to the scanning pattern 5276 for frame coded macroblocks. Both scanning patterns 5276 and 5278 define a scanning order among the transform coefficients 0 to 15.
  • precoder 5012 proceeds in step 5280 with precoding the transform coefficients in arrays 5272 and 5274 into sequences 5282 and 5284 of transform data units. The step is described in more detail with respect to Fig.
  • precoder 5012 sets.significant_coeff_flag for this scanning position to 1. Otherwise, precoder 5012 sets significant_coeff_fiag to zero.
  • the first transform coefficient level at scanning position 0 is 9 so that, as can be seen in the third row in Fig. 57, significant_coeff_flag for this scanning position is 1.
  • precoder 5012 creates syntax elements that specify these transform coefficients.
  • precoder 5012 creates for each transform coefficient from the first scanning position to the scanning position of the last significant transform coefficient a transform data unit composed of either a significant_coeff_flag, if the respective transform coefficient is not significant, and of a significant_coeff_flag, last_significant_coeff_flag and data specifying the value of the respective transform coefficient in case of the transform coefficient being significant. All syntax elements significant_coeff_flag and last_significant_coeff_flag such obtained form the significance map indicating the scanning positions of transform coefficients being significant. Precoder 5012 incorporates sequences 5282 and 5284 into a linear stream of syntax elements, i.e. the precoded video signal 5286. The order of the syntax elements is as follows:
  • the syntax elements corresponding to the significance map are transmitted followed by the values indicating the significant transform coefficients in reverse scanning order.
  • the first syntax element of sequences 5282 and 5284 is the significant_coeff_flag of scanning position 0. If the transform coefficient of the scanning position 0 is significant, the significant_coeff_flag of scanning position 0 is followed by a last_significant_coeff_flag for that scanning position. Thereafter, the significant_coeff_flag of the following scanning position, i.e. scanning position 1, is transmitted and so on.
  • the non-binary values for the non-zero transform coefficients are transmitted in reverse scanning order.
  • the syntax elements in syntax element stream 286 are (beginning with the first syntax elements): 1, 0, 0, 1, 0, 1, 0, 0, 0, 0, 1, 0, 0, 1, 1, 1, -1, 3, -5, 9.
  • the three more syntax elements 5286 is fed to entropy coder 5014 which handles the syntax elements element wise.
  • the following description is now concentrated on the handling of significant_coeff_flag and last_significant_coeff_flag.
  • a flag arrives at entropy coder 5014, it chooses at step 5288 one of two context model sets 5290 and 5292 based on the scanning pattern that was used in order to map the two-dimensional transform coefficient matrix into the sequence of transform data units, which the current syntax element last_signi- ficant_coeff_flag or significant_coeff_flag belongs to.
  • entropy coder 5014 inspects the aforementioned mb_field_decoding_flag of the current macroblock, which the current syntax element belongs to, this flag indicating as to whether the current macroblock is coded in field mode in which case the scanning pattern 5278 has been used, or the current macroblock has been coded in frame mode in which case scanning pattern 5276 has been used.
  • entropy coder 5014 assigns one context model of the chosen one of the context models sets 5290 and 5292 to the current syntax element to be encoded based on its scanning position.
  • the syntax element significant_coeff_flag for instance belonging to a scanning position 2 in sequence 5282 results in a context model of context model set 5290 that is different to the context model assigned to the significant coeff flag at scanning position 2 in sequence 5284.
  • the current syntax element is binary entropy encoded by use of the assigned context model from the chosen context model set. This is done as described in the above embodiments, i.e. be selecting the respective partial bitstream by using the probability estimate of the chosen context.
  • step 5288, 5294 and 5296 performed by entropy coder 5014 from Fig. 55 are explained in more detail.
  • context modeller 5112 determines as to whether the incoming syntax element or bin is a last_significant_coeff_flag. If not, context modeller 5112 determines as to whether the incoming syntax element or bin is a significant_coeff_flag at step 5302. If this is not the case, context modeller 5112 switches to another syntax element handling procedure or subroutine at step 5304. If the current syntax element is a last_significant_coeff_flag, it is checked as to whether this flag belongs to either a frame coded macroblock or field coded macroblock in step 5306.
  • Step 5306 is equal to checking as to whether the scanning position of the transform coefficient to which the last_significant_coeff_flag belongs is determined by scanning pattern 5276 or scanning pattern 5278. The same check is performed in step 5308 if, in step 5302, it is determined that the syntax element is a significant_coeff_flag. Steps 5300, 5302, 5306, and 5308 restrict the possible scenarios at the arrival of a syntax element to 5 possible cases. The first case has already been mentioned and exists when the current syntax element is neither a last_significant_coeff_flag nor a significant_coeff_flag. In this case, another syntax element handling is performed in step 5304.
  • the syntax element is a last_significant_coeff_flag belonging to a field code macroblock in which case, a context index offset value is set to a value offset 1 in step 5310.
  • the current syntax element is a last_significant_coeff_flag belonging to a frame coded macroblock, where an index ctxldx Offset is set to offset2 in step5 312.
  • the syntax element is a significant_coeff_flag belonging to a field coded or frame coded macroblock, where index ctxldx Offset is set to offset3 in step 5340 and to offset4 in step 5316, respectively.
  • the index ctxldx Offset is for pointing into on common list of context models.
  • the values offset 1 to offset4 are different to each other.
  • an index incrementor ctxldxlnc (context index incrementor) is set equal to the scanning position (scanningPos) which the syntax element, i.e., the last_significant_coeff_flag or significant_coeff_flag, belongs to, in step 5318.
  • a context index is determined by the sum of the context index offset ctxldx Offset and the index incrementor ctxldx Inc.
  • the steps 5300 to 5320 are performed by context modeller 5112.
  • the result of step 5320 is a context index ctxidx indicating the probability model to be used for binary entropy coding of the syntax element.
  • context modeller 5112 After the determination of ctxidx, context modeller 5112 passes the variable ctxidx or the probability estimation status indexed by ctxidx along with the syntax element itself, i.e. with last_significant_coeff_flag or significant_coeff_flag, to regular coding engine 118 which, sa outlined above, is confufgured in accordance with any of the above embodiments. Based on these inputs, the regular coding engine 5118 entropy encodes, in step 5322, the syntax element into the bit stream 5124 by using the current probability state of the context model as indexed by ctxidx.
  • regular coding engine 5118 passes the bin value of last_significant_coeff_flag or significant_coeff_flag, respectively, via path 117 back to context modeller 5112, whereupon context modeller 5112 adapts, in step 5324, the context model indexed by ctxidx with respect to its probability estimation state. Thereafter, the process of coding the last_significant_coeff_flag or significant_coeff_flag into the bit stream at the output 5124 ends at 5326.
  • steps 5300 to 5316 belong to step 5288 of Fig. 55, while steps 5318 and 5320 belong to step 5294 and steps 5322 and 5324 belong to step 5296.
  • Fig. 59 shows a pseudo-C code for the parsing process performed in entropy decoder 5400.
  • the parsing process shown in Fig. 59 is appropriate for parsing a significance map along with its corresponding coefficient values for a macroblock provided that the significance map and the coefficient values are coded in the way explained above with respect to Fig. 57, i.e., the significance map following the scanning order and preceding the coefficient values of the significant transform coefficients which are ordered in reverse scanning order.
  • Fig. 59 variables that are written in bold are read from the entropy coded bit stream arriving at entropy decoder 5400.
  • binary entropy decoding is used for reading.
  • the reading process is shown in more detail for bins sigmficant_coeff_flag and a last_significant_coeff_flag in Figs. 60.
  • the parsing process is called residual_block_CABAC as can be seen in line 1 of the pseudo-C code of Fig. 59.
  • the output of the process residual block CABAC is an one-dimensional array coeffLevel indicating the transform coefficient levels for each scanning position like shown in Fig. 57.
  • the process residual block CABAC receives the variable maxNumCoeff which indicates the maximum number of coefficients in a macroblock, in our case being 16.
  • the process then starts in line 2 with reading a flag called coded_block_flag indicating as to whether in the current block all transform coefficient levels are zero. In this case, coded_block_flag is zero otherwise the coded_block_flag is one. In the following it is assumed that coded_block_flag is one so that at least one transform coefficient level is not zero. So, in line 4, a variable NumCoeff is set to maxNumCoeff.
  • the actual value of this transform coefficient is computed in line 20. Then, in the for-loop from lines 21 to 28, the remaining significant transform coefficients are read from the bit stream by use of the significance map as received in lines 4 to 17.
  • the counter of the for-loop i is initialised to NumCoeff-2, i.e., the scanning position immediately preceding the scanning position of the last significant transform coefficient. This counter is decremented after each for-loop-transition. If significant_coeff_flag[i] is one (line 22) the coefficient level of the transform coefficient at scanning position i is determined in lines 23 to 25, these lines being in effect, identical to lines 18 to 20.
  • Fig. 60 now shows the decoding process performed by the entropy decoder 400 each time a bin is to be decoded. Which bin is to be decoded depends on the syntax element which is currently expected by entropy decoder 5400. With regard to the significance map and the corresponding non-zero transform coefficient levels this knowledge results from the parsing process of Fig. 59.
  • step 5500 the decoder 5400 checks as to whether the next bin to decode is a last_significant_coeff_fiag. If this is not the case, the checks in step 5502 as to whether the next bin to decode is a significant_coeff_flag. The next bin to decode is a significant_coeff_flag. If the next bin to decode is either a last_significant_coeff_flag or a significant_coeff_flag, decoder 5400 checks as to whether the flag belongs to a frame coded or field coded macroblock in step 5504 or 5506, respectively.
  • decoder 400 proceeds to another syntax element handling in step 504.
  • the checks of steps 5500, 5502, 5506, and 5508 correspond to steps 5300, 5302, 5306, and 5308 of the encoding process of Fig.
  • ctxldx Offset is set to 1 of offsetl, offset2, offset3, and offset4 in one of steps 5510, 5512, 5514, and 5516, respectively, with steps 5510-5516 corresponding to steps 5310 to 3516 of Fig. 58. Accordingly, ctxldxlnc and ctxldx is determined in steps 5518 and 5520 in the same way as in the encoding process of Fig. 8 in steps 318 and 320 in order to determine the context model to be used in the following entropy decoding.
  • step 5522 the entropy decoder 5400 entropy decodes the actual bin, i.e., last_significant_coeff_flag or significant_coeff_flag, from the entropy coded bit stream by use of the actual probability state of the context module as indexed by ctxldx obtained in steps 5510 to 5520. That is, the respective partial bitstream is selected based on the probability index and the bin retrieved therefrom. The result of this step is the value for the actual bin.
  • step 5524 the ctxldx probability state is adapted or updated, as it was the case in step 5224. Thereafter, the process ends at step 5526.
  • step 5527 i.e. the decoding of mb_filed_decoding_flag, seems to have take place after the decoding of each bin, this is not the case. In fact, this step occurs, at maximum, merely for one macroblock pair.
  • Fig. 61a,b and 62 enable the use of computationally less complex variable length codes in order to entropy-code the individually bitstreams, and effectively cover portions of the probability estimate.
  • the symbols are of binary nature and the VLC codes presented below effectively cover the probability estimate representated by, for example, R LPS , extending within [0;0.5].
  • the embodiments outlined below describe possible implementations for the indivdual entropy coders 10 and decoders 22 in Fig. 1 to 11, respectively. They are suitable for coding of bins, i.e. binary symbols, as they occur in image or video compression applications. Accordingly, these embodiments are also applicable to image or video coding where such binary symbols are splitted-up into the one or more streams of bins 7 to be encoded and bitstreams 24 to be decoded, respectively, where each such bin stream can be considered as a realization of a Bernoulli process.
  • the embodiments described below use one or more of the below-explained various so-called variable-to-variable-codes (v2v-codes) to encode the bin streams.
  • a v2v-code can be considered as two prefix-free codes with the same number of code words.
  • a primary, and a secondary prefix-free code Each code word of the primary prefix-free code is associated with one code word of the secondary prefix-free code.
  • the encoders 10 and decoders 22 operate as follows: To encode a particular sequence of bins 7, whenever a code word of the primary prefix-free code is read from buffer 8, the corresponding code-word of the secondary prefix-free code is written to the bit stream 12. The same procedure is used to decode such a bit stream 24, but with primary and secondary prefix-free code interchanged. That is, to decode a bitstream 24, whenever a code word of the secondary prefix-free code is read from the respective bit stream 24, the corresponding code-word of the primary prefix- free code is written to buffer 20.
  • the codes described below do not necessitate look-up tables.
  • the codes are implementable in form of finite state machines.
  • the v2v-codes presented here can be generated by simple construction rules such that there is no need to store large tables for the code words. Instead, a simple algorithm can be used to carry out encoding or decoding. Three construction rules are described below where two of them can be parameterized. They cover different or even disjoint portions of the afore-mentioned probability interval and are, accordingly, specifically advantegeuous if used together, such as all three codes in parallel (each for different ones of the en/decoders 11 and 22), or two of them.
  • Construction rule 1 'Unary bin pipe' codes or en-/decoders 10 and 22
  • the construction of bin pipe codes is described first.
  • a bin pipe code can be constructed from any prefix-free code with at least three code words. To form a v2v-code, it uses the prefix-free code as primary and secondary code, but with two code words of the secondary prefix-free code interchanged. This means that except for two code words, the bins are written to the bit stream unchanged.
  • a unary bin pipe code is constructed from a special prefix-free code.
  • Construction rule 2 'Unary to rice' codes and Unary to rice en-/decoders 10 and 22:
  • Unary to rice codes use a truncated unary code as primary code. I.e. unary code words are generated starting with T, '01 ', ⁇ ⁇ , ... until 2 n + 1 code words are generated and from the longest code word, the trailing 1 is removed, n is the parameter of the unary to rice code.
  • the secondary prefix-free code is constructed from the code words of the primary prefix-free code as follows. To the primary code word only consisting of 0s, the code word T is assigned. All other code words consist of the concatenation of the code word '0' with the n-bit binary representation of the number of 0s of the corresponding code word of the primary prefix-free code.
  • Construction rule 3 'Three bin' code
  • the three bin code is given as:
  • the primary code symbol sequences
  • the code words are sorted by ascending numbers of Is.
  • three bins are read from the bin stream (i.e.7). If these three bins contain exactly one 1, the code word T is written to the bit stream followed by two bins consisting of the binary representation of the position of the 1 (starting from right with 00). If the three bins contain exactly one 0, the code word ⁇ 1 ⁇ is written to the bit stream followed by two bins consisting of the binary representation of the position of the 0 (starting from the right with 00). The remaining code words '000' and '111' are mapped to '0' and ⁇ 111 , respectively.

Abstract

An encoder for encoding a sequence of symbols is described which comprises an assigner configured to assign a number of parameters to each symbol of the sequence of symbols based on information contained within previous symbols of the sequence of symbols; a plurality of entropy encoders each of which is configured to convert the symbols forwarded to the respective entropy encoder into a respective bitstream; and a selector configured to forward each symbol to a selected one of the plurality of entropy encoders, the selection depending on the number of parameters assigned to the respective symbol.

Description

PROBABILITY INTERVAL PARTIONING ENCODER AND DECODER
The present invention relates to entropy coding and may be used in applications such as, for example, video and audio compression.
The present invention describes a new method and apparatus for entropy encoding and decoding of discrete data. Entropy coding, in general, can be considered as the most generic form of lossless data compression. Lossless compression aims to represent discrete data with fewer bits than needed for the original data representation but without any loss of information. Discrete data can be given in the form of text, graphics, images, video, audio, speech, facsimile, medical data, meteorological data, financial data, or any other form of digital data.
In entropy coding, the specific high-level characteristics of the underlying discrete data source are often neglected. Consequently, any data source is considered to be given as a sequence of source symbols that takes values in a given w-ary alphabet and that is characterized by a corresponding (discrete) probability distribution {pi, ..., pm}. In these abstract settings, the lower bound of any entropy coding method in terms of expected codeword length in bits per symbol is given by the entropy m
#"= -∑A log2 /¾ . (Al)
=1
Huffman codes and arithmetic codes are well-known examples of practical codes capable of approximating the entropy limit (in a certain sense). For a fixed probability distribution, Huffman codes are relatively easy to construct. The most attractive property of Huffman codes is that its implementation can be efficiently realized by the use of variable-length code (VLC) tables. However, when dealing with time- varying source statistics, i.e., changing symbol probabilities, the adaptation of the Huffman code and its corresponding VLC tables is quite demanding, both in terms of algorithmic complexity as well as in terms of implementation costs. Also, in the case of having a dominant alphabet value with j¾ > 0.5, the redundancy of the corresponding Huffman code (without using any alphabet extension such as run length coding) may be quite substantial. Another shortcoming of Huffman codes is given by the fact that in case of dealing with higher-order probability modeling, multiple sets of VLC tables may be required. Arithmetic coding, on the other hand, while being substantially more complex than VLC, offers the advantage of a more consistent and adequate handling when coping with adaptive and higher-order probability modeling as well as with the case of highly skewed probability distributions. Actually, this characteristic basically results from the fact that arithmetic coding provides a mechanism, at least conceptually, to map any given value of probability estimate in a more or less direct way to a portion of the resulting codeword. Being provided with such an interface, arithmetic coding allows for a clean separation between the tasks of probability modeling and probability estimation, on the one hand, and the actual entropy coding, i.e., mapping of a symbols to codewords, on the other hand.
For the sake of clarity, we will restrict the exposition in the following to the case of a binary alphabet, although conceptually, the very basic methods of this invention also apply to the general case of an w-ary alphabet with m > 2. The probability distribution of symbols with values in a binary alphabet can be described by a single parameter p = /?LPSJ where the useful distinction is made between the so-called less probable symbol (LPS) with probability estimates />LPS≤ 0-5 and the more probable symbol (MPS) with PMPS = 1 ~PLPS- Thus, in the general case we have 0 < pus≤ 0-5, where in practical cases, the probability estimate is often bounded from below by a minimum probability value pmm such that pmjn < pws-
Description of the invention
An encoder for encoding a sequence of symbols according to an embodiment of the present invention comprises an assigner configured to assign a number of parameters to each symbol of the sequence of symbols based on information contained within previous symbols of the sequence of symbols; a plurality of entropy encoders each of which is configured to convert the symbols forwarded to the respective entropy encoder into a respective bitstream; and a selector configured to forward each symbol to a selected one of the plurality of entropy encoders, the selection depending on the number of parameters assigned to the respective symbol.
A decoder for reconstructing a sequence of symbols according to an embodiment of the present invention comprises a plurality of entropy decoders, each of which is configured to convert a respective bitstream into symbols; an assigner configured to assign a number of parameters to each symbol of a sequence of symbols to be reconstructed based on information contained within previously reconstructed symbols of the sequence of symbols; and a selector configured to retrieve each symbol of the sequence of symbols to be reconstructed from a selected one of the plurality of entropy decoders, the selection depending on the number of parameters defined to the respective symbol.
Embodiments of an entropy encoding algorithm are described in which a sequence of arbitrary source symbols is mapped onto two or more partial bitstreams, and a decoding algorithm, in which the original source symbols are decoded from the two or more partial bitstreams. At the encoder side, the source symbols are first binarized (if they don't already represent binary symbols) and each bin of the binarizations is associated with a set of parameters. The associated parameters are then mapped onto a small set of indices and all bins that are associated with a particular index are coded by a particular binary entropy encoder and the corresponding codeword or codewords are written to a particular partial bitstream. At the decoder side, the source symbols are decoded by requests for source symbols. The same binarization scheme as for the encoder is used to convert these request for source symbols into requests for bins, and each request for a bin is associated with the same set of parameters as the corresponding bin at the encoder side. The associated parameters are again mapped onto a small set of indices, where the mapping is the same as at the encoder side. And all bin requests that are associated with a particular index are decoded by a particular binary entropy decoder, which reads a codeword or multiple codewords from the corresponding partial bitstream. The encoders and decoders are configured in a way that different binary entropy encoders and decoders use different coding algorithms (i.e., a different mapping between bin sequences and codewords).
In a preferred embodiment of the invention, the set of associated parameters represents a measure for the probability of one of the two bin values for a particular bin (or a set of equivalent measures). The measure for the probability can for example represent a probability state, which can be represented by a value of a set of N values. The probability estimation is decoupled from the actual coding of the bins. The probability measure is then mapped onto a probability index, by which the encoding and decoding of a bin is assigned to a particular binary entropy encoder and decoder. The mapping of the probability measures to probability indexes (and thus the clustering into groups of probability measures) can also be varied over time, e.g. in dependence of already transmitted symbols. For instance, the assignment can be adapted to the number of bins that have been coded with a certain probability measure or probability index in order to ensure that the created partial bitstreams have similar bit rates. In a preferred embodiment of the invention, the number of binary entropy encoders and decoders is less than the number of possible values for the probability measure. A large number of possible probability measures, which allows an accurate estimation of the associated probability, is mapped onto a small number of probability indices, where each of those is associated with a particular binary entropy encoder and decoder. The small number of binary entropy encoders and decoders provides the advantage that the overhead for transmitting the partial bitstream can be kept small and that the synchronization overhead is of encoding and decoding system can be kept small. When using a probability measure as basis for assigning the bins to particular binary entropy encoders and decoders, the binary encoders and decoders can be particularly optimized for a representative of the group of probability measures, which allows a high coding efficiency (similar to the best known entropy coding schemes). In addition, simple encoding and decoding algorithms, which may consist of a simple mapping of bin sequences to codewords and vice versa, can be designed for the representatives of the groups of probability measure. This reduces the complexity of the encoding and decoding system, while still providing a high coding efficiency. It this point, it should be noted that a high granularity of probability measure is required for an accurate estimation of the bin probabilities; any inaccuracy would have a negative impact on the coding efficiency. But for coding the bins, a relatively small number (significantly smaller than the number of possible probability measure values for the probability estimation) of probability groups is sufficient, since the rate increase that results from encoding a bin with a particular associated probability measure by using an entropy coding algorithm that was designed for a probability that is different than but close to the particular probability is very small. And using a small number of binary entropy encoders and decoders provides the above mentioned advantages. In further embodiments of the invention, the set of parameters that is associated with a bin consists of a measure for the probability of one of the two bin values and one or more additional parameters.
In the following, preferred embodiments of the invention is described in more detail referring to the figures and various preferred embodiments of the invention, which provide advantages in respect to several design aspects, are highlighted. Among the figures,
Fig. 1 shows a bock diagram of an encoder according to an embodiment;
Fig. 2 shows a bock diagram of an decoder suitable for decoding bitstream generated by the encoder of Fig. 1, according to an embodiment;
Fig. 3 shows a schematic diagram illustrating a data packet with multiplexed partial bitstreams according to an embodiment;
Fig. 4: shows a schematic diagram illustrating a data packet with an alternative segmentation using fixed-size segments according to a further embodiment;
Fig. 5 shows a bock diagram of an encoder according to an embodiment using partial bitstream interleaving;
Fig. 6 shows a schematic illustrating examples for the status of a codeword buffer at the encoder side of Fig. 5 according to an embodiment;
Fig. 7 shows a bock diagram of a decoder according to an embodiment using partial bitstream interleaving;
Fig. 8 shows a bock diagram of a decoder according to an embodiment using codeword interleaving using a single set of codewords;
Fig. 9 shows a bock diagram of an encoder according to an embodiment using interleaving of fixed-length bit sequences;
Fig. 10 shows a schematic illustrating examples for the status of a global bit buffer at the encoder side of Fig. 9 according to an embodiment;
Fig. 11 shows a bock diagram of a decoder according to an embodiment using interleaving of fixed-length bit sequences;
Fig. 12 shows a graph for illustrating an optimal probability interval discretization into
K = 4 intervals assuming a uniform probability distribution in ;
Fig. 13 shows a schematic diagram illustrating a tree of binary events for an LPB probability of P ~~ and an associated variable length code obtained by the Huffman algorithm;
Fig. 14 shows a graph from which the relative bit rate increase for optimal codes C given a maximum number of table entries ^m may be gathered;
Fig. 15 shows a graph illustrating the rate increase for the theorectically optimal probability interval partitioning into K = \2 intervals (cp. sec. 3) and a real design with
VNB2VLC codes with a maximum number of ^m ^ table entries;
Fig. 16 shows a schematic diagram illustrating an example for conversion of a ternary choice tree into a full binary choice tree;
Fig. 17 shows a block diagram of a system comprising an encoder (left part) and decoder (right part) according to an embodiment;
Fig. 18 is an illustration of the basic principle of coding transform coefficients according to an embodiment using the entropy coding according to any of the embodiments of Fig. 1 1 17,
Fig. 19 shows two examples for coding the significance mapping (the marked symbols are not transferred),
Fig. 20 shows binarization for the magnitudes of the transform coefficients (ABS),
Fig. 21 shows block types and their classification for the H.264/AVC standard,
Fig. 22 shows context modeling for the one-bit symbol CBP4, and
Fig. 23 shows examples of the context modeling for coding the magnitudes of the significant transform coefficients.
Fig. 24 shows a block diagram of an entropy encoder arrangement according to an application example;
Fig. 25 shows a detailed diagram of the initializer from Fig. 24;
Fig. 26 shows a sequence of steps for calculating a reference index to reference an initialization probability information table including probability information for the least probable symbol (LPS) and a value of the corresponding most probable symbol (MPS) as initialization probability information;
Fig. 27 shows a schematic block diagram of an entropy decoder arrangement according to an application example;
Fig. 28 shows a high-level block diagram of a coding environment;
Fig. 29 shows a block diagram of the entropy coding part of the coding environment of
Fig. 28; Fig. 30 shows a schematic diagram illustrating the spatial subdivision of a picture or video frame into macroblock pairsn;
Fig. 31a shows a schematic diagram illustrating the frame mode, in accordance with an embodiment of the present invention.
Fig. 31b shows a schematic diagram illustrating the field mode, in accordance with an embodiment of the present invention.
Fig. 32 shows a flow diagram illustrating the encoding of syntax elements with context assignments based on neighboring syntax elements in accordance with an embodiment of the present invention.
Fig. 33 shows a flow diagram illustrating the binary entropy coding of the syntax elements based on the context model to which it is assigned in accordance with an embodiment of the present invention.
Fig. 34 shows a schematic diagram illustrating the addressing scheme of the macroblocks in accordance with an embodiment of the present invention.
Fig. 35 shows a table illustrating how to obtain the macroblock address mbAddrN indicating the macroblock containing a sample having coordinates xN and yN relative to the upper-left sample of a current macroblock and, additionally, the y coordinate yM for the sample in the macroblock mbAddrN for that sample, dependent on the sample being arranged beyond the top or the left border of the current macroblock, the current macroblock being frame or field coded, and the current macroblock being the top or the bottom macroblock of the current macroblock pair, and, eventually, the macroblock mbAddrA being frame or field coded and the line in which the sample lies having an odd or even line number yN.
Fig. 36 shows a schematic illustrating macroblock partitions, sub-macroblock partitions, macroblock partitions scans, and sub-macroblock partition scans.
Fig. 37 shows a high-level block diagram of a decoding environment in which the present invention may be employed.
Fig. 38 shows a flow diagram illustrating the decoding of the syntax elements coded as shown in Fig. 32 and 33 from the coded bit stream;
Fig. 39 shows a flow diagram illustrating the entropy decoding process and the decoding process of Fig. 3;
Fig. 40 shows a basic coding structure for the emerging H.264/AVC video encoder for a macroblock.
Fig. 41 illustrates a context template consisting of two neighboring syntax elements A and B to the left and on the top of the current syntax element C.
Fig. 42 shows an illustration of the subdivision of a picture into slices.
Fig. 43 shows, to the left, intra_4x4 prediction conducted for samples a-p of a block using samples A_Q, and to the right, "prediction directions for intra_4x4 prediction.
Fig. 44 a part of the binarization coding tree related to the binarization scheme Fig. 46; Fig. 45 a schematic diagram illustrating the binarization of an absolute data value;
Fig. 46 table showing bin strings into which an absolute value is binarized;
Fig. 47 table showing bin strings into which an absolute value is binarized;
Fig. 48 shows a pseudo-C code for performing a binarization;
Fig. 49 shows a schematic diagram illustrating the decoding of an entropy coded bit stream into a data value;
Fig. 50 shows a schematic diagram illustrating the recovering of a data value from a binarization of the data value;
Fig. 51 shows a schematic diagram illustrating the extraction with regard to the suffix part in the process of Fig. 50;
Fig. 52 shows a flow diagram illustrating the encoding of the syntax element mb_field_decoding_flag;
Fig. 53 flow diagram illustrating a process of assigning context models to the mb_field_decoding_flags ;
Fig. 54 shows a flow diagram illustrating the decoding of the syntax element mb_field_decoding_flag from the coded bit stream as derived by the encoding scheme of Fig. 52;
Fig. 55 shows a schematic flow diagram illustrating the encoding of a video frame or picture;
Fig. 56 illustrating the different scanning patterns used for field coded macroblocks and frame coded macroblocks.
Fig. 57 shows a table illustrating a significance map as obtained from the exemplarily chosen transform coefficient levels;
Fig. 58 shows a flow diagram illustrating the encoding of the syntax elements last_significant_coeff_flag and significant_coeff_flag in accordance with an embodiment of the present invention;
Fig. 59 shows a pseudo C-code illustrating the parsing process on decoder side;
Fig. 60 shows a flow diagram illustrating the decoding of the syntax elements significant_coeff_flag and last_significant_coeff_flag from the coded bit stream as derived by the encoding scheme of Fig. 58;
Fig. 61a,b schematically show the process of decoding a run of MPS for an entropy encoder/decoder pair in accordance with an embodiment; and
Fig. 62 schematically shows an example for the quantization of the interval width according to an embodiment.
It is noted that during the description of the figures, elements occurring in several of these Figures are indicated with the same reference sign in each of these Figures and a repeated description of these elements as far as the functionality is concerned is avoided in order to avoid unnecessary repetitions. Nevertheless, the functionalities and descriptions provided with respect to one figure shall also apply to other Figures unless the opposite is explicitly indicated.
An encoder according to an embodiment of the invention is illustrated in Fig. 1. The encoder losslessly converts a stream of source symbols 1 into a set of two or more partial bitstreams 12. In a preferred embodiment of the invention, each source symbol 1 is associated with a category of a set of one or more categories. As an example, the categories can specify the type of the source symbol. In the context of hybrid video coding, a separate category may be associated with macroblock coding modes, block coding modes, reference picture indices, motion vector differences, subdivision flags, coded block flags, quantization parameters, transform coefficient levels, etc. In other application areas such as audio, speech, text, document, or general data coding, different categorizations of source symbols are possible.
In general, each source symbol can take a value of a finite or countable infinite set of values, where the set of possible source symbol values can differ for different source symbol categories. For reducing the complexity of the encoding and decoding algorithm and for allowing a general encoding and decoding design for different source symbols and source symbol categories, the source symbols 1 are converted into ordered sets of binary decisions and these binary decisions are then processed by simple binary coding algorithms. Therefore, the binarizer 2 bijectively maps the value of each source symbol 1 onto a sequence (or string) of bins 3. The sequence of bins 3 represents a set of ordered binary decisions. Each bin 3 or binary decision can take one value of a set of two values, e.g. one of the values 0 and 1. The binarization scheme can be different for different source symbol categories. The binarization scheme for a particular source symbol category can depend on the set of possible source symbol values and/or other properties of the source symbols for the particular category. Table 1 illustrates three example binarization schemes for countable infinite sets. Binarization schemes for countable infinite sets can also be applied for finite sets of symbol values. In particular for large finite sets of symbols values, the inefficiency (resulting from unused sequences of bins) can be negligible, but the universality of such binarization schemes provides an advantage in terms of complexity and memory requirements. For small finite sets of symbol values, it is often preferable (in terms of coding efficiency) to adapt the binarization scheme to the number of possible symbol values. Table 2 illustrates three example binarization schemes for finite sets of 8 values. Binarization schemes for finite sets can be derived from the universal binarization schemes for countable infinite sets by modifying some sequences of bins in a way that the finite sets of bin sequences represent a redundancy-free code (and potentially reordering the bin sequences). As an example, the truncated unary binarization scheme in Table 2 was created by modifying the bin sequence for the source symbol 7 of the universal unary binarization (see Table 1). The truncated and reordered Exp-Golomb binarization of order 0 in Table 2 was created by modifying the bin sequence for the source symbol 7 of the universal Exp-Golomb order 0 binarization (see Table 1) and by reordering the bin sequences (the truncated bin sequence for symbol 7 was assigned to symbol 1). For finite sets of symbols, it is also possible to use non-systematic / non-universal binarization schemes, as exemplified in the last column of Table 2.
Table 1 : Binarization examples for countable infinite sets (or large finite sets).
Figure imgf000011_0001
Table 2: Binarization examples for finite sets.
Figure imgf000011_0002
Each bin 3 of the sequence of bins created by the binarizer 2 is fed into the parameter assigner 4 in sequential order. The parameter assigner assigns a set of one or more parameters to each bin 3 and outputs the bin with the associated set of parameters 5. The set of parameters is determined in exactly the same way at encoder and decoder. The set of parameters may consist of one or more of the following parameters: - a measure for an estimate of the probability for one of the two possible bin values for the current bin,
- a measure for an estimate of the probability for the less probable or more probable bin value for the current bin,
- an identifier specifying an estimate for which of the two possible bin values represents the less probable or more probable bin value for the current bin,
- the category of the associated source symbol,
- a measure for the importance of the associated source symbol,
- a measure for the location of the associated symbol (e.g. in temporal, spatial, or volumetric data sets),
- an identifier specifying the channel code protection for the bin or the associated source symbol,
- an identifier specifying the encryption scheme for the bin or the associated source symbol,
- an identifier specifying a class for the associated symbol,
- the bin number in the sequence of bins for the associated source symbol.
In a preferred embodiment of the invention, the parameter assigner 4 associates each bin 3,5 with a measure for an estimate of the probability for one of the two possible bin values for the current bin. In a further preferred embodiment of the invention, the parameter assigner 4 associates each bin 3,5 with a measure for an estimate of the probability for the less probable or more probable bin value for the current bin and an identifier specifying an estimate for which of the two possible bin values represents the less probable or more probable bin value for the current bin. It should be noted that the probability for the less probable or more probable bin value and the identifier specifying which of the two possible bin values represents the less probable or more probable bin value are equivalent measures for the probability of one of the two possible bin values.
In a further preferred embodiment of the invention, the parameter assigner 4 associates each bin 3,5 with a measure for an estimate of the probability for one of the two possible bin values for the current bin and one or more further parameters (which may be one or more of the above listed parameters). In a further preferred embodiment of the invention, the parameter assigner 4 associates each bin 3,5 with a measure for an estimate of the probability for the less probable or more probable bin value for the current bin, an identifier specifying an estimate for which of the two possible bin values represents the less probable or more probable bin value for the current bin, and one or more further parameters (which may be one or more of the above listed parameters).
In a preferred embodiment of the invention, the parameter assigner 4 determines one or more of the above mentioned probability measures (measure for an estimate of the probability for one of the two possible bin values for the current bin, measure for an estimate of the probability for the less probable or more probable bin value for the current bin, identifier specifying an estimate for which of the two possible bin values represents the less probable or more probable bin value for the current bin) based on a set of one or more already encoded symbols. The encoded symbols that are used for determining the probability measures can include one or more already encoded symbols of the same symbol category, one or more already encoded symbols of the same symbol category that correspond to data sets (such as blocks or groups of samples) of neighboring spatial and/or temporal locations (in relation to the data set associated with the current source symbol), or one or more already encoded symbols of different symbol categories that correspond to data sets of the same and/or neighboring spatial and/or temporal locations (in relation to the data set associated with the current source symbol).
Each bin with an associated set of parameters 5 that is output of the parameter assigner 4 is fed into a bin buffer selector 6. The bin buffer selector 6 potentially modifies the value of the input bin 5 based on the input bin value and the associated parameters 5 and feds the output bin 7 - with a potentially modified value - into one of two or more bin buffers 8. The bin buffer 8 to which the output bin 7 is sent is determined based on the value of the input bin 5 and/or the value of the associated parameters 5.
In a preferred embodiment of the invention, the bin buffer selector 6 does not modify the value of the bin, i.e., the output bin 7 has always the same value as the input bin 5.
In a further preferred embodiment of the invention, the bin buffer selector 6 determines the output bin value 7 based on the input bin value 5 and the associated measure for an estimate of the probability for one of the two possible bin values for the current bin. In a preferred embodiment of the invention, the output bin value 7 is set equal to the input bin value 5 if the measure for the probability for one of the two possible bin values for the current bin is less than (or less than or equal to) a particular threshold; if the measure for the probability for one of the two possible bin values for the current bin is greater than or equal to (or greater than) a particular threshold, the output bin value 7 is modified (i.e., it is set to the opposite of the input bin value). In a further preferred embodiment of the invention, the output bin value 7 is set equal to the input bin value 5 if the measure for the probability for one of the two possible bin values for the current bin is greater than (or greater than or equal to) a particular threshold; if the measure for the probability for one of the two possible bin values for the current bin is less than or equal to (or less than) a particular threshold, the output bin value 7 is modified (i.e., it is set to the opposite of the input bin value). In a preferred embodiment of the invention, the value of the threshold corresponds to a value of 0.5 for the estimated probability for both possible bin values. In a further preferred embodiment of the invention, the bin buffer selector 6 determines the output bin value 7 based on the input bin value 5 and the associated identifier specifying an estimate for which of the two possible bin values represents the less probable or more probable bin value for the current bin. In a preferred embodiment of the invention, the output bin value 7 is set equal to the input bin value 5 if the identifier specifies that the first of the two possible bin values represents the less probable (or more probable) bin value for the current bin, and the output bin value 7 is modified (i.e., it is set to the opposite of the input bin value) if identifier specifies that the second of the two possible bin values represents the less probable (or more probable) bin value for the current bin.
In a preferred embodiment of the invention, the bin buffer selector 6 determines the bin buffer 8 to which the output bin 7 is sent based on the associated measure for an estimate of the probability for one of the two possible bin values for the current bin. In a preferred embodiment of the invention, the set of possible values for the measure for an estimate of the probability for one of the two possible bin values is finite and the bin buffer selector 6 contains a table that associates exactly one bin buffer 8 with each possible value for the estimate of the probability for one of the two possible bin values, where different values for the measure for an estimate of the probability for one of the two possible bin values can be associated with the same bin buffer 8. In a further preferred embodiment of the invention, the range of possible values for the measure for an estimate of the probability for one of the two possible bin values is partitioned into a number of intervals, the bin buffer selector 6 determines the interval index for the current measure for an estimate of the probability for one of the two possible bin values, and the bin buffer selector 6 contains a table that associates exactly one bin buffer 8 with each possible value for the interval index, where different values for the interval index can be associated with the same bin buffer 8. In a preferred embodiment of the invention, input bins 5 with opposite measures for an estimate of the probability for one of the two possible bin values (opposite measure are those which represent probability estimates P and 1 - P) are fed into the same bin buffer 8. In a further preferred embodiment of the invention, the association of the measure for an estimate of the probability for one of the two possible bin values for the current bin with a particular bin buffer is adapted over time, e.g. in order to ensure that the created partial bitstreams have similar bit rates.
In a further preferred embodiment of the invention, the bin buffer selector 6 determines the bin buffer 8 to which the output bin 7 is sent based on the associated measure for an estimate of the probability for the less probable or more probable bin value for the current bin. In a preferred embodiment of the invention, the set of possible values for the measure for an estimate of the probability for the less probable or more probable bin value is finite and the bin buffer selector 6 contains a table that associates exactly one bin buffer 8 with each possible value of the estimate of the probability for the less probable or more probable bin value, where different values for the measure for an estimate of the probability for the less probable or more probable bin value can be associated with the same bin buffer 8. In a further preferred embodiment of the invention, the range of possible values for the measure for an estimate of the probability for the less probable or more probable bin value is partitioned into a number of intervals, the bin buffer selector 6 determines the interval index for the current measure for an estimate of the probability for the less probable or more probable bin value, and the bin buffer selector 6 contains a table that associates exactly one bin buffer 8 with each possible value for the interval index, where different values for the interval index can be associated with the same bin buffer 8. In a further preferred embodiment of the invention, the association of the measure for an estimate of the probability for the less probable or more probable bin value for the current bin with a particular bin buffer is adapted over time, e.g. in order to ensure that the created partial bitstreams have similar bit rates.
Each of the two or more bin buffers 8 is connected with exactly one bin encoder 10 and each bin encoder is only connected with one bin buffer 8. Each bin encoder 10 reads bins from the associated bin buffer 8 and converts a sequence of bins 9 into a codeword 11, which represents a sequence of bits. The bin buffers 8 represent first-in-first-out buffers; bins that are fed later (in sequential order) into a bin buffer 8 are not encoded before bins that are fed earlier (in sequential order) into the bin buffer. The codewords 11 that are output of a particular bin encoder 10 are written to a particular partial bitstream 12. The overall encoding algorithm converts source symbols 1 into two or more partial bitstreams 12, where the number of partial bitstreams is equal to the number of bin buffers and bin encoders. In a preferred embodiment of the invention, a bin encoder 10 converts a variable number of bins 9 into a codeword 11 of a variable number of bits. One advantage of the above- and below- outlined embodiments of the invention is that the encoding of bins can be done in parallel (e.g. for different groups of probability measures), which reduces the processing time for several implementations.
Another advantage of embodiments of the invention is that the bin encoding, which is done by the bin encoders 10, can be specifically designed for different sets of parameters 5. In particular, the bin encoding and encoding can be optimized (in terms of coding efficiency and/or complexity) for different groups of estimated probabilities. On the one hand side, this allows a reduction of the encoding/decoding complexity relative to state-of-the-art entropy coding algorithms with similar coding efficiency. On the other hand side, it allows an improvement of the coding efficiency relative to state-of-the-art entropy coding algorithms with similar encoding/decoding complexity. In a preferred embodiment of the invention, the bin encoders 10 implement different encoding algorithms (i.e. mapping of bin sequences onto codewords) for different groups of measures for an estimate of the probability for one of the two possible bin values 5 for the current bin. In a further preferred embodiment of the invention, the bin encoders 10 implement different encoding algorithms for different groups of measures for an estimate of the probability for the less probable or more probable bin value for the current bin. In a further preferred embodiment, the bin encoders 10 implement different encoding algorithms for different channel protection codes. In a further preferred embodiment, the bin encoders 10 implement different encoding algorithms for different encryption schemes. In a further preferred embodiment, the bin encoders 10 implement different encoding algorithms for different combinations of channel protection codes and groups of measures for an estimate of the probability for one of the two possible bin values 5 for the current bin. In a further preferred embodiment, the bin encoders 10 implement different encoding algorithms for different combinations of channel protection codes and groups of measures for an estimate of the probability for the less probable or more probable bin value 5 for the current bin. In a further preferred embodiment, the bin encoders 10 implement different encoding algorithms for different combinations of encryption schemes and groups of measures for an estimate of the probability for one of the two possible bin values 5 for the current bin. In a further preferred embodiment, the bin encoders 10 implement different encoding algorithms for different combinations of encryption schemes and groups of measures for an estimate of the probability for the less probable or more probable bin value 5 for the current bin.
In a preferred embodiment of the invention, the bin encoders 10 - or one or more of the bin encoders - represent binary arithmetic encoding engines. In a further preferred embodiment of the invention, one or more of the bin encoders represent a binary arithmetic coding engine, wherein the mapping from the representative LPS/LPB probability /?Lps of a given bin buffer to a corresponding code interval width ?LPS - i-e. the interval subdivision of the internal state of the binary arithmetic coding engine, which is defined by the current interval width R and the current interval offset L, identifying, for example, the lower bound of the code interval - is realized by using a table lookup. In a further preferred embodiment, for each table-based binary arithmetic coding engine associated to a given bin buffer, K representative interval width values {Qo, QK-I} are used for representing ?LPS with the choice of K and the representative interval width values {Qo, QK-I) being dependent on the bin buffer. For a choice of K > 1, arithmetic encoding of a bin may involve the substeps of mapping the current interval width R to a quantization index q with values in{6>, K-l ) and performing the interval subdivision by accessing the corresponding partial interval width value Qq from a lookup table with using q as an index. For a choice of K— 1 , i.e., for the case where only one representative interval width value Qo is given, this value Qo may be chosen as a power of two in order to allow decoding of multiple MPS/MPB values entering the corresponding bin buffer within a single renormalization cycle. The resulting codewords of each arithmetic coding engine may be separately transmitted, packetized, or stored, or they may be interleaved for the purpose of transmission or storage as described hereinafter.
That is, a binary arithmetic coding engine 10 could perform the following steps in coding the bins in its bin buffer 8:
1. Receiving valLPS, bin from bin buffer (recall: the respective binary arithmetic coding engine 10 considered here had been chosen to receive "bin" because (or, in other words, "bin" was associated with the respective binary arithmetic coding engine 10) the probability distribution estimate, such as p_state[bin], was associated with that binary arithmetic coding engine 10)
2. Quantization of R:
q_index = Qtab[R»q] (or some other form of quantization)
3. Determination of RLps and R:
RLPS = Rtab [q_index] (note that p_state is not mentioned here, as it is fixed for the binary arithmetic coding engine 10 considered, i.e. p_state [encoder], and Rtab has stored therein pre-calculated values for p[p_state[encoder]] Q[q_index]
R = R- RLPS [that is, R is preliminarily pre-updated as if "bin" was MPS]
4. Calculation of the new partial interval:
if (bin = 1 - valMPS) then
L <- L + R
R - RLPS
5. Renormalization of L and R, writing bits, wherein
q_index describes the index of a quantization value read out of Qtab,
p_state describes the current state (fixed for the binary arithmetic coding engine 10),
RLPS describes the interval width corresponding to the LPS and
valMPS describes the value of the bit corresponding to the MPS.
Accordingly, a binary arithmetic decoding engine 22 could perform the following steps in decoding the bins output into bin buffer 20:
1. Receiving request for a bin (recall: the respective binary arithmetic decoding engine 22 considered here had been chosen to decode "bin" because (or, in other words, "bin" was associated with the respective binary arithmetic decoding engine 22) the probability distribution estimate, such as p_state[bin], was associated with that binary arithmetic decoding engine 22)
2. Quantization of R: q_index = Qtab[R»q] (or some other form of quantization)
3. Determination of RLps and R:
RLPS = Rtab [q_index] (note that p_state is not mentioned here, as it is fixed for the binary arithmetic decoding engine 22 considered, i.e. p_state [encoder], and Rtab has stored therein pre-calculated values for p[p_state[encoder]] Q[q_index]
R = R - RLPS [that is, R is preliminarily pre-updated as if "bin" was MPS]
4. Determination of bin depending on the position of the partial interval:
if (V > R) then
bin - 1 - valMPS (bin is decoded as LPS; bin buffer selector 18 will obtain the actual bin value by use of this bin information and valMPS)
V - V - R
R - RLPS
else
bin <- valMPS (bin is decoded as MPS; bin buffer selector 18 will obtain the
actual bin value by use of this bin information and valMPS)
5. Renormalization of R, reading out one bit and updating V, wherein
q_index describes the index of a quantization value read out of Qtab,
p_state describes the current state (fixed for the binary arithmetic decoding engine 22), RLPS describes the interval width corresponding to the LPS,
valMPS describes the value of the bit corresponding to the MPS, and
V describes a value from the interior of the
current partial interval .
In a further preferred embodiment of the invention, the bin encoders 10 - or one or more of the bin encoders - represent entropy encoders that directly map sequences of input bins 9 onto codewords 10. Such mappings can be efficiently implemented and don't require a complex arithmetic coding engine. The inverse mapping of codewords onto sequences of bins (as done in the decoder) have to be unique in order to guarantee perfect decoding of the input sequence, but the mapping of bin sequences 9 onto codewords 10 doesn't necessarily need to be unique, i.e., it is possible that a particular sequence of bins can be mapped onto more than one sequence of codewords. In a preferred embodiment of the invention, the mapping of sequences of input bins 9 onto codewords 10 is bijective. In a further preferred embodiment of the invention, the bin encoders 10 - or one or more of the bin encoders - represent entropy encoders that directly map variable-length sequences of input bins 9 onto variable-length codewords 10. In a preferred embodiment of the invention, the output codewords represent redundancy-free codes such as general huffman codes or canonical huffman codes.
Two examples for the bijective mapping of bin sequences to redundancy-free codes are illustrated in Table 3. In a further preferred embodiment of the invention, the output codewords represent redundant codes suitable for error detection and error recovery. In a further preferred embodiment of the invention, the output codewords represent encryption codes suitable for encrypting the source symbols.
Table 3: Examples for mappings between bin sequences and codewords. sequence of bins codewords
(bin order is from left to right) (bits order is from left to right)
0000 0000 1
0000 0001 0000
0000 001 0001
0000 01 0010
0000 1 0011
0001 0100
001 0101
01 0110
1 0111 sequence of bins codewords
(bin order is from left to right) (bits order is from left to right)
000 10
01 11
001 010
11 011
1000 0 0001
1001 0010
1010 0011
1000 1 0000 0
1011 0000 1
In a further preferred embodiment of the invention, the bin encoders 10 - or one or more of the bin encoders - represent entropy encoders that directly map variable-length sequences of input bins 9 onto fixed-length codewords 10. In a further preferred embodiment of the invention, the bin encoders 10 - or one or more of the bin encoders - represent entropy encoders that directly map fixed-length sequences of input bins 9 onto variable-length codewords 10.
The decoder according an embodiment of the invention is illustrated in Figure 2. The decoder performs basically the inverse operations of the encoder, so that the (previously encoded) sequence of source symbols 27 is decoded from a set of two or more partial bitstreams 24. The decoder includes two different process flows: A flow for data requests, which replicates the data flow of the encoder, and a data flow, which represents the inverse of the encoder data flow. In the illustration in Fig. 2, the dashed arrows represent the data request flow, while the solid arrows represent the data flow. The building blocks of the decoder basically replicate the building blocks of the encoder, but implement the inverse operations.
The decoding of a source symbol is triggered by a request for a new decoded source symbol 13 that is sent to the binarizer 14. In a preferred embodiment of the invention, each request for a new decoded source symbol 13 is associated with a category of a set of one or more categories. The category that is associated with a request for a source symbol is the same as the category that was associated with the corresponding source symbol during encoding.
The binarizer 14 maps the request for a source symbol 13 into one or more requests for a bin that are sent to the parameter assigner 16. As final response to a request for a bin that is sent to the parameter assigner 16 by the binarizer 14, the binarizer 14 receives a decoded bin 26 from the bin buffer selector 18. The binarizer 14 compares the received sequence of decoded bins 26 with the bin sequences of a particular binarization scheme for the requested source symbol and, if the received sequence of decoded bins 26 matches the binarization of a source symbol, the binarizer empties its bin buffer and outputs the decoded source symbol as final response to the request for a new decoded symbol. If the already received sequence of decoded bins does not match any of the bin sequences for the binarization scheme for the requested source symbol, the binarizer sends another request for a bin to the parameter assigner until the sequence of decoded bins matches one of the bin sequences of the binarization scheme for the requested source symbol. For each request for a source symbol, the decoder uses the same binarization scheme that was used for encoding the corresponding source symbol. The binarization scheme can be different for different source symbol categories. The binarization scheme for a particular source symbol category can depend on the set of possible source symbol values and/or other properties of the source symbols for the particular category.
The parameter assigner assigns a set of one or more parameters to each request for a bin and sends the request for a bin with the associated set of parameters to the bin buffer selector. The set of parameters that are assigned to a requested bin by the parameter assigner is the same that was assigned to the corresponding bin during encoding. The set of parameters may consist of one or more of the parameters that are mentioned in the encoder description.
In a preferred embodiment of the invention, the parameter assigner 16 associates each request for a bin with a measure for an estimate of the probability for one of the two possible bin values for the current requested bin. In a further preferred embodiment of the invention, the parameter assigner 16 associates each request for a bin with a measure for an estimate of the probability for the less probable or more probable bin value for the current requested bin and an identifier specifying an estimate for which of the two possible bin values represents the less probable or more probable bin value for the current requested bin.
In a further preferred embodiment of the invention, the parameter assigner 16 associates each request for a bin 15,17 with a measure for an estimate of the probability for one of the two possible bin values for the current requested bin and one or more further parameters. In a further preferred embodiment of the invention, the parameter assigner 16 associates each request for a bin 15,17 with a measure for an estimate of the probability for the less probable or more probable bin value for the current requested bin, an identifier specifying an estimate for which of the two possible bin values represents the less probable or more probable bin value for the current requested bin, and one or more further parameters (which may one or more of the above listed parameters).
In a preferred embodiment of the invention, the parameter assigner 16 determines one or more of the above mentioned probability measures (measure for an estimate of the probability for one of the two possible bin values for the current requested bin, measure for an estimate of the probability for the less probable or more probable bin value for the current requested bin, identifier specifying an estimate for which of the two possible bin values represents the less probable or more probable bin value for the current requested bin) based on a set of one or more already decoded symbols. The determination of the probability measures for a particular request for a bin replicates the process at the encoder for the corresponding bin. The decoded symbols that are used for determining the probability measures can include one or more already decoded symbols of the same symbol category, one or more already decoded symbols of the same symbol category that correspond to data sets (such as blocks or groups of samples) of neighboring spatial and/or temporal locations (in relation to the data set associated with the current request for a source symbol), or one or more already decoded symbols of different symbol categories that correspond to data sets of the same and/or neighboring spatial and/or temporal locations (in relation to the data set associated with the current request for a source symbol). Each request for a bin with an associated set of parameters 17 that is output of the parameter assigner 16 is fed into a bin buffer selector 18. Based on the associated set of parameters 17, the bin buffer selector 18 sends a request for a bin 19 to one of two or more bin buffers 20 and receives a decoded bin 25 from the selected bin buffer 20. The decoded input bin 25 is potentially modified and the decoded output bin 26 - with a potentially modified value - is send to the binarizer 14 as final response to the request for a bin with an associated set of parameters 17.
The bin buffer 20 to which the request for a bin is forwarded is selected in the same way as the bin buffer to which the output bin of the bin buffer selector at the encoder side was sent.
In a preferred embodiment of the invention, the bin buffer selector 18 determines the bin buffer 20 to which the request for a bin 19 is sent based on the associated measure for an estimate of the probability for one of the two possible bin values for the current requested bin. In a preferred embodiment of the invention, the set of possible values for the measure for an estimate of the probability for one of the two possible bin values is finite and the bin buffer selector 18 contains a table that associates exactly one bin buffer 20 with each possible value of the estimate of the probability for one of the two possible bin values, where different values for the measure for an estimate of the probability for one of the two possible bin values can be associated with the same bin buffer 20. In a further preferred embodiment of the invention, the range of possible values for the measure for an estimate of the probability for one of the two possible bin values is partitioned into a number of intervals, the bin buffer selector 18 determines the interval index for the current measure for an estimate of the probability for one of the two possible bin values, and the bin buffer selector 18 contains a table that associates exactly one bin buffer 20 with each possible value for the interval index, where different values for the interval index can be associated with the same bin buffer 20. In a preferred embodiment of the invention, requests for bins 17 with opposite measures for an estimate of the probability for one of the two possible bin values (opposite measure are those which represent probability estimates P and 1 - P) are forwarded to the same bin buffer 20. In a further preferred embodiment of the invention, the association of the measure for an estimate of the probability for one of the two possible bin values for the current bin request with a particular bin buffer is adapted over time.
In a further preferred embodiment of the invention, the bin buffer selector 18 determines the bin buffer 20 to which the request for a bin 19 is sent based on the associated measure for an estimate of the probability for the less probable or more probable bin value for the current requested bin. In a preferred embodiment of the invention, the set of possible values for the measure for an estimate of the probability for the less probable or more probable bin value is finite and the bin buffer selector 18 contains a table that associates exactly one bin buffer 20 with each possible value of the estimate of the probability for the less probable or more probable bin value, where different values for the measure for an estimate of the probability for the less probable or more probable bin value can be associated with the same bin buffer 20. In a further preferred embodiment of the invention, the range of possible values for the measure for an estimate of the probability for the less probable or more probable bin value is partitioned into a number of intervals, the bin buffer selector 18 determines the interval index for the current measure for an estimate of the probability for the less probable or more probable bin value, and the bin buffer selector 18 contains a table that associates exactly one bin buffer 20 with each possible value for the interval index, where different values for the interval index can be associated with the same bin buffer 20. In a further preferred embodiment of the invention, the association of the measure for an estimate of the probability for the less probable or more probable bin value for the current bin request with a particular bin buffer is adapted over time.
After receiving a decoded bin 25 from the selected bin buffer 20, the bin buffer selector 18 potentially modifies the input bin 25 and sends the output bin 26 - with a potentially modified value - to the binarizer 14. The input/output bin mapping of the bin buffer selector 18 is the inverse of the input/output bin mapping of the bin buffer selector at the encoder side.
In a preferred embodiment of the invention, the bin buffer selector 18 does not modify the value of the bin, i.e., the output bin 26 has always the same value as the input bin 25.
In a further preferred embodiment of the invention, the bin buffer selector 18 determines the output bin value 26 based on the input bin value 25 and the measure for an estimate of the probability for one of the two possible bin values for the current requested bin that is associated with the request for a bin 17. In a preferred embodiment of the invention, the output bin value 26 is set equal to the input bin value 25 if the measure for the probability for one of the two possible bin values for the current bin request is less than (or less than or equal to) a particular threshold; if the measure for the probability for one of the two possible bin values for the current bin request is greater than or equal to (or greater than) a particular threshold, the output bin value 26 is modified (i.e., it is set to the opposite of the input bin value). In a further preferred embodiment of the invention, the output bin value 26 is set equal to the input bin value 25 if the measure for the probability for one of the two possible bin values for the current bin request is greater than (or greater than or equal to) a particular threshold; if the measure for the probability for one of the two possible bin values for the current bin request is less than or equal to (or less than) a particular threshold, the output bin value 26 is modified (i.e., it is set to the opposite of the input bin value). In a preferred embodiment of the invention, the value of the threshold corresponds to a value of 0.5 for the estimated probability for both possible bin values. In a further preferred embodiment of the invention, the bin buffer selector 18 determines the output bin value 26 based on the input bin value 25 and the identifier, specifying an estimate for which of the two possible bin values represents the less probable or more probable bin value for the current bin request, that is associated with the request for a bin 17. In a preferred embodiment of the invention, the output bin value 26 is set equal to the input bin value 25 if the identifier specifies that the first of the two possible bin values represents the less probable (or more probable) bin value for the current bin request, and the output bin value 26 is modified (i.e., it is set to the opposite of the input bin value) if identifier specifies that the second of the two possible bin values represents the less probable (or more probable) bin value for the current bin request.
As described above, the bin buffer selector sends a request for a bin 19 to one of the two or more bin buffers 20. The bin buffers 20 represent first-in-first-out buffers, which are fed with sequences of decoded bins 21 from the connected bin decoders 22. As response to a request for a bin 19 that is sent to a bin buffer 20 from the bin buffer selector 18, the bin buffer 20 removes the bin of its content that was first fed into the bin buffer 20 and sends it to the bin buffer selector 18. Bins that are earlier sent to the bin buffer 20 are earlier removed and sent to the bin buffer selector 18.
Each of the two or more bin buffers 20 is connected with exactly one bin decoder 22 and each bin decoder is only connected with one bin buffer 20. Each bin decoder 22 reads codewords 23, which represent sequences of bits, from a separate partial bitstream 24. The bin decoder converts a codeword 23 into a sequence of bins 21 that is sent to the connected bin buffer 20. The overall decoding algorithm converts two or more partial bitstreams 24 into a number of decoded source symbols, where the number of partial bitstreams is equal to the number of bin buffers and bin decoders and the decoding of source symbols is triggered by requests for new source symbols. In a preferred embodiment of the invention, a bin decoder 22 converts codewords 23 of a variable number of bits into a sequence of a variable number of bins 21. One advantage of embodiments of the invention is that the decoding of bins from the two or more partial bitstreams can be done in parallel (e.g. for different groups of probability measures), which reduces the processing time for several implementations.
Another advantage of embodiments of the invention is that the bin decoding, which is done by the bin decoders 22, can be specifically designed for different sets of parameters 17. In particular, the bin encoding and decoding can be optimized (in terms of coding efficiency and/or complexity) for different groups of estimated probabilities. On the one hand side, this allows a reduction of the encoding/decoding complexity relative to state-of-the-art entropy coding algorithms with similar coding efficiency. On the other hand side, it allows an improvement of the coding efficiency relative to state-of-the-art entropy coding algorithms with similar encoding/decoding complexity. In a preferred embodiment of the invention, the bin decoders 22 implement different decoding algorithms (i.e. mapping of bin sequences onto codewords) for different groups of measures for an estimate of the probability for one of the two possible bin values 17 for the current bin request. In a further preferred embodiment of the invention, the bin decoders 22 implement different decoding algorithms for different groups of measures for an estimate of the probability for the less probable or more probable bin value for the current requested bin. In a further preferred embodiment, the bin decoders 22 implement different decoding algorithms for different channel protection codes. In a further preferred embodiment, the bin decoders 22 implement different decoding algorithms for different encryption schmes. In a further preferred embodiment, the bin decoders 22 implement different decoding algorithms for different combinations of channel protection codes and groups of measures for an estimate of the probability for one of the two possible bin values 17 for the current requested bin. In a further preferred embodiment, the bin decoders 22 implement different decoding algorithms for different combinations of channel protection codes and groups of measures for an estimate of the probability for the less probable or more probable bin value 17 for the current requested bin. In a further preferred embodiment, the bin decoders 22 implement different decoding algorithms for different combinations of encryption schemes and groups of measures for an estimate of the probability for one of the two possible bin values 17 for the current requested bin. In a further preferred embodiment, the bin decoders 22 implement different decoding algorithms for different combinations of encryption schemes and groups of measures for an estimate of the probability for the less probable or more probable bin value 17 for the current requested bin.
The bin decoders 22 do the inverse mapping of the corresponding bin encoders at the encoder side.
In a preferred embodiment of the invention, the bin decoders 22 - or one or more of the bin decoders - represent binary arithmetic decoding engines.
In a further preferred embodiment of the invention, the bin decoders 22 - or one or more of the bin decoders - represent entropy decoders that directly map codewords 23 onto sequences of bins 21. Such mappings can be efficiently implemented and don't require a complex arithmetic coding engine. The mapping of codewords onto sequences of bins has to be unique. In a preferred embodiment of the invention, the mapping of codewords 23 onto sequences of bins 21 is bijective. In a further preferred embodiment of the invention, the bin decoders 10 - or one or more of the bin decoders - represent entropy decoders that directly map variable-length codewords 23 into variable-length sequences of bins 21. In a preferred embodiment of the invention, the input codewords represent redundancy-free codes such as general huffman codes or canonical huffman codes. Two examples for the bijective mapping of redundancy-free codes to bin sequences are illustrated in Table 3. In a further preferred embodiment of the invention, the input codewords represent redundant codes suitable for error detection and error recovery. In a further preferred embodiment of the invention, the input codewords represent encryption codes.
In a further preferred embodiment of the invention, the bin decoders 22 - or one or more of the bin decoders - represent entropy decoders that directly map fixed-length codewords 23 onto variable-length sequences of bins 21. In a further preferred embodiment of the invention, the bin decoders 22 - or one or more of the bin decoders - represent entropy decoders that directly map variable-length codewords 23 onto fixed-length sequences of bins 21.
Thus, Fig. 1 and 2 showed an embodiment for an encoder for encoding a sequence of symbols 3 and a decoder for reconstructing same. The encoder comprises an assigner 4 configured to assign a number of parameters 5 to each symbol of the sequence of symbols. The assignment is based on information contained within previous symbols of the sequence of symbols such as the category of the syntax element 1 to the representation - such as binarization - of which the current symbol belongs and which, according to the syntax structure of the syntax elements 1, is currently be expected which expectation, in turn, is deducible from the history of previous syntax elements 1 and symbols 3. Further, the encoder comprises a plurality of entropy encoders 10 each of which is configured to convert the symbols 3 forwarded to the respective entropy encoder into a respective bitstream 12, and a selector 6 configured to forward each symbol 3 to a selected one of the plurality of entropy encoders 10, the selection depending on the number of parameters 5 assigned to the respective symbol 3. The decoder for reconstructing a sequence of symbols comprises a plurality of entropy decoders 22, each of which is configured to convert a respective bitstream 23 into symbols 21; an assigner 16 configured to assign a number of parameters 17 to each symbol 15 of a sequence of symbols to be reconstructed based on information contained within previously reconstructed symbols of the sequence of symbols (see 26 and 27 in Fig. 2); and a selector 18 configured to retrieve each symbol of the sequence of symbols to be reconstructed from a selected one of the plurality of entropy decoders 22, the selection depending on the number of parameters defined to the respective symbol. The assigner 16 may be configured such that the number of parameters assigned to each symbol comprises, or is, a measure for an estimate of a probability of distribution among the possible symbol values a respective symbol may assume. The sequence of symbols to be reconstructed may be of a binary alphabet and the assigner 16 may be configured such that the estimate of the probability distribution consists of a measure for an estimate of a probability of a less probable or more probable bin value of the two possible bin values of the binary alphabet and an identifier specifying an estimate for which of the two possible bin values represents the less probable or more probable bin value. The assigner 16 may further be configured to internally assign a context to each symbol of the sequence of symbols 15 to be reconstructed based on the information contained within previously reconstructed symbols of the sequence of symbols to be reconstructed with each context having a respective probability distribution estimate associated therewith, and to adapt the probability distribution estimate for each context to an actual symbol statistic based on symbol values of previously reconstructed symbols to which the respective context is assigned. The context may take into account a spatial relationship or neighbourhood of positions to which the syntax elements belong such as in video or picture coding, or even in tables in case of financial applications. Then, the measure for the estimate of the probability distribution for each symbol may be determined based on the probability distribution estimate associated with the context assigned to the respective symbol such as by quantizing the probability distribution estimate associated with the context assigned with the respective symbol to one of a plurality of probability distribution estimate representatives in order to obtain the measure for the estimate of the probability distribution. The selector may be configured such that a bijective association is defined between the plurality of entropy encoders and the plurality of probability distribution estimate representatives. The selector 18 may be configured to change a quantization mapping from a range of the probability distribution estimates to the plurality of probability distribution estimate representatives in a predetermined deterministic way depending on previously reconstructed symbols of the sequence of symbols, over time. That is, selector 18 may change the quantization step sizes, i.e. the intervals of probability distributions mapped onto the individual probability indices bijectively associated with the individual entropy decoders. The plurality of entropy decoders 22, in turn, may be configured to adapt their way of converting symbols into bit streams responsive to a change in the quantization mapping. For example, each entropy decoder 22 may be optimised for, i.e may have an optimal compression rate for, a certain probability distribution estimate within the respective probability distribution estimate quantization interval, and may change its codeword/symbol sequence mapping so as to adapt the position of this certain probability distribution estimate within the respective probability distribution estimate quantization interval upon a change of the latter so as to be optimised. The selector may be configured to change the quantization mapping such that rates by which the symbols are retrieved from the plurality of entropy decoders, are made less dispersed. As to the binarizer 14 it is noted that same me be left away if the syntax elements are already binary. Further, depending on the type of decoder 22, the existence of the buffers 20 is not necessary. Further, the buffers may be integrated within the decoders.
Termination of finite source symbol sequences
In a preferred embodiment of the invention, the encoding and decoding is done for a finite set of source symbols. Often a certain quantity of data such as a still image, a frame or field of a video sequence, a slice of an image, a slice of a frame or a field of a video sequence, or a set of successive audio samples, etc. is coded. For finite sets of source symbols, in general, the partial bitstreams that are created at the encoder side have to be terminated, i.e., it has to be ensured that all source symbols can be decoded from the transmitted or stored partial bitstreams. After the last bin is inserted into the corresponding bin buffer 8, the bin encoder 10 has to ensure that a complete codeword is written to the partial bitstream 12. If the bin encoder 10 represents a binary arithmetic coding engine, the arithmetic codeword has to be terminated. If the bin encoder 10 represents an entropy encoder that implements a direct mapping of bin sequences onto codewords, the bin sequence that is stored in the bin buffer after writing the last bin to the bin buffer might not represent a bin sequence that is associated with a codeword (i.e., it might represent a prefix of two or more bin sequences that are associated with codewords). In such a case, any of the codewords associated with a bin sequence that contains the bin sequence in the bin buffer as prefix has to be written to the partial bitstream (the bin buffer has to be flushed). This could be done by inserting bins with a particular or an arbitrary value into the bin buffer until a codeword is written. In a preferred embodiment of the invention, the bin encoder selects one of the codewords with minimum length (in addition to the property that the associated bin sequence must contain the bin sequence in the bin buffer as prefix). At the decoder side, the bin decoder 22 may decode more bins than required for the last codeword in a partial bitstream; these bins are not requested by the bin buffer selector 18 and are discarded and ignored. The decoding of the finite set of symbols is controlled by requests for decoded source symbols; if no further source symbol is requested for a quantity of data, the decoding is terminated.
Transmission and multiplexing of the partial bitstreams
The partial bitstreams 12 that are created by the encoder can be transmitted separately, or they can be multiplexed into a single bitstream, or the codewords of the partial bitstreams can be interleaved in a single bitstream.
In a preferred embodiment of the invention, each partial bitstream for a quantity of data is written to one data packet. The quantity of data can be an arbitrary set of source symbols such as a still picture, a field or frame of a video sequence, a slice of a still picture, a slice of a field or frame of a video sequence, or a frame of audio samples, etc.
In another preferred embodiment of the invention, two or more of the partial bitstreams for a quantity of data or all partial bitstreams for a quantity of data are multiplexed into one data packet. The structure of a data packet that contains multiplexed partial bitstreams is illustrated in Figure 3. The data packet 100 consists of a header and one partition for the data of each partial bitstream (for the considered quantity of data). The header 100 of the data packet contains indications for the partitioning of the (remainder of the) data packet into segments of bitstream data 102. Beside the indications for the partitioning, the header may contain additional information. In a preferred embodiment of the invention, the indications for the partitioning of the data packet are the locations of the beginning of the data segments in units of bits or bytes or multiples of bits or multiples of bytes. In a preferred embodiment of the invention, the locations of the beginning of the data segments are coded as absolute values in the header of the data packet, either relative to the beginning of the data packet or relative to the end of the header or relative to the beginning of the previous data packet. In a further preferred embodiment of the invention, the locations of the beginning of the data segments are differentially coded, i.e., only the difference between the actual beginning of a data segment and a prediction for the beginning of the data segment is coded. The prediction can be derived based on already known or transmitted information such as the overall size of the data packet, the size of the header, the number of data segments in the data packet, the location of the beginning of preceding data segments. In a preferred embodiment of the invention, the location of the beginning of the first data packet is not coded, but inferred based on the size of the data packet header. At the decoder side, the transmitted partition indications are used for deriving the beginning of the data segments. The data segments are then used as partial bitstreams and the data contained in the data segments are fed into the corresponding bin decoders in sequential order.
There are several alternatives for multiplexing the partial bitstreams into a data packet. One alternative, which can reduce the require side information, in particular for cases in which the sizes of the partial bitstreams are very similar, is illustrated in Fig. 4. The payload of the data packet, i.e., the data packet 110 without its header 111, is partitioned into segments 112 a predefined way. As an example, the data packet payload can be partitioned into segments of the same size. Then each segment is associated with a partial bitstream or with the first part of a partial bitstream 113. If a partial bitstream is greater than the associated data segment, its remainder 114 is placed into the unused space at the end of other data segments. This can be done in a way that the remaining part of a bitstream is inserted in reverse order (starting from the end of the data segment), which reduces the side information. The association of the remainders of the partial bitstreams to data segments and, when more than one remainder is added to a data segment, the start point for one or more of the remainders have to be signaled inside the bitstream, e.g. in the data packet header.
Interleaving of variable-length codewords For some applications, the above described multiplexing of the partial bitstreams (for a quantity of source symbols) in one data packet can have the following disadvantages: On the one hand side, for small data packets, the number of bits for the side information that is required for signaling the partitioning can become significant relative to the actual data in the partial bitstreams, which finally reduces the coding efficiency. On the other hand, the multiplexing may not suitable for applications that require a low delay (e.g. for video conferencing applications). With the described multiplexing, the encoder cannot start the transmission of a data packet before the partial bitstreams have been completely created, since the locations of the beginning of the partitions are not known before. Furthermore, in general, the decoder has to wait until it receives the beginning of the last data segment before it can start the decoding of a data packet. For applications as video conferencing systems, these delays can add-up to an additional overall delay of the system of several video pictures (in particular for bit rates that are close to the transmission bit rate and for encoders/decoders that require nearly the time interval between two pictures for encoding/decoding a picture), which is critical for such applications. In order to overcome the disadvantages for certain applications, the encoder of a preferred embodiment of the invention can be configured in a way that the codewords that are generated by the two or more bin encoders are interlaved into a single bitstream. The bitstream with the interleaved codewords can be directly send to the decoder (when neglecting a small buffer delay, see below). At the decoder side, the two or more bin decoders read the codewords directly from the bitstream in decoding order; the decoding can be started with the first received bit. In addition, no side information is required for signaling the multiplexing (or interleaving) of the partial bitstreams.
The basic structure of an encoder with codeword interleaving is shown in Fig. 5. The bin encoders 10 don't write the codewords directly to the partial bitstreams, but are connected with a single codeword buffer 29, from which codewords are written to the bitstream 34 in coding order. The bin encoders 10 send requests for one or more new codeword buffer entries 28 to the codeword buffer 29 and later send the codewords 30 to the codeword buffer 29, which are stored in the reserved buffer entries. The (in general variable-length) codewords 31 of the codeword buffer 29 are accessed by a codeword writer 32, which writes the corresponding bits 33 to the produced bitstream 34. The codeword buffer 29 operates as a first-in-first-out buffer; codeword entries that are reserved earlier are earlier written to the bitstream.
In a further generalization, multiple codeword buffers and partial bitstreams are possible, where the number of codeword buffers is less than the number of bin encoders. A bin encoder 10 reserves one or more codewords in the codeword buffer 29, whereby the reservation of the one or more codewords in the codeword buffer is triggered by certain events in the connected bin buffer 8. In a preferred embodiment of the invention, the codeword buffer 29 is operated in a way that the decoder can instantaneously decode the bitstream. The coding order in which the codewords are written to the bitstream is the same as the order in which the corresponding codewords are reserved in the codeword buffer. In a preferred embodiment of the invention, each bin encoder 10 reserves one codeword, with the reservation being triggered by a certain event in the connected bin buffer. In another preferred embodiment of the invention, each bin encoder 10 reserves more than one codeword, with the reservation being triggered by a certain event in the connected bin buffer. In further preferred embodiment of the invention, the bin encoders 10 reserve a different amount of codewords, where the amount of codewords that are reserved by a particular bin encoder can be dependent on the particular bin encoder and/or other properties of the particular bin encoder/bin buffer (such as the associated probability measure, the number of already written bits, etc.).
In a preferred embodiment of the invention, the codeword buffer is operated as follows. If a new bin 7 is sent to a particular bin buffer 8 and the number of already stored bins in the bin buffer is zero and there is currently no codeword reserved in the codeword buffer for the bin encoder that is connected with the particular bin buffer, the connected bin encoder 10 sends a request to the codeword buffer, by which one or more codeword entries are reserved in the codeword buffer 29 for the particular bin encoder. The codeword entries can have a variable number of bits; an upper threshold for the number of bits in a buffer entry is usually given by the maximum codeword size for the corresponding bin encoder. The next codeword or the next codewords that are produced by the bin encoder (for which the codeword entry or codeword entries have been reserved) are stored in the reserved entry or entries of the codeword buffer. If all reserved buffer entries in the codeword buffer for a particular bin encoder are filled with codewords and the next bin is sent to the bin buffer that is connected with the particular bin encoder, one or more new codewords are reserved in the codeword buffer for the particular bin encoder, etc. The codeword buffer 29 represents a first-in-first-out buffer in a certain way. Buffer entries are reserved in sequential order. Codewords for which the corresponding buffer entries have been reserved earlier are earlier written to the bitstream. The codeword writer 32 checks the status of the codeword buffer 29, either continuously or after a codeword 30 is written to the codeword buffer 29. If the first buffer entry contains a complete codeword (i.e., the buffer entry is not reserved, but includes a codeword), the corresponding codeword 31 and the corresponding buffer entry are removed from the codeword buffer 20 and the bits of the codeword 33 are written to the bitstream. This process is repeated until the first buffer entry does not contain a codeword (i.e., it is reserved or free). At the end of the decoding process, i.e., if all source symbols of the considered quantity of data have been processed, the codeword buffer must be flushed. For that flushing process, the following is applied for each bin buffer/ bin encoder as first step: If the bin buffer does contain bins, a bin with a particular or an arbitrary value is added until the resulting bin sequence represents a bin sequence that is associated with a codeword (as noted above, one preferred way of adding bins is to add such bin values that produce the shortest possible codeword - or one of those - that is associated with a bin sequence that contains the for the original content of the bin buffer as prefix), then the codeword is written to the next reserved buffer entry for the corresponding bin encoder (and the corresponding) bin buffer is emptied. If more than one buffer entry has been reserved for one or more bin encoders, the codeword buffer may still contain reserved codeword entries. In that case, these codeword entries are filled with arbitrary, but valid codewords for the corresponding bin encoders. In a preferred embodiment of the invention, the shortest valid codeword or one of the shortest valid codewords (if there are multiple) is inserted. Finally, all remaining codewords in the codeword buffer are written to the bitstream.
Two examples for the status of the codeword buffer are illustrated in Fig. 6. In example (a), the codeword buffer contains 2 entries that are filled with a codeword and 5 reserved entries. In addition, the next free buffer entry is marked. The first entry is filled with a codeword (i.e., the bin encoder 2 just wrote a codeword to a previously reserved entry). In the next step, this codeword will be removed from the codeword buffer and written to the bitstream. Then, the first reserved codeword for bin encoder 3 is the first buffer entry, but this entry cannot be removed from the codeword buffer, since it is only reserved, but no codeword has been written to this entry. In the example (b), the codeword buffer contains 3 entries that are filled with a codeword and 4 reserved entries. The first entry is marked as reserved and hence the codeword writer cannot write a codeword to the bitstream. Although 3 codewords are contained in the codeword buffer, the codeword writer has to wait until a codeword is written to the first reserved buffer entry for bin encoder 3. Note that the codewords must be written in the order in which they were reserved, in order to be able to invert the process at the decoder side (see below).
The basic structure of a decoder with codeword interleaving is shown in Fig. 7. The bin decoders 10 don't read the codewords directly from separate partial bitstreams, but are connected to a bit buffer 38, from which the codewords 37 are read in coding order. It should be noted that the bit buffer 38 is not necessarily required, since the codewords could also be directly read from the bitstream. The bit buffer 38 is mainly included in the illustration for clearly separate different aspects of the processing chain. The bits 39 of the bitstream 40 with interleaved codewords are sequentially inserted into the bit buffer 38, which represents a first- in-first-out buffer. If a particular bin decoder 22 receives a request for one or more bin sequences35, the bin decoder 22 reads one or more codewords 37 from the bit buffer 38 via requests for bits 36. The decoder can instantaneously decode the source symbols. Note that the encoder (as described above) must ensure by suitably operating the codeword buffer that the codewords are written in the same order to the bitstream in which they are requested by the bin decoders. At the decoder, the entire decoding process is triggered by requests for source symbols. Parameters as the number of codewords that are reserved at the encoder side by a particular bin encoder and the number of codewords that are read by the corresponding bin decoder must be the same.
In a further generalization, multiple codeword buffers and partial bitstreams are possible, where the number of bit buffers is less than the number of bin decoders. A bin decoder 22 reads one or more codewords from the bit buffer 38 at one time instant, whereby the reading of the one or more codewords from the bit buffer is triggered by certain events in the connected bin buffer 20. In a preferred embodiment of the invention, the decoder is operated in a way that one or more codewords are read when a request for a bin 19 is sent to a particular bin buffer 20 and the bin buffer doesn't contain any bins. But it is also possible to trigger the reading of codewords by other events, e.g. if the number of bins in the bin buffer is below a predefined threshold. In a preferred embodiment of the invention, each bin decoder 22 reads one codeword, with the reading being triggered by a certain event in the connected bin buffer. In another preferred embodiment of the invention, each bin decoder 22 reads more than one codeword, with the reading being triggered by a certain event in the connected bin buffer. In further preferred embodiment of the invention, the bin decoders 22 read a different amount of codewords, where the amount of codewords that are read by a particular bin decoder can be dependent on the particular bin decoder and/or other properties of the particular bin decoder/bin buffer (such as the associated probability measure, the number of already read bits, etc.).
In a preferred embodiment of the invention, the reading of codewords from the bit buffer is operated as follows. If a new bin request 19 is sent from the bin buffer selector 18 to a particular bin buffer 20 and the number of bins in the bin buffer is zero, the connected bin decoder 22 reads one or more codewords 37 from the bit buffer 38, via bit request 36 to the bit buffer 38. The bin decoder 22 converts the read codewords 37 into sequences of bins 21 and stores these bin sequences in the connected bin buffer 20. As final response to the request for a bin 19, the first inserted bin is removed from the bin buffer 20 and sent to the bin buffer selector 18. As response the further bin requests, the remaining bins in the bin buffer are removed until the bin buffer is empty. An additional bin request triggers the bin decoder to read one or more new codewords from the bit buffer, etc. The bit buffer 38 represents a first- in-first-out buffer of a predefined size and is continuously filled with bits 39 from the bitstream 40. In order to ensure that the codewords are written to the bitstream in the same way as they are requested by the decoding process, the codeword buffer at the encoder side can be operated in the way described above.
Thus, each of the plurality of entropy decoders may be a variable length decoder configured to map codewords of fixed lengths to symbol sequences of variable lengths, and a codeword entry such as the output of the codeword buffer 43 may be provided for receiving a single stream of interleaved codewords. The plurality of entropy decoders 22 may be configured to retrieve the codewords from the codeword entry in a sequential order depending on an order in which the symbols of the sequence of symbols to be reconstructed as retrieved by the selector 18 from the plurality of entropy decoders result in a new symbol sequence to be mapped from a new codeword at the respective entropy decoders.
Interleaving of variable-length codewords with a low-delay constraint
The described codeword interleaving does not require that any partitioning information is sent as side information. And since the codewords are interleaved in the bitstream, the delay is in general small. However, it is not guaranteed that a particular delay constraint (e.g. specified by a maximum number of bits that are stored in the codeword buffer) is obeyed. Furthermore, the required buffer size for the codeword buffer can theoretically become very large. When considering the example in Fig. 6(b), it might be possible that no further bins are send to bin buffer 3 and hence the bin encoder 3 will not send any new codeword to the codeword buffer until the flushing process at the end of the data packet is applied. Then all codewords for bin encoders 1 and 2 would have to wait until the end of the data packet, before they can be written to the bitstream. This drawback can be circumvented by adding a further mechanism to the encoding process (and also to the decoding process as described later). The basic concept of that additional mechanism is that if a measure related to the delay or an upper bound of the delay (see below) exceeds a specified threshold, the first reserved buffer entry is filled by flushing the corresponding bin buffer (using a similar mechanism as at the end of a data packet). By such a mechanism, the number of waiting buffer entries is reduced until the associated delay measure is less than the specified threshold. At the decoder side, the bins that have been inserted at the encoder side in order to obey the delay constraint must be discarded. For this discarding of bins basically the same mechanism as at the encoder side can be used. In the following two preferred embodiments for such a delay control are described.
In one preferred embodiment of the invention, the measure for the delay (or an upper bound of the delay) is the number of active buffer entries in the codeword buffer, where the number of active buffer entries is the number of reserved buffer entries plus the number of buffer entries that contain codewords. Note that the first buffer entry is always a reserved buffer entry or a free buffer entry, since if the first buffer entry contains a codeword, this codeword is written to the bitstream. If for example, the maximum allowed buffer delay (as determined by the application) is D bits and the maximum codeword size for all bin encoders is L, a lower bound for the maximum number of codewords that can be contained in the codeword buffer without violating the delay constraint can be calculated by N = D/L. The delay measure D in bits is not required by the system, but the maximum number of codewords N must be known to both encoder and decoder. In a preferred embodiment of the invention, the maximum number of codeword buffer entries N is fixed by the application. In another preferred embodiment of the invention, the maximum number of codeword buffer entries N is signaled inside the bitstream, e.g., in the header of the data packet (or slice header) or in a parameter set, which is included in the bitstream. If a bin encoder 10 sends a request for the reservation of one or more new buffer entries to the codeword buffer 29, the following process is executed before a new codeword buffer entry is reserved (i.e., it is executed multiple times if multiple codeword buffer entries are reserved by one request): If the number of currently active buffer entries plus 1 (taking into account the buffer entry that will be reserved next) is greater than the maximum number of codeword buffer entries N, the first buffer entry (which is reserved) is flushed by the process described in the following until the number of currently active buffer entries plus 1 is less than or equal to the maximum number of codeword buffer entries N. The flushing of a reserved buffer entry is similar to the flushing at the end of a data packet: The bin encoder 10 that has reserved the corresponding first buffer entry is flushed by adding bins with particular or arbitrary values to the connected bin buffer 8 until the resulting bin sequence represents a bin sequence that is associated with a codeword, the codeword is then written to the reserved buffer entry and it is finally added to the bitstream (while emptying the bin buffer and removing the previously reserved buffer entry). As mentioned above, one preferred way for adding bins to the bin buffer is to add those bins that produce the shortest possible codeword. At the decoder side, a similar process is executed for discarding the bins that have been added to obey the delay constraint. Therefore, the decoder maintains a counter C that counts the codewords that have been read from the bit buffer (this counter can be maintained in the bit buffer). This counter C is initialized (e.g. with zero) at the beginning of the decoding of a data packet and is increased by one after a codeword is read. In addition, each bin decoder 22 contains a counter Cx, which stores the value of the codeword counter C before the last codeword was read by the corresponding bin decoder 22. I.e., when a particular bin decoder 22 reads a new codeword, its counter Cx is set equal to C as a first step and then the codeword is read from the bit buffer. When a request for a bin 19 is sent to a particular bin buffer 20 and the difference (C - Cx) between the overall codeword counter C and the counter Cx of the connected bin decoder 22 is greater than the maximum number of codeword buffer entries N, all bins that are currently stored in the particular bin buffer 20 are discarded and ignored. Beside that additional step, the decoding is operated as described above. If the bin buffer 20 to which a request for a bin 19 is sent is empty (either because all bins have already been removed or because the low-delay mechanism did discard all bins in the first step after the bin request has been received), the connected bin decoder 22 reads one or more new codewords from the bit buffer 38 etc.
In another preferred embodiment of the invention, the measure for the delay (or an upper bound of the delay) is the sum of the maximum codeword lengths for the active buffer entries in the codeword buffer, where the maximum codeword length for a particular buffer entry depends on the bin decoded that is associated with that buffer entry. As illustration, the maximum codeword lengths for the buffer entries are indicated in the examples in 6. Note again that the first buffer entry is always a reserved buffer entry or a free buffer entry, since if the first buffer entry contains a codeword, this codeword is written to the bitstream. Let the maximum allowed buffer delay (as determined by the application) be D bits. This maximum buffer delay D must be known to both encoder and decoder. In a preferred embodiment of the invention, the maximum buffer delay D is fixed by the application. In another preferred embodiment of the invention, the maximum buffer delay D is signaled inside the bitstream, e.g., in the header of the data packet (or slice header) or in a parameter set, which is included in the bitstream. It can be signaled in units of bits, or bytes, or a multiple of bits, or a multiple of bytes. If a bin encoder 10 sends a request for the reservation of one or more new buffer entries to the codeword buffer 29, the following process is executed before a new codeword buffer entry is reserved (i.e., it is executed multiple times if multiple codeword buffer entries are reserved by one request).
If the sum of the maximum codeword lengths for all currently active buffer entries plus the maximum codeword length for the buffer entry that will be reserved is greater than the maximum buffer delay D, the first buffer entry (which is reserved) is flushed by the process described above until the sum of the maximum codeword lengths for all active buffer entries plus the maximum codeword length for the buffer entry that will be reserved is less than or equal to the maximum buffer delay D. As an example, let's consider the example in Fig. 6(b). The sum of the maximum codeword lengths for all currently active buffer entries is 29. Let's assume that the maximum buffer delay D is set equal to 32. If the next buffer entry is reserved by bin encoder 2 for which the maximum codeword length is equal to 3, the first buffer entry is not flushed, since 29 + 3 is not greater than 32. But if the next buffer entry is reserved by bin encoder 1 for which the maximum codeword length is equal to 7, the first buffer entry is flushed, since 29 + 7 is greater than 32. The flushing of the reserved buffer entry is done as described above (by adding bin with particular or arbitrary values to the corresponding bin buffer).
At the decoder side, a similar process is executed for discarding the bins that have been added to obey the delay constraint. Therefore, the decoder maintains a counter C that counts the maximum codeword length for the codewords that have been read from the bit buffer (this counter can be maintained in the bit buffer). Note that the maximum codeword lengths that are associated with different bin decoders can be different. The counter C is initialized (e.g. with zero) at the beginning of the decoding of a data packet and it is increased after a codeword is read. This counter is not increased by the actual length of the read codewords, but by its maximum length. I.e., if a codeword is read by a particular bin decoder and the maximum codeword length that is associated with the codeword table used by the particular bin decoder is Lx (a different bin decoder can be associated with a different maximum codeword length), the counter C is increased by Lx. In addition to the overall counter C, each bin decoder 22 contains a counter Cx, which stores the value of the codeword counter C before the last codeword was read by the corresponding bin decoder 22. I.e., when a particular bin decoder 22 reads a new codeword, its counter Cx is set equal to C as a first step and then the codeword is read from the bit buffer. When a request for a bin 19 is sent to a particular bin buffer 20 and the difference (C - Cx) between the overall counter C and the counter Cx of the connected bin decoder 22 is greater than the maximum buffer delay D, all bins that are currently stored in the particular bin buffer 20 are discarded and ignored. Beside that additional step, the decoding is operated as described above. If the bin buffer 20 to which a request for a bin 19 is sent is empty (either because all bins have already been removed or because the low-delay mechanism did discard all bins in the first step after the bin request has been received), the connected bin decoder 22 reads one or more new codewords from the bit buffer 38 etc.
Thus, the plurality of entropy decoders 22 and the selector 18 may be configured to intermittently discard suffixes of symbol sequences so as to not participate in forming the sequence of symbols to be reconstructed. 29. The intermittently discarding may be performed at events where a number of codewords having been retrieved from the codeword entry by the plurality of entropy decoders between two consecutive codeword retrievals of a respective entropy decoder from the codeword entry, fulfils a predetermined criterion. The plurality of entropy encoders and the codeword buffer may, in turn, be configured to intermittently extend currently forwarded but not yet mapped symbols to valid symbol sequences by don't-care symbols having the currently forwarded but not yet mapped symbols as prefix, map the thus extended symbol sequences into codewords, enter the thus obtained codewords into the reserved codeword entries and flush the codeword entries. The intermittently extending, entering and flushing may take place at events where a number of reserved codeword entries plus a number of codeword entries having codewords entered therein fulfils a predetermined criterion. The predetermined criteria may take the maximum lengths of codewords of the plurality of encoder/decoder pairs into account.
For some architectures, the above described preferred embodiment for the codeword interleaving might result in a drawback in terms of the decoding complexity. As illustrated in Fig. 7, all bin decoders 22 read codewords (in the general case, variable-length codewords) from a single bit buffer 38. The reading of the codewords cannot be done in parallel, since the codeword must be read in the correct order. That means, a particular bin decoder must wait until other bin decoders finish the reading of codewords. And when the complexity of the reading of the variable-length codewords is significant in relation to the remainder of the (partially parallelized) decoding process, this access of the variable-length codewords can be a bottleneck for the entire decoding process. There are some variations of the described embodiments of the invention that can be employed for reducing the complexity of the access from the single bit buffer, a few of them will be described in the following. In one preferred embodiment of the invention, there exists a single set of codewords (representing for instance a redundancy-free prefix code) and the set of codewords that is used for each bin decoder 22 is a subset of the single codeword set. Note that different bin decoders 22 can use different subsets of the single codeword set. Even if the codeword sets that are used by some of the bin decoders 22 are the same, their association with bin sequences is different for different bin decoders 22. In a particular embodiment of the invention, the same set of codewords is used for all bin decoders 22. If we have a single codeword set that includes the codeword sets for all bin decoders as subsets, the parsing of the codewords can be done outside the bin decoders, which can reduce the complexity of the codeword access. The encoding process is not changed in relation to the above described process. The modified decoding process is illustrated in Fig. 8. A single codeword reader is fed with bits 46 from the bitstream 40 and parses the - in general variable-length - codewords. The read codewords 44 are inserted into a codeword buffer 43, which represents a first-in-first-out buffer. A bin decoder 22 sends a request for one or more codewords 41 to the codeword buffer 43 and as response to this request, the one or more codewords are removed from the codeword buffer (in sequential order) and send to the corresponding bin decoder 22. Note that with this embodiment of the invention, the potentially complex codeword parsing can be done in a background process and it doesn't need to wait for the bin decoders. The bin decoders access already parsed codewords, the potentially complex codeword parsing is no more part of a request to the overall buffer. Instead already parsed codewords are send to the bin decoders, which can also be implemented in a way that only codeword indices are send to the bin decoders.
Interleaving of fixed-length bit sequences
A further way of reducing the decoder complexity can be achieved when the bin decoders 22 don't read variable-length codewords from the global bit buffer 38, but instead they always read fixed-length sequences of bits from the global bit buffer 38 and add these fixed-length sequences of bits to a local bit buffer, where each bin decoder 22 is connected with a separate local bit buffer. The variable-length codewords are then read from the local bit buffer. Hence, the parsing of variable-length codewords can be done in parallel, only the access of fixed- length sequences of bits has to be done in a synchronized way, but such an access of fixed-length sequences of bits is usually very fast, so that the overall decoding complexity can be reduced for some architectures. The fixed number of bins that are sent to a particular local bit buffer can be different for different local bit buffer and it can also vary over time, depending on certain parameters as events in the bin decoder, bin buffer, or bit buffer. However, the number of bits that are read by a particular access does not depend on the actual bits that are read during the particular access, which is the important difference to the reading of variable-length codewords. The reading of the fixed-length sequences of bits is triggered by certain events in the bin buffers, bin decoders, or local bit buffers. As an example, it is possible to request the reading of a new fixed-length sequence of bits when the number of bits that are present in a connected bit buffer falls below a predefined threshold, where different threshold values can be used for different bit buffers. At the encoder, it has to be insured that the fixed-length sequences of bins are inserted in the same order into the bitstream, in which they are read from the bitstream at the decoder side. It is also possible to combine this interleaving of fixed-length sequences with a low-delay control similar to the ones explained above. In the following, a preferred embodiment for the interleaving of fixed-length sequences of bits is described.
Fig. 9 shows an illustration of the basic encoder structure for the embodiment of the invention that interleaves fixed-length sequences of bits for two or more bin encoders. In contrast to the embodiment depicted in Fig. 5, the bin encoders 10 are not connected with a single codeword buffer. Instead, each bin encoder 10 is connected with a separate bit buffer 48, which stores bits for the corresponding partial bitstream. All bit buffers 48 are connected to a global bit buffer 51. The global bit buffer 51 is connected to a bit writer 53, which removes the bits 52 in coding/decoding order from the global bit buffer and writes the removed bits 54 to the bitstream 55. On a certain events in a particular bit buffer 48 or the connected bin encoder 10 or bin buffer 8, the bit buffer 48 sends a request 49 to the global bit buffer 51 by which a certain number of bits is reserved in the global bit buffer 51. The requests for the reservation of fixed-length bit sequences 49 are processed in sequential order. The global bit buffer 51 represents a first-in-first-out buffer in a certain way; bits that are reserved earlier are earlier written to the bitstream. It should be noted that different bit buffers 48 can reserved a different amount of bits, which can also vary over time based on already coded symbols; but the number of bits that are reserved by a particular request is known at the time at which the request is sent to the global bit buffer.
In a particular embodiment of the invention, the bit buffers 48 and the global bit buffer 51 are operated as described in the following. The amount of bits that is reserved by a particular bit buffer 48 is denoted as Nx. This number of bits Nx can be different for different bit buffers 48 and it can also vary over time. In a preferred embodiment of the invention, the number of bits Nx that are reserved by a particular bit buffer 48 is fixed over time. The reservations for a fixed number Nx of bits 49 are triggered based on the number of bits Mx in the bit buffers 48, the number of bits Nx for the reservation requests, and the associated maximum codeword length Lx. Note that each bin encoder 10 can be associated with a different maximum codeword length Lx. If a bin 7 is sent to a particular bin buffer 8, and the particular bin buffer 8 is empty, and not more than one sequence of Nx bits is reserved in the global bit buffer for the bit buffer 48 that is connected with the particular bin buffer (via a bin encoder), and the difference N - Mx between the number Nx of bits that are reserved by a reservation request of the bit buffer 48 that is connected (via a bin encoder) with the particular bin buffer 8 and the number of bits Mx that are currently present in this bit buffer 48 is less than the maximum codeword length Lx that is associated with the corresponding bin encoder 10, the connected bit buffer 49 sends a request 49 for the reservation of Nx bits to the global bit buffer 51. The global bit buffer 51 reserves Nx bits for the particular bit buffer 48 and increases its pointer for the next reservation. After the Nx bits have been reserved in the global bit buffer, the bin 7 is stored in bin buffer 8. If this single bin does already represent a bin sequence that is associated with a codeword, the bin encoder 10 removes this bin from the bin buffer 8 and writes the corresponding codeword 47 to the connected bit buffer 48. Otherwise (this single bin does already represent a bin sequence that is associated with a codeword), further bins 7 are accepted by the particular bin buffer 8 until the bin buffer 8 contains a bin sequence that is associated with a codeword. In this case, the connected bin encoder 10 removes the bin sequence 9 from the bin buffer 8 and writes the corresponding codeword 47 to the connected bit buffer 48. If the resulting number of bits Mx in the bit buffer 48 is greater than or equal to the number of reserved bits Nx, the Nx bits that were first written to the bit buffer 48 are inserted into the previously reserved space in the global bit buffer 51. For the next bin 7 that is sent to the particular bin buffer 8, the same process as specified above is executed; i.e., it is checked first whether a new number of Nx bits must be reserved in the global bit buffer (if Nx - Mx is less than Lx) and then the bin is inserted into the bin buffer 8, etc. The bit writer writes the fixed-length bit sequences of the global bit buffer in the order in which they have been reserved. If the first fixed-length entry in the global bit buffer 51 contains a fixed-length bit sequence that has been actually inserted in the global bit buffer (i.e., it is not only reserved), the bit writer 53 removes the bits for this bit sequence 52 from the global bit buffer 51 and writes the bits 54 to the bitstream. This process is repeated until the first fixed-length entry in the global bit buffer represents a reserved or a free entry. If the first fixed-length entry in the global bit buffer represents a reserved entry, the bit writer 53 waits until this entry is filled with actual bits before it writes further bits 54 to the bitstream 55.
At the end of a data packet, the bin buffers are flushed as described above. In addition, the bit buffers must be flushed by adding bits with a particular or an arbitrary value until all reserved buffer entries in the global bit buffer are filled and written to the bitstream.
In Fig. 10, two examples for the possible status of the global bit buffer 51 are illustrated. In example (a), a case is illustrated in which different bit buffers/ bin encoders reserve a different number of bits. The global bit buffer contains 3 entries with actually written fixed-length bit sequences and 4 entries with reserved fixed-length bit sequences. The first fixed-length entry already contains actual bits (which must have been just inserted by bit buffer/ bin encoder 2); this entry (i.e., the corresponding 8 bits) can be removed and written to the bitstream. The next entry reserves 10 bits for bin encoder 3, but actual bits haven't been inserted yet. This entry cannot be written to the bitstream; it must be waited until the actual bits are inserted. In the second example (b), all bit buffers/ bin encoders reserved the same number of bits (8 bits). The global bit buffer contains 4 reservations for 8 bit sequences and 3 actually written 8 bit sequences. The first entry contains a reservation for 8 bits for bin encoder 3. Before any new bits can be written to the bitstream, the bit writer has to wait until bit buffer/ bin encoder 3 writes the actual values of the 8 bits into this reserved entry.
Fig. 11 shows an illustration of the basic decoder structure for the embodiment of the invention that interleaves fixed-length sequences of bits. In contrast to the embodiment depicted in Fig. 7, the bin decoders 22 are not connected with a single bit buffer. Instead, each bin decoder 22 is connected with a separate bit buffer 58, which stores bits from the corresponding partial bitstream. All bit buffers 58 are connected to a global bit buffer 61. The bits 62 from the bitstream 63 are inserted into the global bit buffer 61. On a certain events in a particular bit buffer 58 or the connected bin decoder 22 or bin buffer 20, the bit buffer 58 sends a request 59 to the global bit buffer 61 by which a fixed-length sequence of bits 60 is removed from the global bit buffer 61 and inserted into the particular bit buffer 58. The requests for the fixed-length bit sequences 59 are processed in sequential order. The global bit buffer 61 represents a first-in-first-out buffer; bits that are earlier inserted into the global bit buffer are earlier removed. It should be noted that different bit buffers 58 can request a different amount of bits, which can also vary over time based on already decoded symbols; but the number of bits that are requested by a particular request is known at the time at which the request is sent to the global bit buffer. It should be noted that the global bit buffer 61 is not necessarily required, since the codewords could also be directly read from the bitstream. The global bit buffer 61 is mainly included in the illustration for clearly separate different aspects of the processing chain.
In a particular embodiment of the invention, the bit buffers 58 and the global bit buffer 61 are operated as described in the following. The amount of bits that is requested and read by a particular bit buffer 58 is denoted as Nx, it is equal to the amount of bits that is written to the global bit buffer by the corresponding bit buffer at the encoder side. This number of bits Nx can be different for different bit buffers 58 and it can also vary over time. In a preferred embodiment of the invention, the number of bits Nx that are requested and read by a particular bit buffer 58 is fixed over time. The reading of a fixed number Nx of bits 60 is triggered based on the number of bits Mx in the bit buffer 58 and the associated maximum codeword length Lx. Note that each bin decoder 22 can be associated with a different maximum codeword length Lx. If a request for a bin 19 is sent to a particular bin buffer 20, and the particular bin buffer 20 is empty, and the number Mx of bits in the bit buffer 58 that is connected (via a bin decoder) with the particular bin buffer 20 is less than the maximum codeword length Lx that is associated with the corresponding bin decoder 22, the connected bit buffer 58 sends a request 59 for a new sequences of N bits to the global bit buffer 61. As response to this request, the first Nx bits are removed from to global bit buffer 61 and this sequence of Nx bits 60 is sent to the bit buffer 58 from which the request was sent. Finally, this sequence of Nx bits is added to the corresponding bit buffer 58. Then the next codeword 57 is read from this bit buffer, and the connected bin decoder 22 inserts the associated bin sequence 21 into the connected bin buffer 20. As final response to the original request for a bin 19, the first bin is removed from the bin buffer 20 and this decoded bin 25 is sent to the bin buffer selector 18. When the next bin request 19 is sent to the particular bin buffer 20 and the bin buffer is not empty, the next bit is removed from the bin buffer 20. If the bin buffer is empty but the number Mx of bits in the connected bit buffer 58 is greater than or equal to the associated maximum codeword length Lx, the next codeword is read from the bit buffer and a new bin sequence is inserted in the bin buffer, from which the first bit is removed and sent to the bin buffer selector. If the bin buffer is empty and the number Mx of bits in the connected bit buffer 58 is less than the associated maximum codeword length Lx, the next sequence of Nx bits is read from the global bit buffer 61 and inserted into the connected local bit buffer 58, the next codeword is read from the bit buffer, a new bin sequence is inserted in the bin buffer, and the first bin of the sequence is removed and sent to the bin buffer selector. This process is repeated until all source symbols are decoded.
At the end of a data packet, more bins and/or bits than required for decoding the requested source symbols might be inserted into the bin buffer and/or bit buffer. The remaining bins in the bin buffer and the remaining bits in the bit buffer are discarded and ignored.
Interleaving of fixed-length bit sequences with a low-delay constraint
The described embodiment for an entropy encoder and decoder with interleaving of fixed- length bit sequences can also be combined with the scheme for controlling the encoder buffer delay, which is described above. The basic concept is the same as in the embodiment with delay control described above. If a measure related to the delay or an upper bound of the delay (see below) exceeds a specified threshold, the first reserved buffer entry is filled by flushing the corresponding bin buffer (using a similar mechanism as at the end of a data packet) and potentially writing additional bits for filling all bits of the reserved fixed-length buffer entry. By such a mechanism, the number of waiting buffer entries is reduced until the associated delay measure is less than the specified threshold. At the decoder side, the bins and bits that have been inserted at the encoder side in order to obey the delay constraint must be discarded. For this discarding of bins and bits basically the same mechanism as at the encoder side can be used.
In a preferred embodiment of the invention, the measure for the delay (or an upper bound of the delay) is the number of bits in the active buffer entries in the global bit buffer, where the number of active buffer entries is the number of reserved fixed-length buffer entries plus the number of fixed-length buffer entries that contain already written bits. Note that the first buffer entry is always a reserved fixed-length buffer entry or a free buffer entry, since if the first buffer entry contains written bits, these bits are written to the bitstream. Let the maximum allowed buffer delay (as determined by the application) be D bits. This maximum buffer delay D must be known to both encoder and decoder. In a preferred embodiment of the invention, the maximum buffer delay D is fixed by the application. In another preferred embodiment of the invention, the maximum buffer delay D is signaled inside the bitstream, e.g., in the header of the data packet (or slice header) or in a parameter set, which is included in the bitstream. It can be signaled in units of bits, or bytes, or a multiple of bits, or a multiple of bytes. If a bin encoder 10 sends a request for the reservation of a new fixed-length bit sequence to the global bit buffer 51, the following process is executed before a new fixed- length buffer entry is reserved.
If the number of bits in the active buffer entries in the global bit buffer plus the number of bits that will be reserved by the current reservation request is greater than the maximum buffer delay D, the first buffer entry (which is reserved) is flushed by the process described in the following until the number of bits in the active buffer entries in the global bit buffer plus the number of bits that will be reserved by the current reservation request is less than or equal to the maximum buffer delay D. The flushing of a reserved fixed-length buffer entry is similar to the flushing at the end of a data packet: The bin encoder 10 that is connected with the bit buffer 48 that has reserved the corresponding first buffer entry is flushed by adding bins with particular or arbitrary values to the connected bin buffer 8 until the resulting bin sequence represents a bin sequence that is associated with a codeword, the codeword is then inserted into the corresponding bit buffer 48. As mentioned above, one preferred way for adding bins to the bin buffer is to add those bins that produce the shortest possible codeword. If, after the writing of the codeword to the connected bit buffer and a potential insertion of a fixed-length bit sequence into the global bit buffer, there are still bits in the bit buffer (i.e., the written codeword did not completely fill the reserved fixed-length sequence of bits), further bits with particular or arbitrary values are added to the bit buffer until all bits are removed from the bit buffer and written to the reserved buffer entry. Finally, at the end of this process, the completed buffer entry (the first fixed-length entry in the global bit buffer) is removed from the global bit buffer and written to the bitstream. At the decoder side, a similar process is executed for discarding the bins and bits that have been added to obey the delay constraint. Therefore, the decoder maintains a counter C that counts the bits that have been read from the global bit buffer (this counter can be maintained in the global bit buffer). The counter C is initialized (e.g. with zero) at the beginning of the decoding of a data packet and it is increased after a fixed-length sequence of is read. If a fixed-length sequence of Nx bits is read from the global bit buffer 61, the counter C is increased by Nx. In addition to the overall counter C, each bit buffer 58 contains a counter Cx, which stores the value of the bit counter C before the last fixed-length bit sequence was read into the corresponding bit buffer 58. When a particular bit buffer 58 reads a new fixed-length bit sequence, its counter Cx is set equal to C as a first step and then the fixed-length bit sequence is read from the global bit buffer 61. When a request for a bin 19 is sent to a particular bin buffer 20 and the difference (C - Cx) between the overall counter C and the counter Cx of the connected bit buffer 58 is greater than the maximum buffer delay D, all bins that are currently stored in the particular bin buffer 20 and all bits that are stored in the connected bit buffer 58 are discarded and ignored. Beside that additional step, the decoding is operated as described above. If the bin buffer 20 to which a request for a bin 19 is sent is empty (either because all bins have already been removed or because the low-delay mechanism did discard all bins in the first step after the bin request has been received), the connected bin decoder 22 attempts to read a new codeword from the connected bit buffer 58. If the number of bits in the bit buffer 58 is less than the maximum codeword length, a new fixed-length bit sequence is read from the global bit buffer 61, before the codeword is read, etc.
Thus, the above embodiments each represent a novel approach to entropy coding that provides the efficiency and simple modeling capability of arithmetic coding at the complexity level of Huffman coding. The event space of the input sequence of discrete symbols is mapped onto a small set of binary probability intervals. The probability models for the source symbols can be fixed or adaptive while entropy coding using the probability intervals remains fixed and is decoupled from the modeling stage. Each of the probability intervals is coded using a very simple entropy code that has the complexity level of Huffman codes. The excess rate of the event space projection entropy code (ESPEC) is similar to that of arithmetic coding.
Thus, the above embodiments enable a possible way to increase the efficiency of entropy encoding. Entropy coding, in general, can be considered as the most generic form of lossless data compression. Lossless compression aims to represent discrete data with fewer bits than needed for the original data representation but without any loss of information. Discrete data can be given in the form of text, graphics, images, video, audio, speech, facsimile, medical data, meteorological data, financial data, or any other form of digital data. In many coding applications, the original source data are first mapped onto so-called coding symbols and these coding symbols are then entropy coded. The mapping onto coding symbols can include quantization, in which case the overall coding scheme is lossy. A coding symbol s can take any value of an -ary (M≥2 ) alphabet A = {a0,...,aM_x} . For the purpose of coding the symbol s , the alphabet is associated with an estimated probability mass function (pmf) {ρΑαο)> · · · > Ρ, (αΜ-ι )} ^11^ aU dependencies between coding symbols that are not considered in this pmf are neglected. For these abstract settings, the entropy
M-l
H. = -∑P.(.°,)to&*M) (Bl)
=0 is the greatest lower bound for the expected codeword length in bits per symbol, for coding the symbols s , that can be achieved with entropy coding techniques. For decades, Huffman coding and arithmetic coding have dominated practical entropy coding. They are well-known examples of practical codes capable of approximating the entropy limit (in a certain sense).
For a fixed probability distribution, Huffman codes are relatively easy to construct. The most attractive property of Huffman codes is that their implementation can be efficiently realized by the use of variable-length code (VLC) tables. However, when dealing with time-varying source statistics, i.e., changing symbol probabilities, the adaptation of the Huffman code and its corresponding VLC tables is quite demanding, both in terms of algorithmic complexity as well as in terms of implementation costs. Also, in the case of having a dominant alphabet value with ps(at) > 0.5 , the redundancy of the corresponding Huffman code (without using any alphabet extension such as run length coding) may be quite substantial. Another shortcoming of Huffman codes is given by the fact that in case of dealing with higher-order probability modeling, multiple sets of VLC tables may be required.
Arithmetic coding, on the other hand, while being substantially more complex than VLC, offers the advantage of a more consistent and adequate handling when coping with adaptive and higher-order probability modeling as well as with the case of highly skewed probability distributions. Actually, this characteristic basically results from the fact that arithmetic coding provides a mechanism, at least conceptually, to map any given value of probability estimate in a more or less direct way to a portion of the resulting codeword. Being provided with such an interface, arithmetic coding allows for a clean separation between the tasks of probability modeling and probability estimation, on the one hand, and the actual entropy coding, i.e., mapping of symbols to codewords, on the other hand.
Differing from the just-discussed conventional entropy coding schemes, some of the above- described embodiments of the present invention use a kind of event space projection described the mathematical background is described in more detail below.
Consider the sequence of coding symbols {s, 0,.. ,s, w_1} . Each symbol is drawn from an alphabet s, e At. . The alphabets Ai = {a0' , a[,...} contain two or more letters each being associated with a probability estimate ps{am' ) . The probability estimates ps{am' ) are known to encoder and decoder and may be fixed or variable. It is assumed that variable probabilities are simultaneously estimated at encoder and decoder. The alphabets may either be identical for the sequence of symbols or different symbol types are associated with different alphabets. In the latter case, it is assumed that the decoder knows the alphabet of each symbol in the sequence. This assumption is justified as practical source codec descriptions contain a syntax that stipulates the order of symbols and their alphabets.
The sequence of symbols {¾,..., 5·^} is converted into a sequence of binary symbols, which are also referred to as bins. For each symbol st , the binarization
Figure imgf000046_0001
represents a bijective mapping of the alphabet letters am' onto ordered sets of bins b 'm . The binarization mapping γ can be different for different symbols si or symbol categories. Each bin sequence ' for a particular symbol s. consists of one or more bins ¾ . At the decoder side, a symbols st can be reconstructed by the inverse mapping si = (yb' yl(b') given the sequence of bins b ' . As a result of the binarization, a sequence of bins {¾,..., is obtained that represents the sequence of source symbols {s0,...,sN_1} .
All bins Z>y. are associated with the same binary alphabet B = {0,1 } , but the corresponding binary pmfs {pl,p(} , with p( = pQ ] , are usually different. A binary pmf {p0 J,p(} can be described by the less probable bin (LPB) value bJ LPB and its probability pJ LPB (with PJ LPB≤0.5 ). This binary probability description {bj LPB ,pJ LPB } can be directly derived from the probability estimates ps(am' ) for the symbol alphabets given the binarization mappings γ[ . It is also possible (and often preferable) to directly estimate {ύ^ ,ρ^ } simultaneously at encoder and decoder side. Therefore, the bins can be associated with a probability model (which is also referred to as context) based on the syntax and previously coded symbols or bins. And for each probability model, the probability description {bJ LPB ,pj LPB } can be estimated based on the values of the bins that are coded with the probability model. An example for such an binary probability modeling is described with respect to CABAC of H.264.
Since the binary entropy function H(p) = -p \og2{p) - (1 - p)\og2(\ - p) (B3) is symmetric around p=Q.5 , the same binary coder can be used for coding all bins that are associated with the same LPB probability pJ LPB , independent of the value of bJ LPB . Therefore, the sequence of bins {¾,... , b^} is converted into a sequence of coding bins {¾,... , bB c_x } . For each bin Z>;. , the corresponding bijective mapping γα' is specified by
Figure imgf000047_0001
where Θ denotes the exclusive or operator. At the decoder side, the bins b} can be reconstructed given the coding bins bj and the corresponding LPB value bj LPB by the inverse mapping bj = (y^)'1 (bj ) = bj ® bJ LPB . A coding bin bj = 0 specifies that the value of corresponding bin b . is equal to the LPB value bJ LPB and a coding bin b* = 1 specifies that the value of the corresponding bin b} is equal to the more probable bin (MPB) value -bJ LPB .
The sequence of coding bins {¾,... , ¾_,} does uniquely represent the sequence of source symbols {sQ , ... , sN^ } and the corresponding probability estimates, which can be employed for entropy coding, are completely described by the LPB probabilities pJ LPB (with pj LPB < 0.5 ).
Hence, only probabilities in the half-open interval (0,0.5] need to be considered for designing the binary entropy coder for the coding bins bf .
For the actual binary entropy coding, the sequence of coding bins {¾,... , is projected onto a small number of probability intervals Ik . The LPB probability interval (0,0.5] is partitioned into K intervals Ik = {pk,pM] = 0 for k≠j (B5)
Figure imgf000047_0002
The set of K intervals is characterized by K - 1 intervals borders pk with k = \,... , K - 1. Without loss of generality we assume pk < pk+1 for k = 0,... , K . The outer interval borders are fixed and given by p0 = 0 and pK = 0.5 . A simple non-adaptive binary entropy coder is designed for each interval Ik . All coding bin bj with associated LPB probabilities pi LPB e Ik are assigned to the interval Ik and are coded with the corresponding fixed entropy coder.
In the following description, all bins represent coding bins tf and all probabilities p are LPB probabilities pj LPB .
For investigating the impact of the probability interval discretization on the coding efficiency, we assume that we can design an optimal entropy coder for a fixed probability that achieves the entropy bound. Each probability interval Ik = (pk,pk+l] is associated with a representative probability t e Ik and the corresponding optimal entropy coder shall achieve the entropy limit for this representative probability. Under these assumption, the rate for coding a bin with probability p using the optimal entropy coder for the interval representative Pj is given by
R(P, P Y- P log20½ ) - (1 - )log2(l - Pjk )
=H(plk ) + (p - pIk ) H'(pIk ) (B6) where H(p) represents the binary entropy function 3 and
Figure imgf000048_0001
is its first derivative. We further assume that the distribution of the probabilities in the interval
. Then, the expected rate, in bits per bin, for a
Figure imgf000048_0002
given set of K intervals {Ik} with corresponding representative probabilities {p, } can be written as
K-l Pk+i
R = R({Ik},{pIk }) =∑ \ R(p,plk ) f(p) dp (B8)
i=0
The first partial derivative with respect to any representative probability
k = 0,...,K -l , is given by
Figure imgf000048_0003
The equation R = 0 has a single solution
Figure imgf000048_0004
for the representative probability p1 inside the domain of definition Ik . The second partial derivative for this solution
Figure imgf000049_0001
is always greather than zero if
Figure imgf000049_0002
Hence, if condition B12 is fulfilled, the value p) given in eq. BIO is the representative probability for an interval Ik that minimizes the expected overall rate R given the interval boundaries pk and pk+l . Otherwise, no bin is projected to the interval Ik and the representative probability pt e lk can be arbitrarily chosen without any impact on the overall rate R ; but such a configuration should be avoided, since the interval Ik would not be employed for entropy coding.
For finding a condition for optimal interval borders, we investigate the first derivitives of the expected overall rate R with respect to the interval borders pk with k = l,... , K -l . If f (p) > 0 for all p e [pj ,p} ) , the equation R = 0 has a single solution
k
' _ ^ )- ¾¾ ) - ¾H)¾ff¾ )
for the interval border pk inside the domain of definition [_p7 i , pt ) and the second partial derivative for this solution d
R(Pk) = f(p;)( '(plk - H'(plk)) (B14) is always greater than zero, so that
Figure imgf000049_0003
is the interval border pk e [pj ,pIk ) that minimizes the expected overall rate R given the interval representatives Pj ^ and p^ . If there exist probabilities p [pj ,Pj ) with f(p) = 0 , the equation R = 0 has multiple solutions,
*-i k dpk
but pi as given in eq. B13 is still optimal even though further optimal solutions may exist. Given the number of intervals K and the probability distribution / (p) , the interval borders pk , with k = \,... ,K - \ , and the interval representatives p{ , with k - 0,...,K -l , that minimize the expected overall rate R can be obtained by solving the equation system given by eqs. BIO and B13 subject to the conditions B12 for k = 0,... , K - \ . This can be achieved with the following iterative algorithm. Algorithm 1 :
1) Partition the interval (0, 0.5] into K arbitrary intervals Ik = (pk,pk+l] with PQ—0 , pK = 0.5 , and. pk < pk+l for all k = 0,... ,K - l in a way that the conditions B 12 are obeyed for all k = 0,... , K - 1 .
2) Update the representatives pj with k = 0, ... , K - 1 according to eq. B 10
3) Update the interval borders pk with k = 1 , ... , K - 1 according to eq. B 13
4) Repeat the previous two steps until convergence
Figure 12 shows an example for the optimal interval discretization using the described algorithm. For this example, we assumed a uniform probability distribution / (p) = 2 for 0 < p≤0.5 and partitioned the probability interval (0,0.5] into K = 4 intervals. It can be seen that the probability interval discretization leads to a piecewise linear approximation A(p) of the binary entropy function H(p) with A(p)≥ H(p) for all /? e (0,0.5] .
As measure for the impact of the interval discretization on the coding efficiency the expected overall rate increase relative to the entropy limit
Figure imgf000050_0001
can be used. For the particular example of Figure 12 , the expectation value of the entropy H— Jo H(p) f (p) dp is equal to 1/(2 In 2) bit per bin and the rate overhead p is equal to 1.01%. Table 4 lists rate overheads p mi and p ljn for the uniform probability distribution and a linear increasing probability distribution / (p) = 8p with p e (0,0.5] , respectively, for selected numbers of intervals K .
Table 4: Rate overhead vs. the number of probability intervals for the uniform and a linear increasing probability distribution
K 1 2 4 8 12 16 „„, [%] 12.47 3.67 1.01 0.27 0.12 0.07 5.68 1.77 0.50 0.14 0.06 0.04
The investigations in this section showed that the discretization of the LPB probability interval (0, 0.5] into a small number of intervals with a fixed probability (e.g., 8 to 10 intervals) has a very small impact on the coding efficiency.
The above discussed entropy coding for probability intervals, enables thus of individual coders using fixed probabilities.
In the following, we first show how an simple code can be designed for fixed probabilities. Given these results, we develop an algorithms that jointly optimizes the code design and the partitioning of the LPB probability interval (0, 0.5] .
The entropy coding for fixed probabilities p = p} can be done using arithmetic coding or variable length coding. For the latter case, the following approach appears to be simple and very efficient.
We consider a binary entropy coding scheme by which a variable number of bins is mapped onto variable length codewords. For unique decodability, the inverse mapping of a codeword to a bin sequence must be unique. And since we want to design a code that approaches the entropy limit as close as possible, we constrain our considerations to bijective mappings. Such a bijective mapping can be represented by a binary tree where all leaf nodes are associated with codewords, as depicted in Figure 13. The tree edges represent binary events. In the example of Figure 13, the lower edges represent the LPB bin value and the upper edges represent the MPB bin value. The binary tree represents a prefix code for the bins if it is a full binary tree, i.e., if every node is either a leaf or has two descendants. Each leaf node is associated with a probability based on the given LPB probability p . The root node has the probability pwot = 1 . The probability for all other nodes is obtained by multiplying the probability of the corresponding ancestor with p for the LPB descendants and q = l - p for the MPB descendants. Each leaf node L, is characterized by the number of LPB edges a, and the number MPB edges bl from the root node to the leaf node. For a particular LPB probability p , the probability pl for a leaf node is equal to
Pl = P > (Ι -ΡΫ' (B16)
The binary tree T is fully characterized by the number of leaf nodes L and the associated pairs with / = 0,...,J -1.
Given a full binary tree T and a LPB probability p , the optimal assignment of codewords to the leaf nodes can be obtained by the Huffmann algorithm. The resulting variable number of bits to variable length codewords (VNB2VLC) mapping C is characterized by the number of codewords L , which is identical to the number of leaf nodes, and the tuples {α,,δ,,^} for
/ = 0,... , Z -1 , where lt represents the codeword length that is associated with the corresponding leaf node Z, = {al,bl} . It should be noted that there are multiple possibilities for the codeword assignment given the codeword lengths {/,} and the actual codeword assignment is not important as long as the codewords represent a uniquely decodable prefix code. The expected rate R(p,C) in bits per bin for a given code C and an LPB probability p is the ratio of the expected codeword length and the expected number of bins per codeword
R{p,C) = T- ^ = 7-r =2 (B17)
∑ ,(«, + *,) ∑ ¾ (1 - / ¾ («, + ¼)
1=0 1=0
The code design is often limited by factors as the maximum number of codewords L , the maximum number of bins per codeword, or the maximum codeword length, or it is restricted to codes of particular structures (e.g., for allowing optimized parsing). If we assume that the set Sc of usuable codes for a particular application is given, the optimum code C* e Sc for a particular LPB probability p can be found by minimizing the expected rate R(p,C)
C*(/>) = arg min ii(p, C) (B18)
vc<=sc
As faster alternative, the minimization can also proceed over a given set of binary trees ST and for each tree only one VNB2VLC code C that is obtained by the Huffmann algorithm is considered. As an example, we designed VNB2VLC codes for various LPB probabilities p by considering all binary trees T for which the number of leaf nodes L is less than or equal to a given maximum Lm . In Figure 14, the relative rate increase p(p,C*(p)) = R(p,C*(p))/H(p) is plotted over the LPB probability p for selected maximum table sizes Lm . The rate increase p(p) can usually be reduced by allowing larger table sizes. For larger LPB probabilities, a small table size L of 8 to 16 codewords is usually sufficient for keeping the rate increase p(p) reasonably small, but for smaller LPB probabilities (e.g., p < 0.1 ), larger table sizes L are required.
In the previous sections, we considered the optimal probability discretization assuming optimal codes and the code design for fixed LPB probabilities. But since, in general, we cannot achieve the entropy limit with real VNB2VLC codes of limited table sizes, the code design and the partitioning of the LPB probability interval (0, 0.5] must be jointly considered for optaining an optimized entropy coding design. For a given interval Ik = (pk,pk+l], a code Ck of a given set Sc in an optimal code Ck if it minimizes the e
Figure imgf000053_0001
For practical designs, the minimization of the integral in eq. B19 can be simplified, with a minor impact on the coding efficiency, by first determining an optimal representative probability p) for the interval Ik according to eq. BIO and then choosing the optimal code
Ck * of the given set Sc for the representive probability p) according to eq. B 18.
Optimal interval borders pk, with k = \,...,K -I, given the set of codes Ck, with k = 0,...,K-l, can be derived by minimizing the expected overall rate
K-l Pk+l
R = R({pk},{Ck}) =∑\ \ R(P,Ck)f(p)dp (B20)
k=0
Setting the first derivatives with respect to the interval borders equal to zero, if = 0, for k = I,..., K-l, yields p,= pk with Rip^C^Rip^C,) (B21)
Similarly as for eq. B13, it can be shown that pk * is always an optimal solution, but depending on the probability distribution / (p) further optimal solutions might exist. Hence, an optimal interval border pk' between two intervals lk_x and Ik with given associated codes Ck_x and Ck , respectively, is the intersection point of the functions R p,Ck_l) and R p,Ck) .
Consequently, the following interative algorithm can be used for jointly deriving the probability interval partitioning and the associated codes given the number K of probability intervals, the set of possible codes Sc, and the probability distribution f(p), with pe (0,0.5]. Algorithm 2:
1) Derive initial probability interval boundaries pk, with k = 0,...,K, using algorithm 1 specified in sec.3
2) Derive representatives pj for the probability intervals Ik, with k = 0, ... , K - 1 , according to eq. B 10
3) Derive codes Ck e Sc for the interval representatives p} , with k = Q,... , K - \ , according to eq. B18
4) Update the interval borders pk , with k = \,... , K - I , according to eq. B21
5) Repeat the previous three steps until convergence
The steps 2 and 3 in algorithm 2 could also be replaced by a direct derivation of the codes Ck G Sc , with k = 0,... , K - i , based on the interval borders pk , with k = 0, ... , K , according to eq. B 19. And, as mentioned in sec. 4.1, the minimization in step 3 can also proceed over a given set of binary trees ST where for each binary tree T only one VNB2VLC code Ck obtained by the Huffmann algorithm is considered.
As an example, we jointly derived the partitioning into K = 12 probability intervals and corresponding VNB2VLC codes using algorithm 2. At this, the minimization in step 3 of the algorithm was replaced with an equivalent minimization over a given set of binary trees ST where the evaluated code C for each tree T was obtained by the Huffmann algorithm. We considered trees T with a maximum number of Lm = 65 leaf nodes and hence codes C with up to 65 table entries. All binary trees T with up to 16 leaf nodes have been evaluated in the minimization; for trees with more than 16 leaf nodes, we employed a suboptimal search given the best results for trees with a smaller number of leaf nodes.
In Figure 15 the expected rate increase relative to the entropy limit AR(p) = R(p) - H(p) for the code design example is plotted over the LPB probability p . As comparison, we also plotted the expected rate increase AR for the theoretically optimal probability interval discretization (as developed in sec. 3) and the theoretically optimal probability discretization with the additional constraint p} = 0.5 inside the diagram. It can be seen that the joint probability interval discretization and VNB2VLC code design leads to a shifting of the interval borders (the interval borders pk , with k = \ ,... , K - 1 , are given by the local maxima of the AR(p) curves). The relative expected overall rate increase relative to the entropy limit for the design example with real VNB2VLC codes is p = 0.24 %, when assuming a uniform probability distribution / (p) . The corresponding relative rate increases for the theoretically optimal probability interval discretization and the theoretically optimal probability discretization with the additional constraint p, = 0.5 are = 0.12 % and > = 0.13 %, respectively.
Codeword termination can be done as follows. When coding a finite sequence of symbols {¾,... ,¾_,} , each of the K binary encoders processes a finite sequence of coding bins b 1= {bo ,... ,bB c _l}k , with k = 0,... , K - 1 . And it has been to be ensured that, for each of the K binary encoders, all coding bins of the sequence b k c= {bQ°,... ,bB c i )k can be reconstructed given the codeword or sequence of codewords ^(b^) .
When employing arithmetic coding, the arithmetic codeword for the sequence of coding bins has to be terminated in a way, that all coding bins can be decoded given the codeword. For the VNB2VLC codes described above, the bins at the end of the sequence bc k may not represent a bin sequence that is associated with a codeword. In such a case, any codeword that contains the remaining bin sequence as prefix can be written. The overhead can be minimized, if the corresponding codeword that has the minimum length (or one of these codewords) is chosen. At the decoder side, the additionally read bins at the end of the bin sequence, which can be identified given the bitstream syntax and binarization schemes, are discarded.
A simple code design example is presented below. For illustration purposes, we consider the simple example of a source {s) with three letters and fixed associated probabilities of ps(a0) = 0.7 , ps (<¾) = 0.18 , and ps(a2) = 0.12 . The corresponding ternary choice tree can be converted into a full binary tree as shown in Fig. 15 .
A binarization for the full binary tree in Fig. 15 is given in Tab. 5 . The ternary symbol pmf ps is converted into two binary pmfs pb = (0.7, 0.3) and p^ = (0.6,0.4) . For each symbol s in the bit stream, the bin b0 is present. When b0 is equal to 0 , also bx is present. Note that the binarization given in Tab. 2 is identical to an optimal single-letter Huffman code for the source s .
Table 5: Binarization of a three letter source. The LPB probabilities p LPB are 0.3 for the first bin and 0.4 for the second bin
Figure imgf000055_0001
The entropy for the source s is
#(0.7,0.18, 0.12) = 1.1726 bitl symbol (B22) The average code word length of the single-letter Huffman code is given as
£HC = ¾ fc = 1.3 bitl symbol (B23)
;=0 corresponding to a redundancy of pHC = 0.1274 bit/symbol or 10.87% expected rate overhead.
For the particular binarization example with fixed pmfs, the bins b0 and bx already represent coding bins, since for both bins the LPB value bJ LPB is equal to 0. The distribution f(s) of the LPB probabilities is discrete, with / (p) = 0 except for p = 0.3 and p = 0.4 .
Consequently, the optimal probability discretization leads to K = 2 intervals with the representatives _p; = 0.3 and pj = 0.4 . The interval border p1 between these intervals can be arbitarily chosen in [0.3,0.4) .
For encoding the source, he sequence of source symbols is binarized into a sequence of bins. The bin b0 is transmitted for every source symbol. The bin bx is only transmitted when b0 - 0 . The bins b0 and bx are coded separately with constant LPB probabilities of pj = 0.3 and Pj = 0.4 , respectively.
An efficient coding of a binary alphabet with fixed probability can be achieved by a simple
VNB2VLC mapping. Examples for VNB2VLC mappings with small coding tables for the LPB probabilities p LPB = 0.3 and p LPB = 0.4 are given in Tab. 6 and Tab. 7 , respectively.
The VNB2VLC mapping for p LPB = 0.3 yields a redundancy of 0.0069 bit/bin or 0.788% .
For the LPB probability of p LPB = 0.4 , the redundancy is 0.0053 bit/bin or 0.548% .
Table 6: Bin tree and codes for an LPB probability of p LPB = 0.3 . The redundancy of this code is 0.788%
Bin Tree Probability Codes
'1 1' 0.72 = 0.49 '1'
or '01'
0.7 - 0.3 = 0.21
'0' 0;3 '00' Table 7: Bin tree and codes for an LPB probability of p LPB = 0.4. The redundancy of this code is 0.548%
Figure imgf000057_0001
The overall expected rate incurred by the new coding method is lNC =lbo +o.3-lbi
= 1.181 W// f symbol (B24)
The overall redundancy is 0.73% relative to the entropy limit, which represents a significant improvement in comparison to the single-letter Huffman code.
It could be argued that a similar coding efficiency improvement could be obtained by creating a run-length code. For the above example, we could construct a run-length code for the most probable symbol by considering runs of up to two symbols. Each of the events {α0α00αχ0α2λ2} would be associated with a separate codeword. Such a code yields a redundancy of 1.34% relative to the entropy limit. Actually, the VNB2VLC codes can be considered as a generalization of run-length codes for binary symbols (the VNB2VLC code in Tab. 3 does effectively represent a run-length code). For a single symbol alphabet with fixed probabilities, a similar coding efficiency as for the presented approach can also be achieved by creating a code that maps a variable number of source symbols to variable length codewords. The main advantage of the presented approach is its flexibility in mapping arbitrary source symbol sequences with fixed or adapative probability estimates to a small number of simple binary coders that are operated with fixed LPB probabilities.
How to achive an unique decidability is considered next.
With the presented entropy coding scheme, the coding of a sequence of source symbols s = {sQ , ... , sN_} } consists of the following three basic steps:
• symbol binarization b = {¾,...,¾_,} = b(s) yielding a sequence of bins b = {b0,... ,bB_l} • conversion of the sequence of bins into a sequence of coding bins b °- {¾,... , bB_t} = c(b)
• binary entropy coding of the sequence of coding bins b c= {bQ° , ... , bB c_x } using probability interval discretization and K fixed binary coders
The symbol sequence 5 = {J03... , sN_x} is uniquely decodable, if the sequence of coding bins b °= {b0 c,... ,bB c_x} is uniquely decodable and the mappings b and yc are invertible.
Let ye notify the encoder mapping of a sequence of one or more coding bins b c= {b0°,...} onto a sequence of one or more codewords c(bc)= {c0 , ...} c(bc)= re(V) (B25)
For unique decodability of a sequence of coding bins bc given the sequence of codewords c(bc), the encoder mapping ye must have the property that a unique codeword c(bc) is assigned to each possible sequence of coding bins bc:
V Z>iC, b?≠b? → c(bic)≠ c(bf) (B26)
This property is always fulfilled when arithmetic codes or prefix codes are used. It is particularly fulfilled for the VNB2VLC codes described in sec. 4.1 (including the codeword termination described in sec. 4.3), since the VNB2VLC codes represent prefix codes for variable numbers of bins.
However, in the presented entropy coding approach, the sequence of coding bins b 0 is partitioned into K sub-sequences \ , with k = 0,... , K - 1 ,
{ bQ ... , b^} = Yp{bc) (B27) and to each of the sub-sequences b °k , a sequence of codewords Ck(Ak°) is assigned using a particular encoder mapping y * . Consequently, the condition on unique decodability has to be extended. A sequence of coding bins b 0 is uniquely decodable given K sequences of codewords c^b^), with k - 0,...,K- \ , if each sub-sequence of coding bins b^ is uniquely decodable given the corresponding codeword Ck(6k°) and the partitioning rule γ is known to the decoder. The partitioning rule γρ is given by the LPB probability interval discretization {7^} and the LPB probabilities pJ LPB that are associated with the coding bins bj , with 7 = 0,... ,5 - 1 . Hence, the LPB probability interval discretization {Ik} has to be know at the decoder side and the LPB probability pJ LPB for each coding bin bj , with j = 0,... , B - l , has to derived in the same way at encoder and decoder side. For the mapping yc of a sequence of bins onto a sequence of coding bins 0 , each single r , with j = 0,... , B -\ , is converted by the binary mapping b} c = yc J (bj) = bj ® bJ LPB . At the decoder side, the bin sequence can be derived by the binary mappings
Figure imgf000059_0001
with = 0,... , 5 -1 . If the LPB value bJ LPB for each bin b . is derived in the same way at encoder and decoder side, these mappings (yc' )_1 represent the inverses of the corresponding encoder mappings yc J , since b) Θ VLPB = bj Θ bj LPB Θ b{PB = bj ® 0 = bJ (B29) and hence, the conversion yb of a sequence of bins b into a sequence of coding bins bc is invertible.
Finally, we investigate the invertibility of the binarization b=yb($) by which each symbol si , with = 0,... ,N - 1 , is mapped onto a bin sequence b '= (-?,.) . A symbol s,. can be uniquely decoded given the corresponding bin sequence b ' if the binarization mapping γ[ assigns a different bin sequence b J m to each letter am' of the alphabet Ai for the symbol sj . However, this condition is not sufficient, since the partitioning of the bin sequence b = {¾,... , bB_x } into bin sequences b ' that correspond to the symbols , with z = 0,... ,N - 1 , is not known to the decoder. A sufficient condition is given, when for each symbol s,. , the bin sequences b J m that are associated with the letters am' of the corresponding alphabet Ai form a prefix code and the binarization mappings γ for each symbol si , with = 0,...,N - 1 , are known at the decoder side.
The conditions of unique decodability for the presented entropy coding approach can be summarized as follows:
• the binarization mappings γ represent prefix codes and are known to the decoder (in symbol coding order)
• the probability models b- 'ΙΡΒ , p- 'LPB ) for all bins bj are derived in the same way at encoder and decoder side
the partitioning of the LPB probability interval (0,0.5] into K intervals Ik , with k = 0,... , K - 1 , is known to the decoder
the mapping ye k for each probability interval Ik , with k = 0,... , K - l , represents a uniquely decodable code
In the following, we describe examples for the overall encoder and decoder design in more detail. We concentrate on coding schemes, in which the probability models {b LPB , p LPB } for the bins are directly estimated at encoder and decoder side and the K binary coders use
VNB2VLC mappings described in sec. 4.1. Each source symbol s shall be associated with a symbol category cs , which determines the type of the symbol including its range of values.
The order of symbols and associated symbol categories shall be given by the syntax, which is presumed to be known at encoder and decoder side.
The block diagram for an example encoder and decoder design is illustrated in Figure 17. At the encoder side, the symbols s with associated symbol categories cs are fed into the binarizer, which converts each symbol s into a sequence of bins b = Yb°s (s) .
The used binarization scheme
Figure imgf000060_0001
is determined based on the symbol category cs . In addition, the binarizer associates each bin b of a bin sequence s with a probability model indication cb , which specifies the probability model that is used for coding the bin b . The probability model indication cb can be derived based on the symbol category cs , the bin number of the current bin inside the bin sequence s , and/or the values of already coded bins and symbols.
The probability estimator and assigner maintains multiple probability models, which are characterized by pairs of values {b LPB , p LPB } . It received bins b and associated probability model indications cb from the binarizer, and forwards the LPB value b LPB and the LPB probability p LPB of the indicated probability model to the coding bin deriver and the probability quantizer, respectively. Thereafter, the corresponding probability model {b LPB , p LPB } is updated using the value of the received bin b .
The coding bin deriver receives bins b and associated LPB values b LPB from the binarizer and the probability estimator and assigner, respectively, and sends coding bins bc , which are derived by b° = b ®b LPB , to the probability quantizer. The probability quantizer forwards each coding bin b° to one of the K binary encoders. It contains information about the LPB probability interval quantization {Ik} . The LPB probability p LPB , which is associated with a coding bin bc and received from the probability estimator and assigner, is compared to the interval borders {pk} and probability interval index k , for which p LPB e lk , is derived.
Then, the coding bin bc is forwarded to the associated binary encoder.
Each of the K binary encoders consists of a bin buffer and a bin encoder. The bin buffer receives coding bins b° from the probability quantizer and stores them in coding order. The bin encoder implements a particular VNB2VLC mapping and compares the bin sequence in the bin buffer with the bin sequences that are associated with codewords. If the bin sequence in the bin buffer is equal to one of those bin sequences, the bin encoder removes the bin sequence {bc} from the bin buffer and writes the associated codeword ({ba}) to the corresponding codeword stream. At the end of the encoding process for a symbol sequence, for all binary encoders for which the bin buffers are not empty, a terminating codeword is written as described in sec. 4.3.
The K resulting codeword streams can be separately transmitted, packetized, or stored, or they can be interleaved (cp. sec. 6.2) for the purpose of transmission or storage.
At the decoder side, each of the K binary decoders consisting of a bin decoder and a bin buffer receives one codeword stream. The bin decoder reads codewords ({bc}) from the codeword stream and inserts the associated bin sequence {bc} , in coding order, into the bin buffer.
The decoding of the symbol sequence is driven by the underlying syntax. Requests for a symbol s are sent together with the symbol category cs to the binarizer. The binarizer converts these symbol requests into request for bins. A request for a bin is associated with a probability model indication cb , which is derived in the same way as in the encoder, and sent to the probability estimator and assigner. The probability estimator and assigner is operated similar to its counterpart at the encoder side. Based on the probability model indication cb , it identifies a probability model and forwards its LPB value b LPB and LPB probability p LPB to the bin deriver and the probability quantizer, respectively.
The probability quantizer determines one of the K binary decoders based on the LPB probability p LPB , in the same way as the binary encoder is determined at the encoder side, removes the first coding bin b° , in coding order, from the corresponding bin buffer, and forwards it to the bin deriver. The bin deriver receives coding bins bc and associated LPB values b LPB from the probability quantizer and probability estimator and assigner, respectively, and determines the bin values b = bc Θ b LPB . As final response to a bin request sent by the binarizer, the bin deriver send the decoded bin value b to the binarizer and the probability estimator and assigner.
In the probability estimator and assigner, the value of the decoded bin b is used to update the probability model ,{b LPB , p LPB } , which was chosen by the associated value cb , in the same way as at the encoder side. Finally, the binarizer adds the received bin b to the bin sequence s which has been already received for a symbol request and compares this bin sequence 4 with the bin sequences that are associated with symbol values by the binarization scheme . If the bin sequence s matches one of those bin sequences, the corresponding decoded symbol s is output as final response to the symbol request. Otherwise, the binarizer sends further bin requests until the symbol s is decoded.
The decoding of a symbol sequence is terminated if no further symbol requests, which are driven by the syntax, are received. The coding bins b° that may be contained in the bin buffers at the end of the entropy decoding process (as a result of termination codewords) are discarded.
Summarizing the above embodiments, same represent a novel approach to entropy coding. The mapping of a sequence of symbols to a bit stream and the inverse mapping are described. Each symbol carries associated parameter(s) with it that are simultaneously known at encoder and decoder. The entropy codec contains multiple first-in first-out (FIFO) buffers with each of them assigned to subsets of parameter(s) that are associated to the symbols. For given parameter(s) of a symbol, the encoder assigns the symbol to the corresponding FIFO buffer. Since the encoder assignment rule is known at the decoder side, the decoder reads from the FIFO buffer to which the encoder has assigned the symbol.
One example for a parameter assignment to a symbol is an estimate of its probability. The probability estimate may be given or derived from a measurement that can be conducted simultaneously at encoder and decoder. A particular FIFO buffer contains symbols with estimated probability values falling into a subset of probabilities that is chosen, so that the entropy coding can be improved. The improvement may relate to bit rate reduction, complexity reduction, delay reduction or error resilience.
Another example of a parameter assignment to symbols could be an estimate of the symbol probability and another parameter including one or more of the following: semantic of the symbol, its importance, its location (temporal, spatial, volumetric, etc), its channel code protection. A particular FIFO buffer contains symbols related to a subset of probabilities that is chosen, so that the entropy coding can be improved and a benefit can be achieved with regard to the other parameters.
After having described some embodiments rather generally with respect to the data compressed, embodiments are described according to which the above-mentioned coding is used for compressing video data, i.e. is employed in video and/or picture coding.
EXAMPLE 1 The first embodiment is concerned with Coding Transform Coefficients in Picture and/or Video Coders and Decoders and describes a coding of transform coefficients which can particularly be employed as a novel efficient method for coding transform coefficients in the field of video coding, replacing the corresponding part of CABAC in H.264/AVC, for example.
In earlier hybrid block-based standards for video coding, such as, for example, MPEG-2, H.263 and MPEG-4, the blocks of quantized transform coefficients (levels) are mapped by a defined scan process to a vector which is coded to code words having a variable length by using a run-length coding and subsequent mapping.
In MPEG-2, the code words having a variable length are associated to two-dimensional events (RUN, LEVEL), wherein LEVEL represents the quantized value of a (significant) transform coefficient not quantized to zero; the run-length RUN indicates the number of subsequent (non-significant) transform coefficients quantized to zero, which, in the vector of transform coefficients, are immediately in front of the present significant transform coefficients. In addition, code words having a variable length are defined for the two special events EOB and ESCAPE. While the EOB event indicates that there are no further significant transform coefficients in the block, the ESCAPE event signals that the existing event (RUN, LEVEL) cannot be represented by the defined alphabet of code words having a variable length. In this case, the symbols RUN and LEVEL are coded by code words of a fixed length.
In the newer coding standards H.263 and MPEG-4, the association of code words having a variable length takes place on the basis of three-dimensional events (LAST, RUN, LEVEL), wherein the binary symbol LAST indicates whether the present significant transform coefficient is the last significant coefficient within the block or whether further significant transform coefficients follow. By using these three-dimensional events, no additional EOB event is required; an ESCAPE event is used in analogy to MPEG-2, wherein the binary symbol LAST is coded additionally to RUN and LEVEL.
The coding of the transform coefficients realized in MPEG-2, H.263 and MPEG-4 has the following disadvantages:
To each coding event only a code word having an integer length can be associated, an efficient coding of events with probabilities larger than 0.5 cannot take place.
The usage of a fixed table for mapping the coding events to the code word having a variable length for all the transform coefficients within a block does not consider symbol statistics depending on position or frequency.
Adaptation to the actually existing symbol statistics is not possible.
No usage is made of inter-symbol redundancies present. Annex E of the H.263 standard specifies an optional non-adaptive arithmetic coding in which different predetermined model probability distributions are used,
one each for the first, second and third event (LAST, RUN, LEVEL)/ESCAPE another one for all the following events (LAST, RUN, LEVEL)/ESCAPE of a block of transform coefficients,
as well as one each for the symbols LAST, RUN and LEVEL, which are coded after an ESCAPE event.
For the following reasons no appreciable increase of the coding efficiency is, however, possible by this optional arithmetic coding:
The advantage of arithmetic coding that a code word having a non-integer length can be associated to a coding event hardly has any effect on the coding efficiency by using combined events of the form (LAST, RUN, LEVEL).
The advantage of using different probability distributions is eliminated by the fact that adaptation to the actually present symbol statistics is not possible.
In H.264/AVC, a context-adaptive method on the basis of code words having a variable length for coding transform coefficients is specified as the standard method for entropy coding. Here, the coding of a block of transform coefficients is determined by the following characteristics:
Both the number of the significant coefficients within a block and the number of subsequent coefficients quantized to one at the end of the vector of transform coefficients are determined by a symbol COEFF_TOKEN. Depending on the block type and the symbols COEFF_TOKEN already coded/decoded for neighboring blocks, one of five defined code word tables is chosen for coding.
While for the transform coefficients quantized to one at the end of the coefficient vector only a single bit is transferred for specifying the sign, the coding of the values (levels) of the remaining significant transform coefficients takes place in a reverse scan order by means of a combined prefix suffix code word.
If the number of significant transform coefficients is smaller than the number of transform coefficients for the corresponding block, a symbol TOTAL ZEROS will be coded, which indicates the number of transform coefficients quantized to zero which, in the coefficient vector, are in front of the last significant coefficient. For this, 18 code word tables have been specified, which are switched depending on the number of significant coefficients and the block type.
The run-length of the (non-significant) coefficients quantized to zero (RUN) in front of a significant coefficient is coded for each significant transform coefficient in a reverse scan order as long as the sum of RUNs already coded is smaller than TOTAL_ZEROS. Depending on TOTAL_ZEROS and the RUNs already coded/decoded, switching between seven code word tables takes place.
Although this so-called CAVLC method (CAVLC: context-adaptive variable length coding), by context based switching the code word tables, allows a considerably more efficient coding of the transform coefficients than the methods specified in MPEG-2, H.263 and MPEG-4, it basically has the following disadvantages:
Switching between different code word tables takes place depending on symbols already coded/decoded, the code word tables, however, cannot be adjusted to the actual symbol statistics.
By using code words having a variable length, events having symbol probabilities larger than 0.5 cannot be coded efficiently. This limitation especially prevents coding symbols having a smaller value range, by means of which a construction of suitable contexts might be possible for switching between different model probability distributions.
A possible solution for avoiding the disadvantages illustrated of well-known methods for coding transform coefficients in block-based picture and video coders is a combination of an adaptive entropy coding and a suitable context generation for using the inter-symbol redundancies.
Compared to the above-mentioned coding schemes, the application example described next represents scheme for coding transform coefficients in picture and/or video coders and decoders, which eliminates the deficiencies mentioned above and, in particular, keep the amount of calculating required for coding small. By combining this coding with the entropy coding schemes described in the embodiments with respect to Fig. 1 to 17, a very effective coding efficiency results.
In accordance with a first aspect, the coding scheme described below provides a method for coding transform coefficients in picture and/or video coders and decoders, wherein for blocks of (video) pictures containing significant transform coefficients, the coding of transform coefficients takes place in such a way that, for each block,
in a scan process, the positions of significant transform coefficients in the block and subsequently,
in a reverse scan order - starting with the last significant transform coefficients within the block - the values (levels) of the significant transform coefficients
are determined and coded. One preferred embodiment is characterized by each significant transform coefficient of the block other than the last transform coefficient of the block being characterized by a one-bit symbol.
It is also of advantage if the sign for each significant transform coefficient is indicated by a one-bit symbol (SIGN) and the magnitude is indicated by a binary coded symbol (ABS).
In one preferred embodiment, blocks containing significant transform coefficients are characterized by a one-bit symbol CBP4 in connection with other syntax elements, such as, for example, CBP or macro block mode.
It is a special advantage of the method that, by transferring a one-bit symbol SIG for each coefficient of a block and of a one-bit symbol LAST for each significant coefficient of a block, a significance mapping is coded, wherein the transfer takes place in the scan order, SIG serves for identifying significant coefficients and LAST indicates whether there are further significant transform coefficients in the block.
In another preferred embodiment, modeling for the one-bit symbol CBP4, for coding the significance mapping and/or for coding the coefficient magnitudes takes place in a context- dependent way. Thus, block types of transform coefficients having comparable statistics are summarized to block categories.
It is also of advantage that, in a special embodiment of the inventive method, no significance information (SIG, LAST) is transferred for the last scan position of a block.
In a preferred embodiment, the magnitude (ABS) is indicated by a symbol in unary binarization or by a symbol having a prefix part and a suffix part, wherein the prefix part consists of ones and the suffix part is coded in a 0th order exp-golomb code.
In accordance with a second aspect, the current embodiment provides an arrangement having at least one processor and/or chip formed such that method for coding transform coefficients in picture and/or video coders and decoders can be performed, wherein for blocks of (video) pictures containing significant transform coefficients, the coding of transform coefficients takes place in the just-mentioned way.
In accordance with a third aspect, the current application example provides a computer program enabling a computer, after having been loaded into the memory of the computer, to perform a method for coding transform coefficients in picture and/or video coders and decoders, wherein for blocks of (video) pictures containing significant transform coefficients, the coding of transform coefficients takes place in the way described above.
Such computer programs can, for example, (for a fee or free of charge, freely accessible or protected by a password) be provided in a data or communication network for downloading. The computer programs provided in this way can then be utilized by a method in which a computer program according to claim 11 is downloaded from a network for transfer, such as, for example, from the Internet, to data processing means connected to the network.
In accordance with a fourth aspect, the current embodiment provides a computer-readable storage medium on which a program is stored, enabling a computer, after having been loaded into the memory of the computer, to perform a method for coding transform coefficients in picture and/or video coders and decoders, wherein for blocks of (video) pictures containing significant transform coefficients, the coding of transform coefficients takes place in the described way.
The method for coding transform coefficients is especially characterized by the following characteristics:
- A two-dimensional block of transform coefficients is mapped to a one-dimensional vector by a scan process.
- The syntax elements of the EOB symbol, the LAST symbol or the coefficient counter (number of significant coefficients) as well as RUN (number of non-significant coefficients in the scan order) used in well-known methods are replaced by a one-bit symbol CBP4 and a significance mapping.
The levels (magnitudes of the significant coefficients) are coded in a reverse scan order.
Context modeling takes place in a novel manner.
Figure 18 illustrates the coding method. For each block of transform coefficients, a one-bit symbol CBP4 is transferred at first, unless higher order syntax elements (CBP or macro block mode) already indicate that the block considered does not contain any significant transform coefficients. The CBP4 symbol will be zero if there are no significant coefficients in the block. If it is one, a significance mapping specifying the position (in scan order) of the significant transform coefficients will be coded. Subsequently, the magnitude and the signs of the significant coefficients are transferred in a reverse scan order. A detailed description of the coding process will be given below in 1. Afterwards, context modeling for the binary entropy coding will be described in 2, with the entropy coding of the bins being performed with any of the above-outlined embodiments for entropy coding. 1. Description of the coding of the transform coefficients
1.1 Scanning transform coefficients
The transform coefficients of each block are mapped to a vector by means of a scan process (such as, for example, a zig zag scan).
1.2 CBP4 symbol
CBP4 is a one-bit symbol indicating whether there are significant transform coefficients (transform coefficients unequal to zero) in a block. If the CBP4 symbol is zero, no further information for the corresponding block will be transferred.
1.3 Significance mapping
If the CBP4 symbol indicates that the corresponding block contains significant coefficients, a significance mapping will be coded. This takes place by transferring a one-bit symbol (SIG) for each coefficient in the scan order. If a corresponding significance symbol is one (significant coefficient), another one-bit symbol (LAST) will be transferred. This symbol indicates whether the present significant coefficient is the last significant coefficient within a block or whether further significant coefficients follow. Figure 19 shows two examples of the method described for coding the significance mapping. Significance information (SIG, LAST) will never be transferred for the last scan position of a block. If the transfer of the significance mapping has not already been terminated by a LAST symbol of one, it is obvious that the coefficient at the last scan position is significant (see yellow marked position in Fig. 19).
1.4 Level information
The positions of the significant transform coefficients within a block are clearly specified by the significance mapping. The coding of the precise values of the coefficients (levels) takes place by two coding symbols: ABS (magnitude of the coefficients) and SIGN (sign of the coefficients). While SIGN represents a one-bit symbol, binarization according to Figure 20 is used for coding the magnitudes of the coefficients (ABS). For coefficient magnitudes in the interval [1; 14], this binarization corresponds to a unary binarization. The binarization for coefficient magnitudes larger than 14 consists of a prefix part consisting of 14 ones and a suffix part representing a 0th order exp-golomb code for the symbol (ABS- 15). Binarization does not include a representation for coefficient magnitudes (ABS) equaling 0, since significant coefficients (coefficients unequal to zero) will always have a magnitude (ABS) larger than or equal to one. The binarization formed of a prefix part and a suffix part consisting of a 0th order exp-golomb code for coefficient magnitudes larger than 14 has the advantage that a special non-adaptive context with symbol probabilities 0.5 can be used without sacrificing in coding efficiency for all the binary decisions of the suffix part, whereby the amount of calculating for encoding and decoding can be reduced.
The levels are coded in a reverse scan order - beginning with the last significant coefficient within the block; this enables forming suitable contexts for the binary entropycoding.
2. Context modeling
The above coding of the significance map and the value binarizations results in a sequence of symbols which is de/encoded by any of the above embodiments wherein the assigner assigns the parameter for partial bitstream selection based on the probabilities associated with the contexts which, in turn, are determined as outlined below.
In general, different types of transform coefficient blocks are differentiated when considering a picture and/or video coding system. Thus, there are, for example, twelve types of transform coefficient blocks, having different statistics (the left column of table in Figure 21) in the present Final Draft International Standard of the H.264/AVC standard. For most picture sequences and coding conditions, some of the statistics are, however, very similar. In order to keep the number of contexts used small and thus to ensure a quick adaptation to the statistic of the picture sequence to be coded, the block types, in the H.264/AVC standard, can, for example, be classified into five categories (see right column of table in Figure 21). Similar classifications are possible for other picture and/or video coding systems. For each of the- in the case of the H.264/AVC standard - five categories, an individual quantity of contexts is used for the symbols CBP4, SIG, LAST and ABS.
2.1 Context modeling for the CBP4 symbol
For coding the one-bit symbol CBP4, four different contexts are used for each individual category of transform blocks (see Figure 21). The context number for block C to be coded is determined by ctx_number_cbp4 (C) = CBP4 (A) + 2 x CBP4 (B) wherein those neighboring blocks (left and top) of block C considered to be associated to the same block type are designated by A and B. Regarding the H.264/AVC standard, the following 6 block types are differentiated for this conditioning: Luma-DC, Luma-AC, Chroma-U-DC, Chroma-U-AC, Chroma-V-DC and Chroma-V-AC. If the concerning block X (A or B) of transform coefficients does not exist in a neighboring macro block (this is, for example, the case if the present block is coded in the INTRA 16 16 mode, but the neighboring block has been transferred in an INTER mode), CBP4 (X) is set to zero for the neighboring block X. If a neighboring block X (A or B) is outside the picture area or belongs to another slice, the corresponding value CBP4 (X) is replaced by a default value. A default value of one is used for INTRA-coded blocks and a default value of zero is used for INTER-coded blocks.
2.2 Context modeling for coding the significance mapping
For coding the significance mapping, max_koeff-l different contexts are each used per block category (see Figure 21) for coding the symbols SIG and LAST. max_koeff thus designates the number of transform coefficients for the corresponding block category (for H.264/AVC, see Figure 21). The context number is always indicated by the corresponding scan position of the coefficient considered. The context numbers of a coefficient koefffi], which has been scanned as an i-th coefficient, thus result in: ctx_number_sig(koeff[i]) = ctx_number_last(koeff[i]) = i.
2*max_koeff-2 contexts are used for each category of block types for coding the significance mapping.
2.3 Context modeling for coding coefficient magnitudes
The binarization illustrated in Figure 20 is used for coding the magnitudes of the significant transform coefficients. Two different context quantities are used per block category, namely one for coding the first binary decision bin = 1 (marked orange in Figure 20) and another one for coding the binary decisions bin = 2..14 (marked green in Figure 20) of the binarization. The context numbers are thus associated as follows: ctx__number_abs_ 1 bin
= (koeff with ABS>1 coded ? 4:
max(3, number of coded coefficients with ABS=1)), ctx_number_abs_rbins
= max(4, number of coded coefficients with ABS>1)).
The magnitudes of the transform coefficients are transferred in a reverse scan order. The context for the first binary decision is determined by the number of coefficients already transferred (in reverse scan order) having a magnitude of ABS = 1. If more than three coefficients with a magnitude ABS = 1 have already been transferred, context number 3 will always be chosen. As soon as a coefficient having a magnitude ABS > 1 has been transferred, context 4 will be used for all the remaining significant coefficients within the block.
All binary decisions with bin = 2..14 are coded using one and the same context. Thus, the context number is determined by the number of coefficients already coded (in a reverse scan order) having a magnitude of ABS>1, wherein the maximum context number is limited to 4. For illustration, two examples for the context selection, when coding the magnitudes ABS of the significant transform coefficients, are illustrated in Figure 23. For coding the binary decision bin>14 for the coefficient magnitudes and for the signs SIGN, an individual non- adaptive context with the symbol probabilities P0 = Pi = 0.5 is used.
Thus, the application embodiment described with respect to Fig. 18 to 23 related to coding transform coefficients in picture and/or video coders and decoders wherein for blocks of (video) pictures containing significant transform coefficients, the coding of transform coefficients takes place in such a way that, for each block, in a scan process, the positions of significant transform coefficients in the block and subsequently, in a reverse scan order - starting with the last significant transform coefficients within the block - the values (levels) of the significant transform coefficients are determined and coded. Each significant transform coefficient of the block other than the last transform coefficient of the block may be characterized by a one-bit symbol. Further, for each significant transform coefficient, the sign may be indicated by a one-bit symbol (SIGN) and the magnitude is indicated by a binary- coded symbol (ABS). The magnitude may be indicated by a symbol (ABS) in unary binarization or by a symbol (ABS) having a prefix part and a suffix part, wherein the prefix part consists of ones and the suffix part is coded in a 0th order exp-golomb code. Blocks containing significant transform coefficients may be characterized by a one-bit symbol CBP4 in connection with further syntax elements, such as, for example, CBP or macro block mode. Further, by transferring a one-bit symbol (SIG) for each coefficient of a block and a one-bit symbol (LAST) for each significant coefficient of a block, a significance mapping may be coded, wherein the transfer takes place in a scan order, (SIG) serves for identifying significant coefficients and (LAST) indicates whether there are further significant transform coefficients in the block. Modeling, for the one-bit symbol CBP4, for coding the significance mapping and/or for coding the coefficient magnitudes may take place in a context-dependent way. No significance information (SIG, LAST) might be transferred for the last scan position of a block. Block types of transform coefficients having comparable statistics may be summarized to block categories.
Differently stated, the coding of transform coefficient levels may be split into three steps. In the first step, a binary syntax element coded_block_flag is transmitted for each transform block, which signals whether the transform block contains significant transform coefficient levels (i.e., transform coefficients that are non-zero). If this syntax element indicates that significant transform coefficient levels are present, a binary- alued significance map is coded, which specifies which of the transform coefficient levels have non-zero values. And then, in a reverse scan order, the values of the non-zero transform coefficient levels are coded. The significance map is coded as follows. For each coefficient in the scan order, a binary syntax element significant_coeff_flag is coded, which specifies whether the corresponding transform coefficient level is not equal to zero. If the significant_coeff_flag bin is equal to one, i.e., if a non-zero transform coefficient level exists at this scanning position, a further binary syntax element last_significant_coeff_fiag is coded. This bin indicates if the current significant transform coefficient level is the last significant transform coefficient level inside the block or if further significant transform coefficient levels follow in scanning order. If last_significant_coeff_flag indicates that no further significant transform coefficients follow, no further syntax elements are coded for specifying the significance map for the block. In the next step, the values of the significant transform coefficient levels are coded, whose locations inside the block are already determined by the significance map. The values of significant transform coefficient levels are coded in reverse scanning order by using the following three syntax elements. The binary syntax element coeff_abs_greater_one indicates if the absolute value of the significant transform coefficient level is greater than one. If the binary syntax element coeff_abs_greater_one indicates that the absolute value is greater than one, a further syntax element coeff_abs_level_minus_one is sent, which specifies the absolute value of the transform coefficient level minus one. Finally, the binary syntax element coeff_sign_flag, which specifies the sign of the transform coefficient value, is coded for each significant transform coefficient level. It should be noted again that the syntax elements that are related to the significance map are coded in scanning order, whereas the syntax elements that are related to the actual values of the transform coefficients levels are coded in reverse scanning order allowing the usage of more suitable context models.
In the CABAC entropy coding in H.264, all syntax elements for the transform coefficient levels are coded using a binary probability modelling. The non-binary syntax element coeff_abs_level_minus_one is first binarized, i.e., it is mapped onto a sequence of binary decisions (bins), and these bins are sequentially coded. The binary syntax elements significant_coeff_flag, last_significant_coeff_flag, coeff_abs_greater_one, and coeff_sign_flag are directly coded. Each coded bin (including the binary syntax elements) is associated with a context. A context represents a probability model for a class of coded bins. A measure related to the probability for one of the two possible bin values is estimated for each context based on the values of the bins that have been already coded with the corresponding context. For several bins related to the transform coding, the context that is used for coding is selected based on already transmitted syntax elements or based on the position inside a block. The significance map specifies information about the significance (transform coefficient level is different from zero) for the scan positions. In the CABAC entropy coding of H.264, for a block size of 4x4, a separate context is used for each scan position for coding the binary syntax elements significant_coeff_flag and the last_significant_coeff_flag, where different contexts are used for the significant_coeff_flag and the last_significant_coeff_flag of a scan position. For 8x8 blocks, the same context model is used for four successive scan positions, resulting in 16 context models for the significant_coeff_flag and additional 16 context models for the last_significant_coeff_flag.
After coding the significance map, the block is processed in reverse scan order. If a scan position is significant, i.e., the coefficient is different from zero, the binary syntax element coeff_abs_greater_one is transmitted. Initially, the second context model of the corresponding context model set is selected for the coeff_abs_greater_one syntax element. If the coded value of any coeff_abs_greater_one syntax element inside the block is equal to one (i.e., the absolute coefficient is greater than 2), the context modelling switches back to the first context model of the set and uses this context model up to the end of the block. Otherwise (all coded values of coeff_abs_greater_one inside the block are zero and the corresponding absolute coefficient levels are equal to one), the context model is chosen depending on the number of the coeff_abs_greater_one syntax elements equal to zero that have already been coded/decoded in the reverse scan of the considered block. The context model selection for the syntax element coeff_abs_greater_one can be summarized by the following equation, where the current context model index Ct+i is selected based on the previous context model index Ct and the value of the previously coded syntax element coeff_abs_greater_one, which is represented by bint in the equation. For the first syntax element coeff_abs_greater_one inside a block, the context model index is set equal to Q = 1.
Figure imgf000073_0001
The second syntax element for coding the absolute transform coefficient levels, coeff_abs_level_minus_one is only coded, when the coeff_abs_greater__one syntax element for the same scan position is equal to one. The non-binary syntax element coeff_abs_level_minus_one is binarized into a sequence of bins and for the first bin of this binarization; a context model index is selected as described in the following. The remaining bins of the binarization are coded with fixed contexts. The context for the first bin of the binarization is selected as follows. For the first coeff_abs_level_minus_one syntax element, the first context model of the set of context models for the first bin of the coeff_abs_level_minus_one syntax element is selected, the corresponding context model index is set equal to Ct = 0. For each further first bin of the coeff_abs_level_minus_one syntax element, the context modelling switches to the next context model in the set, where the number of context models in set is limited to 5. The context model selection can be expressed by the following formula, where the current context model index Ct+i is selected based on the previous context model index Ct. As mentioned above, for the first syntax element coeff_abs_level_minus_one inside a block, the context model index is set equal to Ct = 0. Note, that different sets of context models are used for the syntax elements coeff _abs_greater_one and coeff_abs_level_minus_one.
Ci+1(C,) = min(Q +l,4)
When combining the just-described application embodiment with the entropy coding embodiments described earlier, the probability assigned to the individual contexts would, for example, be used as, or at least one of the parameters used for selecting the entropy encoder( entropy decoder among the plurality of partial bitstream paths, such as, described above, after having quantized the probability to a smaller set of probability partial bitstreams. In other words, the reconstruct of the sequence of symbols sequentially defining the significance mapping and the subsequent coding of the values of transform coefficients being unequal to zero in reverse scan order may be performed using a decoder according to any of the previously described embodiments.
EXAMPLE 2
According to example 2 for an application of the above-outlined entropy coding embodiments, an improved entropy encoding design with higher compression efficiency may be achieved.
An apparatus for entropy-encoding a symbol sequence of information symbols to obtain entropy-encoded information symbols, the symbol sequence having a start symbol, comprises: an entropy encoder for entropy encoding a symbol of the symbol sequence based on probability information for the symbol, the symbol being part of a symbol set, to produce the entropy-encoded information symbols; a context modeler for deriving the probability information for the symbol based on a context of the symbol, the context including one or more context symbols processed by the entropy encoder prior to processing the symbol, the context modeler including an initializer for initializing the context modeler by determining and providing initialization probability information to the entropy encoder, the initialization probability information to be used by the entropy encoder for processing the start symbol, wherein the initializer is operative to determine the initialization probability information based on an estimation of symbol statistics relating to the start symbol such that an initialization probability distribution is different from an equi-probable distribution for all symbols of the symbol set. The entropy encoder may embodied in accordance with any of the above-mentioned embodiments.
A decoder for entropy decoding entropy-encoded information symbols, the entropy-encoded information symbols being produced by entropyally encoding a symbol of the symbol sequence based on probability information for the symbol, the symbol being part of a symbol set, wherein the probability information for the symbol is derived based on a context of the symbol, the context including one or more context symbols processed earlier, and wherein, for entropyally encoding the start symbol, an initialization probability information was used, the initialization probability information being based on an estimation of a symbol statistics relating to a start symbol and being determined such that an initialization probability distribution is different from an equi-probable distribution for all symbols of the symbol set, comprises: an entropy decoder for entropy decoding the entropy-encoded information symbols to obtain the symbol sequence of information symbols having the start symbol; and a context modeler for obtaining the probability information used when entropy encoding the sequence of information symbols, context modeler including an initializer for obtaining the initialization probability information used when entropy encoding the start symbol. The entropy decoder may embodied in accordance with any of the above-mentioned embodiments.
An improved compression efficiency can be obtained by initializing the context modeler so that the initialization probability information is different from an equi-probable distribution for all symbols of the symbol set in question. Particularly, the initializer is operative to determine the initialization probability information to be used at least for the start symbol of the symbol sequence based on an estimation of a symbol statistics relating to the start symbol so that the initialization probability information is different from an equi-probable distribution for all symbols of the symbol set.
In particular, this "intelligent" initialization is advantageous for coding of video material with widely varying content from different video sources. It has been found out that, in a preferred embodiment of the present invention, the initializing probability information can be derived based on the quantization parameter used for quantizing the to be encoded information. In video coding applications, and, in particular, in video encoding applications in connection with the H.264/AVC video compression standard, the symbol sequence of information symbols to be entropy-encoded is derived from transformed residual values, which are obtained by a time/spatial inter/intra frame prediction mode for example. Additionally, the intelligent initialization can advantageously be based on prior knowledge on the video source material. In particular, it has been found out that the compression efficiency can be enhanced by defining several sets of initialization values. For so-called I- or Si-slices, for example, a single set of initialization values is sufficient. For temporally predictive-coded slices such as the so-called P-, SP- or B-slices, more than one and, preferably, three different sets of initialization values may be defined. For these slices, the selected set of initialization values is to be specified by a certain syntax element in the encoder output bit stream so that a decoder processing an encoded output bit stream is able to correctly initialize itself. Preferably, this syntax element is transmitted as part of the slice header. Now, the encoder has the freedom to select the set of initialization values that is the most suitable one for the symbol statistics for each slice.
Preferably, the initialization values are further diversified by detecting, to which data type or sub-data type, the symbol sequence to be processed by the entropy encoder belongs to. This certain type is indicated by a context index variable. This context index variable is preferably designed so that it can signal one out of up to 399 different context models from which a corresponding number of initialization probability information can be derived so that an optimally adapted initialization for a large number of different symbol sequences, each of which having a start symbol, can be determined.
It has to be noted, that the symbols to be entropy coded have attributed thereto different context models. Preferably, several symbols or bits are to be encoded using one context model out of the plurality of context models. These several symbols or bits form the symbol sequence, wherein the first to be encoded symbol of the symbol sequence, the symbol sequence having attributed thereto a context model, is the start symbol. A context model can include one or even more than one probability models. In case of a context model having more than one probability models, for each change of probability models, a new initialization will take place.
In the following, reference is made to Fig. 24 for illustrating the structure and functionality of an entropy-encoding apparatus or an entropy-encoding method. In particular, Fig. 24 shows an apparatus for entropy-encoding symbol sequences of information symbols to obtain entropy-encoded information symbols, wherein the symbol sequence has a start symbol. The symbol sequence is input into a symbol sequence input of an entropy encoder 1112. The entropy-encoded information symbols are output at an entropy-encoder output 1114, which is also connected to the entropy encoder 1112. The entropy encoder is operative for entropy encoding a symbol of the symbol sequence based on probability information for the symbol. The probability information is input into the entropy encoder 1112 via a probability information input 16. Since the symbol is part of a specified symbol set, which, in the binary case, only includes two different symbols, i.e., a binary "0" and the binary "1", it is sufficient to provide probability information only for one symbol of the symbol set. In accordance with the present invention, it is preferred to provide the probability information of the least probable symbol (LPS), since this number is naturally smaller than the probability for the most probable symbol. Additionally, the probability information provided to the entropy encoder via the probability information input 1116 includes the value of the most probable symbol, from which the value of the least probable symbol and vice versa can be easily derived.
The encoder further includes the context modeler 1118 for deriving the probability information for the symbol based on a context of the symbol, the context including one or more context symbols input into the context modeler by a context input 1120, which have been processed by the entropy encoder 1112 prior to processing the actual symbol.
The context modeler 1118 includes an initializer 1120 for initializing the context modeler by determining and providing initialization probability information to the entropy encoder 1112, when the initialization probability information is to be used by the entropy encoder 1112 for processing the start symbol, i.e., the symbol of the symbol sequence, for which no context symbols exist, i.e. symbols which have been processed by the entropy encoder 1112 prior to the start symbol. This means that, since the entropy encoding scheme is a serial scheme, the start symbol is the first symbol of the symbol sequence.
The initializer is operative to determine the initialization probability information based on an estimation of symbol statistics relating to the start symbol such that the initialization probability is different from an equi-probable distribution for all symbols of the symbol set, to which the start symbol belongs to.
Since the context modeler, and, therefore, the entropy encoder is initialized in an intelligent way based on an estimation of the symbol statistics that are expected for the symbol sequence to be encoded by the entropy encoder, the entropy-encoding apparatus shows an improved start-up performance. Stated in other words, the entropy-encoding apparatus behaves, for the start symbol, as if the start symbol is not the first symbol of a sequence but is an intermediate symbol of a sequence. Therefore, the initialization results in a pre-adaptation of the entropy- encoding apparatus already for the start symbol. Contrary to the straightforward approach, in which the initialization probability information is set to an equi-probable distribution, the entropy encoder is immediately context-pre-adapted. In the straightforward case, the context- adaptation would need several steps corresponding to the number of prior encoded symbols forming the context symbols. It has been found out that the inventive initialization of the context modeler and the entropy encoder results, without any further actions, in bit-rate savings of up to 3 % simply by providing an intelligent initialization probability information compared to a straightforward equi-probable distribution.
Fig. 25 shows a preferred embodiment of the initializer 1020 from Fig. 24. The initializer 1120 includes a retriever for retrieving a quantization parameter for the symbol sequence from a slice header. Additionally, the slice header is also input into a detector for detecting a data type of the slice data. Additionally, a detector 1204 for detecting a sub-data type in the slice data is provided. The detector 1204 is further operative to access different initialization tables based on the detected sub-data type. As it will be described later on, the access to different initialization tables is controlled by means of a context index on the one hand and, preferably, by means of an initialization index on the other hand. Depending on the slice data, there only exists one initialization table with certain context indices for I- and SI- slices. For P- and B-slices, there exist at least two and preferably three initialization tables which are, once again, addressed by the context index, wherein the table of the at least two tables which is actually used is addressed by the initialization index.
Connected to the retriever 1200 on the one hand and the detector 1204 on the other hand, is a calculator 1206, which is operative to calculate the information on the initialization probability which can be the actual initialization probability or, in accordance with the preferred embodiment, a probability state index, i.e., a reference to an initialization probability information table including probability information for the least probable symbol. It has been found out by the inventors that a linear dependence of the initialization probability from the quantization parameter used for quantizing the slice data results in an improved compression efficiency, i.e., in an near to optimum initialization. Therefore, the calculator 1206 in Fig. 25 is operative to apply a linear calculation, wherein the parameter m indicates a gradient of the linear dependence, while the other initialization parameter, i.e., n indicates a y- axis offset. The basic equation processed by the calculator 1206 in Fig. 25 is indicated in block 206 as m x QP + n, wherein QP is the quantization parameter while m and n are the initialization parameters taken out from different initialization tables addressed by a certain context index.
In a preferred embodiment, the calculator 1206 in Fig. 25 performs the sequence of steps as indicated in Fig. 26. In a first step, the equation m x sliceQP + n is calculated. It is to be noted here that in the preferred embodiment of Fig. 26, the value F of m x sliceQP is shifted to the right by four digits. It is to be noted here that, naturally, all values are binary strings. The shift by four digits relates to a representation of the product m x sliceQP in a two-complementary presentation. Then, the result after the shift is added to the other initialization parameter n. Then, the minimum of this result on one hand and the integer value 1126 on the other hand is selected. In case this minimum is larger than 1 , this minimum is selected and assigned to the auxiliary variable preCtxState. Then, a determination is made, if this auxiliary variable is lower than or equal to 63. In case this determination results in a yes answer, the probability state index variable pStateldx is calculated as the difference between 63 and the auxiliary variable. Additionally, it is determined that the value of the most probable symbol (valMPS) is equal to zero. In the other case, i.e., in which the auxiliary variable is greater than 63, the probability state index pStateldx is calculated as the difference between the auxiliary variable on the one hand and the value of 64. Additionally, the value of the most probable symbol is set to one.
It is to be noted that the probability state index is a reference to an initialization probability information table including probability information for the least probable symbol. The value of the probability state index can address 64 different probability information values in the probability information table. The reason for firstly calculating a probability state index and then addressing a table using the probability state index is in an improved design of the actual entropy encoder core. Naturally, by an appropriate mapping, the sequence of the steps shown in Fig. 26 could also directly calculate the initialization probability information.
The quantization parameter SliceQP is used for determining the quantization of transform coefficients. The parameter may take 52 values. Theses values may be arranged so that an increase of 1 in quantization parameter means an increase of quantization step size by approximately 12 % (an increase of 6 means an increase of quantization step size by exactly a factor of 2). It can be noticed that a change of step size by approximately 12 % also means roughly a reduction of bit rate by approximately 12 %.
In the following, reference is made to Fig. 27 showing an inventive entropy decoder. As it is known in the art of entropy coding/decoding, the entropy decoder needs the same probability distribution used by the encoder for a respective encoding step to resolve the transmitted interval (which is naturally represented in binary form) into the decoded bits, i.e., the information symbol sequence having the start symbol. To this end, the inventive entropy decoder includes an entropy decoder 1160 and a context modeler 1162, which also includes an initializer 1164. It is to be seen here that the decoder shown in Fig. 27 is analogously constructed as the encoder in Fig. 24. Again, the context modeler 1162 provides probability information or, when the start symbol of the sequence is concerned, the initialization probability information. This information is produced in the same way as has been described with respect to Fig. 24, i.e., by determining a respective slice quantization parameter, by determining a certain context index and, by determining the corresponding initialization parameters m, n, for calculating the initialization probability information, when the entropy decoder is operative to process the start symbol. It is to be noted here that the entropy encoding/decoding chain is a kind of first-in-first-out pipeline, since the start symbol of the symbol sequence in the encoder is the first symbol to be encoded and is also the first symbol to be decoded. Therefore, the same initialization process for the entropy decoder shown in Fig. 27 can be used, which has been described in connection with the entropy encoder shown in Fig. 24.
Regarding the regular context adaptation mode, the context modeler also performs the same operational steps as have been described with respect to Fig. 24.
In both Fig. 24 and 27, the entropy en/decoder may be implemented as indicated in any of the previous Fig. 1 to 17.
EXAMPLE 3
In example 3, several extensions of the above described embodiments are described.
It is particularly advantageous when the probability estimation underlying the symbol to be encoded or decoded is associated to a probability state Pn using an index p_state.
For example, there are N different representative probability states. For same, transition rules may be given, wherein the transition rules indicate which new state is used based on the currently encoded or decoded symbol for the next symbol to be encoded or decoded. It is hereby of an advantage when a table Next_State_LPS is created which contains the index m of the new probability state Pm when a least probable symbol (LPS) occurs in addition to the index n of the currently given probability state Pn, and/or when a table Next_State_MPS is created which contains the index m of the new probability state Pm when a most probable symbol (MPS) occurs in addition to the index n of the currently given probability state Pn.
One special implementation of the encoding includes the following steps:
1. Determination of the LPS
2. Quantization
p = quantize [p_state]
3. En/decoding En/decode bit using p as parameter for partial bitstream selection
4. Adaptation of p_state:
if (bit = 1 - valMPS) then
p_state <- Next_State_LPS [p_state]
if (p_state = 0) then valMPS <- 1 - valMPS
else
p_state <- Next_State_MPS [p_state]
bit current bit to be decoded/encoded
p_state describes the current probability state of the context assigned
valMPS describes the value of the bit corresponding to the MPS.
It further turns out to be advantageous when the initialization of the probability models is performed depending on a quantization parameter SliceQP and preset model parameters m and n, wherein SliceQP describes the quantization parameter preset at the beginning of a slice and m and n describe the model parameters.
It is also advantageous when the initialization of the probability models includes the following steps:
1. preState = min(max(l, ((m * SliceQP) »4)+n), 2*(N-1))
2. if (preState <=N-1) then
p_state = N-l - preState
valMPS = 0
else
p_state = preState
valMPS = 1,
wherein valMPS describes the value of the bit corresponding to the MPS, SliceQP describes the quantization parameter preset at the beginning of a slice and m and n describe the model parameters.
The theoretical background is to be explained in more detail:
Table-aided probability estimation
As it was already mentioned above, the estimation of the occurrence probability of the symbols to be coded should be as good as possible. In order to enable an adaptation to non- stationary source statistics, this estimation needs to be updated in the course of the coding process. Generally, usually methods are used for this which operate using scaled frequency counters of the coded results. If CLps and CMps designates counters for the occurrence frequencies of LPS and MPS, then using these counters the estimation
Figure imgf000082_0001
may be performed and then the entropy en/decoding may be carried out using this probability. For practical purposes the division required in equation (CI) is disadvantageous. It is often convenient and required, however, to perform a rescaling of the counter readings when a predetermined threshold value Cmax of the overall counter Ciotai = C PS + CLPS is exceeded. With a suitable selection of Cmax the reciprocal values of Cx0tai may be tabulated, so that the division required in equation (CI) may be replaced by a table access and by a multiplication and shift operation. In order to prevent also these arithmetic operations, however, in the present invention a completely table-aided method is used for the probability estimation.
For this purpose in a training phase representative probability states {Ρ I 0 < k < Nmax} are preselected, wherein the selection of the states is on the one hand dependent on the statistics of the data to be coded and on the other hand on the side conditions of the default maximum number Nmax of states. Additionally, transition rules are defined which indicate which new state is to be used for the next symbol to be coded based on the currently coded symbol. These transition rules are provided in the form of two tables: {Next_State_LPSk I 0 < k < Nmax} and {Next_State_MPSk I 0 < k < Nmax}, wherein the tables provide the index m of the new probability state Pm when an LPS or MPS occurs, respectively, for the index n of the currently given probability state. It is to be noted here, that for a probability estimation in the entropy encoder or decoder, respectively, as it is proposed herein, no explicit tabulation of the probability states is required. Rather, the states are only implicitly addressed using their respective indices, as it is described in the following section. In addition to the transition rules it needs to be specified at which probability states the value of the LPS and MPS needs to be exchanged. Generally, there will only be one such excellent state which may be identified using its index p_state.
Next, it is described, how the updating of the probability state p_state is performed depending on the value of the coded bit:
= 1 - valMPS) then
p_state <- Next_State_LPS [p_state]
if (p_state = 0) then valMPS <- 1 - valMPS p_state < - Next_State_MPS [p_state]
Transition tables Next_State_LPS and Next_State_MPS are used which were already mentioned above in the section "table-aided probability estimation".
Coding with uniform probability distribution
In applications in which e.g. signed values are to be coded whose probability distribution is arranged symmetrically around zero, for coding the sign information generally an equal distribution may be assumed. As this information is one the one hand to be embedded in the entropy coded bit stream, while it is on the other hand not sensible to use a relatively compact apparatus of the table-aided probability estimation and interval separation for the case of a probability of p « 0.5, it is for this special case proposed to optionally use a special encoder/decoder procedure which may be illustrated as follows.
In this special case the entropy coding of the bin is performed using a probability distribution according to which all symbol values are equally probable.
Addressing and initializing the probability models
Every probability model, as it is used, is indicated using two parameters: 1) The index p_state that characterizes the probability state of the LPS, and 2) the value valMPS of the MPS. Each of these two variables needs to be initialized at the beginning of the encoding or decoding, respectively, of a completed coding unit (in applications of video coding about one slice). The initialization values may thereby be derived from control information, like e.g. the quantization parameter (of a slice), as it is illustrated by:
1. preState - min(max(l, ((m * SliceQP) »4)+n), 2*(N-1))
2. if (preState <=N- 1 ) then
p_state = N-l - preState
valMPS = 0
else
p_state = preState - N
valMPS = 1,
Forward-controlled initialization process A further possibility of adaptation of the starting distributions of the models is provided by the following method. In order to guarantee a better adaptation of the initializations of the models, in the encoder a selection of predetermined starting values of the models may be provided. These models may be combined into groups of starting distributions and may be addressed using indices, so that in the encoder the adaptive selection of a group of starting values is performed and is transmitted to the decoder in the form of an index as page information. This method is referred to as a forward-controlled initialization process.
EXAMPLE 4
The next example is related to video frame coding and, in particular, to an coding scheme using context assignment based on neighboring syntax elements. The following application example provides a video coding scheme using the above embodiments, which enables a higher compression effectiveness.
This advantage is achieved by encoding a video signal representing at least one video frame, with at least one video frame being composed of picture samples, the picture samples belonging either to a first or a second field being captured at different time instants, the video frame being spatially divided up into macroblock pair regions, each macroblock pair region being associated with a top and bottom macroblock, the encoding comprising the steps of deciding, for each macroblock pair region, as to whether same is of a first or a second distribution type; assigning, for each macroblock pair region, each of the pixel samples in the respective macroblock pair region to a respective one of the top and bottom macroblock of the respective macroblock pair region, in accordance with the distribution type of the respective macroblock pair region, and pre-coding the video signal into a pre-coded video signal, the pre-coding comprising the sub-step of pre-coding a current macroblock of the top and bottom macroblock associated with a current macroblock pair region of the macroblock pair regions to obtain a current syntax element. Thereafter, it is determined, for the current syntax element, a neighboring macroblock at least based upon as to whether the current macroblock pair region is of a first or second distribution type. One of at least two context models is assigned to the current syntax element based on a pre-determined attribute of the neighboring macroblock, wherein each context model is associated with a different probability estimation. Finally, an encoding of a sequence of symbols takes place in accordance with any of the above embodiments, with at least one symbol relating to a representation of the syntax element. The assigner is configured to determine the measure for the estimate of the probability distribution for the respective symbol based on the probability distribution estimate associated with the context assigned to the respective symbol. The selector selects the respective entropy coder and the respective partial bit stream accordingly. Decoding a syntax element from a coded bit stream is also described. The coded bit stream is an entropy encoded version of a pre-coded video signal, the pre-coded video signal being a pre-coded version of a video signal, the video signal representing at least one video frame being composed of picture samples, the picture samples belonging either to a first or a second field being captured at a different time instants, the video frame being spatially divided up into macroblock pair regions, each macroblock pair region being associated with a top and a bottom macroblock, each macroblock pair region being either of a first or a second distribution type, wherein, for each macroblock pair region, each of the pixel samples in the respective macroblock pair region is assigned to a respective one of the top and bottom macroblock of the respective macroblock pair region in accordance with the distribution type of the respective macroblock pair region, wherein the syntax element relates to a current macroblock of the top and bottom macroblock of a current macroblock pair region of the macroblock pair regions. The method comprises determining, for the current syntax element, a neighboring macroblock at least based upon as to whether the current macroblock pair region is of a first or a second distribution type; assigning one of at least two context models to the current syntax element based on a predetermined attribute of the neighboring macroblock, wherein each context model is associated with a different probability estimation; and reconstructing a sequence of symbols among which at least one symbol relates to a representation of the syntax element in accordance with any of the above embodiment is performed. The assigner is configured to determine the measure for the estimate of the probability distribution for the respective symbol based on the probability distribution estimate associated with the context assigned to the respective symbol. The selector selects the respective entropy decoder and the respective partial bit stream, respectively, accordingly.
The below-outlined application example exploits the finding that when, for whatever reason, such as the better effectiveness when coding video frames having non-moving regions and moving regions, macroblock pair regions of a first and a second distribution type, i.e., field and frame coded macroblock pairs, are used concurrently in a video frame, i.e. MBAFF coding is used, the neighborhood between contiguous blocks of pixel samples has to be defined in a way different from considering each macroblock pair as frame macroblock pair for the purpose of context modeling and that the distance of areas covered by a neighboring and a current block could be very large when considering each macroblock pair as a frame macroblock pair. This in turn, could degrade the coding efficiency, since choosing neighboring blocks that are not arranged nearby the current block affects the adaption of the conditional probability models.
Further, the below-outlined application example exploits, that, for sake of a better definition of neighborhood between blocks of picture samples, i.e. the neighboring block which the syntax element to be coded or decoded relates to and the current block based on the attribute of which the assignment of a context model is conducted, and when the neighboring block lies beyond the borders or circumference of the current macroblock containing the current block, it is important to make the determination of the macroblock containing the neighboring block dependent upon as to whether the current macroblock pair region containing the current block is of a first or a second distribution type, i.e., frame or field coded.
The blocks may be a macroblock or some sub-part thereof. In both cases, the determination of a neighboring block comprises at least the determination of a neighboring macroblock a s long as the neighboring block lies beyond the borders of the current macroblock.
Fig. 28 shows a general view of a video encoder environment to which the present invention could be applied. A picture of video frame 2010 is fed to a video precoder 2012. The video precoder treats the picture 2010 in units of so-called macroblocks 10a. Each macroblock contains several picture samples of picture 2010. On each macroblock a transformation into transformation coefficients is performed followed by a quantization into transform coefficient levels. Moreover, intra-frame prediction or motion compensation is used in order not to perform the afore mentioned steps directly on the pixel data but on the differences of same to predicted pixel values, thereby achieving small values which are more easily compressed.
Precoder 2012 outputs the result, i.e., the precoded video signal. All residual data elements in the precoded video signal, which are related to the coding of transform coefficients, such as the transform coefficient levels or a significance map indicating transform coefficient levels skipped, are called residual data syntax elements. Besides these residual data syntax elements, the precoded video signal output by precoder 2012 contains control information syntax elements containing control information as to how each macroblock has been coded and has to be decoded, respectively. In other words, the syntax elements are dividable into two categories. The first category, the control information syntax elements, contains the elements related to a macroblock type, sub-macroblock type, and information on prediction modes both of a spatial and of temporal types as well as slice-based and macroblock-based control information, for example. In the second category, all residual data elements such as a significance map indicating the locations of all significant coefficients inside a block of quantized transform coefficients, and the values of the significant coefficients, which are indicated in units of levels corresponding to the quantizations steps, are combined, i.e., the residual data syntax elements.
The macroblocks into which the picture 2010 is partitioned are grouped into several slices. In other words, the picture 2010 is subdivided into slices. An example for such a subdivision is shown in Fig. 43, in which each block or rectangle represents a macroblock. For each slice, a number of syntax elements are generated by precoder 2012, which form a coded version of the macro blocks of the respective slice.
The precoder 2012 transfers the syntax elements to a final coder stage 2014, which comprises an entropy coder according to any of the embodiments described above with respect to any of Fig. 1 to 17 and explained in more detail with respect to Fig. 29. The final coder stage 2014 generates an codeword stream for each slice. When generating the codeword stream for a slice, the final coding stage 2014 exploits the fact that each syntax element is a data value having a certain meaning in the video signal bit stream that is passed to the entropy coder 2014. The entropy coder 2014 outputs a final compressed coded video bit stream comprising codewords for the slices of picture 2010.
Fig. 29 shows the arrangement for coding the syntax elements into the final code bit stream, the arrangement generally indicated by reference number 2100. The coding arrangement 2100 is divided into three stages, 2100a, 2100b, and 2100c.
The first stage 2100a is the binarization stage and comprises a binarizer 2102. An input of the binarizer 2102 is connected to an input 2104 of stage 2100a via a switch 2106. At the same time, input 2104 forms the input of coding arrangement 2100. The output of binarizer 2102 is connected to an output 2108 of stage 2100a, which, at the same time, forms the input of stage 2100b. Switch 2106 is able to pass syntax elements arriving at input 2104 to either binarizer 2102 or binarization stage output 2108, thereby bypassing binarizer 2102.
The function of switch 2106 is to directly pass the actual syntax element at input 2104 to the binarization stage output 2108 if the syntax element is already in a wanted binarized form. Examples for syntax elements that are not in the correct binarization form, called non-binary valued syntax elements, are motion vector differences and transform coefficient levels. Examples for a syntax element that has not to be binarized since it is already a binary value comprise the MBAFF (MBAFF = Macroblock Adaptive Frame/Field) Coding mode flag or mb_f!eld_decoding_flag, the mb_skip_flag and coded_block_flag to be described later in more detail. Examples for a syntax element that has to be binarized since it is not a binary value comprise syntax elements mbjype, coded_block_pattern, ref_idx_10, ref_idx_ll, mvd_10, mvd_ll, and intro_chroma_pred_mode.
Different binarization schemes are used for the syntax elements to be binarized. For example, a fixed-length binarization process is constructed by using an L-bit unsigned integer bin string of the syntax element value, where L is equal to log2(cMax+l) rounded up to the nearest integer greater than or equal to the sum, with cMax being the maximum possible value of the syntax element. The indexing of the bins for the fl binarization is such that the bin index of zero relates to the least significant bit with increasing values of the bin index towards the most significant bit. Another binarization scheme is a truncated unary binarization scheme where syntax element values C smaller than the largest possible value cMax are mapped to a bit or bin string of length C+l with the bins having a bin index smaller than C being equal to 1 and the bin having the bin index of C being equal to 0, whereas for syntax elements equal to the largest possible value cMax, the corresponding bin string is a bit string of length cMax with all bits equal to one not followed by a zero. Another binarization scheme is a k-th order exponential Golomb binarization scheme, where a syntax element is mapped to a bin string consisting of a prefix bit string and, eventually, a suffix bit string.
The non-binary valued syntax elements are passed via switch 2106 to binarizer 2102. Binarizer 2102 maps the non-binary valued syntax elements to a codeword, or a so-called bin string, so that they are now in a binary form. The term "bin" means the binary decision that have to be made at a node of a coding tree defining the binarization mapping of a non-binary value to a bit string or codeword, when transitioning from the route node of the coding tree to the leaf of the coding tree corresponding to the non-binary value of the non-binary syntax element to be binarized. Thus, a bin string is a sequence of bins or binary decisions and corresponds to a codeword having the same number of bits, each bit being the result of a binary decision.
The bin strings output by binarizer 2102 may not be passed directly to binarization stage output 108 but controllably passed to output 2108 by a bin loop over means 2110 arranged between the output of binarizer 2102 and output 2108 in order to merge the bin strings output by binarizer 2102 and the already binary valued syntax elements bypassing binarizer 2102 to a single bit stream at binarization stage output 2108.
Thus, the binarization stage 2108 is for transferring the syntax elements into a suitable binarized representation. The binarization procedure in binarizer 2102 preferably yields a binarized representation which is adapted to the probability distribution of the syntax elements so as to enable very efficient binary entropy coding.
Stage 2100b is a context modeling stage and comprises a context modeler 2112 as well as a switch 2113. The context modeler 2112 comprises an input, an output, and an optional feedback input. The input of context modeler 2112 is connected to the binarization stage output 2108 via switch 2113. The output of context modeler 2112 is connected to a regular coding input terminal 2114 of stage 2100c. The function of switch 2113 is to pass the bits or bins of the bin sequence at binarization stage output 2108 to either the context modeler 2112 or to a bypass coding input terminal 2116 of stage 2100c, thereby bypassing context modeler 2112. The aim of switch 2113 is to ease the subsequent binary entropy coding performed in stage 100c. To be more precise, some of the bins in the bin string output by binarizer 2102 show heuristically nearly an equi-probable distribution. This means, the corresponding bits are, with a probability of nearly 50%, 1 and, with a probability of nearly 50%, 0, or, in other words, the bits corresponding to this bin in a bin string have a 50/50 chance to be 1 or 0. These bins are fed to the bypass-coding input terminal 2116 and are binary entropy coded by use of an equi-probable probability estimation, which is constant and, therefore, needs no adaption or updating overhead. For all other bins, it has been heuristically determined that the probability distribution of these bins depends on other bins as output by stage 2100a so that it is worthwhile to adapt or update the probability estimation used for binary entropy coding of the respective bin as it will be described in more detail below exemplarily with respect to exemplary syntax elements. The latter bins are thus fed by switch 2113 to the input terminal of context modeler 2112.
Context modeler 2112 manages a set of context models. For each context model, the context modeler 2112 has stored an actual bit or bin value probability distribution estimation. For each bin that arrives at the input of context modeler 2112, the context modeler 2112 selects one of the sets of context models. In other words, the context modeler 2112 assigns the bin to one of the set of context models. The assignment of bins to a context model is such that the actual probability distribution of bins belonging to the same context model show the same or likewise behavior so that the actual bit or bin value probability distribution estimation stored in the context modeler 2112 for a certain context model is a good approximation of the actual probability distribution for all bins that are assigned to this context model. The assignment process in accordance with the present invention exploits the spatial relationship between syntax element of neighboring blocks. This assignment process will be described in more detail below.
When having assigned the context model to an incoming bin the context modeler 2112 passes the bin further to entropy coding stage 2100c together with the probability distribution estimation of the context model, which the bin is assigned to. By this measure, the context modeler 2112 drives the entropy coding stage 2100c to generate a sequence of bits as a coded representation of the bins input in context modeler 2112 by switch 2113 according to the switched bit value probability distribution estimations as indicated by the context modeler 2112.
Moreover, the context modeler 2112 continuously updates the probability distribution estimations for each context model in order to adapt the probability distribution estimation for each context model to the property or attributes of the picture or video frame from which the syntax elements and bins have been derived. The estimation adaptation or estimation update is based on past or prior bits or bin values which the context modeler 2112 receives at the feedback input over a feedback line 2117 from stage 2100c or may temporarily store. Thus, in other words, the context modeler 2112 updates the probability estimations in response to the bin values passed to entropy coding stage 2100c. To be more precise, the context modeler 2112 uses a bin value assigned to a certain context model merely for adaptation or update of the probability estimation that is associated with the context model of this bin value.
Some of the syntax elements, when the same bin or same syntax element occurs several times in the bins passed from stage 2100a may be assigned to different of the context models each time they occur, depending on previously incoming or previously entropy coded bins, and/or depending on other circumstances, such as previously coded syntax elements of neighboring blocks, as is described in more detail below with respect to exemplary syntax elements.
It is clear from the above, that the probability estimation used for binary entropy coding determines the code and its efficiency in the first place, and that it is of paramount importance to have an adequate model that exploits the statistical dependencies of the syntax elements and bins to a large degree so that the probability estimation is always approximating very effectively the actual probability distribution during encoding.
The third stage 2100c of coding arrangement 100 is the entropy coding stage. It comprises a regular coding engine 2118, a bypass-coding engine 2120, and a switch 2122. The regular coding engine 2118 comprises an input and an output terminal. The input terminal of regular coding engine 2118 is connected to the regular coding input terminal 2114. The regular coding engine 2118 binary entropy codes the bin values passed from context modeler 2112 by use of the context model also passed from context modeler 2112 and outputs coded bits. Further, the regular coding engine 2118 passes bin values for context model updates to the feedback input of context modeler 2112 over feedback line 2117. The regular coding engine corresponds to any of the above-mentioned embodiments of Fig. 1 to 17 with however, the assigner of the above embodiments partially integrate within the context modeler as far as the setting of the context for the respective bin and the corresponding probability index is concerned. As mentioned above, assigner, may quantize the probability index before forwarding the thus obtained parameter to the selector 18 which, in turn, selects the respective entropy encoder.
The bypass-coding engine 2112 has also an input and an output terminal, the input terminal being connected to the bypass coding input terminal 2116. The bypass-coding engine 2120 is for binary entropy coding the bin values passed directly from binarization stage output 2108 via switch 2113 by use of a static predetermined probability distribution estimation and also outputs coded bits. The bypass-coding engine may use a 1:1 mapping in units of any number of symbols.
The coded bits output from regular coding engine 2118 and bypass coding engine 2120 are merged to a single bit stream at an output 2124 of coding arrangement 2100 by switch 2122, the bit stream representing a binary entropy coded bit stream of the syntax elements as input in input terminal 2104. Thus, regular coding engine 2118 and bypass coding 2120 cooperate in order to bit wise perform entropy coding based on either an adaptive or a static probability distribution model.
After having described with respect to Figs. 28 and 29 rather generally the operation of coding arrangement 2100, in the following its functioning is described in more detail with respect to the handling of exemplary syntax elements for which an context assignment process based on syntax elements of neighboring blocks is used, in accordance with embodiments. In order to do so, firstly, with regard to Figs. 30 to 31b, the meaning of MBAFF coding is described, in order to enable a better understanding of the definition of neighborhood between a current block and a neighboring block used during assignment of a context model to a syntax element concerning the current block in case of MBAFF.
Fig. 30 shows a picture or decoded video frame 2010. The video frame 2010 is spatially partitioned into macroblock pairs 10b. The macroblock pairs are arranged in an array of rows 2200 and columns 2202. Each macroblock pair consists of two macroblocks 2010a.
In order to be able to address each macroblock 10a, a sequence is defined with respect to macroblocks 2010a. In order to do so, in each macroblock pair, one macroblock is designated the top macroblock whereas the other macroblock in the macroblock pair is designated the bottom macroblock, the meaning of top and bottom macroblock depending on the mode by which a macroblock pair is coded by precoder 2012 (Fig. 28) as will be described with respect to Fig. 31a and 31b. Thus, each macroblock pair row 2200 consists of two macroblock rows, i.e., an top macroblock row 2200a consisting of the top macroblocks in the macroblock pairs of the macroblock pair line 2200 and a bottom macroblock row 200b comprising the bottom macroblocks of the macroblock pairs.
In accordance with the present example, the top macroblock of the top left macroblock pair resides at address zero. The next address, i.e. address 1, is assigned to the bottom macroblock of the top left macroblock pair. The addresses of the top macroblocks of the macroblock pairs in the same, i.e., top macroblock row 2200a, are 2, 4, ... , 2i-2, with the addresses rising from left to right, and with i expressing the picture width in units of macroblocks or macroblock pairs. The addresses 1, 3, 2i-l are assigned to the bottom macroblocks of the macroblock pairs in the top macroblock pair row 2200, the addresses rising from left to right. The next 2i- addresses from 2i to 4i-l are assigned to the macroblocks of the macroblock pairs in the next macroblock pair row from the top and so on, as illustrated in Fig. 30 by the numbers written into the boxes representing the macroblocks 2010a and by the arched rows.
It is emphasized that Fig. 30 does show the spatial subdivision of picture 2010 in units of macroblock pairs rather than in macroblocks. Each macroblock pair 2010b represents a spatial rectangular region of the pictures. All picture samples or pixels (not shown) of picture 2010 lying in the spatial rectangular region of a specific macroblock pair 2010b belong to this macroblock pair. If a specific pixel or picture sample belongs to the top or the bottom macroblock of a macroblock pair depends on the mode by which precoder 2012 has coded the macroblocks in that macroblock pair as it is described in more detail below.
Fig. 31a shows on the left hand side the arrangement of pixels or picture samples belonging to a macroblock pair 2010b. As can be seen, the pixels are arranged in an array of rows and columns. Each pixel shown is indicated by a number in order to ease the following description of Fig. 31a. As can be seen in Fig. 31a, some of the pixels are marked by an "x" while the others are marked "Y". All pixels marked with "x" belong to a first field of the picture while the other pixels marked with "Y" belong to a second field of the picture. Pixels belonging to the same field are arranged in alternate rows of the picture. The picture or video frame can be considered to contain two interleaved fields, a top and a bottom field. The top field comprises the pixels marked with "Y" and contains even-numbered rows 2n+2, 2n+4, 2n+6, ... with 2n being the number of rows of one picture or video frame and n being an integer greater than or equal to 0. The bottom field contains the odd-numbered rows starting with the second line of the frame.
It is assumed that the video frame to which macroblock pair 2010b belongs, is an interlaced frame where the two fields were captured at different time instants, for example the top field before the bottom field. It is now that the pixels or picture samples of a macroblock pair are differently assigned to the top or bottom macroblock of the macroblock pair, depending on the mode by which the respective macroblock pair is precoded by precoder 2012 (Fig. 28). The reason for this being the following.
As described above with respect to Fig. 28, the picture samples of a macroblock, which may be luminance or luma and chrominance or chroma samples, may be either spatially or temporarily predicted by precoder 2012, and the resulting prediction residual is encoded using transform coding in order to yield the residual data syntax elements. It is now that in interlaced frames (and it is assumed that the present video frame is an interlaced frame), with regions of moving objects or camera motion, two adjacent rows of pixels tend to show a reduced degree of statistical dependency when compared to progressive video frames in which both fields are captured at the same time instant. Thus, in cases of such moving objects or camera motion, the pre-coding performed by precoder 2012 which, as stated above, operates on macroblocks, may achieve merely a reduced compression efficiency when a macroblock pair is spatially sub-divided into a top macroblock representing the top half region of the macroblock pair and a bottom macroblock representing the bottom half region of the macroblock pair, since in this case, both macroblocks, the top and the bottom macroblock, comprise both top field and bottom field pixels. In this case, it may be more efficient for precoder 2012 to code each field separately, i.e., to assign top field pixels to the top macroblock and bottom field pixels to the bottom field macroblock.
In order to illustrate as to how the pixels of a macroblock pair are assigned to the top and bottom macroblock of the, Figs. 31a and 31b show on the right hand side the resulting top and bottom macroblock in accordance with the frame and field mode, respectively.
Fig. 31a represents the frame mode, i.e., where each macroblock pair is spatially subdivided in a top and a bottom half macroblock. Fig. 31a shows at 2250 the top macroblock and at 2252 the bottom macroblock as defined when they are coded in the frame mode, the frame mode being represented by double-headed arrow 2254. As can be seen, the top macroblock 250 comprises one half of the pixel samples of the macroblock pair 2010b while the other picture samples are assigned to the bottom macroblock 2252. To be more specific, the picture samples of the top half rows numbered 2n+l to 2n+6 belong to the top macroblock 2250, whereas the picture samples 91 to 96, 01 to 06, 11 to 16 of the bottom half comprising rows 2n+7 to 2n+12 of the macroblock pair 10b belong to the bottom macroblock 2252. Thus, when coded in frame mode, both macroblocks 2250 and 2252 comprise both, picture elements of the first field marked with "x" and captured at a first time instant and picture samples of the second field marked with "Y" and captured at a second, different time instant.
The assignment of pixels as they are output by a camera or the like, to top or bottom macroblocks is slightly different in field mode. When coded in field mode, as is indicated by double headed arrow 2256 in Fig. 31b, the top macroblock 2252 of the macroblock pair 2010b contains all picture samples of the top field, marked with "x", while the bottom macroblock 2254 comprises all picture samples of the bottom field, marked with "Y". Thus, when coded in accordance with field mode 2256, each macroblock in a macroblock pair does merely contain either picture samples of the top field or picture samples of the bottom field rather than a mix of picture samples of the top and bottom field.
Now, after having described the spatial sub-division of a picture into macroblock pairs and the assignment of picture samples in a macroblock pair to either the top or the bottom macroblock of the macroblock pair, the assignment depending on the mode by which the macroblock pair or the macroblocks of the macroblock pair are coded by precoder 2012, reference is again made to Fig. 28 in order to explain the function and meaning of the syntax element mb_field_decoding_flag contained in the precoded video signal output by precoder 2012, and, concurrently, in order to explain the advantages of MBAFF coded frames over just field or frame coded frames.
When the precoder 2012 receives a video signal representing an interlaced video frame, precoder 2012 is free to make the following decisions when coding the video frame 2010:
- It can combine the two fields together to code them as one single coded frame, so that each macroblock pair and each macroblock would be coded in frame mode.
- Alternatively, it could combine the two fields and code them as separate coded fields, so that each macroblock pair and each macroblock would be coded in field mode.
- As a last option, it could combine the two fields together and compress them as a single frame, but when coding the frame it splits the macroblock pairs into either pairs of two field macroblocks or pairs of two frame macroblocks before coding them.
The choice between the three options can be made adaptively for each frame in a sequence. The choice between the first two options is referred to as picture adaptive frame/field (PAFF) coding. When a frame is coded as two fields, each field is partitioned into macroblocks and is coded in a manner very similar to a frame.
If a frame consists of mixed regions where some regions are moving and others are not, it is typically more efficient to code the non-moving regions in frame mode and the moving regions in the field mode. Therefore, the frames/field encoding decision can be made independently for each vertical pair of macroblocks in a frame. This is the third coding option of the above-listed options. This coding option is referred to as macroblock adaptive frame/field (MBAFF) coding. It is assumed in the following that precoder 2012 decides to use just this option. As described above, MBAFF coding allows the precoder to better adapt the coding mode type (filed or frame mode) to the respective areas of scenes. For example, precoder 2012 codes macroblock pairs located at stationary areas of a video scene in frame mode, while coding macroblock pairs lying in areas of a scene showing fast movements in field mode.
As mentioned above, for a macroblock pair that is coded in frame mode, each macroblock contains frame lines. For a macroblock pair that is coded in field mode, the top macroblock contains top field lines and the bottom macroblock contains bottom field lines. The frame/field decision for each macroblock pair is made at the macroblock pair level by precoder 2012, i.e. if the top macroblock is field coded same applies for the bottom macroblock within same macroblock pair. By this measure, the basic macroblock processing structure is kept intact, and motion compensation areas are permitted to be as large as the size of a macroblock.
Each macroblock of a field macroblock pair is processed very similarly to a macroblock within a field in PAFF coding. However, since a mixture of field and frame macroblock pairs may occur within an MBAFF frame, some stages of the pre-coding procedure in precoder 2012, such as the prediction of motion vectors, the prediction of intra prediction modes, intra frame sample prediction, deblocking filtering and context modeling in entropy coding and the zig-zag scanning of transform coefficients are modified when compared to the PAFF coding in order to account for this mixture.
To summarize, the pre-coded video signal output by precoder 2012 depends on the type of coding precoder 2012 has decided to use. In case of MBAFF coding, as it is assumed herein, the pre-coded video signal contains a flag mb_field_decoding_flag for each non-skipped macroblock pair. The flag mb_field_decoding_flag indicates for each macroblock pair it belongs to whether the corresponding macroblocks are coded in frame or field coding mode. On decoder side, this flag is necessary in order to correctly decode the precoded video signal. In case, the macroblocks of a macroblock pair are coded in frame mode, the flag mb_field_decoding_flag is zero, whereas the flag is one in the other case.
Now, while the general mode of operation of the original decoder arrangement of Fig. 29 has been described without referring to a special bin, with respect to Fig. 32, the functionality of this arrangement is now described with respect to the binary entropy coding of the bin strings of exemplary syntax elements for which the spatial relationship between the syntax element of neighboring blocks is used while MBAFF coding mode is active.
The process shown in Fig. 32 starts at the arrival of a bin value of a syntax element at the input of context modeler 2112. That is, eventually, the syntax element had to be binarized in binarizer 2102 if needed, i.e. unless the syntax element is already a binary value. In a first step 2300, context modeler 2112 determines as to whether the incoming bin is a bin dedicated to a context assignment based on neighboring syntax elements, i.e. syntax elements in neighboring blocks. It is recalled that the description of Fig. 32 assumes that MBAFF coding is active. If the determination in step 2300 results in the incoming bin not being dedicated to context assignment based on neighboring syntax elements, another syntax element handling is performed in step 2304. In the second case, context modeler 2112 determines a neighboring block of the current block to which the syntax element of the incoming bin relates. The determination process of step 2306 is described in more detail below with respect to exemplary syntax elements and their bins, respectively. In any case, the determination in step 2306 depends on the current macroblock to which the syntax element of the current bin relates being frame or field coded, as long as the neighboring block in question is external to the macroblock containing the current block.
Next, in step 2308, the context modeler 2112 assigns a context model to the bin based on a predetermined attribute of the neighboring block. The step of assigning 2308 results in a context index ctxidx pointing to the respective entry in a table assigning each context index a probability model, to be used for binary entropy coding of the current bin of the current syntax element.
After the determination of ctxidx, context modeler 2112 passes the variable ctxidx or the probability estimation status indexed by ctxidx along with the current bin itself to regular coding engine 2118. Based on these inputs, the regular coding engine 2118 entropy encodes, in step 2322, the bin by using the current probability state of the context model as indexed by ctxidx. In particular, the index is used to select the respective partial bitstream 2012, for example.
Thereafter, regular coding engine 2118 passes the bin value via path 2117 back to context modeler 2112, whereupon context modeler 2112 adapts, in step 2324, the context model indexed by ctxidx with respect to its probability estimation state. Thereafter, the process of coding the syntax element into the bit stream at the output 2124 ends at 2326.
It is emphasized that the bin string into which the syntax element may be binarized before step 2310 may be composed of both, bins that are encoded, i.e. assigned to the respective partial bitstream 2012 and subsequently coded, by use of the current probability state of context model ctxidx in step 2322 and bins encoded in bypass coding engine 2120 by use of an equi-probable probability estimation although this is not shown in Fig. 32. Rather, Fig. 32 merely concerns the exemplary encoding of one bin of a syntax element.
The steps 2322 and 2324, encompassed by dotted line 2327 in Fig 32, are explained in more detail below with respect to Fig. 33.
Fig. 33 shows, on the left hand side, a flow diagram of the process 2327. On the right hand side, Fig. 33 shows a memory 2328 to which both, the context modeler 2112 and the regular coding engine 2118, have access in order to load, write, and update specific variables. These 55528
95
variables comprise R and L, which define the current state or current probability interval of the binary entropy coder 2100c.
Memory 2328 contains a table 2329, which associates each possible value of ctxidx, e.g. 0 - 2398, a pair of a probability state index σϋ and an MPS value co, both defining the current probability estimation state of the respective context model indexed by the respective context index ctxidx. The probability state σ is an index that uniquely identifies one of a set of possible probability values ρσ. The probability values ρσ are an estimation for the probability of the next bin of that context model to be a least probable symbol (LPS). Which of the possible bin values, i.e., a null or one, is meant by the LPS, is indicated by the value of MPS ω. If ω is 1, LPS is 0 and vice- versa. Thus, the state index and MPS together uniquely define the actual probability state or probability estimation of the respective context model.
In step 2334, the actual bin value is sent to the regular coding engine 2118.
In step 2336, the regular coding engine 2118 checks as to whether the value of the actual bin, i.e. either the already binary syntax element or one bin of a bin string obtained from the current syntax element, is equal to the most probable symbol as indicated by co; or not. If the current bin is the MPS, context modeler 2112 updates the probability estimation state of the current context model by updating Oj. In particular, context modeler 2112 uses a table 2340 which associates each probability state index σ with an updated probability state index in case the actual symbol or bin was the most probable symbol, i.e., σ becomes transIdxMPS(a; ).
After step 2338, the process ends at 2340 where the bin sent to the regular coding engine 2118 is forwarded by the selector thereof, to the respective partial bitstream encoder.
If in step 2336 it is determined that the current bin is the least probable symbol LPS, it is determined if σ, is equal to 0, i.e. if the probability state index indicates equal probability for both, 1 and 0, in step 2344. Then, the value MPS is updated by computing ω, = 1 - ooi. Thereafter, in step 2346, the probability state index is actualized by use of table 2340, which also associates each current probability state index with an updated probability state index in case the actual bin value is the least probable symbol, i.e., amending σ; to become transIdxLPS(oi). After the probability state index σ, and ω; has been adapted in steps 2344 and 2346, the process steps to step 2340 which has already been described.
After having described the encoding process of syntax elements by exploiting the spatial relationship between syntax element of neighboring blocks for context model assignment, the context model assignment and the definition of the neighborhood between a current and a neighboring block is described in more detail below with respect to the following syntax elements contained in the precoded video signal as output by precoder 2012. These syntax elements are listed below.
Name of the syntax Meaning of the syntax element
element
Mb_skip_flag This flag relates to a certain macroblock of a certain slice of a video frame. Mb skip flag equal to 1 specifies that the current macroblock is to be skipped when performing a decoding process on the precoded video signal. Mb_skip_flag equal to 0 specifies that the current macroblock is not skipped. In particular, in the H.264/AVC standard, Mb_skip_flag equal to 1 specifies that for the current macroblock, when decoding a P or SP slice, Mbjype is inferred to be p_skip and the macroblock type is collectively referred to as P macroblock type, and when decoding a B slice, Mbjype is inferred to be B_skip and the macroblock type is collectively referred to as B macroblock type.
Mb_field_ Mb field_decodingJlag equal to 0 specifies that the current decoding_flag macroblock pair is a frame macroblock pair and Mb_field_deco- dingjlag equal to 0 specifies that the macroblock pair is a field macroblock pair. Both macroblocks of a frame macroblock pair are referred to in the present description as frame macroblocks, whereas both macroblocks of a field macroblock pair are referred to in this text as field macroblocks.
Mbjype Mbjype specifies the macroblock type. For example, the semantics of Mbjype in the H.264/AVC standard depends on the slice type. Depending on the slice type, Mbjype can assume values in the range of 0 to 25, 0 to 30, 0 to 48 or 0 - 26, depending on the slice type.
Coded_block_ Coded_block )attern specifies which of a sub-part of the current pattern macroblock contains non-zero transform coefficients. Transform coefficients are the scalar quantities, considered to be in a frequency domain, that are associated with a particular one-dimensional or two-dimensional frequency index in an inverse transform part of the decoding process. To be more specific, each macroblock 2010a - irrespective of the macroblock being a frame coded macroblock (Fig. 31a) or a field coded macroblock (Fig. 31b), is partitioned into smaller sub-parts, the sub-parts being arrays of size 8x 8 pixel samples. Briefly referring to Fig. 31a, the pixels 1 to 8, 11 to 18, 21 to 28,..., 71 to 78 could form the upper left block of luma pixel samples in the top macroblock 2250 of macroblock pair 2010b. This top macroblock 2250 would comprise another three of such blocks, all four blocks arranged in a 2x2 array. The same applies for the bottom macroblock 2252 and also applies for field coded macroblocks as shown in Fig. 4b, where, for example, pixels 1 to 8, 21 to 28, 41 to 48, 141 to 148 would form the upper left block of the top macroblock. Thus, for each macroblock coded, the precoded video signal output by precoder 2012 would comprise one or several syntax elements coded_block_pattern. The transformation from spatial domain to frequency domain, could be performed on these 8x8 sub-parts or on some smaller units, for example, 4x4 sub- arrays, wherein each 8x8 sub-part comprises 4 smaller 4x4 partitions. The present description mainly concerns luma pixel samples. Nevertheless, the same could also apply accordingly for chroma pixel samples.
ref ldxJO/ This syntax element concerns the prediction of the pixel samples of refjdxjl a macroblock during encoding and decoding. In particular, ref_Idx_10, when present in the precoded video signal output by precoder 12, specifies an index in a list 0 of a reference picture to be used for prediction. The same applies for ref_Idx_ll but with respect to another list of the reference picture.
mvd_10/mvd_ll mvd_10 specifies the difference between a vector component to be used for motion prediction and the prediction of the vector component. The same applies for mvd_ll, the only difference being, that same are applied to different reference picture lists. ref_Idx_10, ref_Idx_ll, mvd_10 and mvd_ll all relate to a particular macroblock partition. The partitioning of the macroblock is specified by Mb type.
intra_chroma_ Intra_chroma_pred_mode specifies the type of spatial prediction pred_mode used for chroma whenever any part of the luma macroblock is intra- coded. In intra prediction, a prediction is derived from the decoded samples of the same decoded picture or frame. Intra prediction is contrary to inter prediction where a prediction is derived from decoded samples of reference pictures other than the current decoded picture.
coded_block_ coded_block_flag relates to blocks of the size of 4x4 picture flag samples. If coded_block_flag is equal to 0, the block contains no non-zero transform coefficients. If coded block flag is equal to 1, the block contains at least one non-zero transform coefficient.
As can be gathered from the above table, some of these syntax elements relate to a current macroblock in the whole, whereas others relate to sub-parts, i.e., sub-macroblocks or partitions thereof, of a current macroblock. In a similar way, the assignment of a context model to these syntax elements is dependent on syntax elements of either neighboring macroblocks, neighboring sub-macroblocks or neighboring partitions thereof. Fig. 36 illustrates the partition of macroblocks (upper row) and sub-macroblocks (lower row). The partitions are scanned for inter prediction as shown in Fig. 36. The outer rectangles in Fig. 36 refer to the samples in a macroblock or sub-macroblock, respectively. The inner rectangles refer to the partitions. The number in each inner rectangle specifies the index of the inverse macroblock partition scan or inverse sub-macroblock partition scan.
Before describing in detail the dependency of the context model assignment on the syntax element of neighboring blocks, with respect to Fig. 34, it is described, how the addresses of the top macroblock of the macroblock pair to the left and above the current macroblock pair may be computed, since these are the possible candidates, which comprise the syntax element in the block to the left of and above the current block containing the current syntax element to be encoded. In order to illustrate the spatial relationships, in Fig. 34, a portion of six macroblock pairs of a video frame is shown, wherein each rectangle region in Fig. 34 corresponds to one macroblock and the first and the second two vertically adjacent macroblocks in each column form a macroblock pair.
In Fig. 34, CurrMbAddr denotes the macroblock address of the top macroblock of the current macroblock pair, the current syntax element is associated with or relates to. The current macroblock pair is encompassed by bold lines. In other words, they from the border of a macroblock pair. mbAddrA and mbAddrB denote the addresses of the top macroblocks of the macroblock pairs to the left and above the current macroblock pair, respectively.
In order to compute the addresses of the top macroblock of the neighboring macroblock pair to the left and above the current macroblock pair, context modeler 2112 computes
MbAddrA = 2 · {CurrMbAddr/2 - 1)
MbAddrB - 2 · (CurrMbAddr/2 - PicWidthlnMbs) where PicWidthlnMbs specifies the picture within units of macroblocks. The equations given above can be understood by looking at Fig. 30. It is noted that in Fig. 30 the picture width in units of macroblocks has been denoted i. It is further noted that the equations given above are also true when the current macroblock address CurrMbAddress is interchanged with the odd numbered macroblock address of the bottom macroblock of the current macroblock pair, i.e., CurrMbAddress+l? because in the equation above, "/" denotes an integer division with transaction of the result towards zero. For example, 7/4 and -11 -A are truncated to 1 and -7/4 and 7/-1 are truncated to -1.
Now, after having described how to compute neighboring macroblocks, it is briefly recalled that each macroblock contains 16x16 luma samples. These luma samples are divided up into four 8x8 luma blocks. These luma blocks may be further subdivided into 4x4 luma blocks. Furthermore, for the following description, each macroblock further comprises 8x8 luma samples, i.e., the pixel width of the chroma samples being doubled compared to luma samples. These 8x8 chroma samples of a macroblock are divided up into four 4x4 luma blocks. The blocks of a macroblock are numbered. Accordingly, the four 8x8 luma blocks each have a respective block address uniquely indicating each 8x8 block in the macroblock. Next, each pixel sample in a macroblock belongs to a position (x, y) wherein (x, y) denotes the luma or chroma location of the upper-left sample of the current block in relation to the upper-left luma or chroma sample of the macroblock. For example, with respect to luma samples, the pixel 2023 in top macroblock 2252 in Fig. 31b would have the pixel position (2, 1), i.e., third column, second row.
After having described this, the derivation process of ctxidx for at least some of the bins of syntax elements listed in the above table is described.
With respect to the syntax element mb_skip_flag, the context modeler assignment depends on syntax elements relating to neighboring macroblocks. Thus, in order to determine the context index ctxidx the addresses mbAddrA and mbAddrB are determined as described above. Then, let condTermN (with N being either A or B) be a variable that is set as follows:
If mbAddrN is not available or mb_skip_flag for the macroblock mbAddrN is equal to 1, conTermN is set to 0
Otherwise, condTermN is set to 1. ctxidx is derived based on an context index incrementor ctxldxlnc = conTermA + condTermB. -
For the syntax element mb_field_decoding_flag, ctxidx is determined as follows:
Let condTermN (with N being either A or B) be a variable that is set as follows.
If any of the following conditions is true, then condTermN is set to 0, mbAddrN is not available
the macroblock mbAddrN is a frame macroblock.
Otherwise, condTermN is set to 1. ctxidx is derived based on an context index incrementor ctxldxlnc = condTermA + condTermB wherein mbaddrN is not available, if
(((CurrMbAddr/2)%(PicWidthInMbs)) = 0).
For the syntax element Mbjype, ctxidx is determined dependent on the availability of macroblocks mbAddrN (with N being either A or B), and the syntax element Mbjype of this neighboring macroblocks.
With respect to the other syntax element listed in the above table, the dependency of the context modeler assignment is defined accordingly, wherein for syntax elements, which relate to blocks smaller than a macroblock, the assignment is also dependent on syntax element relating to such smaller blocks being smaller than macroblocks. For example, for the syntax element coded_block_pattern, the context index assignment is dependent not only on the availability of macroblock MbAddrN and the syntax element Mbjype of the macroblock MbAddrN but also on the syntax element Coded_block_pattern of the neighboring block. Further, it is worth noting that the syntax elements listed above are all dependent on the respective syntax element of the neighboring block. Differing thereto, the context model assignment of syntax elements mvd_10, mvd_ll, ref_idx 0 and ref_idx_ll is not dependent on the respective syntax elements of the neighboring block. The context modeler assignment for intra_chroma_pred_mode is dependent on mbAddrN availability, macroblock mbAddrN being coded in inter prediction mode or not, Mb_type for the macroblock mbAddrN and the syntax element intra_chroma_pred_mode for the macroblock MbAddrN. The syntax element coded_block_flag context model assignment is dependent on the availability of MbAddrN, the current macroblock being coded in inter prediction mode, Mb_type for the macroblock mbAddrN and the syntax element coded_block_flag of the neighboring block.
In the following, it is described, how a neighboring block is determined. In particular, this involves computing mbAddrN and the block index indexing the sub-part of the macroblock MbAddrN, this sub-part being the neighboring block of the current block. The neighborhood for slices using macroblock adaptive frames/field coding as described in the following in accordance with an embodiment of the present invention is defined in a way that guarantees that the areas covered by neighboring blocks used for context modeling in context adaptive binary entropy coding inside an MBAFF-frame adjoin to the area covered by the current block. This generally improves the coding efficiency of a context adaptive entropy coding scheme as it is used here in connection with the coding of MBAFF-slices in comparison to considering each macroblock pair as frame macroblock pair for the purpose of context modeling as described in the introductory portion of the specification, since the conditional probabilities estimated during the coding process are more reliable.
The general concept of defining the neighborhood between a current and a reference block is described in the following section 1.1. In section 1.2, a detailed description, which specifies how the neighboring blocks, macroblocks, or partitions to the left of and above the current block, macroblock, or partition are obtained for the purpose of context modeling in context adaptive binary entropy coding, is given.
1.1. General concept neighborhood definition
Let (x, y) denote the luma or chroma location of the upper-left sample of the current block in relation to the upper-left luma or chroma sample of the picture CurrPic. The variable CurrPic specifies the current frame, which is obtained by interleaving the top and the bottom field, if the current block is part of a macroblock pair coded in frame mode (rnb_field_decoding_fIag is equal to 0). If the current block is or is part of a top field macroblock, CurrPic specifies the top field of the current frame; and if the current block is or is part of a bottom field macroblock, CurrPic specifies the bottom field of the current frame.
Let (xA, yA) and (xB, yB) denote the luma or chroma location to the left of and above the location (x, y), respectively, inside the picture CurrPic. The locations (xA, yA) and (xB, yB) are specified by
(xA, yA) = (x -l, y)
(xB, yB) = (x, y -l)
The block to the left of the current block is defined as the block that contains the luma or chroma sample at location (xA, yA) relative to the upper-left luma or chroma sample of the picture CurrPic and the block above the current block is defined as the block that contains the luma or chroma sample at location (xB, yB) relative to the upper-left luma or chroma sample of the picture CurrPic. If (xA, yA) or (xB, yB) specify a location outside the current slice, the corresponding block is marked as not available. 1.2. Detailed description of neighborhood definition
The algorithm described in Sec. 1.2.1 specifies a general concept for MBAFF-slices that describes how a luma sample location expressed in relation to the upper-left luma sample of the current macroblock is mapped onto a macroblock address, which specifies the macroblock that covers the corresponding luma sample, and a luma sample location expressed in relation to the upper-left luma sample of that macroblock. This concept is used in the following Sec. 1.2.2 - 1.2.6.
The Sec. 1.2.2 - 1.2.6 describe how the neighboring macroblocks, 8x8 luma blocks, 4x4 luma blocks, 4x4 chroma block, and partitions to the left of and above a current macroblock, 8x8 luma block, 4x4 luma block, 4x4 chroma block, or partition are specified. These neighboring macroblock, block, or partitions are needed for the context modeling of CABAC for the following syntax elements: mb_skip_flag, mb_type, coded_block_pattern, intra_chroma_pred_mode, coded_block_fiag, ref_idx_10, ref_idx_ll, mvdJLO, mvd_ll.
1.2.1 Specification of neighboring sample locations
Let (xN, yN) denote a given luma sample location expressed in relation to the upper-left luma sample of the current macroblock with the macroblock address CurrMbAddr. It is recalled that in accordance with the present embodiment each macroblock comprises 16x16 luma samples. xN and yN lie within— 1...16. Let mbAddrN be the macroblock address of the macroblock that contains (xN, yN), and let (xW,yW) be the, location (xN, yN) expressed in relation to the upper-left luma sample of the macroblock mbAddrN (rather than relative to the upper-left luma sample of the current macroblock).
Let mbAddrA and mbAddrB specify the macroblock address of the top macroblock of the macroblock pair to the left of the current macroblock pair and the top macroblock of the macroblock pair above the current macroblock pair, respectively. Let PicWidthlnMbs be a variable that specifies the picture width in units of macroblocks. mbAddrA and mbAddrB are specified as follows. mbAddrA = 2* (CurrMbAddr/2 -1)
If mbAddrA is- less than 0, or if (CurrMbAddr/2) % PicWidthlnMbs is equal to 0, or if the macroblock with address mbAddrA belongs to a different slice than the current slice, mbAddrA is marked as not available. mbAddrB = 2* (CurrMbAddr/2 - PicWidthlnMbs) If mbAddrB is less than 0, or if the macroblock with address mbAddrB belongs to a different slice than the current slice, mbAddrB is marked as not available.
The Table in Fig. 35 specifies the macroblock address mbAddrN and a variable yM in the following two ordered steps:
1. Specification of a macroblock address mbAddrX (fifth column)depending on (xN, yN) (first and second column) and the following variables:
The variable currMbFrameFlag (third column) is set to 1, if the current macroblock with address CurrMbAddr is a part of a frame macroblock pair; otherwise it is set to 0.
The variable mbl sTopMbFlag (forth column) is set to 1 , if CurrMbAddr%2 is equal to 0; otherwise it is set to 0.
2. Depending on the availability of mbAddrX (fifth column), the following applies:
If mbAddrX (which can be either mbAddrA or mbAddrB) is marked as not available, mbAddrN is marked as not available.
Otherwise (mbAddrX is available), mbAddrN is marked as available and Table 1 specifies mbAddrN and yM depending on (xN, yN) (first and second column), currMbFrameFlag (third column), mblsTopMbFlag (forth column), and the variable mbAddrXFrameFlag (sixth column), which is derived as follows: rribAddrXFrameFlag is set to 1, if the macroblock mbAddrX is a frame macroblock; otherwise it is set to 0.
Unspecified values of the above flags in Table 1 indicate that the value of the corresponding flags is not relevant for the current table rows.
To summarize: in the first four columns, the input values xN, yN, currMbFrameFlag and MblsTopMbFlag are entered. In particular, the possible input values for parameters xN and yN are -1 to 16, inclusive. These parameters determine mbAddrX listed in the fifth column, i.e. the macroblock pair containing the wanted luma sample. The next two columns, i.e., the sixth and the seventh column, are needed to obtain the final output mbAddrN and yN. These further input parameters are MbAddrXFrameFlag indicating as to whether a macroblock pair indicated by mbAddrX is frame or field coded, and some additional conditions concerning as to whether yN is even or odd numbered or is greater than or equal to 8 or not. As can be seen, when xN and yN are both positive or zero, i.e., the wanted pixel sample lies within the current macroblock relative to which xN and yN are defined, the output macroblock address does not change, i.e., it is equal to CurrMbAddr. Moreover, yM is equal yM. This changes when the input xM and yM indicates a pixel sample lying outside the current macroblock, i.e., to the left (xN < 0) all to the top of the current macroblock (yN < 0).
Outgoing from the result of the table of Fig. 35, the neighboring luma location (xW, yW) relative to the upper-left luma sample of the macroblock-mbAddrN is specified as xW=(xN+16)%16
yW=(yM+16)%16.
It is emphasized that the aforementioned considerations pertained for illustrative purposes merely luma samples. The considerations are slightly different when considering chroma samples since a macroblock contains merely 8x8 chroma samples.
1.2.2 Specification of neighboring macroblocks
The specification of the neighboring macroblocks to the left of and above the current macroblock is used for the context modeling of CAB AC for the following syntax elements: mb_skip_flag, mb_type, coded_block_pattern, intra_chroma_prediction_mode, and coded_block_flag.
Let mbAddrA be the macroblock address of the macroblock to the left of the current macroblock, and mbAddrB be the macroblock address of the macroblock above the cvtrrent macroblock. mbAddrA, mbAddrB, and their availability statuses are obtained as follows:
• mbAddrA and its availability status are obtained as described in Sec. 1.2.1 given the luma location (xN, yN) = (-1, 0).
• mbAddrB and its availability status are obtained as described in Sec. 1.2.1 given the luma location (xN, yN) ='(0, -1).
1.2.3 Specification of neighboring 8x8 luma blocks The specification of the neighboring 8x8 luma blocks to the left of and above the current 8x8 luma block is used for the context modeling of CABAC for the syntax element coded_block_pattern.
Let luma8x8BIkIdx be the index of the current 8x8 luma block inside the current macroblock CurrMbAddr. An embodiment of the assignment of block index luma8x8BlkIdx to the respective blocks within a macroblock is shown in Fig. 9 (upper-right corner).
Let mbAddrA be the macroblock address of the macroblock that contains the 8x8 luma block to the left of the current 8x8 luma block, and let mbAddrB be the macroblock address of the macroblock that contains the 8x8 luma block above the current 8x8 luma block. Further, let luma8x8BlkIdxA be the 8x8 luma block index (inside the macroblock mbAddrA) of the 8x8 luma block to the left of the current 8x8 luma block, and let luma8x8BlkIdXB be the 8x8 luma block index (inside the macroblock mbAddrB) of the 8x8 luma block above the current 8x8 luma block. mbAddrA, mbAddrB, luma8X8B dxA, luma8x8BlkIdxB, and their availability statuses are obtained as follows:
• Let (xC, yC) be the luma location of the upper-left sample of the current 8x8 luma block relative to the upper-left luma sample of the current macroblock.
• mbAddrA, its availability status, and the luma location (xW, yW) are obtained as described in Sec. 1.2.1 given the luma location ( N, yN) = (xC -1, yC). If mbAddrA is available, then luma8x8BlkIdxA is set in a way that it refers to the 8x8 luma block inside the macroblock mbAddrA that covers the luma location (xW, yW); otherwise, iumaSxSBlkldA is marked as not available.
• mbAddrB, its availability status, and the luma location (xW, yW) are obtained as described in Sec. 1.2.1 given the luma location (xN, yN) = (xC, yC -1). If mbAddrB is available, then luma8x8BlkIdxB is set in a way that it refers to the 8x8 luma block inside the macroblock mbAddrB that covers the luma location (xW, y W); otherwise,
luma8x8BlkldxB is marked as not available.
1.2.4 Specification of neighboring 4x4 luma blocks
The specification of the neighboring 4x4 luma blocks to the left of and above the current 4x4 luma block is used for the context modeling of CABAC for the syntax element coded_block_flag. Let luma4x4BlkIdx be the index (in decoding order) of the current 4x4 luma block inside the current macroblock CurrMbAddr. For example, luma4x4BlkIdx could be defined as luma8x8BlkIdx of the 8x8 block containing the 4x4block multiplied by 4 plus the partition number as shown in the bottom-right corner of Fig. 9.
Let mbAddrA be the macroblock address of the macroblock that contains the 4x4 luma block to the left of the current 4x4 luma block, and let mbAddrB be the macroblock address of the macroblock that contains the 4x4 luma block above the current 4x4 luma block. Further, let luma4x4BlkIdxA be the 4x4 luma block index (inside the macroblock mbAddrA) of the 4x4 luma block to the left of the current 4x4 luma block, and let luma4x4BlkIdxB be the 4x4 luma block index (inside the macroblock mbAddrB) of the 4x4 luma block above the current 4x4 luma block. mbAddrA, mbAddrB, luma4x4BlkIdxA, luma4x4BlkIdxB, and their availability statuses are obtained as follows:
• Let (xC, yC) be the luma location of the upper-left sample of the current 4x4 luma block relative to the upper-left luma sample of the current macroblock.
• mbAddrA, its availability status, and the luma location (xW, yW) are obtained as described in Sec. 1.2.1 given the luma location (xN, yN) = (xC -1, yC). if mbAddrA is available, then luma4x4BlkIdxA is set in a way that it refers to the 4x4 luma block inside the macroblock mbAddrA that covers the luma location (xW, yW); otherwise, luma4x4BlkIdxA is marked as not available.
• mbAddrB, its availability status, and the luma location (xW, yW) are obtained as described in Sec. 1.2.1 given the luma location (xN, yN) = (xC, yC - 1). If mbAddrB is available, then luma4x4BlkIdxB is set in a way that it refers to the 4x4 luma block inside the macroblock mbAddrB that covers the luma location (xW, yW); otherwise, Iuma4x4BlkIdxB is marked as not available.
1.2.5 Specification of neighboring 4x4 chroma blocks
The specification of the neighboring 4x4 chroma blocks to the left of and above the current 4x4 chroma block is used for the context modeling of CABAC for the syntax element coded_block_flag.
Let chroma4x4BlkIdx be the index (in decoding order) of the current 4x4 chroma block inside the current macroblock CurrMbAddr. Let mbAddrA be the macroblock address of the macroblock that contains the 4x4 chroma block to the left of the current 4x4 chroma block, and let mbAddrB be the macroblock address of the macroblock that contains the 4x4 chroma block above the current 4x4 chroma block. Further, let chroma4x4BlkIdxA be the 4x4 chroma block index (inside the macroblock mbAddrA) of the 4x4 chroma block to the left of the current 4x4 chroma block, and let chroma4x4BlkIdxB be the 4x4 chroma block index (inside the macroblock mbAddrB) of the 4x4 chroma block above the current 4x4 chroma block. mbAddrA, mbAddrB, chroma4x4BlkIdxA, chroma4x4BlkIdxB, and their availability statuses are obtained as follows:
Given luma8x8BlkIdx = chroma4x4BlkIdx, the variables mbAddrA, mbAddrB, luma8x8BlkIdxA, luma8x8BlkIdxB, and their availability statuses are obtained as described in Sec. 1.2.3.
If luma8x8BIkIdxA is available, chroma4x4BIkIdxA is set equal to luma8x8BlkIdxA; otherwise chroma4x4BlkIdxA is marked as not available.
If luma8x8BlkIdxB is available, chroma4x4BlkIdxB is set equal to luma8x8BlkIdxB; otherwise chroma4x4BlkIdxB is marked as not available.
1.2.6 Specification of neighboring partitions
The specification of the neighboring partitions to the left of and above the current partition is used for the context modeling of CAB AC for the following syntax elements: ref_idx_10, ref_idx_ll, mvd_10, mvd_ll.
Let mbPartldx and subMbPartldx be the macroblock partition and sub-macroblock partition indices that specify the current partition inside the current macroblock CurrMbAddr. An example for such partition indices is shown in Fig. 36.
Let mbAddrA be the macroblock address of the macroblock that contains the partition to the left of the current partition, and let mbAddrB be the macroblock address of the macroblock that contains the partition above the current partition. Further, let mbPartldxA and subMbPartldxA be the macroblock partition and sub-macroblock partition indices (inside the macroblock mbAddrA) of the partition to the left of the current partition, and let mbPartldxB and subMbPartldxB be the macroblock partition and sub-macroblock partition indices (inside the macroblock mbAddrB) of the partition above the current partition. mbAddrA, mbAddrB, mbPartldxA, subMbPartldxA, mbPartldxB, subMbPartldxB, and their availability statuses are obtained as follows:
• Let (xC, yC) be the luma location of the upper-left sample of the current partition given by mbPartldx and subMbPartldx relative to the upper-left luma sample of the current macroblock.
• mbAddrA, its availability status, and the luma location (xW, yW) are obtained as described in Sec. 1.2.1 given the luma location (xN,yN)=(xC-l, yC). If mbAddrA is not available, mbPartldxA and subMbPartldxA are marked as not available; otherwise mbPartldxA is set in a way that it refers to the macroblock partition inside the macroblock mbAddrA that covers the luma location (xW, y W), and subMbPartldxA is set in a way that it refers to the sub-macroblock partition inside the macroblock partition mbPartldxA (inside the macroblock mbAddrA) that covers the luma location (xW, yW).
• mbAddrB, its availability status, and the luma location (xW, yW) are obtained as described in Sec. 1.2.1 given the luma location ( N, yN):=(xC, yC -1). If mbAddrB is not available, mbPartldxB and subMbPartldxB are marked as not available; otherwise mbPartldxB is set in a way that it refers to the macroblock partition inside the macroblock mbAddrB that covers the luma location (xW, yW), and subMbPartldxB is set in a way that it refers to the sub-macroblock partition inside the macroblock partition mbPartldxB (inside the macroblock mbAddrB) that covers the luma location (xW, yW).
After having described how to encode the above syntax elements or the bin strings or part of their bins into an entropy coded bit stream, the decoding of said bit stream and the retrieval of the bins is described with respect to Figs. 37 to 39
Fig. 37 shows a general view of a video decoder environment to which the present invention could be applied. An entropy decoder 400 receives the entropy coded bit stream as described above and treats it as will be described in more detail below with respect to Figs. 38 - 39. In particular, the entropy decoder 2400 decodes the entropy coded bit stream by binary entropy decoding in order to obtain the precoded video signal and, in particular, syntax elements contained therein and passes same to a precode decoder 2402. The precode decoder 2402 uses the syntax elements, such as motion vector components and flags, such as the above listed syntax elements, in order to retrieve, macroblock by macroblock and then slice after slice, the picture samples of pixels of the video frames 2010.
Fig. 38 now shows the decoding process performed by the entropy decoder 2400 each time a bin is to be decoded. Which bin is to be decoded depends on the syntax element which is currently expected by entropy decoder 2400. This knowledge results from respective parsing regulations.
In the decoding process, first, in step 2500, the decoder 2400 checks as to whether the next bin to decode is a bin of a syntax element of the type corresponding to context model assignment based on neighboring syntax elements. If this is not the case, decoder 2400 proceeds to another syntax element handling in step 2504. However, if the check result in step 2500 is positive, decoder 2400 performs in steps 2506 and 2508 a determination of the neighboring block of the current block which the current bin to decode belongs to and an assignment of a context model to the bin based on a predetermined attribute of the neighboring block determined in step 2506, wherein steps 2506 and 2508 correspond to steps 2306 and 2308 of encoding process of Fig. 32. The result of these steps is the context index ctxldx. Accordingly, the determination of ctxldx is performed in steps 2506 and 2508 in the same way as in the encoding process of Fig. 32 in steps 2306 and 2308 in order to determine the context model to be used in the following entropy decoding.
Then, in step 2522, the entropy decoder 2400 entropy decodes the actual bin, from the entropy coded bit stream by use of the actual probability state of the context module as indexed by ctxldx obtained in steps 2510 to 2520. That is, the entropy decoder 2400 comprises a regular decoding engine that corresponds to any of the above-described entropy decoders having a plurality of entropy decoders inside, each associated with a respective partial bitstream. In step 2522 the bin is requested from that partial bitstream which is indexed by the probability index or, at least, depending therefrom. The result of this step is the value for the actual bin.
Thereafter, in step 2524, the ctxldx probability state is adapted or updated, as it was the case in step2 224. Thereafter, the process ends at step 2526.
Of course, the individual bins that are obtained by the process shown in Fig. 38 represent the syntax element value merely in case the syntax element is of a binary type. Otherwise, a step corresponding to the binarization has to be performed in reverse manner in order to obtain from the bin strings the actual value of the syntax element.
Fig. 39 shows the steps 2522 and 2524 being encompassed by dotted line2 527 in more detail on the left hand side. On the right hand side, indicated with 2564, Fig. 38 shows a memory and its content to which entropy decoder 2400 has access in order to load, store and update variables. As can be seen, entropy decoder manipulates or manages the same variables as entropy coder 2014 since entropy decoder 2400 emulates the encoding process as will be described in the following. In a first step 2566, decoder 2400 obtains the value of the current bin from the partial bitstream indexed by the probability index associated with the respective context.
However, the value is yet indicated in relative terms with respect the values MPS and LPS, Therefore, in step 2570 decoder 2400 checks as to whether the value of the entropy coding codeword in the entropy coded bit stream was the most probable symbol as indicated by G>I; in which case decoder 2400 accordingly sets the bin value to the value of co; in step 2572. In the other case, decoder 2400 sets the symbol to be 1-ω, in step 2574.
After step 2572, in step 2576, the decoder 2400 adapts or updates the probability state of the current context model i as defined by o>j and G>; by transitioning the probability state index coj as was described with respect to step 2338 in Fig. 36. Thereafter, the process 2527 ends at step 2578.
After step 2574, decoder 2400 adapts or updates the probability state in steps 2582 and 2584 by computing ω; = Ι-ω, in step 2582, if ω; is equal to 0, and transitioning the probability state index Oj to a new probability state index in the same way as described with respect to step 2346 in Fig. 36. Thereafter, the process ends at step 2578.
After having described specific embodiments, it is noted that same are not restricted to these embodiments. In particular, the assignment in accordance with steps 2308 and 2408 does not have to be dependent on syntax elements of neighboring blocks, i.e., syntax elements contained in the precoded video signal output by precoder 2012. Rather, the assignment may be dependent on other attributes of the neighboring blocks. Moreover, the definition of neighborhoods between neighboring blocks is described with respect to the table of Fig. 35 may be varied. Further, the pixel samples of the two interlaced fields could be arranged in another way than described above.
Moreover, other block sizes than 4x4 blocks could be used as a basis for the transformation, and, although in the above embodiment the transformation was applied to picture sample differences to a prediction, the transformation could be as well applied to the picture sample itself without performing a prediction. Furthermore, the type of transformation is not critical. DCT could be used as well as a FFT or wavelet transformation. Furthermore, the example outlined is not restricted to binary entropy encoding/decoding. Rather, same can be applied to multi-symbol entropy encoding as well. Additionally, the sub-divisions of the video frame into slices, macroblock pairs, macroblocks, picture elements etc. was for illustrating purposes only, and this is not to restrict the scope of the examples. In the following, reference is made to Fig. 40 to show, in more detail than in Fig. 28, the complete setup of a video encoder engine including an entropy-encoder as it is shown in Fig. 40 in block 2800 in which the aforementioned entropy coding of syntax elements by use of a context assignment based on neighboring syntax elements is used. In particular, Fig. 40 shows the basic coding structure for the emerging H.264/AVC standard for a macroblock. The input video signal is, split into macroblocks, each macroblock having 16 x 16 luma pixels. Then, the association of macroblocks to slice groups and slices is selected, and, then, each macroblock of each slice is processed by the network of operating blocks in Fig. 40. It is to be noted here that an efficient parallel processing of macroblocks is possible, when there are various slices in the picture. The association of macroblocks to slice groups and slices is performed by means of a block called coder control 2802 in Fig. 40. There exist several slices, which are defined as follows:
• I slice: A slice in which all macroblocks of the slice are coded using intra prediction.
• P slice: In addition, to the coding types of the I slice, some macroblocks of the P slice can also be coded using inter prediction with at most one motion-compensated prediction signal per prediction block.
• B slice: In addition, to the coding types available in a P slice, some macroblocks of the B slice can also be coded using inter prediction with two motion-compensated prediction signals per prediction block.
The above three coding types are very similar to those in previous standards with the exception of the use of reference pictures as described below. The following two coding types for slices are new:
• SP slice: A so-called switching P slice that is coded such that efficient switching between different precoded pictures becomes possible.
• SI slice: A so-called switching I slice that allows an exact match of a macroblock in an SP slice for random access and error recovery purposes.
Slices are a sequence of macroblocks, which are processed in the order of a raster scan when not using flexible macroblock ordering (FMO). A picture maybe split into one or several slices as shown in Fig. 42. A picture is therefore a collection of one or more slices. Slices are self-contained in the sense that given the active sequence and picture parameter sets, their syntax elements can be parsed from the bit stream and the values of the samples in the area of the picture that the slice represents can be correctly decoded without use of data from other slices provided that utilized reference pictures are identical at encoder and decoder. Some information from other slices maybe needed to apply the deblocking filter across slice boundaries.
FMO modifies the way how pictures are partitioned into slices and macroblocks by utilizing the concept of slice groups. Each slice group is a set of macroblocks defined by a macroblock to slice group map, which is specified by the content of the picture parameter set and some information from slice headers. The macroblock to slice group map consists of a slice group identification number for each macroblock in the picture, specifying which slice group the associated macroblock belongs to. Each slice group can be partitioned into one or more slices, such that a slice is a sequence of macroblocks within the same slice group that is processed in the order of a raster scan within the set of macroblocks of a particular slice group. (The case when FMO is not in use can be viewed as the simple special case of FMO in which the whole picture consists of a single slice group.)
Using FMO, a picture can be split into many macroblock-scanning patterns such as interleaved slices, a dispersed macroblock allocation, one or more "foreground" slice groups and a "leftover" slice group, or a checker-board type of mapping.
Each macroblock can be transmitted in one of several coding types depending on the slice- coding type. In all slice-coding types, the following types of intra coding are supported, which are denoted as Intra_4x4 or Intra_16xl6 together with chroma prediction and I_PCM prediction modes.
The Intra_4x4 mode is based on predicting each 4x4 luma block separately and is well suited for coding of parts of a picture with significant detail. The Intra_16xl6 mode, on the other hand, does prediction of the whole 16x16 luma block and is more suited for coding very smooth areas of a picture.
In addition, to these two types of luma prediction, a separate chroma prediction is conducted. As an alternative to Intra_4x4 and Intra_16xl6, the I_PCM coding type allows the encoder to simply bypass the prediction and transform coding processes and instead directly send the values of the encoded samples. The I_PCM mode serves the following purposes:
1. It allows the encoder to precisely represent the values of the samples
2. It provides a way to accurately represent the values of anomalous picture content without significant data expansion 3. It enables placing a hard limit on the number of bits a decoder must handle for a macroblock without harm to coding efficiency.
In contrast to some previous video coding standards (namely H.263+ and MPEG-4 Visual), where intra prediction has been conducted in the transform domain, intra prediction in H.264/AVC is always conducted in the spatial domain, by referring to the bins of neighboring samples of previously coded blocks which are to the left and/or above the block to be predicted. This may incur error propagation in environments with transmission errors that propagate due to motion compensation into inter-coded macroblocks. Therefore, a constrained intra coding mode can be signaled that allows prediction only from intra-coded neighboring macroblocks.
When using the Intra_4x4 mode, each 4x4 block is predicted from spatially neighboring samples as illustrated on the left-hand side of Fig. 43. The 16 samples of the 4x4 block, which are labeled as a-p, are predicted using prior decoded samples in adjacent blocks labeled as A- Q. For each 4x4 block one of nine prediction modes can be utilized. In addition, to "DC" prediction (where one value is used to predict the entire 4x4 block), eight directional prediction modes are specified as illustrated on the right-hand side of Fig. 41. Those modes are suitable to predict directional structures in a picture such as edges at various angles.
In addition, to the intra macroblock coding types, various predictive or motion-compensated coding types are specified as P macroblock types. Each P macroblock type corresponds to a specific partition of the macroblock into the block shapes used for motion-compensated prediction. Partitions with luma block sizes of 16x16, 16x8, 8x16, and 8x8 samples are supported by the syntax. In case partitions with 8x8 samples are chosen, one additional syntax element for each 8x8 partition is transmitted. This syntax element specifies whether the corresponding 8x8 partition is further partitioned into partitions of 8x4, 4x8, or 4x4 luma samples and corresponding chroma samples.
The prediction signal for each predictive-coded MxN luma block is obtained by displacing an area of the corresponding reference picture, which is specified by a translational motion vector and a picture reference index. Thus, if the macroblock is coded using four 8x8 partitions and each 8x8 partition is further split into four 4x4 partitions, a maximum of sixteen motion vectors may be transmitted for a single P macroblock.
The quantization parameter SliceQP is used for determining the quantization of transform coefficients in H.264/AVC. The parameter can take 52 values. These values are arranged so that an increase of 1 in quantization parameter means an increase of quantization step size by approximately 12 % (an increase of 6 means an increase of quantization step size by exactly a factor of 2). It can be noticed that a change of step size by approximately 12 % also means roughly a reduction of bit rate by approximately 12 %.
The quantized transform coefficients of a block generally are scanned in a zig-zag fashion and transmitted using entropy coding methods. The 2x2 DC coefficients of the chroma component are scanned in raster-scan order. All inverse transform operations in H.264/AVC can be implemented using only additions and bit-shifting operations of 16-bit integer values. Similarly, only 16-bit memory accesses are needed for a good implementation of the forward transform and quantization process in the encoder.
The entropy encoder 2800 in Fig. 40 in accordance with a coding arrangement described above with respect to Fig. 29. A context modeler feeds a context model, i.e., a probability information, to an entropy encoder, which is also referred to as the regular coding engine. The to be encoded bit, i.e. a bin, is forwarded from the context modeler to the regular coding engine. This bin value is also fed back to the context modeler so that a context model update can be obtained. A bypass branch is provided, which includes an entropy encoder, which is also called the bypass coding engine. The bypass coding engine is operative to entropy encode the input bin values. Contrary to the regular coding engine, the bypass coding engine is not an adaptive coding engine but works preferably with a fixed probability model without any context adaption. A selection of the two branches can be obtained by means of switches. The binarizer device is operative to binarize non-binary valued syntax elements for obtaining a bin string, i.e., a string of binary values. In case the syntax element is already a binary value syntax element, the binarizer is bypassed.
In the co-called regular coding mode, prior to the actual entropy coding process the given binary decision, which, in the sequel, we will refer to as a bin, enters the context modeling stage, where a probability model is selected such that the corresponding choice may depend on previously encoded syntax elements or bins. Then, after the assignment of a context model the bin value along with its associated model is passed to the regular coding engine, where the final stage of entropy encoding together with a subsequent model updating takes place (see Fig. 29).
Alternatively, the bypass coding mode is chosen for selected bins in order to allow a speedup of the whole encoding (and decoding) process by means of a simplified coding engine without the usage of an explicitly assigned model. This mode is especially effective when coding the bins of the primary suffix of those syntax elements, concerning components of differences of motion vectors and transform coefficient levels. In the following, the three main functional building blocks, which are binarization, context modeling, and binary entropy coding in the encoder of Fig. 40, along with their
EXAMPLE 5
In this example, example 4 is further extended and Fig. 34, 35, 46 - 49. In accordance with the present example, binarizing a data value is performed by binarizing the minimum of the data value and a predetermined cut-off value in accordance with a first binarization scheme, in order to yield a primary prefix, and, if the data value is greater than the cut-off value, binarizing a difference of the data value minus the predetermined cut-off value in accordance with a second binarization scheme to obtain a binary suffix, the first binarization scheme being different from the second binarization scheme, and appending the primary suffix to the primary prefix.
A data value is recovered from a binarized representation of the data value, the binarized representation of the data value being a codeword having a primary prefix, which is a binarization of the minimum of the data value and a predetermined cut-off value in accordance with a first binarization scheme and, if the data value is greater than the predetermined cut-off value, a primary suffix appended to the primary prefix, the primary suffix being a binarization of the difference of the data value minus the predetermined cut-off value in accordance with a second binarization scheme, the method comprising extracting, from the primary prefix, the minimum, and, if the minimum is equal to the cut-off value, extracting, from the primary suffix, the difference from the data value minus the predetermined cut-off value . Then, the predetermined cut-off value is added to the difference, to obtain the data value, and, if the minimum is smaller than the cut-off value, the minimum is regarded as the data value.
A coding of a data value into a coded bit stream comprises binarizing the minimum of the data value and a predetermined cut-off value in accordance with a first binarization scheme, in order to yield a primary prefix, and, if the data value is greater than the cut-off value, binarizing a difference of the data value minus the predetermined cut-off value in accordance with a second binarization scheme to obtain a binary suffix, the first binarization scheme being different from the second binarization scheme, and appending the primary suffix to the primary prefix. Then, for each bit in the codeword, if the bit of the code word is part of the primary prefix, binary entropy coding the bit by means of a adaptively varying bit value probability estimation, is performed according to any of the above embodiments with probability estimation dependent partial bitstream selection, or, if the bit of the code word is part of the primary suffix, binary coding the bit by means of a static bit value probability estimation, is performed, thereby obtaining the coded bit stream.
Decoding a coded bit stream which represents a binarized representation of the data value is described, the binarized representation of the data value being a codeword having a primary prefix, which is a binarization of the minimum of the data value and a predetermined cut-off value in accordance with a first binarization scheme and, if the data value is greater than the predetermined cut-off value, a primary suffix appended to the primary prefix, the primary suffix being a binarization of the difference of the data value minus the predetermined cut-off value in accordance with a second binarization scheme, the method comprising, for each bit in the codeword, if the bit of the codeword is part of the primary prefix, determining the bit by binary entropy decoding the coded bit stream by means of a adaptively varying bit value probability estimation, namely by retrieving the bit via the partial bitstream associated with the probability estimation, and, if the bit of the codeword is part of the primary suffix, determining the bit by binary decoding the bit by means of a static bit value probability estimation, thereby obtaining the codeword. Then, the minimum is extracted from the primary prefix. If the minimum is equal to the cut-off value, the difference from the data value minus the predetermined cut-off value is extracted from the primary suffix and the predetermined cut-off value is added to the difference, to obtain the data value. If the minimum is smaller than the cut-off value, the minimum is regarded as the data value.
A very effective compression of data values, and, in particular, of components of motion vector differences or transform coefficient level values, forming syntax elements in a video signal, is achieved by using a binarization scheme for preparing the syntax elements for the entropy coding, the binarization scheme substantially being a combination of two different binarization schemes, and by using binary entropy coding instead of m-ary entropy coding for coding the binarized syntax elements.
The advantage of using a binary entropy coding engine instead of a m-ary entropy coder operating on an original m-ary source alphabet is that the complexity of the probability estimation is reduced since the determination of the probabilities for the two possible bit values can be defined by just one probability estimation value. Adaptive m-ary entropy coding, for m > 2, would be in general a computationally complex operation requiring a number of fairly complex operations to perform the update of the probability estimation.
In accordance with a specific embodiment, a concatenation of a truncated unary binarization scheme and a kth order exponential Golomb binarization scheme is used. This concatenated scheme, which is referred to as unary/kth order Exp-Golomb (UEGk) binarization, is applied to motion vector differences and absolute values of transform coefficient levels. The design of these concatenated binarization schemes is motivated by the following observations. First, the unary code is the simplest prefix-free code in terms of implementation cost. Secondly, it permits a fast adaptation of the individual symbol probabilities in a subsequent context modeling stage, since the arrangement of the nodes in the corresponding tree is typically such that with increasing distance of the internal nodes from the root node the corresponding binary probabilities are less skewed. These observations are accurate for small values of the absolute motion vector differences and transform coefficient levels. For larger values, there is not much use of an adaptive modeling having led to the idea of concatenating an adapted truncated unary tree as a prefix and a static Exp-Golomb code tree as a suffix. Typically, for larger values, the EGk suffix part represents already a fairly good fit to the observed probability distribution, as already mentioned above.
In order to illustrate the relation between "bits" or bins of the codewords to which binarizer 2102 maps the syntax elements, on the one hand, and "bins", on the other hand, Fig. 44 shows a binary code tree corresponding to the binarization scheme later described with respect to Fig. 46. The code tree, generally indicated with reference number 3150 comprises a root node 3152 and several internal nodes 3154, the root node 3152 and the internal nodes 3154 being designated by Cj, with the index i being an integer associated with a corresponding node. Each of these nodes defines a conditional branch or binary decision called a "bin", with a left branch corresponding to the bin value 0 and the right branch corresponding to the bin value 1.
The tree 3150 also includes terminal nodes 3156. The numbers assigned to the terminal nodes 3156 in Fig. 44 correspond to the data value of the syntax element, which, in accordance with the binarization scheme of Fig. 46, corresponds to the codeword or bit string resulting form the bits or bin values when following the path from root node 3152 to the respective terminal node 3156. For example, the data value 17 corresponds, as shown in Fig. 46, to the codeword bit string 11111111111111101. The path comprises root node 3152 and the internal nodes 3154 C2 to C1 . Each node 3152 or 3154 may be seen as a binary decision called "bin". The concatenation of the binary decision for traversing the tree 3150 from the root node 3152 to the respective terminal node 3156 represents the "bin string" of the corresponding symbol value or syntax element value.
Each binary decision has an actual probability distribution indicating the probability as to whether the decision yields a binary zero (left branch in Fig. 44) and a binary one (right branch in Fig. 44). In other words, the actual probability distribution of a binary decision defines the probability that the syntax element path takes the zero or one branch at this binary decision. Accordingly, the actual probability for the syntax element to assume a certain data value is equal to the product of the respective probabilities at the nodes arranged along the path from root node 3152 to the respective terminal node 3156.
For binary entropy coding, each binary decision or node 3152, 3154 of tree 3150, i.e. each bin, is associated with a bit value probability estimation or a context model, i.e. Q. As is described in more detail below, not all of the nodes or bins is associated with adaptive probability estimation models but may be associated with a static probability estimation, thereby reducing the context modeling overhead, i.e. the management for adapting the estimation according to past bins in order to increase the compression ratio.
Next, with respect to Fig. 45, an embodiment of a binarization scheme in accordance with an embodiment of the present invention is described. Fig. 45 shows schematically the steps binarizer 3102 performs on the non-binary valued syntax elements. In particular, Fig. 45 is an embodiment for binarizing the absolute values of transform coefficient levels, i.e., an integer quantity representing the value associated with a particular two-dimensional frequency index in the decoding process prior to scaling for computation of a transform coefficient value.
The binarization scheme of Fig. 45 is a concatenation of a truncated unary binarization scheme, on the one hand, and a kth order Exp-Golomb binarization scheme, on the other hand. A truncated unary binarization scheme is based a unary binarization scheme. In accordance with an unary binarization scheme, for each unsigned integer valued symbol x greater than or equal to zero the unary codeword to which x is mapped consists of x "1" bits plus a terminating "0" bit. Differing from the unary binarization scheme, a truncated unary (TU) code is only defined for x with 0 < x < S, where for x < S the code is given by the unary code, whereas for x = S the terminating "0" bit is neglected such that the TU code of x = S is given by a codeword consisting of x "1" bits only, i.e. without a terminating bit of "0". In accordance with the exponential Golomb binarization scheme, generally, a data value x is mapped to a codeword consisting of a prefix and a suffix. Mere precisely, in accordance with the kth order Exp-Golomb binarization scheme, data values are mapped to code words having at least a prefix and, depending on the data value, a suffix. The prefix part of the EGk code word consists of a unary code corresponding to the value l(x), with l(x) = |_log2(x/2k + lj, k indicating the order of the code and Ld being maximum integer smaller than or equal to real value r. The EGk suffix part is computed as the binary representation of x + 2k(l-21(x)) using k + l(x) significant bits.
As can be seen from Fig. 45, the absolute data value x to be binarized is compared to the cutoff value S at 3200 to determine the minimum of x and S. S is the cut-off value defining for which absolute data values x the truncated unary code is to be used exclusively, i.e. without using the Golomb binarization scheme for the residual. In other words, for all absolute data values x less than the cut-off values S the truncated unary binarization is used exclusively in order to map the absolute data value x to a codeword, the codeword accordingly merely consisting of a primary prefix. If the absolute data value x is greater than or equal to S, a primary suffix is appended to the primary prefix, the primary suffix being created in accordance with a zero order Exp-Golomb binarization scheme from x-S, as will be described further below.
The minimum between x and S is binarized in accordance with the truncated unary binarization scheme at 3202. The result is a primary prefix 204, which forms at least a part of the final codeword 3206 to which x is mapped to. The primary prefix thus consists of a unary binarization of the value of Min(S, Abs(x)), where Min(a,b) denotes the minimum of values a and b, and where Abs(x) denotes the absolute value of x. In other words, primary prefix consists of S ones if x is greater than or equal to S, and consists of x leading ones followed by a terminating zero if x is less tha S.
The absolute data value x is compared at 3208 to the cut-off value S. If a comparison yields that x is smaller than S, the binarization ends at 3210 and, therefore, the codeword 3206, the absolute data value x is mapped to in accordance with the binarization scheme of Fig. 45, just consists of the primary prefix 3204. Otherwise, i.e., if x is equal to or greater than S, the cutoff value S is subtracted from the absolute data value x at 3212. The resulting difference x minus S, sometimes called a residual, is processed at 3214 in accordance with a kth order Exp- Golomb binarization scheme, with k equal to zero. The kth order Exp-Golomb binarization scheme will later be described in more detail with respect to Fig. 48.
The result of the binarization 3214 of the difference "x minus S" is appended to the primary prefix 3204 as the primary suffix 3208. Therefore, if the absolute data value x is greater than or equal to the cut-off value S the resulting codeword 206 consists of a primary prefix 3204 followed by a primary suffix 3208, wherein the primary suffix 3208 is a kth order Exp- Golomb binarization of x - S. As will be apparent from the further description, the primary suffix 3208 itself consists of a prefix and a suffix, hereinafter called a secondary prefix 3218 and a secondary suffix 3220.
Experimental studies have shown, that for absolute values of transform coefficient levels, S = 14 and k = 0 yield relatively good compression ratios when passing such binarized transform coefficient level values to a binary entropy coder even if just the bits of the codewords 3206 being part of the primary prefix are entropy coded using an adaptive probability model, thereby reducing the coding overhead.
Fig. 46 shows a table in which absolute values of transform coefficient levels in the left column 3215 are listed with their corresponding codewords in accordance with the binarization if Fig. 45 to the right at 3252, when S = 14 und k = 0. It is noted that the binarization and subsequent coding process is applied to the syntax element "transform coefficient level values minus 1", since zero valued, i.e. non-significant, transform coefficient levels are encoded using the aforementioned significances map. Therefore, with respect to Fig. 45, in the table of Fig. 46 at 3250 the values of x plus 1 is listed. The table of Fig. 46 shows the result of the binarization scheme of Fig. 45 for the values S = 14 and k = 0.
As can be seen, the codeword or bin string, an absolute value is mapped to, comprises at least a TU prefix, the TU prefix part of the codewords or bin strings being highlighted in Fig. 46 at 3254. As can be seen, absolute values below 15 (=x+l) do not have an EGO suffix as shown in 3256. Further, as shown by dotted line 3258 each EGO suffix 3256 comprises a secondary prefix which is on the left side of dotted line 3258, and all EGO suffixes 3256, except the one of codeword belonging to absolute value 15, comprise a secondary suffix arranged to the right side of dotted line 3258.
In the same way as Fig. 46, Fig. 47 shows a table of pairs of absolute values and bin strings as they are obtained in accordance with the binarization scheme of Fig. 45 with k = 3 and S = 9. The elements of the table of Fig. 47 being equal to those of Fig. 46 are indicated by the same reference numbers as in Fig. 46, and a repetitive description of these elements is omitted. The sign bit is not shown in table Fig. 47.
The values for k and S according to the binarization scheme in accordance with Fig. 47 has shown good compression effectiveness when applied to absolute components of motion vector differences. For motion vector differences UEGk binarization as shown in Fig. 47 is constructed as follows. Let us assume the value mvd of a motion vector component is given. For the prefix part of the UEGk bin string, a TU binarization (3202 in Fig. 45) is invoked using the absolute value of mvd with a cut-off value of S = 9. If mvd is equal to zero, the bin string consists only of the prefix code word "0", 270. If the condition | mvd\ > 9 holds (3208 in Fig. 45), the suffix is constructed (3214 in Fig. 45) as an EG3 codeword for the value of I //rai l -9, to which the sign of mvd is appended using the sign bit "1" (not shown) for a negative mvd and the sign bit "0" (not shown) otherwise. For mvd values with 0< | mvd | <9, the suffix consists only of the sign bit.
Assuming that the components of a motion vector difference represents the prediction error at quarter-sample accuracy, the prefix part always corresponds to a maximum error component of +/- 2 samples. With the choice of the Exp-Golomb parameter k = 3, the suffix code words are given such that a geometrical increase of the prediction error in units of 2 samples is captured by a linear increase in the corresponding suffix code word length. After having described the functionality of binarizer 3102 in Fig. 29 with respect to Figs. 45 to 47 in a rather coarse way, a possible implementation for the k* order Exp-Golomb binarization 3214 in Fig. 45 is described with respect to Fig. 48 in more detail. Fig. 48 shows a pseudo-C code illustrating the process of constructing the primary suffix part of the code word representing the binarization of absolute data value x in accordance with the binarization scheme of Fig. 45. First, at 3300 an auxiliary parameter k is initialised to the order of the Exp-Golomb binarization used. In case of binarization of components of motion vector differences in accordance with Fig. 47, for example, k is initialized in step 3300 to the value of 3.
If a comparison in line 3302 between the absolute value of the motion vector difference (mvd) and the cut-off value, here 9, reveals that the absolute value of mvd is greater than or equal to 9, in a step 3304, the cut-off value S=9 is subtracted from the absolute value of mvd to obtain sufS. Furthermore, a parameter stopLoop is initialized in step 3306 to the value of zero. Next, in 3308, a conditional if-operation is performed as long as the stopLoop-parameter has not changed his value to a positive value greater than 0, as is checked in line 310. If the conditional if-operation reveals that sufS is greater than or equal to the kth power of 2 (312d), at 3312 a bit having the bit value 1 is output (3312a), the kth power of 2 is subtracted from sufS (3312b), and k is incremented (3312c). Otherwise, if sufS is smaller than the kth power of 2 (3312e), a bit having the bit value of zero is output (3312f), a binary representation of sufS using' a number of bits equal to the actual value of k (3312g) is output (3312h), and the parameter stopLoop is set to a value of 1 (3312e) thereby resulting in stopping the iterative execution of the if-operation 3308.
All steps 3304 to 3312 are performed if the comparison in step 3302 reveals that the absolute value of mvd is greater than or equal to 9. Irrespective of the comparison in 3302, at 3314 an a sign bit of zero is output if mvd is greater than zero and an a sign bit of 1 is output if mvd is smaller than zero, the sign bit not being shown in Fig. 47.
As becomes clear from the pseudo-C code of Fig. 48, in case of third order Exp-Golomb binarization scheme, the suffix part at least comprises four bits, one being the zero bit of 3312f, and the other three being the bits of the binary representation output in 3312h, as can be seen also in Fig. 47, neglecting the sign bit at the end of each codeword.
After having described the binarization and entropy coding on the encoder side in accordance with embodiments of the present invention, Fig. 49 shows as an embodiment of the present invention, the steps performed on decoder side in order to decode the entropy coded bit stream. Firstly, in step 3400, the decoder receiving the entropy coded bit stream determines whether the next bin or bit to be determined from the coded bit stream is an primary prefix bit or an primary suffix bit. The decoder knows from the way of the predetermined syntax of the entropy coded bit stream, as to whether the next bin or bit expected is a bit of a primary prefix or a primary suffix of a binarized syntax element.
If, in step 3400, it is determined that the expected bit is a bit of the primary prefix, the decoder determines the bit value for this bit by binary entropy decoding based on an adaptively varying probability estimation or context model. That is, the decoder selects the respective partial bit stream described above, using the respective probability estimation for selection. The decoder updates this adaptively varying probability estimation based on past decoded bins indicated by the respective context model type.
If the bit is a bit of the primary suffix, in 3404 the next bit or bin value is determined by performing entropy decoding based on a static probability model. This may involve a 1:1 mapping of codewords to bins including the actual bit.
The decoder repeats, 3408, the steps 3400 to 3404 if there are bits left to achieve the bit values for the bits of a whole codeword. If there are no bits left at step 3408, the decoding process ends at step 3410. The decoder knows when the binary representation of a syntax element ends, i.e. which decoded bit is the last of a actual syntax element, from following the respective binarization code tree, such as the code tree shown in Fig. 44 in case of the syntax element being a absolute transform coefficient level minus 1.
Fig. 50 shows the process of recovering from the codeword as obtained by the process of Fig. 49 the data value, which has been binarized into the codeword. The bit values as obtained by the process of Fig. 49 form a bit stream 3450. The decoder knows, as described above, when a new code word is on the bit stream 3450, the instance of a new codeword being indicated by way of a signal 3452.
In a minimum extracting step 3454 then a value m is extracted from the binary code word by counting the leading ones in the primary prefix of the code word, the count being equal to m. If m is equal to S, what is checked in step 3456 (the value of S is known to the decoder because he knows which syntax element is coded by which binarization scheme), in a difference extracting step 3458 the bits following to the bits of the primary prefix from the bit stream 3450 are received, which form the primary suffix part of the code word. In the difference extracting step 3458 the difference between the data value x to be recovered and the cut-off value S is recovered from the primary suffix of the code word in the bit stream 3450. In step 3460, S is added to the difference x-S recovered in step 3358 to obtain the value x, the value for the actually encoded syntax element.
An example for a pseudo-C code illustrating minimum extraction step 454 could be: m = -l; /Initializing counting parameter
for (b=0;(b-l);m++) { /Bitwise reading the leading bits of
b=read_bits(l); /codeword and, before each reading,
if (m=S) /incrementing k, until the actually
b=0; /read bit is zero or m=S
}
The substeps performed by in difference extracting step 3458 are shown in Fig. 51. In a step 3500 the difference extractor counts the leading bits in the primary suffix to obtain a value m. Next, k, which is 3 for motion vector difference and is 0 for a transform coefficient level value, is compared to zero in step 3502. If k is equal to zero, in step 3504, m is compared to zero in step 3504. If comparison yield that m = 0, it is deduced in step 3506 that the data value x is equal to the cut-off value S, which is 9 in the case of motion vector differences and 14 in the case of transform coefficient levels.
If k is not equal to zero and/or m is not equal to zero, in step 3508 the next k plus m following bits are read from the bit stream 3450, i.e. the bits following the terminating bit of the secondary prefix. Thereby, a binary (k+m) bit long representation of A=x-S+2k (l-2m) is achieved. As can easily gathered form this equation, A=x-S if m=0.
When a comparison of m with 0 in step 3510 reveals that m = 0, from the binary representation A the value of 2k(l-2m) is subtracted to obtain the value of x-S.
EXAMPLE 6
In this example, example 4 is further extended and Fig. 34, 35, 36, 37a,b, 39, 45, 46 - 49 also apply. In the present example, a coding scheme for syntax elements contained in a pre-coded video signal is provided, which enables a higher compression effectiveness.
A method for encoding a syntax element contained in a precoded video signal into a coded bit stream is described, the precoded video signal representing at least one video frame, the syntax element being associated with a predetermined portion of the video frame and being indicative of as to whether the predetermined portion of the video frame is precoded in a first or a second way into the precoded video signal, the method comprising investigating as to whether a neighboring portion of the video frame neighboring the predetermined portion is precoded in the first way or the second way, in order to obtain a binary value; assigning one of at least two context models to the predetermined portion of the video frame based on the binary value, wherein each context model is associated with a different probability estimation; Finally, an encoding of a sequence of symbols take place in accordance with any of the above embodiments, with at least one symbol relating to a representation of the syntax element. The assigner is configured to determine the measure for the estimate of the probability distribution for the respective symbol based on the probability distribution estimate associated with the context assigned to the respective symbol. The selector selects the respective entropy coder accordingly.
In accordance with a second aspect of the present invention, this aspect is achieved by a method for decoding a syntax element from a coded bit stream, the syntax element being contained in a precoded video signal, the precoded video signal representing at least one video frame, the syntax element being associated with a predetermined portion of the video frame and being indicative of as to whether the predetermined portion of the video frame is precoded in a first or a second way into the precoded video signal, the method comprising investigating as to whether a neighboring portion of the video frame neighboring the predetermined portion is precoded in the first way or the second way, in order to obtain a binary value; assigning one of at least two context models to the predetermined portion of the video frame based on the binary value, wherein each context model is associated with a different probability estimation; and reconstructing a sequence of symbols among which at least one symbol relates to a representation of the syntax element in accordance with any of the above embodiment is performed. The assigner is configured to determine the measure for the estimate of the probability distribution for the respective symbol based on the probability distribution estimate associated with the context assigned to the respective symbol. The selector selects the respective entropy decoder accordingly.
The present example is based on the finding that the coding of a pre-coded video signal may be rendered more compressive, when a syntax element associated with a pre-determined portion of the video frame or picture and being indicative of as to whether the pre-determined portion of the video frame is pre-coded in a first or a second way, i.e., a control information syntax element, is assigned to one of at least two context models, the assignment being dependent on a neighbouring portion of the video frame being pre-coded in the first or second way. The syntax element may then be entropy encoded into the coded bit stream based on the probability estimation with which the one context model is associated. By this measure, the dependency of the actual or real probability distribution of the syntax element on the syntax element being of the same type but associated with neighbouring portions of the video frame is exploited and the actual probability distribution can be approximated more precisely.
Thus, it is an advantage of the present invention, that, due to the more precisely adjustable probability estimation, the compression ratio of the coded bit stream to the pre-coded video signal may be enhanced.
In other words, in accordance with an embodiment of the present invention, the syntax element of a specific type may be assigned to different ones of the at least two context models each time it occurs in the pre-coded video signal. The selection of the context model is performed each time the syntax element occurs and is dependent on as to whether the neighbouring portion of the video frame is pre-coded in the first or the second way. In effect, by this measure, all the syntax elements of this type occurring in the precoded video signal are grouped into a few context model groups, each syntax element belonging to one context model group, each context model group being assigned to a different one of the context models, and each context model being associated with a different probability estimation. Therefore, each group is entropy coded based on a different probability estimation, namely the one belonging to the context model it is assigned to.
In accordance with another embodiment of the present invention, the probability estimation of each context model is adapted to the previously coded syntax elements, the adaption being performed separately for each context model. In other words, each syntax element just having been entropy coded is used to adapt the probability estimation of the context model to which this syntax element is assigned. Hence, following syntax elements of the same type and assigned to the same context model are entropy coded by means of an adapted probability estimation. Since, by this measure, the context models are adapted or managed independently of each other, the adaptation yields a better approximation to the actual or real probability estimation and therefore results in a better compression ratio.
In accordance with a specific embodiment of the present invention, the spatially dependent context model selection is used for coding the frame/field coding mode flags in a pre-coded video signal. The frame/field coding mode flags indicate as to whether the respective video frame portion is coded in frame or field coding mode. To be more precisely, in accordance with this embodiment, the video frame is an interlaced frame where some of its picture samples are captured at one time instant while the other picture samples are captured at a different time instant, the first picture samples being called a first field and the second picture samples being called a second field. Moreover, the video frame is spatially subdivided into an array of macroblock pairs, each macroblock pairs consisting of two macroblocks. When coded in field mode, one macroblock of a macroblock pair merely comprises the picture samples of the first field or the picture samples of the second field within that macroblock pair, while the other macroblock of that macroblock pair contains or represents the picture samples of the other field within the macroblock pair. When coded in frame mode, the macroblock contains both, first and second field picture samples, in particular, the spatially top or bottom ones within a macroblock pair, while the other macroblock of the same macroblock pair represents the other picture samples within the macroblock pair.
As will be described in more detail with respect to the syntax element mb_field_decoding_fiag, when the same bin or same syntax element occurs several times in the bins passed from stage 2100a it may be assigned to different of the context models depending on previously incoming or previously entropy coded bins.
After having described the operation of coding arrangement 2100, in the following its functioning is described in more detail with respect to the handling of the syntax element mb_field_decoding_flag in accordance with embodiments of the present invention. In order to do so, firstly, with regard to Figs. 30 to 3 lb, the meaning of this flag is explained.
Now, the assignment depending on the mode by which the macroblock pair or the macroblocks of the macroblock pair are coded by precoder 2012, reference is again made to Fig. 28 in order to explain the function and meaning of the syntax element mb_field_decoding_fiag contained in the precoded video signal output by precoder 4012.
- When the precoder 4012 receives a video signal representing an interlaced video frame, precoder 4012 is free to make the following decisions when coding the video frame 4010:
- It can combine the two fields together to code them as one single coded frame, so that each macroblock pair and each macroblock would be coded in frame mode. - Alternatively, it could combine the two fields and code them as separate coded fields, so that each macroblock pair and each macroblock would be coded in field mode.
As a last option, it could combine the two fields together and compress them as a single frame, but when coding the frame it splits the macroblock pairs into either pairs of two field macroblocks or pairs of two frame macroblocks before coding them.
The choice between the three options can be made adaptively for each frame in a sequence. The choice between the first two options is referred to as picture adaptive frame/field (PAFF) coding. When a frame is coded as two fields, each field is partitioned into macroblocks and is coded i a manner very similar to a frame.
If a frame consists of mixed regions where some regions are moving and others are not, it is typically more efficient to code the non-moving regions in frame mode and the moving regions in the field mode. Therefore, the frames/field encoding decision can be made independently for each vertical pair of macroblocks in a frame. This is the third coding option of the above-listed options. This coding option is referred to as macroblock adaptive frame/field (MBAFF) coding. It is assumed in the following that precoder 4012 decides to use just this option. As described above, MBAFF coding allows the precoder to better adapt the coding mode type (filed or frame mode) to the respective areas of scenes. For example, precoder 4012 codes macroblock pairs located at stationary areas of a video scene in frame mode, while coding macroblock pairs lying in areas of a scene showing fast movements in field mode.
As mentioned above, for a macroblock pair that is coded in frame mode, each macroblock contains frame lines. For a macroblock pair that is coded in field mode, the top macroblock contains top field lines and the bottom macroblock contains bottom field lines. The frame/field decision for each macroblock pair is made at the macroblock pair level by precoder 4012, i.e. if the top macroblock is field coded same applies for the bottom macroblock within same macroblock pair. By this measure, the basic macroblock processing structure is kept intact, and motion compensation areas are permitted to be as large as the size of a macroblock.
Each macroblock of a field macroblock pair is processed very similarly to a macroblock within a field in PAFF coding. However, since a mixture of field and frame macroblock pairs may occur within an MBAFF frame, some stages of the pre-coding procedure in precoder 4012, such as the prediction of motion vectors, the prediction of intra prediction modes, intra frame sample prediction, deblocking filtering and context modelling in entropy coding and the zick-zack scanning of transform coefficients are modified when compared to the PAFF coding in order to account for this mixture.
To summarize, the pre-coded video signal output by precoder 4012 depends on the type of coding precoder 4012 has decided to use. In case of MBAFF coding, as it is assumed herein, the pre-coded video signal contains a flag mb_field_decoding_flag for each non-skipped macroblock pair. The flag mb_field_decoding_flag indicates for each macroblock pair it belongs to whether the corresponding macroblocks are coded in frame or field coding mode. On decoder side, this flag is necessary in order to correctly decode the precoded video signal. In case, the macroblocks of a macroblock pair are coded in frame mode, the flag mb_field_decoding_fiag is zero, whereas the flag is one in the other case.
Now, while the general mode of operation of the original decoder arrangement of Fig. 29 has been described without referring to a special bin, with respect to Fig. 52, the functionality of this arrangement is now described with respect to the binary entropy coding of the flag mb_field_decoding_fiag.
The process shown in Fig. 52 starts at the time, where the flag mb_field_decoding_flag arrives at the input of the context modeller 4112. In a first step 4300 context modeller 4112 determines as to whether the incoming syntax element or bin is the mb_field_decoding_flag. If not, context modeller 4112 switches to another syntax element handling procedure or subroutine at step 302. If the incoming bin is an mb_field_decoding_flag the context modeller 4112 determines in step 304 as to which context model the flag belongs to.
Step 4304 is shown in more detail in Fig. 53. When determining the context model corresponding to the current mb_field_decoding_flag, context modeller 4112 first computes, instep 4306, the addresses of the top macroblock of the macroblock pair to the left and above the current macroblock pair.
In the next steps 4308 and 4309, context modeller 4112 determines the availability of the top macroblock of the macroblock pair to the left and above the current macroblock pair, respectively. In step 4308, the neighbouring macroblock A to the left, is marked as available, when mbAddrA is smaller than 0. The same applies for step 4309 with respect to mbAddrB. Thus, MbAddrN, with N being either A or B, are the addresses of the top macroblocks of the respective neighbouring macroblock pair in case the macroblock pair does really exist and denotes concurrently the availability status of the respective neighbouring macroblock pair A and B, respectively. In an alternative embodiment, the macroblock A or B is only marked as available, when additionally to the above condition the following is true: The macroblock with address mbAddrN belongs to same slice as the current slice.
When in step 4308/4309 it is determined that macroblock N, with N being either A or B, is marked as available, in step 4310 and step 4311, respectively, it is determined as to whether macroblock N (N being equal to A or B) is a macroblock coded in field mode or a macroblock coded in frame mode. In case macroblock A is coded in field mode a flag condTermflagA is set to 1 in step 4312. The same applies for macroblock B so that condTermflagB is set to 1 in step 4313 if macroblock B is coded in field mode. In all other cases, i.e., when mbAddrA is smaller than 0 (4308) or macroblock A is coded in frame mode (4310), the flag condTermFlagA is set to 0 in step 4314. Accordingly, when mbAddrB is smaller than 0 (4309) or macroblock B is coded in frame mode (4311), the flag condTermFlagB is set to 0 in step 4315.
CondTermFlagA and condTermFlagB are used in step 4316 by context modeller 4112 to compute context index ctxIdxMbField indicating the probability model to be used for binary entropy coding of mb_field_decoding_flag. In particular, context modeller computes ctxIdxMbField = condTermFlagA + condTermFlagB, the result being a number of { 0, 1 ,2 } .
Referring back to Fig. 52, after the determination in step 4304, context modeller 4112 passes the variable ctxIdxMbField (context index macroblock field) or the probability estimation status indexed by CtxIdxMbField along with mb_field_decoding_flag to regular coding engine 4118 which, in turn, is contracted in accordance with any of the above described embodiments. That is, the assigner of the above embodiments stes the parameter for selcting the parial bitstream in accordance with the probability determined vie the context determined, in turn, in step 4304, whereupon the selector sends the bins of ctxIdxMbField to the respective partial bitstream path. In step 4318, the flag mb_field_is then encoded into the respective partial bitstream selcted in accordance with the current probability state of the context model as indexed by ctxIdxMbField.
Thereafter, regular coding engine 4118 passes the bin value of mb_field_decoding_flag via path 4117 back to context modeller 4112, whereupon context modeller 4112 adapts, in step 4320, the context model indexed by ctxIdxMbField with respect to its probability estimation state. Thereafter the process of coding the flag mb_field_decoding_flag into the bit stream at the output 4124 ends at 4322. Fig. 54 shows the decoding process performed by the entropy decoder 4400 in more detail. First, in a step 4500, the decoder 4400 checks as to whether the next bin to decode is an mb_field_decoding_flag. If this is not the case, decoder 400 switches to another bin handling process (not shown) in step 4502. If yes, decoder 4400 determines in step 4504 the context model to which the flag belongs. Step 4504 is performed in the same way as step 4304 in the encoding process, or as shown in Fig. 53.
Then, in step 4506, the actual bin, i.e., mb_field_decoding_flag, is obtained from the respective partial bitstream entropy decoder by indexing unsing the actual probability state of the context model as indexed by ctxIdxMbField obtained in step 4504. The result of this step is the value for mb_field_decoding_flag. Thereafter, in step 4508, the ctxIdxMbField probability state is adapted or updated, as it was the case in step 4320. Thereafter, the process ends at step 4510.
The determination in step 4504 yields the same index ctxIdxMbField since the macroblocks are stepped through in the precoded video signal in an appropriate order that guarantees that the flags mb_field_decoding_flag of macroblocks A and B (Fig. 22) have already been previously precoded. Thus, the context model can be deduced in step 4504 on the basis of previously decoded flags mb_field_decoding_flag of neighbouring macroblocks as was the case on encoder side.
It is noted that the above example is not restricted to a flag indicating frame/field mode coding of a macroblock or macroblock pair. Rather, the above description could be applied to other syntax elements as well.
With respect to step 4310, the following is noted. Normally, a macroblock is coded into the precoded video signal. Nevertheless, it might be that a macroblock is not coded or skipped because, for instance, the corresponding transform coefficient levels of this macroblock are all smaller than a certain quantization threshold. Thus, for some macroblocks there is no mb_field_decoding_flag available. Thus, it might be that step 4310 cannot be performed by merely investigating the mb_field_decoding_flag of the respective neighbouring macroblock if the latter is a skipped one. When mb_field_decoding_flag is not present for either macroblock of a neighbouring macroblock pair (containing A or B in Fig. 22) (when the bottom macroblock has a mb_field_decoding_flag, this one is chosen to be mb_field_decoding_flag of A or B), the following applies (it is assumed that the macroblock pair containing A is missing): • if there is a neighbouring macroblock pair to the left (of A) in the same slice, the value of mb_field_decoding_flag (of A) should be inferred to be equal to the value of mb_field_decoding_flag for the neighbouring macroblock pair to the left of the current macroblock pair (containing A),
• if there is no neighbouring macroblock pair to the left (of A) in the same slice, and there is a neighbouring macroblock pair above (of A) in the same slice, the value of mb_field_decoding_flag (of A) shall be inferred to be equal to the value of mb_field_decoding_flag of the neighbouring macroblock pair above the current macroblock pair (containing A),
• otherwise (if there is no neighbouring macroblock pair either on the left or above the current macroblock pair (containing A), the value of mb_field_decoding_flag (of A) shall be inferred to be equal to zero.
Of course, the above steps also apply for top macroblock B, by merely switching A to B.
Furthermore, other neighbouring relationships than those described with respect to Figs. 22 and 34 might be used in order to determine the context model for entropy encoding the mb_field_decoding_flag.
Furthermore, as already noted, the above embodiments are also applicable to other control syntax elements or to other bins. Furthermore, the present invention is not restricted to binary entropy coding but could applied as well to other symbol entropy coding. Additionally, the subdivisions of the video frame into slices, macroblock pairs, macroblocks, picture elements etc. was for illustrating proposes only, and is not to restrict the scope of the invention to this special case.
EXAMPLE 7
In this example, example 4 is further extended and Fig. 34, 35, 36, 37a,b, 39, 45, 46 - 49 also apply. In accordance with example 7, a method for encoding a video frame or picture comprises dividing up the video frame or the picture in portions of a first type and portions of a second type, wherein the portions of the first type are associated with a first scanning pattern, and the portions of the second type are associated with a second scanning pattern that is different from the first scanning pattern; transforming data corresponding to a predetermined of the portions of the video frame or picture into a two-dimensional array of transform coefficients, wherein a scanning order is defined among the transform coefficients by the scanning pattern of the predetermined portion, the scanning order assigning each transform coefficient a unique scanning position; precoding a predetermined of the transform coefficients in order to obtain a transform data unit; choosing one of a first and a second set of context models, depending on as to whether the predetermined portion is a portion of a type being associated with the first or the second scanning pattern, the first and the second set being different to each other; assigning one context model of the chosen one of the first and the second set of context models to the transform data unit based on the scanning position assigned to the predetermined transform coefficient, wherein each context model is associated with a different probability estimation; and encoding the transform data unit or a sub-unit thereof into a coded bit stream based on the probability estimation with which the assigned context model is associated using any of the above embodiments.
According to example 7, a decoding a transform data unit or a sub-unit thereof from a coded bit stream is described, the transform data unit being a precoded version of a predetermined transform coefficient of transform coefficients which are the result of a transformation of data corresponding to a predetermined portion of portions of a video frame or picture, the portions being either a portion of a first type being associated with a first scanning pattern or a portion of a second type being associated with a second scanning pattern, the method comprising choosing one of a first and a second set of context models, depending on as to whether the predetermined portion is a portion of a type being associated with the first or the second scanning pattern, the first and the second set being different to each other; assigning one context model of the chosen one of the first and the second set of context models to the transform data unit or the sub-unit thereof based on the scanning position assigned to the predetermined transform coefficient, wherein each context model is associated with a different probability estimation; and decoding the transform data unit or the sub-unit thereof from the coded bit stream based on the probability estimation with which the assigned context model is associated using any of the above embodiments.
The finding is exploited that the transform data units or sub-units thereof obtained from a video frame or picture may be coded in a more compressive way, when different sets of context models are used for coding transform data units or sub-units thereof related to portions of the video frame or picture associated with a first scanning pattern as compared to transform data units or sub-units thereof related to portions of the video frame or picture associated with a different scanning pattern. By this measure, the context models can be adapted more precisely to the actual symbol statistics, which may be different between transform data units belonging to the same scanning position due to the different scannmg patterns. Since the actual probability distribution can be approximated more precisely, the compression ratio of the coded bit stream may be enhanced. Furthermore, since different context models are used for transform data units related to picture portions associated with different scanning patterns, the transform data units contributing to the update of a probability estimation of that context model show a more uniform actual probability distribution so that the updated probability estimation does more precisely approximate the actual probability distribution. Similarly, providing different context model sets enables initialising the probability estimation of the context models more precisely to the actual probability distribution. Thus, compression ratio of the coded bit stream may further be enhanced.
After having described the macroblock adaptive frame/field coding, the general handling of frame and field macroblocks during the coding scheme in accordance with an embodiment of the present invention is described with respect to Fig. 55. Fig. 55 shows the steps of encoding frame and field macroblocks as they are performed by precoder 2012 and entropy coder 2014 of Fig. 28.
Fig. 55 shows the process of encoding the video frame or picture 5010 which has already been divided up into macroblock pairs 5010M and 5010b2 as described with respect to Fig. 34. In Fig. 55 exemplarily merely two macroblock pairs 5010M and 5010b2 are shown. It is assumed that macroblock pair 5010M is field coded while macroblock pair 5010b2 is frame coded. Pixel samples belonging to the macroblock pairs 501 Obi and 5010b2 are passed to precoder 5012 where they are processed in step 5270. Step 5270 comprises subtracting from the picture samples of the macroblock pairs 501 Obi and 5010b2 predicted pixel samples and transformation the differences from a spatial domain into a spectrum domain. The results of step 5220 are several two-dimensional arrays of transform coefficients for each macroblock pair 5010M and 5010b2, each two-dimensional array of transformation coefficients belonging to a sub-part of a macroblock of the macroblock pairs 501 Obi and 5010b2, such as a sub-part of 4x4 pixel samples or pixel samples differences. Fig. 55 exemplarily shows one such two- dimensional array of transform coefficients as derived from macroblock 501 Obi indicated 5272 and another two-dimensional array of transform coefficients 5274 as derived from macroblock 5010b2. For example, the arrays 5272 and 5274 are 4x4 arrays. They may be obtained by a DCT applied to a 4x4 block of prediction residuals on one of the macroblocks of the macroblock pairs 501 Obi and 5010b2. For example, the macroblocks are 16x16 samples big. In this case, 16 4x4 blocks of predictional residuals would be transformed into 16 4x4 transformation coefficient arrays for each macroblock. The DCT may be a 4x4 discrete cosine transform or a separable integer transform with similar properties. Inverse- transform mismatches on decoder side may be avoided by defining the transformation matrix such that the inverse transform is defined by exact integer operations. Now, in the following it is assumed that the 4x4 block of pixel samples resulting in the array 5272 of transform coefficients corresponds to the block of pixels 1 to 4, 21 to 24, 41 to 44 and 61 to 64 in the macroblock 2252 as shown in Fig. 31b, and that the 4x4 block of pixel samples which array 5274 is related to comprises the pixel samples 1 to 4, 11 to 14, 21 to 24 and 31 to 34 of macroblock 2250 as shown in Fig. 31a.
As can be seen, the spatial relationship between the pixels in the 4x4 blocks to which the transformation of step 5270 is applied is different in field macroblocks and frame macroblocks. In particular, the pixel pitch along the columns is doubled in case of frame macroblocks compared to field macroblocks. Therefore, two transform coefficients assuming the same array position in arrays 5272 and 5274 relate to different frequencies. Due to this, as mentioned above, different scanning patterns are used in order to define a scanning order among the transform coefficients in arrays 5272 and 5274. The reason for this is that transform coefficients of lower frequencies are more likely to be zero. Thus, different scanning orders are used in order to 'sort' the transform coefficients in accordance with their frequency they relate to.
Fig. 56 shows an example for the scanning patterns used for frame and field macroblocks. At 5274, a two-dimensional array of transform coefficients is obtained from a frame macroblock as illustrated. At each array position a number from 0 to 15 is written, the number indicating the scanning position of the respective transform coefficient. The arrows 5276 represent the scanning pattern. As can be seen, the scanning pattern 5276 used to define a scanning order among the transform coefficients related to a frame coded macroblock are scanned in a zigzag scanning starting at one corner of the array 5274 and running diagonally to the corner arranged diametrically opposite to the starting corner. It is assumed in Fig. 56 that the transform coefficient in the upper left corner of array 5274 having scanning position 0 represents the DC component in column and row direction while the transform coefficient having scanning position 5015 arranged diametrically opposite thereto represents the highest frequency part in column and row direction.
At the right side of Fig. 56, the scanning pattern for transform coefficient arrays 5272 related to field macroblocks as illustrated. Again, each array position of the 4x4 transformation coefficient array 5272 is designated by its scanning position number. The scanning pattern in case of field macroblocks is illustrated by arrows 5278. As can be seen, the scanning pattern 5278 for field coded macroblocks is different to the scanning pattern 5276 for frame coded macroblocks. Both scanning patterns 5276 and 5278 define a scanning order among the transform coefficients 0 to 15. Referring back to Fig. 55, precoder 5012 proceeds in step 5280 with precoding the transform coefficients in arrays 5272 and 5274 into sequences 5282 and 5284 of transform data units. The step is described in more detail with respect to Fig. 57, showing a table having four rows. In the first row, the scanning positions from 0 to 8 are listed. In a second row, an example for possible transformation coefficient levels is given for the scanning positions 0 to 8. It is assumed that the transform coefficient levels at scanning positions greater than 8 are all 0.
Now, when performing the precoding of 5280 precoder 5012 steps through the transform coefficients in scanning order beginning at scanning positions 0, and performs the following steps on each transformation coefficient:
1) If the transform coefficient is not 0, i.e., the transform coefficient is significant, precoder 5012 sets.significant_coeff_flag for this scanning position to 1. Otherwise, precoder 5012 sets significant_coeff_fiag to zero. As can be seen, the first transform coefficient level at scanning position 0 is 9 so that, as can be seen in the third row in Fig. 57, significant_coeff_flag for this scanning position is 1.
2. For significant transform coefficients, it is proved as to whether this transform coefficient is the last significant transform coefficient among the transform coefficients in scanning order. If this is the case, a last_significant_coeff flag for the current scanning position created and set to 1. Otherwise, last_significant_coeff_flag for the current scanning position is set to 0. As can be seen in the last row of Fig. 57, in the present example, the last transform coefficient level being significant is the transform coefficient at scanning position 8. Therefore, the last_significant_coeff_flag for this transform coefficient is set to 1, whereas the last_significant_coeff_flag for the other significant transform coefficient at scanning position 0, 2, 3, and 6 is set to 0.
3. For all significant transform coefficient, precoder 5012 creates syntax elements that specify these transform coefficients.
To summarise, precoder 5012 creates for each transform coefficient from the first scanning position to the scanning position of the last significant transform coefficient a transform data unit composed of either a significant_coeff_flag, if the respective transform coefficient is not significant, and of a significant_coeff_flag, last_significant_coeff_flag and data specifying the value of the respective transform coefficient in case of the transform coefficient being significant. All syntax elements significant_coeff_flag and last_significant_coeff_flag such obtained form the significance map indicating the scanning positions of transform coefficients being significant. Precoder 5012 incorporates sequences 5282 and 5284 into a linear stream of syntax elements, i.e. the precoded video signal 5286. The order of the syntax elements is as follows:
In general, first, the syntax elements corresponding to the significance map are transmitted followed by the values indicating the significant transform coefficients in reverse scanning order. In particular, the first syntax element of sequences 5282 and 5284 is the significant_coeff_flag of scanning position 0. If the transform coefficient of the scanning position 0 is significant, the significant_coeff_flag of scanning position 0 is followed by a last_significant_coeff_flag for that scanning position. Thereafter, the significant_coeff_flag of the following scanning position, i.e. scanning position 1, is transmitted and so on. After having transmitted the last_significant_coeff_flag of the last scanning position having a nonzero transform coefficient, the non-binary values for the non-zero transform coefficients are transmitted in reverse scanning order. In the example of Fig. 57, the syntax elements in syntax element stream 286 are (beginning with the first syntax elements): 1, 0, 0, 1, 0, 1, 0, 0, 0, 1, 0, 0, 1, 1, 1, -1, 3, -5, 9.
The three more syntax elements 5286 is fed to entropy coder 5014 which handles the syntax elements element wise. The following description is now concentrated on the handling of significant_coeff_flag and last_significant_coeff_flag. When such a flag arrives at entropy coder 5014, it chooses at step 5288 one of two context model sets 5290 and 5292 based on the scanning pattern that was used in order to map the two-dimensional transform coefficient matrix into the sequence of transform data units, which the current syntax element last_signi- ficant_coeff_flag or significant_coeff_flag belongs to. In order to perform step 5288, entropy coder 5014 inspects the aforementioned mb_field_decoding_flag of the current macroblock, which the current syntax element belongs to, this flag indicating as to whether the current macroblock is coded in field mode in which case the scanning pattern 5278 has been used, or the current macroblock has been coded in frame mode in which case scanning pattern 5276 has been used.
In the next step, entropy coder 5014 assigns one context model of the chosen one of the context models sets 5290 and 5292 to the current syntax element to be encoded based on its scanning position. By this measure, the syntax element significant_coeff_flag for instance belonging to a scanning position 2 in sequence 5282 results in a context model of context model set 5290 that is different to the context model assigned to the significant coeff flag at scanning position 2 in sequence 5284. This accounts for the different probability distributions for the syntax element significant coeff flag at these scanning positions due to the different scanning pattern used to define the scanning positions in sequences 5282 and 5284. Finally, in step 5296, the current syntax element is binary entropy encoded by use of the assigned context model from the chosen context model set. This is done as described in the above embodiments, i.e. be selecting the respective partial bitstream by using the probability estimate of the chosen context.
In the following, with respect to Fig. 58, step 5288, 5294 and 5296 performed by entropy coder 5014 from Fig. 55 are explained in more detail.
The process shown in Fig. 58 starts at the time, where a syntax element arrives at the input of context modeller 5112. In a first step 5300, context modeller 5112 determines as to whether the incoming syntax element or bin is a last_significant_coeff_flag. If not, context modeller 5112 determines as to whether the incoming syntax element or bin is a significant_coeff_flag at step 5302. If this is not the case, context modeller 5112 switches to another syntax element handling procedure or subroutine at step 5304. If the current syntax element is a last_significant_coeff_flag, it is checked as to whether this flag belongs to either a frame coded macroblock or field coded macroblock in step 5306. Step 5306 is equal to checking as to whether the scanning position of the transform coefficient to which the last_significant_coeff_flag belongs is determined by scanning pattern 5276 or scanning pattern 5278. The same check is performed in step 5308 if, in step 5302, it is determined that the syntax element is a significant_coeff_flag. Steps 5300, 5302, 5306, and 5308 restrict the possible scenarios at the arrival of a syntax element to 5 possible cases. The first case has already been mentioned and exists when the current syntax element is neither a last_significant_coeff_flag nor a significant_coeff_flag. In this case, another syntax element handling is performed in step 5304. In the second case, the syntax element is a last_significant_coeff_flag belonging to a field code macroblock in which case, a context index offset value is set to a value offset 1 in step 5310. In a third case the current syntax element is a last_significant_coeff_flag belonging to a frame coded macroblock, where an index ctxldx Offset is set to offset2 in step5 312. Similarly, in a forth and fifth case, the syntax element is a significant_coeff_flag belonging to a field coded or frame coded macroblock, where index ctxldx Offset is set to offset3 in step 5340 and to offset4 in step 5316, respectively. The index ctxldx Offset is for pointing into on common list of context models. The values offset 1 to offset4 are different to each other.
After steps 5310 to 5316, an index incrementor ctxldxlnc (context index incrementor) is set equal to the scanning position (scanningPos) which the syntax element, i.e., the last_significant_coeff_flag or significant_coeff_flag, belongs to, in step 5318. Afterwards, in step 5320, a context index is determined by the sum of the context index offset ctxldx Offset and the index incrementor ctxldx Inc. The steps 5300 to 5320 are performed by context modeller 5112. The result of step 5320 is a context index ctxidx indicating the probability model to be used for binary entropy coding of the syntax element.
After the determination of ctxidx, context modeller 5112 passes the variable ctxidx or the probability estimation status indexed by ctxidx along with the syntax element itself, i.e. with last_significant_coeff_flag or significant_coeff_flag, to regular coding engine 118 which, sa outlined above, is confufgured in accordance with any of the above embodiments. Based on these inputs, the regular coding engine 5118 entropy encodes, in step 5322, the syntax element into the bit stream 5124 by using the current probability state of the context model as indexed by ctxidx.
Thereafter, regular coding engine 5118 passes the bin value of last_significant_coeff_flag or significant_coeff_flag, respectively, via path 117 back to context modeller 5112, whereupon context modeller 5112 adapts, in step 5324, the context model indexed by ctxidx with respect to its probability estimation state. Thereafter, the process of coding the last_significant_coeff_flag or significant_coeff_flag into the bit stream at the output 5124 ends at 5326.
As may be clear from the above, steps 5300 to 5316 belong to step 5288 of Fig. 55, while steps 5318 and 5320 belong to step 5294 and steps 5322 and 5324 belong to step 5296.
Before explaining in more detail the actual decoding process, the parsing process is illustrated with respect to Fig. 59 which shows a pseudo-C code for the parsing process performed in entropy decoder 5400. The parsing process shown in Fig. 59 is appropriate for parsing a significance map along with its corresponding coefficient values for a macroblock provided that the significance map and the coefficient values are coded in the way explained above with respect to Fig. 57, i.e., the significance map following the scanning order and preceding the coefficient values of the significant transform coefficients which are ordered in reverse scanning order.
In Fig. 59, variables that are written in bold are read from the entropy coded bit stream arriving at entropy decoder 5400. For reading, binary entropy decoding is used. The reading process is shown in more detail for bins sigmficant_coeff_flag and a last_significant_coeff_flag in Figs. 60. The parsing process is called residual_block_CABAC as can be seen in line 1 of the pseudo-C code of Fig. 59. The output of the process residual block CABAC is an one-dimensional array coeffLevel indicating the transform coefficient levels for each scanning position like shown in Fig. 57. As an input, the process residual block CABAC receives the variable maxNumCoeff which indicates the maximum number of coefficients in a macroblock, in our case being 16. The process then starts in line 2 with reading a flag called coded_block_flag indicating as to whether in the current block all transform coefficient levels are zero. In this case, coded_block_flag is zero otherwise the coded_block_flag is one. In the following it is assumed that coded_block_flag is one so that at least one transform coefficient level is not zero. So, in line 4, a variable NumCoeff is set to maxNumCoeff. This means that when in the following when-do-loop no last_significant_coeff_flag being equal to one is read, the scanning position of the last significant transform coefficient is equal to the maximum number of coefficients maxNumCoeff. In line 5, a counter i is initialised to zero. In a following while-do-loop starting at line 6, firstly, one bin is read and inputted into significant_coeff_flag. With regard to Fig. 59, significant_coeff_flag is a one-dimensional array. The bin read in line 7 is put into position i of array significant_coeff_flag, i.e., significant_coeff_flag[i]. If significant_coeff_flag[i] is one (line 8), i.e., the coefficient at scanning position i is significant, the next bin is read from the bit stream, i.e., last_significant_coeff_fiag[i]. If last_significant_coeff_flag[i] is one (line 10), i.e., the transform coefficient at scanning position i is the last significant transform coefficient in scanning order, the variable NumCoeff is set to i+1 in line 11 and coeffLevel also being an one-dimensional array is set to zero at all positions greater than or equal to i+1 (lines 12 and 13). This means that beyond the scanning position where last_significant_coeff_flag is one, the transform coefficient of all following scanning positions are zero.
After both if-clauses (lines 14 and 15) the counter i is incremented in line 16. At the end, at line 17, it is proved as to whether counter i is greater than or equal to NumCoeff- 1. If this is the case, the while-do-loop is repeated once again. Thus, in lines 5 to 17, the significance map is read from the bit stream. Following, at line 18, the first significant transform coefficient level minus 1 is read, i.e., coeff_ups_level_minusl, followed by the sign flag coeff_sign_flag. To be more specific, both syntax elements coeff_abs_level_ups_minusl and coeff_sign_flag read in lines 18 and 19 concern the last significant transform coefficient in scanning order. The actual value of this transform coefficient is computed in line 20. Then, in the for-loop from lines 21 to 28, the remaining significant transform coefficients are read from the bit stream by use of the significance map as received in lines 4 to 17. The counter of the for-loop i is initialised to NumCoeff-2, i.e., the scanning position immediately preceding the scanning position of the last significant transform coefficient. This counter is decremented after each for-loop-transition. If significant_coeff_flag[i] is one (line 22) the coefficient level of the transform coefficient at scanning position i is determined in lines 23 to 25, these lines being in effect, identical to lines 18 to 20. If significant_coeff_flag[i] is zero (line 26), the coefficient level of the scanning position is set to zero in line 27. Fig. 60 now shows the decoding process performed by the entropy decoder 400 each time a bin is to be decoded. Which bin is to be decoded depends on the syntax element which is currently expected by entropy decoder 5400. With regard to the significance map and the corresponding non-zero transform coefficient levels this knowledge results from the parsing process of Fig. 59.
In the decoding process, first, in step 5500, the decoder 5400 checks as to whether the next bin to decode is a last_significant_coeff_fiag. If this is not the case, the checks in step 5502 as to whether the next bin to decode is a significant_coeff_flag. The next bin to decode is a significant_coeff_flag. If the next bin to decode is either a last_significant_coeff_flag or a significant_coeff_flag, decoder 5400 checks as to whether the flag belongs to a frame coded or field coded macroblock in step 5504 or 5506, respectively. If the next bin to decode is none of a last_significant_coeff_flag and significant_coeff_flag, decoder 400 proceeds to another syntax element handling in step 504. The checks of steps 5500, 5502, 5506, and 5508 correspond to steps 5300, 5302, 5306, and 5308 of the encoding process of Fig. 58, and accordingly, depending on the next bin to decode being a last_significant_coeff_flag or a significant_coeff_flag and the corresponding macroblock being frame coded or field coded, ctxldx Offset is set to 1 of offsetl, offset2, offset3, and offset4 in one of steps 5510, 5512, 5514, and 5516, respectively, with steps 5510-5516 corresponding to steps 5310 to 3516 of Fig. 58. Accordingly, ctxldxlnc and ctxldx is determined in steps 5518 and 5520 in the same way as in the encoding process of Fig. 8 in steps 318 and 320 in order to determine the context model to be used in the following entropy decoding.
Then, in step 5522, the entropy decoder 5400 entropy decodes the actual bin, i.e., last_significant_coeff_flag or significant_coeff_flag, from the entropy coded bit stream by use of the actual probability state of the context module as indexed by ctxldx obtained in steps 5510 to 5520. That is, the respective partial bitstream is selected based on the probability index and the bin retrieved therefrom. The result of this step is the value for the actual bin. Thereafter, in step 5524, the ctxldx probability state is adapted or updated, as it was the case in step 5224. Thereafter, the process ends at step 5526.
Although in the above decoding process, step 5527, i.e. the decoding of mb_filed_decoding_flag, seems to have take place after the decoding of each bin, this is not the case. In fact, this step occurs, at maximum, merely for one macroblock pair.
After having described embodiments according to which the even previously coding is used for compressing video data, reference is made to Fig. 61a,b and 62 in order to describe an even further embodiment for implementing embodiments of the present invention which renders the implmentation especially effective in terms of a good trade-off between compression rate on the one hand and look-up table and compatation overhead on the other hand. In particular, the foil wing embodiments enable the use of computationally less complex variable length codes in order to entropy-code the individually bitstreams, and effectively cover portions of the probability estimate. In the embodiments described below, the symbols are of binary nature and the VLC codes presented below effectively cover the probability estimate representated by, for example, RLPS, extending within [0;0.5].
In particular, the embodiments outlined below describe possible implementations for the indivdual entropy coders 10 and decoders 22 in Fig. 1 to 11, respectively. They are suitable for coding of bins, i.e. binary symbols, as they occur in image or video compression applications. Accordingly, these embodiments are also applicable to image or video coding where such binary symbols are splitted-up into the one or more streams of bins 7 to be encoded and bitstreams 24 to be decoded, respectively, where each such bin stream can be considered as a realization of a Bernoulli process. The embodiments described below use one or more of the below-explained various so-called variable-to-variable-codes (v2v-codes) to encode the bin streams. A v2v-code can be considered as two prefix-free codes with the same number of code words. A primary, and a secondary prefix-free code. Each code word of the primary prefix-free code is associated with one code word of the secondary prefix-free code. In accordance with the below-outlined embodiments, at least some of the encoders 10 and decoders 22, operate as follows: To encode a particular sequence of bins 7, whenever a code word of the primary prefix-free code is read from buffer 8, the corresponding code-word of the secondary prefix-free code is written to the bit stream 12. The same procedure is used to decode such a bit stream 24, but with primary and secondary prefix-free code interchanged. That is, to decode a bitstream 24, whenever a code word of the secondary prefix-free code is read from the respective bit stream 24, the corresponding code-word of the primary prefix- free code is written to buffer 20.
Advantageously, the codes described below do not necessitate look-up tables. The codes are implementable in form of finite state machines. The v2v-codes presented here, can be generated by simple construction rules such that there is no need to store large tables for the code words. Instead, a simple algorithm can be used to carry out encoding or decoding. Three construction rules are described below where two of them can be parameterized. They cover different or even disjoint portions of the afore-mentioned probability interval and are, accordingly, specifically advantegeuous if used together, such as all three codes in parallel (each for different ones of the en/decoders 11 and 22), or two of them. With the construction rules described below, it is possible to design a set of v2v-codes, such that for Bernoulli processes with arbitrary probability p, one of the codes performs well in terms of excess code length. As stated above, the encoding and decoding of the streams 12 and 24 respectively, can either be performed independently for each stream or in an interleaved manner. This, however, is not specific to the presented classes of v2v-codes and therefore, only the encoding and decoding of a particular codeword is described for each of the three construction rules in the following. However, it is emphasized, that all of the above embodiments concerning the interleaving solutions are also combinable with the presently described codes or en- and decoders 10 and 22, respectively.
Construction rule 1: 'Unary bin pipe' codes or en-/decoders 10 and 22
Unary bin pipe codes (PIPE = probability interval partitioning entropy) are a special version of the so-called 'bin pipe' codes, i.e. codes suitable for coding of any of the individual bitstreams 12 and 24, each transferring data of a binary symbol statistics belonging to a certain probability sub-interval of the afore-mentioned probability range [0;0.5]. The construction of bin pipe codes is described first. A bin pipe code can be constructed from any prefix-free code with at least three code words. To form a v2v-code, it uses the prefix-free code as primary and secondary code, but with two code words of the secondary prefix-free code interchanged. This means that except for two code words, the bins are written to the bit stream unchanged. With this technique, only one prefix-free code needs to be stored along with the information, which two code words are interchanged and thus, memory consumption is reduced. Note, that it only makes sense to interchange code words of different length since otherwise, the bit stream would have the same length as the bin stream (neglecting effects that can occur at the end of the bin stream).
Due to this construction rule, an outstanding property of the bin pipe codes is, that if primary and secondary prefix-free code are interchanged (while the mapping of the code words is retained), the resulting v2v-code is identical to the original v2v-code. Therefore, the encoding algorithm and decoding algorithm are identical for bin-pipe codes.
A unary bin pipe code is constructed from a special prefix-free code. This special prefix-free code is constructed as follows/First, a prefix-free code consisting of n unary code words is generated starting with 'Ο , '00 Γ, Ό00Γ, ... until n code words are produced, n is the parameter for the unary bin pipe code. From the longest code word, the trailing 1 is removed. This corresponds to a truncated unary code (but without the code word Ό'). Then, n - 1 unary code words are generated starting with ΊΟ', Ί ΙΟ', Ί Ι ΙΟ', ... until n - 1 code words are produced. From the longest of these code words, the trailing 0 is removed. The union set of these two prefix-free codes are used as input to generate the unary bin pipe code. The two code words that are interchanged are the one only consisting of 0s and the one only consisting of Is. Example for n = 4:
Nr Primary Secondary
1 0000 111
2 0001 0001
3 001 001
4 01 01
5 10 10
6 110 110
7 111 0000
Construction rule 2: 'Unary to rice' codes and Unary to rice en-/decoders 10 and 22:
Unary to rice codes use a truncated unary code as primary code. I.e. unary code words are generated starting with T, '01 ', ΌΟ Γ, ... until 2n + 1 code words are generated and from the longest code word, the trailing 1 is removed, n is the parameter of the unary to rice code. The secondary prefix-free code is constructed from the code words of the primary prefix-free code as follows. To the primary code word only consisting of 0s, the code word T is assigned. All other code words consist of the concatenation of the code word '0' with the n-bit binary representation of the number of 0s of the corresponding code word of the primary prefix-free code.
Example for n = 3:
Nr Primary Secondary
1 1 0000
2 01 0001
3 001 0010
4 0001 0011
5 00001 0100
6 000001 0101
7 0000001 0110
8 00000001 0111
9 00000000 1
Note, that this is identical to mapping an infinite unary code to a rice code with rice parameter 2n.
Construction rule 3: 'Three bin' code
The three bin code is given as:
Nr Primary Secondary
1 000 0
2 001 100 3 010 101
4 100 110
5 110 11100
6 101 11101
7 on 11110
8 111 11111
It has the property, that the primary code (symbol sequences) is of fixed length (always three bins) and the code words are sorted by ascending numbers of Is.
An efficient implementation of three bin code is described next. An encoder and decoder for the three bin code can be implemented without storing tables in the following way.
In the encoder (any of 10), three bins are read from the bin stream (i.e.7). If these three bins contain exactly one 1, the code word T is written to the bit stream followed by two bins consisting of the binary representation of the position of the 1 (starting from right with 00). If the three bins contain exactly one 0, the code word Ί 1 Γ is written to the bit stream followed by two bins consisting of the binary representation of the position of the 0 (starting from the right with 00). The remaining code words '000' and '111' are mapped to '0' and Ί 111 , respectively.
In the decoder (any of 22), one bin or bit is read from the respective bitstream 24. If it equals Ό', the code word '000' is decoded to the bin stream 21. If it equals Ί', two more bins are read from the bit stream 24. If these two bits do not equal '11', they are interpreted as the binary representation of a number and two 0s and one 1 is decoded to the bit stream such that the position of the 1 is determined by the number. If the two bits equal Ί Γ, two more bits are read and interpreted as binary representation of a number. If this number is smaller than 3, two Is and one 0 are decoded and the number determines the position of the 0. If it equals 3, Ί 11 ' is decoded to the bin stream.
An efficient implementation of unary bin pipe codes is described next. An encoder and decoder for unary bin pipe codes can be efficiently implemented by using a counter. Due to the structure of bin pipe codes, encoding and decoding of bin pipe codes is easy to implement:
In the encoder (any of 10), if the first bin of a code word equals 'Ο', bins are processed until a T occurs or until n 0s are read (including the first '0' of the code word). If a T occurred, the read bins are written to the bit stream unchanged. Otherwise (i.e. n 0s were read), n - 1 Is are written to the bit stream. If the first bin of the code word equals T, bins are processed until a '0' occurs or until n - 1 Is are read (including the first T of the code word). If a '0' occurred, the read bins are written to the bit stream unchanged. Otherwise (i.e. n - 1 Is were read), n 0s are written to the bit stream.
In the decoder (any of 22), the same algorithm is used as for the encoder, since this is the same for bin pipe codes as described above.
An efficient implementation of unary to rice codes is described next. An encoder and decoder for unary to rice codes can be efficiently implemented by using a counter as will be described now.
In the encoder (any of 10), bins are read from the bin stream (i.e. 7) until a 1 occurs or until 2" 0s are read. The number of 0s is counted. If the counted number equals 2n, the code word T is written to the bit stream. Otherwise, '0' is written, followed by the binary representation of the counted number, written with n bits.
In the decoder (any of 22), one bit is read. If it equals T, 2° 0s are decoded to the bin string. If it equals Ό', n more bits are read and interpreted as binary representation of a number. This number of 0s is decoded to the bin stream, followed by a T.
In other words, the just-described embodiments describe an encoder for encoding a sequence of symbols 3, comprising an assigner 16 configured to assign a number of parameters 5 to each symbol of the sequence of symbols based on information contained within previous symbols of the sequence of symbols; a plurality of entropy encoders 10 each of which is configured to convert the symbols 7 forwarded to the respective entropy encoder 10 into a respective bitstream 12; and a selector 6 configured to forward each symbol 3 to a selected one of the plurality of entropy encoders 10, the selection depending on the number of parameters 5 assigned to the respective symbol 3. According to the just-outlined embodiments, at least a first subset of the entropy encoders may be a variable length encoder configured to map symbol sequences of variable lengths within the stream of symbols 7 to codewords of variable lengths to be inserted in bitsream 12, respectively, with each of the entropy coders 10 of the first subset using a bijective mapping rule according to which code words of a primary prefix-free code with (2n-l) > 3 code words are mapped to code words of a secondary prefix-free code which is identical to the primary prefix code such that all but two of the code words of the primary prefix-free code are mapped to identical code words of the secondary prefix-free code while the two code words of the primary and secondary prefix-free codes have different lengths and are mapped onto each other in an interchanged manner, wherein the entropy encoders may use different n so as to covers different portions of an interval of the above-mentoined probability interval. The first prefix-free code may be constructed such that the codewords of the first prefix-free code are (a,b)2, (a,a,b)3, (a,...,a,b)n, (a,...,a)n, (b,a)2, (b,b,a)3, (b,...,b,a)n-1, (b,...,b)n-1, and the two codewords mapped onto each other in the interchanged manner are (a,...,a)n and (b,...,b)n-1 with b≠ a and a,b e {0,1 }. However, alternatives are feasible.
In other words, each of a first subset of entropy encoders may be configured to, in converting the symbols forwarded to the respective entropy encoder into the respective bitstream, examine a first symbol forwarded to the respective entropy encoder, to determine as to whether (1) the first symbol equals a e {0,1}, in which case the respective entropy encoder is configured to examine the following symbols forwarded to the respective entropy encoder to determine as to whether (1.1) b with b≠ a and b e {0,1} occurs within the next n-1 symbols following the first symbol, in which case the respective entropy encoder is configured to write a codeword to the respective bitstream, which equals the first symbol followed by the following symbols forwarded to the respective entropy encoder, up to the symbol b; (1.2) no b occurs within the next n-1 symbols following the first symbol, in which case the respective entropy encoder is configured to write a codeword to the respective bitstream, which equals (b,...,b)n-1; or (2) the first symbol equals b, in which case the respective entropy encoder is configured to examine the following symbols forwarded to the respective entropy encoder to determine as to whether (2.1) a occurs within the next n-2 symbols following the first symbol, in which case the respective entropy encoder is configured to write a codeword to the respective bitstream, which equals the first symbol followed by the following symbols forwarded to the respective entropy encoder up to the symbol a; or (2.2) no a occurs within the next n-2 symbols following the first symbol, in which case the respective entropy encoder is configured to write a codeword to the respective bitstream, which equals (a,...,a)n.
Additionally or alternatively, a second subset of the entropy encoders 10 may be a variable length encoder configured to map symbol sequences of variable lengths to codewords of fixed lengths, respectively, with each of the entropy coders of the second subset using a bijective mapping rule according to which code words of a primary truncated unary code with 2n+l code words of the type {(a), (ba), (bba), ... ,(b...ba), (bb...b)} with b≠ a and a,b e {0,1 } are mapped to code words of a secondary prefix-free code such that the codeword (bb...b) of the primary truncated unary code is mapped onto codeword (c) of the secondary prefix-free code and all other codewords {(a), (ba), (bba), ... ,(b...ba)} of the primary truncated unary code are mapped onto codewords having (d) with c≠ d and c,d e {0,1 } as a prefix and a n-bit word as suffix, wherein the entropy encoders use different n. Each of the second subset of entropy encoders may be configured such that the n-bit word is an n-bit representation of the number of b's in the respective codeword of the primary truncated unary code. However, alternatives are feasible. Again, from the perspective of the mode of operation of the respective encoder 10, each of the second subset of entropy encoders may be configured to, in converting the symbols forwarded to the respective entropy encoder into the respective bitstream, count a number of b's in a sequence of symbols forwarded to the respective entropy encoder, until an a occurs, or until the number of the sequence of symbols forwarded to the respective entropy encoder reaches 2n with all 2n symbols of the sequence being b, and (1) if the number of b's equals 2n, write c with c e {0,1 } as codeword of a secondary prefix-free code to the respective bitstream, and (2) if the number of b's is lower than 2", write a codeword of the secondary prefix-free code to the respective bitstream, which has (d) with c≠ d and d e {0,1} as prefix and a n-bit word determined depending on the number of b's as suffix.
Also additionally or alternatively, a predetermined one of the entropy encoders 10 may be a variable length encoder configured to map symbol sequences of fixed lengths to codewords of variable lengths, respectively, the predetermined entropy coder using a bijective mapping rule according to which 23 code words of length 3 of a primary code are mapped to code words of a secondary prefix-free code such that the codeword (aaa)3 of the primary code with a e {0,1 } is mapped onto codeword (c) with c e {0,1 }, all three codewords of the primary code having exactly one b with b≠a and b <≡ {0,1 } are mapped onto codewords having (d) with c≠ d and d e {0,1 } as a prefix and a respective first 2-bit word out of a first set of 2-bit words as a suffix, all three codewords of the primary code having exactly one a are mapped onto codewords having (d) as a prefix and a concatenation of a first 2-bit word not being an element of the first set and a second 2-bit word out of a second set of 2-bit words, as a suffix, and wherein the codeword (bbb)3 is mapped onto a codeword having (d) as a prefix and a concatenation of the first 2-bit word not being an element of the first set and a second 2-bit word not being an element of the second set, as a suffix. The first 2-bit word of the codewords of the primary code having exactly one b may be a 2-bit representation of a position of the b in the respective codeword of the primary code, and the second 2-bit word of the codewords of the primary code having exactly one a may be a 2-bit representation of a position of the a in the respective codeword of the primary code. However, alternatives are feasible.
Again, the predetermined one of the entropy encoders may be configured to, in converting the symbols forwarded to the predetermined entropy encoder into the respective bitstream, examine the symbols to the predetermined entropy encoder in triplets as to whether (1) the triplet consists of a's, in which case the predetermined entropy encoder is configured to write the codeword (c) to the respective bitstream, (2) the triplet exactly comprises one b, in which case the predetermined entropy encoder is configured to write a codeword having (d) as a prefix and a 2-bit representation of a position of the b in the triplet as a suffix, to the respective bitstream; (3) the triplet exactly comprises one a, in which case the predetermined entropy encoder is configured to write a codeword having (d) as a prefix and a concatenation of the first 2-bit word not being an element of the first set and a 2-bit representation of a position of the a in the triplet as a suffix, to the respective bitstream; or (4) the triplet consists of b's, in which case the predetermined entropy encoder is configured to write a codeword having (d) as a prefix and a concatenation of the first 2-bit word not being an element of the first set and the first 2-bit word not being an element of the second set as a suffix, to the respective bitstream.
Regarding the decoding side, just-described embodiments disclose a decoder for reconstructing a sequence of symbols 26, comprising a plurality of entropy decoders 22, each of which is configured to convert a respective bitstream 24 into symbols 21; an assigner 16 configured to assign a number of parameters to each symbol 26 of a sequence of symbols to be reconstructed based on information contained within previously reconstructed symbols of the sequence of symbols; and a selector 18 configured to retrieve each symbol 25 of the sequence of symbols to be reconstructed from a selected one of the plurality of entropy decoders, the selection depending on the number of parameters defined to the respective symbol. According to the just-described embodiments at least a first subset of the entropy decoders 22 are variable length decoders configured to map codewords of variable lengths to symbol sequences of variable lengths, respectively, with each of the entropy decoders 22 of the first subset using a bijective mapping rule according to which code words of a primary prefix-free code with (2n-l) > 3 code words are mapped to code words of a secondary prefix- free code which is identical to the primary prefix code such that all but two of the code words of the primary prefix-free code are mapped to identical code words of the secondary prefix- free code while the two code words of the primary and secondary prefix-free codes have different lengths and are mapped onto each other in an interchanged manner, wherein the entropy encoders use different n. The first prefix-free code may be constructed such that the codewords of the first prefix-free code are (a,b)2, (a,a,b)3, (a,...,a,b)n, (a,...,a)n, (b,a)2, (b,b,a)3, (b,...,b,a)n-1, (b,...,b)n-1, and the two codewords mapped onto each other in the interchanged manner may be (a,...,a)n and (b,...,b)n-1 with b≠ a and a,b e {0,1}. However, alternatives are feasible.
Each of the first subset of entropy encoders may be configured to, in converting the respective bitstream into the symbols, examine a first bit of the respective bitstream, to determine as to whether (1) the first bit equals a 0 {0,1}, in which case the respective entropy encoder is configured to examine the following bits of the respective bitstream to determine as to whether (1.1) b with b≠ a and b 0 {0,1} occurs within the next n-1 bits following the first bit, in which case the respective entropy decoder is configured to reconstruct a symbol sequence, which equals the first bit followed by the following bits of the respective bitstream, up to the bit b; or (1.2) no b occurs within the next n-1 bits following the first bit, in which case the respective entropy decoder is configured to reconstruct a symbol sequence, which equals (b,...,b)n-1; or (2) the first bit equals b, in which case the respective entropy decoder is configured to examine the following bits of the respective bitstream to determine as to whether (2.1) a occurs within the next n-2 bits following the first bit, in which case the respective entropy decoder is configured to reconstruct a symbol sequence, which equals the first bit followed by the following bits of the respective bitstream up to the symbol a; or (2.2) no a occurs within the next n-2 bits following the first bit, in which case the respective entropy decoder is configured to reconstruct a symbol sequence, which equals (a,...,a)n.
Additionally or alternatively, at least a second subset of the entropy decoders 22 may be a variable length decoder configured to map codewords of fixed lengths to symbol sequences of variable lengths, respectively, with each of the entropy decoders of the second subset using a bijective mapping rule according to which code words of a secondary prefix-free code are mapped onto code words of a primary truncated unary code with 2n+l code words of the type {(a), (ba), (bba), ... ,(b...ba), (bb...b)} with b≠ a and a,b e {0,1} such that codeword (c) of the secondary prefix-free code is mapped onto the codeword (bb...b) of the primary truncated unary code and codewords having (d) with c≠ d and c,d e {0,1 } as a prefix and a n-bit word as suffix, are mapped to a respective one of the other codewords {(a), (ba), (bba), ... ,(b...ba)} of the primary truncated unary code, wherein the entropy decoders use different n. Each of the second subset of entropy decoders may be configured such that the n-bit word is an n-bit representation of the number of b's in the respective codeword of the primary truncated unary code. However, alternatives are feasible.
Each of a second subset of entropy decoders may be a variable length decoder configured to map codewords of fixed lengths to symbol sequences of variable lengths, respectively, and configured to, in converting the bitstream of the respective entropy decoder into the symbols, examine a first bit of the respective bitstream to determine as to whether (1) same equals c with c e {0,1}, in which case the respective entropy decoder is configured to reconstruct a symbol sequence which equals (bb...b)2 n with b e {0,1}; or (2) same equals d with c≠ d and c,d e {0,1}, in which case the respective entropy decoder is configured to determine a n-bit word from n further bits of the respective bitstream, following the first bit, and reconstruct a symbol sequence therefrom which is of the type {(a), (ba), (bba), ... ,(b...ba), (bb...b)} with b≠ a and b e {0,1 } with the number of b's depending on the n-bit word.
Additionally or alternatively, a predetermined one of the entropy decoders 22 may be a variable length decoders configured to map codewords of variable lengths to symbol sequences of fixed lengths, respectively, the predetermined entropy decoder using a bijective mapping rule according to which code words of a secondary prefix-free code are mapped to 2 code words of length 3 of a primary code such that codeword (c) with c e {0,1 } is mapped to the codeword (aaa)3 of the primary code with a e {0,1 }, codewords having (d) with c≠ d and d€ {0,1 } as a prefix and a respective first 2-bit word out of a first set of three 2-bit words as a suffix are mapped onto all three codewords of the primary code having exactly one b with b≠a and b e {0,1 }, codewords having (d) as a prefix and a concatenation of a first 2-bit word not being an element of the first set and a second 2-bit word out of a second set of three 2-bit words, as a suffix are mapped onto all three codewords of the primary code having exactly one a, and a codeword having (d) as a prefix and a concatenation of the first 2-bit word not being an element of the first set and a second 2-bit word not being an element of the second set, as a suffix is mapped onto the codeword (bbb)3. The first 2-bit word of the codewords of the primary code having exactly one b may be a 2-bit representation of a position of the b in the respective codeword of the primary code, and the second 2-bit word of the codewords of the primary code having exactly one a may be a 2-bit representation of a position of the a in the respective codeword of the primary code. However, alternatives are feasible.
The predetermined one of the entropy decoders may be a variable length decoder configured to map codewords of variable lengths to symbol sequences of three symbols each, respectively, and configured to, in converting the bitstream of the respective entropy decoder into the symbols, examine the first bit of the respective bitstream to determine as to whether (1) the first bit of the respective bitstream equals c with c e {0,1 }, in which case the predetermined entropy decoder is configured to reconstruct a symbol sequence which equals (aaa)3 with a 0 {0,1 }, or (2) the first bit of the respective bitstream equals d with c≠ d and d e {0,1 }, in which case the predetermined entropy decoder is configured to determine a first 2- bit word from 2 further bits of the respective bitstream, following the first bit, and examine the first 2-bit word to determine as to whether (2.1) the first 2-bit word is no element of a first set of three 2-bit words, in which case the predetermined entropy decoder is configured to reconstruct a symbol sequence which has exactly one b with b≠a. and b 0 {0,1 }, with the position of b in the respective symbol sequence depending on the first 2-bit word, or (2.2) the first 2-bit word is element of the first set, in which case the predetermined entropy decoder is configured to determine a second 2-bit word from 2 further bits of the respective bitstream, following the two bits from which the first 2-bit word has been determined, and examine the second 2-bit word to determine as to whether (3.1) the second 2-bit word is no element of a second set of three 2-bit words, in which case the predetermined entropy decoder is configured to reconstruct a symbol sequence which has exactly one a, with the position of a in the respective symbol sequence depending on the second 2-bit word, or (3.2) the second 2-bit word is element of a second set of three 2-bit words, in which case the predetermined entropy decoder is configured to reconstruct a symbol sequence which equals (bbb)3 .
Next, a table-based binary arithmetic encoding and decoding scheme which could be used for one or several or even all of the encoders/decoders 10 and 22 is described, i.e. for each of these encoder/decoder pairs responsible for a certain probability interval of the Probability Interval Partitioning resulting from the distrbution of the symbols to the individual pathes.
Advantegeously, the below explained embodiments for encoders/decoders 10 and 22 are able to process multiple MPS's within a single renormalization cycle of a binary arithmetic decoding engine.
In order to understand the principles and advantages of the present embodiment, imagine a plurality of the encoder/decoder 10 and 22 pairs were binary arithmetic encoders/decoders of the following type. The mapping from the representative LPS/LPB probability /?LPS of a given bin buffer 8 to a corresponding code interval width i?u s - i-e. the interval subdivision of the internal state of the binary arithmetic coding engine, which is defined by the current interval width R and the current interval offset L, identifying, for example, the lower bound of the code interval - is realized by using a table look-up. Imagine, for each such table-based binary arithmetic coding engine 10 associated to a given bin buffer 8, K representative interval width values {Qo, QK-I) were used for representing R with the choice of K and the representative interval width values {Qo, QK-I) being dependent on the bin buffer 8. Thus, arithmetic encoding of a bin 9 may involve the substeps of mapping the current interval width R to a quantization index q with values in{0, K-l} and performing the interval subdivision by accessing product values for a product of the corresponding partial interval width value Qq times the probability estimate p for LPS associated with the current bin and the entropy encoder to which this bin has been assigned by assigner, from a lookup table Rtab with using q as an index. The resulting codewords of each such arithmetic coding engine 10 may be separately transmitted, packetized, or stored, or they may be interleaved for the purpose of transmission or storage as described above with respect to the various embodiments such as Fig. 1-11.
That is, such a binary arithmetic encoding engine 10 could perform the following steps in coding the bins 9 in its bin buffer 8:
1. Receiving valLPS, and bin from bin buffer 8 (recall: the respective binary arithmetic coding engine 10 considered here had been chosen or selected to receive the "bin" (by assigner 16 and selector 6) or, in other words, "bin" was associated with the respective binary arithmetic coding engine 10 considered here, because the probability distribution estimate, such as p_state[bin] as determined by assigner 4, was associated with that binary arithmetic coding engine 10 curently considered)
2. Quantization of R:
q_index = Qtab[R»q] (or some other form of quantization; K quantization levels are available) 3. Determination of RLPS and RMPS :
RLPS - Rtab [q_index] (note that p_state, i.e the probability of bin being LPS, is not mentioned here, as it is fixed for the binary arithmetic coding engine 10 considered, i.e. p_state is fixed for the current encoder 10 and may be denoted p_state [encoder], and Rtab has stored therein pre-calculated values for p[p_state[encoder]] · Q [q_index]
RMPS = R— RLPS (That is, the encoder 10 divides the interval into two portions)
4. Calculation of the new partial interval:
if (bin = 1 - valMPS) then {
L r- L + RMPS (That is, the lower portion of the current interval R immediately extending from L is associated with MPS) else
Figure imgf000154_0001
5. Renormalization of L and R, writing bits,
while ( R < Rmin) {
R 4- R « 1
L <- L « 1
Write one bit to the respective bitstream 12 }
wherein
q_index describes the index of a quantization value read out of Qtab,
p_state describes the current state (fixed for and different among the binary arithmetic coding engines),
RLPS describes the interval width corresponding to the LPS and
RMPS describes the interval width corresponding to the MPS and
Rmin describes the minimum allowed value for R,
valMPS describes the value of the bit corresponding to the MPS.
Accordingly, a binary arithmetic decoding engine 22 could perform the following steps in decoding the bins output into bin buffer:
1. Receiving request for a bin or bit from the respective partial bitstream 24 (recall: the respective binary arithmetic decoding engine 22 considered here had been chosen to decode this "bin" or "bit" (or, in other words, "bin" was associated with the respective binary arithmetic decoding engine 22), namely by the assigner 16 and selector 18 because the probability distribution estimate, such as p_state[bin], was associated with that binary arithmetic decoding engine by assigner 16)
2. Quantization of R:
q_index = Qtab[R»q] (or some other form of quantization)
3. Determination of RLPS and R PS:
RLPS = Rtab [q_index] (note that p_state is not mentioned here, as it is fixed for the binary arithmetic decoding engine considered, i.e. p_state [encoder], and Rtab has stored therein pre-calculated values for p[p_state[encoder]] Q[q_index]
Figure imgf000155_0001
4. Determination of bin depending on the position of the partial interval:
Figure imgf000155_0002
bin 4- 1 - valMPS (bin is decoded as LPS; bin buffer selector 18 will obtain the actual bin value by use of this bin information and valMPS)
V <- V - RMPS
R r- RLPS}
else {
bin— valMPS (bin is decoded as MPS; bin buffer selector will obtain the actual bin value by use of this bin information and valMPS)
R <— RMPS}
5. Renormalization of R, reading out one bit and updating V,
while ( R < Rmin) {
R <- R « 1
bit— read one bit from respective partial bitsream 24
V <- (V « l) + bit } wherein
q_index describes the index of a quantization value read out of Qtab,
p_state describes the current state (fixed for the binary arithmetic decoding engine),
RLps describes the interval width corresponding to the LPS,
RMPS describes the interval width corresponding to the MPS,
Rmin describes the minimum allowed value for R,
valMPS describes the value of the bit corresponding to the MPS, and
V describes a value from the interior of the current partial interval.
The following outlined embodiment for a bin decoder 22 achieves an improvement of the decoding process in terms of throughput as is described below. In particular, coding of a run of MPS, leaves L unchanged while R changes merely slowly. As p state is contant for the encoder/decoder pairs 10/22, it's possible to decode a run of MPS up to the next renormalization necessary without having to loop through steps 1 to 5 including the renormalization check in step 5 for all symbols of the run of MPS individually. The following discussion focuses on the construction of one encoder/decoder pair associated with one corresponding pjstate. However, more than one of the decoders 22 may be constructed accordingly, with, however, different p_state or Rtab, respectively.
Let Do, DK-I denote the K decision thresholds that are used to derive the value q_index from the range value R in step 3 ("Quantization of R"). For R being in the interval [DB Di+i), the value of q_index equal to i is derived. This is illustrated in Fig. 62. The value range of R is, per definition and per effect of the renormalization, between Rmjn and Rmax. As shown herein, Do, K-I may divide the interval [Rmin, Rmax) into subintervals of substantially equal length.
To each interval with interval index i = 0...K-1, a representative interval width value out of set {Qo, QK-J}, belongs, and, as described above, an value for RLps out of {RLPS(0), ... , RLps(K-l)} representing, thus, a tabulated value ?LPS for the given representative interval width value and the estimated probability for LPS associated with the bin and the entropy en- /decoder, respectively. Note that D0 is equal to Rmin.
For sufficiently small values of i?n>s relative to the decision interval width Di+j - Dh multiple MPS can be decoded within a single cycle (steps 1-5 as outlined above). This is illustrated in Fig. 61a,b. However, the exact number of MPS that can be decoded simultaneously, depends on R, Dj, and V, provided that the bin buffer selector 18 requests multiple bins from the same bin buffer 20. The maximum number A s of multiple MPS is given as follows. Let
ST = max{K + l, (1) then
R-T
N MPS (2)
R LPS
The first alternative "T^V+1" of formula (1) is illustrated in Fig. 61a, while the second, i.e. "T=DrRLps" is illustrated in Fig. 61b.
Using this notation, step 4 of the binary arithmetic decoding engine as outlined above can be modified as follows: Determination of bin depending on the position of the partial interval:
if (V > RMPS) then [no modification for the LPS path] {
bin <— 1 - valMPS (bin is decoded as LPS; bin buffer selector will obtain the actual bin value by use of this bin information and valMPS)
Figure imgf000157_0001
R «- RLPS >
else {
Determination of JVMPS using formulas (1) and (2)
R <- R -NMPS * RLPS
for i=0 to NMPS -l
bin <- valMPS (NMPS bins are decoded as MPS; bin buffer selector will obtain the actual bin values by use of this bin information and valMPS) }
If arbitrary values were admitted for RLPS, this would require the application of integer division and multiplication operations in the first two substeps of the MPS path in the modified step 4 (corresponding to the "else" branch).
In order to simplify these operations, powers of two are used as the tabulated RLPS values in Rtab [q_index] corresponding to the respective product representatives for products between {Qo,—, QK-J} and PLPS of the respective en-/decoder, according to the present embodiment. In addition, for those bin encoders/decoders 10 and 22 for which the assumption holds that ?LPS is small relative to the decision interval width Di+j - D, the number of K is kept small; typically with a choice of K < 2.
With the choice of Rtab [1...K] (corresponding to ?LPs) = 2 equation (2) is simplified to
Figure imgf000157_0002
while the update of R in substep 2 of the MPS branch in the modified step 4 is simplified to R <- R - ( NMPs « a )
As an illustration, we consider the case of 10 different bin encoders/decoder pairs 10, 22. Note that in our illustrative case, R is represented with 9 bits such that Rmjn = 256. Table 8 contains an example of the tabulated RLPS values with the illustrative choice of K = 4 for that case without using the present embodiment. Table 8: Tabulated RLPS values for 10 bin encoders/decoders (each corresponding to a single
Figure imgf000158_0001
Table 9 shows an example for the case where the present embodiment was applied for the design of the tabulated RLPS values for at least those three bin encoders/decoders that correspond to the last three rows of Table 9, i.e. rows number 8-10 which correspond to the lowest p_state values. For the bin encoders/decoders corresponding to row number 6 and 7 the principle was partially applied with the result that the RLPS values for at least two columns in each of these rows are represented by powers of two.
Table 9: Tabulated RLPS values for 10 bin encoders/decoders (each corresponding to a single row), where at least the 3 bin encoders/decoders corresponding to the 3 last rows (8-10) are constructed along the principle of the present embodiment
Row/Columns 1 2 3 4
1 128 168 200 230
2 126 156 186 214
3 116 146 172 198
4 100 124 148 170
5 80 98 116 134
6 64 64 96 96
7 48 48 64 64
8 32 32 32 32
9 16 16 16 16
10 8 8 8 8 In Table 9, each line corresponds to table Rtab[1...4] for ten different entropy en-/decoder pairs 10/22.
Thus, the embodiments described with respect to Fig. 61a to 62, inter alias describe a decoder for reconstructing a sequence of symbols 26, comprising a plurality of entropy decoders 22, each of which is configured to convert a respective bitstream 24 into symbols 21; an assigner 16 configured to assign a number of parameters to each symbol 26 of a sequence of symbols to be reconstructed based on information contained within previously reconstructed symbols of the sequence of symbols; and a selector 18 configured to retrieve each symbol 25 of the sequence of symbols to be reconstructed from a selected one of the plurality of entropy decoders, the selection depending on the number of parameters defined to the respective symbol, wherein each of at least a subset of the entropy decoders may be a binary arithmetic decoding engine having an internal state which is defined by a current interval width R and a value V within the current interval width R, the binary arithmetic decoding engine being configured to, in converting the respective bitstream into the symbols, (1) using K intervals [Dj, Di+i) with i=0...K-l, map the current interval width R onto a quantization index q_index so that R falls into interval [Dqjndex, Dqjndex+i); (2) perform an interval subdivision by using a table Rtab and the quantization index q_index as an index into the table Rtab so as to obtain the partial interval width RLps = Rtab[q_index] corresponding to LPS, with Rtab[z] being a power of two for i=0...K-l, wherein the partial interval width RMPS corresponding to MPS is R - RLPS 5 (3) check as to whether ?MPS≤ an if RMPS≤ V, reconstruct a current symbol to be equal to LPS, and update Vto be V-RMPS, an R to be RLPS, and if not ?MPs≤ V, reconstruct the current symbol and NMPS-1 subsequent symbols to be equal to MPS with N PS = (R - max{V+l, Dq_index-RLps})/RLPs and update R to be R - NMps ' R LPS; and (4) renormalize R und update V using the respective bitstream. The binary arithmetic decoding engine of each of the subset of the entropy decoders may be configured to, in renormalizing R und updating V, (4.1) check as to whether R < Rmjn with Rmjn = A), and if R < Rmj„, (4.2) increase R to be R-2y using minimal y so that R gets equal to or greater than Rmjn, and (4.3) update V by shifting V y bit positions to the left, and (4.4) set the y least significant bits of V according to next y bits of the respective bitstreams in line. The subset of the entropy decoders may comprise a plurality of entropy decoders each having a different table Rtab. If so, Rtab[i] for the plurality of decoders may be selected such that for each entropy decoder, there exists a pair of PLPS and a vector {Q0, QK-i) with Qt e [A Α+;) such that 0.99 pLPs &<Rtab[/]<1.01 PLPS-O for i=0...K, with LPS being mutually different among the plurality of entropy decoders. K may be mutually different among the plurality of entropy decoders. Among the plurality of entropy decoders, for at least one may K=l or £¾=Q0 for i=l ...K, wherein the binary arithmetic decoding engine of that entropy decoder may be configured to, in converting the respective bitstream into the symbols, perform an interval subdivision by setting RLPS = Qo; check as to whether i?Mps ≤ V, and if i?Mps ≤ V, reconstruct a current symbol to be equal to LPS, and update Vto be K- -MPS, and R to be RLPS, and if not ?MPs≤ V, reconstruct the current symbol and NMPS-1 subsequent symbols to be equal to MPS with N PS = (R - max{V+l, Rmin- RLPS})/RLPS and update R to be R - JVMPS · R LPS; and renormalize R und update V using the respective bitstream. The binary arithmetic decoding engine of each of the subset of the entropy decoders may be configured to, compute NMPS by shifting the result of the term (R - max{V+l, Dqjndex-RLPs}) log2(RLPs) bits to the right. Also, each of at least a subset of the entropy decoders may be a binary arithmetic decoding engine having an internal state which is defined by a current interval width R and a value V within the current interval width R, the binary arithmetic decoding engine being configured to, in converting the respective bitstream into the symbols, perform an interval subdivision by setting the the partial interval width RLPS corresponding to LPS to be equal to Q0, wherein the partial interval width RMPS corresponding to MPS is R - RLPS; check as to whether J?MPS < V, and if 7?MPS ≤ V, reconstruct a current symbol to be equal to LPS, and update Vto be V-RM?S, and R to be RLPS, and if not ?MPs≤ V, reconstruct the current symbol and NMPS-1 sybsequent symbols to be equal to MPS with NMPS = (R - max{V+l , D¾_index-RLPs}) RLPS and update R to be R - NMPS ' R LPS; and renormalize R und update V using the respective bitstream.
Although some aspects have been described in the context of an apparatus, it is clear that these aspects also represent a description of the corresponding method, where a block or device corresponds to a method step or a feature of a method step. Analogously, aspects described in the context of a method step also represent a description of a corresponding block or item or feature of a corresponding apparatus. Some or all of the method steps may be executed by (or using) a hardware apparatus, like for example, a microprocessor, a programmable computer or an electronic circuit. In some embodiments, some one or more of the most important method steps may be executed by such an apparatus.

Claims

Claims
1. Encoder for encoding a sequence of symbols, comprising an assigner configured to assign a number of parameters to each symbol of the sequence of symbols based on information contained within previous symbols of the sequence of symbols; a plurality of entropy encoders each of which is configured to convert the symbols forwarded to the respective entropy encoder into a respective bitstream; and a selector configured to forward each symbol to a selected one of the plurality of entropy encoders, the selection depending on the number of parameters assigned to the respective symbol.
2. Encoder according to claim 1 wherein the assigner is configured such that the number of parameters assigned to each symbol comprises, or is, a measure for an estimate of a probability distribution among the possible values the respective symbols may assume.
3. Encoder according to claim 2 wherein the sequence of symbols are of a binary alphabet, and the assigner is configured such that the estimate of the probability distribution consists of a measure for an estimate of a probability of a less probable or more probable bin value and an identifier specifying an estimate for which of the two possible bin values represents the less probable or more probable bin value.
4. Encoder according to claims 2 or 3 wherein the assigner is configured to assign a context to each symbol of the sequence of symbols based on the information contained within previous symbols of the sequence of symbols with each context having a respective probability distribution estimate associated therewith, and to adapt the probability distribution estimate of each context to an actual symbol statistic based on symbol values of previous symbols to which the respective context is assigned, and to determine the measure for the estimate of the probability distribution for each symbol based on the probability distribution estimate associated with the context assigned to the respective symbol.
5. Encoder according to claim 4, wherein the assigner is configured to, in determining the measure for the estimate of the probability distribution for each symbol, quantize the probability distribution estimate associated with the context assigned to the respective symbol to one of a plurality of probability distribution estimate representatives in order to obtain the measure for the estimate of the probability distribution and wherein the selector is configured such that a bijective association is defined between the plurality of entropy encoders and the plurality of probability distribution estimate representatives.
6. Encoder according to claim 5, wherein the selector is configured to change a quantization mapping from a range of the probability distribution estimate to the plurality of probability distribution estimate representatives in a predetermined deterministic way depending on previous symbols of the sequence of symbols over time.
7. Encoder according to claim 6, wherein the plurality of entropy encoders is configured to adapt their way of converting symbols into bit streams responsive to a change in the quantization mapping.
8. Encoder according to claim 5 or 6 wherein the selector is configured to change the quantization mapping such that bitrates of the bit streams into which the entropy encoders convert to symbols, are rendered less dispersed.
9. Encoder according to any of the preceding claims, wherein at least one of the plurality of the entropy encoders has a symbol input buffer associated therewith, wherein the selector is configured to forward the symbols to the at least one entropy encoder via the associated symbol input buffer.
10. Encoder according to any of the preceding claims wherein at least one entropy encoder is a variable length coder configured to map symbol sequences to codewords.
11. Encoder according to any of the preceding claims wherein each of the plurality of entropy encoders is a variable length coder configured to map symbol sequences of variable lengths to codewords of fixed lengths.
12. Encoder according to claim 11 further comprising a codeword buffer, wherein the codeword buffer is configured to reserve a sequence of codeword entries for the codewords from the plurality of entropy encoders in a sequential order depending on an order in which the symbols of the sequence of symbols forwarded by the selector to the plurality of entropy encoders result in a beginning of a new symbol sequence to be mapped to a codeword at the respective entropy encoder, and to remove codewords entered into the codeword entries in the sequential order to obtain a single stream of interleaved codewords from the plurality of bitstreams, wherein each entropy encoder is configured to sequentially enter his codewords into the codeword entries reserved for the respective entropy encoder.
13. Encoder according to claim 12, wherein the plurality of entropy encoders and the codeword buffer are configured to intermittently extend currently forwarded but not yet mapped symbols to valid symbol sequences by don't-care symbols having the currently forwarded but not yet mapped symbols as prefix, map the thus extended symbol sequences into codewords, enter the thus obtained codewords into the reserved codeword entries and flush the codeword entries.
14. Encoder according to claim 12, wherein the plurality of entropy encoders and the codeword buffer are configured to perform the intermittently extending, entering and flushing at events where a number of reserved codeword entries plus a number of codeword entries having codewords entered therein fulfils a predetermined criterion.
15. Encoder according to any of the preceding claims wherein at least a first subset of the entropy encoders is a variable length coder configured to map symbol sequences of variable lengths to codewords of variable lengths, respectively, with each of the entropy coders of the first subset using a bijective mapping rule according to which code words of a primary prefix- free code with (2n-l) > 3 code words are mapped to code words of a secondary prefix-free code which is identical to the primary prefix-free code such that all but two of the code words of the primary prefix-free code are mapped to identical code words of the secondary prefix- free code while the two code words of the primary and secondary prefix-free codes have different lengths and are mapped onto each other in an interchanged manner, wherein the entropy encoders use different n.
16. Encoder according to claim 15, wherein the first prefix-free code is constructed such that the codewords of the first prefix-free code are (a,b)2, (a,a,b)3, (a,...,a,b)n, (a,...,a)n, (b,a)2, (b,b,a)3, (b,...,b,a)n-1, (b,...,b)n-i, and the two codewords mapped onto each other in the interchanged manner are (a, ... ,a)n and (b, ... ,b)n-1 with b≠ a and a,b e {0, 1 } .
17. Encoder according to any of claims 1 to 16, wherein each of a first subset of entropy encoders is configured to, in converting the symbols forwarded to the respective entropy encoder into the respective bitstream, examine a first symbol forwarded to the respective entropy encoder, to determine as to whether the first symbol equals a e {0,1}, in which case the respective entropy encoder is configured to examine the following symbols forwarded to the respective entropy encoder to determine as to whether b with b≠ a and b e {0,1 } occurs within the next n-1 symbols following the first symbol, in which case the respective entropy encoder is configured to write a codeword to the respective bitstream, which equals the first symbol followed be the following symbols forwarded to the respective entropy encoder, up to the symbol b; no b occurs within the next n-1 symbols following the first symbol, in which case the respective entropy encoder is configured to write a codeword to the respective bitstream, which equals (b, ... ,b)n- 1 ; the first symbol equals b, in which case the respective entropy encoder is configured to examine the following symbols forwarded to the respective entropy encoder to determine as to whether a occurs within the next n-2 symbols following the first symbol, in which case the respective entropy encoder is configured to write a codeword to the respective bitstream, which equals the first symbol followed be the following symbols forwarded to the respective entropy encoder up to the symbol a; no a occurs within the next n-2 symbols following the first symbol, in which case the respective entropy encoder is configured to write a codeword to the respective bitstream, which equals (a,...,a)n.
18. Encoder according to any of claims 1 to 17 wherein at least a second subset of the entropy encoders is a variable length coder configured to map symbol sequences of variable lengths to codewords of fixed lengths, respectively, with each of the entropy coders of the second subset using a bijective mapping rule according to which code words of a primary truncated unary code with 2n+l code words of the type {(a), (ba), (bba), ... ,(b...ba), (bb...b)} with b≠ a and a,b e {0,1} are mapped to code words of a secondary prefix-free code such that the codeword (bb...b) of the primary truncated unary code is mapped onto codeword (c) of the secondary prefix-free code and all other codewords {(a), (ba), (bba), ... ,(b...ba)} of the primary truncated unary code are mapped onto codewords having (d) with c≠ d and c,d e {0, 1 } as a prefix and a n-bit word as suffix, wherein the entropy encoders use different n.
19. Encoder according to claim 18, wherein wherein each of the second subset of entropy encoders is configured such that the n-bit word is a n-bit representation of the number of b's in the respective codeword of the primary truncated unary code.
20. Encoder according to any of claims 1 to 19 wherein each of a second subset of entropy encoders is configured to, in converting the symbols forwarded to the respective entropy encoder into the respective bitstream, count a number of b's in a sequence of symbols forwarded to the respective entropy encoder, until an a occurs, or until the number of the sequence of symbols forwarded to the respective entropy encoder reaches 2" with all 2n symbols of the sequence being b, and if the number of b's equals 2n, write c with c e {0,1} as codeword of a secondary prefix-free code to the respective bitstream, and if the number of b's is lower than 2n, write a codeword of the secondary prefix-free code to the respective bitstream, which has (d) with c≠ d and d e {0,1} as prefix and a n-bit word determined depending on the number of b's as suffix.
21. Encoder according to any of claims 1 to 20 wherein a predetermined one of the entropy encoders is a variable length coder configured to map symbol sequences of fixed lengths to codewords of variable lengths, respectively, the predetermined entropy coder using a bijective mapping rule according to which 23 code words of length 3 of a primary code are mapped to code words of a secondary prefix-free code such that the codeword (aaa)3 of the primary code with a e {0,1} is mapped onto codeword (c) with c € {0,1}, all three codewords of the primary code having exactly one b with b≠a and b e {0,1 } are mapped onto codewords having (d) with c≠ d and d e {0,1} as a prefix and a respective first 2-bit word out of a first set of 2-bit words as a suffix, all three codewords of the primary code having exactly one a are mapped onto codewords having (d) as a prefix and a concatenation of a first 2-bit word not being an element of the first set and a second 2-bit word out of a second set of 2-bit words, as a suffix, and wherein the codeword (bbb)3 is mapped onto a codeword having (d) as a prefix and a concatenation of the first 2-bit word not being an element of the first set and a second 2-bit word not being an element of the second set, as a suffix.
22. Encoder according to claim 21, wherein the first 2-bit word of the codewords of the primary code having exactly one b is a 2-bit representation of a position of the b in the respective codeword of the primary code, and the second 2-bit word of the codewords of the primary code having exactly one a is a 2-bit representation of a position of the a in the respective codeword of the primary code.
23. Encoder according to any of claims 1 to 22 wherein a predetermined one of the entropy encoders is configured to, in converting the symbols forwarded to the predetermined entropy encoder into the respective bitstream, examine the symbols to the predetermined entropy encoder in triplets as to whether the triplet consists of a's, in which case the predetermined entropy encoder is configured to write the codeword (c) to the respective bitstream, the triplet exactly comprises one b, in which case the predetermined entropy encoder is configured to write a codeword having (d) as a prefix and a 2-bit representation of a position of the b in the triplet as a suffix, to the respective bitstream; the triplet exactly comprises one a, in which case the predetermined entropy encoder is configured to write a codeword having (d) as a prefix and a concatenation of the first 2-bit word not being an element of the first set and a 2-bit representation of a position of the a in the triplet as a suffix, to the respective bitstream; or the triplet consists of b's, in which case the predetermined entropy encoder is configured to write a codeword having (d) as a prefix and a concatenation of the first 2-bit word not being an element of the first set and the first 2-bit word not being an element of the second set as a suffix, to the respective bitstream.
24. Method for encoding a sequence of symbols, comprising assigning a number of parameters to each symbol of the sequence of symbols based on information contained within previous symbols of the sequence of symbols; and forwarding each symbol to a selected one of a plurality of entropy encoders, the selection depending on the number of parameters assigned to the respective symbol, and each of the plurality of entropy encoders being configured to convert symbols forwarded to the respective entropy encoder into a respective bitstream.
25. Decoder for reconstructing a sequence of symbols, comprising: a plurality of entropy decoders, each of which is configured to convert a respective bitstream into symbols; an assigner configured to assign a number of parameters to each symbol of a sequence of symbols to be reconstructed based on information contained within previously reconstructed symbols of the sequence of symbols; and a selector configured to retrieve each symbol of the sequence of symbols to be reconstructed from a selected one of the plurality of entropy decoders, the selection depending on the number of parameters defined to the respective symbol.
26. Decoder according to claim 25, wherein the assigner is configured such that the number of parameters assigned to each symbol comprises, or is, a measure for an estimate of a probability of distribution among the possible symbol values a respective symbol may assume.
27. Decoder according to claim 26, wherein the sequence of symbols to be reconstructed is of a binary alphabet and the assigner is configured such that the estimate of the probability distribution consists of a measure for an estimate of a probability of a less probable or more probable bin value of the two possible bin values of the binary alphabet and an identifier specifying an estimate for which of the two possible bin values represents the less probable or more probable bin value.
28. Decoder according to claim 26 or 27, wherein the assigner is configured to internally assign a context to each symbol of the sequence of symbols to be reconstructed based on the information contained within previously reconstructed symbols of the sequence of symbols to be reconstructed with each context having a respective probability distribution estimate associated therewith, and to adapt the probability distribution estimate for each context to an actual symbol statistic based on symbol values of previously reconstructed symbols to which the respective context is assigned, and to determine the measure for the estimate of the probability distribution for each symbol based on the probability distribution estimate associated with the context assigned to the respective symbol.
29. Decoder according to claim 28, wherein the assigner is configured to, in determining the measure for the estimate of the probability distribution for each symbol, quantize the probability distribution estimate associated with the context assigned with the respective symbol to one of a plurality of probability distribution estimate representatives in order to obtain the measure for the estimate of the probability distribution and wherein the selector is configured such that a bijective association is defined between the plurality of entropy encoders and the plurality of probability distribution estimate representatives.
30. Decoder according to claim 39, wherein the selector is configured to change a quantization mapping from a range of the probability distribution estimates to the plurality of probability distribution estimate representatives in a predetermined deterministic way depending on previously reconstructed symbols of the sequence of symbols, over time.
31. Decoder according to claim 30, wherein the plurality of entropy decoders is configured to adapt their way of converting symbols into bit streams responsive to a change in the quantization mapping.
32. Decoder according to claims 30 or 31 , wherein the selector is configured to change the quantization mapping such that rates by which the symbols are retrieved from the plurality of entropy decoders, are made less dispersed.
33. Decoder according to any of claims 25 to 32, wherein at least one entropy decoder has a symbol output buffer associated therewith, wherein the selector is configured to retrieve the symbols from the at least one entropy decoder via the associated symbol output buffer.
34. Decoder according to any of claims 25 to 33, wherein the at least one entropy decoder is a variable length decoder configured to map codewords to symbol sequences.
35. Decoder according to any of claims 25 to 34, wherein each of the plurality of entropy decoders is a variable length decoder configured to map codewords of fixed lengths to symbols sequences of variable lengths.
36. Decoder according to claim 35, further comprising a codeword entry for receiving a single stream of interleaved codewords, wherein the plurality of entropy decoders are configured to retrieve the codewords from the codeword entry in a sequential order depending on an order in which the symbols of the sequence of symbols to be reconstructed as retrieved by the selector from the plurality of entropy decoders result in a new symbol sequence to be mapped from a new codeword at the respective entropy decoders.
37. Decoder according to claim 36, wherein the plurality of entropy decoders and the selector are configured to intermittently discard suffixes of symbol sequences so as to not participate in forming the sequence of symbols to be reconstructed.
38. Decoder according to claim 37, wherein the plurality of entropy decoders and the selectors are configured to perform the intermittently discarding at events where a number of codewords having been retrieved from the codeword entry by the plurality of entropy decoders between two consecutive codeword retrievals of a respective entropy decoder from the codeword entry, fulfils a predetermined criterion.
39. Decoder according to any of claims 36 to 38, further comprising a codeword buffer at codeword entry, configured to parse the interleaved codewords and distribute same to the entropy decoders in units of codewords upon requests for retrieval from the decoders.
40. Decoder according to any of claims 25 to 39 wherein at least a first subset of the entropy decoders is a variable length decoder configured to map codewords of variable lengths to symbol sequences of variable lengths, respectively, with each of the entropy decoders of the first subset using a bijective mapping rule according to which a code words of a primary prefix-free code with (2n-l) > 3 code words are mapped to code words of a secondary prefix-free code which is identical to the primary prefic code such that all but two of the code words of the primary prefix-free code are mapped to identical code words of the secondary prefix-free code while the two code words of the primary and secondary prefix-free codes have different langths and are mapped onto each other in an interchanged manner, wherein the entropy encoders use different n.
41. Decoder according to claim 40, wherein the first prefix-free code is constructed such that the codewords of the first prefix-free code are (a,b)2, (a,a,b)3, (a,...,a,b)n, (a,...,a)n, (b,a)2, (b,b,a)3, (b,...,b,a)n-1, (b,...,b)n-1, and the two codewords mapped onto each other in the interchanged manner are (a,...,a)n and (b,...,b)n-1 with b≠ a and a,b <≡ {0,1}.
42. Decoder according to any of claims 25 to 41, wherein each of a first subset of entropy encoders is a variable length decoder configured to map codewords of variable lengths to symbol sequences of variable lengths, arid configured to, in converting the respective bitstream into the symbols, examine a first bit of the respective bitstream, to determine as to whether the first bit equals a e {0,1 }, in which case the respective entropy encoder is configured to examine the following bits of the respective bitstream to determine as to whether b with b≠ a and b e {0,1 } occurs within the next n-1 bits following the first bit, in which case the respective detropy decoder is configured to reconstruct a symbol sequence, which equals the first bit followed be the following bits of the respective bitstream, up to the bit b; no b occurs within the next n-1 bits following the first bit, in which case the respective entropy decoder is configured to reconstruct a symbol sequence, which equals (b,...,b)n-1; the first bit, equals b, in which case the respective entropy decoder is configured to examine the following bits of the respective bitstream to determine as to whether a occurs within the next n-2 bits following the first bit, in which case the respective entropy decoder is configured to reconstruct a symbol sequence, which equals the first bit followed be the following bits of the respective bitstream up to the symbol a; no a occurs within the next n-2 bits following the first bit, in which case the respective entropy decoder is configured to reconstruct a symbol sequence, which equals (a,...,a)n.
43. Decoder according to any of claims 35 to 42 wherein at least a second subset of the entropy decoders is a variable length decoder configured to map to codewords of fixed lengths to symbol sequences of variable lengths, respectively, with each of the entropy decoders of the second subset using a bijective mapping rule according to which code words of a secondary prefix-free code ar mapped onto code words of a primary truncated unary code with 2n+l code words of the type {(a), (ba), (bba), ... ,(b...ba), (bb...b)} with b≠ a and a,b e {0,1 } such that codeword (c) of the secondary prefix-free code is mapped onto the codeword (bb...b) of the primary truncated unary code and codewords having (d) with c≠ d and c,d e {0,1 } as a prefix and a n-bit word as suffix, are mapped to a respective one of the other codewords {(a), (ba), (bba), ... ,(b...ba)} of the primary truncated unary code, wherein the entropy decoders use different n.
44. Decoder according to claim 43, wherein each of the second subset of entropy decoders is configured such that the n-bit word is a n-bit representation of the number of b's in the respective codeword of the primary truncated unary code.
45. Decoder according to any of claims 25 to 44 wherein each of a second subset of entropy decoders is a variable length decoder configured to map codewords of fixed lengths to symbol sequences of variable lengths, respectively, and configured to, in converting the bitstream of the respective entropy decoder into the symbols, examine a first bit of the respective bitstream to determine as to whether same equals c with c e {0,1 }, in which case the respective entropy decoder is configured to reconstruct a symbol sequence which equals (bb...b)2" with b e {0, 1 } ; same equals d with c≠ d and c,d e {0,1 }, in which case the respective entropy decoder is configured to determine a n-bit word from n further bits of the respective bitstream, following the first bit, and reconstruct a symbol sequence therefrom which is of the type {(a), (ba), (bba), ... ,(b...ba), (bb...b)} with b≠ a and b e {0,1 } with the number of b's depending on the n-bit word.
46. Decoder according to any of claims 25 to 45 wherein a predetermined one of the entropy decoders is a variable length decoder configured to map codewords of variable lengths to symbol sequences of fixed lengths, respectively, the predetermined entropy decoder using a bijective mapping rule according to which code words of a secondary prefix-free code are mapped to 23 code words of length 3 of a primary code such that codeword (c) with c e {0,1 } is mapped to the codeword (aaa)3 of the primary code with a e {0,1 }, codewords having (d) with c≠ d and d€ {0,1} as a prefix and a respective first 2-bit word out of a first set of three 2-bit words as a suffix are mapped onto all three codewords of the primary code having exactly one b with b≠a and b e {0,1 }, codewords having (d) as a prefix and a concatenation of a first 2-bit word not being an element of the first set and a second 2-bit word out of a second set of three 2-bit words, as a suffix are mapped onto all three codewords of the primary code having exactly one a, and a codeword having (d) as a prefix and a concatenation of the first 2-bit word not being an element of the first set and a second 2-bit word not being an element of the second set, as a suffix is mapped onto the codeword (bbb)3.
47. Decoder according to claim 46, wherein the first 2-bit word of the codewords of the primary code having exactly one b is a 2-bit representation of a position of the b in the respective codeword of the primary code, and the second 2-bit word of the codewords of the primary code having exactly one a is a 2-bit representation of a position of the a in the respective codeword of the primary code.
48. Decoder according to any of claims 25 to 47- wherein a predetermined one of the entropy decoders is a variable length decoder configured to map codewords of variable lengths to symbol sequences of three symbols each, respectively, and configured to, in converting the bitstream of the respective entropy decoder into the symbols, examine the first bit of the respective bitstream to determine as to whether the first bit of the respective bitstream equals c with c e {0,1 }, in which case the predetermined entropy decoder is configured to reconstruct a symbol sequence which equals (aaa)3 with a 0 {0,1 }, or the first bit of the respective bitstream equals d with c≠ d and d e {0,1 }, in which case the predetermined entropy decoder is configured to determine a first 2-bit word from 2 further bits of the respective bitstream, following the first bit, and examine the first 2-bit word to determine as to whether the first 2-bit word is no element of a first set of three 2-bit words, in which case the predetermined entropy decoder is configured to reconstruct a symbol sequence which has exactly one b with b≠a and b e {0,1 }, with the position of b in the respective symbol sequence depending on the first 2-bit word, or the first 2-bit word is element of the first set, in which case the predetermined entropy decoder is configured to determine a second 2-bit word from 2 further bits of the respective bitstream, following the two bits from which the first 2-bit word has been determined, and examine the second 2-bit word to determine as to whether the second 2-bit word is no element of a second set of three 2-bit words, in which case the predetermined entropy decoder is configured to reconstruct a symbol sequence which has exactly one a, with the position of a in the respective symbol sequence depending on the second 2-bit word, or the second 2-bit word is element of a second set of three 2-bit words, in which case the predetermined entropy decoder is configured to reconstruct a symbol sequence which equals (bbb)3 .
49. Decoder according to any of claims 25 to 41 wherein each of at least a subset of the entropy decoders is a binary arithmetic decoding engine having an internal state which is defined by a current interval width R and a value V within the current interval width R, the binary arithmetic decoding engine being configured to, in converting the respective bitstream into the symbols, using K intervals [£>*, Α+;) with z-0...i -l, map the current interval width R onto a quantization index q_index so that R falls into interval [Dq_inciex, Dgjncjex+i); perform an interval subdivision by using a table Rtab and the quantization index q_index as an index into the table Rtab so as to obtain the the partial interval width RLPS = Rtab[q_index] corresponding to LPS, with Rtab[z] being a power of two for i=0...K-l, wherein the partial interval width R PS corresponding to MPS is R - RLPS; check as to whether ?MPS≤ and if ?MPS≤ V, reconstruct a current symbol to be equal to LPS, and update Vto be V-RMps, and R to be RLPS, an if not RMPS≤ reconstruct the current symbol and NMPS-1 subsequent symbols to be equal to MPS with NMPS = (R - max{V+l, Dq_index-RLPs}) RLPS and update R to be R - NMps · R LPS; and renormalize R und update V using the respective bitstream.
50. Decoder according to claim 49 wherein the binary arithmetic decoding engine of each of the subset of the entropy decoders is configured to, in renormalizing R und updating V, check as to whether R < Rmin with Rmin = D0, and if R < Rmin, increase R to be R-2Y using minimal y so that R gets equal to or greater than Rmin, and update V by shifting V y bit positions to the left, and set the y least significant bits of V according to next y bits of the respective bitstreams in line.
51. Decoder according to claim 49 or 50 wherein the subset of the entropy decoders comprises a plurality of entropy decoders each having a different table Rtab.
52. Decoder according to claim 51 wherein Rtab[i] for the"plurality of decoders is selected such that for each entropy decoder, there exists a pair of PLPS and a vector {Qo, QK-J} with Qt e [Di, +i) such that 0.99 pLps a<Rtab[ ]<1.01 PLpS a for i=0...K, with pLps being mutually different among the plurality of entropy decoders.
53. Decoder according to claim 51 or 52 wherein K is mutually different among the plurality of entropy decoders.
54. Decoder according to any of claims 51 to 53 wherein among the plurality of entropy decoders, for at least one K=l or
Figure imgf000173_0001
...K, wherein the binary arithmetic decoding engine of that entropy decoder is configured to, in converting the respective bitstream into the symbols, perform an interval subdivision by setting RLPS = Qo', check as to whether Rues≤ and if RMPS≤ K reconstruct a current symbol to be equal to LPS, and update Vto be V-RMPS, and R to be RLPS, and if not J?MPS < V, reconstruct the current symbol and NMps-l sybsequent symbols to be equal to MPS with NMPS = (R - max{V+l, Rmin- RLPS}) RLPS and update R to be R - iVMps ' R LPS; and renormalize R und update V using the respective bitstream.
55. Decoder according to any of claims 49 to 54 wherein the binary arithmetic decoding engine of each of the subset of the entropy decoders is configured to, compute NMps by shifting the result of the term (R - max{V+l, Dqjndex-RLPs}) log2(RLPs) bits to the right.
56. Decoder according to any of claims 25 to 41 wherein each of at least a subset of the entropy decoders is a binary arithmetic decoding engine having an internal state which is defined by a current interval width R and a value V within the current interval width R, the binary arithmetic decoding engine being configured to, in converting the respective bitstream into the symbols, perform an interval subdivision by setting the the partial interval width RLPS corresponding to LPS to be equal to Q0, wherein the partial interval width RMps corresponding to MPS is R - RLPS; check as to whether .RMPS≤ V, and if J?MPS≤ V, reconstruct a current symbol to be equal to LPS, and update Vto be V-Rups, and R to be RLps, and if not ?MPs≤ V, reconstruct the current symbol and NMPS-1 sybsequent symbols to be equal to MPS with N PS = (R - max{V+l, Dq_index-RLPs}) LPs and update R to be R - NM?s · R LPS and renoraialize R und update V using the respective bitstream.
57. A device for decoding a sequence of symbols sequentially representing syntax elements of a significance mapping and then symbol representations of values of transform coefficients being unequal to zero for blocks of (video) pictures containing transform coefficients being unequal to zero, wherein the syntax elements of the significance mapping specify for the positions of the transform coefficients in a scan order as to whether at the respective position a transform coefficient being unequal to zero or not, and the symbol representations of values of transform coefficients being unequal to zero are represented by the sequence of symbols in a reverse scan order - starting from the last transform coefficient being unequal to zero, the device being configured to reconstruct the sequence of symbols using a decoder according to any of claims 25 to 56.
58. Apparatus for entropy decoding entropy-encoded information symbols, the entropy- encoded information symbols being produced by entropy encoding a symbol of a sequence of symbols based on probability information for the symbol, the symbol being part of a symbol set, wherein the probability information for the symbol is derived based on a context of the symbol, the context including one or more context symbols processed earlier, and wherein, for entropy encoding the start symbol, an initialization probability information was used, the initialization probability information being based on an estimation of a symbol statistics relating to a start symbol and being determined such that an initialization probability distribution is different from an equi-probable distribution for all symbols of the symbol set, comprising: a decoder for reconstructing the sequence of symbols according to any of claims 25 to 56, the decoder being configured to obtain the probability information having been used when entropy encoding the sequence of information symbols, the decoder including an initializer for obtaining the initialization probability information having been used when entropy encoding the start symbol.
59. Decoder according to claim 28 or 29, wherein the assigner is configured to initialise the probability distribution estimate for each context to a respective initialisation probability distribution estimate in advance of a first occurrence of a symbol to which the respective contexts are assigned, within the sequence of symbols to be reconstructed, the initialization probability distribution estimate being based on symbol statistics relating to a start symbol and being determined such that an initialization probability distribution is different from an equi-probable distribution for all symbols of a symbol set underlying the symbols.
60. Decoder according to claim 28 or 29, wherein the sequence of symbols to be reconstructed comprises at least one symbol of a symbol representation of a predetermined syntax element, the sequence of symbols to be reconstructed being a pre-coded video signal being a pre-coded version of a video signal, the video signal representing at least one video frame being composed of picture samples, the picture samples belonging either to a first or a second field being captured at a different time instants, the video frame being spatially divided up into macroblock pair regions, each macroblock pair region being associated with a top and a bottom macroblock, each macroblock pair region being either of a first or a second distribution type, wherein, for each macroblock pair region, each of the pixel samples in the respective macroblock pair region is assigned to a respective one of the top and bottom macroblock of the respective macroblock pair region in accordance with the distribution type of the respective macroblock pair region, wherein the predetermined syntax element relates to a predetermined macroblock of the top and bottom macroblock of a predetermined macroblock pair region of the macroblock pair regions, wherein the assigner is configured to determine, for the predetermined syntax element, a neighboring macroblock, at least based upon as to whether the predetermined macroblock pair region is of a first or a second distribution type; assigning one of at least two contexts to the at least one symbol of the symbol representation of the predetermined syntax element based on a predetermined attribute of the neighboring macroblock, wherein each context is associated with a different probability distribution estimation; and determine the measure for the estimate of the probability distribution for the at least one symbol based on the probability distribution estimate associated with the context assigned to the at least one symbol.
61. Decoder according to claim 60 wherein assigner is configured to, in determining, for the predetermined syntax element, a neighboring macroblock, determine a neighboring macroblock to the left of the predetermined macroblock at least based upon as to whether the predetermined macroblock pair region is of the frame or field coded distribution type such that if the predetermined macroblock pair region is of the field coded distribution type, the neighboring macroblock to the left of the predetermined macroblock is determined to be a bottom macroblock of a macroblock pair region to the left of the predetermined macroblock pair region, if the macroblock pair region to the left is also of the field coded distribution type and the predetermined macroblock is the bottom macroblock of the predetermined macroblock pair region, and the neighboring macroblock to the left of the predetermined macroblock is determined to be a top macroblock of the macroblock pair region to the left of the predetermined macroblock pair region, if the macroblock pair region to the left is of the frame coded distribution type, or if the macroblock pair region to the left is of the field coded distribution type with the predetermined macroblock being the top macroblock of the predetermined macroblock pair region, and if the predetermmed macroblock pair region is of the frame coded distribution type, the neighboring macroblock to the left of the predetermined macroblock is determined to be the bottom macroblock of the macroblock pair region to the left of the predetermined macroblock pair region, if the macroblock pair region to the left is also of the frame coded distribution type and the predetermined macroblock is the bottom macroblock of the predetermined macroblock pair region, and the neighboring macroblock to the left of the predetermined macroblock is determined to be the top macroblock of the macroblock pair region to the left of the predetermined macroblock pair region, if the macroblock pair region to the left is of the field coded distribution type, or if the macroblock pair region to the left is of the frame coded distribution type with the predetermined macroblock being the top macroblock of the predetermined macroblock pair region and a neighboring macroblock to the top of the predetermined macroblock at least based upon as to whether the predetermined macroblock pair region is of a frame or field coded distribution type such that if the predetermined macroblock pair region is of the frame coded distribution type the neighboring macroblock to the top of the predetermined macroblock is determined to be a top macroblock of the predetermined macroblock pair region if the predetermined macroblock is the bottom macroblock of the predetermined macroblock pair region, and a bottom macroblock of the macroblock pair region to the top of the predetermined macroblock pair region if the predetermined macroblock is the top macroblock of the predetermined macroblock pair region, if the predetermined macroblock pair region is of the field coded distribution type and the predetermined macroblock is the top macroblock of the predetermined macroblock pair region, the neighboring macroblock to the top of the predetermined macroblock is determined to be the bottom macroblock of the macroblock pair region to the top of the predetermined macroblock pair region, if the macroblock pair region to the top of the predetermined macroblock pair region is of the frame coded distribution type, the top macroblock of the macroblock pair region to the top of the predetermined macroblock pair region, if the macroblock pair region to the top of the predetermined macroblock pair region is of the field coded distribution type, if the predetermined macroblock pair region is of the field coded distribution type and the predetermined macroblock is the bottom macroblock of the predetermined macroblock pair region, the neighboring macroblock to the top of the predetermined macroblock is determined to be the bottom macroblock of the macroblock pair region to the top of the predetermined macroblock pair region; assign the one of the at least two context models to the predetermined syntax element of the predetermined macroblock based on a predetermined attribute of the neighboring macroblock to the left of the predetermined macroblock and the neighboring macroblock to the top of the predetermined macroblock.
62. Decoder according to claim 60 or 61 wherein the predetermined syntax element relates to a predetermined macroblock of the top and bottom macroblock of a predetermined macroblock pair region of the macroblock pair regions and is an chroma prediction indicator indicating a type of spatial prediction used for chroma information of the predetermined macroblock, and wherein the assigner is configured to, in assigning the one of the at least two contexts, perform the assignment based on an availability of the neighboring macroblock indicating as to whether the predetermined macroblock and the neighboring macroblock belong to the same slice of the video frame or to different slices of the video frame, a macroblock type indicator of the neighboring macroblock, specifying a macroblock prediction mode and a partitioning of the neighboring macroblock used for prediction, the neighboring macroblock being inter or intra coded in the coded bit stream, and a synta element specifying, for the neighboring macroblock, a type of spatial prediction used for chroma information of the neighboring macroblock.
63. Decoder according to any of claims 60 to 62 wherein the first and second fields are top and bottom fileds of alternate rows of the video frame, respectively, and the first and second distribution types are frame coded or field coded distribution types, respectively, wherein each of the pixel samples in the respective macroblock pair region is assigned to a respective one of the top aiid bottom macroblock of the respective macroblock pair region in accordance with the distribution type of the respective macroblock pair region such that, in accordance with the frame coded distribution type, pixel samples lying in an upper part of the respective macroblock pair region are assigned to the top macroblock and pixel samples lying in a lower part of the respective macroblock pair region are assigned to the bottom macroblock, whereas, in accordance with the field coded distribution type, pixel samples in the respective macroblock pair region belonging to the top field are assigned to the top macroblock and pixel samples in the respective macroblock pair region belonging to the bottom field are assigned to the bottom macroblock, wherein each of the macroblocks is associated with a respective one of the syntax elements, the predetermined syntax element relating to a predetermined macroblock of the top and bottom macroblock of a predetermined macroblock pair region of the macroblock pair regions, wherein assigner is configured to, in determining, for the predetermined syntax element, a neighboring macroblock, determine a neighboring macroblock to the left of the predetermined macroblock based upon as to whether the predetermined macroblock pair region is of the frame or field coded distribution type, a macroblock pair region to the left is of the frame or field coded distribution type, and the predetermined macroblock is the top or bottom macroblock of the predetermined macroblock pair region such that if the predetermined macroblock pair region is of the field coded distribution type, the macroblock pair region to the left is also of the field coded distribution type, and the predetermined macroblock is the bottom macroblock of the predetermined macroblock pair region, then the neighboring macroblock to the left of the predetermined macroblock is determined to be a bottom macroblock of the macroblock pair region to the left of the predetermined macroblock pair region, if the predetermined macroblock pair region is of the field coded distribution type, the macroblock pair region to the left is of the frame coded distribution type, and the predetermined macroblock is the bottom macroblock of the predetermined macroblock pair region, then the neighboring macroblock to the left of the predetermined macroblock is determined to be a top macroblock of the macroblock pair region to the left of the predetermined macroblock pair region, if a line number of a luma location of an upper-left sample of the predetermined subpart, lies within a top half of the predetermined macroblock, and the neighboring macroblock to the left of the predetermined macroblock is determined to be the bottom macroblock of the macroblock pair region to the left of the predetermined macroblock pair region, if the line number of the luma location of the upper-left sample of the predetermined subpart, lies within a bottom half of the predetermined macroblock, if the predetermined macroblock pair region is of the field coded distribution type, a macroblock pair region to the left is also of the field coded distribution type, and the predetermined macroblock is the top macroblock of the predetermined macroblock pair region, then the neighboring macroblock to the left of the predetermined macroblock is determined to be the top macroblock of the macroblock pair region to the left of the predetermined macroblock pair region, if the predetermined macroblock pair region is of the field coded distribution type, the macroblock pair region to the left is of the frame coded distribution type, and the predetermined macroblock is the top macroblock of the predetermined macroblock pair region, then the neighboring macroblock to the left of the predetermined macroblock is determined to be the top macroblock of the macroblock pair region to the left of the predetermined macroblock pair region, if the line number of the luma location of the upper-left sample of the predetermined subpart, lies with n the top half of the predetermined macroblock, and the neighboring macroblock to the left of the predetermined macroblock is determined to be the bottom macroblock of the macroblock pair region to the left of the predetermined macroblock pair region, if the line number of the luma location of the upper-left sample of the predetermined subpart, lies within the bottom half of the predetermined macroblock, if the predetermined macroblock pair region is of the frame coded distribution type, a macroblock pair region to the left is also of the frame coded distribution type, and the predetermined macroblock is the top macroblock of the predetermined macroblock pair region, then the neighboring macroblock to the left of the predetermined macroblock is determined to be the top macroblock of the macroblock pair region to the left of the predetermined macroblock pair region, if the predetermined macroblock pair region is of the frame coded distribution type, the macroblock pair region to the left is of the field coded distribution type, and the predetermined macroblock is the top macroblock of the predetermined macroblock pair region, then the neighboring macroblock to the left of the predetermined macroblock is determined to be the top macroblock of the macroblock pair region to the left of the predetermined macroblock pair region, if the predetermined macroblock pair region is of the frame coded distribution type, a macroblock pair region to the left is also of the frame coded distribution type, and the predetermined macroblock is the bottom macroblock of the predetermined macroblock pair region, then the neighboring macroblock to the left of the predetermined macroblock is determined to be the bottom macroblock of the macroblock pair region to the left of the predetermined macroblock pair region, if the predetermined macroblock pair region is of the frame coded distribution type, the macroblock pair region to the left is of the field coded distribution type, and the predetermined macroblock is the bottom macroblock of the predetermined macroblock pair region, then the neighboring macroblock to the left of the predetermined macroblock is determined to be the top macroblock of the macroblock pair region to the left of the predetermined macroblock pair region; and assign the one of the at least two context models to the predetermined syntax element based on a predetermined attribute of the neighboring macroblock to the left of the predetermined macroblock.
64. Decoder according to any of claims 60 to 63, wherein the syntax element is a skip indicator specifying, for the respective macroblock, as to whether the respective macroblock is to be skipped when decoding the pre-coded video signal; the syntax element is a type indicator specifying a macroblock prediction mode and a partitioning of the respective macroblock used for prediction; the syntax element is a pattern indicator specifying, for the respective macroblock, which of sub-parts of the respective macroblock may contain non-zero transform coefficients and which may not; the syntax element is an chroma prediction indicator indicating, for the respective macroblock, a type of spatial prediction used for chroma information of the respective macroblock; the predetermined macroblock is transformed from spatial domain to frequency domain in units of sub-parts into which the predetermined macroblock is partitioned, or smaller sub-units partitioning, in turn, the sub-parts, and the predetermined syntax element specifies, for a predetermined sub-part of the predetermined macroblock, as to whether the predetermined sub-part corresponds to a set of transform coefficients comprising at least one non-zero transform coefficient; the syntax element specifies, for a predetermined one of the partitions for a respective macroblock, a difference between a motion vector component and a prediction of the motion vector component; or wherein the predetermined syntax element relates to a predetermined macroblock of the top and bottom macroblocks of a predetermined macroblock pair region of the macroblock pair regions and specifies, for a predetermined one of partitions of the predetermined macroblock, an index of a reference picture used in motion prediction of the predetermined macroblock.
65. Decoder according to cany of claim 25 to 56, wherein the symbols of the sequence of symbols to be reconstructed are bins and comprise a binary representation of a data value, the binary representation of the data value having a primary prefix which is a binarization of the minimum of the data value and a predetermined cut-off value in accordance with a first binarization scheme and, if the data value is greater than the predetermined cut-off value, a primary suffix appended to the primary prefix, the primary suffix being a binarization of the difference of the data value minus the predetermined cut-off value in accordance with a second binarization scheme, wherein the assigner and the selector are configured such that for each bin in the binary representation, if the bin is part of the primary prefix, the assigner determines the paramter for the bin and the selector performs the the retrival and selction based thereon; and if the bin of the codeword is part of the primary suffix, the selctor retrives the bin from a decoding stage of the decoder configured to decode the bin accordoing to a static bit value probability estimation, thereby obtaining the binary representationof the data value.
66. Decoder according to claim 25 or 56, wherein the sequence of symbols to be reconstructed comprises at least one symbol of a symbol representation of a predetermined syntax element, the sequence of symbols to be reconstructed being a pre-coded video representing at least one video frame, the syntax element being associated with a predetermined portion of the video frame and being indicative of as to whether the predetermined portion of the video frame is precoded in a first or a second way into the precoded video signal, wherein the assigner is configured to investigate as to whether a neighboring portion of the video frame neighboring the predetermined portion is precoded in the first way or the second way, in order to obtain a binary value, and assign one of at least two contexts to the at least one symbol of the symbol representation of the predetermined syntax element based on the' binary value, wherein each context model is associated with a different probability estimation; and determine the measure for the estimate of the probability distribution for the at least one symbol based on the probability distribution estimate associated with the context assigned to the at least one symbol.
67. Decoder according to claim 19 or 20, wherein the sequence of symbols to be reconstructed comprises at least one symbol of a symbol representation of transform data unit or a sub-unit thereof, the transform data unit being a precoded version of a predetermined transform coefficient of transform coefficients which are the result of a transformation of data corresponding to a predetermined portion of portions of a video frame or picture, the portions being either a portion of a first type being associated with a first scanning pattern or a portion of a second type being associated with a second scanning pattern, wherein the assigner is configured to choose one of a first and a second set of context models, depending on as to whether the predetermined portion is a portion of a type being associated with the first or the second scanning pattern, the first and the second set being different to each other; assign one context model of the chosen one of the first and the second set of context models to the at least one symbol of the symbol representation of the transform data unit or the sub-unit thereof based on the scanning position assigned to the predetermined transform coefficient, wherein each context model is associated with a different probability estimation; and determine the measure for the estimate of the probability distribution for the at least one symbol based on the probability distribution estimate associated with the context assigned to the at least one symbol.
68. Method for reconstructing a sequence of symbols, comprising: assigning a number of parameters to each symbol of a sequence of symbols to be reconstructed based on information contained within previously reconstructed symbols of the sequence of symbols; and retrieving each symbol of the sequence of symbols to be reconstructed from a selected one of a plurality of entropy decoders, the selection depending on the number of parameters defined to the respective symbol, and each of the plurality of entropy decoders being configured to convert a respective bitstream into symbols;
69. Computer readable digital storage medium having stored thereon a computer program having a program code for performing, when running on a computer, a method according to claim 24 or 68.
PCT/EP2011/055528 2010-04-13 2011-04-08 Probability interval partioning encoder and decoder WO2011128268A1 (en)

Priority Applications (5)

Application Number Priority Date Filing Date Title
JP2013504210A JP5676744B2 (en) 2010-04-13 2011-04-08 Entropy coding
EP11714643.1A EP2559166B1 (en) 2010-04-13 2011-04-08 Probability interval partioning encoder and decoder
CN201180029224.XA CN103119849B (en) 2010-04-13 2011-04-08 Probability interval partition encoding device and decoder
PL11714643T PL2559166T3 (en) 2010-04-13 2011-04-08 Probability interval partioning encoder and decoder
US13/647,956 US8907823B2 (en) 2010-04-13 2012-10-09 Entropy coding

Applications Claiming Priority (6)

Application Number Priority Date Filing Date Title
EPPCT/EP2010/054828 2010-04-13
EP10159793.8 2010-04-13
EP2010054828 2010-04-13
EP10159793 2010-04-13
US201161432875P 2011-01-14 2011-01-14
US61/432,875 2011-01-14

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US13/647,956 Continuation US8907823B2 (en) 2010-04-13 2012-10-09 Entropy coding

Publications (1)

Publication Number Publication Date
WO2011128268A1 true WO2011128268A1 (en) 2011-10-20

Family

ID=44070634

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/EP2011/055528 WO2011128268A1 (en) 2010-04-13 2011-04-08 Probability interval partioning encoder and decoder

Country Status (8)

Country Link
US (1) US8907823B2 (en)
EP (1) EP2559166B1 (en)
JP (1) JP5676744B2 (en)
CN (1) CN103119849B (en)
HU (1) HUE037656T2 (en)
PL (1) PL2559166T3 (en)
TW (1) TWI514779B (en)
WO (1) WO2011128268A1 (en)

Cited By (38)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2012172114A1 (en) 2011-06-16 2012-12-20 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Context initialization in entropy coding
DE102011085451B3 (en) * 2011-10-28 2013-04-11 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Encoding device for encoding multiple symbols in encoded bitstream, has multiplexer, which is arranged to attach block coded bits with each other for encoded bitstream so that series is predetermined
CN103096062A (en) * 2011-11-01 2013-05-08 珠海扬智电子科技有限公司 Arithmetic decoding method
WO2013067436A1 (en) * 2011-11-04 2013-05-10 Huawei Technologies Co., Ltd. Binarization of prediction residuals for lossless video coding
WO2013070487A1 (en) * 2011-11-08 2013-05-16 Sony Corporation Binarisation of last position for higher throughput
WO2013158642A1 (en) * 2012-04-16 2013-10-24 Qualcomm Incorporated Transform coefficient coding
CN103460701A (en) * 2012-01-20 2013-12-18 索尼公司 Complexity reduction of significance map coding
JP2014504077A (en) * 2010-12-03 2014-02-13 クゥアルコム・インコーポレイテッド Individual coding of the position of the last significant coefficient of a video block in video coding
WO2014049982A1 (en) * 2012-09-28 2014-04-03 三菱電機株式会社 Video encoding device, video decoding device, video encoding method and video decoding method
JP2014509158A (en) * 2011-03-08 2014-04-10 クゥアルコム・インコーポレイテッド Coding transform coefficients for video coding
JP2014195269A (en) * 2011-11-19 2014-10-09 Blackberry Ltd Multilevel effectiveness mapping scanning
CN104104949A (en) * 2013-04-12 2014-10-15 浙江大学 Entropy encoding and decoding method of direction information syntax element and device
JP2014535222A (en) * 2011-11-03 2014-12-25 クゥアルコム・インコーポレイテッドQualcomm Incorporated Initializing context states and probabilities for context-adaptive entropy coding
JP2015504292A (en) * 2012-01-19 2015-02-05 クゥアルコム・インコーポレイテッドQualcomm Incorporated Context-adaptive entropy coding with reduced initial value set
JP2015039231A (en) * 2010-04-23 2015-02-26 エム アンド ケー ホールディングス インコーポレイテッド Device and method for encoding video
JP2015510319A (en) * 2012-01-20 2015-04-02 フラウンホッファー−ゲゼルシャフト ツァ フェルダールング デァ アンゲヴァンテン フォアシュンク エー.ファオ Transform coefficient coding
EP2773109A4 (en) * 2011-10-31 2015-05-06 Samsung Electronics Co Ltd Method and apparatus for determining a context model for transform coefficient level entropy encoding and decoding
CN104737542A (en) * 2013-01-30 2015-06-24 英特尔公司 Content adaptive entropy coding for next generation video
JP2015149774A (en) * 2011-06-28 2015-08-20 サムスン エレクトロニクス カンパニー リミテッド Method and apparatus for coding video and method and apparatus for decoding video accompanied with arithmetic coding
US9154792B2 (en) 2011-11-08 2015-10-06 Qualcomm Incorporated Progressive coding of position of last significant coefficient
CN105027560A (en) * 2012-01-21 2015-11-04 摩托罗拉移动有限责任公司 Method of determining binary codewords for transform coefficients
US9237344B2 (en) 2012-03-22 2016-01-12 Qualcomm Incorporated Deriving context for last position coding for video coding
US9313498B2 (en) 2012-04-16 2016-04-12 Qualcomm Incorporated Sign hiding techniques for quantized transform coefficients in video coding
US9357185B2 (en) 2011-11-08 2016-05-31 Qualcomm Incorporated Context optimization for last significant coefficient position coding
US9516349B2 (en) 2011-07-12 2016-12-06 Futurewei Technologies, Inc. Pixel-based intra prediction for coding in HEVC
AU2015201781B2 (en) * 2011-10-31 2017-01-19 Samsung Electronics Co., Ltd. Method and apparatus for determining a context model for transform coefficient level entropy encoding and decoding
RU2611992C2 (en) * 2012-01-17 2017-03-01 Квэлкомм Инкорпорейтед Increasing throughput capacity for cabac coefficients level encoding
CN106941610A (en) * 2017-04-25 2017-07-11 西安电子科技大学 Based on the two-value ROI mask coding methods for improving block encription
RU2630750C1 (en) * 2014-02-20 2017-09-12 Гурулоджик Микросистемс Ой Device and method for encoding and decoding initial data
RU2639266C1 (en) * 2012-03-26 2017-12-20 ДжейВиСи КЕНВУД КОРПОРЕЙШН Device for decoding images, method for decoding images and long-term record medium
RU2633175C9 (en) * 2012-04-13 2018-03-19 ДжейВиСи КЕНВУД КОРПОРЕЙШН Picture coding device, picture coding method and picture coding program, and picture decoding device, picture decoding method and picture decoding program
US10340946B2 (en) 2015-11-01 2019-07-02 Gurulogic Microsystems Oy Encoders, decoders, and methods
US10645388B2 (en) 2011-06-16 2020-05-05 Ge Video Compression, Llc Context initialization in entropy coding
WO2020145698A1 (en) * 2019-01-13 2020-07-16 엘지전자 주식회사 Method and device for processing video signal
WO2022000089A1 (en) * 2020-06-30 2022-01-06 National Research Council Of Canada Vector space model for form data extraction
US11330272B2 (en) 2010-12-22 2022-05-10 Qualcomm Incorporated Using a most probable scanning order to efficiently code scanning order information for a video block in video coding
CN114731446A (en) * 2019-09-24 2022-07-08 弗劳恩霍夫应用研究促进协会 Coding concept for sequences of information values
CN114731407A (en) * 2019-09-24 2022-07-08 弗劳恩霍夫应用研究促进协会 Arithmetic encoder and decoder, video encoder and decoder, method for encoding or decoding symbols and method for encoding or decoding video content

Families Citing this family (119)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8077064B2 (en) * 2010-02-26 2011-12-13 Research In Motion Limited Method and device for buffer-based interleaved encoding of an input sequence
US20120082228A1 (en) * 2010-10-01 2012-04-05 Yeping Su Nested entropy encoding
US10104391B2 (en) 2010-10-01 2018-10-16 Dolby International Ab System for nested entropy encoding
US8537038B1 (en) * 2010-10-22 2013-09-17 Smith Micro Software, Inc. Efficient compression method for sorted data representations
US8755620B2 (en) 2011-01-12 2014-06-17 Panasonic Corporation Image coding method, image decoding method, image coding apparatus, image decoding apparatus, and image coding and decoding apparatus for performing arithmetic coding and/or arithmetic decoding
US8687904B2 (en) * 2011-01-14 2014-04-01 Panasonic Corporation Image coding method, image coding apparatus, image decoding method, image decoding apparatus, and image coding and decoding apparatus which include arithmetic coding or arithmetic decoding
US9066104B2 (en) 2011-01-14 2015-06-23 Google Inc. Spatial block merge mode
WO2012134246A2 (en) 2011-04-01 2012-10-04 엘지전자 주식회사 Entropy decoding method, and decoding apparatus using same
US9379736B2 (en) * 2011-06-03 2016-06-28 Qualcomm Incorporated Context-adaptive coding video data
US9258565B1 (en) * 2011-06-29 2016-02-09 Freescale Semiconductor, Inc. Context model cache-management in a dual-pipeline CABAC architecture
US20130083856A1 (en) * 2011-06-29 2013-04-04 Qualcomm Incorporated Contexts for coefficient level coding in video compression
US8798139B1 (en) 2011-06-29 2014-08-05 Zenverge, Inc. Dual-pipeline CABAC encoder architecture
US9060173B2 (en) 2011-06-30 2015-06-16 Sharp Kabushiki Kaisha Context initialization based on decoder picture buffer
US9338465B2 (en) * 2011-06-30 2016-05-10 Sharp Kabushiki Kaisha Context initialization based on decoder picture buffer
CN104041035B (en) * 2011-06-30 2017-06-27 华为技术有限公司 Lossless coding and coherent signal method for expressing for composite video
US11647197B2 (en) 2011-06-30 2023-05-09 Velos Media, Llc Context initialization based on slice header flag and slice type
US8891616B1 (en) 2011-07-27 2014-11-18 Google Inc. Method and apparatus for entropy encoding based on encoding cost
CN103299307B (en) * 2011-08-23 2016-08-03 华为技术有限公司 For estimating method of estimation and the estimator of the probability distribution of quantization index
US9300975B2 (en) 2011-09-11 2016-03-29 Texas Instruments Incorporated Concurrent access shared buffer in a video encoder
JPWO2013046504A1 (en) * 2011-09-29 2015-03-26 パナソニックIpマネジメント株式会社 Arithmetic decoding device, image decoding device, and arithmetic decoding method
US9185428B2 (en) 2011-11-04 2015-11-10 Google Technology Holdings LLC Motion vector scaling for non-uniform motion vector grid
MX346677B (en) * 2011-11-04 2017-03-29 Sharp Kk * Arithmetic decoding device, image decoding device, arithmetic coding device, image coding device, and arithmetic decoding method.
EP2779648A4 (en) * 2011-11-07 2016-06-01 Panasonic Ip Corp America Image encoding method, image decoding method, image encoding device, image decoding device, and image encoding/decoding device
PT2777285T (en) 2011-11-11 2017-12-01 Ge Video Compression Llc Adaptive partition coding
EP3468184B1 (en) 2011-11-11 2022-07-27 GE Video Compression, LLC Effective wedgelet partition coding using prediction
EP2777286B1 (en) 2011-11-11 2017-01-04 GE Video Compression, LLC Effective wedgelet partition coding
CN104247427B (en) 2011-11-11 2018-10-12 Ge视频压缩有限责任公司 device and method for coding and decoding
EP2803190B1 (en) 2012-01-09 2017-10-25 Dolby Laboratories Licensing Corporation Hybrid reference picture reconstruction method for multiple layered video coding systems
US9503717B2 (en) * 2012-01-09 2016-11-22 Texas Instruments Incorporated Context adaptive binary arithmetic coding (CABAC) with scalable throughput and coding efficiency
WO2013106987A1 (en) * 2012-01-16 2013-07-25 Mediatek Singapore Pte. Ltd. Methods and apparatuses of bypass coding and reducing contexts for some syntax elements
US9729884B2 (en) 2012-01-18 2017-08-08 Lg Electronics Inc. Method and device for entropy coding/decoding
KR101647242B1 (en) * 2012-01-19 2016-08-09 미쓰비시덴키 가부시키가이샤 Image decoding device, image and coding device, image decoding method, image coding method and storage medium
AU2012200319B2 (en) 2012-01-19 2015-11-26 Canon Kabushiki Kaisha Method, apparatus and system for encoding and decoding the significance map for residual coefficients of a transform unit
CN104956674B (en) * 2012-01-20 2019-02-15 谷歌技术控股有限责任公司 Device and method for the context reduction in the position encoding and decoding of most end significant coefficient
AU2012200345B2 (en) * 2012-01-20 2014-05-01 Canon Kabushiki Kaisha Method, apparatus and system for encoding and decoding the significance map residual coefficients of a transform unit
US9531990B1 (en) 2012-01-21 2016-12-27 Google Inc. Compound prediction using multiple sources or prediction modes
US11039138B1 (en) 2012-03-08 2021-06-15 Google Llc Adaptive coding of prediction modes using probability distributions
US8737824B1 (en) * 2012-03-09 2014-05-27 Google Inc. Adaptively encoding a media stream with compound prediction
US9584802B2 (en) * 2012-04-13 2017-02-28 Texas Instruments Incorporated Reducing context coded and bypass coded bins to improve context adaptive binary arithmetic coding (CABAC) throughput
CN105007496B (en) * 2012-04-15 2019-05-10 三星电子株式会社 The method and apparatus that video is decoded
CA2807382A1 (en) * 2012-04-16 2013-10-16 Research In Motion Limited Modified contexts for level coding of video data
WO2013159112A1 (en) * 2012-04-20 2013-10-24 The Board Of Regents Of The University Of Texas System Systems and methods for simultaneous compression and encryption
KR20130122353A (en) * 2012-04-30 2013-11-07 삼성전자주식회사 Bitrate estimation device and bitrate estimation method thereof
ES2936333T3 (en) 2012-05-25 2023-03-16 Sun Patent Trust Video image coding method, video image decoding method, video image coding device, video image decoding device and video image encoding-decoding device
WO2013175736A1 (en) * 2012-05-25 2013-11-28 パナソニック株式会社 Video encoding method, video encoding device, video decoding method, video decoding device, and video encoding/decoding device
CA3217710A1 (en) 2012-05-25 2013-11-28 Sun Patent Trust Image encoding and decoding using context adaptive binary arithmetic coding with a bypass mode
EP4020990A1 (en) 2012-06-04 2022-06-29 Sun Patent Trust Video image encoding device, video image decoding device
MY192340A (en) 2012-06-11 2022-08-17 Samsung Electronics Co Ltd Method and apparatus for encoding videos sharing sao parameter according to color component
US11317101B2 (en) 2012-06-12 2022-04-26 Google Inc. Inter frame candidate selection for a video encoder
US9781447B1 (en) 2012-06-21 2017-10-03 Google Inc. Correlation based inter-plane prediction encoding and decoding
US9185414B1 (en) 2012-06-29 2015-11-10 Google Inc. Video encoding using variance
US9774856B1 (en) 2012-07-02 2017-09-26 Google Inc. Adaptive stochastic entropy coding
US9167268B1 (en) 2012-08-09 2015-10-20 Google Inc. Second-order orthogonal spatial intra prediction
US9015681B1 (en) 2012-08-10 2015-04-21 The Mathworks, Inc. Creating and using array creation routines or a family of routines separate from algorithmic code
US9146750B2 (en) * 2012-08-10 2015-09-29 The Mathworks, Inc. Mechanism for replacing an array creation routine within code
US9244667B2 (en) 2012-08-10 2016-01-26 The Mathworks, Inc. Mechanism for transforming between type-independent and type-specific code
US9344742B2 (en) 2012-08-10 2016-05-17 Google Inc. Transform-domain intra prediction
US9380298B1 (en) 2012-08-10 2016-06-28 Google Inc. Object-based intra-prediction
US9485515B2 (en) 2013-08-23 2016-11-01 Google Inc. Video coding using reference motion vectors
US9503746B2 (en) * 2012-10-08 2016-11-22 Google Inc. Determine reference motion vectors
FR2999760A1 (en) * 2012-12-17 2014-06-20 France Telecom METHOD FOR ENCODING AND DECODING IMAGES, CORRESPONDING ENCODING AND DECODING DEVICE AND COMPUTER PROGRAMS
US9628790B1 (en) 2013-01-03 2017-04-18 Google Inc. Adaptive composite intra prediction for image and video compression
EP2946553B1 (en) * 2013-01-16 2019-01-02 BlackBerry Limited Transform coefficient coding for context-adaptive binary entropy coding of video
JP2016513825A (en) * 2013-03-14 2016-05-16 マサチューセッツ インスティテュート オブ テクノロジー Safety communication method and apparatus
US9509998B1 (en) 2013-04-04 2016-11-29 Google Inc. Conditional predictive multi-symbol run-length coding
GB2513110A (en) * 2013-04-08 2014-10-22 Sony Corp Data encoding and decoding
US10277909B2 (en) * 2013-05-15 2019-04-30 Vid Scale, Inc. Single loop decoding based interlayer prediction
FR3005816B1 (en) * 2013-05-17 2019-11-29 Jean-Claude Colin METHOD FOR ENCODING, ESPECIALLY COMPRESSED IMAGES, IN PARTICULAR BY "RANGE CODER" OR ARITHMETIC COMPRESSION.
US9374578B1 (en) 2013-05-23 2016-06-21 Google Inc. Video coding using combined inter and intra predictors
US9247251B1 (en) 2013-07-26 2016-01-26 Google Inc. Right-edge extension for quad-tree intra-prediction
US9392288B2 (en) 2013-10-17 2016-07-12 Google Inc. Video coding using scatter-based scan tables
US9179151B2 (en) 2013-10-18 2015-11-03 Google Inc. Spatial proximity context entropy coding
US9609343B1 (en) 2013-12-20 2017-03-28 Google Inc. Video coding using compound prediction
US9558109B2 (en) * 2014-04-04 2017-01-31 Samsung Israel Research Corporation Method and apparatus for flash memory arithmetic encoding and decoding
US10142636B2 (en) 2014-06-09 2018-11-27 Sony Corporation Communication system with coding mechanism and method of operation thereof
US9337862B2 (en) 2014-06-09 2016-05-10 Tidal Systems, Inc. VLSI efficient Huffman encoding apparatus and method
US9854253B2 (en) * 2014-06-30 2017-12-26 Qualcomm Incorporated Method for motion vector difference (MVD) and intra block copy vector difference (BVD) coding of screen content video data
US11018795B2 (en) * 2014-09-29 2021-05-25 The Regents Of The University Of California Methods and apparatus for coding for interference network
WO2016056977A1 (en) * 2014-10-06 2016-04-14 Telefonaktiebolaget L M Ericsson (Publ) Coding and deriving quantization parameters
US20160100177A1 (en) * 2014-10-06 2016-04-07 Qualcomm Incorporated Non-uniform exponential-golomb codes for palette mode coding
US10069518B2 (en) * 2014-10-24 2018-09-04 Qualcomm Incorporated Uneven bit distributions for encoder parsing
US9781424B2 (en) * 2015-01-19 2017-10-03 Google Inc. Efficient context handling in arithmetic coding
GB2539239B (en) * 2015-06-10 2017-10-04 Gurulogic Microsystems Oy Encoders, decoders and methods utilizing mode symbols
WO2017041271A1 (en) * 2015-09-10 2017-03-16 Mediatek Singapore Pte. Ltd. Efficient context modeling for coding a block of data
US10158874B2 (en) * 2015-09-30 2018-12-18 Apple Inc. Parallel bypass and regular bin coding
US20170180757A1 (en) * 2015-12-18 2017-06-22 Blackberry Limited Binarizer selection for image and video coding
WO2017114450A1 (en) * 2015-12-31 2017-07-06 Mediatek Inc. Method and apparatus of prediction binary tree structure for video and image coding
US10708164B2 (en) * 2016-05-03 2020-07-07 Qualcomm Incorporated Binarizing secondary transform index
US9712830B1 (en) * 2016-09-15 2017-07-18 Dropbox, Inc. Techniques for image recompression
US10264264B2 (en) * 2016-09-24 2019-04-16 Apple Inc. Multi-bin decoding systems and methods
GB2556886B (en) 2016-11-23 2019-05-15 Imagination Tech Ltd Encoding and decoding variable length instructions
US9819359B1 (en) * 2016-12-11 2017-11-14 Microsoft Technology Licensing, Llc Multi-symbol, multi-format, parallel symbol decoder for hardware decompression engines
US10757412B2 (en) * 2017-01-03 2020-08-25 Avago Technologies International Sales Pte. Limited Architecture flexible binary arithmetic coding system
US10244261B2 (en) * 2017-01-26 2019-03-26 Google Llc Transform coefficient coding using level maps
EP3358754A1 (en) * 2017-02-02 2018-08-08 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Antenna array codebook with beamforming coefficients adapted to an arbitrary antenna response of the antenna array
US10595033B2 (en) 2017-02-15 2020-03-17 Sony Corporation Variable length coding of header data for image compression
US10225562B1 (en) * 2017-08-21 2019-03-05 Google Llc Embedding information about EOB positions
US10735736B2 (en) * 2017-08-29 2020-08-04 Google Llc Selective mixing for entropy coding in video compression
US10791341B2 (en) * 2017-10-10 2020-09-29 Qualcomm Incorporated Binary arithmetic coding with progressive modification of adaptation parameters
CN107819573A (en) * 2017-10-17 2018-03-20 东北大学 High dimension safety arithmetic coding method
WO2019111012A1 (en) * 2017-12-06 2019-06-13 V-Nova International Ltd Method and apparatus for decoding a received set of encoded data
DE102018108419A1 (en) * 2018-04-10 2019-10-10 Carl Zeiss Microscopy Gmbh Methods and apparatus for compression and decompression of drive curves
EP3562162A1 (en) * 2018-04-27 2019-10-30 InterDigital VC Holdings, Inc. Method and apparatus for video encoding and decoding based on neural network implementation of cabac
CN112106363A (en) * 2018-05-10 2020-12-18 夏普株式会社 System and method for performing binary arithmetic coding in video coding
US20230024560A1 (en) * 2018-05-10 2023-01-26 Sharp Kabushiki Kaisha Systems and methods for performing binary arithmetic coding in video coding
EP3588800A1 (en) 2018-06-29 2020-01-01 FRAUNHOFER-GESELLSCHAFT zur Förderung der angewandten Forschung e.V. Antenna array codebook with beamforming coefficients adapted to an arbitrary antenna response of the antenna array
US11483575B2 (en) * 2018-08-24 2022-10-25 Hfi Innovation Inc. Coding transform coefficients with throughput constraints
US11006150B2 (en) 2018-09-24 2021-05-11 Tencent America LLC Method and apparatus for video coding
KR102613155B1 (en) * 2018-09-24 2023-12-13 프라운호퍼-게젤샤프트 추르 푀르데룽 데어 안제반텐 포르슝 에 파우 Efficient coding of transform coefficients using or suitable for a combination with dependent scalar quantization
CN109474281B (en) * 2018-09-30 2022-07-08 湖南瑞利德信息科技有限公司 Data encoding and decoding method and device
US10402172B1 (en) 2019-02-28 2019-09-03 Qrypt, Inc. Multi-source entropy and randomness aggregation and distribution network
GB2619640A (en) * 2019-07-05 2023-12-13 V Nova Int Ltd Quantization of residuals in video coding
CN112398484B (en) * 2019-08-15 2024-04-23 华为技术有限公司 Coding method and related equipment
US11095311B2 (en) * 2019-11-27 2021-08-17 Qualcomm Incorporated Quantization codeword selection for low cost parity checking
AR121126A1 (en) 2020-02-29 2022-04-20 Beijing Bytedance Network Tech Co Ltd SIGNALING OF SYNTAX ELEMENTS FOR INDICATION OF REFERENCE IMAGE
US11115049B1 (en) * 2020-08-24 2021-09-07 Innogrit Technologies Co., Ltd. Hardware friendly data decompression
US11924437B2 (en) * 2022-03-28 2024-03-05 Bitmovin Gmbh Variable framerate encoding using content-aware framerate prediction for high framerate videos
US20240015331A1 (en) * 2022-07-05 2024-01-11 Tencent America LLC Systems and methods including syntax design for multi-symbol arithmetic coding
CN116684003B (en) * 2023-07-27 2023-10-24 南京中科齐信科技有限公司 Quantum communication-based railway line air-ground comprehensive monitoring method and system

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5381145A (en) * 1993-02-10 1995-01-10 Ricoh Corporation Method and apparatus for parallel decoding and encoding of data
US5717394A (en) * 1993-02-10 1998-02-10 Ricoh Company Ltd. Method and apparatus for encoding and decoding data
US6943710B2 (en) * 2002-05-02 2005-09-13 Faunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Method and arrangement for arithmetic encoding and decoding binary states and a corresponding computer program and a corresponding computer-readable storage medium
WO2008087466A1 (en) * 2007-01-17 2008-07-24 Rosen Stefanov Run-length encoding of binary sequences followed by two independent compressions

Family Cites Families (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5357250A (en) * 1992-11-20 1994-10-18 International Business Machines Corporation Adaptive computation of symbol probabilities in n-ary strings
JP3230933B2 (en) * 1994-08-19 2001-11-19 株式会社リコー Data decompression device, data decompression method, decoding device, decoding method, encoding device, and entropy decoder
CA2156889C (en) 1994-09-30 1999-11-02 Edward L. Schwartz Method and apparatus for encoding and decoding data
WO2003092302A1 (en) * 2002-04-26 2003-11-06 Nec Corporation Moving image transferring system, moving image encoding apparatus, moving image decoding apparatus, and moving image transferring program
EP1467491B1 (en) * 2002-05-02 2007-01-24 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Arithmetical coding of transform coefficients
JP3871995B2 (en) * 2002-10-25 2007-01-24 富士ゼロックス株式会社 Encoding device and decoding device
KR100636229B1 (en) * 2005-01-14 2006-10-19 학교법인 성균관대학 Method and apparatus for adaptive entropy encoding and decoding for scalable video coding
CA2590705A1 (en) 2005-01-14 2006-07-20 Sungkyunkwan University Methods of and apparatuses for adaptive entropy encoding and adaptive entropy decoding for scalable video encoding
WO2007065351A1 (en) * 2005-12-05 2007-06-14 Huawei Technologies Co., Ltd. Binarizing method and device thereof
US7498960B2 (en) * 2007-04-19 2009-03-03 Analog Devices, Inc. Programmable compute system for executing an H.264 binary decode symbol instruction
US7990297B1 (en) * 2010-02-26 2011-08-02 Research In Motion Limited Encoding and decoding methods and devices employing dual codesets

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5381145A (en) * 1993-02-10 1995-01-10 Ricoh Corporation Method and apparatus for parallel decoding and encoding of data
US5717394A (en) * 1993-02-10 1998-02-10 Ricoh Company Ltd. Method and apparatus for encoding and decoding data
US6943710B2 (en) * 2002-05-02 2005-09-13 Faunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Method and arrangement for arithmetic encoding and decoding binary states and a corresponding computer program and a corresponding computer-readable storage medium
WO2008087466A1 (en) * 2007-01-17 2008-07-24 Rosen Stefanov Run-length encoding of binary sequences followed by two independent compressions

Non-Patent Citations (8)

* Cited by examiner, † Cited by third party
Title
AL-HASHIMI B M ET AL: "Variable-length input huffman coding for system-on-a-chip test", IEEE TRANSACTIONS ON COMPUTER AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, IEEE SERVICE CENTER, PISCATAWAY, NJ, US, vol. 22, no. 6, 1 June 2003 (2003-06-01), pages 783 - 796, XP011096901, ISSN: 0278-0070, DOI: 10.1109/TCAD.2003.811451 *
BORMANS J ET AL: "Video coding with H.264/AVC: tools, performance, and complexity", IEEE CIRCUITS AND SYSTEMS MAGAZINE, IEEE SERVICE CENTER, NEW YORK, NY, US, vol. 4, no. 1, 1 January 2004 (2004-01-01), pages 7 - 28, XP011111220, ISSN: 1531-636X, DOI: DOI:10.1109/MCAS.2004.1286980 *
DAVID SALOMON: "Variable-length Codes for Data Compression", 31 December 2007, SPRINGER VERLAG, ISBN: 978-1-84628-958-3, pages: 1 - 7, XP009151666 *
HE D ET AL: "Video Coding Technology Proposal by RIM", no. JCTVC-A120, 15 April 2010 (2010-04-15), pages 1 - 29, XP002607631, Retrieved from the Internet <URL:http://wftp3.itu.int/av-arch/jctvc-site/2010_04_A_Dresden/> [retrieved on 20101028] *
KAR-MING CHEUNG: "Probability quantization for multiplication-free binary arithmetic coding", INFORMATION THEORY, 1995. PROCEEDINGS., 1995 IEEE INTERNATIONAL SYMPOS IUM ON WHISTLER, BC, CANADA 17-22 SEPT. 1995, NEW YORK, NY, USA,IEEE, US, 17 September 1995 (1995-09-17), pages 376, XP010192389, ISBN: 978-0-7803-2453-4, DOI: DOI:10.1109/ISIT.1995.550363 *
MERHAV N ET AL: "Modeling and low-complexity adaptive coding for image prediction residuals", PROCEEDINGS OF THE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP) LAUSANNE, SEPT. 16 - 19, 1996; [PROCEEDINGS OF THE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP)], NEW YORK, IEEE, US, vol. 1, 16 September 1996 (1996-09-16), pages 353 - 356, XP010202667, ISBN: 978-0-7803-3259-1, DOI: 10.1109/ICIP.1996.560829 *
PASTUSZAK G: "A High-Performance Architecture of the Double-Mode Binary Coder for H.264.AVC", IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, IEEE SERVICE CENTER, PISCATAWAY, NJ, US, vol. 18, no. 7, 1 July 2008 (2008-07-01), pages 949 - 960, XP011225651, ISSN: 1051-8215, DOI: DOI:10.1109/TCSVT.2008.920743 *
WINKEN (FRAUNHOFER HHI) M ET AL: "Video coding technology proposal by Fraunhofer HHI", 1. JCT-VC MEETING; 15-4-2010 - 23-4-2010; DRESDEN; (JOINT COLLABORATIVE TEAM ON VIDEO CODING OF ISO/IEC JTC1/SC29/WG11 AND ITU-T SG.16 ); URL: HTTP://WFTP3.ITU.INT/AV-ARCH/JCTVC-SITE/,, no. XP030007556, 24 April 2010 (2010-04-24), XP030007557 *

Cited By (153)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2015039231A (en) * 2010-04-23 2015-02-26 エム アンド ケー ホールディングス インコーポレイテッド Device and method for encoding video
JP2014504077A (en) * 2010-12-03 2014-02-13 クゥアルコム・インコーポレイテッド Individual coding of the position of the last significant coefficient of a video block in video coding
US11330272B2 (en) 2010-12-22 2022-05-10 Qualcomm Incorporated Using a most probable scanning order to efficiently code scanning order information for a video block in video coding
US11006114B2 (en) 2011-03-08 2021-05-11 Velos Media, Llc Coding of transform coefficients for video coding
US10499059B2 (en) 2011-03-08 2019-12-03 Velos Media, Llc Coding of transform coefficients for video coding
US10397577B2 (en) 2011-03-08 2019-08-27 Velos Media, Llc Inverse scan order for significance map coding of transform coefficients in video coding
US9338449B2 (en) 2011-03-08 2016-05-10 Qualcomm Incorporated Harmonized scan order for coding transform coefficients in video coding
US9197890B2 (en) 2011-03-08 2015-11-24 Qualcomm Incorporated Harmonized scan order for coding transform coefficients in video coding
JP2014511657A (en) * 2011-03-08 2014-05-15 クゥアルコム・インコーポレイテッド Coding transform coefficients for video coding
US11405616B2 (en) 2011-03-08 2022-08-02 Qualcomm Incorporated Coding of transform coefficients for video coding
JP2014509158A (en) * 2011-03-08 2014-04-10 クゥアルコム・インコーポレイテッド Coding transform coefficients for video coding
JP2014511655A (en) * 2011-03-08 2014-05-15 クゥアルコム・インコーポレイテッド Coding transform coefficients for video coding
JP2014511656A (en) * 2011-03-08 2014-05-15 クゥアルコム・インコーポレイテッド Coding transform coefficients for video coding
US9930371B2 (en) 2011-06-16 2018-03-27 Ge Video Compression, Llc Entropy coding of motion vector differences
US9743090B2 (en) 2011-06-16 2017-08-22 Ge Video Compression, Llc Entropy coding of motion vector differences
KR102631388B1 (en) 2011-06-16 2024-01-31 지이 비디오 컴프레션, 엘엘씨 Context initialization in entropy coding
US11838511B2 (en) 2011-06-16 2023-12-05 Ge Video Compression, Llc Entropy coding supporting mode switching
EP4277276A1 (en) 2011-06-16 2023-11-15 GE Video Compression, LLC Entropy coding of motion vector differences
US11533485B2 (en) 2011-06-16 2022-12-20 Ge Video Compression, Llc Entropy coding of motion vector differences
US11516474B2 (en) 2011-06-16 2022-11-29 Ge Video Compression, Llc Context initialization in entropy coding
KR20220122800A (en) * 2011-06-16 2022-09-02 지이 비디오 컴프레션, 엘엘씨 Context initialization in entropy coding
KR102437504B1 (en) 2011-06-16 2022-08-29 지이 비디오 컴프레션, 엘엘씨 Context initialization in entropy coding
US10021393B2 (en) 2011-06-16 2018-07-10 Ge Video Compression, Llc Entropy coding of motion vector differences
WO2012172113A1 (en) 2011-06-16 2012-12-20 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Entropy coding of motion vector differences
EP4033672A1 (en) 2011-06-16 2022-07-27 GE Video Compression, LLC Context initialization in entropy coding
US11277614B2 (en) 2011-06-16 2022-03-15 Ge Video Compression, Llc Entropy coding supporting mode switching
EP3930330A1 (en) 2011-06-16 2021-12-29 GE Video Compression, LLC Entropy coding of motion vector differences
US11012695B2 (en) 2011-06-16 2021-05-18 Ge Video Compression, Llc Context initialization in entropy coding
KR20200138444A (en) * 2011-06-16 2020-12-09 지이 비디오 컴프레션, 엘엘씨 Context initialization in entropy coding
KR102187981B1 (en) 2011-06-16 2020-12-07 지이 비디오 컴프레션, 엘엘씨 Context initialization in entropy coding
US10819982B2 (en) 2011-06-16 2020-10-27 Ge Video Compression, Llc Entropy coding supporting mode switching
US10645388B2 (en) 2011-06-16 2020-05-05 Ge Video Compression, Llc Context initialization in entropy coding
US10630987B2 (en) 2011-06-16 2020-04-21 Ge Video Compression, Llc Entropy coding supporting mode switching
US10630988B2 (en) 2011-06-16 2020-04-21 Ge Video Compression, Llc Entropy coding of motion vector differences
KR20190136116A (en) * 2011-06-16 2019-12-09 지이 비디오 컴프레션, 엘엘씨 Context initialization in entropy coding
KR102052471B1 (en) 2011-06-16 2019-12-05 지이 비디오 컴프레션, 엘엘씨 Context initialization in entropy coding
US10440364B2 (en) 2011-06-16 2019-10-08 Ge Video Compression, Llc Context initialization in entropy coding
US10432940B2 (en) 2011-06-16 2019-10-01 Ge Video Compression, Llc Entropy coding of motion vector differences
KR101619333B1 (en) 2011-06-16 2016-05-18 지이 비디오 컴프레션, 엘엘씨 Context initialization in entropy coding
US10432939B2 (en) 2011-06-16 2019-10-01 Ge Video Compression, Llc Entropy coding supporting mode switching
US10425644B2 (en) 2011-06-16 2019-09-24 Ge Video Compression, Llc Entropy coding of motion vector differences
EP3343781A1 (en) 2011-06-16 2018-07-04 GE Video Compression, LLC Context initialization in entropy coding
US10306232B2 (en) 2011-06-16 2019-05-28 Ge Video Compression, Llc Entropy coding of motion vector differences
US10298964B2 (en) 2011-06-16 2019-05-21 Ge Video Compression, Llc Entropy coding of motion vector differences
EP3471415A1 (en) 2011-06-16 2019-04-17 GE Video Compression, LLC Entropy coding of motion vector differences
US10230954B2 (en) 2011-06-16 2019-03-12 Ge Video Compression, Llp Entropy coding of motion vector differences
US10148962B2 (en) 2011-06-16 2018-12-04 Ge Video Compression, Llc Entropy coding of motion vector differences
KR101882848B1 (en) 2011-06-16 2018-07-27 지이 비디오 컴프레션, 엘엘씨 Context initialization in entropy coding
US10063858B2 (en) 2011-06-16 2018-08-28 Ge Video Compression, Llc Entropy coding of motion vector differences
US10057603B2 (en) 2011-06-16 2018-08-21 Ge Video Compression, Llc Entropy coding supporting mode switching
KR20180086522A (en) * 2011-06-16 2018-07-31 지이 비디오 컴프레션, 엘엘씨 Context initialization in entropy coding
US10313672B2 (en) 2011-06-16 2019-06-04 Ge Video Compression, Llc Entropy coding supporting mode switching
US9973761B2 (en) 2011-06-16 2018-05-15 Ge Video Compression, Llc Context initialization in entropy coding
US9936227B2 (en) 2011-06-16 2018-04-03 Ge Video Compression, Llc Entropy coding of motion vector differences
US9628827B2 (en) 2011-06-16 2017-04-18 Ge Video Compression, Llc Context initialization in entropy coding
KR101730587B1 (en) 2011-06-16 2017-04-26 지이 비디오 컴프레션, 엘엘씨 Context initialization in entropy coding
KR20170047406A (en) * 2011-06-16 2017-05-04 지이 비디오 컴프레션, 엘엘씨 Context initialization in entropy coding
WO2012172114A1 (en) 2011-06-16 2012-12-20 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Context initialization in entropy coding
US9930370B2 (en) 2011-06-16 2018-03-27 Ge Video Compression, Llc Entropy coding of motion vector differences
US9918104B2 (en) 2011-06-16 2018-03-13 Ge Video Compression, Llc Entropy coding of motion vector differences
US9686568B2 (en) 2011-06-16 2017-06-20 Ge Video Compression, Llc Context initialization in entropy coding
US9768804B1 (en) 2011-06-16 2017-09-19 Ge Video Compression, Llc Context initialization in entropy coding
US9762913B2 (en) 2011-06-16 2017-09-12 Ge Video Compression, Llc Context initialization in entropy coding
US10091510B2 (en) 2011-06-28 2018-10-02 Samsung Electronics Co., Ltd. Method and apparatus for coding video and method and apparatus for decoding video accompanied with arithmetic coding
US9668001B2 (en) 2011-06-28 2017-05-30 Samsung Electronics Co., Ltd. Method and apparatus for coding video and method and apparatus for decoding video accompanied with arithmetic coding
US9554157B2 (en) 2011-06-28 2017-01-24 Samsung Electronics Co., Ltd. Method and apparatus for coding video and method and apparatus for decoding video accompanied with arithmetic coding
US9565455B2 (en) 2011-06-28 2017-02-07 Samsung Electronics Co., Ltd. Method and apparatus for coding video and method and apparatus for decoding video accompanied with arithmetic coding
US10547842B2 (en) 2011-06-28 2020-01-28 Samsung Electronics Co., Ltd. Method and apparatus for coding video and method and apparatus for decoding video accompanied with arithmetic coding
JP2015149774A (en) * 2011-06-28 2015-08-20 サムスン エレクトロニクス カンパニー リミテッド Method and apparatus for coding video and method and apparatus for decoding video accompanied with arithmetic coding
JP2015149773A (en) * 2011-06-28 2015-08-20 サムスン エレクトロニクス カンパニー リミテッド Method and apparatus for coding video and method and apparatus for decoding video accompanied with arithmetic coding
US9258571B2 (en) 2011-06-28 2016-02-09 Samsung Electronics Co., Ltd. Method and apparatus for coding video and method and apparatus for decoding video accompanied with arithmetic coding
JP2015149770A (en) * 2011-06-28 2015-08-20 サムスン エレクトロニクス カンパニー リミテッド Method and apparatus for coding video and method and apparatus for decoding video accompanied with arithmetic coding
JP2015149771A (en) * 2011-06-28 2015-08-20 サムスン エレクトロニクス カンパニー リミテッド Method and apparatus for coding video and method and apparatus for decoding video accompanied with arithmetic coding
JP2015149772A (en) * 2011-06-28 2015-08-20 サムスン エレクトロニクス カンパニー リミテッド Method and apparatus for coding video and method and apparatus for decoding video accompanied with arithmetic coding
US9516349B2 (en) 2011-07-12 2016-12-06 Futurewei Technologies, Inc. Pixel-based intra prediction for coding in HEVC
US10244262B2 (en) 2011-07-12 2019-03-26 Futurewei Technologies, Inc. Pixel-based intra prediction for coding in HEVC
DE102011085451B3 (en) * 2011-10-28 2013-04-11 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Encoding device for encoding multiple symbols in encoded bitstream, has multiplexer, which is arranged to attach block coded bits with each other for encoded bitstream so that series is predetermined
US9307263B2 (en) 2011-10-31 2016-04-05 Samsung Electronics Co., Ltd. Method and apparatus for determining a context model for transform coefficient level entropy encoding and decoding
EP2773109A4 (en) * 2011-10-31 2015-05-06 Samsung Electronics Co Ltd Method and apparatus for determining a context model for transform coefficient level entropy encoding and decoding
EP3145197A1 (en) * 2011-10-31 2017-03-22 Samsung Electronics Co., Ltd. Method and apparatus for determining a context model for transform coefficient level entropy encoding and decoding
EP3166317A1 (en) * 2011-10-31 2017-05-10 Samsung Electronics Co., Ltd. Method and apparatus for determining a context model for transform coefficient level entropy encoding and decoding
CN107529060B (en) * 2011-10-31 2020-04-21 三星电子株式会社 Method and apparatus for determining context model and computer recordable medium
EP3846479A1 (en) * 2011-10-31 2021-07-07 Samsung Electronics Co., Ltd. Determining a context model for transform coefficient level entropy encoding
CN107529060A (en) * 2011-10-31 2017-12-29 三星电子株式会社 Determine the method and apparatus and computer recordable media of context model
US9521430B2 (en) 2011-10-31 2016-12-13 Samsung Electronics Co., Ltd. Method and apparatus for determining a context model for transform coefficient level entropy encoding and decoding
EP3399760A1 (en) * 2011-10-31 2018-11-07 Samsung Electronics Co., Ltd. Method and apparatus for determining a context model for transform coefficient level entropy encoding and decoding
US9532072B2 (en) 2011-10-31 2016-12-27 Samsung Electronics Co., Ltd. Method and apparatus for determining a context model for transform coefficient level entropy encoding and decoding
AU2015201781B2 (en) * 2011-10-31 2017-01-19 Samsung Electronics Co., Ltd. Method and apparatus for determining a context model for transform coefficient level entropy encoding and decoding
US9503752B2 (en) 2011-10-31 2016-11-22 Samsung Electronics Co., Ltd. Method and apparatus for determining a context model for transform coefficient level entropy encoding and decoding
US9414088B2 (en) 2011-10-31 2016-08-09 Samsung Electronics Co., Ltd. Method and apparatus for determining a context model for transform coefficient level entropy encoding and decoding
CN103096062A (en) * 2011-11-01 2013-05-08 珠海扬智电子科技有限公司 Arithmetic decoding method
JP2014535222A (en) * 2011-11-03 2014-12-25 クゥアルコム・インコーポレイテッドQualcomm Incorporated Initializing context states and probabilities for context-adaptive entropy coding
US9484952B2 (en) 2011-11-03 2016-11-01 Qualcomm Incorporated Context state and probability initialization for context adaptive entropy coding
US9253508B2 (en) 2011-11-04 2016-02-02 Futurewei Technologies, Inc. Differential pulse code modulation intra prediction for high efficiency video coding
US9503750B2 (en) 2011-11-04 2016-11-22 Futurewei Technologies, Inc. Binarization of prediction residuals for lossless video coding
WO2013067436A1 (en) * 2011-11-04 2013-05-10 Huawei Technologies Co., Ltd. Binarization of prediction residuals for lossless video coding
US9813733B2 (en) 2011-11-04 2017-11-07 Futurewei Technologies, Inc. Differential pulse code modulation intra prediction for high efficiency video coding
WO2013070487A1 (en) * 2011-11-08 2013-05-16 Sony Corporation Binarisation of last position for higher throughput
US9154792B2 (en) 2011-11-08 2015-10-06 Qualcomm Incorporated Progressive coding of position of last significant coefficient
US9357185B2 (en) 2011-11-08 2016-05-31 Qualcomm Incorporated Context optimization for last significant coefficient position coding
JP2014195269A (en) * 2011-11-19 2014-10-09 Blackberry Ltd Multilevel effectiveness mapping scanning
RU2611992C2 (en) * 2012-01-17 2017-03-01 Квэлкомм Инкорпорейтед Increasing throughput capacity for cabac coefficients level encoding
US9654772B2 (en) 2012-01-19 2017-05-16 Qualcomm Incorporated Context adaptive entropy coding with a reduced initialization value set
JP2017225132A (en) * 2012-01-19 2017-12-21 クゥアルコム・インコーポレイテッドQualcomm Incorporated Context adaptive entropy coding with reduced initialization value set
JP2015504292A (en) * 2012-01-19 2015-02-05 クゥアルコム・インコーポレイテッドQualcomm Incorporated Context-adaptive entropy coding with reduced initial value set
US9712844B2 (en) 2012-01-20 2017-07-18 Ge Video Compression, Llc Transform coefficient coding
RU2641235C2 (en) * 2012-01-20 2018-01-16 ДжиИ Видео Компрешн, ЭлЭлСи Coding of transformation coefficients
US10757447B2 (en) 2012-01-20 2020-08-25 Ge Video Compression, Llc Transform coefficient coding
CN103460701A (en) * 2012-01-20 2013-12-18 索尼公司 Complexity reduction of significance map coding
US11968395B2 (en) 2012-01-20 2024-04-23 Ge Video Compression, Llc Transform coefficient coding
CN105847827B (en) * 2012-01-20 2019-03-08 索尼公司 The complexity of availability graph code reduces
US11616982B2 (en) 2012-01-20 2023-03-28 Ge Video Compression, Llc Transform coefficient coding
US10462487B2 (en) 2012-01-20 2019-10-29 Ge Video Compression, Llc Transform coefficient coding
US10582219B2 (en) 2012-01-20 2020-03-03 Ge Video Compression, Llc Transform coefficient coding
US10271068B2 (en) 2012-01-20 2019-04-23 Ge Video Compression, Llc Transform coefficient coding
US10045049B2 (en) 2012-01-20 2018-08-07 Ge Video Compression Llc Transform coefficient coding
JP2015510319A (en) * 2012-01-20 2015-04-02 フラウンホッファー−ゲゼルシャフト ツァ フェルダールング デァ アンゲヴァンテン フォアシュンク エー.ファオ Transform coefficient coding
RU2708967C2 (en) * 2012-01-20 2019-12-12 ДжиИ Видео Компрешн, ЭлЭлСи Encoding conversion coefficients
CN105847827A (en) * 2012-01-20 2016-08-10 索尼公司 Complexity reduction of significance map coding
CN105027560A (en) * 2012-01-21 2015-11-04 摩托罗拉移动有限责任公司 Method of determining binary codewords for transform coefficients
US9237344B2 (en) 2012-03-22 2016-01-12 Qualcomm Incorporated Deriving context for last position coding for video coding
RU2668390C1 (en) * 2012-03-26 2018-09-28 ДжейВиСи КЕНВУД КОРПОРЕЙШН Decoding image device, method for decoding images and long-term computer readable record medium
RU2679215C9 (en) * 2012-03-26 2019-08-21 ДжейВиСи КЕНВУД КОРПОРЕЙШН Picture coding device, picture coding method and non-transitory computer readable recording medium
RU2679216C1 (en) * 2012-03-26 2019-02-06 ДжейВиСи КЕНВУД КОРПОРЕЙШН Decoding image device, method for decoding images and long-term computer readable record medium
RU2639266C1 (en) * 2012-03-26 2017-12-20 ДжейВиСи КЕНВУД КОРПОРЕЙШН Device for decoding images, method for decoding images and long-term record medium
RU2639266C9 (en) * 2012-03-26 2018-05-14 ДжейВиСи КЕНВУД КОРПОРЕЙШН Picture decoding device, picture decoding method and non-transitory recording medium
RU2679215C1 (en) * 2012-03-26 2019-02-06 ДжейВиСи КЕНВУД КОРПОРЕЙШН Image encoding device, image encoding method and long-term computer-readable recording medium
RU2679216C9 (en) * 2012-03-26 2019-08-08 ДжейВиСи КЕНВУД КОРПОРЕЙШН Picture decoding device, picture decoding method and non-transitory computer readable recording medium
RU2689808C9 (en) * 2012-04-13 2019-11-05 ДжейВиСи КЕНВУД КОРПОРЕЙШН Picture decoding device, picture decoding method and reception device
RU2713848C1 (en) * 2012-04-13 2020-02-07 ДжейВиСи КЕНВУД КОРПОРЕЙШН Picture coding device, picture coding method and transmission device
RU2705650C9 (en) * 2012-04-13 2020-12-07 ДжейВиСи КЕНВУД КОРПОРЕЙШН Picture decoding device, picture decoding method and reception device
RU2668399C9 (en) * 2012-04-13 2019-06-19 ДжейВиСи КЕНВУД КОРПОРЕЙШН Picture decoding device, picture decoding method and reception device
RU2668399C1 (en) * 2012-04-13 2018-09-28 ДжейВиСи КЕНВУД КОРПОРЕЙШН Image decoding device, method of decoding image and receiving device
RU2633175C9 (en) * 2012-04-13 2018-03-19 ДжейВиСи КЕНВУД КОРПОРЕЙШН Picture coding device, picture coding method and picture coding program, and picture decoding device, picture decoding method and picture decoding program
WO2013158642A1 (en) * 2012-04-16 2013-10-24 Qualcomm Incorporated Transform coefficient coding
US9313498B2 (en) 2012-04-16 2016-04-12 Qualcomm Incorporated Sign hiding techniques for quantized transform coefficients in video coding
US9124872B2 (en) 2012-04-16 2015-09-01 Qualcomm Incorporated Coefficient groups and coefficient coding for coefficient scans
CN104247420A (en) * 2012-04-16 2014-12-24 高通股份有限公司 Transform coefficient coding
US9621921B2 (en) 2012-04-16 2017-04-11 Qualcomm Incorporated Coefficient groups and coefficient coding for coefficient scans
WO2014049982A1 (en) * 2012-09-28 2014-04-03 三菱電機株式会社 Video encoding device, video decoding device, video encoding method and video decoding method
CN104737542A (en) * 2013-01-30 2015-06-24 英特尔公司 Content adaptive entropy coding for next generation video
CN104104949B (en) * 2013-04-12 2018-01-09 浙江大学 A kind of entropy encoding/decoding method and apparatus of directional information syntactic element
CN104104949A (en) * 2013-04-12 2014-10-15 浙江大学 Entropy encoding and decoding method of direction information syntax element and device
RU2630750C1 (en) * 2014-02-20 2017-09-12 Гурулоджик Микросистемс Ой Device and method for encoding and decoding initial data
US10003356B2 (en) 2014-02-20 2018-06-19 Gurulogic Microsystems Oy Devices and methods of source-encoding and decoding of data
US10340946B2 (en) 2015-11-01 2019-07-02 Gurulogic Microsystems Oy Encoders, decoders, and methods
CN106941610A (en) * 2017-04-25 2017-07-11 西安电子科技大学 Based on the two-value ROI mask coding methods for improving block encription
CN106941610B (en) * 2017-04-25 2019-12-24 西安电子科技大学 Binary ROI mask coding method based on improved block coding
WO2020145698A1 (en) * 2019-01-13 2020-07-16 엘지전자 주식회사 Method and device for processing video signal
US11856196B2 (en) 2019-01-13 2023-12-26 Lg Electronics Inc. Method and device for processing video signal
CN114731407A (en) * 2019-09-24 2022-07-08 弗劳恩霍夫应用研究促进协会 Arithmetic encoder and decoder, video encoder and decoder, method for encoding or decoding symbols and method for encoding or decoding video content
CN114731446A (en) * 2019-09-24 2022-07-08 弗劳恩霍夫应用研究促进协会 Coding concept for sequences of information values
WO2022000089A1 (en) * 2020-06-30 2022-01-06 National Research Council Of Canada Vector space model for form data extraction

Also Published As

Publication number Publication date
JP2013528025A (en) 2013-07-04
CN103119849A (en) 2013-05-22
US20130027230A1 (en) 2013-01-31
TWI514779B (en) 2015-12-21
CN103119849B (en) 2017-06-16
US8907823B2 (en) 2014-12-09
EP2559166B1 (en) 2017-11-01
EP2559166A1 (en) 2013-02-20
PL2559166T3 (en) 2018-04-30
TW201143307A (en) 2011-12-01
JP5676744B2 (en) 2015-02-25
HUE037656T2 (en) 2018-09-28

Similar Documents

Publication Publication Date Title
EP2559166B1 (en) Probability interval partioning encoder and decoder
US8947273B2 (en) Entropy coding
US10826524B2 (en) Entropy encoding and decoding scheme

Legal Events

Date Code Title Description
WWE Wipo information: entry into national phase

Ref document number: 201180029224.X

Country of ref document: CN

121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 11714643

Country of ref document: EP

Kind code of ref document: A1

WWE Wipo information: entry into national phase

Ref document number: 2013504210

Country of ref document: JP

NENP Non-entry into the national phase

Ref country code: DE

REEP Request for entry into the european phase

Ref document number: 2011714643

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 2011714643

Country of ref document: EP