US20080175502A1 - Method and system for providing arithmetic code normalization and byte construction - Google Patents

Method and system for providing arithmetic code normalization and byte construction Download PDF

Info

Publication number
US20080175502A1
US20080175502A1 US11/625,417 US62541707A US2008175502A1 US 20080175502 A1 US20080175502 A1 US 20080175502A1 US 62541707 A US62541707 A US 62541707A US 2008175502 A1 US2008175502 A1 US 2008175502A1
Authority
US
United States
Prior art keywords
data
bits
input
variable
subsets
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US11/625,417
Other versions
US7397402B1 (en
Inventor
Yendo Hu
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Google Technology Holdings LLC
Original Assignee
General Instrument Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by General Instrument Corp filed Critical General Instrument Corp
Priority to US11/625,417 priority Critical patent/US7397402B1/en
Assigned to GENERAL INSTRUMENT CORPORATION reassignment GENERAL INSTRUMENT CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: HU, YENDO
Application granted granted Critical
Publication of US7397402B1 publication Critical patent/US7397402B1/en
Publication of US20080175502A1 publication Critical patent/US20080175502A1/en
Assigned to GENERAL INSTRUMENT HOLDINGS, INC. reassignment GENERAL INSTRUMENT HOLDINGS, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: GENERAL INSTRUMENT CORPORATION
Assigned to MOTOROLA MOBILITY LLC reassignment MOTOROLA MOBILITY LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: GENERAL INSTRUMENT HOLDINGS, INC.
Assigned to Google Technology Holdings LLC reassignment Google Technology Holdings LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: MOTOROLA MOBILITY LLC
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H03ELECTRONIC CIRCUITRY
    • H03MCODING; DECODING; CODE CONVERSION IN GENERAL
    • H03M7/00Conversion of a code where information is represented by a given sequence or number of digits to a code where the same, similar or subset of information is represented by a different sequence or number of digits
    • H03M7/30Compression; Expansion; Suppression of unnecessary data, e.g. redundancy reduction
    • H03M7/40Conversion to or from variable length codes, e.g. Shannon-Fano code, Huffman code, Morse code
    • H03M7/4006Conversion to or from arithmetic code

Definitions

  • This disclosure generally relates to the field of video data processing. More particularly, the disclosure relates to Context Adaptive Binary Arithmetic Coding (“CABAC”) for digital video encoders.
  • CABAC Context Adaptive Binary Arithmetic Coding
  • Video signals generally include data corresponding to one or more video frames.
  • Each video frame is composed of an array of picture elements, which are called pixels.
  • a typical color video frame having a standard resolution may be composed of over several hundreds of thousands of pixels, which are arranged in arrays of blocks.
  • Each pixel is characterized by pixel data indicative of a hue (predominant color), saturation (color intensity), and luminance (color brightness).
  • the hue and saturation characteristics may be referred to as the chrominance.
  • the pixel data includes chrominance and luminance. Therefore, the pixel data may be represented by groups of four luminance pixel blocks and two chrominance pixel blocks. These groups are called macroblocks (“MBs”).
  • MBs macroblocks
  • the video frame As a video frame generally includes many pixels, the video frame also includes a large number of MBs. Thus, digital signals representing a sequence of video frame data, which usually include many video frames, have a large number of bits. However, the available storage space and bandwidth for transmitting these digital signals is limited. Therefore, compression processes are used to more efficiently transmit or store video data.
  • VLC variable length coding
  • Huffman Huffman coding
  • blocks of symbols coding arithmetic coding
  • CABAC Context Adaptive Binary Arithmetic Coding
  • CABAC techniques are capable of losslessly compressing syntax elements in a video stream using the probabilities of syntax elements in a given context.
  • the CABAC process will take in syntax elements representing all elements within a macroblock. Further, the CABAC process constructs a compress bit sequence by building out the following structure: the sequential set of fields for the macroblock based on the chosen macroblock configuration, the specific syntax element type and value for each of the fields within this field sequence, and the context address for each of the syntax elements.
  • the CABAC process will then perform binarization of the syntax elements, update the context weights, arithmetically encode the binarizations of syntax elements (“bins”), and subsequently pack the bits into bytes through the syntax element processing component.
  • the components of the CABAC process include: the CABAC weight initialization mode selection module, the macroblock syntax sequence generator, the binarization engine, the context address generator, the context weight update engine, the arithmetic coder, the bit packetizer, and the Network Abstraction Layer (“NAL”) header generator.
  • the CABAC engine within a video encoder may accomplish two goals within the encoding process: (1) to carry out compressed data resource prediction for mode decision purposes; and (2) to losslessly compress the data for signal output delivery.
  • the compressed data resource prediction task predicts the amount of bits required given a set of specific encoding modes for a given macroblock. Potential mode decision implementations may have up to eight modes to select from.
  • the computational demand on the CABAC engine to support the mode decision task is significant.
  • the weight update, arithmetic encoder and the bit packing components of the CABAC engine may require a significant amount of non-trivial computational and processing resources in a sequential processor implementation. Given that high performance encoding systems require multiple macro block rate distortion iterations of encoding per macro block, the CABAC process may impose an unreasonable resource demand on a processor-based solution. Prior implementations typically compromise on mode decision CABAC resource estimation accuracy by limiting the CABAC to bin level accuracy.
  • a system capable of processing one binary symbol per clock cycle requires a matching back end-receiving engine capable of also processing the results on every cycle.
  • the back end tasks consist of a value normalization task, which may generate up to eight bits of data, and a bit packing task, which groups the bits into bytes.
  • the implementation solutions for the normalization and bit packing tasks are complex and computationally demanding.
  • the first category includes routines that can generate at most one bit per cycle. This approach may utilize up to eight cycles to process one binary symbol as a single binary symbol may generate up to eight bits.
  • the second category includes routines that achieve single cycle per binary symbol using a method that does not optimally handle all cases of the carry from the input data and the adder.
  • a process extracts a plurality of subsets of bits from a first input. Each of the subsets of bits has a bit width equaling a number of leading zeros from a second input variable. Further, the process stores, in a memory, a consecutive sequence of the plurality of subsets. In addition, the process reads the consecutive sequence of the plurality of subsets from the memory if a third input release flag is established.
  • a process stores a consecutive set of variable bit width data into a first in first out buffer.
  • the variable bit width data has a width that is determined by a number of leading zeroes from an input variable. Further, the process reads the data from the first in first out buffer if the receiving data contains only ones.
  • a process stores a consecutive set of data from a first input variable into a memory. Further, the process receives a subsequent data set from the first input variable. In addition, the process reads the consecutive set of data from the memory if the subsequent data set includes one or more binary bits having a value of zero.
  • FIG. 1 illustrates a CABAC process
  • FIG. 2 illustrates an arithmetic coder normalization process
  • FIG. 3 illustrates an arithmetic coder normalization process that can process a new input data on every clock cycle.
  • FIG. 4 illustrates a process for code normalization and byte construction.
  • FIG. 5 illustrates another process for code normalization and byte construction.
  • FIG. 6 illustrates yet another process for code normalization and byte construction.
  • FIG. 7 illustrates a block diagram of a station or system that implements a code normalizer and byte construction engine.
  • a method and system are disclosed, which provide an improved video digital data compression capable of providing a single cycle normalization for real-time digital video encoders, such as an MPEG-4 or an H-264 series encoder.
  • the method and system may be utilized by the back end processor within the arithmetic encoder. As a result, normalization and payload to byte packing may be accomplished.
  • FIG. 1 illustrates a CABAC process 100 .
  • the CABAC process 100 selects a CABAC weight initialization mode.
  • the CABAC process 100 generates an MB syntax sequence.
  • the CABAC process 106 converts a syntax to binary.
  • the term binarization may be utilized to denote the process block 106 .
  • the CABAC process 100 performs a context address determination.
  • the term ctxldx generation may be utilized to denote the process block 108 .
  • the CABAC process 100 performs a context weight update.
  • the CABAC process 100 performs an arithmetic encoding.
  • the CABAC process 100 performs a bit packetizing.
  • the CABAC process 100 performs a NAL header construction. An elementary stream results from the CABAC process 100 .
  • FIG. 2 illustrates an arithmetic coder normalization process 200 .
  • the arithmetic coder normalization process 200 can be utilized for the MPEG4 standard to process data at the bit level.
  • the arithmetic coder normalization process 200 may utilize up to eight loop iterations to process a single binary input symbol from the front end arithmetic coder. Accordingly, an upper bound is placed on the computational demand.
  • the arithmetic coder normalization process begins at a process block 202 . Further, at a process block 204 , the arithmetic coder normalization process 200 receives a codeLow input variable and a codeRange input variable.
  • the codeLow input variable includes ten bits and the codeRange input variable includes nine bits.
  • the arithmetic coder normalization process 200 increments an internal index. For example, the arithmetic coder normalization process 200 may increment an internal index “t” by one.
  • the arithmetic coder normalization process 200 extracts the most significant bit from the codeLow input variable for a carry bit.
  • the arithmetic coder normalization process 200 sets a variable to hold the number of leading zeros of the codeRange input variable. For example, the variable may be entitled shftCnt.
  • the arithmetic coder normalization process 200 extracts a block of bits from the codeLow input variable. This is accomplished by discarding the most significant bits of the codeLow input variable and removing all the leading zeros to form a variable bit width block of bits.
  • the variable bit width block of bits is then stored in a payload array at location t, which may be referred to by the variable payload[t].
  • the arithmetic coder normalization process 200 determines if the contents of the variable payload[t], i.e., the bits, include only ones or both ones and zeroes. If the variable payload(t) includes both ones and zeroes, the arithmetic coder normalization process 200 proceeds to a process block 216 .
  • the arithmetic coder normalization process begins with the first entry of the payload array. A carry is added to the first entry in the payload array. The payload is then outputted without the resulting carry. The arithmetic coder normalization process 200 then adds the carry from the addition of the first entry in the payload array to the second entry in the payload array.
  • the payload is then outputted without the resulting carry.
  • the arithmetic coder normalization process 200 then proceeds to a process block 220 to reset the payload array by setting the variable t to zero.
  • the arithmetic coder normalization process 200 then ends at a process block 230 .
  • the arithmetic coder normalization process 200 determines, at the decision block 212 , that the contents of the variable payload[t] include only ones, the arithmetic coder normalization process proceeds from the decision block 212 to the process block 222 .
  • the carry bit is examined.
  • the arithmetic coder normalization process 200 then proceeds to a decision block 224 to determine if the input carry bit equals one. If the arithmetic coder normalization process 200 determines that the input carry bit equals one, the arithmetic coder normalization process 200 proceeds to a process block 226 .
  • the arithmetic coder normalization process 200 then proceeds to a process block 228 .
  • the arithmetic coder normalization process 200 resets the index to negative one.
  • the arithmetic coder normalization process 200 then ends at a process block 230 .
  • the arithmetic coder normalization process 200 determines, at the decision block 224 , that the input carry bit does not equal one, the arithmetic coder normalization process 200 ends at the process block 230 .
  • FIG. 3 illustrates a normalization and bit packing engine 300 capable of receiving one codeLow input and one code range input per cycle.
  • This approach is based on binarization of syntax element (“bin”) level processing.
  • a hardware solution is capable of providing a normalization and bit packing to bytes operation. This approach significantly reduces the hardware resources utilized by current systems. Specifically, this approach utilizes logic instead of memory lookup tables to resolve decision making tasks.
  • the normalization and bit packing engine 300 receives two distinct variables: a codeLow variable 302 and a codeRange variable 304 , on every clock cycle.
  • a leading zero detector 306 generates an output that is equal to the number of leading zero binary bits in the codeRange variable 304 . This output is registered in a latch shiftCnt 308 .
  • a bus splitter 310 outputs a carry bit and a dchunk variable. The carry bit is extracted from the most significant bit of the codeLow variable 302 . Further, the dchunk variable, which includes the second through ninth lower bits of the codeLow variable 302 , is then shifted right by shiftCnt variable 308 through a shift latch 312 .
  • the output dchunkRa of this shift latch 312 is then further shifted by bitPos 16 _ 1 variable through a bitPos 16 _ 1 shift latch 314 to align the data to fit into an output preparation register 316 .
  • the output preparation register 316 is utilized to hold data until there are enough output bits to form a full byte. In another embodiment, a plurality of output preparation registers 316 may be utilized.
  • a bit position calculator 318 generates a bitPos 16 _ 1 variable and a byte ready flag based on the input to the shiftCnt variable.
  • the bitPos 16 _ 1 variable identifies where the dchunkR should reside within the output preparation register 316 .
  • the byte ready flag identifies when the least significant byte 320 is ready for output.
  • the bitPos 16 _ 1 shift latch 314 outputs dchunk 16 , which is then sent to a logical or gate 322 along with the output from the output preparation register 316 .
  • the output from the logical or gate 322 is then sent to an adder 324 along with a shifted carry bit from a shift latch 326 to form both the output byte 328 and the new data for the output preparation register 316 .
  • the shifted carry bit is generated by the shift latch 326 , which shifts the logically conditioned carry bit utilizing oneFlag_d, a delayed carry flag carry_d 6 , and a delayed carry flag carry_d 5 .
  • the oneFlag_d is generated by first providing dchunk to an all ones detector 330 . If dchunk is all ones, the all ones detector 330 outputs oneFlag and provides oneFlag to a latch 332 . The latch 332 shifts oneFlag and outputs oneFlag_d.
  • the oneFlag_d is provided along with a delayed carry flag carry_d 6 to a first gate 330 . Further, the output of the first gate 334 is provided along with a delayed carry flag carry_d 5 to a second gate 336 .
  • the output of the adder 324 is split into a plurality of bytes through a bit splitter 338 .
  • the bit splitter 338 splits the output of the adder 324 into three bytes. Further, in one embodiment, the bit splitter 338 is a twenty four bit splitter.
  • the most significant byte is provided to an output byte register 340 , which may be denoted by the term outByte.
  • the two least significant bytes are routed through a multiplexor 342 to feed the inputs of the output preparation register 316 . Based on the byteRdy flag, the multiplexor 342 selects one of the two lower output bytes from the adder 324 for the middle byte 344 of the output preparation register 316 .
  • FIG. 4 illustrates a process 400 for code normalization and byte construction.
  • the process 400 extracts a plurality of subsets of bits from a first input. Each of the subsets of bits has a bit width equaling a number of leading zeros from a second input variable.
  • the process 400 stores, in a memory, a consecutive sequence of the plurality of subsets.
  • the process 400 reads the consecutive sequence of the plurality of subsets from the memory if a third input release flag is established.
  • FIG. 5 illustrates another process 500 for code normalization and byte construction.
  • the process 500 stores a consecutive set of variable bit width data into a first in first out buffer.
  • the variable bit width data has a width that is determined by a number of leading zeroes from an input variable.
  • the process 500 reads the data from the first in first out buffer if the receiving data contains only ones.
  • FIG. 6 illustrates yet another process 600 for code normalization and byte construction.
  • the process 600 stores a consecutive set of data from a first input variable into a memory. Further, at a process block 604 , the process 600 receives a subsequent data set from the first input variable. In addition, at a process block 606 , the process 600 reads the consecutive set of data from the memory if the subsequent data set includes one or more binary bits having a value of zero.
  • FIG. 7 illustrates a block diagram of a station or system 700 that implements a code normalizer and byte construction engine.
  • the station or system 700 is implemented using a general purpose computer or any other hardware equivalents.
  • the station or system 700 comprises a processor (“CPU”) 710 , a memory 720 , e.g., random access memory (“RAM”) and/or read only memory (ROM), a normalization and byte construction module 740 , and various input/output devices 730 , (e.g., storage devices, including but not limited to, a tape drive, a floppy drive, a hard disk drive or a compact disk drive, a receiver, a transmitter, a speaker, a display, an image capturing sensor, e.g., those used in a digital still camera or digital video camera, a clock, an output port, a user input device (such as a keyboard, a keypad, a mouse, and the like, or a microphone for capturing speech commands)).
  • a user input device
  • the code normalization and byte construction module 740 may be implemented as one or more physical devices that are coupled to the CPU 710 through a communication channel.
  • the normalization and byte construction module 740 may be represented by one or more software applications (or even a combination of software and hardware, e.g., using application specific integrated circuits (ASIC)), where the software is loaded from a storage medium, (e.g., a magnetic or optical drive or diskette) and operated by the CPU in the memory 720 of the computer.
  • ASIC application specific integrated circuits
  • the normalization and byte construction module 740 (including associated data structures) of the present invention may be stored on a computer readable medium, e.g., RAM memory, magnetic or optical drive or diskette and the like.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)
  • Compression, Expansion, Code Conversion, And Decoders (AREA)

Abstract

A method and system are provided for code normalization and byte construction. A plurality of subsets of bits is extracted from a first input. Each of the subsets of bits has a bit width equaling a number of leading zeros from a second input variable. Further, a consecutive sequence of the plurality of subsets is stored in a memory. In addition, the consecutive sequence of the plurality of subsets is read from the memory if a third input release flag is established.

Description

    BACKGROUND
  • 1. Field
  • This disclosure generally relates to the field of video data processing. More particularly, the disclosure relates to Context Adaptive Binary Arithmetic Coding (“CABAC”) for digital video encoders.
  • 2. General Background
  • Video signals generally include data corresponding to one or more video frames. Each video frame is composed of an array of picture elements, which are called pixels. A typical color video frame having a standard resolution may be composed of over several hundreds of thousands of pixels, which are arranged in arrays of blocks. Each pixel is characterized by pixel data indicative of a hue (predominant color), saturation (color intensity), and luminance (color brightness). The hue and saturation characteristics may be referred to as the chrominance. Accordingly, the pixel data includes chrominance and luminance. Therefore, the pixel data may be represented by groups of four luminance pixel blocks and two chrominance pixel blocks. These groups are called macroblocks (“MBs”). As a video frame generally includes many pixels, the video frame also includes a large number of MBs. Thus, digital signals representing a sequence of video frame data, which usually include many video frames, have a large number of bits. However, the available storage space and bandwidth for transmitting these digital signals is limited. Therefore, compression processes are used to more efficiently transmit or store video data.
  • Compression of digital video signals for transmission or for storage has become widely practiced in a variety of contexts. For example, multimedia environments for video conferencing, video games, Internet image transmissions, digital TV, and the like utilize compression. Coding and decoding are accomplished with coding processors. Examples of such coding processors include general computers, special hardware, multimedia boards, or other suitable processing devices. Further, the coding processors may utilize one of a variety of coding techniques, such as variable length coding (“VLC”), fixed coding, Huffman coding, blocks of symbols coding, and arithmetic coding. An example of arithmetic coding is Context Adaptive Binary Arithmetic Coding (“CABAC”).
  • CABAC techniques are capable of losslessly compressing syntax elements in a video stream using the probabilities of syntax elements in a given context. The CABAC process will take in syntax elements representing all elements within a macroblock. Further, the CABAC process constructs a compress bit sequence by building out the following structure: the sequential set of fields for the macroblock based on the chosen macroblock configuration, the specific syntax element type and value for each of the fields within this field sequence, and the context address for each of the syntax elements. The CABAC process will then perform binarization of the syntax elements, update the context weights, arithmetically encode the binarizations of syntax elements (“bins”), and subsequently pack the bits into bytes through the syntax element processing component.
  • The components of the CABAC process include: the CABAC weight initialization mode selection module, the macroblock syntax sequence generator, the binarization engine, the context address generator, the context weight update engine, the arithmetic coder, the bit packetizer, and the Network Abstraction Layer (“NAL”) header generator. The CABAC engine within a video encoder may accomplish two goals within the encoding process: (1) to carry out compressed data resource prediction for mode decision purposes; and (2) to losslessly compress the data for signal output delivery. The compressed data resource prediction task predicts the amount of bits required given a set of specific encoding modes for a given macroblock. Potential mode decision implementations may have up to eight modes to select from. The computational demand on the CABAC engine to support the mode decision task is significant.
  • The weight update, arithmetic encoder and the bit packing components of the CABAC engine may require a significant amount of non-trivial computational and processing resources in a sequential processor implementation. Given that high performance encoding systems require multiple macro block rate distortion iterations of encoding per macro block, the CABAC process may impose an unreasonable resource demand on a processor-based solution. Prior implementations typically compromise on mode decision CABAC resource estimation accuracy by limiting the CABAC to bin level accuracy.
  • A system capable of processing one binary symbol per clock cycle requires a matching back end-receiving engine capable of also processing the results on every cycle. The back end tasks consist of a value normalization task, which may generate up to eight bits of data, and a bit packing task, which groups the bits into bytes. The implementation solutions for the normalization and bit packing tasks are complex and computationally demanding.
  • Current implementations of the normalization function for the CABAC arithmetic coder fall into two categories. The first category includes routines that can generate at most one bit per cycle. This approach may utilize up to eight cycles to process one binary symbol as a single binary symbol may generate up to eight bits. The second category includes routines that achieve single cycle per binary symbol using a method that does not optimally handle all cases of the carry from the input data and the adder.
  • SUMMARY
  • In one aspect of the disclosure, a process extracts a plurality of subsets of bits from a first input. Each of the subsets of bits has a bit width equaling a number of leading zeros from a second input variable. Further, the process stores, in a memory, a consecutive sequence of the plurality of subsets. In addition, the process reads the consecutive sequence of the plurality of subsets from the memory if a third input release flag is established.
  • In another aspect, a process stores a consecutive set of variable bit width data into a first in first out buffer. The variable bit width data has a width that is determined by a number of leading zeroes from an input variable. Further, the process reads the data from the first in first out buffer if the receiving data contains only ones.
  • In yet another aspect, a process stores a consecutive set of data from a first input variable into a memory. Further, the process receives a subsequent data set from the first input variable. In addition, the process reads the consecutive set of data from the memory if the subsequent data set includes one or more binary bits having a value of zero.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The above-mentioned features of the present disclosure will become more apparent with reference to the following description taken in conjunction with the accompanying drawings wherein like reference numerals denote like elements and in which:
  • FIG. 1 illustrates a CABAC process.
  • FIG. 2 illustrates an arithmetic coder normalization process.
  • FIG. 3 illustrates an arithmetic coder normalization process that can process a new input data on every clock cycle.
  • FIG. 4 illustrates a process for code normalization and byte construction.
  • FIG. 5 illustrates another process for code normalization and byte construction.
  • FIG. 6 illustrates yet another process for code normalization and byte construction.
  • FIG. 7 illustrates a block diagram of a station or system that implements a code normalizer and byte construction engine.
  • DETAILED DESCRIPTION
  • A method and system are disclosed, which provide an improved video digital data compression capable of providing a single cycle normalization for real-time digital video encoders, such as an MPEG-4 or an H-264 series encoder. The method and system may be utilized by the back end processor within the arithmetic encoder. As a result, normalization and payload to byte packing may be accomplished.
  • FIG. 1 illustrates a CABAC process 100. At a process block 102, the CABAC process 100 selects a CABAC weight initialization mode. Further, at a process block 104, the CABAC process 100 generates an MB syntax sequence. In addition, at a process block 106, the CABAC process 106 converts a syntax to binary. The term binarization may be utilized to denote the process block 106. Further, at a process block 108, the CABAC process 100 performs a context address determination. The term ctxldx generation may be utilized to denote the process block 108. At a process block 110, the CABAC process 100 performs a context weight update. Further, at a process block 112, the CABAC process 100 performs an arithmetic encoding. In addition, at a process block 114, the CABAC process 100 performs a bit packetizing. Finally, at a process block 116, the CABAC process 100 performs a NAL header construction. An elementary stream results from the CABAC process 100.
  • FIG. 2 illustrates an arithmetic coder normalization process 200. In one embodiment, the arithmetic coder normalization process 200 can be utilized for the MPEG4 standard to process data at the bit level. The arithmetic coder normalization process 200 may utilize up to eight loop iterations to process a single binary input symbol from the front end arithmetic coder. Accordingly, an upper bound is placed on the computational demand. The arithmetic coder normalization process begins at a process block 202. Further, at a process block 204, the arithmetic coder normalization process 200 receives a codeLow input variable and a codeRange input variable. In one embodiment, the codeLow input variable includes ten bits and the codeRange input variable includes nine bits. Further, at a process block 206, the arithmetic coder normalization process 200 increments an internal index. For example, the arithmetic coder normalization process 200 may increment an internal index “t” by one. In addition, at a process block 208, the arithmetic coder normalization process 200 extracts the most significant bit from the codeLow input variable for a carry bit. At a process block 210, the arithmetic coder normalization process 200 sets a variable to hold the number of leading zeros of the codeRange input variable. For example, the variable may be entitled shftCnt. Further, at a next process block 212, the arithmetic coder normalization process 200 extracts a block of bits from the codeLow input variable. This is accomplished by discarding the most significant bits of the codeLow input variable and removing all the leading zeros to form a variable bit width block of bits. The variable bit width block of bits is then stored in a payload array at location t, which may be referred to by the variable payload[t].
  • At a decision block 214, the arithmetic coder normalization process 200 determines if the contents of the variable payload[t], i.e., the bits, include only ones or both ones and zeroes. If the variable payload(t) includes both ones and zeroes, the arithmetic coder normalization process 200 proceeds to a process block 216. At the process block 216, the arithmetic coder normalization process begins with the first entry of the payload array. A carry is added to the first entry in the payload array. The payload is then outputted without the resulting carry. The arithmetic coder normalization process 200 then adds the carry from the addition of the first entry in the payload array to the second entry in the payload array. The payload is then outputted without the resulting carry. The arithmetic coder normalization process 200 works through the entries payload array in a similar manner until the entry in payload(t-1) is processed. The iterations through these entries in the payload array may be denoted by the following code: for (i=0; i<t; i++) {payload[i]+=carry; Output(payload[i]}. Once the entry in payload[t-1] is processed, the arithmetic coder normalization process 200 proceeds to a process block 218 where the most recent payload is moved to the base of the array, which may be denoted by payload[0]=payload[t]. The arithmetic coder normalization process 200 then proceeds to a process block 220 to reset the payload array by setting the variable t to zero. The arithmetic coder normalization process 200 then ends at a process block 230.
  • If the arithmetic coder normalization process 200 determines, at the decision block 212, that the contents of the variable payload[t] include only ones, the arithmetic coder normalization process proceeds from the decision block 212 to the process block 222. At the process block 222, the carry bit is examined. The arithmetic coder normalization process 200 then proceeds to a decision block 224 to determine if the input carry bit equals one. If the arithmetic coder normalization process 200 determines that the input carry bit equals one, the arithmetic coder normalization process 200 proceeds to a process block 226. At the process block 226, the arithmetic coder normalization process 200 outputs all payload entries from index zero to index t sequentially beginning with the index zero. This approach can be denoted by the following code: for (i=0; i<=t; i++) {Output(payload[i])}. The arithmetic coder normalization process 200 then proceeds to a process block 228. At the process block 228, the arithmetic coder normalization process 200 resets the index to negative one. The arithmetic coder normalization process 200 then ends at a process block 230.
  • If the arithmetic coder normalization process 200 determines, at the decision block 224, that the input carry bit does not equal one, the arithmetic coder normalization process 200 ends at the process block 230.
  • FIG. 3 illustrates a normalization and bit packing engine 300 capable of receiving one codeLow input and one code range input per cycle. This approach is based on binarization of syntax element (“bin”) level processing. In one embodiment, a hardware solution is capable of providing a normalization and bit packing to bytes operation. This approach significantly reduces the hardware resources utilized by current systems. Specifically, this approach utilizes logic instead of memory lookup tables to resolve decision making tasks.
  • The normalization and bit packing engine 300 receives two distinct variables: a codeLow variable 302 and a codeRange variable 304, on every clock cycle. A leading zero detector 306 generates an output that is equal to the number of leading zero binary bits in the codeRange variable 304. This output is registered in a latch shiftCnt 308. A bus splitter 310 outputs a carry bit and a dchunk variable. The carry bit is extracted from the most significant bit of the codeLow variable 302. Further, the dchunk variable, which includes the second through ninth lower bits of the codeLow variable 302, is then shifted right by shiftCnt variable 308 through a shift latch 312. The output dchunkRa of this shift latch 312 is then further shifted by bitPos16_1 variable through a bitPos16_1 shift latch 314 to align the data to fit into an output preparation register 316. The output preparation register 316 is utilized to hold data until there are enough output bits to form a full byte. In another embodiment, a plurality of output preparation registers 316 may be utilized.
  • A bit position calculator 318 generates a bitPos16_1 variable and a byte ready flag based on the input to the shiftCnt variable. The bitPos16_1 variable identifies where the dchunkR should reside within the output preparation register 316. The byte ready flag identifies when the least significant byte 320 is ready for output. The bitPos16_1 shift latch 314 outputs dchunk16, which is then sent to a logical or gate 322 along with the output from the output preparation register 316. The output from the logical or gate 322 is then sent to an adder 324 along with a shifted carry bit from a shift latch 326 to form both the output byte 328 and the new data for the output preparation register 316. The shifted carry bit is generated by the shift latch 326, which shifts the logically conditioned carry bit utilizing oneFlag_d, a delayed carry flag carry_d6, and a delayed carry flag carry_d5.
  • The oneFlag_d is generated by first providing dchunk to an all ones detector 330. If dchunk is all ones, the all ones detector 330 outputs oneFlag and provides oneFlag to a latch 332. The latch 332 shifts oneFlag and outputs oneFlag_d.
  • The oneFlag_d is provided along with a delayed carry flag carry_d6 to a first gate 330. Further, the output of the first gate 334 is provided along with a delayed carry flag carry_d5 to a second gate 336.
  • The output of the adder 324 is split into a plurality of bytes through a bit splitter 338. In one embodiment, the bit splitter 338 splits the output of the adder 324 into three bytes. Further, in one embodiment, the bit splitter 338 is a twenty four bit splitter. The most significant byte is provided to an output byte register 340, which may be denoted by the term outByte. The two least significant bytes are routed through a multiplexor 342 to feed the inputs of the output preparation register 316. Based on the byteRdy flag, the multiplexor 342 selects one of the two lower output bytes from the adder 324 for the middle byte 344 of the output preparation register 316.
  • FIG. 4 illustrates a process 400 for code normalization and byte construction. At a process block 402, the process 400 extracts a plurality of subsets of bits from a first input. Each of the subsets of bits has a bit width equaling a number of leading zeros from a second input variable. Further, at a next process block 404, the process 400 stores, in a memory, a consecutive sequence of the plurality of subsets. In addition, at a process block 406, the process 400 reads the consecutive sequence of the plurality of subsets from the memory if a third input release flag is established.
  • FIG. 5 illustrates another process 500 for code normalization and byte construction. At a process block 502, the process 500 stores a consecutive set of variable bit width data into a first in first out buffer. The variable bit width data has a width that is determined by a number of leading zeroes from an input variable. Further, at a process block 504, the process 500 reads the data from the first in first out buffer if the receiving data contains only ones.
  • FIG. 6 illustrates yet another process 600 for code normalization and byte construction. At a process block 602, the process 600 stores a consecutive set of data from a first input variable into a memory. Further, at a process block 604, the process 600 receives a subsequent data set from the first input variable. In addition, at a process block 606, the process 600 reads the consecutive set of data from the memory if the subsequent data set includes one or more binary bits having a value of zero.
  • FIG. 7 illustrates a block diagram of a station or system 700 that implements a code normalizer and byte construction engine. In one embodiment, the station or system 700 is implemented using a general purpose computer or any other hardware equivalents. Thus, the station or system 700 comprises a processor (“CPU”) 710, a memory 720, e.g., random access memory (“RAM”) and/or read only memory (ROM), a normalization and byte construction module 740, and various input/output devices 730, (e.g., storage devices, including but not limited to, a tape drive, a floppy drive, a hard disk drive or a compact disk drive, a receiver, a transmitter, a speaker, a display, an image capturing sensor, e.g., those used in a digital still camera or digital video camera, a clock, an output port, a user input device (such as a keyboard, a keypad, a mouse, and the like, or a microphone for capturing speech commands)).
  • It should be understood that the code normalization and byte construction module 740 may be implemented as one or more physical devices that are coupled to the CPU 710 through a communication channel. Alternatively, the normalization and byte construction module 740 may be represented by one or more software applications (or even a combination of software and hardware, e.g., using application specific integrated circuits (ASIC)), where the software is loaded from a storage medium, (e.g., a magnetic or optical drive or diskette) and operated by the CPU in the memory 720 of the computer. As such, the normalization and byte construction module 740 (including associated data structures) of the present invention may be stored on a computer readable medium, e.g., RAM memory, magnetic or optical drive or diskette and the like.
  • It is understood that the normalization and byte construction engine described herein may also be applied in other type of encoders. Those skilled in the art will appreciate that the various adaptations and modifications of the embodiments of this method and apparatus may be configured without departing from the scope and spirit of the present method and system. Therefore, it is to be understood that, within the scope of the appended claims, the present method and apparatus may be practiced other than as specifically described herein.

Claims (20)

1. A method comprising:
extracting a plurality of subsets of bits from a first input, each of the subsets of bits having a bit width equaling a number of leading zeros from a second input variable;
storing, in a memory, a consecutive sequence of the plurality of subsets; and
reading the consecutive sequence of the plurality of subsets from the memory if a third input release flag is established.
2. The method of claim 1, wherein the consecutive sequence of the plurality of subsets is read from the memory in the same order that the consecutive sequence of the plurality of subsets is stored in the memory.
3. The method of claim 1, wherein each of the plurality of subsets of bits has a variable length.
4. The method of claim 3, further comprising concatenating the plurality of subsets of bits to form a stream of bits.
5. The method of claim 4, further comprising sending the stream of bits in a plurality of constant width blocks.
6. The method of claim 5, further comprising storing remaining bits that do not completely fill the constant width blocks in the plurality of constant width blocks as a subset of bits for a next set of input data.
7. The method of claim 1, wherein the third input release flag is the most significant bit of the first input.
8. The method of claim 5, wherein each of the subset of bits from the first input is formed by removing the most significant bit of the first input and continuing to remove the least significant bits of the input until the remaining bits are equal to the number of leading zeros from the second input variable.
9. The method of claim 1, further wherein the memory utilizes an array data structure for storage.
10. A method comprising,
storing a consecutive set of variable bit width data into a first in first out buffer, the variable bit width data having a width that is determined by a number of leading zeroes from an input variable; and
reading the data from the first in first out buffer if the receiving data contains only ones.
11. The method of claim 10, further comprising concatenating the variable length blocks of data to form a stream of bits.
12. The method of claim 11, further comprising sending the stream of bits in constant bit width blocks.
13. The method of claim 12, further comprising storing remaining bits that do not completely fill the constant width blocks as the first variable block width data for a next set of input data.
14. A method comprising:
storing a consecutive set of data from a first input variable into a memory;
receiving a subsequent data set from the first input variable; and
reading the consecutive set of data from the memory if the subsequent data set includes one or more binary bits having a value of zero.
15. The method of claim 14, further comprising reading the consecutive set of data from the memory if the most significant bit of the subsequent data set equals one.
16. The method of claim 14, wherein the first input variable stores data having a variable bit width.
17. The method of claim 16, wherein variable bit width data has a width that is determined by a number of leading zeroes from a second input variable.
18. The method of claim 14, further comprising adding an input carry flag to the consecutive set of data if the subsequent data set includes one or more bits having a value of zero.
19. The method of claim 18, further comprising adding a carry bit of the consecutive set of data to the subsequent data set.
20. The method of claim 15, further comprising determining a data set to store in the first input variable after the reading has completed based on data that causes a condition to be met so that the reading is initiated.
US11/625,417 2007-01-22 2007-01-22 Method and system for providing arithmetic code normalization and byte construction Expired - Fee Related US7397402B1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US11/625,417 US7397402B1 (en) 2007-01-22 2007-01-22 Method and system for providing arithmetic code normalization and byte construction

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US11/625,417 US7397402B1 (en) 2007-01-22 2007-01-22 Method and system for providing arithmetic code normalization and byte construction

Publications (2)

Publication Number Publication Date
US7397402B1 US7397402B1 (en) 2008-07-08
US20080175502A1 true US20080175502A1 (en) 2008-07-24

Family

ID=39589608

Family Applications (1)

Application Number Title Priority Date Filing Date
US11/625,417 Expired - Fee Related US7397402B1 (en) 2007-01-22 2007-01-22 Method and system for providing arithmetic code normalization and byte construction

Country Status (1)

Country Link
US (1) US7397402B1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102123144A (en) * 2011-02-22 2011-07-13 中兴通讯股份有限公司 Streaming media data transmission method and streaming transmission device
US20110268174A1 (en) * 2008-12-31 2011-11-03 Coship Electronics Co., Ltd. Package output method and device for arithmetic coding

Families Citing this family (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8189683B2 (en) * 2006-11-28 2012-05-29 General Instrument Corporation Method and system for providing single cycle context weight update leveraging context address look ahead
US8213499B2 (en) * 2007-04-04 2012-07-03 General Instrument Corporation Method and apparatus for context address generation for motion vectors and coefficients
US20080247459A1 (en) * 2007-04-04 2008-10-09 General Instrument Corporation Method and System for Providing Content Adaptive Binary Arithmetic Coder Output Bit Counting
US7535387B1 (en) * 2007-09-10 2009-05-19 Xilinx, Inc. Methods and systems for implementing context adaptive binary arithmetic coding
US8335256B2 (en) * 2008-11-14 2012-12-18 General Instrument Corporation Motion compensation in video coding
US9681133B2 (en) * 2012-03-29 2017-06-13 Intel Corporation Two bins per clock CABAC decoding
WO2015143602A1 (en) * 2014-03-24 2015-10-01 华为技术有限公司 Method for transmitting physical layer data and data transmission device

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5111415A (en) * 1989-11-06 1992-05-05 Hewlett-Packard Company Asynchronous leading zero counter employing iterative cellular array
US5751233A (en) * 1995-05-19 1998-05-12 Canon Kabushiki Kaisha Decoding apparatus and method therefor
US5798953A (en) * 1994-08-24 1998-08-25 Advanced Micro Devices, Inc. Apparatus and method for determining a number of digits leading a particular digit
US6970114B2 (en) * 1999-06-01 2005-11-29 Wend Llc Gate-based zero-stripping, varying length datum segment and arithmetic method and apparatus
US7286066B1 (en) * 2005-12-16 2007-10-23 Nvidia Corporation Acceleration of bitstream decoding

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5111415A (en) * 1989-11-06 1992-05-05 Hewlett-Packard Company Asynchronous leading zero counter employing iterative cellular array
US5798953A (en) * 1994-08-24 1998-08-25 Advanced Micro Devices, Inc. Apparatus and method for determining a number of digits leading a particular digit
US5751233A (en) * 1995-05-19 1998-05-12 Canon Kabushiki Kaisha Decoding apparatus and method therefor
US6970114B2 (en) * 1999-06-01 2005-11-29 Wend Llc Gate-based zero-stripping, varying length datum segment and arithmetic method and apparatus
US7286066B1 (en) * 2005-12-16 2007-10-23 Nvidia Corporation Acceleration of bitstream decoding

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110268174A1 (en) * 2008-12-31 2011-11-03 Coship Electronics Co., Ltd. Package output method and device for arithmetic coding
US8687701B2 (en) * 2008-12-31 2014-04-01 Shenzhen Coship Electronics Co., Ltd. Method and device for arithmetic code package outputting
CN102123144A (en) * 2011-02-22 2011-07-13 中兴通讯股份有限公司 Streaming media data transmission method and streaming transmission device

Also Published As

Publication number Publication date
US7397402B1 (en) 2008-07-08

Similar Documents

Publication Publication Date Title
US7397402B1 (en) Method and system for providing arithmetic code normalization and byte construction
US8213499B2 (en) Method and apparatus for context address generation for motion vectors and coefficients
JP7170669B2 (en) Coding the last significant coefficient flag
US7385534B2 (en) Methods and apparatus for processing variable length coded data
US7526029B2 (en) General purpose compression for video images (RHN)
US6639945B2 (en) Method and apparatus for implementing motion detection in video compression
US6292114B1 (en) Efficient memory mapping of a huffman coded list suitable for bit-serial decoding
US8189683B2 (en) Method and system for providing single cycle context weight update leveraging context address look ahead
WO2000002319A1 (en) Real time algorithms and architectures for coding images compressed by dwt-based techniques
JP6464192B2 (en) System and method for flatness detection for display stream compression (DSC)
EP2651133A2 (en) High performance renormalization for binary arithmetic video coding
WO2011005105A1 (en) Computer implemented early skip
US20080247459A1 (en) Method and System for Providing Content Adaptive Binary Arithmetic Coder Output Bit Counting
WO2012078281A1 (en) Video processing architecture
TWI705693B (en) Apparatus and method for vector-based entropy coding for display stream compression
EP3456044A2 (en) Method and device for context-adaptive binary arithmetic coding a sequence of binary symbols representing a syntax element related to video data
EP1292152B1 (en) Image processing apparatus, and image processing method
US9407933B2 (en) Simultaneous and loopless vector calculation of all run-level pairs in video compression
US10609382B2 (en) Method and apparatus for compressing video data
CN1849826A (en) Method and apparatus for scalable signal processing
US10225562B1 (en) Embedding information about EOB positions
Murat Key architectural optimizations for hardware efficient JPEG-LS encoder
KR20200065367A (en) Image processing device and frame buffer compressor
US10887593B2 (en) Parallel table-based bit rate estimator
EP0843483A2 (en) A method for decoding encoded video data

Legal Events

Date Code Title Description
AS Assignment

Owner name: GENERAL INSTRUMENT CORPORATION, PENNSYLVANIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:HU, YENDO;REEL/FRAME:018785/0071

Effective date: 20070118

STCF Information on status: patent grant

Free format text: PATENTED CASE

FPAY Fee payment

Year of fee payment: 4

AS Assignment

Owner name: MOTOROLA MOBILITY LLC, ILLINOIS

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:GENERAL INSTRUMENT HOLDINGS, INC.;REEL/FRAME:030866/0113

Effective date: 20130528

Owner name: GENERAL INSTRUMENT HOLDINGS, INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:GENERAL INSTRUMENT CORPORATION;REEL/FRAME:030764/0575

Effective date: 20130415

AS Assignment

Owner name: GOOGLE TECHNOLOGY HOLDINGS LLC, CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MOTOROLA MOBILITY LLC;REEL/FRAME:034294/0001

Effective date: 20141028

FPAY Fee payment

Year of fee payment: 8

FEPP Fee payment procedure

Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

LAPS Lapse for failure to pay maintenance fees

Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STCH Information on status: patent discontinuation

Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362

FP Lapsed due to failure to pay maintenance fee

Effective date: 20200708