US20080144723A1 - Rate control for multi-layer video design - Google Patents
Rate control for multi-layer video design Download PDFInfo
- Publication number
- US20080144723A1 US20080144723A1 US11/416,858 US41685806A US2008144723A1 US 20080144723 A1 US20080144723 A1 US 20080144723A1 US 41685806 A US41685806 A US 41685806A US 2008144723 A1 US2008144723 A1 US 2008144723A1
- Authority
- US
- United States
- Prior art keywords
- multimedia data
- layer
- encoding
- quality
- encoded
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/30—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/30—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
- H04N19/31—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability in the temporal domain
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/115—Selection of the code volume for a coding unit prior to coding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/124—Quantisation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/146—Data rate or code amount at the encoder output
- H04N19/15—Data rate or code amount at the encoder output by monitoring actual compressed data size at the memory before deciding storage at the transmission buffer
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/177—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a group of pictures [GOP]
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/189—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding
- H04N19/192—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding the adaptation method, adaptation tool or adaptation type being iterative or recursive
- H04N19/194—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding the adaptation method, adaptation tool or adaptation type being iterative or recursive involving only two passes
Definitions
- This invention relates to a method and apparatus for encoding and decoding scalable multimedia data with efficient reuse of base layer modules for construction of enhancement layer frames.
- IP Internet Protocol
- Wired connections include dial-up, ISDN, cable, xDSL, fiber, LAN (local area network), WAN (wide area network) and others.
- the transmission mode can be either uni-cast or multi-cast.
- PDA personal digital assistant
- laptop desktop, set-top box, TV, HDTV (high-definition television), mobile phone and others, requires bitstreams of different bandwidths simultaneously for the same content.
- the connection bandwidth could vary quickly with the time (from 9.6 kbps to 100 Mbps and above), and can be faster than a server's reaction.
- bitstreams are embedded as subsets of the macro-bitstream.
- scalable coding is for robust video transmissions on error prone channels. Error protection and error concealment can be performed. A more reliable transmission channel or a better error protection can be applied to base-layer bits that contain the most significant information.
- SNR signal to noise ratio
- MPEG-x high-density polyethylene
- H.261, H.262, H.263, and H.264 high-density polyethylene
- MCP motion-compensated prediction
- a video is typically divided into a series of groups of pictures (GOP), where each GOP begins with an intra-coded frame (I) followed by an arrangement of forward predicted frames (P) and bidirectional predicted frames (B). Both P-frames and B-frames are inter-frames.
- the B frame is the key to temporal scalability in most MPEG like coders.
- some profiles such as the MPEG-4 Simple profile and the H.264 Baseline Profile, do not support B frames.
- the MPEG-4 standard is described in ISO/IEC 14496-2.
- the H.264 standard is described in ISO/IEC 14496-10.
- profiles and levels provide a means of defining subsets of the syntax and semantics based on the decoder capabilities required to decode a particular bitstream.
- a profile is a defined sub-set of the entire bitstream syntax.
- a level is a defined set of constraints imposed on parameters in the bitstream. For any given profile, levels generally correspond to decoder processing load and memory capability. So profiles and levels specify restrictions on bitstreams and hence place limits on the capabilities of decoding the bitstreams.
- a decoder may be deemed to be conformant to a given profile at a given level if it is able to properly decode all allowed values of all syntactic elements of that profile at that level.
- a video encoding and decoding system that enables scalable coding of a video stream into multiple layers, is described.
- the quality of the base layer encoding and the enhancement layer encoding can be modified so that the transmission bandwidths for the base and enhancement layers satisfy a mathematical relationship, such as a ratio.
- One example is a method of encoding multimedia data into at least first and second layers, where each of the first and second layers has an associated transmission bandwidth.
- the method includes pre-encoding the multimedia data to generate one or more characteristic parameters of the multimedia data, and encoding the multimedia data to generate the at least first and second layers, where the associated bandwidths are determined at least in part based on the characteristic parameters so that the associated bandwidths substantially satisfy a mathematical relationship.
- a multimedia data encoding apparatus configured to encode the multimedia data into at least first and second layers, where each of the first and second layers has a corresponding transmission bandwidth.
- the apparatus includes a pre-encoder configured to pre-encode the multimedia data to generate one or more characteristic parameters of the multimedia data, and an encoder configured to encode the multimedia data to generate the at least first and second layers, where the corresponding transmission bandwidths are determined at least in part based on the characteristic parameters so that the corresponding bandwidths substantially satisfy a mathematical relationship.
- Another example is a multimedia data encoding apparatus configured to encode the multimedia data into at least first and second layers, where each of the first and second layers has a corresponding transmission bandwidth, the apparatus includes means for pre-encoding the multimedia data to generate one or more characteristic parameters of the multimedia data, and means for encoding the multimedia data to generate the at least first and second layers, where the corresponding transmission bandwidths are determined at least in part based on the characteristic parameters so that the corresponding bandwidths substantially satisfy a mathematical relationship.
- Another example is a computer readable medium including instructions which when executed cause a system to perform a method of encoding multimedia data into at least first and second layers, where each of the first and second layers has a corresponding transmission bandwidth.
- the method includes pre-encoding the multimedia data to generate one or more characteristic parameters of the multimedia data, and encoding the multimedia data to generate the at least first and second layers, where the corresponding transmission bandwidths are determined at least in part based on the characteristic parameters so that the corresponding bandwidths substantially satisfy a mathematical relationship.
- Another example is a processor configured to pre-encode multimedia data to generate one or more characteristic parameters of the multimedia data, and to encode the multimedia data to generate the at least first and second layers.
- the corresponding transmission bandwidths are determined at least in part based on the characteristic parameters so that the corresponding bandwidths substantially satisfy a mathematical relationship.
- FIG. 1 is a block diagram of a communications system for encoding and decoding streaming pictures.
- FIG. 2A is a diagram illustrating a conventional MPEG-4 Simple Profile data stream.
- FIG. 2B is a diagram illustrating a conventional encoded data stream that enables temporal scalability.
- FIG. 3 is an illustration of an example of a P Frame construction process in MPEG-4.
- FIG. 4 is a bandwidth allocation diagram depicting a bandwidth allocation methodology.
- FIG. 5 is a flow chart showing a method for encoding multimedia data.
- FIG. 6 is a graphical representation of a bitstream.
- a method, apparatus and system configured to provide scalable video coded data where the quality of the base layer encoding and the quality of the enhancement layer encoding can be adjusted so that the transmission bandwidth for the base and enhancement layers satisfy a desired mathematical relationship, such as a ratio, is described.
- Video frames can be pre-encoded at a pre-encode quality for the purpose of characterizing the raw data. Then, based on the characterization of the raw data, the frames can be encoded such that over a super-frame (e.g. one second of multimedia data), the bandwidth for the base and enhancement layers satisfy a particular mathematical relationship, such as a 1:1 ratio.
- examples may be described as a process, which is depicted as a flowchart, a flow diagram, a structure diagram, or a block diagram. Although a flowchart may describe the operations as a sequential process, many of the operations can be performed in parallel or concurrently and the process can be repeated. In addition, the order of the operations may be re-arranged.
- a process is terminated when its operations are completed.
- a process may correspond to a method, a function, a procedure, a subroutine, a subprogram, etc. When a process corresponds to a function, its termination corresponds to a return of the function to the calling function or the main function.
- a pair of video coding standards known as MPEG-x and H.26x, describe data processing and manipulation techniques (referred to herein as hybrid coding) that are well suited to the compression and delivery of video, audio and other information using fixed or variable length source coding techniques.
- the above-referenced standards, and other hybrid coding standards and techniques compress, illustratively, video information using intra-frame coding techniques (such as, for example, run-length coding, Huffman coding and the like) and inter-frame coding techniques (such as, for example, forward and backward predictive coding, motion compensation and the like).
- intra-frame coding techniques such as, for example, run-length coding, Huffman coding and the like
- inter-frame coding techniques such as, for example, forward and backward predictive coding, motion compensation and the like.
- hybrid video coding systems are characterized by prediction-based compression encoding of video frames with intra- and/or inter-frame motion compensation encoding.
- Intra-frame coding refers to encoding a picture (a field or a frame) without reference to any other picture, but the Intra-coded picture can be used as a reference for other pictures.
- the terms intra-frame, intra-coded frame and I Frame are all examples of video-objects formed with intra-coding that are used throughout this application.
- Inter or predictive coding refers to encoding a picture (a field or a frame) with reference to another picture. Compared to the Intra-coded picture, the Inter-coded or predicted picture may be coded with greater efficiency. Examples of inter-coded pictures that will be used throughout this application are predicted frames (either forward or backward predicted, also referred to as P frames), and bi-directional predicted frames (also referred to as B Frames). Other terms for inter-coding include high-pass coding, residual coding, motion compensated interpolation and others that are well known to those of ordinary skill in the art.
- a technique known as scalable coding, can divide encoded data into different layers in the bitstream such as, for example, a base layer and an enhancement layer.
- Scalable coding is useful in dynamic channels, where scalable bitstreams can be adapted to match fluctuations in network bandwidth.
- scalable coding can add robustness through unequal error protection of the base layer and the enhancement layer. Better error protection can be applied to the more important layer.
- FIG. 1 is a block diagram of a communications system for encoding and decoding streaming pictures.
- the system 100 includes encoder device 105 and decoder device 110 .
- Encoder device 105 further includes encoding component 120 , pre-encoding component 125 , memory component 130 and communications component 175 .
- Encoder device 105 is able to receive data from external source 135 using communication logic contained in communications component 175 .
- External source 135 could be, for example, external memory, the Internet, a live video and/or audio feed, and receiving the data can include wired and/or wireless communications.
- the data contained in external source 135 can be in a raw (not encoded) or encoded state.
- Pre-encoding component 125 is configured to perform an initial encoding of the data and to generate one or more parameters characterizing the raw data. Characterizing parameters such as a complexity metric can be generated by pre-encoding component 125 .
- Encoding component 120 is configured to encode the raw data based on the characterizing parameters generated in the pre-encoding component 125 .
- Each of pre-encoding component 125 and encoding component 120 are configured to encode both intra-coded portions of frames and predicted portions of frames, including forward prediction, backward prediction and bi-directional prediction.
- Each of pre-encoding component 125 and encoding component 120 is configured to determine which portions of multimedia data should be intra-coded, and which portions should be inter-coded.
- Each of pre-encoding component 125 and encoding component 120 is configured to determine which portions of video are low priority and should be placed in the enhancement layer, and which portions are high priority and should be placed in a base layer.
- Each of pre-encoding component 125 and encoding component 120 is configured to encode overhead information that can include a map of intra-coded and inter-coded portions, such as, for example a macro-block map where macroblocks (or sub-macroblocks) are identified as being intra-coded or inter-coded (also identifying which type of inter-coding including, for example forward, backward or bi-directional) and to which frame(s) inter-coded portions are referenced.
- the encoded frames are stored in memory component 130 or external memory.
- the external memory can be contained within external source 135 or a separate memory component (not shown).
- Communications component 175 contains logic used for data transmission (Tx) in conjunction with network 140 .
- Network 140 can be part of a wired system such as telephone, cable, and fiber optic, or a wireless system.
- network 140 can comprise, for example, part of a code division multiple access (CDMA or CDMA2000) communication system or alternately, the system can be a frequency division multiple access (FDMA) system, an orthogonal frequency division multiple access (OFDMA) system, a time division multiple access (TDMA) system such as GSMIGPRS (General Packet Radio Service)/EDGE (enhanced data GSM environment) or TETRA (Terrestrial Trunked Radio) mobile telephone technology for the service industry, a wideband code division multiple access (WCDMA), a high data rate (1xEV-DO or 1xEV-DO Gold Multicast) system, or in general any wireless communication system employing a combination of techniques.
- the encoded frames are transmitted (Tx) over network 140 .
- Decoder device 110 contains similar components to encoder device 105 , including decoding component 150 , memory component 160 and communications component 180 . Decoder device 110 receives encoded data that has been transmitted over network 140 or from external storage 165 . Communications component 180 contains logic used for receiving (Rx) encoded data in conjunction with network 140 , as well as logic for receiving encoded data from external storage 165 . External storage 165 could be, for example, external RAM or ROM, or a remote server. Decoding component 150 is used to decode intra-coded data, and inter-coded data. Decoding component 150 also decodes both P frames (forward or backward predicted) as well as B frames.
- P frames forward or backward predicted
- Decoding component 150 also contains logic used for decoding the overhead information, including the macroblock map discussed above, and translating this information so as to decode the appropriate macroblocks in the appropriate way.
- Decoding component 150 also contains logic used to skip decoding of enhancement layer data including bi-directionally predicted frames and/or low priority multimedia data. Decoding of enhancement layer data can be skipped if, for example, the enhancement layer data is not received or received with poor or insufficient SNR, and/or if there is insufficient processing power at the decoder device.
- the decoded frames can be displayed with display component 170 or stored in internal memory 160 or external storage 165 .
- Display component 170 can be an integrated part of the decoding device that contains such parts as video display hardware and logic, including a display screen, or it can be an external peripheral device. The decoding processes performed by decoder device 110 are more fully described below.
- FIG. 2A is a diagram illustrating a conventional MPEG-4 Simple Profile data stream, which depicts frame dependencies for a GOP.
- GOP 210 is made up of initial I Frame 212 , followed by several forward predicted P frames 214 .
- the dependency of P frames on a previous I or P frame can limit the temporal scalability afforded to a system (systems such as those conforming to the MPEG-4 Simple and H.264 Baseline Profiles) that may only support forward predicted frames.
- P frame removal may result in, for example, video jitter or the inability of the decoder to continue decoding until the next I frame 216 , which marks the beginning of the next GOP.
- FIG. 2B is a diagram illustrating a conventional encoded data stream that enables temporal scalability, which depicts the frame dependencies of a GOP.
- GOP 220 is made up of I frame 222 A, forward predicted P frames 224 , and bi-directional predicted B frames 226 .
- Each B frame can combine forward and backward motion vectors and residual errors referenced to I frame 222 A or forward predicted P frames 224 (backward predicted P frames could also be used but are not shown in this example).
- I frame 222 B marks the beginning of the next GOP. As shown in FIG.
- B frames 226 only one B frame 226 is contained between I frame 222 A and P frame 224 or between two P frames 224 .
- Several B frames could be inserted between reference frames to allow for greater flexibility in temporal scalability. Since no other frames may depend on the B frame as a reference frame, one could remove B frames 226 without loss of information regarding the decoding of other frames. This characteristic of B frames 226 can allow B frames 226 to be inserted into a bitstream, where the B frames 226 can be removed at the option of an encoder, a transcoder or a decoder to accommodate channel conditions, bandwidth limitations, battery power as well as other considerations. For example, if there are three B frames between reference frames, then one could remove all three B frames and reduce the frame rate by three quarters or one could keep the B frame in the middle and remove the other two to reduce the frame rate by one half. The data rate could decrease accordingly.
- FIG. 3 is an illustration of an example of a P Frame construction process in, for example, MPEG-4.
- Process 300 includes current picture 305 made up of 5 ⁇ 5 macroblocks, where the number of macroblocks in this example is arbitrary.
- a macroblock is made up of 16 ⁇ 16 pixels. Pixels can be defined by an 8-bit luminance value (Y) and two 8-bit chrominance values (Cr and Cb).
- Y, Cr and Cb components can be stored in a 4:2:0 format, where the Cr and Cb components are down-sampled by 2 in the X and the Y directions.
- each macroblock would consist of 256 Y components, 64 Cr components and 64 Cb components.
- Macroblock 315 of current picture 305 is predicted from reference picture 310 at a different time point than current picture 305 .
- a search is made in reference picture 310 to locate best matching macroblock 320 that is closest, in terms of Y, Cr and Cb values to current macroblock 315 being encoded.
- Methods of searching for best matching macroblock 320 include a) minimizing SAD (sum of absolute pixel differences) between current macroblock 315 and reference picture 310 macroblocks, b) minimizing SSD (sum of squared pixel differences), and c) minimum cost in a rate distortion sense, and others.
- the location of best matching macroblock 320 in reference picture 310 is encoded in motion vector 325 .
- Reference picture 310 can be an I Frame or P Frame that the decoder will have reconstructed prior to the construction of current picture 305 .
- Best matching macroblock 320 is subtracted from current macroblock 315 (a difference for each of the Y, Cr and Cb components is calculated) resulting in residual error 330 .
- Residual error 330 is encoded with 2D Discrete Cosine Transform (DCT) 335 and then quantized 340 .
- Quantization 340 can be performed to provide spatial compression by, for example, allotting fewer bits to the high frequency coefficients while allotting more bits to the low frequency coefficients.
- the quantized coefficients of residual error 330 along with motion vector 325 and reference picture 310 identifying information, are encoded information representing current macroblock 315 .
- the encoded information can be stored in memory for future use or operated on for purposes of, for example, error correction or image enhancement, or transmitted over network 345 .
- the encoded quantized coefficients of residual error 330 can be used to reconstruct current macroblock 315 in the encoder for use as part of a reference frame for subsequent motion estimation and compensation.
- the encoder can emulate the procedures of a decoder for this P Frame reconstruction. The emulation of the decoder will result in both the encoder and decoder working with the same reference picture.
- An example of a reconstruction process, whether done in an encoder, for further inter-coding, or in a decoder, is presented here. Reconstruction of a P Frame can be started after the reference frame (or a portion of a picture or frame that is being referenced) is reconstructed.
- the encoded quantized coefficients are dequantized 350 and then 2D Inverse DCT, or IDCT, 355 is performed resulting in decoded or reconstructed residual error 360 .
- Encoded motion vector 325 is used to locate the already reconstructed best matching macroblock 365 in the already reconstructed reference picture 310 .
- Reconstructed residual error 360 is then added to reconstructed best matching macroblock 365 to form reconstructed macroblock 370 .
- Reconstructed macroblock 370 can be stored in memory, displayed independently or in a picture with other reconstructed macroblocks, or processed further for image enhancement.
- B Frames can exploit temporal redundancy between a region in a current picture and a best matching prediction region in a previous picture and a best matching prediction region in a subsequent picture.
- the subsequent best matching prediction region and the backward best matching prediction region are combined to form a combined bi-directional predicted region.
- the difference between the current picture region and the best matching combined bi-directional prediction region is a residual error (or prediction error).
- the locations of the best matching prediction region in the subsequent reference picture and the best matching prediction region in the previous reference picture can be encoded in two motion vectors.
- the information of each of the I Frames, P Frames, and B Frames may be encoded in multiple layers, for example, a base layer may have certain minimum information, such as header information, and motion vector and residual error information for predicted frames, as well as I Frame data.
- An enhancement layer may have information which may not be necessary, but when optionally decoded enhances the quality of the displayed data.
- the enhancement layer may have I Frame data and residual data for P Frames and B Frames encoded with a higher quality than the data in the base layer.
- the quality of the encoded data may be adjusted through modification of various encoding parameters, such as the quantization parameter (QP), which determines a quantization level of the encoded data.
- QP quantization parameter
- the QP may be increased, resulting in encoded data which has higher quantization, and hence fewer bits and therefore a decreased bandwidth requirement.
- the enhancement layer may also include additional frames, such as B Frames, that are not included in the base layer. Both the base and enhancement layers may be transmitted to the decoder.
- bandwidth it is desirable to allocate bandwidth according to a specified mathematical relationship, such as a ratio, or a ratio plus a constant.
- a ratio the ratio of the bandwidth of the enhancement layer to the bandwidth of the base layer is 1:1, wherein the bandwidth allocated to the base layer is equal to the bandwidth of the enhancement layer.
- Other ratios may also be desirable, such as 1.5:1, 2:1
- Some embodiments encode I Frame data and P and B Frame residual data of the GOP at the enhancement layer quality. The size of encoded data then determines a desired bandwidth for the total base plus enhancement layers. The bandwidth for the base layer is then determined based on the desired mathematical relationship between the base layer bandwidth and the enhancement layer bandwidth.
- the bandwidth for each of the base layer elements can be determined.
- the bandwidth of certain fixed quality base layer elements may be independent of the bandwidth of the entire base layer.
- at least one of the header information and the motion vector information of the predicted frames may be encoded at a fixed quality into the base layer.
- the header and motion vector information may be encoded at the same high quality as the enhancement layer. The fixed quality for the header and motion vector information determines a size, and therefore, a bandwidth requirement for these fixed quality base layer elements.
- variable quality base layer elements such as the I Frame data and the P and B Frame residues. Accordingly, the variable quality base layer elements may be encoded with a quality such that the resulting data may be transmitted with the remaining bandwidth.
- FIG. 4 is a bandwidth allocation diagram depicting the bandwidth allocation methodology described above.
- Encoded enhancement layer information 410 and encoded base layer information 420 is depicted, where the horizontal dimension of each represents the allotted bandwidth of each. As shown, in this example, the total bandwidth of each layer is identical.
- the enhancement layer contains I Frame information 412 and B and P Frame information 414 encoded at a high, enhancement layer quality.
- the base layer contains header information 422 , motion vector information 424 , I Frame information 426 , and B and P Frame information 428 .
- the I Frame information 426 and the B and P Frame information 428 are each encoded with a base layer quality, the base layer quality being less than that of the enhancement layer.
- B Frame information is not encoded into the base layer.
- the decoder can decode B Frame information if it receives and is able to process the enhancement layer.
- the ratio of enhancement layer bandwidth to base layer bandwidth is not 1:1, but another ratio.
- the I Frame information 412 is encoded with a quality different from that of the B and P Frame residue information 414 .
- the I Frame information 426 is encoded with a quality different from that of the B and P Frame residue information 428 .
- the bandwidths of the enhancement layer and the base layer are substantially identical over the GOP.
- the 1:1 ratio does not hold for at least some smaller portions of the data, such as for frames or for macroblocks.
- the base layer bandwidth for the entire GOP minus the bandwidth allocated to the fixed quality elements is allocated to the variable quality elements according to characteristic parameters of each of the variable quality elements determined in a pre-encode operation.
- FIG. 5 is a flow chart showing a method for encoding multimedia data.
- a mathematical relationship between the bandwidth of the enhancement layer and the bandwidth of the base layer is determined.
- the mathematical relationship is a ratio in a memory location, for example a 1:1 ratio, where the enhancement layer bandwidth is substantially equal to the base layer bandwidth.
- the mathematical relationship is determined based on an algorithm.
- the multimedia data is pre-encoded.
- the multimedia data may be pre-encoded at a high quality. Characterizing parameters of the pre-encoded data are measured.
- Various complexity metrics of the encoded data can be determined by pre-encoding the multimedia data. For example, the size of encoded frames is measured and stored. Other parameters may also be measured and stored. The parameters of various portions of the data may be measured. For example, complexity of the encoded information may be measured for GOPs, frames, slices, macroblocks, and sub-macroblocks.
- a GOP is a super-frame, containing one second of multimedia data.
- the multimedia data is encoded according to the characterizing parameters measured at 520 .
- the I Frame data and the B and P Frame residues are encoded at a high enhancement layer quality into the enhancement layer.
- the enhancement layer and base layer bandwidths for the GOP are then determined.
- the bandwidths of fixed quality elements such as header information and motion vector information, is also determined.
- the BW vqe is allocated among the variable quality elements of the base layer according to the characterizing parameters, such as a complexity metric, of each of the variable quality elements determined in the pre-encode operation of 520 .
- the complexity metric determined in the pre-encode operation of 520 may be a size for each encoded frame.
- Size (i) is the size of the frame i encoded in 520
- Size total is to total of the sizes of all of the frames of the GOP encoded in 520 .
- the result is a bandwidth allocation for each frame of the GOP. Once each frame has an allocated bandwidth, the information for each frame is encoded with a maximum quality which will allow the encoded frame to be transmitted with the allocated bandwidth.
- FIG. 6 is a graphical representation of a bitstream that can result from a pre-encode operation on a multimedia GOP, and the results of subsequent encode operations which generate a bitstream for an enhancement layer for the GOP and a bitstream for a base layer for the GOP.
- the GOP comprises a GOP header, and three frames; an I Frame; a B Frame; and a P Frame.
- the I Frame comprises a header
- the B and P Frames each comprise a header and a motion vector.
- the pre-encoded bitstream includes all of the header information, the motion vector information, the I Frame data, and the residue information from the B and P Frames.
- the pre-encode operation encodes at a high quality, producing a bitstream with quality substantially identical to the quality of the enhancement layer video.
- the enhancement layer bitstream comprises the I Frame data, and the B and P Frame residue information, encoded at an enhancement layer quality.
- the enhancement layer may be generated with higher or lower quality than the pre-encoded data.
- the base layer bitstream has substantially the same size, and therefore bandwidth, as the enhancement layer bitstream.
- the base layer also includes header and motion vector information, which may be encoded at an enhancement layer quality
- the I Frame data and the B and P Frame header and residue information would be encoded at a quality less than that of the enhancement layer.
- the allocation of bandwidth for each of the I Frame data, the B Frame residue, and the P Frame residue in the base layer can be weighted according to the complexity of the corresponding pre-encoded data as determined by the pre-encoding process.
- a multimedia data encoding apparatus can be configured to encode the multimedia data into at least first and second layers. Each of the first and second layers has a corresponding transmission bandwidth.
- the apparatus includes means for pre-encoding the multimedia data to generate one or more characterizing parameters of the multimedia data, and means for encoding the multimedia data to generate the at least first and second layers.
- the means for pre-encoding can be an encoder configured to encode with a desired quality, and the means for encoding can be another or the same encoder configured to encode with the same or with another quality.
- the transmission bandwidths are determined at least in part based on the characterizing parameters so that the bandwidths substantially satisfy a mathematical relationship.
- a total bandwidth allotment for both the base layer and the enhancement layer is determined by a multiplexer (MUX), which determines bandwidth for multiple applications. It is advantageous when the sum of the base layer bandwidth and the enhancement layer bandwidth do not go beyond the limit set by the MUX.
- the MUX tries to allocate bits to different programs in a fair manner. The fairness is defined as different programs have similar quality. So the total bandwidth can be allocated to different programs more effectively. Since the MUX may cut bit rates for certain programs because of the limitation of total bandwidth, it is advantageous to have 2 pass coding on a superframe (SF) basis, where a SF is 1s of multimedia data.
- SF superframe
- the base layer and the enhancement layer need to keep, for example, a 1:1 ratio on a SF basis but for each frame there may be no such 1:1 restriction on the ratio of the base layer frame and the enhancement layer frame. Therefore, the 2 pass coding on a SF basis will also be helpful to allocate bits to each frame based on frame complexity while having 1:1 ratio on a SF basis.
- the encoding can target the enhancement layer quality, target the base layer quality, or another quality and multiply the base layer bit rate by 2 to get the bit rate for the sum of the base layer and enhancement layer.
- the enhancement layer quality can be optimized first.
- Pre-encoding is a characterization of the multimedia data, and will provide a reference quality of the base layer and enhancement layer. In this example, based on the characterization, the base layer quality will be inferred subjected to the requirement of 1:1 ratio.
- the pre-encoding operations There are a number of reasons for the pre-encoding operations. While encoding multimedia data at a preferred quality, a total desired bandwidth is determined, and can be given to the MUX. Another reason is to gather information for rate control for the subsequent encoding. After finishing the pre-encoding of the whole SF, the encoder will have a full picture of which parts of the SF need more bandwidth. Moreover, the rate control model, which is an essential tool for bit allocation can be more accurately built. Usually in single-pass coding, the model based on the encoding of neighboring macroblocks (MBs) is used for the current MB.
- MBs neighboring macroblocks
- the apparatus can estimate the base layer and the enhancement layer SF size.
- the MB header information such as MB mode, spatial prediction direction can be encoded into the base layer.
- the coefficient information can be encoded into both the enhancement layer and the base layer.
- ⁇ 1 is estimated based on the 2 nd pass coding.
- ⁇ qp is the difference between the 1 st pass QP and the preferred base layer QP.
- the motion information will be sent at the base layer.
- the number of bits in this part is known exactly. Similar to I frames, the coefficient bits will be estimated based on an exponential function.
- f 1,i (header) represents the bits for MB mode and motion information in the pre-encode
- f 1,i (Inter coeff) represents the bits for the coefficients of inter MBs in the pre-encode
- f 1,i (Intra coeff) represents the bits for the coefficients of intra MBs in the pre-encode
- ⁇ qp P,Inter is the QP difference between BL and the pre-encode Inter MBs
- ⁇ qp P,Intra is the QP difference between BL and the pre-encode Intra MBs
- ⁇ P,Inter and ⁇ P,Intra are coefficients decided by experiments.
- B frames if they only exist at one layer, their size can be added directly into that layer's SF size. If they have two layers, their contribution to each layer is estimated in the same way as P frames.
- the estimated BL SF size ⁇ tilde over (S) ⁇ b is defined as
- the estimated EL SF size ⁇ tilde over (S) ⁇ e is defined as
- the bit rate request sent to MUX will be MAX( ⁇ tilde over (S) ⁇ b , ⁇ tilde over (S) ⁇ e ) ⁇ 2. Based on the required bit rate requested by different programs, the MUX will return the upper bound for each program.
- the complexity metric is based on statistics of the residual after temporal prediction and the rate-control model is built upon the encoding of neighboring MBs. All of these could cause the discrepancy between the model and the real scenario. Instead, by using the pre-encoding results, we can build an accurate complexity metric reflecting the coding complexity for the encode operation. A rate-control model based on this scheme can help allocate bits more effectively at the encoding operation.
- the metric can be based on the number of bits of each MB in the pre-encoding operation. Since the different MBs could use different QPs, we normalize all MBs to the same QP using an exponential relationship and derive the corresponding coefficient bits. This process is advantageous because reducing inter MB QP will significantly increase the coefficient bits, which may not reflect the true motion complexity. For the same reason, the inter MB coefficient bits are adjusted in the first frame (according to the encoding order) of a SF if it is a P frame. The adjustment can be based on the MB QP of the first P frame and the QP of the reference area. The header bits may have no change. Then the complexity metric of each MB is the total bits of its header and coefficients. The complexity metric of each frame is the sum of the complexity metric of its MBs.
- the encode operation bit allocation at the base layer can be more accurate to avoid this kind of problem.
- a one-second buffer can be used to smooth out variations in the instantaneous output rate inside an SF. But unlike in the conventional leaky bucket model, neighboring SFs have independent buffers. So the smooth-out effect only happens inside an SF.
- the projected frame size is the
- N(j) is the remaining unencoded frame of the current SF.
- ⁇ (n j ) is the complexity metric of the j th frame decided in the pre-encoding.
- a basic unit can be either one or multiple MBs.
- the size of the basic unit determines how frequently QP can be adjusted. However, too much QP adjustment increases overhead.
- MBs of similar statistics can be grouped together and assigned a single QP. Based on inter mode or intra mode decided by the pre-encoding, MBs with the same mode can be grouped into a basic unit, because the statistics of inter MBs and intra MBs are different. MBs can also be grouped based on the coefficient information of the pre-encode operation. If one MB in a P frame has a non-zero coefficient, we will let this MB be a basic unit.
- b ⁇ ( BU k ) f r ⁇ ( k - 1 ) ⁇ ⁇ ⁇ ( BU k ) ⁇ ⁇ ( n j , k ) ,
- ⁇ (BU k ) is the complexity metric of the basic unit to be encoded
- ⁇ (n j,k ) is the complexity metric of the remaining frame.
- Q b MAX_BASE_QP, where ⁇ circumflex over (b) ⁇ 2,hdr (BU k ) is the estimated number of the encoded base layer header bits, ⁇ circumflex over (b) ⁇ 2,coeff (BU k ) is the estimated number of the encoded base layer coefficient bits, and MAX_BASE_QP is the maximum base layer QP.
- MAX_BASE_QP is chosen to make the base layer provide the minimum acceptable quality.
- each frame with in the GOP can be allocated a bandwidth according to the remaining available bandwidth for the GOP, the characterizing parameters of the frame, and the characterizing parameters of the remaining frames of the GOP, as described above.
- macroblocks within one or more frames can be allocated a bandwidth according to the remaining available bandwidth for the one or more frames, the characterizing parameters of the macroblock, and the characterizing parameters of the remaining macroblocks or other portions of the one or more frames. Numerous other embodiments applying the principles to various portions and combinations of portions of multimedia data are also possible.
- DSP digital signal processor
- ASIC application specific integrated circuit
- FPGA field programmable gate array
- a general purpose processor may be a microprocessor, but in the alternative, the processor may be any conventional processor, controller, microcontroller, or state machine.
- a processor may also be implemented as a combination of computing devices, e.g., a combination of a DSP and a microprocessor, a plurality of microprocessors, one or more microprocessors in conjunction with a DSP core, or any other such configuration.
- a software module may reside in RAM memory, flash memory, ROM memory, EPROM memory, EEPROM memory, registers, hard disk, a removable disk, a CD-ROM, or any other form of storage medium known in the art.
- a storage medium is coupled to the processor such that the processor can read information from, and write information to, the storage medium.
- the storage medium may be integral to the processor.
- the processor and the storage medium may reside in an Application Specific Integrated Circuit (ASIC).
- the ASIC may reside in a wireless modem.
- the processor and the storage medium may reside as discrete components in the wireless modem.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
Abstract
Description
- The present application for patent claims priority to Provisional Application No. 60/677,608 entitled “Rate Control For Two Layer Video Design” filed May 3, 2005, and assigned to the assignee hereof and hereby expressly incorporated by reference herein.
- 1. Field
- This invention relates to a method and apparatus for encoding and decoding scalable multimedia data with efficient reuse of base layer modules for construction of enhancement layer frames.
- 2. Background
- Due to the explosive growth and great success of the Internet and wireless communication, as well as increasing demand for multimedia services, streaming media over the Internet and mobile/wireless channels has drawn tremendous attention. In heterogeneous Internet Protocol (IP) networks, video is provided by a server and can be streamed by one or more clients. Wired connections include dial-up, ISDN, cable, xDSL, fiber, LAN (local area network), WAN (wide area network) and others. The transmission mode can be either uni-cast or multi-cast. The variety of individual client devices, including PDA (personal digital assistant), laptop, desktop, set-top box, TV, HDTV (high-definition television), mobile phone and others, requires bitstreams of different bandwidths simultaneously for the same content. The connection bandwidth could vary quickly with the time (from 9.6 kbps to 100 Mbps and above), and can be faster than a server's reaction.
- Similar to the heterogeneous IP network is mobile/wireless communication. Transport of multimedia content over mobile/wireless channels is very challenging because these channels are often severely impaired due to multi-path fading, shadowing, inter-symbol interference, and noise disturbances. Some other reasons such as mobility and competing traffic also cause bandwidth variations and loss. The channel noise and the number of users being served determine the time-varying property of channel environments. In addition to environmental conditions, the destination network can vary from second to third generation cellular networks to broadband data-only networks due to geographic location as well as mobile roaming. All these variables call for adaptive rate adjustment for multimedia content, even on the fly. Thus, successful transmission of video over heterogeneous wired/wireless networks requires efficient coding, as well as adaptability to varying network conditions, device characteristics, and user preferences, while also being resilient to losses.
- To meet different user requirements and to adapt to channel variation, one could generate multiple independent versions of bitstreams, each meeting one class of constraints based on transmission bandwidth, user display and computational capability, but this is not efficient for server storage and multicast application. In scalable coding, where a single macro-bitstream accommodating high-end users is built at the server, the bitstreams for low-end applications are embedded as subsets of the macro-bitstream. As such, a single bitstream can be adapted to diverse application environments by selectively transmitting sub-bitstreams. Another advantage provided by scalable coding is for robust video transmissions on error prone channels. Error protection and error concealment can be performed. A more reliable transmission channel or a better error protection can be applied to base-layer bits that contain the most significant information.
- There are spatial, temporal and signal to noise ratio (SNR) scalabilities in hybrid coders like MPEG-1, MPEG-2, MPEG-4 (collectively referred to as MPEG-x), H.261, H.262, H.263, and H.264 (collectively referred to as H.26x). In hybrid coding, temporal redundancy is removed by motion-compensated prediction (MCP). A video is typically divided into a series of groups of pictures (GOP), where each GOP begins with an intra-coded frame (I) followed by an arrangement of forward predicted frames (P) and bidirectional predicted frames (B). Both P-frames and B-frames are inter-frames. The B frame is the key to temporal scalability in most MPEG like coders. However, some profiles, such as the MPEG-4 Simple profile and the H.264 Baseline Profile, do not support B frames. The MPEG-4 standard is described in ISO/IEC 14496-2. The H.264 standard is described in ISO/IEC 14496-10.
- In MPEG-4, profiles and levels provide a means of defining subsets of the syntax and semantics based on the decoder capabilities required to decode a particular bitstream. A profile is a defined sub-set of the entire bitstream syntax. A level is a defined set of constraints imposed on parameters in the bitstream. For any given profile, levels generally correspond to decoder processing load and memory capability. So profiles and levels specify restrictions on bitstreams and hence place limits on the capabilities of decoding the bitstreams. In general, a decoder may be deemed to be conformant to a given profile at a given level if it is able to properly decode all allowed values of all syntactic elements of that profile at that level.
- A video encoding and decoding system that enables scalable coding of a video stream into multiple layers, is described. The quality of the base layer encoding and the enhancement layer encoding can be modified so that the transmission bandwidths for the base and enhancement layers satisfy a mathematical relationship, such as a ratio.
- One example is a method of encoding multimedia data into at least first and second layers, where each of the first and second layers has an associated transmission bandwidth. The method includes pre-encoding the multimedia data to generate one or more characteristic parameters of the multimedia data, and encoding the multimedia data to generate the at least first and second layers, where the associated bandwidths are determined at least in part based on the characteristic parameters so that the associated bandwidths substantially satisfy a mathematical relationship.
- Another example is a multimedia data encoding apparatus configured to encode the multimedia data into at least first and second layers, where each of the first and second layers has a corresponding transmission bandwidth. The apparatus includes a pre-encoder configured to pre-encode the multimedia data to generate one or more characteristic parameters of the multimedia data, and an encoder configured to encode the multimedia data to generate the at least first and second layers, where the corresponding transmission bandwidths are determined at least in part based on the characteristic parameters so that the corresponding bandwidths substantially satisfy a mathematical relationship.
- Another example is a multimedia data encoding apparatus configured to encode the multimedia data into at least first and second layers, where each of the first and second layers has a corresponding transmission bandwidth, the apparatus includes means for pre-encoding the multimedia data to generate one or more characteristic parameters of the multimedia data, and means for encoding the multimedia data to generate the at least first and second layers, where the corresponding transmission bandwidths are determined at least in part based on the characteristic parameters so that the corresponding bandwidths substantially satisfy a mathematical relationship.
- Another example is a computer readable medium including instructions which when executed cause a system to perform a method of encoding multimedia data into at least first and second layers, where each of the first and second layers has a corresponding transmission bandwidth. The method includes pre-encoding the multimedia data to generate one or more characteristic parameters of the multimedia data, and encoding the multimedia data to generate the at least first and second layers, where the corresponding transmission bandwidths are determined at least in part based on the characteristic parameters so that the corresponding bandwidths substantially satisfy a mathematical relationship.
- Another example is a processor configured to pre-encode multimedia data to generate one or more characteristic parameters of the multimedia data, and to encode the multimedia data to generate the at least first and second layers. The corresponding transmission bandwidths are determined at least in part based on the characteristic parameters so that the corresponding bandwidths substantially satisfy a mathematical relationship.
-
FIG. 1 is a block diagram of a communications system for encoding and decoding streaming pictures. -
FIG. 2A is a diagram illustrating a conventional MPEG-4 Simple Profile data stream. -
FIG. 2B is a diagram illustrating a conventional encoded data stream that enables temporal scalability. -
FIG. 3 is an illustration of an example of a P Frame construction process in MPEG-4. -
FIG. 4 is a bandwidth allocation diagram depicting a bandwidth allocation methodology. -
FIG. 5 is a flow chart showing a method for encoding multimedia data. -
FIG. 6 is a graphical representation of a bitstream. - A method, apparatus and system configured to provide scalable video coded data, where the quality of the base layer encoding and the quality of the enhancement layer encoding can be adjusted so that the transmission bandwidth for the base and enhancement layers satisfy a desired mathematical relationship, such as a ratio, is described. Video frames can be pre-encoded at a pre-encode quality for the purpose of characterizing the raw data. Then, based on the characterization of the raw data, the frames can be encoded such that over a super-frame (e.g. one second of multimedia data), the bandwidth for the base and enhancement layers satisfy a particular mathematical relationship, such as a 1:1 ratio.
- In the following description, specific details are given to provide a thorough understanding of the examples. However, it will be understood by one of ordinary skill in the art that the examples may be practiced without these specific details. For example, electrical components may be shown in block diagrams in order not to obscure the examples in unnecessary detail. In other instances, such components, other structures and techniques may be shown in detail to further explain the examples.
- It is also noted that the examples may be described as a process, which is depicted as a flowchart, a flow diagram, a structure diagram, or a block diagram. Although a flowchart may describe the operations as a sequential process, many of the operations can be performed in parallel or concurrently and the process can be repeated. In addition, the order of the operations may be re-arranged. A process is terminated when its operations are completed. A process may correspond to a method, a function, a procedure, a subroutine, a subprogram, etc. When a process corresponds to a function, its termination corresponds to a return of the function to the calling function or the main function.
- A pair of video coding standards, known as MPEG-x and H.26x, describe data processing and manipulation techniques (referred to herein as hybrid coding) that are well suited to the compression and delivery of video, audio and other information using fixed or variable length source coding techniques. In particular, the above-referenced standards, and other hybrid coding standards and techniques, compress, illustratively, video information using intra-frame coding techniques (such as, for example, run-length coding, Huffman coding and the like) and inter-frame coding techniques (such as, for example, forward and backward predictive coding, motion compensation and the like). Specifically, in the case of video processing systems, hybrid video coding systems are characterized by prediction-based compression encoding of video frames with intra- and/or inter-frame motion compensation encoding.
- Intra-frame coding refers to encoding a picture (a field or a frame) without reference to any other picture, but the Intra-coded picture can be used as a reference for other pictures. The terms intra-frame, intra-coded frame and I Frame are all examples of video-objects formed with intra-coding that are used throughout this application.
- Inter or predictive coding refers to encoding a picture (a field or a frame) with reference to another picture. Compared to the Intra-coded picture, the Inter-coded or predicted picture may be coded with greater efficiency. Examples of inter-coded pictures that will be used throughout this application are predicted frames (either forward or backward predicted, also referred to as P frames), and bi-directional predicted frames (also referred to as B Frames). Other terms for inter-coding include high-pass coding, residual coding, motion compensated interpolation and others that are well known to those of ordinary skill in the art.
- A technique, known as scalable coding, can divide encoded data into different layers in the bitstream such as, for example, a base layer and an enhancement layer. Scalable coding is useful in dynamic channels, where scalable bitstreams can be adapted to match fluctuations in network bandwidth. In error-prone channels, scalable coding can add robustness through unequal error protection of the base layer and the enhancement layer. Better error protection can be applied to the more important layer.
-
FIG. 1 is a block diagram of a communications system for encoding and decoding streaming pictures. Thesystem 100 includesencoder device 105 anddecoder device 110.Encoder device 105 further includesencoding component 120,pre-encoding component 125,memory component 130 andcommunications component 175.Encoder device 105 is able to receive data fromexternal source 135 using communication logic contained incommunications component 175.External source 135 could be, for example, external memory, the Internet, a live video and/or audio feed, and receiving the data can include wired and/or wireless communications. The data contained inexternal source 135 can be in a raw (not encoded) or encoded state.Pre-encoding component 125 is configured to perform an initial encoding of the data and to generate one or more parameters characterizing the raw data. Characterizing parameters such as a complexity metric can be generated bypre-encoding component 125.Encoding component 120 is configured to encode the raw data based on the characterizing parameters generated in thepre-encoding component 125. Each ofpre-encoding component 125 andencoding component 120 are configured to encode both intra-coded portions of frames and predicted portions of frames, including forward prediction, backward prediction and bi-directional prediction. - Each of
pre-encoding component 125 andencoding component 120 is configured to determine which portions of multimedia data should be intra-coded, and which portions should be inter-coded. Each ofpre-encoding component 125 andencoding component 120 is configured to determine which portions of video are low priority and should be placed in the enhancement layer, and which portions are high priority and should be placed in a base layer. Each ofpre-encoding component 125 andencoding component 120 is configured to encode overhead information that can include a map of intra-coded and inter-coded portions, such as, for example a macro-block map where macroblocks (or sub-macroblocks) are identified as being intra-coded or inter-coded (also identifying which type of inter-coding including, for example forward, backward or bi-directional) and to which frame(s) inter-coded portions are referenced. After encoding, the encoded frames are stored inmemory component 130 or external memory. The external memory can be contained withinexternal source 135 or a separate memory component (not shown). -
Communications component 175 contains logic used for data transmission (Tx) in conjunction withnetwork 140.Network 140 can be part of a wired system such as telephone, cable, and fiber optic, or a wireless system. In the case of wireless communication systems,network 140 can comprise, for example, part of a code division multiple access (CDMA or CDMA2000) communication system or alternately, the system can be a frequency division multiple access (FDMA) system, an orthogonal frequency division multiple access (OFDMA) system, a time division multiple access (TDMA) system such as GSMIGPRS (General Packet Radio Service)/EDGE (enhanced data GSM environment) or TETRA (Terrestrial Trunked Radio) mobile telephone technology for the service industry, a wideband code division multiple access (WCDMA), a high data rate (1xEV-DO or 1xEV-DO Gold Multicast) system, or in general any wireless communication system employing a combination of techniques. The encoded frames are transmitted (Tx) overnetwork 140. The encoding processes performed byencoder device 105 are more fully described below. -
Decoder device 110 contains similar components toencoder device 105, includingdecoding component 150,memory component 160 andcommunications component 180.Decoder device 110 receives encoded data that has been transmitted overnetwork 140 or fromexternal storage 165.Communications component 180 contains logic used for receiving (Rx) encoded data in conjunction withnetwork 140, as well as logic for receiving encoded data fromexternal storage 165.External storage 165 could be, for example, external RAM or ROM, or a remote server.Decoding component 150 is used to decode intra-coded data, and inter-coded data.Decoding component 150 also decodes both P frames (forward or backward predicted) as well as B frames. -
Decoding component 150 also contains logic used for decoding the overhead information, including the macroblock map discussed above, and translating this information so as to decode the appropriate macroblocks in the appropriate way.Decoding component 150 also contains logic used to skip decoding of enhancement layer data including bi-directionally predicted frames and/or low priority multimedia data. Decoding of enhancement layer data can be skipped if, for example, the enhancement layer data is not received or received with poor or insufficient SNR, and/or if there is insufficient processing power at the decoder device. After decoding, the decoded frames can be displayed withdisplay component 170 or stored ininternal memory 160 orexternal storage 165.Display component 170 can be an integrated part of the decoding device that contains such parts as video display hardware and logic, including a display screen, or it can be an external peripheral device. The decoding processes performed bydecoder device 110 are more fully described below. - In a typical MPEG decoder, predictive coded pixel blocks (i.e., blocks that comprise one or more motion vectors and a residual error component) are decoded with respect to a reference frame (where an intra-frame or another predicted frame can serve as a reference frame).
FIG. 2A is a diagram illustrating a conventional MPEG-4 Simple Profile data stream, which depicts frame dependencies for a GOP.GOP 210 is made up of initial I Frame 212, followed by several forward predicted P frames 214. The dependency of P frames on a previous I or P frame can limit the temporal scalability afforded to a system (systems such as those conforming to the MPEG-4 Simple and H.264 Baseline Profiles) that may only support forward predicted frames. Removing any of the P frames 214 may result in a loss of information that may be crucial in decoding other P frames. P frame removal may result in, for example, video jitter or the inability of the decoder to continue decoding until the next I frame 216, which marks the beginning of the next GOP. - Use of bi-directional predicted data in an enhancement layer of a scalable video coding system could be useful in dynamic channels as discussed above.
FIG. 2B is a diagram illustrating a conventional encoded data stream that enables temporal scalability, which depicts the frame dependencies of a GOP.GOP 220 is made up of I frame 222A, forward predicted P frames 224, and bi-directional predicted B frames 226. Each B frame can combine forward and backward motion vectors and residual errors referenced to I frame 222A or forward predicted P frames 224 (backward predicted P frames could also be used but are not shown in this example). I frame 222B marks the beginning of the next GOP. As shown inFIG. 2B , only oneB frame 226 is contained between I frame 222A andP frame 224 or between two P frames 224. Several B frames could be inserted between reference frames to allow for greater flexibility in temporal scalability. Since no other frames may depend on the B frame as a reference frame, one could remove B frames 226 without loss of information regarding the decoding of other frames. This characteristic of B frames 226 can allowB frames 226 to be inserted into a bitstream, where the B frames 226 can be removed at the option of an encoder, a transcoder or a decoder to accommodate channel conditions, bandwidth limitations, battery power as well as other considerations. For example, if there are three B frames between reference frames, then one could remove all three B frames and reduce the frame rate by three quarters or one could keep the B frame in the middle and remove the other two to reduce the frame rate by one half. The data rate could decrease accordingly. - P Frames (or any inter-coded sections) can exploit temporal redundancy between a region in a current picture and a best matching prediction region in a reference picture. The difference between the current region and the best matching reference prediction region is known as residual error (or prediction error). The location of the best matching prediction region in the reference frame can be encoded in a motion vector.
FIG. 3 is an illustration of an example of a P Frame construction process in, for example, MPEG-4.Process 300 includescurrent picture 305 made up of 5×5 macroblocks, where the number of macroblocks in this example is arbitrary. A macroblock is made up of 16×16 pixels. Pixels can be defined by an 8-bit luminance value (Y) and two 8-bit chrominance values (Cr and Cb). In MPEG, Y, Cr and Cb components can be stored in a 4:2:0 format, where the Cr and Cb components are down-sampled by 2 in the X and the Y directions. Hence, each macroblock would consist of 256 Y components, 64 Cr components and 64 Cb components.Macroblock 315 ofcurrent picture 305 is predicted fromreference picture 310 at a different time point thancurrent picture 305. A search is made inreference picture 310 to locatebest matching macroblock 320 that is closest, in terms of Y, Cr and Cb values tocurrent macroblock 315 being encoded. Methods of searching forbest matching macroblock 320, known by those of ordinary skill in the art, include a) minimizing SAD (sum of absolute pixel differences) betweencurrent macroblock 315 andreference picture 310 macroblocks, b) minimizing SSD (sum of squared pixel differences), and c) minimum cost in a rate distortion sense, and others. The location ofbest matching macroblock 320 inreference picture 310 is encoded inmotion vector 325.Reference picture 310 can be an I Frame or P Frame that the decoder will have reconstructed prior to the construction ofcurrent picture 305.Best matching macroblock 320 is subtracted from current macroblock 315 (a difference for each of the Y, Cr and Cb components is calculated) resulting inresidual error 330.Residual error 330 is encoded with 2D Discrete Cosine Transform (DCT) 335 and then quantized 340.Quantization 340 can be performed to provide spatial compression by, for example, allotting fewer bits to the high frequency coefficients while allotting more bits to the low frequency coefficients. The quantized coefficients ofresidual error 330, along withmotion vector 325 andreference picture 310 identifying information, are encoded information representingcurrent macroblock 315. The encoded information can be stored in memory for future use or operated on for purposes of, for example, error correction or image enhancement, or transmitted overnetwork 345. - The encoded quantized coefficients of
residual error 330, along with encodedmotion vector 325 can be used to reconstructcurrent macroblock 315 in the encoder for use as part of a reference frame for subsequent motion estimation and compensation. The encoder can emulate the procedures of a decoder for this P Frame reconstruction. The emulation of the decoder will result in both the encoder and decoder working with the same reference picture. An example of a reconstruction process, whether done in an encoder, for further inter-coding, or in a decoder, is presented here. Reconstruction of a P Frame can be started after the reference frame (or a portion of a picture or frame that is being referenced) is reconstructed. The encoded quantized coefficients are dequantized 350 and then 2D Inverse DCT, or IDCT, 355 is performed resulting in decoded or reconstructedresidual error 360. Encodedmotion vector 325 is used to locate the already reconstructed best matchingmacroblock 365 in the already reconstructedreference picture 310. Reconstructedresidual error 360 is then added to reconstructed best matchingmacroblock 365 to form reconstructed macroblock 370. Reconstructed macroblock 370 can be stored in memory, displayed independently or in a picture with other reconstructed macroblocks, or processed further for image enhancement. - B Frames (or any section coded with bi-directional prediction) can exploit temporal redundancy between a region in a current picture and a best matching prediction region in a previous picture and a best matching prediction region in a subsequent picture. The subsequent best matching prediction region and the backward best matching prediction region are combined to form a combined bi-directional predicted region. The difference between the current picture region and the best matching combined bi-directional prediction region is a residual error (or prediction error). The locations of the best matching prediction region in the subsequent reference picture and the best matching prediction region in the previous reference picture can be encoded in two motion vectors.
- As described above, the information of each of the I Frames, P Frames, and B Frames may be encoded in multiple layers, for example, a base layer may have certain minimum information, such as header information, and motion vector and residual error information for predicted frames, as well as I Frame data. An enhancement layer may have information which may not be necessary, but when optionally decoded enhances the quality of the displayed data. For example, the enhancement layer may have I Frame data and residual data for P Frames and B Frames encoded with a higher quality than the data in the base layer. The quality of the encoded data may be adjusted through modification of various encoding parameters, such as the quantization parameter (QP), which determines a quantization level of the encoded data. For example, the QP may be increased, resulting in encoded data which has higher quantization, and hence fewer bits and therefore a decreased bandwidth requirement. The enhancement layer may also include additional frames, such as B Frames, that are not included in the base layer. Both the base and enhancement layers may be transmitted to the decoder.
- In some systems, it is desirable to allocate bandwidth according to a specified mathematical relationship, such as a ratio, or a ratio plus a constant. For example, in some embodiments, the ratio of the bandwidth of the enhancement layer to the bandwidth of the base layer is 1:1, wherein the bandwidth allocated to the base layer is equal to the bandwidth of the enhancement layer. Other ratios may also be desirable, such as 1.5:1, 2:1
- In some embodiments, it is desirable to achieve a 1:1 ratio of bandwidths for each GOP. Some embodiments encode I Frame data and P and B Frame residual data of the GOP at the enhancement layer quality. The size of encoded data then determines a desired bandwidth for the total base plus enhancement layers. The bandwidth for the base layer is then determined based on the desired mathematical relationship between the base layer bandwidth and the enhancement layer bandwidth.
- Once the bandwidth for the base layer is known, the bandwidth for each of the base layer elements can be determined. In some embodiments, the bandwidth of certain fixed quality base layer elements may be independent of the bandwidth of the entire base layer. For example, in some embodiments, at least one of the header information and the motion vector information of the predicted frames may be encoded at a fixed quality into the base layer. For example, in some embodiments, the header and motion vector information may be encoded at the same high quality as the enhancement layer. The fixed quality for the header and motion vector information determines a size, and therefore, a bandwidth requirement for these fixed quality base layer elements. Once the bandwidth requirement for the fixed quality base layer elements is known, the remaining base layer bandwidth can be allotted to the remaining variable quality base layer elements, such as the I Frame data and the P and B Frame residues. Accordingly, the variable quality base layer elements may be encoded with a quality such that the resulting data may be transmitted with the remaining bandwidth.
-
FIG. 4 is a bandwidth allocation diagram depicting the bandwidth allocation methodology described above. Encodedenhancement layer information 410 and encodedbase layer information 420 is depicted, where the horizontal dimension of each represents the allotted bandwidth of each. As shown, in this example, the total bandwidth of each layer is identical. The enhancement layer containsI Frame information 412 and B andP Frame information 414 encoded at a high, enhancement layer quality. The base layer containsheader information 422,motion vector information 424, I Frameinformation 426, and B andP Frame information 428. In order to generate the base layer such that it can be transmitted with the same bandwidth as the enhancement layer, theI Frame information 426 and the B andP Frame information 428 are each encoded with a base layer quality, the base layer quality being less than that of the enhancement layer. - Various modifications can be made to the above described scheme. For example, in some embodiments, B Frame information is not encoded into the base layer. In such embodiments, the decoder can decode B Frame information if it receives and is able to process the enhancement layer. In some embodiments, the ratio of enhancement layer bandwidth to base layer bandwidth is not 1:1, but another ratio. In some embodiments the
I Frame information 412 is encoded with a quality different from that of the B and PFrame residue information 414. Similarly, in some embodiments, theI Frame information 426 is encoded with a quality different from that of the B and PFrame residue information 428. - In the example of
FIG. 4 , the bandwidths of the enhancement layer and the base layer are substantially identical over the GOP. In some embodiments, the 1:1 ratio does not hold for at least some smaller portions of the data, such as for frames or for macroblocks. For example, in some embodiments, the base layer bandwidth for the entire GOP minus the bandwidth allocated to the fixed quality elements is allocated to the variable quality elements according to characteristic parameters of each of the variable quality elements determined in a pre-encode operation. -
FIG. 5 is a flow chart showing a method for encoding multimedia data. At 510 a mathematical relationship between the bandwidth of the enhancement layer and the bandwidth of the base layer is determined. In some embodiments, the mathematical relationship is a ratio in a memory location, for example a 1:1 ratio, where the enhancement layer bandwidth is substantially equal to the base layer bandwidth. In some embodiments, the mathematical relationship is determined based on an algorithm. - At 520 the multimedia data is pre-encoded. The multimedia data may be pre-encoded at a high quality. Characterizing parameters of the pre-encoded data are measured. Various complexity metrics of the encoded data can be determined by pre-encoding the multimedia data. For example, the size of encoded frames is measured and stored. Other parameters may also be measured and stored. The parameters of various portions of the data may be measured. For example, complexity of the encoded information may be measured for GOPs, frames, slices, macroblocks, and sub-macroblocks. In some embodiments, a GOP is a super-frame, containing one second of multimedia data.
- At 530 the multimedia data is encoded according to the characterizing parameters measured at 520. In one example, the I Frame data and the B and P Frame residues are encoded at a high enhancement layer quality into the enhancement layer. According to methods described above, the enhancement layer and base layer bandwidths for the GOP are then determined. According to methods described above, the bandwidths of fixed quality elements, such as header information and motion vector information, is also determined.
- Next, the base layer bandwidth for the entire GOP (BWBL) minus the bandwidth allocated to the fixed quality elements of the base layer (BWfqe) is allocated to the variable quality elements of the base layer (BWvqe). Accordingly, BWvqe=BWBL−BWfqe. The BWvqe is allocated among the variable quality elements of the base layer according to the characterizing parameters, such as a complexity metric, of each of the variable quality elements determined in the pre-encode operation of 520. For example, the complexity metric determined in the pre-encode operation of 520 may be a size for each encoded frame. The allocated bandwidth for frame i (BWvqe(i)) may be a portion of the total BWvqe weighted according to the complexity of the frame i as determined in the pre-encode operation of 520. Accordingly, BWvqe(i)=BWvqe×Size(i)/Sizetotal, where Size(i) is the size of the frame i encoded in 520, and Sizetotal is to total of the sizes of all of the frames of the GOP encoded in 520. The result is a bandwidth allocation for each frame of the GOP. Once each frame has an allocated bandwidth, the information for each frame is encoded with a maximum quality which will allow the encoded frame to be transmitted with the allocated bandwidth.
-
FIG. 6 is a graphical representation of a bitstream that can result from a pre-encode operation on a multimedia GOP, and the results of subsequent encode operations which generate a bitstream for an enhancement layer for the GOP and a bitstream for a base layer for the GOP. In this example, the GOP comprises a GOP header, and three frames; an I Frame; a B Frame; and a P Frame. The I Frame comprises a header, and the B and P Frames each comprise a header and a motion vector. - As shown, the pre-encoded bitstream includes all of the header information, the motion vector information, the I Frame data, and the residue information from the B and P Frames. In this example, the pre-encode operation encodes at a high quality, producing a bitstream with quality substantially identical to the quality of the enhancement layer video. The enhancement layer bitstream comprises the I Frame data, and the B and P Frame residue information, encoded at an enhancement layer quality. In some embodiments, the enhancement layer may be generated with higher or lower quality than the pre-encoded data. As discussed above, in this example, the base layer bitstream has substantially the same size, and therefore bandwidth, as the enhancement layer bitstream. However, because the base layer also includes header and motion vector information, which may be encoded at an enhancement layer quality, the I Frame data and the B and P Frame header and residue information would be encoded at a quality less than that of the enhancement layer. The allocation of bandwidth for each of the I Frame data, the B Frame residue, and the P Frame residue in the base layer can be weighted according to the complexity of the corresponding pre-encoded data as determined by the pre-encoding process.
- In some embodiments a multimedia data encoding apparatus can be configured to encode the multimedia data into at least first and second layers. Each of the first and second layers has a corresponding transmission bandwidth. The apparatus includes means for pre-encoding the multimedia data to generate one or more characterizing parameters of the multimedia data, and means for encoding the multimedia data to generate the at least first and second layers. The means for pre-encoding can be an encoder configured to encode with a desired quality, and the means for encoding can be another or the same encoder configured to encode with the same or with another quality. The transmission bandwidths are determined at least in part based on the characterizing parameters so that the bandwidths substantially satisfy a mathematical relationship.
- In some embodiments a total bandwidth allotment for both the base layer and the enhancement layer is determined by a multiplexer (MUX), which determines bandwidth for multiple applications. It is advantageous when the sum of the base layer bandwidth and the enhancement layer bandwidth do not go beyond the limit set by the MUX. The MUX tries to allocate bits to different programs in a fair manner. The fairness is defined as different programs have similar quality. So the total bandwidth can be allocated to different programs more effectively. Since the MUX may cut bit rates for certain programs because of the limitation of total bandwidth, it is advantageous to have 2 pass coding on a superframe (SF) basis, where a SF is 1s of multimedia data.
- Furthermore, the base layer and the enhancement layer need to keep, for example, a 1:1 ratio on a SF basis but for each frame there may be no such 1:1 restriction on the ratio of the base layer frame and the enhancement layer frame. Therefore, the 2 pass coding on a SF basis will also be helpful to allocate bits to each frame based on frame complexity while having 1:1 ratio on a SF basis.
- During the pre-encode, the encoding can target the enhancement layer quality, target the base layer quality, or another quality and multiply the base layer bit rate by 2 to get the bit rate for the sum of the base layer and enhancement layer. Which approach to choose depends on the encoder design and which layer is chosen to be optimized first. Because in some embodiments, both layers can be received, the enhancement layer quality can be optimized first. Pre-encoding is a characterization of the multimedia data, and will provide a reference quality of the base layer and enhancement layer. In this example, based on the characterization, the base layer quality will be inferred subjected to the requirement of 1:1 ratio.
- There are a number of reasons for the pre-encoding operations. While encoding multimedia data at a preferred quality, a total desired bandwidth is determined, and can be given to the MUX. Another reason is to gather information for rate control for the subsequent encoding. After finishing the pre-encoding of the whole SF, the encoder will have a full picture of which parts of the SF need more bandwidth. Moreover, the rate control model, which is an essential tool for bit allocation can be more accurately built. Usually in single-pass coding, the model based on the encoding of neighboring macroblocks (MBs) is used for the current MB.
- Based on the pre-encoding results, the apparatus can estimate the base layer and the enhancement layer SF size.
- For I frames, there are only Intra MBs. The MB header information such as MB mode, spatial prediction direction can be encoded into the base layer. The coefficient information can be encoded into both the enhancement layer and the base layer. The estimated base layer frame size after the encoding is {circumflex over (f)}b 2,i=f1,i(header)+α1 Δqpf1,i(Intra coeff), where f1,i(header) and f1,i(Intra coeff) represent the number of header bits and the number of the coefficient bits of frame i in the 1st pass coding respectively. α1 is estimated based on the 2nd pass coding. Δqp is the difference between the 1st pass QP and the preferred base layer QP.
- The estimated EL frame size in the 2nd pass coding is {circumflex over (f)}e 2,i=(1−α1 Δqp)f1,i(Intra coeff). For P frames, the motion information will be sent at the base layer. The number of bits in this part is known exactly. Similar to I frames, the coefficient bits will be estimated based on an exponential function.
- Therefore, the BL frame size is defined as {circumflex over (f)}b 2,i=f1,i(header)+αP,Inter Δqp
P,Inter f1,i(Inter coeff)+αP,Intra ΔqpP,Intra f1,i(Intra coeff), Where f1,i(header) represents the bits for MB mode and motion information in the pre-encode, f1,i(Inter coeff) represents the bits for the coefficients of inter MBs in the pre-encode, f1,i(Intra coeff) represents the bits for the coefficients of intra MBs in the pre-encode, ΔqpP,Inter is the QP difference between BL and the pre-encode Inter MBs, ΔqpP,Intra is the QP difference between BL and the pre-encode Intra MBs, and αP,Inter and αP,Intra are coefficients decided by experiments. - The EL frame size is defined as {circumflex over (f)}e 2,i=(1−αP,Inter Δqp
P,Inter )f1,i(Inter coeff). For B frames, if they only exist at one layer, their size can be added directly into that layer's SF size. If they have two layers, their contribution to each layer is estimated in the same way as P frames. - The estimated BL SF size {tilde over (S)}b is defined as
-
-
- Usually in single-pass coding, the complexity metric is based on statistics of the residual after temporal prediction and the rate-control model is built upon the encoding of neighboring MBs. All of these could cause the discrepancy between the model and the real scenario. Instead, by using the pre-encoding results, we can build an accurate complexity metric reflecting the coding complexity for the encode operation. A rate-control model based on this scheme can help allocate bits more effectively at the encoding operation.
- The metric can be based on the number of bits of each MB in the pre-encoding operation. Since the different MBs could use different QPs, we normalize all MBs to the same QP using an exponential relationship and derive the corresponding coefficient bits. This process is advantageous because reducing inter MB QP will significantly increase the coefficient bits, which may not reflect the true motion complexity. For the same reason, the inter MB coefficient bits are adjusted in the first frame (according to the encoding order) of a SF if it is a P frame. The adjustment can be based on the MB QP of the first P frame and the QP of the reference area. The header bits may have no change. Then the complexity metric of each MB is the total bits of its header and coefficients. The complexity metric of each frame is the sum of the complexity metric of its MBs.
- In the encoding operation, for each layer, there is rate control for a target bit rate, Wb(0) for the base layer and We(0) for the enhancement layer. During 2 layer rate control, the base layer and enhancement layer overhead with respect to single layer coding should be minimized. Dramatically changing QP at the base layer can introduce a significant amount of overhead. Based on the statistics collected during the pre-encoding, the encode operation bit allocation at the base layer can be more accurate to avoid this kind of problem.
- In the base layer encoding, a one-second buffer can be used to smooth out variations in the instantaneous output rate inside an SF. But unlike in the conventional leaky bucket model, neighboring SFs have independent buffers. So the smooth-out effect only happens inside an SF.
- Wb(j) is the remaining base layer bandwidth of the current SF after encoding the (j−1)th frame. At the start of the SF, j=1. Wb(j) is updated frame by frame as follows: Wb(i)=Wb(i−1)−Sb(j−1), where Sb(j−1) is the number of bits generated by the (j−1)th frame.
- The projected frame size is
-
- where N(j) is the remaining unencoded frame of the current SF. β(nj) is the complexity metric of the jth frame decided in the pre-encoding.
- A basic unit can be either one or multiple MBs. The size of the basic unit determines how frequently QP can be adjusted. However, too much QP adjustment increases overhead. MBs of similar statistics can be grouped together and assigned a single QP. Based on inter mode or intra mode decided by the pre-encoding, MBs with the same mode can be grouped into a basic unit, because the statistics of inter MBs and intra MBs are different. MBs can also be grouped based on the coefficient information of the pre-encode operation. If one MB in a P frame has a non-zero coefficient, we will let this MB be a basic unit.
- fr(k) denotes the number of remaining bits after encoding basic unit k in the frame j. Initially, fr(0)=R(j). The number of coefficient bits spent on the current basic unit is calculated as
-
- where β(BUk) is the complexity metric of the basic unit to be encoded, and β(nj,k) is the complexity metric of the remaining frame.
- Meanwhile, the first pass generated basic unit header size b1,hdr (BUk) should also be considered. If b1,hdr(BUk)<b(BUk), {circumflex over (b)}2,hdr(BUk)=b1,hdr(BUk), and {circumflex over (b)}b2,coeff(BUk)=b(BUk)−{circumflex over (b)}2,hdr(BUk). Qb can then be calculated: Qb=MAX_BASE_QP, where {circumflex over (b)}2,hdr(BUk) is the estimated number of the encoded base layer header bits, {circumflex over (b)}2,coeff(BUk) is the estimated number of the encoded base layer coefficient bits, and MAX_BASE_QP is the maximum base layer QP. Usually, MAX_BASE_QP is chosen to make the base layer provide the minimum acceptable quality.
- The basic unit QP is assigned according to the exponential model R=αQ-Q
1 b1,coeff(BUk), where b1,coeff(BUk) is the estimated number of the 1st pass coefficient bits, Q1 is the 1st pass encoding QP, and α is a parameter updated on a frame-by-frame basis using linear regression. Intra MBs and Inter MBs use different α. fr(k) will be updated according to the number of bits usedb (BUk), fr(k)=fr(k−1)−b (BUk). - The above discussion focuses on determining frame sizes, however, the principles and methods discussed with respect to frames may be applied to any other portions of the data, such as macroblocks and basic units, which comprise macroblocks of similar characteristics. In some embodiments, the principles and methods discussed with respect to frame size determination can be applied to portions of the multimedia data of different sizes. For example, each frame with in the GOP can be allocated a bandwidth according to the remaining available bandwidth for the GOP, the characterizing parameters of the frame, and the characterizing parameters of the remaining frames of the GOP, as described above. In addition, macroblocks within one or more frames can be allocated a bandwidth according to the remaining available bandwidth for the one or more frames, the characterizing parameters of the macroblock, and the characterizing parameters of the remaining macroblocks or other portions of the one or more frames. Numerous other embodiments applying the principles to various portions and combinations of portions of multimedia data are also possible.
- Those of ordinary skill in the art would understand that information and signals may be represented using any of a variety of different technologies and techniques. For example, data, instructions, commands, information, signals, bits, symbols, and chips that may be referenced throughout the above description may be represented by voltages, currents, electromagnetic waves, magnetic fields or particles, optical fields or particles, or any combination thereof.
- Those of ordinary skill would further appreciate that the various illustrative logical blocks, modules, and algorithm steps described in connection with the examples disclosed herein may be implemented as electronic hardware, computer software, or combinations of both. To clearly illustrate this interchangeability of hardware and software, various illustrative components, blocks, modules, circuits, and steps have been described above generally in terms of their functionality. Whether such functionality is implemented as hardware or software depends upon the particular application and design constraints imposed on the overall system. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the disclosed methods.
- The various illustrative logical blocks, modules, and circuits described in connection with the examples disclosed herein may be implemented or performed with a general purpose processor, a digital signal processor (DSP), an application specific integrated circuit (ASIC), a field programmable gate array (FPGA) or other programmable logic device, discrete gate or transistor logic, discrete hardware components, or any combination thereof designed to perform the functions described herein. A general purpose processor may be a microprocessor, but in the alternative, the processor may be any conventional processor, controller, microcontroller, or state machine. A processor may also be implemented as a combination of computing devices, e.g., a combination of a DSP and a microprocessor, a plurality of microprocessors, one or more microprocessors in conjunction with a DSP core, or any other such configuration.
- The steps of a method or algorithm described in connection with the examples disclosed herein may be embodied directly in hardware, in a software module executed by a processor, or in a combination of the two. A software module may reside in RAM memory, flash memory, ROM memory, EPROM memory, EEPROM memory, registers, hard disk, a removable disk, a CD-ROM, or any other form of storage medium known in the art. A storage medium is coupled to the processor such that the processor can read information from, and write information to, the storage medium. In the alternative, the storage medium may be integral to the processor. The processor and the storage medium may reside in an Application Specific Integrated Circuit (ASIC). The ASIC may reside in a wireless modem. In the alternative, the processor and the storage medium may reside as discrete components in the wireless modem.
- The previous description of the disclosed embodiments is provided to enable any person skilled in the art to make or use the present invention. Various modifications to these embodiments will be readily apparent to those skilled in the art, and the generic principles defined herein may be applied to other examples and additional elements may be added without departing from the spirit or scope of the invention. Thus, the present invention is not intended to be limited to the embodiments shown herein but is to be accorded the widest scope consistent with the principles and novel features disclosed herein.
Claims (48)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/416,858 US7974341B2 (en) | 2005-05-03 | 2006-05-02 | Rate control for multi-layer video design |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US67760805P | 2005-05-03 | 2005-05-03 | |
US11/416,858 US7974341B2 (en) | 2005-05-03 | 2006-05-02 | Rate control for multi-layer video design |
Publications (2)
Publication Number | Publication Date |
---|---|
US20080144723A1 true US20080144723A1 (en) | 2008-06-19 |
US7974341B2 US7974341B2 (en) | 2011-07-05 |
Family
ID=37308706
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/416,858 Expired - Fee Related US7974341B2 (en) | 2005-05-03 | 2006-05-02 | Rate control for multi-layer video design |
Country Status (7)
Country | Link |
---|---|
US (1) | US7974341B2 (en) |
EP (1) | EP1884027A4 (en) |
JP (2) | JP2008541570A (en) |
KR (1) | KR100942395B1 (en) |
CN (1) | CN101208873B (en) |
TW (1) | TW200718215A (en) |
WO (1) | WO2006119436A2 (en) |
Cited By (44)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20080063051A1 (en) * | 2006-09-08 | 2008-03-13 | Mediatek Inc. | Rate control method with frame-layer bit allocation and video encoder |
US20080144713A1 (en) * | 2006-12-13 | 2008-06-19 | Viasat, Inc. | Acm aware encoding systems and methods |
US20080152003A1 (en) * | 2006-12-22 | 2008-06-26 | Qualcomm Incorporated | Multimedia data reorganization between base layer and enhancement layer |
US20080205529A1 (en) * | 2007-01-12 | 2008-08-28 | Nokia Corporation | Use of fine granular scalability with hierarchical modulation |
US20080267296A1 (en) * | 2007-04-24 | 2008-10-30 | Samsung Electronics Co., Ltd. | Method and apparatus for concealing an error of an image using residual data |
US20090041100A1 (en) * | 2006-12-13 | 2009-02-12 | Viasat, Inc. | Link aware mobile data network |
US20090060033A1 (en) * | 2006-12-13 | 2009-03-05 | Viasat, Inc. | Predictive adaptive coding and modulation |
US20090073876A1 (en) * | 2006-12-13 | 2009-03-19 | Viasat, Inc. | Adaptive coding and modulation aware network load balancing |
US20100008419A1 (en) * | 2008-07-10 | 2010-01-14 | Apple Inc. | Hierarchical Bi-Directional P Frames |
US20100189179A1 (en) * | 2009-01-29 | 2010-07-29 | Microsoft Corporation | Video encoding using previously calculated motion information |
US20100189183A1 (en) * | 2009-01-29 | 2010-07-29 | Microsoft Corporation | Multiple bit rate video encoding using variable bit rate and dynamic resolution for adaptive video streaming |
US20100260045A1 (en) * | 2006-12-13 | 2010-10-14 | Viasat, Inc. | Video and data network load balancing with video placeholder |
US20100260043A1 (en) * | 2006-12-13 | 2010-10-14 | Viasat, Inc. | Video and data network load balancing |
US20100260259A1 (en) * | 2006-12-13 | 2010-10-14 | Viasat, Inc. | Acm and fixed coding and modulation of hierarchical layers |
US20100262708A1 (en) * | 2009-04-08 | 2010-10-14 | Nokia Corporation | Method and apparatus for delivery of scalable media data |
US20100316126A1 (en) * | 2009-06-12 | 2010-12-16 | Microsoft Corporation | Motion based dynamic resolution multiple bit rate video encoding |
US20100329328A1 (en) * | 2007-06-26 | 2010-12-30 | Nokia, Inc. | Using scalable codecs for providing channel zapping information to broadcast receivers |
US7961665B2 (en) | 2006-12-13 | 2011-06-14 | Viasat, Inc. | Terminal aware multicasting |
US20110216828A1 (en) * | 2008-11-12 | 2011-09-08 | Hua Yang | I-frame de-flickering for gop-parallel multi-thread viceo encoding |
US20110222604A1 (en) * | 2008-11-13 | 2011-09-15 | Thomson Licensing | Multiple thread video encoding using gop merging and bit allocation |
US20110293009A1 (en) * | 2010-05-27 | 2011-12-01 | Freescale Semiconductor Inc. | Video processing system, computer program product and method for managing a transfer of information between a memory unit and a decoder |
US20120192240A1 (en) * | 2011-01-20 | 2012-07-26 | Roi Sasson | Participant aware configuration for video encoder |
WO2012099589A1 (en) * | 2011-01-20 | 2012-07-26 | Vidyo, Inc. | Participant aware configuration for video encoder |
US20130022120A1 (en) * | 2011-07-21 | 2013-01-24 | Texas Instruments Incorporated | Methods and systems for chroma residual data prediction |
US20130028316A1 (en) * | 2010-01-06 | 2013-01-31 | Dolby Laboratories Licensing Corporation | High Performance Rate Control for Multi-Layered Video Coding Applications |
US20130064292A1 (en) * | 2010-05-17 | 2013-03-14 | Sk Telecom Co., Ltd. | Image coding/decoding device using coding block in which intra block and inter block are mixed, and method thereof |
US8411571B2 (en) | 2006-12-13 | 2013-04-02 | Viasat, Inc. | Video and data network load balancing with video drop |
TWI394462B (en) * | 2010-06-22 | 2013-04-21 | Univ Nat Taiwan | Rate control method of perceptual-based rate-distortion optimized bit allocation |
US8457958B2 (en) | 2007-11-09 | 2013-06-04 | Microsoft Corporation | Audio transcoder using encoder-generated side information to transcode to target bit-rate |
US8576858B2 (en) | 2006-12-13 | 2013-11-05 | Viasat, Inc. | Multiple transmission paths for hierarchical layers |
US20140003492A1 (en) * | 2012-07-02 | 2014-01-02 | Qualcomm Incorporated | Video parameter set for hevc and extensions |
US8705616B2 (en) | 2010-06-11 | 2014-04-22 | Microsoft Corporation | Parallel multiple bitrate video encoding to reduce latency and dependences between groups of pictures |
US20140376617A1 (en) * | 2013-06-24 | 2014-12-25 | Kyeong Ho Yang | Rate control algorithm for scalable video encoding with disposable p-frames |
US20150172680A1 (en) * | 2013-12-16 | 2015-06-18 | Arris Enterprises, Inc. | Producing an Output Need Parameter for an Encoder |
US20150237369A1 (en) * | 2008-11-11 | 2015-08-20 | Samsung Electronics Co., Ltd. | Moving picture encoding/decoding apparatus and method for processing of moving picture divided in units of slices |
TWI508531B (en) * | 2014-06-04 | 2015-11-11 | Hon Hai Prec Ind Co Ltd | Video encoding device and method |
US20150350686A1 (en) * | 2014-05-29 | 2015-12-03 | Apple Inc. | Preencoder assisted video encoding |
US9467700B2 (en) | 2013-04-08 | 2016-10-11 | Qualcomm Incorporated | Non-entropy encoded representation format |
US9591318B2 (en) | 2011-09-16 | 2017-03-07 | Microsoft Technology Licensing, Llc | Multi-layer encoding and decoding |
US9986303B2 (en) * | 2014-03-31 | 2018-05-29 | JVC Kenwood Corporation | Video image coding data transmitter, video image coding data transmission method, video image coding data receiver, and video image coding data transmission and reception system |
US11089343B2 (en) | 2012-01-11 | 2021-08-10 | Microsoft Technology Licensing, Llc | Capability advertisement, configuration and control for video coding and decoding |
US11595650B2 (en) | 2017-03-31 | 2023-02-28 | Intel Corporation | Optimization of multi-sink Wi-Fi display with intelligent multi-session encoding |
US11860936B2 (en) * | 2008-10-14 | 2024-01-02 | Disney Enterprises, Inc. | Method and system for producing customized content |
WO2024058880A1 (en) * | 2022-09-13 | 2024-03-21 | Google Llc | Dynamic quantization parameter for encoding a video frame |
Families Citing this family (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2007008286A1 (en) * | 2005-07-11 | 2007-01-18 | Thomson Licensing | Method and apparatus for macroblock adaptive inter-layer intra texture prediction |
US8948309B2 (en) * | 2005-07-26 | 2015-02-03 | Broadcom Corporation | Method and system for redundancy-based decoding of video content in a wireless system |
KR100772878B1 (en) * | 2006-03-27 | 2007-11-02 | 삼성전자주식회사 | Method for assigning Priority for controlling bit-rate of bitstream, method for controlling bit-rate of bitstream, video decoding method, and apparatus thereof |
US7970066B1 (en) | 2006-12-27 | 2011-06-28 | Marvell International Ltd. | Tracking automatic gain control of orthogonal frequency domain multiplexing systems |
FR2913844B1 (en) * | 2007-03-16 | 2009-06-05 | Canon Kk | METHOD AND DEVICE FOR ALLOCATING FLOW IN A VIDEO SEQUENCE |
EP2059050A1 (en) * | 2007-10-31 | 2009-05-13 | Thomson Licensing | Method for coding a subsequent sequence of pictures by help of a preceding sequence of pictures |
US8213506B2 (en) * | 2009-09-08 | 2012-07-03 | Skype | Video coding |
GB2476271B (en) * | 2009-12-17 | 2015-09-02 | Skype | Coding data streams |
EP2805504A1 (en) * | 2012-01-18 | 2014-11-26 | Luca Rossato | Distinct encoding and decoding of stable information and transient/stochastic information |
US20130208809A1 (en) * | 2012-02-14 | 2013-08-15 | Microsoft Corporation | Multi-layer rate control |
US9565431B2 (en) * | 2012-04-04 | 2017-02-07 | Qualcomm Incorporated | Low-delay video buffering in video coding |
KR102001415B1 (en) * | 2012-06-01 | 2019-07-18 | 삼성전자주식회사 | Rate control Method for multi-layer video encoding, and video encoder and video signal processing system using method thereof |
US9510004B2 (en) | 2013-06-12 | 2016-11-29 | Microsoft Technology Licensing, Llc | Multi-layered rate control for scalable video coding |
FR3041851B1 (en) * | 2015-09-24 | 2018-02-16 | B<>Com | ASSOCIATED FLOW ALLOCATION METHOD, DEVICE, ENCODER AND COMPUTER PROGRAM |
CN113132728B (en) * | 2019-12-31 | 2023-06-27 | 上海海思技术有限公司 | Coding method and coder |
US11824653B2 (en) | 2021-12-17 | 2023-11-21 | Lenovo (Singapore) Pte. Ltd. | Radio access network configuration for video approximate semantic communications |
US11917206B2 (en) | 2021-12-17 | 2024-02-27 | Lenovo (Singapore) Pte. Ltd. | Video codec aware radio access network configuration and unequal error protection coding |
Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5515377A (en) * | 1993-09-02 | 1996-05-07 | At&T Corp. | Adaptive video encoder for two-layer encoding of video signals on ATM (asynchronous transfer mode) networks |
US6141486A (en) * | 1993-01-13 | 2000-10-31 | Hitachi America, Ltd. | Methods and apparatus for recording digital data including sync block and track number information for use during trick play operation |
US6349284B1 (en) * | 1997-11-20 | 2002-02-19 | Samsung Sdi Co., Ltd. | Scalable audio encoding/decoding method and apparatus |
US7227894B2 (en) * | 2004-02-24 | 2007-06-05 | Industrial Technology Research Institute | Method and apparatus for MPEG-4 FGS performance enhancement |
US7262721B2 (en) * | 2005-01-14 | 2007-08-28 | Samsung Electronics Co., Ltd. | Methods of and apparatuses for adaptive entropy encoding and adaptive entropy decoding for scalable video encoding |
US7400889B2 (en) * | 2003-04-01 | 2008-07-15 | Telefonaktiebolaget Lm Ericsson (Publ) | Scalable quality broadcast service in a mobile wireless communication network |
US7424730B2 (en) * | 1998-12-04 | 2008-09-09 | Microsoft Corporation | Multimedia presentation latency minimization |
US7463683B2 (en) * | 2000-10-11 | 2008-12-09 | Koninklijke Philips Electronics N.V. | Method and apparatus for decoding spatially scaled fine granular encoded video signals |
US7671893B2 (en) * | 2004-07-27 | 2010-03-02 | Microsoft Corp. | System and method for interactive multi-view video |
US7676722B2 (en) * | 2004-03-31 | 2010-03-09 | Sony Corporation | Multimedia content delivery using pre-stored multiple description coded video with restart |
Family Cites Families (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CA2208950A1 (en) * | 1996-07-03 | 1998-01-03 | Xuemin Chen | Rate control for stereoscopic digital video encoding |
US6292512B1 (en) * | 1998-07-06 | 2001-09-18 | U.S. Philips Corporation | Scalable video coding system |
US6167084A (en) * | 1998-08-27 | 2000-12-26 | Motorola, Inc. | Dynamic bit allocation for statistical multiplexing of compressed and uncompressed digital video signals |
US6263022B1 (en) * | 1999-07-06 | 2001-07-17 | Philips Electronics North America Corp. | System and method for fine granular scalable video with selective quality enhancement |
JP2004515132A (en) | 2000-11-23 | 2004-05-20 | コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ | Video encoding method and corresponding encoder |
GB0031538D0 (en) | 2000-12-22 | 2001-02-07 | Pa Consulting Services | Video layer mapping |
WO2003036978A1 (en) * | 2001-10-26 | 2003-05-01 | Koninklijke Philips Electronics N.V. | Method and apparatus for spatial scalable compression |
MXPA05002511A (en) | 2002-09-05 | 2005-08-16 | Agency Science Tech & Res | A method and an apparatus for controlling the rate of a video sequence; a video encoding device. |
JP2007509525A (en) * | 2003-10-20 | 2007-04-12 | サムスン エレクトロニクス カンパニー リミテッド | Bit rate control method and apparatus for uniform visual image quality |
-
2006
- 2006-05-02 US US11/416,858 patent/US7974341B2/en not_active Expired - Fee Related
- 2006-05-03 JP JP2008510209A patent/JP2008541570A/en not_active Withdrawn
- 2006-05-03 EP EP06759048A patent/EP1884027A4/en not_active Ceased
- 2006-05-03 TW TW095115762A patent/TW200718215A/en unknown
- 2006-05-03 CN CN2006800228882A patent/CN101208873B/en not_active Expired - Fee Related
- 2006-05-03 WO PCT/US2006/017164 patent/WO2006119436A2/en active Application Filing
- 2006-05-03 KR KR1020077028262A patent/KR100942395B1/en not_active IP Right Cessation
-
2012
- 2012-05-10 JP JP2012108243A patent/JP5384694B2/en not_active Expired - Fee Related
Patent Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6141486A (en) * | 1993-01-13 | 2000-10-31 | Hitachi America, Ltd. | Methods and apparatus for recording digital data including sync block and track number information for use during trick play operation |
US5515377A (en) * | 1993-09-02 | 1996-05-07 | At&T Corp. | Adaptive video encoder for two-layer encoding of video signals on ATM (asynchronous transfer mode) networks |
US6349284B1 (en) * | 1997-11-20 | 2002-02-19 | Samsung Sdi Co., Ltd. | Scalable audio encoding/decoding method and apparatus |
US7424730B2 (en) * | 1998-12-04 | 2008-09-09 | Microsoft Corporation | Multimedia presentation latency minimization |
US7463683B2 (en) * | 2000-10-11 | 2008-12-09 | Koninklijke Philips Electronics N.V. | Method and apparatus for decoding spatially scaled fine granular encoded video signals |
US7400889B2 (en) * | 2003-04-01 | 2008-07-15 | Telefonaktiebolaget Lm Ericsson (Publ) | Scalable quality broadcast service in a mobile wireless communication network |
US7227894B2 (en) * | 2004-02-24 | 2007-06-05 | Industrial Technology Research Institute | Method and apparatus for MPEG-4 FGS performance enhancement |
US7676722B2 (en) * | 2004-03-31 | 2010-03-09 | Sony Corporation | Multimedia content delivery using pre-stored multiple description coded video with restart |
US7671893B2 (en) * | 2004-07-27 | 2010-03-02 | Microsoft Corp. | System and method for interactive multi-view video |
US7262721B2 (en) * | 2005-01-14 | 2007-08-28 | Samsung Electronics Co., Ltd. | Methods of and apparatuses for adaptive entropy encoding and adaptive entropy decoding for scalable video encoding |
Cited By (81)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20080063051A1 (en) * | 2006-09-08 | 2008-03-13 | Mediatek Inc. | Rate control method with frame-layer bit allocation and video encoder |
US8135063B2 (en) * | 2006-09-08 | 2012-03-13 | Mediatek Inc. | Rate control method with frame-layer bit allocation and video encoder |
US10470236B2 (en) | 2006-12-13 | 2019-11-05 | Viasat, Inc. | Opportunistic progressive encoding |
US20100260259A1 (en) * | 2006-12-13 | 2010-10-14 | Viasat, Inc. | Acm and fixed coding and modulation of hierarchical layers |
US8456986B2 (en) | 2006-12-13 | 2013-06-04 | Viasat, Inc. | Video and data network load balancing |
US20090041100A1 (en) * | 2006-12-13 | 2009-02-12 | Viasat, Inc. | Link aware mobile data network |
US20090060033A1 (en) * | 2006-12-13 | 2009-03-05 | Viasat, Inc. | Predictive adaptive coding and modulation |
US20090073876A1 (en) * | 2006-12-13 | 2009-03-19 | Viasat, Inc. | Adaptive coding and modulation aware network load balancing |
US11083037B2 (en) | 2006-12-13 | 2021-08-03 | Viasat, Inc. | Opportunistic progressive encoding |
US11570838B2 (en) | 2006-12-13 | 2023-01-31 | Viasat, Inc. | Opportunistic progressive encoding |
US8358690B2 (en) | 2006-12-13 | 2013-01-22 | Viasat, Inc. | Predictive adaptive coding and modulation |
US20100260045A1 (en) * | 2006-12-13 | 2010-10-14 | Viasat, Inc. | Video and data network load balancing with video placeholder |
US20100260043A1 (en) * | 2006-12-13 | 2010-10-14 | Viasat, Inc. | Video and data network load balancing |
US9036716B2 (en) | 2006-12-13 | 2015-05-19 | Viasat, Inc. | Link aware mobile data network |
US20080144713A1 (en) * | 2006-12-13 | 2008-06-19 | Viasat, Inc. | Acm aware encoding systems and methods |
US8576858B2 (en) | 2006-12-13 | 2013-11-05 | Viasat, Inc. | Multiple transmission paths for hierarchical layers |
US8395993B2 (en) | 2006-12-13 | 2013-03-12 | Viasat, Inc. | Video and data network load balancing with video placeholder |
US8411572B2 (en) | 2006-12-13 | 2013-04-02 | Viasat, Inc. | ACM and fixed coding and modulation of hierarchical layers |
US7944872B2 (en) | 2006-12-13 | 2011-05-17 | Viasat, Inc. | Adaptive coding and modulation aware network load balancing |
US7961665B2 (en) | 2006-12-13 | 2011-06-14 | Viasat, Inc. | Terminal aware multicasting |
US8411571B2 (en) | 2006-12-13 | 2013-04-02 | Viasat, Inc. | Video and data network load balancing with video drop |
US8630355B2 (en) * | 2006-12-22 | 2014-01-14 | Qualcomm Incorporated | Multimedia data reorganization between base layer and enhancement layer |
US20080152003A1 (en) * | 2006-12-22 | 2008-06-26 | Qualcomm Incorporated | Multimedia data reorganization between base layer and enhancement layer |
US20080205529A1 (en) * | 2007-01-12 | 2008-08-28 | Nokia Corporation | Use of fine granular scalability with hierarchical modulation |
US20080267296A1 (en) * | 2007-04-24 | 2008-10-30 | Samsung Electronics Co., Ltd. | Method and apparatus for concealing an error of an image using residual data |
US8989260B2 (en) * | 2007-06-26 | 2015-03-24 | Nokia Corporation | Using scalable codecs for providing channel zapping information to broadcast receivers |
US9661378B2 (en) | 2007-06-26 | 2017-05-23 | Nokia Corporation | Using scalable codecs for providing channel zapping information to broadcast receivers |
US20100329328A1 (en) * | 2007-06-26 | 2010-12-30 | Nokia, Inc. | Using scalable codecs for providing channel zapping information to broadcast receivers |
US8457958B2 (en) | 2007-11-09 | 2013-06-04 | Microsoft Corporation | Audio transcoder using encoder-generated side information to transcode to target bit-rate |
US20100008419A1 (en) * | 2008-07-10 | 2010-01-14 | Apple Inc. | Hierarchical Bi-Directional P Frames |
US11860936B2 (en) * | 2008-10-14 | 2024-01-02 | Disney Enterprises, Inc. | Method and system for producing customized content |
US20150237369A1 (en) * | 2008-11-11 | 2015-08-20 | Samsung Electronics Co., Ltd. | Moving picture encoding/decoding apparatus and method for processing of moving picture divided in units of slices |
US20110216828A1 (en) * | 2008-11-12 | 2011-09-08 | Hua Yang | I-frame de-flickering for gop-parallel multi-thread viceo encoding |
US9210431B2 (en) * | 2008-11-13 | 2015-12-08 | Thomson Licensing | Multiple thread video encoding using GOP merging and bit allocation |
US20110222604A1 (en) * | 2008-11-13 | 2011-09-15 | Thomson Licensing | Multiple thread video encoding using gop merging and bit allocation |
US8396114B2 (en) | 2009-01-29 | 2013-03-12 | Microsoft Corporation | Multiple bit rate video encoding using variable bit rate and dynamic resolution for adaptive video streaming |
US8311115B2 (en) | 2009-01-29 | 2012-11-13 | Microsoft Corporation | Video encoding using previously calculated motion information |
WO2010088031A3 (en) * | 2009-01-29 | 2010-10-21 | Microsoft Corporation | Video encoding using previously calculated motion information |
US20100189183A1 (en) * | 2009-01-29 | 2010-07-29 | Microsoft Corporation | Multiple bit rate video encoding using variable bit rate and dynamic resolution for adaptive video streaming |
US20100189179A1 (en) * | 2009-01-29 | 2010-07-29 | Microsoft Corporation | Video encoding using previously calculated motion information |
US20100262708A1 (en) * | 2009-04-08 | 2010-10-14 | Nokia Corporation | Method and apparatus for delivery of scalable media data |
US8270473B2 (en) | 2009-06-12 | 2012-09-18 | Microsoft Corporation | Motion based dynamic resolution multiple bit rate video encoding |
US20100316126A1 (en) * | 2009-06-12 | 2010-12-16 | Microsoft Corporation | Motion based dynamic resolution multiple bit rate video encoding |
US20130028316A1 (en) * | 2010-01-06 | 2013-01-31 | Dolby Laboratories Licensing Corporation | High Performance Rate Control for Multi-Layered Video Coding Applications |
US8908758B2 (en) * | 2010-01-06 | 2014-12-09 | Dolby Laboratories Licensing Corporation | High performance rate control for multi-layered video coding applications |
US20130064292A1 (en) * | 2010-05-17 | 2013-03-14 | Sk Telecom Co., Ltd. | Image coding/decoding device using coding block in which intra block and inter block are mixed, and method thereof |
US9420293B2 (en) * | 2010-05-17 | 2016-08-16 | Sk Telecom Co., Ltd. | Image coding/decoding device using coding block in which intra block and inter block are mixed, and method thereof |
US9143793B2 (en) * | 2010-05-27 | 2015-09-22 | Freescale Semiconductor, Inc. | Video processing system, computer program product and method for managing a transfer of information between a memory unit and a decoder |
US20110293009A1 (en) * | 2010-05-27 | 2011-12-01 | Freescale Semiconductor Inc. | Video processing system, computer program product and method for managing a transfer of information between a memory unit and a decoder |
US8705616B2 (en) | 2010-06-11 | 2014-04-22 | Microsoft Corporation | Parallel multiple bitrate video encoding to reduce latency and dependences between groups of pictures |
TWI394462B (en) * | 2010-06-22 | 2013-04-21 | Univ Nat Taiwan | Rate control method of perceptual-based rate-distortion optimized bit allocation |
US8925027B2 (en) * | 2011-01-20 | 2014-12-30 | Vidyo, Inc. | Participant aware configuration for video encoder |
US20120192240A1 (en) * | 2011-01-20 | 2012-07-26 | Roi Sasson | Participant aware configuration for video encoder |
WO2012099589A1 (en) * | 2011-01-20 | 2012-07-26 | Vidyo, Inc. | Participant aware configuration for video encoder |
US10805619B2 (en) | 2011-07-21 | 2020-10-13 | Texas Instruments Incorporated | Methods and systems for chroma residual data prediction |
US20130022120A1 (en) * | 2011-07-21 | 2013-01-24 | Texas Instruments Incorporated | Methods and systems for chroma residual data prediction |
US11323724B2 (en) | 2011-07-21 | 2022-05-03 | Texas Instruments Incorporated | Methods and systems for chroma residual data prediction |
US9948938B2 (en) * | 2011-07-21 | 2018-04-17 | Texas Instruments Incorporated | Methods and systems for chroma residual data prediction |
US9591318B2 (en) | 2011-09-16 | 2017-03-07 | Microsoft Technology Licensing, Llc | Multi-layer encoding and decoding |
US9769485B2 (en) | 2011-09-16 | 2017-09-19 | Microsoft Technology Licensing, Llc | Multi-layer encoding and decoding |
US11089343B2 (en) | 2012-01-11 | 2021-08-10 | Microsoft Technology Licensing, Llc | Capability advertisement, configuration and control for video coding and decoding |
US20140003492A1 (en) * | 2012-07-02 | 2014-01-02 | Qualcomm Incorporated | Video parameter set for hevc and extensions |
US20170094277A1 (en) * | 2012-07-02 | 2017-03-30 | Qualcomm Incorporated | Video parameter set for hevc and extensions |
US9635369B2 (en) * | 2012-07-02 | 2017-04-25 | Qualcomm Incorporated | Video parameter set including HRD parameters |
US20140003491A1 (en) * | 2012-07-02 | 2014-01-02 | Qualcomm Incorporated | Video parameter set for hevc and extensions |
US20140003493A1 (en) * | 2012-07-02 | 2014-01-02 | Qualcomm Incorporated | Video parameter set for hevc and extensions |
US9716892B2 (en) * | 2012-07-02 | 2017-07-25 | Qualcomm Incorporated | Video parameter set including session negotiation information |
US9602827B2 (en) * | 2012-07-02 | 2017-03-21 | Qualcomm Incorporated | Video parameter set including an offset syntax element |
US9467700B2 (en) | 2013-04-08 | 2016-10-11 | Qualcomm Incorporated | Non-entropy encoded representation format |
US9565437B2 (en) | 2013-04-08 | 2017-02-07 | Qualcomm Incorporated | Parameter set designs for video coding extensions |
US9485508B2 (en) | 2013-04-08 | 2016-11-01 | Qualcomm Incorporated | Non-entropy encoded set of profile, tier, and level syntax structures |
US20170118474A1 (en) * | 2013-06-24 | 2017-04-27 | Kyeong Ho Yang | Rate control algorithm for scalable video encoding with disposable p-frames |
US10178389B2 (en) * | 2013-06-24 | 2019-01-08 | Dialog Corporation | Rate control algorithm for scalable video encoding with disposable P-frames |
US9578342B2 (en) * | 2013-06-24 | 2017-02-21 | Dialogic Corporation | Rate control algorithm for scalable video encoding with disposable P-frames |
US20140376617A1 (en) * | 2013-06-24 | 2014-12-25 | Kyeong Ho Yang | Rate control algorithm for scalable video encoding with disposable p-frames |
US20150172680A1 (en) * | 2013-12-16 | 2015-06-18 | Arris Enterprises, Inc. | Producing an Output Need Parameter for an Encoder |
US9986303B2 (en) * | 2014-03-31 | 2018-05-29 | JVC Kenwood Corporation | Video image coding data transmitter, video image coding data transmission method, video image coding data receiver, and video image coding data transmission and reception system |
US20150350686A1 (en) * | 2014-05-29 | 2015-12-03 | Apple Inc. | Preencoder assisted video encoding |
TWI508531B (en) * | 2014-06-04 | 2015-11-11 | Hon Hai Prec Ind Co Ltd | Video encoding device and method |
US11595650B2 (en) | 2017-03-31 | 2023-02-28 | Intel Corporation | Optimization of multi-sink Wi-Fi display with intelligent multi-session encoding |
WO2024058880A1 (en) * | 2022-09-13 | 2024-03-21 | Google Llc | Dynamic quantization parameter for encoding a video frame |
Also Published As
Publication number | Publication date |
---|---|
JP5384694B2 (en) | 2014-01-08 |
KR20080015831A (en) | 2008-02-20 |
EP1884027A2 (en) | 2008-02-06 |
US7974341B2 (en) | 2011-07-05 |
JP2008541570A (en) | 2008-11-20 |
CN101208873A (en) | 2008-06-25 |
WO2006119436A2 (en) | 2006-11-09 |
TW200718215A (en) | 2007-05-01 |
WO2006119436A3 (en) | 2007-11-01 |
KR100942395B1 (en) | 2010-02-17 |
CN101208873B (en) | 2012-04-25 |
JP2012178864A (en) | 2012-09-13 |
EP1884027A4 (en) | 2012-02-29 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US7974341B2 (en) | Rate control for multi-layer video design | |
US8320464B2 (en) | Method and apparatus to construct bi-directional predicted frames for temporal scalability | |
CA2562172C (en) | Method and apparatus for frame prediction in hybrid video compression to enable temporal scalability | |
JP4109113B2 (en) | Switching between bitstreams in video transmission | |
KR100950273B1 (en) | Scalable video coding with two layer encoding and single layer decoding | |
US8442122B2 (en) | Complexity scalable video transcoder and encoder | |
KR100964778B1 (en) | Multiple layer video encoding | |
KR100770873B1 (en) | Method and apparatus of controling bit rate in image encoding | |
KR100718467B1 (en) | Method and device for video down-sampling transcoding | |
Wu | Rate-distortion based optimal bit allocation for video streaming applications |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: QUALCOMM INCORPORATED, CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:CHEN, PEISONG;TIAN, TAO;RAVEENDRAN, VIJAYALAKSHMI R.;REEL/FRAME:018035/0290 Effective date: 20060707 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
FPAY | Fee payment |
Year of fee payment: 4 |
|
FEPP | Fee payment procedure |
Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
LAPS | Lapse for failure to pay maintenance fees |
Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
STCH | Information on status: patent discontinuation |
Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362 |
|
FP | Lapsed due to failure to pay maintenance fee |
Effective date: 20190705 |