WO2007114622A2 - Method and apparatus for encoding/decoding fgs layers using weighting factor - Google Patents
Method and apparatus for encoding/decoding fgs layers using weighting factor Download PDFInfo
- Publication number
- WO2007114622A2 WO2007114622A2 PCT/KR2007/001599 KR2007001599W WO2007114622A2 WO 2007114622 A2 WO2007114622 A2 WO 2007114622A2 KR 2007001599 W KR2007001599 W KR 2007001599W WO 2007114622 A2 WO2007114622 A2 WO 2007114622A2
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- enhanced layer
- weighted average
- weight
- denotes
- current frame
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 49
- 230000007423 decrease Effects 0.000 claims description 23
- 239000010410 layer Substances 0.000 description 175
- 230000006870 function Effects 0.000 description 10
- 230000008569 process Effects 0.000 description 10
- 238000013139 quantization Methods 0.000 description 10
- 230000002123 temporal effect Effects 0.000 description 10
- 230000005540 biological transmission Effects 0.000 description 7
- 238000007906 compression Methods 0.000 description 7
- 230000006835 compression Effects 0.000 description 7
- 238000010586 diagram Methods 0.000 description 7
- 230000008859 change Effects 0.000 description 5
- 230000003044 adaptive effect Effects 0.000 description 4
- 238000004590 computer program Methods 0.000 description 4
- 238000005516 engineering process Methods 0.000 description 4
- 238000012545 processing Methods 0.000 description 4
- 238000013144 data compression Methods 0.000 description 3
- 238000004891 communication Methods 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000005192 partition Methods 0.000 description 2
- 238000005070 sampling Methods 0.000 description 2
- 238000007792 addition Methods 0.000 description 1
- 238000003491 array Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000018109 developmental process Effects 0.000 description 1
- 239000000284 extract Substances 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 238000010295 mobile communication Methods 0.000 description 1
- 230000001151 other effect Effects 0.000 description 1
- 239000002356 single layer Substances 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/30—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
- H04N19/34—Scalability techniques involving progressive bit-plane based encoding of the enhancement layer, e.g. fine granular scalability [FGS]
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/132—Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
- H04N19/51—Motion estimation or motion compensation
- H04N19/577—Motion compensation with bidirectional frame interpolation, i.e. using B-pictures
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/587—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal sub-sampling or interpolation, e.g. decimation or subsequent interpolation of pictures in a video sequence
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/59—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial sub-sampling or interpolation, e.g. alteration of picture size or resolution
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/593—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques
Definitions
- Methods and apparatuses consistent with the present invention relate to video compression technology. More particularly, the present invention relates to a method and apparatus for encoding/decoding Fine Granular Scalability (FGS) layers by using weighted average sums in a coding technology of FGS layers using an adaptive reference scheme.
- FGS Fine Granular Scalability
- Multimedia data usually have a large volume which requires a large capacity medium for storage of the data and a wide bandwidth for transmission of the data. Therefore, it is indispensable to use a compression coding scheme in order to transmit multimedia data including text, image, and audio data.
- the basic principle of data compression lies in a process of removing redundancy in data.
- Data compression can be achieved by removing the spatial redundancy such as repetition of the same color or entity in an image, the temporal redundancy such as repetition of the same sound in audio data or nearly no change between temporally adjacent pictures in a moving image stream, or the perceptional redundancy based on the fact that the human visual and perceptional capability is insensitive to high frequencies.
- Data compression can be classified into loss/lossless compression according to whether the source data are lost or not, in-frame/inter-frame compression according to whether the compression is independent to each frame, and symmetric/ non-symmetric compression according to whether time necessary for the compression and restoration is the same.
- the temporal repetition is removed by temporal filtering based on motion compensation and the spatial repetition is removed by spatial transform.
- Transmission media which are necessary in order to transmit multimedia data generated after redundancies in the data are removed, show various levels of performance.
- Currently used transmission media include media having various transmission speeds, from an ultra high-speed communication network capable of transmitting several tens of mega bit data per second to a mobile communication network having a transmission speed of 384 kbps.
- the scalable video coding scheme that is, a scheme for transmitting the multimedia data at a proper data rate according to the transmission environment or in order to support transmission media of various speeds, is more proper for the multimedia environment.
- the scalable video coding includes a spatial scalability for controlling a resolution of a video, a Signal-to-Noise Ratio (SNR) scalability for controlling a screen quality of a video, a temporal scalability for controlling a frame rate, and combinations thereof.
- SNR Signal-to-Noise Ratio
- Standardization of the scalable video coding as described above has been already progressed in Moving Picture Experts Group-21 (MPEG-4) part 10.
- MPEG-4 Moving Picture Experts Group-21
- the scalability may be based on multiple layers including a base layer, a first enhanced layer (enhanced layer 1), a second enhanced layer (enhanced layer 2), etc., which have different resolutions (QCIF, CIF, 2CIR, etc.) or different frame rates.
- the motion vector includes a motion vector (former), which is individually obtained and used for each layer, and a motion vector (latter), which is obtained for one layer and is then also used for other layers (either as it is or after up/down sampling).
- FIG. 1 is a view illustrating a scalable video codec using a multi-layer structure.
- a base layer is defined to have a frame rate of Quarter Common Intermediate Format (QCIF)- 15Hz
- a first enhanced layer is defined to have a frame rate of Common Intermediate Format (CIF)-30Hz
- a second enhanced layer is defined to have a frame rate of Standard Definition (SD)-60 Hz. If a CIF 0.5 Mbps stream is required, it is possible to cut and transmit the bit stream so that the bit rate is changed to 0.5 Mbps in CIF_30Hz_0.7 Mbps of the first enhanced layer. In this way, the spatial, temporal, and SNR scalability can be implemented.
- QCIF Quarter Common Intermediate Format
- CIF Common Intermediate Format
- SD Standard Definition
- the SVM 3.0 employs not only the "Inter-prediction” and the “di- rectional intra-prediction,” which are used for prediction of blocks or macro-blocks constituting a current frame in the conventional H.264, but also the scheme of predicting a current block by using a correlation between a current block and a lower layer block corresponding to the current block.
- This prediction scheme is called “Intra_BL prediction,” and an encoding mode using this prediction is called “Intra_BL mode.”
- FIG. 2 is a schematic view for illustrating the three prediction schemes described above, which include an intra-prediction (®) for a certain macro-block 14 of a current frame 11, an inter-prediction ( ⁇ ) using a macro-block 15 of a frame 12 located at a position temporally different from that of the current frame 11, and an intra_BL prediction ( ⁇ ) using texture data for an area 16 of a base layer frame 13 corresponding to the macro-block 14.
- ® intra-prediction
- ⁇ inter-prediction
- ⁇ intra_BL prediction
- FIG. 3 is a block diagram illustrating the concept of a conventional coding of an FGS layer according to an adaptive reference scheme.
- FGS layers of frames are encoded by using an adaptive reference scheme.
- FGS layers of P frames of closed loops include a base layer, a first enhanced layer, and a second enhanced layer.
- the FGS layers are coded by using temporal prediction signals generated by adaptively referring to both a reference frame of the base layer and a reference frame of the enhanced layer.
- Equation (1) ⁇ denotes a predetermined weight known as a leaky factor
- D denotes a restored block of the base layer at the current frame t (that is, a block included in the frame 60)
- D ' denotes a restored block of the second enhanced layer at the previous frame t-1 (that is, a block included in the frame 50)
- R ' denotes the residual data (generated from frame 61) of the first enhanced layer at the current frame t.
- Equation (1) showing the process of generating the prediction signal, it is possible to control drift due to partial decoding by referring to the reference frame of the base layer and is also possible to obtain a high coding efficiency by using the reference frame of the enhanced layer.
- an embodiment of the present invention has been made to solve the above-mentioned problems occurring in the prior art, and an object of the present invention is to provide a method and apparatus for encoding/decoding FGS layers by using weighted average sums, which can control drift and simultaneously improve the coding efficiency in coding of frames of all FGS layers.
- a method of encoding FGS layers by using weighted average sums including (a) calculating a first weighted average sum by using a restored block of an n enhanced layer of a previous frame and a restored block of a base layer of a current frame; (b) calculating a second weighted average sum by using a restored block of the n enhanced layer of a next frame and a restored block of a base layer of the current frame; (c) generating a prediction signal of the n enhanced layer of the current frame by adding residual data of an (n - 1) enhanced layer of the current frame to a sum of the first weighted average sum and the second weighted average sum; and (d) encoding th residual data of the n enhanced layer, which is obtained by subtracting the generated prediction signal of the n enhanced layer from the restored block of the n' enhanced layer of the current frame.
- a method of decoding FGS layers by using weighted average sums including (a) calculating a first weighted average sum by using a restored block of an n enhanced layer of a previous frame and a restored block of a base layer of a current frame; (b) calculating a second weighted average sum by using a restored block of the n enhanced layer of a next frame and a restored block of a base layer of the current th frame; (c) generating a prediction signal of the n enhanced layer of the current frame by adding residual data of an (n - 1 ⁇ )th enhanced layer of the current frame to a sum of the first weighted average sum and the second weighted average sum; and (d) th generating a restored block of the n enhanced layer by adding the generated prediction signal of the n enhanced layer to residual data of the n enhanced layer.
- an encoder for encoding FGS layers by using weighted average sums including a first weighted average sum calculator calculating a first weighted average sum by using a restored block of an n enhanced layer of a previous frame and a restored block of a base layer of a current frame; a second weighted average sum calculator calculating a second weighted average sum by using a restored block of the n' enhanced layer of a next frame and a restored block of a base layer of the current frame; a prediction signal generator generating a prediction signal of the n enhanced layer of the current frame by adding residual data of an (n - 1) enhanced layer of the current frame to a sum of the first weighted average sum and the second weighted average sum; and a residual data generator generating residual data of the n enhanced layer by subtracting the generated prediction signal of the n enhanced layer from the restored block of the n enhanced layer of the current frame.
- a decoder for decoding FGS layers by using weighted average sums including a first weighted average sum calculator calculating a first weighted average sum by using a restored block of an n enhanced layer of a previous frame and a restored block of a base layer of a current frame; a second weighted average sum calculator calculating a second weighted average sum by using a restored block of the n enhanced layer of a next frame and a restored block of a base layer of the current frame; a prediction signal generator generating a prediction signal of the n enhanced layer of the current frame by adding residual data of an (n-1) enhanced layer of the current frame to a sum of the first weighted average sum and the second weighted average sum; and an enhanced layer restorer generating a restored block of the n enhanced layer by adding the generated prediction signal of the n enhanced layer to residual data of the n enhanced layer.
- FIG. 1 is a view illustrating a scalable video codec using a multi-layer structure
- FIG. 2 is a schematic view for illustrating three prediction schemes in a scalable video codec
- FIG. 3 is a block diagram illustrating the concept of a conventional coding of an FGS layer according to an adaptive reference scheme
- FIG. 4 is a flowchart illustrating the entire flow of a method of encoding FGS layers by using weighted average sums according to an exemplary embodiment of the present invention
- FIG. 5 is a flowchart illustrating the entire flow of a method of decoding FGS layers by using weighted average sums according to an exemplary embodiment of the present invention
- FIG. 6 illustrates the concept of an encoding of FGS layers by using weighted average sums according to an exemplary embodiment of the present invention
- FIG. 7 is a block diagram of an FGS encoder 100 for encoding FGS layers by using weighted average sums according to an exemplary embodiment of the present invention.
- FIG. 8 is a block diagram of an FGS decoder 200 for decoding FGS layers by using weighted average sums according to an exemplary embodiment of the present invention.
- Mode for the Invention is a block diagram of an FGS decoder 200 for decoding FGS layers by using weighted average sums according to an exemplary embodiment of the present invention.
- These computer program instructions may also be stored in a computer usable or computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer usable or computer-readable memory produce an article of manufacture including instruction means that implement the function specified in the flowchart block or blocks.
- the computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions that execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart block or blocks.
- each block of the flowchart illustrations may represent a module, segment, or portion of code, which includes one or more executable instructions for implementing the specified logical function(s). It should also be noted that in some alternative implementations, the functions noted in the blocks may occur out of the order. For example, two blocks shown in succession may in fact be executed substantially concurrently or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved.
- a base layer refers to a video sequence which has a frame rate lower than the maximum frame rate of a bit stream actually generated in a scalable video encoder and a resolution lower than the maximum resolution of the bit stream.
- the base layer has a predetermined frame rate and a predetermined solution, which are lower than the maximum frame rate and the maximum resolution, and the base layer need not have the lowest frame rate and the lowest resolution of the bit stream.
- the FGS layers may exist between the base layer and the enhanced layer.
- FIG. 4 is a flowchart illustrating the entire flow of a method of encoding FGS layers by using weighted average sums according to an embodiment of the present invention. The method shown in FIG. 4 will be described hereinafter with reference to FIG. 6 which illustrates the concept of an encoding of FGS layers by using weighted average sums according to an embodiment of the present invention.
- a first weighted average sum is calculated by using a restored block 111 of the base layer of the current frame t and a restored block 103 of the n enhanced layer of the previous frame t-1 (operation S 102).
- the first weighted average sum can be obtained by Equation (2) below.
- Equation (2) ⁇ denotes a predetermined first weight or leaky factor
- D ' denotes the restored block 111 of the base layer of the current frame t
- D n ' denotes the restored block 103 of the n enhanced layer of the previous frame t- 1.
- Equation (3) Equation (3)
- Equation (3) ⁇ denotes a predetermined second weight or leaky factor
- D denotes the restored block 111 of the base layer of the current frame t
- D n denotes the restored block 123 of the n enhanced layer of the next frame t+1.
- the first weighted average sum and the second weighted average sum are added, so as to reflect both of the two weighted average sums. At this time, it is preferred, but not necessary, to calculate an arithmetic mean of the two average sums rather than to simply add the first weighted average sum and the second weighted average sum.
- Equation (4) P ' denotes the prediction signal of the n enhanced layer of the n current frame t, and R n-l ' denotes the residual data of the (n-1) enhanced layer of the current frame t (the residual data is generated from the frame 112).
- Equation (4) It is noted from Equation (4) that two weights or leaky factors ⁇ and ⁇ are used during the process of obtaining the prediction signal of the n enhanced layer.
- the first and second weights can be derived from syntax factors existing in the header of the slice including macro-blocks to be coded, and adaptively change from 0 to 1 depending on characteristic information of the macro-blocks of the n enhanced layer of the current frame t.
- the characteristic information includes, for example, information about prediction direction of the macro-block, information about a Coded Block Pattern (CBP) value, and information about a Motion Vector Difference (MVD) value for the macro-block.
- CBP Coded Block Pattern
- MVD Motion Vector Difference
- the first weighted average sum is calculated by using the restored block 111 of the base layer of the current frame t and the restored block 103 of the n enhanced layer of the previous frame t-1 (operation S202). Then, the second weighted average sum is calculated by using the restored block 111 of the base layer of the current frame t and the restored block 123 of the n enhanced layer of the next frame t+1 (operation S204). Then, the first weighted average sum and the second weighted average sum are added and are then divided by 2, and the residual data of the (n- 1) enhanced layer of the current frame is added to the quotient of the division (operation S206), so that a prediction signal of the n enhanced layer of the current frame (operation S208). Operations S202 to S208 are similar to operations S 102 to S 108 described above in the encoding process shown in FIG. 4, so more detailed description thereof will be omitted here.
- the n n n n residual data R n ' of the n enhanced layer corresponds to residual data generated as a result of decoding and de-quantization of the FGS layer bit stream generated during the encoding process.
- an encoder and a decoder for performing the encoding and decoding will be described with reference to FIGS. 7 and 8. [60] From among the elements of the invention shown in FIGS. 7 and 8, the "unit" or
- module refers to a software element or a hardware element, such as a Field Pro- grammable Gate Array (FPGA) or an Application Specific Integrated Circuit (ASIC), which performs a predetermined function.
- FPGA Field Pro- grammable Gate Array
- ASIC Application Specific Integrated Circuit
- the module may be constructed either to be stored in an addressable storage medium or to execute one or more processors. Therefore, the module includes, for example, software elements, object-oriented software elements, class elements or task elements, processes, functions, properties, procedures, sub-routines, segments of a program code, drivers, firmware, micro-codes, circuits, data, database, data structures, tables, arrays, and parameters.
- the elements and functions provided by the modules may be either combined into a smaller number of elements or modules or divided into a larger number of elements or modules.
- FIG. 7 is a block diagram of an FGS encoder 100 for encoding FGS layers by using weighted average sums according to an embodiment of the present invention.
- a first weighted average sum calculator 110 calculates the first weighted average sum by adding a product obtained by multiplying the restored block data of the n enhanced layer of the previous frame by the first weight ⁇ and a product obtained by multiplying of the restored block data of the base layer of the current frame by a value 1- ⁇ .
- a second weighted average sum calculator 120 calculates the second weighted average sum
- a prediction signal generator 130 calculates an arithmetic mean of the first weighted average sum and the second weighted average sum by adding them and then dividing the sum of them by two, and then adds the residual data R n-l ' of the (n-1) enhanced layer of the current frame to the arithmetic mean, thereby obtaining the prediction signal R n ' of the n enhanced layer.
- the residual data R n-l ' of the (n-l) enhanced layer For the residual data R n-l ' of the (n-l) enhanced layer, the residual data R n ' for the next frame generated by a residual data generator
- the residual data generator 140 subtracts the prediction signal P n ' of the n enhanced layer generated by the prediction signal generator 130 from the input data D n ' of the restored block.
- the residual data R n ' of the n enhanced layer are obtained, and the obtained residual data R n ' are then input to either the prediction signal generator 130 as described above or a quantizer 150 which will be described below.
- the quantizer 150 quantizes the residual data obtained by the residual data generator
- the quantization refers to an operation of converting a Discrete Cosine Transform (DCT) coefficient expressed by a certain real value to discrete values with predetermined intervals according to a quantization table and then matching the converted discrete values with corresponding indexes.
- DCT Discrete Cosine Transform
- An entropy coder 160 generates an FGS layer bit stream through lossless coding of the quantized coefficient generated by the quantizer 150.
- the lossless coding schemes include various schemes, such as Huffman coding, arithmetic coding, variable length coding, etc.
- FIG. 8 is a block diagram of a FGS decoder 200 for decoding FGS layers by using weighted average sums according to an embodiment of the present invention.
- An entropy decoder 260 decodes an FGS layer bit stream in a video signal from the
- the FGS encoder 100 The entropy decoder 260 extracts texture data through lossless coding of the FGS layer bit stream.
- a de-quantizer 250 de-quantizes the texture data.
- the de-quantization corresponds to an inverse process of the quantization performed by the FGS encoder 100, in which values matching the indexes generated through the quantization process are restored from the indexes by using the quantization table used in the quantization process.
- the de-quantizer 250 generates the residual data R n ' of the n enhanced layer.
- a first weighted average sum calculator 210, a second weighted average sum calculator 220, and a prediction signal generator 230 in the FGS decoder 200 have the same functions as those of the first weighted average sum calculator 110, the second weighted average sum calculator 120, and the prediction signal generator 130 of the FGS encoder 100 described above, so a detailed description of the first weighted average sum calculator 210, the second weighted average sum calculator 220, and the prediction signal generator 230 will be omitted here.
- An enhanced layer restorer 240 adds the prediction signal P n ' of the n enhanced layer generated by the prediction signal generator 230 to the residual data R n ' of the n enhanced layer generated by the de-quantizer 250, thereby generating the data D n ' of the restored block of the n enhanced layer. As a result, the enhanced layer restorer 240 generates the restored FGS layer data.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
- Compression, Expansion, Code Conversion, And Decoders (AREA)
Abstract
Provided is a method of encoding FGS layers by using weighted average sums. Method includes calculating a first weighted average sum by using a restored block of n enhanced layer of a previous frame and a restored block of a base layer of a current frame; calculating a second weighted average sum by using a restored block of n enhanced layer of a next frame and a restored block of a base layer of the current frame; generating a prediction signal of n enhanced layer of the current frame by adding residual data of (n 1) enhanced layer of the current frame to a sum of the first weighted average sum and the second weighted average sum; and encoding residual data of n' enhanced layer, which is obtained by subtracting the generated prediction signal of n enhanced layer from the restored block of n' enhanced layer of the current frame.
Description
Description
METHOD AND APPARATUS FOR ENCODING/DECODING FGS LAYERS USING WEIGHTING FACTOR
Technical Field
[1] Methods and apparatuses consistent with the present invention relate to video compression technology. More particularly, the present invention relates to a method and apparatus for encoding/decoding Fine Granular Scalability (FGS) layers by using weighted average sums in a coding technology of FGS layers using an adaptive reference scheme. Background Art
[2] According to developments in information communication technologies including the Internet, multimedia services capable of supporting various types of information, such as text, image, music, etc., are increasing. Multimedia data usually have a large volume which requires a large capacity medium for storage of the data and a wide bandwidth for transmission of the data. Therefore, it is indispensable to use a compression coding scheme in order to transmit multimedia data including text, image, and audio data.
[3] The basic principle of data compression lies in a process of removing redundancy in data. Data compression can be achieved by removing the spatial redundancy such as repetition of the same color or entity in an image, the temporal redundancy such as repetition of the same sound in audio data or nearly no change between temporally adjacent pictures in a moving image stream, or the perceptional redundancy based on the fact that the human visual and perceptional capability is insensitive to high frequencies. Data compression can be classified into loss/lossless compression according to whether the source data are lost or not, in-frame/inter-frame compression according to whether the compression is independent to each frame, and symmetric/ non-symmetric compression according to whether time necessary for the compression and restoration is the same. In the typical video coding schemes, the temporal repetition is removed by temporal filtering based on motion compensation and the spatial repetition is removed by spatial transform.
[4] Transmission media, which are necessary in order to transmit multimedia data generated after redundancies in the data are removed, show various levels of performance. Currently used transmission media include media having various transmission speeds, from an ultra high-speed communication network capable of transmitting several tens of mega bit data per second to a mobile communication network having a transmission speed of 384 kbps. In such an environment, it can be
said that the scalable video coding scheme, that is, a scheme for transmitting the multimedia data at a proper data rate according to the transmission environment or in order to support transmission media of various speeds, is more proper for the multimedia environment.
[5] In a broad sense, the scalable video coding includes a spatial scalability for controlling a resolution of a video, a Signal-to-Noise Ratio (SNR) scalability for controlling a screen quality of a video, a temporal scalability for controlling a frame rate, and combinations thereof.
[6] Standardization of the scalable video coding as described above has been already progressed in Moving Picture Experts Group-21 (MPEG-4) part 10. In the work to set the standardization of the scalable video coding, there have been various efforts to implement scalability on a multi-layer basis. For example, the scalability may be based on multiple layers including a base layer, a first enhanced layer (enhanced layer 1), a second enhanced layer (enhanced layer 2), etc., which have different resolutions (QCIF, CIF, 2CIR, etc.) or different frame rates.
[7] As is in the coding with a single layer, it is necessary to obtain a Motion Vector
(MV) for removing the temporal redundancy for each layer in the coding with multilayers. The motion vector includes a motion vector (former), which is individually obtained and used for each layer, and a motion vector (latter), which is obtained for one layer and is then also used for other layers (either as it is or after up/down sampling).
[8] FIG. 1 is a view illustrating a scalable video codec using a multi-layer structure.
First, a base layer is defined to have a frame rate of Quarter Common Intermediate Format (QCIF)- 15Hz, a first enhanced layer is defined to have a frame rate of Common Intermediate Format (CIF)-30Hz, and a second enhanced layer is defined to have a frame rate of Standard Definition (SD)-60 Hz. If a CIF 0.5 Mbps stream is required, it is possible to cut and transmit the bit stream so that the bit rate is changed to 0.5 Mbps in CIF_30Hz_0.7 Mbps of the first enhanced layer. In this way, the spatial, temporal, and SNR scalability can be implemented.
[9] As noted from FIG. 1, it is possible to presume that the frames 10, 20, and 30 of respective layers having the same temporal position have similar images. Therefore, there is a known scheme in which a texture of a current layer is predicted from a texture of a lower layer either directly or through up-sampling, and a difference between the predicted value and the texture of the current layer is encoded. In "Scalable Video Model 3.0 of ISO/IEC 21000-13 Scalable Video Coding (hereinafter, referred to as SVM 3.0)," the scheme as described above is defined as an "Intra_BL prediction."
[10] As described above, the SVM 3.0 employs not only the "Inter-prediction" and the "di-
rectional intra-prediction," which are used for prediction of blocks or macro-blocks constituting a current frame in the conventional H.264, but also the scheme of predicting a current block by using a correlation between a current block and a lower layer block corresponding to the current block. This prediction scheme is called "Intra_BL prediction," and an encoding mode using this prediction is called "Intra_BL mode."
[11] FIG. 2 is a schematic view for illustrating the three prediction schemes described above, which include an intra-prediction (®) for a certain macro-block 14 of a current frame 11, an inter-prediction (©) using a macro-block 15 of a frame 12 located at a position temporally different from that of the current frame 11, and an intra_BL prediction (©) using texture data for an area 16 of a base layer frame 13 corresponding to the macro-block 14. In the scalable video coding standard as described above, one advantageous scheme is selected and used from among the three prediction schemes for each macro-block.
[12] FIG. 3 is a block diagram illustrating the concept of a conventional coding of an FGS layer according to an adaptive reference scheme. In the current H.264 SE (Scalable Extension), FGS layers of frames are encoded by using an adaptive reference scheme. Referring to FIG. 3, it is assumed that FGS layers of P frames of closed loops include a base layer, a first enhanced layer, and a second enhanced layer. Then, the FGS layers are coded by using temporal prediction signals generated by adaptively referring to both a reference frame of the base layer and a reference frame of the enhanced layer.
[13] More specifically, in order to encode a frame 62 of the second enhanced layer existing in the current frame t, it is necessary to obtain a temporal prediction signal P by calculating a weighted average of a frame 60 including reconstructed blocks of the base layer at the current frame t and a frame 50 including reference blocks of the second enhanced layer existing in the previous frame t-1 and then adding residual data R ' to the weighted average.
[14] ^p; = « x j^ - + α _ acy x jyo + j^ (D
[15] In Equation (1), α denotes a predetermined weight known as a leaky factor, D denotes a restored block of the base layer at the current frame t (that is, a block included in the frame 60), D ' denotes a restored block of the second enhanced layer at the previous frame t-1 (that is, a block included in the frame 50), and R ' denotes the residual data (generated from frame 61) of the first enhanced layer at the current frame t.
[16] By subtracting the temporal prediction signal P ' obtained by using Equation (1) from the restored block D ' at the current frame t, it is possible to obtain residual data
R ' = D ' - P ' of the second enhanced layer. Then, by quantizing and entropy-coding the calculated residual data R ', it is possible to generate a bit stream. Meanwhile, the weight α can be derived by referring to a syntax factor of the slice header. Disclosure of Invention
Technical Problem
[17] In Equation (1) showing the process of generating the prediction signal, it is possible to control drift due to partial decoding by referring to the reference frame of the base layer and is also possible to obtain a high coding efficiency by using the reference frame of the enhanced layer. However, there has been a need for a new technology for adaptively changing and using the leaky factor or the weight according to various characteristics of the block. Technical Solution
[18] Accordingly, an embodiment of the present invention has been made to solve the above-mentioned problems occurring in the prior art, and an object of the present invention is to provide a method and apparatus for encoding/decoding FGS layers by using weighted average sums, which can control drift and simultaneously improve the coding efficiency in coding of frames of all FGS layers.
[19] Further to the above object, the present invention has additional technical objects not described above, which can be clearly understood by those skilled in the art from the following description.
[20] According to an aspect of the present invention, there is provided a method of encoding FGS layers by using weighted average sums, the method including (a) calculating a first weighted average sum by using a restored block of an n enhanced layer of a previous frame and a restored block of a base layer of a current frame; (b) calculating a second weighted average sum by using a restored block of the n enhanced layer of a next frame and a restored block of a base layer of the current frame; (c) generating a prediction signal of the n enhanced layer of the current frame by adding residual data of an (n - 1) enhanced layer of the current frame to a sum of the first weighted average sum and the second weighted average sum; and (d) encoding th residual data of the n enhanced layer, which is obtained by subtracting the generated prediction signal of the n enhanced layer from the restored block of the n' enhanced layer of the current frame.
[21] According to another aspect of the present invention, there is provided a method of decoding FGS layers by using weighted average sums, the method including (a) calculating a first weighted average sum by using a restored block of an n enhanced layer of a previous frame and a restored block of a base layer of a current frame; (b) calculating a second weighted average sum by using a restored block of the n
enhanced layer of a next frame and a restored block of a base layer of the current th frame; (c) generating a prediction signal of the n enhanced layer of the current frame by adding residual data of an (n - 1 \)th enhanced layer of the current frame to a sum of the first weighted average sum and the second weighted average sum; and (d) th generating a restored block of the n enhanced layer by adding the generated prediction signal of the n enhanced layer to residual data of the n enhanced layer.
[22] According to still another aspect of the present invention, there is provided an encoder for encoding FGS layers by using weighted average sums, the encoder including a first weighted average sum calculator calculating a first weighted average sum by using a restored block of an n enhanced layer of a previous frame and a restored block of a base layer of a current frame; a second weighted average sum calculator calculating a second weighted average sum by using a restored block of the n' enhanced layer of a next frame and a restored block of a base layer of the current frame; a prediction signal generator generating a prediction signal of the n enhanced layer of the current frame by adding residual data of an (n - 1) enhanced layer of the current frame to a sum of the first weighted average sum and the second weighted average sum; and a residual data generator generating residual data of the n enhanced layer by subtracting the generated prediction signal of the n enhanced layer from the restored block of the n enhanced layer of the current frame.
[23] According to yet another aspect of the present invention, there is provided a decoder for decoding FGS layers by using weighted average sums, the decoder including a first weighted average sum calculator calculating a first weighted average sum by using a restored block of an n enhanced layer of a previous frame and a restored block of a base layer of a current frame; a second weighted average sum calculator calculating a second weighted average sum by using a restored block of the n enhanced layer of a next frame and a restored block of a base layer of the current frame; a prediction signal generator generating a prediction signal of the n enhanced layer of the current frame by adding residual data of an (n-1) enhanced layer of the current frame to a sum of the first weighted average sum and the second weighted average sum; and an enhanced layer restorer generating a restored block of the n enhanced layer by adding the generated prediction signal of the n enhanced layer to residual data of the n enhanced layer.
[24] Particulars of other embodiments are incorporated in the following description and attached drawings. Brief Description of the Drawings
[25] The above and other objects and features of the present invention will be more apparent from the following detailed description taken in conjunction with the ac-
companying drawings, in which:
[26] FIG. 1 is a view illustrating a scalable video codec using a multi-layer structure;
[27] FIG. 2 is a schematic view for illustrating three prediction schemes in a scalable video codec;
[28] FIG. 3 is a block diagram illustrating the concept of a conventional coding of an FGS layer according to an adaptive reference scheme;
[29] FIG. 4 is a flowchart illustrating the entire flow of a method of encoding FGS layers by using weighted average sums according to an exemplary embodiment of the present invention;
[30] FIG. 5 is a flowchart illustrating the entire flow of a method of decoding FGS layers by using weighted average sums according to an exemplary embodiment of the present invention;
[31] FIG. 6 illustrates the concept of an encoding of FGS layers by using weighted average sums according to an exemplary embodiment of the present invention;
[32] FIG. 7 is a block diagram of an FGS encoder 100 for encoding FGS layers by using weighted average sums according to an exemplary embodiment of the present invention; and
[33] FIG. 8 is a block diagram of an FGS decoder 200 for decoding FGS layers by using weighted average sums according to an exemplary embodiment of the present invention. Mode for the Invention
[34] Advantages and features of the present invention, and ways to achieve them will be apparent from exemplary embodiments of the present invention as will be described below together with the accompanying drawings. However, the scope of the present invention is not limited to such exemplary embodiments, and the present invention may be realized in various forms. The exemplary embodiments to be described below are nothing but the ones provided to bring the disclosure of the present invention to perfection and assist those skilled in the art to completely understand the present invention. The present invention is defined only by the scope of the appended claims. Also, the same reference numerals are used to designate the same elements throughout the specification.
[35] The present invention is described hereinafter with reference to block diagrams or flowcharts for illustrating apparatuses and methods for encoding/decoding FGS layers by using a predetermined weighted average sum according to exemplary embodiments of the present invention. It will be understood that each block of the flowchart illustrations, and combinations of blocks in the flowchart illustrations, can be implemented by computer program instructions. These computer program instructions
can be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions specified in the flowchart block or blocks. These computer program instructions may also be stored in a computer usable or computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer usable or computer-readable memory produce an article of manufacture including instruction means that implement the function specified in the flowchart block or blocks. The computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions that execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart block or blocks.
[36] And each block of the flowchart illustrations may represent a module, segment, or portion of code, which includes one or more executable instructions for implementing the specified logical function(s). It should also be noted that in some alternative implementations, the functions noted in the blocks may occur out of the order. For example, two blocks shown in succession may in fact be executed substantially concurrently or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved.
[37] As used herein, a base layer refers to a video sequence which has a frame rate lower than the maximum frame rate of a bit stream actually generated in a scalable video encoder and a resolution lower than the maximum resolution of the bit stream. In other words, the base layer has a predetermined frame rate and a predetermined solution, which are lower than the maximum frame rate and the maximum resolution, and the base layer need not have the lowest frame rate and the lowest resolution of the bit stream. Although the following description is given mainly for the macro-block, the scope of the present invention is not limited to the macro-block but can be applied to slice, frame, etc. as well as the macro-block.
[38] Further, the FGS layers may exist between the base layer and the enhanced layer.
Further, when there are two or more enhanced layers, the FGS layers may exist between a lower layer and an upper layer. As used herein, a current layer in order to obtain a prediction signal refers to the n enhanced layer, and a layer one step lower than the n enhanced layer refers to the (n-1) enhanced layer. Although the base layer is used as an example of the lower layer, it is just one embodiment and does not limit the present invention.
[39] FIG. 4 is a flowchart illustrating the entire flow of a method of encoding FGS layers by using weighted average sums according to an embodiment of the present invention. The method shown in FIG. 4 will be described hereinafter with reference to FIG. 6 which illustrates the concept of an encoding of FGS layers by using weighted average sums according to an embodiment of the present invention.
[40] First, a first weighted average sum is calculated by using a restored block 111 of the base layer of the current frame t and a restored block 103 of the n enhanced layer of the previous frame t-1 (operation S 102). The first weighted average sum can be obtained by Equation (2) below.
(2)
[42] In Equation (2), α denotes a predetermined first weight or leaky factor, D ' denotes the restored block 111 of the base layer of the current frame t, and D n ' denotes the restored block 103 of the n enhanced layer of the previous frame t- 1.
[43] After obtaining the first weighted average sum by using Equation (2), it is necessary to calculate the second weighted average sum. To this end, the second weighted average sum is calculated by using a restored block 111 of the base layer of the current frame t and a restored block 123 of the n enhanced layer of the next frame t+1 (operation S 102). The first weighted average sum can be obtained by Equation (3) below.
(3)
[45] In Equation (3), β denotes a predetermined second weight or leaky factor, D denotes the restored block 111 of the base layer of the current frame t, and D n denotes the restored block 123 of the n enhanced layer of the next frame t+1. [46] After obtaining the second weighted average sum by using Equation (3), the first weighted average sum and the second weighted average sum are added, so as to reflect both of the two weighted average sums. At this time, it is preferred, but not necessary, to calculate an arithmetic mean of the two average sums rather than to simply add the first weighted average sum and the second weighted average sum. Then, residual data of the (n-1) enhanced layer of the current frame t must be added to the arithmetic mean of the first weighted average sum and the second weighted average sum (operation S 106). Then, a prediction signal of the n enhanced layer of the current frame t is generated (operation S 108). The obtained prediction signal can be defined by Equation (4) below.
C47]
(4)
[48] In Equation (4), P ' denotes the prediction signal of the n enhanced layer of the n current frame t, and R n-l ' denotes the residual data of the (n-1) enhanced layer of the current frame t (the residual data is generated from the frame 112). [49] Finally, residual data R n 'of the n enhanced layer is obtained by subtracting the generated prediction signal P ' of the n enhanced layer of the current frame t from the n restored block D ' of the n' enhanced layer of the current frame t (R ' = D ' - P '), and n n n n is then encoded (operation Sl 10).
[50] Meanwhile, the block 112 of the (n-l) enhanced layer of the current frame t in FIG.
6 generates a prediction signal by referring to the block 102 of the previous frame t-1, the block 122 of the next frame t+1, and the block 111 of the base layer, and the block 11 of the base layer of the current frame t generates a prediction signal by referring to blocks 101 and 121 of the previous frame and the next frame.
[51] It is noted from Equation (4) that two weights or leaky factors α and β are used during the process of obtaining the prediction signal of the n enhanced layer. The first and second weights can be derived from syntax factors existing in the header of the slice including macro-blocks to be coded, and adaptively change from 0 to 1 depending on characteristic information of the macro-blocks of the n enhanced layer of the current frame t.
[52] The characteristic information includes, for example, information about prediction direction of the macro-block, information about a Coded Block Pattern (CBP) value, and information about a Motion Vector Difference (MVD) value for the macro-block.
[53] First, how the weights change according to the information about the prediction direction of the macro-block will be discussed hereinafter. When the prediction direction for partitions of the macro-block (or sub macro-block partitions) to be coded is bi-directional, the ratio of referring to the frames 103 and 123 of the n enhanced layer increases, while the ratio of referring to the frame 111 of the base layer decreases. Therefore, in Equation (4), the first weight and the second weight increase when the prediction direction is bi-directional, while the first weight and the second weight decrease when the prediction direction is uni-directional or in an intra-prediction mode.
[54] Second, how the weights change according to the information about a CBP value will be discussed hereinafter. It is presumed that it is determined from the CBP value that there are a small number of included non-zero transform coefficients. At this time, in the inter-mode in which frames located at temporally different positions are referred,
the ratio of reference between frames will increase. Therefore, the ratio of referring to the frames 103 and 123 of the n enhanced layer increases, while the ratio of referring to the frame 111 of the base layer decreases. As a result, in Equation (4), the first weight and the second weight increase in the inter-prediction mode, while the first weight and the second weight decrease in the intra-prediction mode.
[55] Third, how the weights change according to the information about an MVD value for the macro-block will be discussed hereinafter. When the MVD has a small value, the ratio of reference between frames will increase. Therefore, the ratio of referring to the frames 103 and 123 of the n enhanced layer increases, while the ratio of referring to the frame 111 of the base layer decreases. As a result, in Equation (4), the first weight and the second weight increase as the MVD value decreases, while the first weight and the second weight decrease as the MVD value increases.
[56] Hereinafter, a method of decoding FGS layers by using weighted average sums according to an embodiment of the present invention will be described with reference to FIGS. 5 and 6.
[57] First, the first weighted average sum is calculated by using the restored block 111 of the base layer of the current frame t and the restored block 103 of the n enhanced layer of the previous frame t-1 (operation S202). Then, the second weighted average sum is calculated by using the restored block 111 of the base layer of the current frame t and the restored block 123 of the n enhanced layer of the next frame t+1 (operation S204). Then, the first weighted average sum and the second weighted average sum are added and are then divided by 2, and the residual data of the (n- 1) enhanced layer of the current frame is added to the quotient of the division (operation S206), so that a prediction signal of the n enhanced layer of the current frame (operation S208). Operations S202 to S208 are similar to operations S 102 to S 108 described above in the encoding process shown in FIG. 4, so more detailed description thereof will be omitted here.
[58] When the prediction signal P ' of the n enhanced layer has been generated through n operations S202 to S208, the generated prediction signal P ' of the n enhanced layer n is added to the residual data R ' of the n' enhanced layer, thereby producing the n restored block D ' of the n' enhanced layer (D ' = P ' + R ') (operation 210). The n n n n residual data R n ' of the n enhanced layer corresponds to residual data generated as a result of decoding and de-quantization of the FGS layer bit stream generated during the encoding process. [59] Hereinafter, an encoder and a decoder for performing the encoding and decoding will be described with reference to FIGS. 7 and 8. [60] From among the elements of the invention shown in FIGS. 7 and 8, the "unit" or
"module" refers to a software element or a hardware element, such as a Field Pro-
grammable Gate Array (FPGA) or an Application Specific Integrated Circuit (ASIC), which performs a predetermined function. However, the unit or module does not always have a meaning limited to software or hardware. The module may be constructed either to be stored in an addressable storage medium or to execute one or more processors. Therefore, the module includes, for example, software elements, object-oriented software elements, class elements or task elements, processes, functions, properties, procedures, sub-routines, segments of a program code, drivers, firmware, micro-codes, circuits, data, database, data structures, tables, arrays, and parameters. The elements and functions provided by the modules may be either combined into a smaller number of elements or modules or divided into a larger number of elements or modules.
[61] FIG. 7 is a block diagram of an FGS encoder 100 for encoding FGS layers by using weighted average sums according to an embodiment of the present invention.
[62] A first weighted average sum calculator 110 calculates the first weighted average sum
by adding a product obtained by multiplying the restored block data of the n enhanced layer of the previous frame by the first weight α and a product obtained by multiplying of the restored block data of the base layer of the current frame by a value 1- α.
[63] Similarly, a second weighted average sum calculator 120 calculates the second weighted average sum
. f + l
AXDΓ + O - ^ XD; by adding a product obtained by multiplying the restored block data of the n enhanced layer of the next frame by the second weight β and a product obtained by multiplying of the restored block data of the base layer of the current frame by a value 1-β.
[64] A prediction signal generator 130 calculates an arithmetic mean of the first weighted average sum and the second weighted average sum by adding them and then dividing the sum of them by two, and then adds the residual data R n-l ' of the (n-1) enhanced layer of the current frame to the arithmetic mean, thereby obtaining the prediction signal R n ' of the n enhanced layer. For the residual data R n-l ' of the (n-l) enhanced layer, the residual data R n ' for the next frame generated by a residual data generator
140 is used. [65] Meanwhile, when data D n ' of the block of the n enhanced layer of the current frame restored by the FGS decoder 200, which will be described later, has been input to the
FGS encoder 100, the residual data generator 140 subtracts the prediction signal P n ' of the n enhanced layer generated by the prediction signal generator 130 from the input data D n ' of the restored block. As a result, the residual data R n ' of the n enhanced layer are obtained, and the obtained residual data R n ' are then input to either the prediction signal generator 130 as described above or a quantizer 150 which will be described below.
[66] The quantizer 150 quantizes the residual data obtained by the residual data generator
140. The quantization refers to an operation of converting a Discrete Cosine Transform (DCT) coefficient expressed by a certain real value to discrete values with predetermined intervals according to a quantization table and then matching the converted discrete values with corresponding indexes. The value obtained by the quantization as described above is called "quantized coefficient."
[67] An entropy coder 160 generates an FGS layer bit stream through lossless coding of the quantized coefficient generated by the quantizer 150. The lossless coding schemes include various schemes, such as Huffman coding, arithmetic coding, variable length coding, etc.
[68] FIG. 8 is a block diagram of a FGS decoder 200 for decoding FGS layers by using weighted average sums according to an embodiment of the present invention.
[69] An entropy decoder 260 decodes an FGS layer bit stream in a video signal from the
FGS encoder 100. The entropy decoder 260 extracts texture data through lossless coding of the FGS layer bit stream.
[70] A de-quantizer 250 de-quantizes the texture data. The de-quantization corresponds to an inverse process of the quantization performed by the FGS encoder 100, in which values matching the indexes generated through the quantization process are restored from the indexes by using the quantization table used in the quantization process. By the de-quantization, the de-quantizer 250 generates the residual data R n ' of the n enhanced layer.
[71] Meanwhile, a first weighted average sum calculator 210, a second weighted average sum calculator 220, and a prediction signal generator 230 in the FGS decoder 200 have the same functions as those of the first weighted average sum calculator 110, the second weighted average sum calculator 120, and the prediction signal generator 130 of the FGS encoder 100 described above, so a detailed description of the first weighted average sum calculator 210, the second weighted average sum calculator 220, and the prediction signal generator 230 will be omitted here.
[72] An enhanced layer restorer 240 adds the prediction signal P n ' of the n enhanced layer generated by the prediction signal generator 230 to the residual data R n ' of the n enhanced layer generated by the de-quantizer 250, thereby generating the data D n ' of the restored block of the n enhanced layer. As a result, the enhanced layer restorer
240 generates the restored FGS layer data.
[73] It is obvious to one skilled in the art that the scope of an apparatus for encoding/ decoding FGS layers by using weighted average sums according to the present invention as described above includes a computer-readable recoding medium on which program codes for executing the above-mentioned method in a computer are recorded. Industrial Applicability
[74] According to the present invention, it is possible to improve the coding efficiency and simultaneously control drift in the coding of frames for all FGS layers.
[75] The effects of the present invention are not limited to the above-mentioned effects, and other effects not mentioned above can be clearly understood from the definitions in the claims by one skilled in the art.
[76] Although exemplary embodiments of the present invention have been described for illustrative purposes, those skilled in the art will appreciate that various modifications, additions and substitutions are possible, without departing from the scope and spirit of the invention as disclosed in the accompanying claims. Therefore, the embodiments described above should be understood as illustrative not restrictive in all aspects. The present invention is defined only by the scope of the appended claims and must be construed as including the meaning and scope of the claims, and all changes and modifications derived from equivalent concepts of the claims.
Claims
[1] A method of encoding Fine Granular Scalability (FGS) layers by using weighted average sums, the method comprising: calculating a first weighted average sum by using a restored block of an n enhanced layer of a previous frame and a restored block of a base layer of a current frame; calculating a second weighted average sum by using a restored block of an n enhanced layer of a next frame and the restored block of the base layer of the current frame; generating a prediction signal of an n enhanced layer of the current frame by adding residual data of an (n- 1) enhanced layer of the current frame to a sum of the first weighted average sum and the second weighted average sum; and encoding residual data of the n enhanced layer, obtained by subtracting the generated prediction signal of the n enhanced layer from a restored block of the n enhanced layer of the current frame.
wherein α denotes a predetermined first weight,D ' denotes the restored block of t-l the base layer of the current frame t, and D denotes the restored block of the n th enhanced layer of the previous frame t-l.
[4] The method of claim 1, wherein the prediction signal P ' of the n enhanced n layer of the current frame is defined by:
pt = {<* * D? + (! - <*)* D[) + {β * DT + (ι - β)* D[) , p<
wherein D ' denotes the restored block of the base layer of the current frame t, D denotes the restored block of the n enhanced layer of the previous frame t-l, t+l
D denotes the restored block of the n enhanced layer of the next frame t+l,
and R n-l ' d αeennootteess t mhee rreessiidαuuaali d αaattaa oofr t mhee ( tnn-- 1i) enhanced layer of the current frame t.
[5] The method of claim 4, wherein the first weighted average sum and the second weighted average sum have values each adaptively changing from 0 to 1 depending on characteristic information of macro-blocks of the n enhanced layer of the current frame.
[6] The method of claim 5, wherein the characteristic information comprises information about prediction direction of the macro-block, and the first weight and the second weight increase when the prediction direction is bi-directional, while the first weight and the second weight decrease when the prediction direction is uni-directional or in an intra-prediction mode.
[7] The method of claim 5, wherein the characteristic information comprises information about a Coded Block Pattern (CBP) value, and, when it is determined from the CBP value that there are a small number of included non-zero transform coefficients, the first weight and the second weight increase in an inter-prediction mode, while the first weight and the second weight decrease in an intra- prediction mode.
[8] The method of claim 5, wherein the characteristic information comprises information about a Motion Vector Difference (MVD) value for the macro-block, and the first weight and the second weight increase as the MVD value decreases, while the first weight and the second weight decrease as the MVD value increases.
[9] A computer-readable recording medium having recorded with program codes for executing the method of claim 1 in a computer.
[10] A method of decoding Fine Granular Scalability (FGS) layers by using weighted average sums, the method comprising: calculating a first weighted average sum by using a restored block of an n enhanced layer of a previous frame and a restored block of a base layer of a current frame; calculating a second weighted average sum by using a restored block of the n enhanced layer of a next frame and the restored block of the base layer of the current frame; generating a prediction signal of an n enhanced layer of the current frame by adding residual data of an (n-l) enhanced layer of the current frame to a sum of the first weighted average sum and the second weighted average sum; and generating a restored block of the n enhanced layer by adding the generated prediction signal of the n enhanced layer to residual data of the n enhanced layer.
[11] The method of claim 10, wherein the first weighted average sum is obtained by: e^e x f ~> — ι— Cl — Ge y x y > wherein α denotes a predetermined first weight, D ' denotes the restored block of the base layer of the current frame t, and D '" denotes the restored block of n the n enhanced layer of the previous frame t- 1.
[13] The method of claim 10, wherein the prediction signal P ' of the n enhanced n layer of the current frame is defined by:
D« _ {« x DT + α - «) x PQ} + {Ø * DT + <i - β) * D[} , D,
1 n ~ T + K n-I
wherein D ' denotes the restored block of the base layer of the current frame t, D t-l denotes the restored block of the n enhanced layer of the previous frame t-l, denotes the restored block of the n enhanced layer of the next frame t+ and R ' denotes the residual data of the (n- 1) enhanced layer of the current
[14] The method of claim 13, wherein the first weighted average sum and the second weighted average sum have values each adaptively changing from 0 to 1 depending on characteristic information of macro-blocks of the n enhanced layer of the current frame.
[15] The method of claim 14, wherein the characteristic information comprises information about prediction direction of the macro-block, and the first weight and the second weight increase when the prediction direction is bi-directional, while the first weight and the second weight decrease when the prediction direction is uni-directional or in an intra-prediction mode.
[16] The method of claim 14, wherein the characteristic information comprises information about a Coded Block Pattern (CBP) value, and, when it is determined from the CBP value that there are a small number of included non-zero transform coefficients, the first weight and the second weight increase in an inter-prediction mode, while the first weight and the second weight decrease in an intra- prediction mode.
[17] The method of claim 14, wherein the characteristic information comprises information about a Motion Vector Difference (MVD) value for the macro-block, and the first weight and the second weight increase as the MVD value decreases, while the first weight and the second weight decrease as the MVD value increases.
[18] A computer-readable recording medium in which program codes for executing the method of claim 10 in a computer are recorded.
[19] An encoder for encoding Fine Granular Scalability (FGS) layers by using weighted average sums, the encoder comprising: a first weighted average sum calculator which calculates a first weighted average sum by using a restored block of an n enhanced layer of a previous frame and a restored block of a base layer of a current frame; a second weighted average sum calculator which calculates a second weighted average sum by using a restored block of an n enhanced layer of a next frame and the restored block of the base layer of the current frame; a prediction signal generator which generates a prediction signal of an n enhanced layer of the current frame by adding residual data of an (n-1) enhanced layer of the current frame to a sum of the first weighted average sum and the second weighted average sum; and a residual data generator which generates residual data of the nl enhanced layer by subtracting the generated prediction signal of the n enhanced layer from a restored block of the n enhanced layer of the current frame.
[20] The encoder of claim 19, wherein the first weighted average sum calculator calculates the first weighted average sum by:
GC X J∑y~ λ -+- O- — «O X JZ)0
wherein α denotes a predetermined first weight, D ' denotes the restored block of the base layer of the current frame t, and D n ' denotes the restored block of the n enhanced layer of the previous frame t- 1.
[21] The encoder of claim 19, wherein the second weighted average sum calculator calculates the second weighted average sum by:
/3 χ U>T -+- α - m x JD0
wherein β denotes a predetermined second weight, D ' denotes the restored block of the base layer of the current frame t, and D t+ denotes the restored n block of the n enhanced layer of the next frame t+ 1.
[22] The encoder of claim 19, wherein the prediction signal generator generates the prediction signal P ' of the n enhanced layer of the current frame by:
wherein D ' denotes the restored block of the base layer of the current frame t, D t-l denotes the restored block of the n enhanced layer of the previous frame t-l, denotes the restored block of the n enhanced layer of the next frame t+ and R ' denotes the residual data of the (n 1) enhanced layer of the current
[23] The encoder of claim 22, wherein the first weighted average sum and the second weighted average sum have values each adaptively changing from 0 to 1 depending on characteristic information of macro-blocks of the n enhanced layer of the current frame.
[24] The encoder of claim 23, wherein the characteristic information comprises information about prediction direction of the macro-block, and the first weight and the second weight increase when the prediction direction is bi-directional, while the first weight and the second weight decrease when the prediction direction is uni-directional or in an intra-prediction mode.
[25] The encoder of claim 23, wherein the characteristic information comprises information about a Coded Block Pattern (CBP) value, and, when it is determined from the CBP value that there are a small number of included non-zero transform coefficients, the first weight and the second weight increase in an inter-prediction mode, while the first weight and the second weight decrease in an intra- prediction mode.
[26] The encoder of claim 23, wherein the characteristic information comprises information about a Motion Vector Difference (MVD) value for the macro-block, and the first weight and the second weight increase as the MVD value decreases, while the first weight and the second weight decrease as the MVD value increases.
[27] A decoder for decoding Fine Granular Scalability (FGS) layers by using weighted average sums, the decoder comprising: a first weighted average sum calculator which calculates a first weighted average sum by using a restored block of an n enhanced layer of a previous frame and a restored block of a base layer of a current frame; a second weighted average sum calculator which calculates a second weighted
average sum by using a restored block of an n enhanced layer of a next frame and the restored block of the base layer of the current frame; a prediction signal generator which generates a prediction signal of an n enhanced layer of the current frame by adding residual data of an (n-1) enhanced layer of the current frame to a sum of the first weighted average sum and the second weighted average sum; and an enhanced layer restorer which generates a restored block of the n enhanced layer by adding the generated prediction signal of the n enhanced layer to residual data of the n enhanced layer.
[28] The decoder of claim 27, wherein the first weighted average sum calculator calculates the first weighted average sum by: cc x /)';' + (l - cf) x JJ0
wherein α denotes a predetermined first weight, D ' denotes the restored block of the base layer of the current frame t, and D n ' denotes the restored block of the n enhanced layer of the previous frame t- 1.
[29] The decoder of claim 27, wherein the second weighted average sum calculator calculates the second weighted average sum by:
wherein β denotes a predetermined second weight, D ' denotes the restored block of the base layer of the current frame t, and D n t+ denotes the restored block of the n enhanced layer of the next frame t+ 1.
[30] The decoder of claim 27, wherein the prediction signal generator generates the prediction signal P ' of the n enhanced layer of the current frame by: n
wherein D ' denotes the restored block of the base layer of the current frame t, D t-l denotes the restored block of the n enhanced layer of the previous frame t-l, denotes the restored block of the n enhanced layer of the next frame t+ and R ' denotes the residual data of the (n-1) enhanced layer of the current
[31] The decoder of claim 30, wherein the first weighted average sum and the second weighted average sum have values each adaptively changing from 0 to 1 depending on characteristic information of macro-blocks of the n enhanced
layer of the current frame.
[32] The decoder of claim 31, wherein the characteristic information comprises information about prediction direction of the macro-block, and the first weight and the second weight increase when the prediction direction is bi-directional, while the first weight and the second weight decrease when the prediction direction is uni-directional or in an intra-prediction mode.
[33] The decoder of claim 31, wherein the characteristic information comprises information about a Coded Block Pattern (CBP) value, and, when it is determined from the CBP value that there are a small number of included non-zero transform coefficients, the first weight and the second weight increase in an inter-prediction mode, while the first weight and the second weight decrease in an intra- prediction mode.
[34] The decoder of claim 31, wherein the characteristic information comprises information about a Motion Vector Difference (MVD) value for the macro-block, and the first weight and the second weight increase as the MVD value decreases, while the first weight and the second weight decrease as the MVD value increases.
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP07745762A EP2008463A2 (en) | 2006-04-06 | 2007-04-02 | Method and apparatus for encoding/decoding fgs layers using weighting factor |
MX2008012636A MX2008012636A (en) | 2006-04-06 | 2007-04-02 | Method and apparatus for encoding/decoding fgs layers using weighting factor. |
JP2009504118A JP2009532979A (en) | 2006-04-06 | 2007-04-02 | Method and apparatus for encoding and decoding an FGS layer using a weighted average |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US78958306P | 2006-04-06 | 2006-04-06 | |
US60/789,583 | 2006-04-06 | ||
KR10-2006-0069355 | 2006-07-24 | ||
KR1020060069355A KR100781525B1 (en) | 2006-04-06 | 2006-07-24 | Method and apparatus for encoding and decoding FGS layers using weighting factor |
Publications (2)
Publication Number | Publication Date |
---|---|
WO2007114622A2 true WO2007114622A2 (en) | 2007-10-11 |
WO2007114622A3 WO2007114622A3 (en) | 2007-12-13 |
Family
ID=38805228
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/KR2007/001599 WO2007114622A2 (en) | 2006-04-06 | 2007-04-02 | Method and apparatus for encoding/decoding fgs layers using weighting factor |
Country Status (7)
Country | Link |
---|---|
US (1) | US20070274388A1 (en) |
EP (1) | EP2008463A2 (en) |
JP (1) | JP2009532979A (en) |
KR (1) | KR100781525B1 (en) |
CN (1) | CN101467456A (en) |
MX (1) | MX2008012636A (en) |
WO (1) | WO2007114622A2 (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2013184810A1 (en) * | 2012-06-08 | 2013-12-12 | Qualcomm Incorporated | Bi-layer texture prediction for video coding |
US10212419B2 (en) | 2012-10-01 | 2019-02-19 | Ge Video Compression, Llc | Scalable video coding using derivation of subblock subdivision for prediction from base layer |
Families Citing this family (36)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
FR2903556B1 (en) * | 2006-07-04 | 2008-10-03 | Canon Kk | METHODS AND DEVICES FOR ENCODING AND DECODING IMAGES, A TELECOMMUNICATIONS SYSTEM COMPRISING SUCH DEVICES AND COMPUTER PROGRAMS USING SUCH METHODS |
KR20080004340A (en) * | 2006-07-04 | 2008-01-09 | 한국전자통신연구원 | Method and the device of scalable coding of video data |
US20080013623A1 (en) * | 2006-07-17 | 2008-01-17 | Nokia Corporation | Scalable video coding and decoding |
US8630355B2 (en) * | 2006-12-22 | 2014-01-14 | Qualcomm Incorporated | Multimedia data reorganization between base layer and enhancement layer |
KR100968204B1 (en) * | 2007-01-11 | 2010-07-06 | 전자부품연구원 | Method for image prediction of multi-view video codec and computer readable recording medium therefor |
US20090060035A1 (en) * | 2007-08-28 | 2009-03-05 | Freescale Semiconductor, Inc. | Temporal scalability for low delay scalable video coding |
US8326075B2 (en) | 2008-09-11 | 2012-12-04 | Google Inc. | System and method for video encoding using adaptive loop filter |
US20100104015A1 (en) * | 2008-10-24 | 2010-04-29 | Chanchal Chatterjee | Method and apparatus for transrating compressed digital video |
KR101233627B1 (en) * | 2008-12-23 | 2013-02-14 | 한국전자통신연구원 | Apparatus and method for scalable encoding |
US8503528B2 (en) | 2010-09-15 | 2013-08-06 | Google Inc. | System and method for encoding video using temporal filter |
US9532059B2 (en) | 2010-10-05 | 2016-12-27 | Google Technology Holdings LLC | Method and apparatus for spatial scalability for video coding |
US8693547B2 (en) * | 2011-04-06 | 2014-04-08 | Google Inc. | Apparatus and method for coding using motion vector segmentation |
US8780996B2 (en) | 2011-04-07 | 2014-07-15 | Google, Inc. | System and method for encoding and decoding video data |
US8780971B1 (en) | 2011-04-07 | 2014-07-15 | Google, Inc. | System and method of encoding using selectable loop filters |
US8781004B1 (en) | 2011-04-07 | 2014-07-15 | Google Inc. | System and method for encoding video using variable loop filter |
US8989256B2 (en) | 2011-05-25 | 2015-03-24 | Google Inc. | Method and apparatus for using segmentation-based coding of prediction information |
CN102209079A (en) * | 2011-06-22 | 2011-10-05 | 北京大学深圳研究生院 | Transmission control protocol (TCP)-based adaptive network control transmission method and system |
US8885706B2 (en) | 2011-09-16 | 2014-11-11 | Google Inc. | Apparatus and methodology for a video codec system with noise reduction capability |
US20130107949A1 (en) | 2011-10-26 | 2013-05-02 | Intellectual Discovery Co., Ltd. | Scalable video coding method and apparatus using intra prediction mode |
US9247257B1 (en) | 2011-11-30 | 2016-01-26 | Google Inc. | Segmentation based entropy encoding and decoding |
US9094681B1 (en) | 2012-02-28 | 2015-07-28 | Google Inc. | Adaptive segmentation |
US9131073B1 (en) | 2012-03-02 | 2015-09-08 | Google Inc. | Motion estimation aided noise reduction |
US9392274B2 (en) * | 2012-03-22 | 2016-07-12 | Qualcomm Incorporated | Inter layer texture prediction for video coding |
CN110035286B (en) * | 2012-07-09 | 2021-11-12 | Vid拓展公司 | Codec architecture for multi-layer video coding |
US9344729B1 (en) | 2012-07-11 | 2016-05-17 | Google Inc. | Selective prediction signal filtering |
US9332276B1 (en) | 2012-08-09 | 2016-05-03 | Google Inc. | Variable-sized super block based direct prediction mode |
US9380298B1 (en) | 2012-08-10 | 2016-06-28 | Google Inc. | Object-based intra-prediction |
US9467692B2 (en) | 2012-08-31 | 2016-10-11 | Qualcomm Incorporated | Intra prediction improvements for scalable video coding |
JP5952733B2 (en) * | 2012-12-28 | 2016-07-13 | 日本電信電話株式会社 | Video encoding method, video decoding method, video encoding device, video decoding device, video encoding program, video decoding program, and recording medium |
KR20140089486A (en) * | 2013-01-04 | 2014-07-15 | 삼성전자주식회사 | Motion compensation method and motion compensation apparatus for encoding and decoding of scalable video |
KR101361317B1 (en) | 2013-02-01 | 2014-02-11 | 오철욱 | System for storage section of moving picture and method thereof |
US10102613B2 (en) | 2014-09-25 | 2018-10-16 | Google Llc | Frequency-domain denoising |
CN107113425A (en) * | 2014-11-06 | 2017-08-29 | 三星电子株式会社 | Method for video coding and equipment and video encoding/decoding method and equipment |
KR102351029B1 (en) * | 2017-05-16 | 2022-01-13 | 엘지전자 주식회사 | Intra prediction mode-based image processing method and apparatus therefor |
CN108833923B (en) * | 2018-06-20 | 2022-03-29 | 腾讯科技(深圳)有限公司 | Video encoding method, video decoding method, video encoding device, video decoding device, storage medium and computer equipment |
US11943478B2 (en) * | 2019-09-19 | 2024-03-26 | Telefonaktiebolaget Lm Ericsson (Publ) | Allowing a matrix based intra prediction block to have multiple transform blocks |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6788740B1 (en) * | 1999-10-01 | 2004-09-07 | Koninklijke Philips Electronics N.V. | System and method for encoding and decoding enhancement layer data using base layer quantization data |
US6792044B2 (en) * | 2001-05-16 | 2004-09-14 | Koninklijke Philips Electronics N.V. | Method of and system for activity-based frequency weighting for FGS enhancement layers |
US20050195896A1 (en) * | 2004-03-08 | 2005-09-08 | National Chiao Tung University | Architecture for stack robust fine granularity scalability |
Family Cites Families (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6023301A (en) * | 1995-07-14 | 2000-02-08 | Sharp Kabushiki Kaisha | Video coding device and video decoding device |
JP3676525B2 (en) * | 1996-10-30 | 2005-07-27 | 日本ビクター株式会社 | Moving picture coding / decoding apparatus and method |
US6148026A (en) * | 1997-01-08 | 2000-11-14 | At&T Corp. | Mesh node coding to enable object based functionalities within a motion compensated transform video coder |
WO2001041451A1 (en) * | 1999-11-29 | 2001-06-07 | Sony Corporation | Video/audio signal processing method and video/audio signal processing apparatus |
US6690728B1 (en) * | 1999-12-28 | 2004-02-10 | Sony Corporation | Methods and apparatus for motion estimation in compressed domain |
US6510177B1 (en) * | 2000-03-24 | 2003-01-21 | Microsoft Corporation | System and method for layered video coding enhancement |
US7194035B2 (en) * | 2003-01-08 | 2007-03-20 | Apple Computer, Inc. | Method and apparatus for improved coding mode selection |
US20060012719A1 (en) * | 2004-07-12 | 2006-01-19 | Nokia Corporation | System and method for motion prediction in scalable video coding |
KR20060122671A (en) * | 2005-05-26 | 2006-11-30 | 엘지전자 주식회사 | Method for scalably encoding and decoding video signal |
-
2006
- 2006-07-24 KR KR1020060069355A patent/KR100781525B1/en not_active IP Right Cessation
-
2007
- 2007-02-02 US US11/701,392 patent/US20070274388A1/en not_active Abandoned
- 2007-04-02 EP EP07745762A patent/EP2008463A2/en not_active Withdrawn
- 2007-04-02 JP JP2009504118A patent/JP2009532979A/en active Pending
- 2007-04-02 CN CNA2007800212361A patent/CN101467456A/en active Pending
- 2007-04-02 WO PCT/KR2007/001599 patent/WO2007114622A2/en active Application Filing
- 2007-04-02 MX MX2008012636A patent/MX2008012636A/en not_active Application Discontinuation
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6788740B1 (en) * | 1999-10-01 | 2004-09-07 | Koninklijke Philips Electronics N.V. | System and method for encoding and decoding enhancement layer data using base layer quantization data |
US6792044B2 (en) * | 2001-05-16 | 2004-09-14 | Koninklijke Philips Electronics N.V. | Method of and system for activity-based frequency weighting for FGS enhancement layers |
US20050195896A1 (en) * | 2004-03-08 | 2005-09-08 | National Chiao Tung University | Architecture for stack robust fine granularity scalability |
Cited By (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2013184810A1 (en) * | 2012-06-08 | 2013-12-12 | Qualcomm Incorporated | Bi-layer texture prediction for video coding |
US10212419B2 (en) | 2012-10-01 | 2019-02-19 | Ge Video Compression, Llc | Scalable video coding using derivation of subblock subdivision for prediction from base layer |
US10212420B2 (en) | 2012-10-01 | 2019-02-19 | Ge Video Compression, Llc | Scalable video coding using inter-layer prediction of spatial intra prediction parameters |
US10218973B2 (en) | 2012-10-01 | 2019-02-26 | Ge Video Compression, Llc | Scalable video coding using subblock-based coding of transform coefficient blocks in the enhancement layer |
EP2904783B1 (en) * | 2012-10-01 | 2019-07-24 | GE Video Compression, LLC | Scalable video coding using inter-layer prediction contribution to enhancement layer prediction |
US10477210B2 (en) | 2012-10-01 | 2019-11-12 | Ge Video Compression, Llc | Scalable video coding using inter-layer prediction contribution to enhancement layer prediction |
EP3618436A1 (en) * | 2012-10-01 | 2020-03-04 | GE Video Compression, LLC | Scalable video coding using inter-layer prediction contribution to enhancement layer prediction |
US10694182B2 (en) | 2012-10-01 | 2020-06-23 | Ge Video Compression, Llc | Scalable video coding using base-layer hints for enhancement layer motion parameters |
US11477467B2 (en) | 2012-10-01 | 2022-10-18 | Ge Video Compression, Llc | Scalable video coding using derivation of subblock subdivision for prediction from base layer |
US11575921B2 (en) | 2012-10-01 | 2023-02-07 | Ge Video Compression, Llc | Scalable video coding using inter-layer prediction of spatial intra prediction parameters |
US11589062B2 (en) | 2012-10-01 | 2023-02-21 | Ge Video Compression, Llc | Scalable video coding using subblock-based coding of transform coefficient blocks in the enhancement layer |
US12010334B2 (en) | 2012-10-01 | 2024-06-11 | Ge Video Compression, Llc | Scalable video coding using base-layer hints for enhancement layer motion parameters |
Also Published As
Publication number | Publication date |
---|---|
MX2008012636A (en) | 2008-10-13 |
EP2008463A2 (en) | 2008-12-31 |
KR20070100081A (en) | 2007-10-10 |
WO2007114622A3 (en) | 2007-12-13 |
JP2009532979A (en) | 2009-09-10 |
US20070274388A1 (en) | 2007-11-29 |
CN101467456A (en) | 2009-06-24 |
KR100781525B1 (en) | 2007-12-03 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP2008463A2 (en) | Method and apparatus for encoding/decoding fgs layers using weighting factor | |
US7889793B2 (en) | Method and apparatus for effectively compressing motion vectors in video coder based on multi-layer | |
JP4891234B2 (en) | Scalable video coding using grid motion estimation / compensation | |
KR101033548B1 (en) | Video encoding method, video decoding method, video encoder, and video decoder, which use smoothing prediction | |
KR100714689B1 (en) | Method for multi-layer based scalable video coding and decoding, and apparatus for the same | |
EP2008469B1 (en) | Multilayer-based video encoding method and apparatus thereof | |
KR100703740B1 (en) | Method and apparatus for effectively encoding multi-layered motion vectors | |
US20070086520A1 (en) | Intra-base-layer prediction method satisfying single loop decoding condition, and video coding method and apparatus using the prediction method | |
US20060233250A1 (en) | Method and apparatus for encoding and decoding video signals in intra-base-layer prediction mode by selectively applying intra-coding | |
JP2011101410A (en) | Device and method for generating coded video sequence and for decoding coded video sequence while using intermediate layer residual value prediction | |
JP2006304307A (en) | Method for adaptively selecting context model for entropy coding and video decoder | |
EP1659797A2 (en) | Method and apparatus for compressing motion vectors in video coder based on multi-layer | |
KR100834757B1 (en) | Method for enhancing entropy coding efficiency, video encoder and video decoder thereof | |
WO2008007929A1 (en) | Method and apparatus for encoding and decoding video signal of fgs layer by reordering transform coefficients | |
WO2006109985A1 (en) | Method and apparatus for encoding and decoding video signals in intra-base-layer prediction mode by selectively applying intra-coding | |
JP5122288B2 (en) | Apparatus and method for generating an encoded video sequence using intermediate layer residual value prediction and decoding the encoded video sequence | |
WO2007024106A1 (en) | Method for enhancing performance of residual prediction and video encoder and decoder using the same | |
Eeckhaut et al. | A hardware-friendly wavelet entropy codec for scalable video | |
JP4327789B2 (en) | Image encoding apparatus, program, and computer-readable recording medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 200780021236.1 Country of ref document: CN |
|
WWE | Wipo information: entry into national phase |
Ref document number: MX/a/2008/012636 Country of ref document: MX |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2009504118 Country of ref document: JP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2007745762 Country of ref document: EP |