WO2006056720A1 - Compression video par modification de quantification par zones d'images - Google Patents
Compression video par modification de quantification par zones d'images Download PDFInfo
- Publication number
- WO2006056720A1 WO2006056720A1 PCT/FR2005/050982 FR2005050982W WO2006056720A1 WO 2006056720 A1 WO2006056720 A1 WO 2006056720A1 FR 2005050982 W FR2005050982 W FR 2005050982W WO 2006056720 A1 WO2006056720 A1 WO 2006056720A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- video signal
- images
- image data
- original video
- quantization
- Prior art date
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/59—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial sub-sampling or interpolation, e.g. alteration of picture size or resolution
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/124—Quantisation
- H04N19/126—Details of normalisation or weighting functions, e.g. normalisation matrices or variable uniform quantisers
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/132—Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/136—Incoming video signal characteristics or properties
- H04N19/14—Coding unit complexity, e.g. amount of activity or edge presence estimation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/17—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
- H04N19/176—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/186—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/46—Embedding additional information in the video signal during the compression process
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/85—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/90—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using coding techniques not provided for in groups H04N19/10-H04N19/85, e.g. fractals
Definitions
- the present invention relates to the field of video signal processing systems, these systems can be transmitters (camera, encoders), receivers (decoders, screens) transmission nodes, or storage where can be made a transformation of the signal , such as a transcoding.
- Video compression systems aim to reduce the volume of the video signal in order to transmit it or to store it.
- the volume of the original video signal depends on several parameters. In the first place, these parameters are those relating to the digitization of the video signal.
- the digital video signal is defined as a temporal succession of images, each image consisting of pixels, each pixel being encoded on a number of bits, this number of bits per pixel being known as color depth.
- a pixel of a color video signal is defined by its three color components that are expressed in a given color space (the RGB color space, the YUV color space, etc.).
- the parameters relating to the digitization of the video signal are the image frequency, the spatial resolution of the image in pixels and the number of bits describing the pixel on each of its color components.
- the current compression techniques do indeed provide for modifying the quantization of the video signal, but this modification takes place in the frequency domain, on the DCT (Discrete Cosine Transform) coefficients, for the "discrete cosine transformation", of the signal .
- the data is spatially decorrelated and is no longer related to the major characteristic color depth parameter of the screen to display the video signal.
- the quantization of the DCT coefficients is less effective than it could be since it does not take into account the strong spatial correlation or the application context that represents the screen of the terminal.
- the method according to the invention takes advantage of the spatial correlation as well as it takes advantage of the temporal correlation.
- MPEG-4 ISO-IEC 14496-2 defines tools that make it possible to indicate that a portion of a given image has been compressed with a spatial resolution reduced by half in each of the horizontal and vertical directions relative to the resolution of the video sequence.
- the MPEG standard plans to analyze the input signal to identify the VOP (Video Object Planes) corresponding to specific objects in the image, and change the spatial resolution associated with certain VOPs.
- the standard provides for a uniform modification of the YUV components.
- PCT application WO 03/107678 describes a non-uniform modification of these components and an associated syntax.
- VOPs as defined in MPEG-4 are obtained semi-automatically, and the definition of a VOP uses an operator.
- the modification of the spatial resolution according to the VOPs therefore requires this intervention which is not compatible with an efficient compression method.
- the modification of the spatial resolution according to these VOPs is not the most suitable modification for the compression.
- US Patent 6,490,319 to Yang relating to a video processing method for controlling the contrast between a region of interest of a video image, and the background of a video image.
- a video controller allows the user to control the levels of quantification of the area of interest and the background of the image.
- An interface is provided to the user to enable him to perform such control. This interface is presented as a slider.
- the region of interest can be defined by the user or depend on the video content. In any case, the region of interest for which the contrast can be changed by changing the quantization level is set before the quantization level is changed. The division according to a region of interest therefore does not depend on the level of quantification that will be applied to the region of interest.
- this coding method has the disadvantage of not being automatic and depending on the action of the user on the image. This in particular results in the fact that only two distinct quantization levels are applicable to the video image. On the other hand, if the user wishes to modify the region of interest and / or the quantization level of the image, he must make a new selection of the image and operate the quantization cursor again.
- This encoding method therefore has the disadvantage of not being automatically applied to the video signal. It also has the disadvantage of not being temporally variable, depending on the number of the image in the video signal.
- the present invention therefore intends to solve the disadvantages of the prior art as defined above.
- the present invention relates to a method of encoding an original video signal comprising a temporal succession of images comprising data space-quantized images for obtaining a modified video signal, characterized in that it comprises the steps of:
- said cutting of the images is variable according to the number of the image in said temporal succession of images, which makes it possible to adapt the quantization to the image succession.
- said new quantization is variable according to the number of the image in said temporal succession of images, and this, whether the cutting remains identical or not, again to adapt the compression temporally.
- the method according to the invention further comprises a step of analyzing said original video signal, and said division and / or said new quantification depend on said analysis.
- the selected areas and the associated quantization are optimized for compression.
- the analysis of the images concerns at least the spatial proximity, the movement characteristics and the residual energy characteristics of the image blocks.
- said division and / or said new quantization are predetermined. According to yet another variant of the method, said division and / or said new quantization depend on an action external to the method, such as, for example, the characteristics of the receiving terminal.
- the new quantization concerns the spatial resolution according to the image components.
- said images comprise a set of digital image data sampled in pixels within said image, and said new quantization is a resampling of the spatial resolution according to at least one of said digital image data, the data digital images being for example the color components.
- the new quantization relates to the number of bits on which the image data are coded, and in this case, said images comprise a set of digital data sampled in pixels, and are quantized over a number of bits. of origin, and said new quantization is a new quantization of said number of bits according to at least one of said digital image data or the manner of quantizing the image data on these bits, the digital image data being for example the color components.
- said images comprise a set of digital data sampled in pixels, and quantized over a number of original bits, and said new quantization is a new quantization of said number of bits. bits according to at least one of said data digital images or how to quantify the image data on these bits, and a resampling of the spatial resolution according to at least one of said digital image data, for example the color components.
- said images of the original video signal are cut into a plurality of predefined size blocks, and in that said image areas obtained during the cutting step correspond to a plurality of said adjacent blocks.
- the invention also relates to an apparatus for coding an original video signal comprising a temporal succession of images comprising spatially quantized image data for obtaining a modified video signal, characterized in that it comprises means for receiving said original video signal, means for splitting the images of said video signal into a set of image areas, means for generating a new quantization of the image data in the spatial domain, variable according to the zones of said set of zones, for obtaining a modified video signal.
- the invention also relates to a recording medium on which a series of images of a video scene coded according to the method according to the invention are stored.
- FIG. 1 generally illustrates the description of a video signal in images
- FIG. 2a is an example of a division of a succession of images; in image areas in which quantization varies,
- FIG. 2b represents a VOP slicing as defined in the prior art
- FIG. 3 is a diagram illustrating the implementation of the invention in an encoder, the FIG. 4 details the compression entity described in FIG. 3
- FIG. 5 gives a more detailed description of the invention according to an embodiment that can be envisaged in a video coder taking up the compression scheme based on spatial prediction and temporal prediction
- FIG. 6 gives a description of the quantization modification process according to the invention
- FIGS. 7a and 7b are examples of algorithms for the definition of the zones according to the invention.
- Table 1 describes the variables taken into account in an exemplary algorithm implemented by the present invention.
- the subject of the invention is a method for modifying the quantization of the color components of a video signal in order to improve the video compression / decompression systems and more generally the video signal processing systems.
- the video signal is described in Figure 1.
- This figure gives a structure of a video sequence as commonly used today.
- the video sequence of structure as described in the figure, consists of a periodic succession of I- images of the same size HxL.
- j indicates the time reference.
- An image is composed of pixels, the number of which defines the spatial resolution of the image.
- an image at a time j is an array of HxL pixels.
- a video signal can be described as a temporal succession of two-dimensional or three-dimensional images.
- the periodicity of the temporal succession, as well as the geometry and the image size can be variable. These images can result from a natural shooting through a sensor such as a camera, or a computer program for image synthesis, or drawings, or any other material and / or software and / or human. These images can equally be formed of a combination of images generated separately and combined by any editing process.
- a color dot is represented by three components: the RGB components for example.
- the RGB components for example.
- Other triplets of components than those mentioned above are defined and can be used just as much, these triplets of components being linked to each other by a linear combination.
- the color components C1, C2 and C3 In the example image structure considered in Figure 1, we consider the simplest, namely each pixel P x , y of the image has a value given by a triplet C1, C2 and C3.
- the three color components C1, C2 and C3 are quantized over a generally identical number of bits. Quantification is usually linear.
- the object of the invention is to modify the quantization of C1, C2 and C3 for a group of zones Z k into a predefined quantization Qi, and for a given duration of images, the duration of images ranging from one image to the number of images constituting the total video sequence.
- a zone Z k will have its components quantized chronologically according to, for example, the type Qi during 100 images, then Q m during 1000 images, then Q n during 50 images etc. It is the object of this invention that the zones can be made and discarded as and when advancement in the video sequence.
- the modification of the quantization relates to the spatial resolution of the original video signal.
- the spatial resolution is modified according to at least one of the color components.
- this modification may be variable depending on the zones obtained by the method of the invention, and depending on the image or image group chosen, for a temporal variation. In this way, a spatio-temporal variation of the spatial resolution of the video signal is obtained.
- quantization modification will be used to refer to the modifications relating to the number of bits on which the color components are quantized or how to quantify the image data on these bits, which corresponds to to a quantification in depth, the modifications relating to the spatial resolution of the image, which corresponds to a spatial quantization as well as the modifications on these two types of quantization in a temporal manner, therefore according to the index of the image.
- quantization modifications applied to the number of quantization bits are also applicable to a modification of the spatial resolution, or to a combination of these two quantization modifications.
- control entity can be based on different tools. These tools allow the analysis of the video signal which consists of studying each of the components forming the video signal to be modified, in a global manner-that is to say on the image-or local-that is to say on an area of the image.
- This analysis of the video signal calculates its complexity, namely its texture, likewise this analysis relates to its range, namely the set of values taken by each of the components, likewise this analysis relates to its cleanliness, therefore to the power of the noise included in the signal, likewise this analysis relates to its variation, in particular temporal, that is to say from one image to the following ones. This analysis step will be described in more detail later.
- the quality of service sought is a decision element.
- the entity may use all or part of these tools.
- Other decision criteria can be used as well.
- Figure 1 depicts a structure of the video as a temporal succession of images (j being the time index).
- Each image is an array of HxL pixels (along x and y axes), each pixel P x , y of coordinates (x, y) can be described at most by three values Cl x , y (j), C2 x , y (j), C3 x , y (j).
- FIG. 2 describes the invention by giving an example of image zones Z k .
- the controlling entity makes the following decisions with regard to area-based quantifications after modification
- Quantization Q 3 applied on Z 2 , Z 5 and Z 6 Quantification Q 4 applied on Z 3 and Z 7
- Qi Cl, C2 and C3 quantized linearly over 8 bits
- Q 3 C quantized linearly over 6 bits
- Q 4 C quantized nonlinearly over 6 bits
- FIG. 3 a general view is given of the example of an embodiment of the invention, in which the video signal is caused to be compressed by an encoder.
- the signal is received by the control entity that analyzes the signal to generate elements techniques such as its complexity, the range of values taken by each component, its cleanliness, its temporal variation, etc. This analysis can be done after the "Pre-treatment" process.
- the same original video signal composed by Cl, C2 and C3 is transformed by the "Preprocessing” entity.
- the output of this entity is also a video signal described by three color component signals.
- These signals are compressed by the "Compression” entity which generates a signal that is called a bitstream.
- this compression entity may comprise functions of MPEG or H.26x compression standards, such as spatial prediction, temporal prediction, transformation (DCT, DWT, etc.), quantization coefficients, entropy coding, to name only the main ones.
- the control entity by analyzing the original video signal and / or other parameters, calculates areas
- the control entity reduces the quantization of each of the components C1, C2 and C3.
- the reduction consists of reducing the number of bits describing Cl, C2 and C3 and thus of changing to a number
- N1, N2 and N3 (for each component) less than 8, 8 being the quantification assumed in the original video signal.
- the quantization modification takes place on signals representing each of the components C1, C2 and C3, as much in the entity of "pre-treatment" as in that of
- the control entity adds information to the outgoing bitstream of the "Compression" entity. such as those describing the zones and their quantization, as well as scene change information. It is understood that this additional information can be conveyed differently from the sender to the receiver, and they may not be routed, depending on the application. Other information can be used just as much.
- the raster bit stream enriched or not this information is then transmitted or stored.
- FIG. 5 a more detailed example of an embodiment in which the video signal is made to be compressed by an encoder is given.
- the quantization modification is performed in the compression entity.
- the compression entity is assumed to be based on a schema as detailed in Figure 4.
- Figure 5 only the main functions and only the main signals are shown.
- the control signals, the additional information signals (including in particular the choice of quantization modification, the zones formed in the image), the signal of the motion vectors, the storage entities ("Buf Im") are not indicated.
- the signal (1) of the original video passes through the Preprocessing entity (A) and becomes the signal (2) to be compressed.
- This signal is processed by the Spatial Prediction entity (B) and the Temporal Prediction entity (F).
- the Spatial Prediction entity (B) generates three signals: the signal (3) Image, the signal (10) Spatial Prediction and the signal (4) Residual Spatial.
- the signal Image is the set of pixels of the original image transmitted to the other compression steps (DCT, DWT or other type of equivalent transformation, quantization, reorganization, entropy encoding).
- the signal Spatial Prediction is the set of pixels considered as good prediction for spatially adjacent pixels and used as such.
- the Residual Spatial signal is the set of pixels represented by the difference between their original value and the value of the corresponding Prediction pixel.
- the signals (3) and (4) are then processed by the Quantization Modification entity (C) which is the subject of the present invention.
- the two resulting signals (5) and (6) are respectively the Image signal and the Spatial Residual signal after Modification of Quantization. These two signals are then transmitted to the following compression steps (DCT, DWT or other type of equivalent transformation, quantization, reorganization, entropy encoding).
- DCT compression steps
- DWT or other type of equivalent transformation, quantization, reorganization, entropy encoding
- the purpose of this entity is to operate the inverse function of the quantization modification entity (C).
- the resulting signals (7) and (8) are respectively the Image and Spatial Residual signal after resizing and are directed to the Reconstruction entity (E) which likewise receives the signal (10).
- the objective of (E) is to re-form the original pixels, in particular for the residual pixels. For said residual pixels, this re-formation of the original pixels is based on a simple addition of the value of the residual pixel and the value of the corresponding prediction pixel.
- the signal (9) Reconstructed Spatial then serves as a signal of reference for Spatial Prediction (B) as well as signal (2) Image.
- this signal (8) serves as a reference signal for the temporal prediction (F).
- the Time Prediction (F) entity generates two signals: the Spatiotemporal Prediction (15) signal and the Spatio-Temporal Residual (11) signal.
- the Spatio-temporal Prediction signal is the set of pixels considered as good prediction for spatio-temporally adjacent pixels and used as such.
- the Spatio-Temporal Residual signal is the set of pixels represented by the difference between their original value and the value of the corresponding Prediction pixel.
- the signal (11) is then processed by the Quantification Modification entity (G) object of the present invention.
- the resulting signal (12) is therefore the residual Spatio-temporal signal after quantization modification.
- This signal is then transmitted to the following compression steps (DCT, DWT or other type of equivalent transformation, quantization, reorganization, entropy encoding).
- This same signal (12) is also processed by the entity (D) of Scaling. The purpose of this entity is to operate the inverse function of the quantization modification entity (G).
- the resulting signal (13) is the Spatio-Temporal Residual signal after Scaling and is directed to the Reconstruction entity (E) which likewise receives the Prediction signal (15).
- the Spatiotemporal Rebuilt signal (14) then serves as a reference signal for the Temporal Prediction (F) as well as the signal (2) Image. Likewise, this Reconstructed signal (14) serves as a reference signal for Spatial Prediction (B).
- the Reconstruction entity (E) also receives signals (30) from the other functions of compression, including signals that have traveled the path DCT, Q, Q "1 and DCT “ 1 .
- Output Input x Q
- Input is the value of a pixel at the input of the entity (D) and Output is the value of this same pixel at the output, Q being a constant whose value varies under the control of a control entity such as shown in Figure 3.
- this entity can, in addition to the re-forming function of the original pixel as already described, include other useful functions but not indispensable.
- this entity can, in addition to the re-forming function of the original pixel as already described, include other useful functions but not indispensable.
- the standardization function can be described according to the formula:
- P represents the pixel after uniformization function
- P ' represents the pixel before uniformization function
- x, y are the spatial coordinates of the pixel
- j is the temporal index of the pixel (the index of the image to which it belongs)
- x m , Yn and x m ' , Yn ' are the spatial coordinates of pixels, pixels taken into account in the uniformization function
- ji ' and ji are the temporal indices of pixels, pixels taken into account in the function of uniformity
- C is a linear combination of all or part of pixels.
- FIG. 6 gives a description of the Quantification Modification phase and its surrounding elements.
- the quantization modification can be done on the color component signals passing through the Preprocessing entity as shown in FIG. it can be done on the input color component signals of the two spatial and temporal prediction entities, or it can be done on the signals to be processed by the following compression steps (DCT, DWT or other equivalent transformation type) , quantization, reorganization, entropy encoding) as described in Figure 5, etc.
- the Quantization Editing Signal Selection and Quantization Editing entity (H) receives three signals: a signal (16), this signal being generally the Image signal ((2) according to FIG.
- Entity (J) is the entity that performs Quantification Modification. This entity
- the entity (J) therefore receives the signal (17) already mentioned and the signal (22) carrying the choice of quantization changes from the entity (H).
- the entity (K) is a simple Multiplexer, controlled through the signal (21) from the entity (H).
- the entity (K) will switch to the output (signal (19)) either the modified quantization signal (18) or the original signal (17). This switching takes place at the pixel rate.
- the signal (19) to be processed by the following entities and the signal (22) carrying the quantization modification choices.
- the quantization modification takes place at least on the residual signals of the pixels located in the areas concerned.
- the pixels that predict the neighboring pixels may not be affected by the quantization change.
- the signal (17) is the same as the signal (16).
- the signals (16) and (17) are distinct.
- the information conveyed by the signal (22) is taken again to be inserted possibly in the bit stream at the output of the Screening entity as shown in FIG. 3, and this as part of the additional information such as as previously described, just as this signal (22) can be used by other entities of the compression.
- the quantization modification takes place at least on the residual signals of the pixels situated in the zones concerned.
- the pixels that predict the neighboring pixels may not be affected by the quantization change.
- the images are cut into blocks of predefined sizes.
- the coder decides for each block of the type of prediction noted Type_Pred and the mode noted Mode_Pred for this prediction which will be operated on it.
- the motion estimator receives a video signal. It calculates for the blocks to be predicted temporally the error on the block between the pixels of the potential predictor blocks and the pixels of the current block. It then retains as block predictor the one that gives the minimum error, which will be noted Pred error. He deduces the corresponding motion vector, which will be noted MV.
- Error_Pred For spatially intra-predicted blocks, a prediction error that will also be noted Error_Pred is usually calculated to choose among the different spatial prediction modes.
- the signal after passing through a) i.e. by quantization modification and rescaling functions or by track b) that is, by the modification functions of quantization, DCT, quantization and its inverse function coefficients and rescaling, is reconstructed by the entity (E).
- the reconstructed signal along one of two channels a) or b) is compared to the original signal, and an error between the two signals is calculated. This error will be noted Error_a_Posteriori.
- each block of an image at a given time index has a set of variables that are: Type_Pred, Mode_Pred, Error_a_Posteriori, Error_Pred and MV, except the intra blocks for which the variable MV does not exist.
- Err ⁇ ur_a_Post ⁇ r ⁇ or ⁇ Err ⁇ ur_Pr ⁇ d and MV thresholds respectively S_EaPi, S_Ej and S_MVk.
- Type_Pred and Mode_Pred have meanwhile discrete values defined in the cited norms.
- Table 1 above is a summary table of the variables of a block and their values taken for the zone formed of block (s). Other possibilities may be provided such replace the average values by sums, in which case the thresholds at which the values obtained by these sums will be compared will have to be multiplied by the factor corresponding to the number of pixels included in the blocks in the zone.
- external actions are considered, such as the characteristics of the screen of the target terminal, user configurations (for example the operator), a command issued by the debit controller entity, requests from the network for more or less debit and more generally context information of the application.
- external parameters discrete values are defined.
- FIGS. 7a and 7b describe an example of zone formation algorithm.
- This algorithm is an automatic zone renewal algorithm in each image.
- the principle of this algorithm is that neighboring blocks spatially are gathered in a zone if their Error Preds and their MVs are close in values. For this, the blocks are "scanned" in the natural sense.
- a block may have as neighbors several areas formed of at least one block.
- the difference of the MVs is then compared with a first threshold S_MV1, if this difference does not exceed the threshold, the block is included in the zone and the variables of this zone are updated by calculating the average values; if this difference exceeds the threshold, we compare the signs of MV of the zone and MV of the block: if the signs are different, we go to the next neighboring zone, otherwise we continue the process.
- the algorithm part described in FIG. 7b is applied: the given block is included in any neighboring zone having the same type and the same mode of prediction as this block, if the difference of Error_Pred of the block and of Pred error of the zone does not exceed S E2 in absolute value.
- An option not described in FIGS. 7a and 7b provides for selecting from the zones that meet the criteria for inclusion of the processed block the zone for which the value Err ⁇ ur_Pr ⁇ d is the lowest. To implement this option, for example, it will suffice during the processing of a block, to record the data relating to each area meeting the inclusion criteria and then to compare the values of the Pred errors and to make the right inclusion decision. If more than one zone meets this criterion, then these zones will only form a new zone including the processed block, and the variables of this zone will be updated by calculating the average values.
- any block that does not find a neighboring area that meets the inclusion criteria forms a new area.
- a new quantization is chosen according to the time index of the zone, the external actions and the result of the comparison tests of the variables Type_Pred, Mode_Pred, MV and Error_Pred with the discrete values that the first two variables are likely to take or with the respective thresholds S_MVk and S_Ej for the last two variables. It is well understood that the new quantification chosen may be the same as the original one.
- the method that is the subject of this invention as well as the various steps constituting it can be implemented in a hardware and / or software way, in any video signal processing system, video signal transmission or video signal storage.
- the invention relates more particularly to systems where the compression as well as the optimization of the video signal with respect to the screens of the terminals are of primary importance.
Abstract
Description
Claims
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
FR0412425A FR2878395A1 (fr) | 2004-11-23 | 2004-11-23 | Procede de traitement de signal video |
FR0412425 | 2004-11-23 | ||
FR0505036 | 2005-05-19 | ||
FR0505036A FR2878384A1 (fr) | 2004-11-23 | 2005-05-19 | Compression video par modification de quantification par zones d'images |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2006056720A1 true WO2006056720A1 (fr) | 2006-06-01 |
Family
ID=36029454
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/FR2005/050982 WO2006056720A1 (fr) | 2004-11-23 | 2005-11-23 | Compression video par modification de quantification par zones d'images |
Country Status (2)
Country | Link |
---|---|
FR (1) | FR2878384A1 (fr) |
WO (1) | WO2006056720A1 (fr) |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH04266286A (ja) * | 1991-02-21 | 1992-09-22 | Seiko Epson Corp | テレビ電話用映像信号圧縮装置 |
US5162898A (en) * | 1988-10-06 | 1992-11-10 | Sharp Kabushiki Kaisha | Color image data compressing apparatus and method |
EP0541302A2 (fr) * | 1991-11-08 | 1993-05-12 | AT&T Corp. | Quantification améliorée du signal vidéo pour un environnement de codage MPEG ou similaires |
US5396292A (en) * | 1992-02-26 | 1995-03-07 | Nec Corporation | Encoding apparatus for motion video signals |
JPH0879757A (ja) * | 1994-08-31 | 1996-03-22 | Sanyo Electric Co Ltd | 動画像符号化装置 |
US6490319B1 (en) * | 1999-06-22 | 2002-12-03 | Intel Corporation | Region of interest video coding |
WO2003107678A1 (fr) * | 2002-06-18 | 2003-12-24 | Koninklijke Philips Electronics N.V. | Procede de codage video et dispositifs de codage et de decodage correspondants |
-
2005
- 2005-05-19 FR FR0505036A patent/FR2878384A1/fr active Pending
- 2005-11-23 WO PCT/FR2005/050982 patent/WO2006056720A1/fr not_active Application Discontinuation
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5162898A (en) * | 1988-10-06 | 1992-11-10 | Sharp Kabushiki Kaisha | Color image data compressing apparatus and method |
JPH04266286A (ja) * | 1991-02-21 | 1992-09-22 | Seiko Epson Corp | テレビ電話用映像信号圧縮装置 |
EP0541302A2 (fr) * | 1991-11-08 | 1993-05-12 | AT&T Corp. | Quantification améliorée du signal vidéo pour un environnement de codage MPEG ou similaires |
US5396292A (en) * | 1992-02-26 | 1995-03-07 | Nec Corporation | Encoding apparatus for motion video signals |
JPH0879757A (ja) * | 1994-08-31 | 1996-03-22 | Sanyo Electric Co Ltd | 動画像符号化装置 |
US6490319B1 (en) * | 1999-06-22 | 2002-12-03 | Intel Corporation | Region of interest video coding |
WO2003107678A1 (fr) * | 2002-06-18 | 2003-12-24 | Koninklijke Philips Electronics N.V. | Procede de codage video et dispositifs de codage et de decodage correspondants |
Non-Patent Citations (5)
Title |
---|
MURCHING A M ET AL: "Adaptive subsampling of color images", PROCEEDINGS OF THE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP) AUSTIN, NOV. 13 - 16, 1994, LOS ALAMITOS, IEEE COMP. SOC. PRESS, US, vol. VOL. 3 CONF. 1, 13 November 1994 (1994-11-13), pages 963 - 966, XP010146490, ISBN: 0-8186-6952-7 * |
MUSMANN H G ET AL: "ADVANCES IN PICTURE CODING", PROCEEDINGS OF THE IEEE, IEEE. NEW YORK, US, vol. 73, no. 4, 1 April 1985 (1985-04-01), pages 523 - 548, XP002057834, ISSN: 0018-9219 * |
PATENT ABSTRACTS OF JAPAN vol. 017, no. 057 (E - 1315) 4 February 1993 (1993-02-04) * |
PATENT ABSTRACTS OF JAPAN vol. 1996, no. 07 31 July 1996 (1996-07-31) * |
SINGHAL A ET AL: "SPATIALLY ADAPTIVE RENDERING OF IMAGES FOR DISPLAY ON MOBILE DEVICES", PROCEEDINGS. IS&T'S PICS CONFERENCE. ANNUAL CONFERENCE, PROCEEDINGS OF CONFERENCE OF THE SOCIETY FOR IMAGING SCIENCE AND TECHNOLOGY, XX, XX, 13 May 2003 (2003-05-13), pages 360 - 365, XP008047825 * |
Also Published As
Publication number | Publication date |
---|---|
FR2878384A1 (fr) | 2006-05-26 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9565439B2 (en) | System and method for enhancing data compression using dynamic learning and control | |
EP1766997B1 (fr) | Procédé et dispositif de codage d'une sequence d'images vidéo en coefficients de sous-bandes de fréquences de differentes resolutions spatiales | |
EP3318061B1 (fr) | Procede d'encodage d'image et equipement pour la mise en oeuvre du procede | |
EP2981088A1 (fr) | Procédé de codage et décodage d'images, dispositif de codage et décodage et programmes d'ordinateur correspondants | |
EP3225029B1 (fr) | Procede d'encodage d'image et equipement pour la mise en oeuvre du procede | |
EP3442228A1 (fr) | Procédé de décodage d'images | |
EP2279621B1 (fr) | Procédé de codage, de décodage, codeur et décodeur | |
EP1721470A1 (fr) | Procede de codage et de decodage d'une sequence d'images par analyse temporelle hierarchique | |
FR2951345A1 (fr) | Procede et dispositif de traitement d'une sequence video | |
KR102312337B1 (ko) | Ai 부호화 장치 및 그 동작방법, 및 ai 복호화 장치 및 그 동작방법 | |
Lee et al. | DPICT: Deep progressive image compression using trit-planes | |
EP3139608A1 (fr) | Procédé de compression d'un flux de données vidéo | |
EP2443835A1 (fr) | Codage de vecteurs mouvement par competition de predicteurs | |
EP1702473B1 (fr) | Procede de codage d une sequence d images | |
FR2955995A1 (fr) | Procede et dispositif de traitement d'une sequence video | |
EP3202147B1 (fr) | Procédé de transcodage de données vidéo à fusion d'unités de codage, programme informatique, module de transcodage et équipement de télécommunications associés | |
HUE033524T2 (en) | Procedure for compressing data | |
WO2006056720A1 (fr) | Compression video par modification de quantification par zones d'images | |
WO2018073523A1 (fr) | Procédé de codage et de décodage de paramètres d'image, dispositif de codage et de décodage de paramètres d'image et programmes d'ordinateur correspondants | |
CN116074528A (zh) | 视频编码方法及装置、编码信息调度方法及装置 | |
FR2957744A1 (fr) | Procede de traitement d'une sequence video et dispositif associe | |
Liu et al. | Foveation embedded DCT domain video transcoding | |
FR2907989A1 (fr) | Procede et dispositif d'optimisation de la compression d'un flux video | |
Nasiri | Machine learning based optimization for VVC low bitrate coding | |
Chen et al. | Vesper: Learning to Manage Uncertainty in Video Streaming |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AK | Designated states |
Kind code of ref document: A1 Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BW BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE EG ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KM KN KP KR KZ LC LK LR LS LT LU LV LY MA MD MG MK MN MW MX MZ NA NG NI NO NZ OM PG PH PL PT RO RU SC SD SE SG SK SL SM SY TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW |
|
AL | Designated countries for regional patents |
Kind code of ref document: A1 Designated state(s): BW GH GM KE LS MW MZ NA SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IS IT LT LU LV MC NL PL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 05819413 Country of ref document: EP Kind code of ref document: A1 |
|
WWW | Wipo information: withdrawn in national office |
Ref document number: 5819413 Country of ref document: EP |