GB2519070A - Data encoding and decoding - Google Patents

Data encoding and decoding Download PDF

Info

Publication number
GB2519070A
GB2519070A GB1317406.5A GB201317406A GB2519070A GB 2519070 A GB2519070 A GB 2519070A GB 201317406 A GB201317406 A GB 201317406A GB 2519070 A GB2519070 A GB 2519070A
Authority
GB
United Kingdom
Prior art keywords
rdpcm
array
data values
video data
data
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
GB1317406.5A
Other versions
GB201317406D0 (en
Inventor
James Alexander Gamei
Karl James Sharman
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Sony Corp
Original Assignee
Sony Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Sony Corp filed Critical Sony Corp
Priority to GB1317406.5A priority Critical patent/GB2519070A/en
Priority to GB1317803.3A priority patent/GB2519094A/en
Priority to GBGB1318593.9A priority patent/GB201318593D0/en
Publication of GB201317406D0 publication Critical patent/GB201317406D0/en
Priority to GB1320772.5A priority patent/GB2519177A/en
Priority to PCT/GB2014/052917 priority patent/WO2015049491A1/en
Priority to JP2016519369A priority patent/JP6609753B2/en
Priority to CN201480054385.8A priority patent/CN105612744B/en
Priority to US15/024,705 priority patent/US10939103B2/en
Priority to RU2016117252A priority patent/RU2679991C2/en
Priority to EP14777128.1A priority patent/EP3039867B1/en
Publication of GB2519070A publication Critical patent/GB2519070A/en
Withdrawn legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/90Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using coding techniques not provided for in groups H04N19/10-H04N19/85, e.g. fractals
    • H04N19/91Entropy coding, e.g. variable length coding [VLC] or arithmetic coding
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T9/00Image coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • H04N19/107Selection of coding mode or of prediction mode between spatial and temporal predictive coding, e.g. picture refresh
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/124Quantisation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/129Scanning of coding units, e.g. zig-zag scan of transform coefficients or flexible macroblock ordering [FMO]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/13Adaptive entropy coding, e.g. adaptive variable length coding [AVLC] or context adaptive binary arithmetic coding [CABAC]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/132Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/136Incoming video signal characteristics or properties
    • H04N19/137Motion inside a coding unit, e.g. average field, frame or block difference
    • H04N19/139Analysis of motion vectors, e.g. their magnitude, direction, variance or reliability
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/146Data rate or code amount at the encoder output
    • H04N19/147Data rate or code amount at the encoder output according to rate distortion criteria
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/157Assigned coding mode, i.e. the coding mode being predefined or preselected to be further used for selection of another element or parameter
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/157Assigned coding mode, i.e. the coding mode being predefined or preselected to be further used for selection of another element or parameter
    • H04N19/159Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/176Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/513Processing of motion vectors
    • H04N19/517Processing of motion vectors by encoding
    • H04N19/52Processing of motion vectors by encoding by predictive encoding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/587Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal sub-sampling or interpolation, e.g. decimation or subsequent interpolation of pictures in a video sequence
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/593Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial prediction techniques
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • H04N19/61Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/85Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression
    • H04N19/88Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression involving rearrangement of data among different coding units, e.g. shuffling, interleaving, scrambling or permutation of pixel data or permutation of transform coefficient data among different blocks
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/90Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using coding techniques not provided for in groups H04N19/10-H04N19/85, e.g. fractals
    • H04N19/94Vector quantisation

Abstract

A method and associated apparatus for encoding/decoding video data using a residual differential pulse code modulation (RDCPM) encoder/decoder and an entropy encoder/decoder to entropy encode/decode the RDPCM data in a processing order such that any RDPCM data value in the processing order is independent on a RDPCM value later in the processing order. Further disclosed is a method and apparatus for encoding/decoding data using RDPCM and entropy encoding in which the processing order comprises a horizontal or vertical scan. In a further embodiment of the invention the RDPCM decoder may clip the RDPCM decoded data to an acceptable numerical range. The invention may be used in a High Efficiency Video Coding (HEVC) compliant codec to facilitate the efficient use of a rate-distortion-optimised quantisation (RDOQ) scheme in conjunction with RDPCM coding.

Description

DATA ENCODING AND DECODING
Field of the Disclosure
This disclosure relates to data encoding and decoding.
Description of the Related Art
The "background" description provided herein is for the purpose of generally presenting the context of the disclosure. Work of the presently named inventors, to the extent it is described in this background section, as well as aspects of the description which may not otherwise qualify as prior art at the time of filing, are neither expressly or impliedly admitted as prior art against the present disclosure.
There are several video data compression and decompression systems which involve transforming video data into a frequency domain representation, quantising the frequency domain coefficients and then applying some form of entropy encoding to the quantised coefficients.
The transformation into the spatial frequency domain at the encoder side corresponds to an inverse transformation at the decoder side. Example transformations include the so-called discrete cosine transformation (DCT) and the so-called discrete sine transformation (DST). In some examples the transformations are carried out by matrix-multiplying an array of input samples (derived from the video data to be coded) by a matrix of transformation coefficients to generate frequency-transformed data.
Frequency-transformed data is converted back to sample data, from which output video data can be derived, by matrix-multiplying an array of the frequency-transformed data by a matrix of inverse-transformation coefficients.
Some standards and draft standards, such as the so-called High Efficiency Video Coding (HEVC) standards, define encoding and decoding modes in which a spatial frequency transformation is not in fact used. These are sometimes referred to as "transform skip" or "trans-quant skip" ("trans-quant" being an abbreviation for "transformation and quantisation") modes.
Summary
An aspect of this disclosure is defined by claim 1.
Further respective aspects and features are defined in the appended claims.
It is to be understood that both the foregoing general description and the following detailed description are exemplary, but not restrictive of, the present
disclosure.
Brief Description of the Drawings
A more complete appreciation of the disclosure and many of the attendant advantages thereof will be readily obtained as the same becomes better understood by reference to the following detailed description of embodiments, when considered in connection with the accompanying drawings, wherein: Figure 1 schematically illustrates an audio/video (AN) data transmission and reception system using video data compression and decompression; Figure 2 schematically illustrates a video display system using video data decompression; Figure 3 schematically illustrates an audio/video storage system using video data compression and decompression; Figure 4a schematically illustrates a video camera using video data compression; Figure 4b schematically illustrates an example video camera in more detail; Figure 4c schematically illustrates another example video camera; Figures 4d and 4e schematically illustrate date carriers; Figure 5 provides a schematic overview of a video data compression and decompression apparatus; Figure 6 schematically illustrates the generation of predicted images; Figure 7 schematically illustrates a largest coding unit ([CU); Figure 8 schematically illustrates a set of four coding units (CU); Figures 9 and 10 schematically illustrate the coding units of Figure 8 sub-divided into smaller coding units; Figure 11 schematically illustrates an array of prediction units (PU); Figure 12 schematically illustrates an array of transform units (TU); Figure 13 schematically illustrates an array of samples; Figure 14 schematically illustrates an array of frequency-separated coefficients; Figure 15 schematically illustrates a trend within the array of Figure 14; Figure 16 schematically illustrates a reverse scan; Figures 17a-17c schematically illustrate rounding techniques; Figure 18 schematically illustrates an RDOQ apparatus; Figures 19a and 19b schematically illustrate the operation of the apparatus of Figure 18; Figure 20 schematically illustrates part of an encoder; Figures 21 and 22 schematically illustrate an RDPCM operation; Figures 23a and 23b schematically illustrate trends in RDPCM data arrays; Figures 24a and 24b schematically illustrate changes to a scanning order for vertically encoded RDPCM data; Figures 25a and 25b schematically illustrate changes to a scanning order for horizontally encoded RDPCM data; Figures 26a and 26b schematically illustrate changes to a scanning order for inter-image encoded RDPCM data; Figure 27 is a schematic flowchart illustrating a method of processing; Figure 28 schematically illustrates part of a decoder; Figures 29-31 are schematic flowcharts indicating aspects of the operation of an encoder andlor decoder; Figure 32 schematically indicates a part of an encoder and/or decoder operating in respect of intra-image encoded blocks; Figure 33 schematically indicates a part of an encoder and/or decoder operating in respect of inter-image encoded blocks; Figure 34 schematically indicates a part of an encoder and/or decoder operating in respect of intra-image encoded blocks; and Figure 35 schematically indicates a part of an encoder and/or decoder operating in respect of inter-image encoded blocks.
Description of the Embodiments
Referring now to the drawings, Figures 1-4e are provided to give schematic illustrations of apparatus or systems making use of the compression and/or decompression apparatus to be described below in connection with embodiments.
All of the data compression and/or decompression apparatus is to be described below may be implemented in hardware, in software running on a general-purpose data processing apparatus such as a general-purpose computer, as programmable hardware such as an application specific integrated circuit (ASIC) or field programmable gate array (FPGA) or as combinations of these. In cases where the embodiments are implemented by software and/or firmware, it will be appreciated that such software and/or firmware, and non-transitory machine-readable data storage media by which such software and/or firmware are stored or otherwise provided, are considered as embodiments.
Figure 1 schematically illustrates an audio/video data transmission and reception system using video data compression and decompression.
An input audio/video signal 10 is supplied to a video data compression apparatus which compresses at least the video component of the audio/video signal 10 for transmission along a transmission route 30 such as a cable, an optical fibre, a wireless link or the like. The compressed signal is processed by a decompression apparatus 40 to provide an output audio/video signal 50. For the return path, a compression apparatus 60 compresses an audio/video signal for transmission along the transmission route 30 to a decompression apparatus 70.
The compression apparatus 20 and decompression apparatus 70 can therefore form one node of a transmission link. The decompression apparatus 40 and decompression apparatus 60 can form another node of the transmission link. Of course, in instances where the transmission link is uni-directional, only one of the nodes would require a compression apparatus and the other node would only require a decompression apparatus.
Figure 2 schematically illustrates a video display system using video data decompression. In particular, a compressed audio/video signal 100 is processed by a decompression apparatus 110 to provide a decompressed signal which can be displayed on a display 120. The decompression apparatus 110 could be implemented as an integral part of the display 120, for example being provided within the same casing as the display device. Alternatively, the decompression apparatus 110 might be provided as (for example) a so-called set top box (STB), noting that the expression "set-top" does not imply a requirement for the box to be sited in any particular orientation or position with respect to the display 120; it is simply a term used in the art to indicate a device which is connectable to a display as a peripheral device.
Figure 3 schematically illustrates an audio/video storage system using video data compression and decompression. An input audio/video signal 130 is supplied to a compression apparatus 140 which generates a compressed signal for storing by a store device 150 such as a magnetic disk device, an optical disk device, a magnetic tape device, a solid state storage device such as a semiconductor memory or other storage device. For replay, compressed data is read from the store device 150 and passed to a decompression apparatus 160 for decompression to provide an output audio/video signal 170.
It will be appreciated that the compressed or encoded signal, and a storage medium or data carrier storing that signal, are considered as embodiments. Reference is made to Figures 4d and 4e described below.
Figure 4a schematically illustrates a video camera using video data compression.
In Figure 4a, and image capture device 180, such as a charge coupled device (COD) image sensor and associated control and read-out electronics, generates a video signal which is passed to a compression apparatus 190. A microphone (or plural microphones) 200 generates an audio signal to be passed to the compression apparatus 190. The compression apparatus 190 generates a compressed audio/video signal 210 to be stored and/or transmitted (shown generically as a schematic stage 220).
The techniques to be described below relate primarily to video data compression.
It will be appreciated that many existing techniques may be used for audio data compression in conjunction with the video data compression techniques which will be described, to generate a compressed audio/video signal. Accordingly, a separate discussion of audio data compression will not be provided. It will also be appreciated that the data rate associated with video data, in particular broadcast quality video data, is generally very much higher than the data rate associated with audio data (whether compressed or uncompressed). It will therefore be appreciated that uncompressed audio data could accompany compressed video data to form a compressed audio/video signal. It will further be appreciated that although the present examples (shown in Figures 1-4e) relate to audio/video data, the techniques to be described below can find use in a system which simply deals with (that is to say, compresses, decompresses, stores, displays and/or transmits) video data. That is to say, the embodiments can apply to video data compression without necessarily having any associated audio data handling at all.
Figure 4b schematically illustrates an example video camera apparatus 183 in more detail. Those features numbered in common with Figure 4a will not be described further. Figure 4b is an example of the camera of Figure 4a (in the case that the unit 220 of Figure 4a provides a storage capability) in which the compressed data are first buffered by a buffer 221 and then stored in a storage medium 222 such as a magnetic disk, an optical disk, flash memory, a so-called solid-state disk drive (SSD) or the like.
Note that the arrangement of Figure 4b can be implemented as a single (physical) unit 182.
Figure 4c schematically illustrates another example video camera in which, in place of the storage arrangement of Figure 4b, a network interface 223 is provided in order to allow the compressed data to be transmitted to another unit (not shown). The network interface 223 can also allow for incoming data to be received by the video camera, such as control data. Note that the arrangement of Figure 4b can be implemented as a single (physical) unit 183.
Figures 4d and 4e schematically illustrate data carriers, for example for use as the storage medium 222 and carrying compressed data which has been compressed according to the compression techniques described in the present application. Figure 4d shows a schematic example of a removable non-volatile storage medium 225 implemented as solid state memory such as flash memory. Figure 4e shows a schematic example of a removable non-volatile storage medium 226 implemented as a disk medium such as an optical disk.
Figure 5 provides a schematic overview of a video data compression and decompression apparatus.
Successive images of an input video signal 300 are supplied to an adder 310 and to an image predictor 320. The image predictor 320 will be described below in more detail with reference to Figure 6. The adder 310 in fact performs a subtraction (negative addition) operation, in that it receives the input video signal 300 on a "+" input and the output of the image predictor 320 on a "-" input, so that the predicted image is subtracted from the input image. The result is to generate a so-called residual image signal 330 representing the difference between the actual and projected images.
One reason why a residual image signal is generated is as follows. The data coding techniques to be described, that is to say the techniques which will be applied to the residual image signal, tends to work more efficiently when there is less "energy" in the image to be encoded. Here, the term "efficiently" refers to the generation of a small amount of encoded data; for a particular image quality level, it is desirable (and considered "efficient") to generate as little data as is practicably possible. The reference to "energy" in the residual image relates to the amount of information contained in the residual image. If the predicted image were to be identical to the real image, the difference between the two (that is to say, the residual image) would contain zero information (zero energy) and would be very easy to encode into a small amount of encoded data. In general, if the prediction process can be made to work reasonably well, the expectation is that the residual image data will contain less information (less energy) than the input image and so will be easier to encode into a small amount of encoded data.
The residual image data 330 is supplied to a transform unit 340 which generates a discrete cosine transform (DCT) representation of the residual image data. The DCT technique itself is well known and will not be described in detail here. There are however aspects of the techniques used in the present apparatus which will be described in more detail below, in particular relating to the selection of different blocks of data to which the DCT operation is applied. These will be discussed with reference to Figures 7-12 below.
Note that in some embodiments, a discrete sine transform (DST) is used instead of a DCT. In other embodiments, no transform might be used. This can be done selectively, so that the transform stage is, in effect, bypassed, for example under the control of a "transform skip" command or mode.
The output of the transform unit 340, which is to say, a set of transform coefficients for each transformed block of image data, is supplied to a quantiser 350.
Various quantisation techniques are known in the field of video data compression, ranging from a simple multiplication by a quantisation scaling factor through to the application of complicated lookup tables under the control of a quantisation parameter.
The general aim is twofold. Firstly, the quantisation process reduces the number of possible values of the transformed data. Secondly, the quantisation process can increase the likelihood that values of the transformed data are zero. Both of these can make the entropy encoding process work more efficiently in generating small amounts of compressed video data.
A controller 345 controls the operation of the transform unit 340 and the quantiser 350 (and their respective inverse units), according to techniques to be discussed further below. Note that the controller 345 may also control other aspects of the operation of the apparatus of Figure 5.
A data scanning process is applied by a scan unit 360. The purpose of the scanning process is to reorder the quantised transformed data so as to gather as many as possible of the non-zero quantised transformed coefficients together, and of course therefore to gather as many as possible of the zero-valued coefficients together. These features can allow so-called run-length coding or similar techniques to be applied efficiently. So, the scanning process involves selecting coefficients from the quantised transformed data, and in particular from a block of coefficients corresponding to a block of image data which has been transformed and quantised, according to a "scanning order" so that (a) all of the coefficients are selected once as part of the scan, and (b) the scan tends to provide the desired reordering. One example scanning order which can tend to give useful results is a so-called zigzag scanning order.
The scanned coefficients are then passed to an entropy encoder (EE) 370.
Again, various types of entropy encoding may be used. Two examples are variants of the so-called CABAC (Context Adaptive Binary Arithmetic Coding) system and variants of the so-called CAVLC (Context Adaptive Variable-Length Coding) system. In general terms, CABAC is considered to provide a better efficiency, and in some studies has been shown to provide a 10-20% reduction in the quantity of encoded output data for a comparable image quality compared to CAVLC. However, CAVLC is considered to represent a much lower level of complexity (in terms of its implementation) than CABAC. Note that the scanning process and the entropy encoding process are shown as separate processes, but in fact can be combined or treated together. That is to say, the reading of data into the entropy encoder can take place in the scan order.
Corresponding considerations apply to the respective inverse processes.
The output of the entropy encoder 370, along with additional data, for example defining the manner in which the predictor 320 generated the predicted image, provides a compressed output video signal 380.
However, a return path is also provided because the operation of the predictor 320 itself depends upon a decompressed version of the compressed output data.
The reason for this feature is as follows. At the appropriate stage in the decompression process a decompressed version of the residual data is generated. This decompressed residual data has to be added to a predicted image to generate an output image (because the original residual data was the difference between the input image and a predicted image). In order that this process is comparable, as between the compression side and the decompression side, the predicted images generated by the predictor 320 should be the same during the compression process and during the decompression process. Of course, at decompression, the apparatus does not have access to the original input images, but only to the decompressed images. Therefore, at compression, the predictor 320 bases its prediction (at least, for inter-image encoding) on decompressed versions of the compressed images.
The entropy encoding process carried out by the entropy encoder 370 is considered to be "lossless", which is to say that it can be reversed to arrive at exactly the same data which was first supplied to the entropy encoder 370. So, the return path can be implemented before the entropy encoding stage. Indeed, the scanning process carried out by the scan unit 360 is also considered lossless, but in the present embodiment the return path 390 is from the output of the quantiser 350 to the input of a complimentary inverse quantiser 420.
In general terms, an entropy decoder 410, the reverse scan unit 400, an inverse quantiser 420 and an inverse transform unit 430 provide the respective inverse functions of the entropy encoder 370, the scan unit 360, the quantiser 350 and the transform unit 340. For now, the discussion will continue through the compression process; the process to decompress an input compressed video signal corresponds to the return path of the compression process.
In the compression process, the scanned coefficients are passed by the return path 390 from the quantiser 350 to the inverse quantiser 420 which carries out the inverse operation of the scan unit 360. An inverse quantisation and inverse transformation process are carried out by the units 420, 430 to generate a compressed-decompressed residual image signal 440.
The image signal 440 is added, at an adder 450, to the output of the predictor 320 to generate a reconstructed output image 460. This forms one input to the image predictor 320.
Turning now to the process applied to a received compressed video signal 470, the signal is supplied to the entropy decoder 410 and from there to the chain of the reverse scan unit 400, the inverse quantiser 420 and the inverse transform unit 430 before being added to the output of the image predictor 320 by the adder 450. In straightforward terms, the output 460 of the adder 450 forms the output decompressed video signal 480. In practice, further filtering may be applied before the signal is output.
Figure 6 schematically illustrates the generation of predicted images, and in particular the operation of the image predictor 320.
There are two basic modes of prediction: so-called intra-image prediction and so-called inter-image, or motion-compensated (MC), prediction.
Intra-image prediction bases a prediction of the content of a block of the image on data from within the same image. This corresponds to so-called I-frame encoding in other video compression techniques. In contrast to I-frame encoding, where the whole image is intra-encoded, in the present embodiments the choice between intra-and inter-encoding can be made on a block-by-block basis, though in other embodiments the choice is still made on an image-by-image basis.
Motion-compensated prediction makes use of motion information which attempts to define the source, in another adjacent or nearby image, of image detail to be encoded in the current image. Accordingly, in an ideal example, the contents of a block of image data in the predicted image can be encoded very simply as a reference (a motion vector) pointing to a corresponding block at the same or a slightly different position in an adjacent image.
Returning to Figure 6, two image prediction arrangements (corresponding to intra-and inter-image prediction) are shown, the results of which are selected by a multiplexer 500 under the control of a mode signal 510 so as to provide blocks of the predicted image for supply to the adders 310 and 450. The choice is made in dependence upon which selection gives the lowest "energy" (which, as discussed above, may be considered as information content requiring encoding), and the choice is signalled to the encoder within the encoded output data stream. Image energy, in this context, can be detected, for example, by carrying out a trial subtraction of an area of the two versions of the predicted image from the input image, squaring each pixel value of the difference image, summing the squared values, and identifying which of the two versions gives rise to the lower mean squared value of the difference image relating to that image area.
The actual prediction, in the intra-encoding system, is made on the basis of image blocks received as part of the signal 460, which is to say, the prediction is based upon encoded-decoded image blocks in order that exactly the same prediction can be made at a decompression apparatus. However, data can be derived from the input video signal 300 by an intra-mode selector 520 to control the operation of the intra-image predictor 530.
For inter-image prediction, a motion compensated (MC) predictor 540 uses motion information such as motion vectors derived by a motion estimator 550 from the input video signal 300. Those motion vectors are applied to a processed version of the reconstructed image 460 by the motion compensated predictor 540 to generate blocks of the inter-image prediction.
The processing applied to the signal 460 will now be described. Firstly, the signal is filtered by a filter unit 560. This involves applying a "deblocking" filter to remove or at least tend to reduce the effects of the block-based processing carried out by the transform unit 340 and subsequent operations. Also, an adaptive loop filter is applied using coefficients derived by processing the reconstructed signal 460 and the input video signal 300. The adaptive loop filter is a type of filter which, using known techniques, applies adaptive filter coefficients to the data to be filtered. That is to say, the filter coefficients can vary in dependence upon various factors. Data defining which filter coefficients to use is included as part of the encoded output data stream.
The filtered output from the filter unit 560 in fact forms the output video signal 480. It is also buffered in one or more image stores 570; the storage of successive images is a requirement of motion compensated prediction processing, and in particular the generation of motion vectors. To save on storage requirements, the stored images in the image stores 570 may be held in a compressed form and then decompressed for use in generating motion vectors. For this particular purpose, any known compression I decompression system may be used. The stored images are passed to an interpolation filter 580 which generates a higher resolution version of the stored images; in this example, intermediate samples (sub-samples) are generated such that the resolution of the interpolated image is output by the interpolation filter 580 is 8 times (in each dimension) that of the images stored in the image stores 570. The interpolated images are passed as an input to the motion estimator 550 and also to the motion compensated predictor 540.
In embodiments, a further optional stage is provided, which is to multiply the data values of the input video signal by a factor of four using a multiplier 600 (effectively just shifting the data values left by two bits), and to apply a corresponding divide operation (shift right by two bits) at the output of the apparatus using a divider or right-shifter 610.
So, the shifting left and shifting right changes the data purely for the internal operation of the apparatus. This measure can provide for higher calculation accuracy within the apparatus, as the effect of any data rounding errors is reduced.
The way in which an image is partitioned for compression processing will now be described. At a basic level, and image to be compressed is considered as an array of blocks of samples. For the purposes of the present discussion, the largest such block under consideration is a so-called largest coding unit ([CU) 700 (Figure 7), which represents a square array of 64 x 64 samples. Here, the discussion relates to luminance samples. Depending on the chrominance mode, such as 4:4:4, 4:2:2, 4:2:0 or 4:4:4:4 (GBR plus key data), there will be differing numbers of corresponding chrominance samples corresponding to the luminance block.
Three basic types of blocks will be described: coding units, prediction units and transform units. In general terms, the recursive subdividing of the LCUs allows an input picture to be partitioned in such a way that both the block sizes and the block coding parameters (such as prediction or residual coding modes) can be set according to the specific characteristics of the image to be encoded.
The [CU may be subdivided into so-called coding units (CU). Coding units are always square and have a size between 8x8 samples and the full size of the LCU 700.
The coding units can be arranged as a kind of tree structure, so that a first subdivision may take place as shown in Figure 8, giving coding units 710 of 32x32 samples; subsequent subdivisions may then take place on a selective basis so as to give some coding units 720 of 16x16 samples (Figure 9) and potentially some coding units 730 of 8x8 samples (Figure 10). Overall, this process can provide a content-adapting coding tree structure of CU blocks, each of which may be as large as the [CU or as small as 8x8 samples. Encoding of the output video data takes place on the basis of the coding unit structure.
Figure 11 schematically illustrates an array of prediction units (PU). A prediction unit is a basic unit for carrying information relating to the image prediction processes, or in other words the additional data added to the entropy encoded residual image data to form the output video signal from the apparatus of Figure 5. In general, prediction units are not restricted to being square in shape. They can take other shapes, in particular rectangular shapes forming half of one of the square coding units, as long as the coding unit is greater than the minimum (8x8) size. The aim is to allow the boundary of adjacent prediction units to match (as closely as possible) the boundary of real objects in the picture, so that different prediction parameters can be applied to different real objects. Each coding unit may contain one or more prediction units.
Figure 12 schematically illustrates an array of transform units (TU). A transform unit is a basic unit of the transform and quantisation process. Transform units are always square and can take a size from 4x4 up to 32x32 samples. Each coding unit can contain one or more transform units. The acronym SDIP-P in Figure 12 signifies a so-called short distance intra-prediction partition. In this arrangement only one dimensional transforms are used, so a 4xN block is passed through N transforms with input data to the transforms being based upon the previously decoded neighbouring blocks and the previously decoded neighbouring lines within the current SDIP-P.
Figure 13 schematically illustrates an array of samples. These in fact represent samples of the residual image 330 discussed above, and are drawn at their appropriate relative spatial positions, which is to say that a sample 800 lies at the upper left corner of an array of 4 x 4 adjacent samples, and a sample 810 lies at the bottom right corner of such an array, with respect to positions of those samples within an image.
It will of course be appreciated that the 4 x 4 array of Figure 13 is just an example; the techniques and attributes discussed here can apply to arrays of various different sizes such asS x 8, 16 x 16, 32 x 32 and so on.
The samples of Figure 13 are processed by the transform unit 340 to generate frequency-separated coefficients. Figure 14 schematically illustrates an array of such frequency-separated coefficients. Here, the position of a coefficient within the array represents the spatial frequencies corresponding to that coefficient. By convention, a so-called DC coefficient 820 occupies the upper-left array position. Moving towards the right within the array of Figure 14 indicates an increasing horizontal frequency component, and moving towards the bottom of the array of Figure 14 indicates an increasing vertical frequency component. Note that although it is just a convention to represent the array of coefficients in this manner (rather than, for example, having the DC coefficient in the bottom-right corner), the ordering of the coefficients is technically relevant to other parts of the process. One reason for this is schematically illustrated by a broken-line arrow 830 in Figure 14, which indicates that moving from the bottom-right to the top-left positions within the array of Figure 14 corresponds to a generally increasing importance of the coefficients to image reconstruction. That is to say, in order to reconstruct the array or block of samples of Figure 13, the most important one of the coefficients of Figure 14 is the DC coefficient 820, followed in terms of importance by the lower horizontal and vertical frequency coefficients.
In general terms, this trend of importance can also correspond to a trend in terms of coefficient magnitude, in that the magnitude of the DC coefficient can tend to be the largest within the set of coefficients derived from a block of samples. Figure 15 schematically illustrates such a trend within the array of Figure 14, in which smaller values tend to be towards the lower right of the array and larger values tend to be towards the upper left of the array. Of course, a specific individual array of coefficients may differ from this general trend.
One way in which the trends discussed above are technically relevant relates to the entropy encoding stage carried out by the entropy encoder 370 in Figure 5. A general principle applied to this type of entropy encoding is that it operates more efficiently by dealing with smaller-magnitude data values within a block before dealing with larger-magnitude data values. Accordingly, in the case that frequency separated coefficients are subjected to entropy encoding (whether or not an intervening quantisation stage is used) a so-called "reverse scan" pattern is used to select data in an appropriate order for entropy encoding. Figure 16 schematically illustrates an example of a reverse scan. This example relates to a so-called reverse-diagonal scan S which starts with the coefficient 840 in the lower-right corner of the array of coefficients and progresses up to the DC coefficient 820 according to the diagonal scanning pattern shown. This pattern means that, generally speaking, smaller value coefficients are encoded before larger value coefficients. This is an example of the entropy encoder is operable in the second mode to apply a processing order starting from an array position corresponding to that of a highest frequency component and ending at an array position corresponding to that of a lowest frequency component.
The order in which the coefficients are handled is also relevant to a so-called rate-distortion-optimised quantisation (RDOQ) process. In order to explain this, firstly some aspects of the quantisation process will be described, and then the RDOQ system will be discussed.
In embodiments, quantisation is applied as a division of the coefficient magnitudes (the coefficient signs being preserved separately) in a transform unit (TV) by a quantisation divisor qStep derived from a quantisation parameter applicable to that block, Qp, according to: qstep = baseaStep[Qp % 6] cc mt (Qp/6) Here, the percentage sign % indicates a modulus function, so that A % B is equal to the remainder when A is divided by B. The cc sign indicates a left bit-shift by the value given after the cc sign, so that, for example, cc 3 represents a left shift by three bit positions. The variable baseastep is a function of an index, varying between 0 and 5, represented by Op % 6. So, a (potentially) different value of baseQStep applies according to the relevant index Qp % 6. The function mt indicates an integer value.
For ease of calculation, the reciprocal of baseOStep is pre-calculated in some embodiments of an encoder, and left-shifted by a number of bits such as 14 bits, to give values of inverseQStep which are then multiplied by the coefficient magnitudes. A reason for this is that multiplication operations are more straightforward to implement in some arrangements than division operations.
This gives a combined operation as follows: output = ((input x inverseoStep) + rounding)>> qBits where qBits = 14 + mt (Qp16) + transformShift Here, transformShift represents a compensation for any additional shifting imposed by the DCT process.
In some embodiments, so-called scaling lists are used as a mechanism to further vary the degree of quantisation applicable to different coefficients in the TLJ. Each coefficient position is associated with a corresponding entry in the scaling list. The scaling list value is used to alter the quantisation divisor. In some embodiments, a scaling list value of decimal 16 corresponds to no change, which is to say that it leaves the divisor unchanged. In some embodiments, the scaling list values are used to modify the value inverseoStep used in the equations above, so that: inverseOStep = (inverseoStep c< 4)! scaling list value A variable "rounding" was introduced above. Before the right-shift by qBits, a value is added in order to round the final result. Rounding can be applied by a number of methods. Figures 17a-17c schematically illustrate example rounding techniques.
Note that these diagrams show a typical rounding point of 0.5, but other rounding points may be used. Figure 17a schematically illustrates so-called positive rounding, in which a value which is between two integers is rounded to the greater of the two surrounding integers. Figure 17b schematically illustrates so-called infinity rounding, in which a value which is between two integers is rounded to the nearest integer of the higher magnitude. Figure 17c schematically illustrates so-called even rounding, in which a value which is between two integers is rounded to the nearest even integer. Note that even rounding is not used in embodiments of the HEVC system.
These are all examples of conventional rounding. As an alternative, RDOQ can be used. This is a technique for selecting a rounding regime based upon an estimate of a cost function, so as to aim towards achieving a minimum cost for the selected regime, where the cost depends upon both the error obtained using that rounding regime and also the number of bits required to encode data rounded according to that rounding regime.
Figure 18 schematically illustrates an RDOQ apparatus comprising a tester 850 which executes tests of various rounding regimes as discussed below, and a selector 860 which selects a rounding regime providing the lowest cost function. Note that in practice, these functions can be carried out by the controller 345. Figures 1 9a and 1 9b schematically illustrate the operation of the apparatus of Figure 18 and will be described below.
As mentioned, RDOQ can be used instead of conventional rounding. The steps for a TU are as follows: For each quantised coefficient of magnitude X, candidate integers are determined as follows: X1 = mt (X) and X2 = int(X) + 1, if (X%1)> 0.5 = mt (X)-1 and X2 = int(X), otherwise If X2 is less than 3, a third candidate X0 is provided. X0 has the value of 0.
If X is less than 0.5, X0 is selected. Otherwise, for each of the candidate integers, an absolute error (error(X)) from the true value of X is calculated, the number of bits requires to encode the candidate is estimated, and the cost function of each candidate X is calculated as: cost(X1) = ((error (X1) <c qBits)2 x errorScale) + (bits (X1) x lambda) Here, errorScale is a scaling factor applied is a weighting against the square of the error, and lambda is a scaling factor applied as a weighting against the estimate of the number of bits.
The candidate with the lowest cost is selected.
Figures 1 9a and 1 9b schematically illustrate examples of the error values, in the case that X1 = 3 and X2 = 4 (Figure 19a) and in the case that X1 = 2 and X2 = 3 (Figure 19b). In both of these example illustrations, X2 is clearly the better choice from merely the point of view of reducing errors, but if X1 requires fewer bits than X2 to encode, then X1 may in fact be the better choice overall. This choice is determined by calculation of the cost function.
For each group of coefficients in the TU, a cost is calculated for the case where all coefficients are quantised to 0. If this is lower than the total cost of coding the group, all 16 coefficients are replaced by 0.
The TU is scanned in reverse-scan order, corresponding to the order of entropy encoding as discussed above. For each coefficient X, the cost C of having X as the last non-zero coefficient in the TU is calculated. The coefficient XL with the lowest cost C is selected to be the last. A check is included such that XL cannot precede (in scan order) any coefficient with a magnitude greater than 1.
Figure 20 schematically illustrates part of an encoder and represents a modification of the operation of the transform unit 340 of Figure 5. Note that a corresponding modification would be provided to the inverse transform unit 430.
A selector 900, which may in fact be implemented as part of the functionality of the controller 345, is provided to select (using a schematic switch 910) between three modes of operation, namely a mode in which a frequency transform is used (routing the data via the transform unit 340), a transform skip mode in which no transform is employed, and a so-called RDPCM (residual differential pulse code modulation) mode using an RDFCM unit 342. Data which has been handled by any of the three routes is combined by an output multiplexer 920. Note that the selector 900 may also control part of the operation of the entropy encoder 370; this feature will be discussed further below.
Note also that the transform operation can be omitted in a so-called trans-quant bypass mode, in which neither the transform nor the quantisation operations are applied.
Figures 21 and 22 schematically illustrate an RDPCM operation. The operation is applied to a block of residual data 330, an example of which is schematically illustrated in Figure 21. In Figure 21, an example block of 4 x 4 samples is used, but the techniques are applicable to other block sizes. Each sample in the block is given a unique letter reference from A to P in order that the effect of the RDPCM process may be illustrated in Figure 22.
RDPCM may be applied in a horizontal (left to right) sense or a vertical (top to bottom) sense. Figure 22 schematically illustrates the horizontal RDPCM process, but it will be appreciated that the same techniques are applicable to the vertical process.
Referring to Figure 22, a left-hand column of samples (A, E, I, M) remain unchanged by the coding process. After that, each successive sample along a row is encoded in terms of the difference between that sample and a reconstructed version of the sample to its left. Here, the term "reconstructed" refers to the value of that sample after it has been encoded and then decoded (so that both the encode and decode are working with respect to data available to the decoder). An asterisk (*) is used to signify a reconstructed version of a sample so that, for example, B*represents a reconstructed version of the sample B. So, the sample B is encoded by the value (B -A*), the sample K is encoded by the value (K -J*) and so on.
RDPCM coding can be used in intra-image-encoding or inter-image-encoding. In examples of the HEVC system, in intra-image-encoding, horizontal RDPCM is used in a horizontal prediction mode, which in turn determines the scan order as a reverse vertical scan, and vertical RDPCM is used in a vertical prediction mode, which again in turn determines this can order as a reverse horizontal scan. In inter-image-encoding, either horizontal or vertical RDPCM may be selected (for example, by the control at 345 in response to calculation of a cost function) and the scan mode is always reverse diagonal.
Figures 23a and 23b schematically illustrate trends in RDPCM encoded data arrays. In particular, Figure 23a relates to horizontal RDPCM encoding, and schematically illustrates the fact that the left-hand column 930 of data are encoded unchanged (and so can have any value), and then remaining data tend to have lower values as they represent only difference quantities. Figure 23b schematically illustrates the case for vertical RDPCM encoding, in which the uppermost row 940 of data values are encoded unchanged, whereas the remaining data values are represented by differences which would tend to be lower in magnitude.
A significant aspect of the RDPCM data relates to the dependency of encoded data items upon previously encoded data items. In particular, each column of samples in horizontal RDPCM cannot be encoded or decoded without the reconstructed values of the immediately left-hand column, and each row of samples in vertical RDPCM cannot be encoded or decoded without the reconstructed values of the immediately above row of samples. So, while this is not necessarily a limitation on many aspects of the operation of an encoder or a decoder, it is recognised in the present embodiments that this dependency can in fact prevent successful use of the RDOQ system discussed above. Reasons for this will now be described.
RDOQ relies upon a scanning operation through a block of data using the same scan order as that applied by the entropy encoding function. Assuming a conventional reverse scan direction is used, this would imply that the RDOQ process would require access to data values which are not yet available, because each data value has a chain of dependency heading leftwards (in the case of horizontal RDPCM) or upwards (in the case of vertical RDPCM) in the array.
One solution to this problem might appear to be just to reverse the order in which the RDPCM operation is carried out, so that a vertical RDPCM operation starts with the bottom row and works upward, and a horizontal RDPCM operation starts with the right column and works leftward. However, such an arrangement is not considered desirable, particularly in the case of intra-image predicted data, in which the left column and top row of samples are considered the most reliable, in that they represent spatial positions which are closest to the image positions of those reference samples from which the samples are predicted. Accordingly, instead, a change in the entropy encoding order, and in turn to the order of processing of RDOQ, is proposed. This change will be described below.
Figures 24a and 24b schematically illustrate changes to a scanning order for vertically encoded RDPCM data in the case of intra-image encoding. In particular, Figure 24a schematically illustrates a reverse horizontal scan order, in which the numbers 1... 16, in that order, illustrate successive scan positions starting with the bottom right scan position. The scan order of Figure 24a would normally be applied to a block of data having a vertical prediction direction. However, if RDPCM is used, a forward horizontal scan order as shown in Figure 24b is used, starting with the top left scan position.
Figures 25a and 25b schematically illustrate changes to a scanning order for horizontally encoded RDPCM data. In particular, Figure 25a schematically illustrates a reverse vertical scan order, in which the numbers 1... 16, in that order, illustrate successive scan positions, again starting with the bottom right scan position. The scan order of Figure 25a would normally be applied to a block of data having a horizontal prediction direction. However, if RDPCM is used, a forward vertical scan order as shown in Figure 25b is used, starting with the top left scan position.
Figures 26a and 26b schematically illustrate changes to a scanning order for inter-image encoded RDPCM data in the case of intra-image encoding. In particular, Figure 26a schematically illustrates a reverse diagonal scan order, in which the numbers 1... 16, in that order, illustrate successive scan positions starting with the bottom right scan position. The scan order of Figure 26a would normally be applied to a block of data using inter-image encoding. However, if RDPCM is used, a forward diagonal scan order as shown in Figure 26b is used, starting with the top left scan position.
A common feature of all of the scan orders is shown in Figures 24b, 25b and 26b is that, for any scan position within the forward scan order, the preceding scanned data values (represented by smaller numbers in the scan order) are always available, having already been encoded and decoded by the RDPCM system. In other words, for any RDPCM data value in the processing order, the RDPCM data value does not have a dependency on an RDPCM data value later in the processing order. Accordingly, by using a forward scan order rather than a reverse scan order when RDPCM is used, RDOQ can still be operated even with RDPCM data.
Figure 27 is a schematic flowchart illustrating a method of processing comprising the steps of test encoding a block of data at a step 1000 including using an RDPCM code and a forward scan order as discussed above, applying an RDOQ process at a step 1010, and, at a step 1020, encoding the block using the encoding regime selected by the RDOQ process.
Figure 28 schematically illustrates part of a decoder (or the decoding path of an encoder) corresponding to the part of the encoder shown schematically in Figure 20.
The controller 345 determines controls a schematic switch 1100 so as to route data to an RDPCM decoder 1110, to a bypass route 1120 or to the inverse transform unit 430.
Data resulting from any of these routes is forwarded for further processing by a multiplexer 1130.
The inverse quantiser 420 is shown in a dotted line, because it may be omitted form the processing chain (under the control of the controller 345) in a trans-quant bypass mode.
Note that in Figures 20 and 28, the control operations apply not only to the switches 910, 1100 but also to the entropy encoder and decoder operations. This is because when an RDPCM mode is selected, the scan order used by entropy encoding and decoding is a forward scan order opposite to the usual reverse scan order.
The techniques described above relate to a relationship between the entropy encoding (and decoding) scan order and the use of RDPCM encoding. The techniques are applicable to RDPCM modes including (without limitation) so-called transform skip modes and trans-quant bypass (omitting frequency transformation and quantisation) modes.
In some examples, an operation referred to as "rotation" or re-ordering may be used in the encoding and decoding processes.
Rotation refers to an operation whereby data values in a square array of data values are rotated by 180° around the geometric centre of that array. So, for example, in the case of a 4 x 4 array of the type discussed above, having a numbering scheme from number 1 in the top-left corner, progressing left to right along each row in turn down to 16 in the bottom-right corner, the axis of the 180° rotation would be at the geometric centre of the array, which is to say the junction between the array elements initially numbered 6,7, 10 and 11. A 180° rotation will exchange each array data value with another array data value rotationally displaced from that array element by 180°. For example, the array data values numbered 1 and 16 are interchanged; those numbered 2 and 15 are interchanged; and so on. It will therefore be appreciated that the rotation operation is both lossless and entirely reversible.
Rotation has been proposed for use in some previously proposed encoders and decoders. (Note that if rotation is used at the encoder side, it also has to be used at the decoder side and in the decoder path of the encoder; if rotation is not used at the encoder side, it should not be used either at the decoder side or in the decoder path of the encoder). In particular, rotation may be selected for use in respect of transform-skip blocks. A reason why rotation is available (though not compulsory) for use with transform-skip blocks in some previously proposed systems is as follows. The entropy encoding process is arranged to scan "backwards", which is to say that entropy encoding operates from the lower right region of a block towards the upper left region of a block. This is based on the principle that entropy encoding operates potentially more efficiently if it starts with smaller data values and progresses towards larger data values within a block. However, in a transform-skip mode, it is the residual data which is being encoded by the entropy encoding process. Generally speaking, when using intra-prediction, the residual data values are smaller in magnitude (representing a better or closer prediction of the corresponding pixel) at block positions which are closest in image position to the reference samples from which the prediction is made. In an intra-image encoded block, the reference samples are located down the left side and along the top of the block to be predicted. So, in transform-skip processing of intra-image encoded blocks, the residual values are likely to be of smaller magnitude towards the upper left of the block and of larger magnitude towards the lower right of the block. For this reason, it may be advantageous to use the rotation arrangement in respect of such transform-skip intra-image encoded blocks so that the entropy encoding process operates on the smaller magnitude values first. Rotation can be selected, for example by a controller such as the controller 345, for example in response to testing the number of bits and the bit error rate generated with and without the rotation step included in the process.
Where a frequency transformer is used, embodiments are operable in a first mode (such as transform-skip) in which RDPCM coding is used, and in a second mode in which the frequency transformer is operable to apply a frequency transformation instead of the RDFCM coder applying an RDPCM operation.
If rotation is selected, its use can be flagged in the output data stream such that a decoder may select rotation response to detection of such a flag.
In previously proposed systems, rotation is not obligatory and the selection of rotation within the encoding and decoding path is entirely separate from any decision at the encoder to use RDPCM. In these previously proposed systems, rotation is used only for intra-image encoded 4 x 4 blocks. For bigger blocks than this, rotation is not used.
Variations upon the use of rotation and other features, relating to embodiments of the present technology, will now be discussed. These are initially summarised in Figures 29-31, which are schematic flowcharts indicating aspects of the operation of an encoder and/or decoder, and then will be discussed in more detail below.
Figure 29 relates to circumstances in which rotation is applied. If, at a schematic step 1200, RDPCM or CDPCM (coefficient differential pulse code modulation-see below) is in use in respect of a transform-skip block, then at a step 1210 rotation is always used. Note that this contrasts with previously proposed arrangements in which (a) the use of rotation is entirely optional, and (b) the selection of rotation is independent of the selection of a DF'CM mode.
Figure 30 relates to an aspect of an inverse RDPCM or CDPCM operation, which is to say an operation carried out at a decoder or in a decoding path of an encoder.
Where such an operation is used (indicated by a schematic step 1220), a clipping operation is also performed at a schematic step 1230 so as to clip the output data from the inverse RDPCM or CDPCM operation to a number of bits representing the appropriate or acceptable range of valid data generated by the operation 1220. The number of bits applicable to embodiments of the present disclosure will be discussed below.
Figure 31 relates to a scan order (at the entropy encoding and decoding stage) applicable to inter-image encoded RDPCM or CDPCM TUs. In previously proposed arrangements, inter-image encoded TUs are subject to entropy encoding according to a reverse diagonal scan order. However, in embodiments of the present technology, if (at a schematic step 1240) an inter-image encoded RDPCM or CDPCM TU is being processed, then at a schematic step 1250 a horizontal or vertical scan mode is used at the entropy encoding and decoding stage. Further details of the appropriate scan mode will be discussed below.
If rotation is introduced into an encoder or decoder, the outcomes according to previously proposed arrangements are shown in Figures 32 and 33. These will now be described in order to illustrate potential problems with these arrangements. Where parts or features of these diagrams correspond to identity numbered parts of previously described diagrams, those parts or features will not be described in detail again.
Figure 32 schematically indicates a part of a previously proposed encoder and/or decoder operating in respect of intra-image encoded blocks. As mentioned, in many respects this arrangement is similar to that shown in Figure 5. An RDPCM encoder 1300 supplies encoded data to a rotation (reordering) unit 1310. Here, the rotation operation (1310) is carried out in respect of 4 x 4 transform-skip and trans-quant bypass TUs. The transform 340 and quantisation 350 units are shown, but one or both of these would be expected to be bypassed in either a transform-skip mode (in which the transform unit 340 is not used) or a trans-quant bypass mode in which neither unit is used. A single unit (labelled as 360, 370, 400, 410) indicates the operation of scanning and entropy encoding and decoding. Optional inverse quantisation inverse transform units are shown, and data is then passed to an inverse rotation unit 1320 which restores the original order as an inverse operation to that provided by the rotation unit 1310.
Finally, an inverse RDPCM unit (decoder) 1330 carries out the inverse operation to the RDPCM encoder 1300. Note that as discussed above, the RDPCM encoder 1300 requires the decoded residual values from the decoder 1330 in order to encode the difference values.
Figure 33 schematically indicates a part of a previously proposed encoder and/or decoder operating in respect of inter-image encoded blocks. Here, once again the rotation operation (1310) is carried out in respect of 4 x 4 transform-skip and trans-quant bypass TUs. Otherwise, the units operate in a similar manner but the ordering is slightly different in comparison to Figure 32.
There are a number of perceived problems with these previously proposed arrangements.
A first problem relates to the use of different configurations for inter-and intra-image encoded blocks. Perceived advantages of this previously proposed arrangement are that the inter-image encoding path benefits from the increased accuracy of the different ordering. It also benefits from a shortened encoder feedback path. However, the differences between the inter and intra paths are unwelcome: the use of rotation prior to the encoder's inter-RDPCM is peculiar. Indeed, rather than an RDPCM process (operating on the residual samples), in effect, the inter-encoding path should be considered to be a "coefficient DPCM" (CDPCM) process (even though a transform is not used) because it is applied after the transform-skip in the encoder.
A second problem relates to an inability to apply RDOQ. The rate-distortion-optimised quantisation (RDOQ) algorithm, as discussed above, cannot be applied to the inter-image encoded case in this previously proposed arrangement -the differential coding is applied in forward scan order, whereas the entropy coder (which RDOQ models), will process the coefficients in reverse scan order. RDOQ can actually be used for the intra case, but only if rotation is used (and then therefore only for 4x4 transform-skip TU5).
A third problem relates to numerical limits of the inverse DPCM process, in either inter-or intra-image encoding. The bit-stream is restricted in that it may not contain values that are more than MTDR+1' signed bits (typically 16-bits) in size. MTDR is short for a system variable representing the maximum dynamic range of the transform process, but in this context it represents the magnitude of an appropriate or allowable numerical range for the data. The only other clipping is at the output of the inverse quantiser (again, which clips to MTDR-f1' signed bits for non-transquant-bypass TU5) and when the reconstruction value is formed. In the intra-encoding configuration of Figure 32, this means that the inputs of the inverse RDPCM could be MTDR+1' signed bits in size, even though in reality the values should be of the order of residual coefficients, or n+1' signed bits in size. In addition, the inverse RDPCM module utilises an accumulator, and therefore theoretically, the number of bits in the output (and accumulator) could be as much as the bit depth of the input RDPCM data + log2(size) bits (where "size" represents a linear dimension of the block size, in samples (so a 4 x 4 block would have a size of 4). For 32x32 TUs, this can be 5 bits more than the input, which could be as much as 21 bits (signed).
A fourth problem relates to restrictions on parallelisation. When the so-called sample adaptive intra-prediction technique was originally proposed as a tool in the HEVC system, parallelisation was straightforward as the process was only applied to lossless blocks. An encoder could easily generate differences between the source residual data; a decoder could easily generate accumulations, either column-by-column or row-by-row, contrary to the direction of RDPCM. i.e. for a horizontal RDPCM, the data could be processed column by column. Note, however, that the block could not start to be decoded until the top-left coefficient was decoded, which since the order is reverse-scan, is decoded last. Hence a buffer to temporarily hold the TU is required.
But for the RDPCM system shown in Figure 32, the amount of parallelisation which can be implemented is limited. Again, an encoder can, for a horizontal RDPCM arrangement, process the data column by column. However, as before, a decoder must wait until the top-left coefficient was decoded; for non 4x4 blocks or when rotation is disabled this will be the last in the TU. In addition, for inter-processing, the scan order is always diagonal, and therefore coefficients will not be decoded in the appropriate order.
For 4x4 intra-encoded blocks with rotation, the coefficients are decoded by the entropy decoder in an appropriate order -for horizontal RDFCM, the coefficients are decoded column by column because the intra prediction direction that indicates the use of horizontal RDPCM also indicates the use of the vertical scan order. The combination with rotation means that the first set of values decoded is actually the first column, and therefore RDPCM can be processed column-by-column without delay.
These problems may be alleviated by the use of the so-called unified arrangements of Figures 34 and 35, incorporating one or more of the techniques discussed above with reference to Figures 29-31. Again, where blocks or features correspond to those discussed earlier, they will not be discussed in detail again.
Figure 34 schematically indicates a part of an encoder and/or decoder operating in respect of intra-image encoded blocks.
This arrangement aims to unify the inter and intra cases, to always give an encoder the option to apply RDOQ for the transform-skip case and to improve the processing order of coefficients to increase parallelisation and reduce buffer usage. This can be achieved in one of two ways (as examples), as described in respect of Figures 34 and 35.
In these configurations, always applying rotation after the RDPCM module allows RDOQ to be used -the order the coefficients are encoded in is now effectively from the top left down to the bottom right (either horizontal, vertical or diagonal scan), which is an order in which RDPCM can produce coefficients.
In addition, to simplify the processing of inter-RDPCM coefficients, the encoding of inter RDPCM coefficients is no longer forced to use diagonal scan, but instead uses vertical scanning for horizontal inter RDPCM, and horizontal scanning for vertical inter RPCDM (similar to the intra behaviour).
Also, to prevent the accumulators in the IRDPCM module being excessively large, a clipping operation is used.
RDPCM is applied to the residual if required, then the TU is rotated if RDFCM was applied, or if an SPS flag indicates that such transform-skipped / transquant-bypass TUs should be rotated (SPS is used in previously proposed systems as a flag indicating whether rotation should be applied; in the previously proposed systems rotation is available only to 4 x 4 transform-skip or trans-quant bypass TUs). Following this, the TU is transformed (or the transform is bypassed in a transform-skip mode) and quantised.
Clipping would be applied during the accumulation of RDPCM values in the inverse RDPCM module, limiting the accumulator and downstream values to just ni-i signed bits. Note that rotation could be applied after the quantiser, or can even be considered as a change to the scan order in the entropy coder, effectively scanning in forward order rather than reverse.
Accordingly, Figure 34 provides an example of video data encoding apparatus (and a corresponding method) operable to encode an array of input video data values, the apparatus comprising a residual differential pulse code modulation (RDPCM) coder configured to apply a differential pulse code modulation operation to the array of input video data values to generate an array of RDPCM data values; and an entropy encoder operable to entropy-encode the RDPCM data in a processing order, with respect to the array of RDPCM data values, such that, for any RDPCM data value in the processing order, the RDPCM data value does not have a dependency on an RDPCM data value later in the processing order. These techniques are applicable to inter-encoded data, in which: the array of input video data values comprises an array of residual values generated as differences between respective video samples of a current image and predicted versions of those video samples, the predicted versions being derived from one or more images other than the current image; and for any such array of input video data values, the processing order is always such that, for any RDPCM data value in the processing order, the RDPCM data value does not have a dependency on an RDPCM data value later in the processing order. The techniques are also applicable to intra-encoded data in which the array of input video data values comprises an array of residual values generated as differences between respective video samples of a current image and predicted versions of those video samples, the predicted versions being derived from the current image; and for any such array of input video data values, the processing order is always such that, for any RDPCM data value in the processing order, the RDPCM data value does not have a dependency on an RDPCM data value later in the processing order. In intra-encoded data the predicted versions are derived from one or more reference pixels from reference regions adjacent to the image position of the respective video samples. In an alternative, sometimes referred to as intra block copy, the predicted versions are derived from one or more reference pixels from reference regions (in the current image) defined with respect to the image position of the respective video samples by one or more motion vectors.
Figure 35 schematically indicates a part of an encoder and/or decoder operating in respect of inter-image encoded blocks, in the case that a transform is applied first (so the operation is referred to as CDPCM rather than RDPCM, but the principles remain the same).
The TU is first transformed (or transform-skip is applied), then CDPCM is applied by a CDFCM encoder 1300' (if required), then the TU is rotated (again only if CDPCM was applied, or if a SPS flag indicates that such transform-skipped / transquant-bypass TUs should be rotated), and quantised. Again, clipping would be applied during the accumulation of CDPCM values in the inverse CDPCM module 1330', limiting the accumulator and downstream values to just MTDR-i-1 signed bits. Note that rotation could be applied after the quantiser, or can even be considered as a change to the scan order in the entropy coder, effectively scanning in forward order rather than reverse.
Data Signals It will be appreciated that data signals generated by the variants of coding apparatus discussed above, and storage or transmission media carrying such signals, are considered to represent embodiments of the present disclosure.
In so far as embodiments of the disclosure have been described as being implemented, at least in part, by software-controlled data processing apparatus, it will be appreciated that a non-transitory machine-readable medium carrying such software, such as an optical disk, a magnetic disk, semiconductor memory or the like, is also considered to represent an embodiment of the present disclosure.
It will be apparent that numerous modifications and variations of the present disclosure are possible in light of the above teachings. It is therefore to be understood that within the scope of the appended claims, the technology may be practiced otherwise than as specifically described herein.
All of the techniques described above have application in encoding apparatus and methods, and/or in decoding apparatus and methods. Techniques discussed In relation to encoding have corresponding features on the decoding side. Techniques discussed in relation to decoding have application in the decoding path of the encoder.
Accordingly, it will be appreciated that the description above of encoding techniques and apparatus, which techniques and apparatus include a decoding path, are also relevant as examples of decoding techniques and apparatus.
Respective features and/or embodiments of the present disclosure are defined by the following numbered clauses: 1. Video data encoding apparatus operable to encode an array of input video data values, the apparatus comprising: a residual differential pulse code modulation (RDFCM) coder configured to apply a differential pulse code modulation operation to the array of input video data values to generate an array of RDF'CM data values; and an entropy encoder operable to entropy-encode the RDFCM data in a processing order, with respect to the array of RDPCM data values, such that, for any RDPCM data value in the processing order, the RDPCM data value does not have a dependency on an RDPCM data value later in the processing order.
2. Apparatus according to clause 1, in which: the array of input video data values comprises an array of residual values generated as differences between respective video samples of a current image and predicted versions of those video samples, the predicted versions being derived from one or more images other than the current image; and for any such array of input video data values, the processing order is always such that, for any RDPCM data value in the processing order, the RDPCM data value does not have a dependency on an RDPCM data value later in the processing order.
3. Apparatus according to clause 1 or clause 2, in which: the array of input video data values comprises an array of residual values generated as differences between respective video samples of a current image and predicted versions of those video samples, the predicted versions being derived from the current image; and for any such array of input video data values, the processing order is always such that, for any RDPCM data value in the processing order, the RDPCM data value does not have a dependency on an RDPCM data value later in the processing order.
4. Apparatus according to clause 3, in which the predicted versions are derived from one or more reference pixels from reference regions adjacent to the image position of the respective video samples.
5. Apparatus according to clause 3, in which the predicted versions are derived from one or more reference pixels from reference regions defined with respect to the image position of the respective video samples by one or more motion vectors.
6. Apparatus according to any one of the preceding clauses, comprising a frequency transformer, the apparatus being operable in a first mode in which RDPCM coding is used, and in a second mode in which the frequency transformer is operable to apply a frequency transformation instead of the RDPCM coder applying an RDPCM operation.
7. Apparatus according to clause 6, in which: the frequency transformer is operable in the second mode to generate an array of frequency transformed data; and the entropy encoder is operable in the second mode to apply a processing order starting from an array position corresponding to that of a highest frequency component and ending at an array position corresponding to that of a lowest frequency component.
8. Apparatus according to clause 7, comprising a data reordering unit configured to apply a reordering operation to an array of data which has been subject to RDPCM coding.
9. Apparatus according to clause 8, in which: the first mode is a transform-skip mode in which a frequency transformation is not used; and the data reordering unit is configured so as always to operate on the RDPCM coded data in the transform-skip mode.
10. Apparatus according to any one of the preceding clauses in which, for RDFCM data, the processing order is a horizontally scanned or a vertically scanned processing order.
11. Apparatus according to any one of the preceding clauses, the apparatus comprising, in a decoding path, a residual differential pulse code modulation (RDPCM) decoder configured to apply a differential pulse code modulation operation to the entropy decoded video data values to generate an array of RDPCM data values, the RDPCM decoder being operable to clip the RDPCM decoded data to an acceptable numerical range for such data.
12. Video data encoding apparatus operable to encode an array of inter-image encoded input video data values, the apparatus comprising: a residual differential pulse code modulation (RDPCM) coder configured to apply a differential pulse code modulation operation to the array of input video data values to generate an array of RDPCM data values; and an entropy encoder operable to entropy-encode the RDPCM data in a processing order, with respect to the array of RDPCM data values, the processing order being a horizontally scanned or a vertically scanned processing order.
13. Video data encoding apparatus operable to encode an array of input encoded video data values, the apparatus comprising, in a decoding path, a residual differential pulse code modulation (RDPCM) decoder configured to apply a differential pulse code modulation operation to the entropy decoded video data values to generate an array of RDPCM data values, the RDPCM decoder being operable to clip the RDPCM decoded data to an acceptable numerical range for such data.
14. Apparatus according to any one of the preceding clauses, comprising a rate-distortion optiniised quantisation processor operable, in respect of an array of data, according to the same processing order as that applied by the entropy encoder.
15. Video data decoding apparatus operable to decode an array of input encoded video data values, the apparatus comprising: an entropy decoder; and a residual differential pulse code modulation (RDPCM) decoder configured to apply a differential pulse code modulation operation to the entropy decoded video data values to generate an array of RDPCM data values; in which the entropy decoder is operable to entropy-decode the input encoded video data in a processing order, with respect to the array of RDPCM data values, such that, for any RDPCM data value in the processing order, the RDPCM data value does not have a dependency on an RDPCM data value later in the processing order.
16. Apparatus according to clause 15, in which: the array of input encoded video data values comprises an array of inter-image encoded video data values; and for any such array of input video data values, the processing order is always such that, for any RDPCM data value in the processing order, the RDPCM data value does not have a dependency on an RDPCM data value later in the processing order.
17. Apparatus according to clause 15 or clause 16, in which: the array of input video data values comprises an array of intra-image encoded data values; and for any such array of input video data values, the processing order is always such that, for any RDPCM data value in the processing order, the RDPCM data value does not have a dependency on an RDPCM data value later in the processing order.
18. Apparatus according to clause 16, in which the processing order is a horizontally scanned or a vertically scanned processing order.
19. Apparatus according to any one of clauses 15 to 18, in which the RDPCM decoder is operable to clip the RDPCM decoded data to an acceptable numerical range for such data.
20. Video data decoding apparatus operable to decode an array of input encoded video data values, the apparatus comprising: an entropy decoder; and a residual differential pulse code modulation (RDPCM) decoder configured to apply a differential pulse code modulation operation to the entropy decoded video data values to generate an array of RDPCM data values, the RDPCM decoder being operable to clip the RDFCM decoded data to an acceptable numerical range for such data.
21. Video data decoding apparatus operable to decode an array of input encoded video data values, the apparatus comprising: an entropy decoder operable to entropy-decode the input encoded video data in a processing order, with respect to the array of input encoded video data values, the processing order being a horizontally scanned or a vertically scanned processing order; and a residual differential pulse code modulation (RDFCM) decoder configured to apply a differential pulse code modulation operation to the entropy decoded video data values to generate an array of RDPCM data values.
22. Video data capture, storage, transmission and/or recording apparatus comprising apparatus according to any one of the preceding clauses.
23. A video data encoding method for encoding an array of input video data values, the method comprising: applying a residual differential pulse code modulation (RDPCM) operation to the array of input video data values to generate an array of RDPCM data values; and entropy-encoding the RDPCM data in a processing order, with respect to the array of RDPCM data values, such that, for any RDPCM data value in the processing order, the RDPCM data value does not have a dependency on an RDPCM data value later in the processing order.
24. A video data encoding method for encoding an array of inter-image encoded input video data values, the method comprising: applying a residual differential pulse code modulation (RDPCM) operation to the array of input video data values to generate an array of RDPCM data values; and entropy-encoding the RDPCM data in a processing order, with respect to the array of RDPCM data values, the processing order being a horizontally scanned or a vertically scanned processing order.
25. A video data encoding method for encoding an array of input encoded video data values, the method comprising: in a decoding path, applying a residual differential pulse code modulation (RDPCM) operation to the entropy decoded video data values to generate an array of RDPCM data values; and clipping the RDPCM decoded data to an acceptable numerical range for such data.
26. A video data decoding method for decoding an array of input encoded video data values, the method comprising: entropy decoding the input encoded video data values in a processing order, with respect to the array of RDPCM data values, such that, for any RDPCM data value in the processing order, the RDFCM data value does not have a dependency on an RDPCM data value later in the processing order; and applying a residual differential pulse code modulation (RDPCM) operation to the entropy decoded video data values to generate an array of RDPCM data values.
27. A video data decoding method for decoding an array of input encoded video data values, the method comprising: entropy decoding the input encoded video data values; applying a residual differential pulse code modulation (RDPCM) operation to the entropy decoded video data values to generate an array of RDPCM data values; and clipping the RDPCM decoded data to an acceptable numerical range for such data.
28. A video data decoding method for decoding an array of input encoded video data values, the method comprising: entropy-decoding the input encoded video data values data in a processing order, with respect to the array of data values, the processing order being a horizontally scanned or a vertically scanned processing order; and applying a residual differential pulse code modulation (RDPCM) operation to the entropy decoded video data values to generate an array of RDPCM data values.
29. Computer software which, when executed by a computer, causes the computer to implement the method of any one of clauses 23 to 28.
30. A non-transitory machine-readable storage medium on which computer software according to clause 29 is stored.

Claims (30)

  1. CLAIMS1. Video data encoding apparatus operable to encode an array of input video data values, the apparatus comprising: a residual differential pulse code modulation (RDFCM) coder configured to apply a differential pulse code modulation operation to the array of input video data values to generate an array of RDF'CM data values; and an entropy encoder operable to entropy-encode the RDFCM data in a processing order, with respect to the array of RDPCM data values, such that, for any RDPCM data value in the processing order, the RDPCM data value does not have a dependency on an RDPCM data value later in the processing order.
  2. 2. Apparatus according to claim 1, in which: the array of input video data values comprises an array of residual values generated as differences between respective video samples of a current image and predicted versions of those video samples, the predicted versions being derived from one or more images other than the current image; and for any such array of input video data values, the processing order is always such that, for any RDPCM data value in the processing order, the RDPCM data value does not have a dependency on an RDPCM data value later in the processing order.
  3. 3. Apparatus according to claim 1, in which: the array of input video data values comprises an array of residual values generated as differences between respective video samples of a current image and predicted versions of those video samples, the predicted versions being derived from the current image; and for any such array of input video data values, the processing order is always such that, for any RDPCM data value in the processing order, the RDPCM data value does not have a dependency on an RDPCM data value later in the processing order.
  4. 4. Apparatus according to claim 3, in which the predicted versions are derived from one or more reference pixels from reference regions adjacent to the image position of the respective video samples.
  5. 5. Apparatus according to claim 3, in which the predicted versions are derived from one or more reference pixels from reference regions defined with respect to the image position of the respective video samples by one or more motion vectors.
  6. 6. Apparatus according to claim 1, comprising a frequency transformer, the apparatus being operable in a first mode in which RDFCM coding is used, and in a second mode in which the frequency transformer is operable to apply a frequency transformation instead of the RDPCM coder applying an RDFCM operation.
  7. 7. Apparatus according to claim 6, in which: the frequency transformer is operable in the second mode to generate an array of frequency transformed data; and the entropy encoder is operable in the second mode to apply a processing order starting from an array position corresponding to that of a highest frequency component and ending at an array position corresponding to that of a lowest frequency component.
  8. 8. Apparatus according to claim 7, comprising a data reordering unit configured to apply a reordering operation to an array of data which has been subject to RDPCM coding.
  9. 9. Apparatus according to claim 8, in which: the first mode is a transform-skip mode in which a frequency transformation is not used; and the data reordering unit is configured so as always to operate on the RDPCM coded data in the transform-skip mode.
  10. 10. Apparatus according to claim 1, in which, for RDPCM data, the processing order is a horizontally scanned or a vertically scanned processing order.
  11. 11. Apparatus according to claim 1, the apparatus comprising, in a decoding path, a residual differential pulse code modulation (RDPCM) decoder configured to apply a differential pulse code modulation operation to the entropy decoded video data values to generate an array of RDPCM data values, the RDPCM decoder being operable to clip the RDPCM decoded data to an acceptable numerical range for such data.
  12. 12. Video data encoding apparatus operable to encode an array of inter-image encoded input video data values, the apparatus comprising: a residual differential pulse code modulation (RDPCM) coder configured to apply a differential pulse code modulation operation to the array of input video data values to generate an array of RDFCM data values; and an entropy encoder operable to entropy-encode the RDPCM data in a processing order, with respect to the array of RDFCM data values, the processing order being a horizontally scanned or a vertically scanned processing order.
  13. 13. Video data encoding apparatus operable to encode an array of input encoded video data values, the apparatus comprising, in a decoding path, a residual differential pulse code modulation (RDPCM) decoder configured to apply a differential pulse code modulation operation to the entropy decoded video data values to generate an array of RDPCM data values, the RDPCM decoder being operable to clip the RDPCM decoded data to an acceptable numerical range for such data.
  14. 14. Apparatus according to claim 1, comprising a rate-distortion optimised quantisation processor operable, in respect of an array of data, according to the same processing order as that applied by the entropy encoder.
  15. 15. Video data decoding apparatus operable to decode an array of input encoded video data values, the apparatus comprising: an entropy decoder; and a residual differential pulse code modulation (RDPCM) decoder configured to apply a differential pulse code modulation operation to the entropy decoded video data values to generate an array of RDPCM data values; in which the entropy decoder is operable to entropy-decode the input encoded video data in a processing order, with respect to the array of RDPCM data values, such that, for any RDPCM data value in the processing order, the RDPCM data value does not have a dependency on an RDPCM data value later in the processing order.
  16. 16. Apparatus according to claim 15, in which: the array of input encoded video data values comprises an array of inter-image encoded video data values; and for any such array of input video data values, the processing order is always such that, for any RDPCM data value in the processing order, the RDPCM data value does not have a dependency on an RDPCM data value later in the processing order.
  17. 17. Apparatus according to claim 15, in which: the array of input video data values comprises an array of intra-image encoded data values; and for any such array of input video data values, the processing order is always such that, for any RDPCM data value in the processing order, the RDPCM data value does not have a dependency on an RDPCM data value later in the processing order.
  18. 18. Apparatus according to claim 16, in which the processing order is a horizontally scanned or a vertically scanned processing order.
  19. 19. Apparatus according to claim 15, in which the RDPCM decoder is operable to clip the RDPCM decoded data to an acceptable numerical range for such data.
  20. 20. Video data decoding apparatus operable to decode an array of input encoded video data values, the apparatus comprising: an entropy decoder; and a residual differential pulse code modulation (RDPCM) decoder configured to apply a differential pulse code modulation operation to the entropy decoded video data values to generate an array of RDPCM data values, the RDPCM decoder being operable to clip the RDPCM decoded data to an acceptable numerical range for such data.
  21. 21. Video data decoding apparatus operable to decode an array of input encoded video data values, the apparatus comprising: an entropy decoder operable to entropy-decode the input encoded video data in a processing order, with respect to the array of input encoded video data values, the processing order being a horizontally scanned or a vertically scanned processing order; and a residual differential pulse code modulation (RDPCM) decoder configured to apply a differential pulse code modulation operation to the entropy decoded video data values to generate an array of RDPCM data values.
  22. 22. Video data capture, storage, transmission and/or recording apparatus comprising apparatus according to claim 1.
  23. 23. A video data encoding method for encoding an array of input video data values, the method comprising: applying a residual differential pulse code modulation (RDPCM) operation to the array of input video data values to generate an array of RDPCM data values; and entropy-encoding the RDPCM data in a processing order, with respect to the array of RDPCM data values, such that, for any RDPCM data value in the processing order, the RDPCM data value does not have a dependency on an RDPCM data value later in the processing order.
  24. 24. A video data encoding method for encoding an array of inter-image encoded input video data values, the method comprising: applying a residual differential pulse code modulation (RDPCM) operation to the array of input video data values to generate an array of RDFCM data values; and entropy-encoding the RDPCM data in a processing order, with respect to the array of RDPCM data values, the processing order being a horizontally scanned or a vertically scanned processing order.
  25. 25. A video data encoding method for encoding an array of input encoded video data values, the method comprising: in a decoding path, applying a residual differential pulse code modulation (RDPCM) operation to the entropy decoded video data values to generate an array of RDPCM data values; and clipping the RDPCM decoded data to an acceptable numerical range for such data.
  26. 26. A video data decoding method for decoding an array of input encoded video data values, the method comprising: entropy decoding the input encoded video data values in a processing order, with respect to the array of RDPCM data values, such that, for any RDPCM data value in the processing order, the RDPCM data value does not have a dependency on an RDPCM data value later in the processing order; and applying a residual differential pulse code modulation (RDPCM) operation to the entropy decoded video data values to generate an array of RDPCM data values.
  27. 27. A video data decoding method for decoding an array of input encoded video data values, the method comprising: entropy decoding the input encoded video data values; applying a residual differential pulse code modulation (RDPCM) operation to the entropy decoded video data values to generate an array of RDPCM data values; and clipping the RDPCM decoded data to an acceptable numerical range for such data.
  28. 28. A video data decoding method for decoding an array of input encoded video data values, the method comprising: entropy-decoding the input encoded video data values data in a processing order, with respect to the array of data values, the processing order being a horizontally scanned or a vertically scanned processing order; and applying a residual differential pulse code modulation (RDPCM) operation to the entropy decoded video data values to generate an array of RDPCM data values.
  29. 29. Computer software which, when executed by a computer, causes the computer to implement the method of claim 23.
  30. 30. A non-transitory machine-readable storage medium on which computer software according to claim 29 is stored.
GB1317406.5A 2013-10-01 2013-10-01 Data encoding and decoding Withdrawn GB2519070A (en)

Priority Applications (10)

Application Number Priority Date Filing Date Title
GB1317406.5A GB2519070A (en) 2013-10-01 2013-10-01 Data encoding and decoding
GB1317803.3A GB2519094A (en) 2013-10-01 2013-10-08 Data encoding and decoding
GBGB1318593.9A GB201318593D0 (en) 2013-10-01 2013-10-21 Data encoding and decoding
GB1320772.5A GB2519177A (en) 2013-10-01 2013-11-25 Data encoding and decoding
EP14777128.1A EP3039867B1 (en) 2013-10-01 2014-09-26 Video data encoding and decoding
PCT/GB2014/052917 WO2015049491A1 (en) 2013-10-01 2014-09-26 Video data encoding and decoding
JP2016519369A JP6609753B2 (en) 2013-10-01 2014-09-26 Data encoding and decoding
CN201480054385.8A CN105612744B (en) 2013-10-01 2014-09-26 video data encoding and decoding
US15/024,705 US10939103B2 (en) 2013-10-01 2014-09-26 Data encoding and decoding
RU2016117252A RU2679991C2 (en) 2013-10-01 2014-09-26 Video data encoding and decoding

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
GB1317406.5A GB2519070A (en) 2013-10-01 2013-10-01 Data encoding and decoding

Publications (2)

Publication Number Publication Date
GB201317406D0 GB201317406D0 (en) 2013-11-13
GB2519070A true GB2519070A (en) 2015-04-15

Family

ID=49585155

Family Applications (4)

Application Number Title Priority Date Filing Date
GB1317406.5A Withdrawn GB2519070A (en) 2013-10-01 2013-10-01 Data encoding and decoding
GB1317803.3A Withdrawn GB2519094A (en) 2013-10-01 2013-10-08 Data encoding and decoding
GBGB1318593.9A Ceased GB201318593D0 (en) 2013-10-01 2013-10-21 Data encoding and decoding
GB1320772.5A Withdrawn GB2519177A (en) 2013-10-01 2013-11-25 Data encoding and decoding

Family Applications After (3)

Application Number Title Priority Date Filing Date
GB1317803.3A Withdrawn GB2519094A (en) 2013-10-01 2013-10-08 Data encoding and decoding
GBGB1318593.9A Ceased GB201318593D0 (en) 2013-10-01 2013-10-21 Data encoding and decoding
GB1320772.5A Withdrawn GB2519177A (en) 2013-10-01 2013-11-25 Data encoding and decoding

Country Status (7)

Country Link
US (1) US10939103B2 (en)
EP (1) EP3039867B1 (en)
JP (1) JP6609753B2 (en)
CN (1) CN105612744B (en)
GB (4) GB2519070A (en)
RU (1) RU2679991C2 (en)
WO (1) WO2015049491A1 (en)

Families Citing this family (23)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2519070A (en) * 2013-10-01 2015-04-15 Sony Corp Data encoding and decoding
RU2635064C2 (en) 2013-10-14 2017-11-08 ЭйджЭфАй Инновэйшн Инк. Method of residual differential pulse code modulation for expanding high-efficient video coding (hevc) range
US10856009B2 (en) * 2014-09-04 2020-12-01 Mediatek Inc. Method of block vector clipping and coding for screen content coding and video coding
US10853502B1 (en) * 2015-03-04 2020-12-01 Micro Focus Llc Systems and methods for reducing computational difficulty of cryptographic operations
WO2018120019A1 (en) * 2016-12-30 2018-07-05 上海寒武纪信息科技有限公司 Compression/decompression apparatus and system for use with neural network data
WO2018121670A1 (en) * 2016-12-30 2018-07-05 上海寒武纪信息科技有限公司 Devices for compression/decompression, system, chip, and electronic device
US10778971B2 (en) * 2018-06-01 2020-09-15 Tencent America LLC Method and apparatus for video coding
FR3081656A1 (en) * 2018-06-27 2019-11-29 Orange METHODS AND DEVICES FOR ENCODING AND DECODING A DATA STREAM REPRESENTATIVE OF AT LEAST ONE IMAGE.
US11265544B2 (en) * 2018-09-18 2022-03-01 Sony Corporation Apparatus and method for image compression based on optimal sequential encoding scheme
GB2578633B (en) * 2018-11-02 2021-10-13 Advanced Risc Mach Ltd Encoding data arrays
KR20210094530A (en) 2018-11-29 2021-07-29 베이징 바이트댄스 네트워크 테크놀로지 컴퍼니, 리미티드 Interaction between in-screen block copy mode and cross-screen prediction tools
CA3130472A1 (en) 2019-02-24 2020-08-27 Beijing Bytedance Network Technology Co., Ltd. Independent coding of palette mode usage indication
WO2020182069A1 (en) * 2019-03-10 2020-09-17 Mediatek Inc. Mehtod and apparatus of the quantization matrix computation and representation for video coding
JP7277608B2 (en) * 2019-04-23 2023-05-19 北京字節跳動網絡技術有限公司 Context modeling and selection of multiple transformation matrices
CN110149514B (en) * 2019-05-31 2023-05-05 上海联影医疗科技股份有限公司 Compression method and device of scanning data, acquisition system and medical imaging system
CN114175662B (en) 2019-07-20 2023-11-24 北京字节跳动网络技术有限公司 Condition dependent codec with palette mode usage indication
CN114145013B (en) 2019-07-23 2023-11-14 北京字节跳动网络技术有限公司 Mode determination for palette mode coding and decoding
KR20220036911A (en) * 2019-07-29 2022-03-23 베이징 바이트댄스 네트워크 테크놀로지 컴퍼니, 리미티드 Palette mode coding in prediction process
US11375196B2 (en) * 2019-09-19 2022-06-28 Tencent America LLC Method of context modeling for entropy coding of transform coefficient flags with transform skip, and coding method with the same
US11418814B2 (en) 2020-05-22 2022-08-16 Tencent America LLC Constraint on syntax elements for still picture profiles
US11722678B2 (en) 2020-08-25 2023-08-08 Tencent America LLC Method and apparatus for transform skip coefficients coding
CN114286095B (en) * 2020-09-27 2023-02-07 腾讯科技(深圳)有限公司 Video decoding method, device and equipment
CN113706493B (en) * 2021-08-20 2024-03-22 北京航空航天大学 Digital pathology full-section segmentation method based on DCT frequency self-adaptive selection

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103024383A (en) * 2012-12-14 2013-04-03 北京工业大学 Intra-frame lossless compression coding method based on HEVC (high efficiency video coding) frame
WO2013067435A1 (en) * 2011-11-04 2013-05-10 Huawei Technologies Co., Ltd. Differential pulse code modulation intra prediction for high efficiency video coding

Family Cites Families (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7483581B2 (en) 2001-07-02 2009-01-27 Qualcomm Incorporated Apparatus and method for encoding digital image data in a lossless manner
KR100566290B1 (en) * 2003-09-18 2006-03-30 삼성전자주식회사 Image Scanning Method By Using Scan Table and Discrete Cosine Transform Apparatus adapted it
WO2005036528A1 (en) * 2003-10-10 2005-04-21 Agency For Science, Technology And Research Method for encoding a digital signal into a scalable bitstream; method for decoding a scalable bitstream.
KR100703283B1 (en) * 2004-03-15 2007-04-03 삼성전자주식회사 Image encoding apparatus and method for estimating motion using rotation matching
US20070110151A1 (en) 2005-11-14 2007-05-17 Ess Technology, Inc. System and method for video frame buffer compression
CN101115200B (en) * 2007-04-20 2010-05-19 西安交通大学 Effective motion vector telescopic encoding method
US8295342B2 (en) * 2007-11-14 2012-10-23 International Business Machines Corporation Method and system for efficient video compression with low-complexity encoder
US8295356B2 (en) * 2008-03-07 2012-10-23 International Business Machines Corporation Method and system for coding mode selection in video compression systems
TW201028018A (en) * 2009-01-07 2010-07-16 Ind Tech Res Inst Encoder, decoder, encoding method and decoding method
US9001889B2 (en) * 2009-05-16 2015-04-07 Thomson Licensing Methods and apparatus for improved quantization rounding offset adjustment for video encoding and decoding
PL2476255T3 (en) * 2009-09-10 2019-03-29 Guangdong Oppo Mobile Telecommunications Corp., Ltd. Speedup techniques for rate distortion optimized quantization
US20110292247A1 (en) 2010-05-27 2011-12-01 Sony Corporation Image compression method with random access capability
JP2012175615A (en) * 2011-02-24 2012-09-10 Sony Corp Image processing apparatus and image processing method
US20120307890A1 (en) * 2011-06-02 2012-12-06 Microsoft Corporation Techniques for adaptive rounding offset in video encoding
GB2496193A (en) 2011-11-07 2013-05-08 Sony Corp Context adaptive data encoding and decoding
GB2519070A (en) * 2013-10-01 2015-04-15 Sony Corp Data encoding and decoding

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2013067435A1 (en) * 2011-11-04 2013-05-10 Huawei Technologies Co., Ltd. Differential pulse code modulation intra prediction for high efficiency video coding
CN103024383A (en) * 2012-12-14 2013-04-03 北京工业大学 Intra-frame lossless compression coding method based on HEVC (high efficiency video coding) frame

Also Published As

Publication number Publication date
WO2015049491A1 (en) 2015-04-09
EP3039867A1 (en) 2016-07-06
US20160241852A1 (en) 2016-08-18
RU2679991C2 (en) 2019-02-14
GB201318593D0 (en) 2013-12-04
GB201317803D0 (en) 2013-11-20
GB2519177A (en) 2015-04-15
GB201320772D0 (en) 2014-01-08
GB2519094A (en) 2015-04-15
US10939103B2 (en) 2021-03-02
CN105612744B (en) 2018-12-21
JP2016536848A (en) 2016-11-24
JP6609753B2 (en) 2019-11-27
EP3039867B1 (en) 2018-06-27
GB201317406D0 (en) 2013-11-13
RU2016117252A (en) 2017-11-13
RU2016117252A3 (en) 2018-07-19
CN105612744A (en) 2016-05-25

Similar Documents

Publication Publication Date Title
US10939103B2 (en) Data encoding and decoding
US10958938B2 (en) Data encoding and decoding
US9973777B2 (en) Data encoding and decoding apparatus, method and storage medium
US11924430B2 (en) Data encoding and decoding
GB2585042A (en) Image data encoding and decoding
CN111684798A (en) Data encoding and decoding
WO2020260876A1 (en) Image data encoding and decoding
AU2020301626A1 (en) Image data encoding and decoding
WO2020065293A1 (en) Image data encoding and decoding

Legal Events

Date Code Title Description
WAP Application withdrawn, taken to be withdrawn or refused ** after publication under section 16(1)