WO2017129023A1 - 解码方法、编码方法、解码设备和编码设备 - Google Patents

解码方法、编码方法、解码设备和编码设备 Download PDF

Info

Publication number
WO2017129023A1
WO2017129023A1 PCT/CN2017/071602 CN2017071602W WO2017129023A1 WO 2017129023 A1 WO2017129023 A1 WO 2017129023A1 CN 2017071602 W CN2017071602 W CN 2017071602W WO 2017129023 A1 WO2017129023 A1 WO 2017129023A1
Authority
WO
WIPO (PCT)
Prior art keywords
pixel
image block
component
signal
target image
Prior art date
Application number
PCT/CN2017/071602
Other languages
English (en)
French (fr)
Inventor
曾兵
陈宸
朱树元
缪泽翔
张红
赵寅
Original Assignee
华为技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 华为技术有限公司 filed Critical 华为技术有限公司
Publication of WO2017129023A1 publication Critical patent/WO2017129023A1/zh

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/136Incoming video signal characteristics or properties
    • H04N19/137Motion inside a coding unit, e.g. average field, frame or block difference
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/13Adaptive entropy coding, e.g. adaptive variable length coding [AVLC] or context adaptive binary arithmetic coding [CABAC]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/176Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • H04N19/61Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding

Definitions

  • the embodiments of the present invention relate to the field of video coding and decoding and compression, and more particularly, to a decoding method, an encoding method, a decoding device, and an encoding device.
  • Digital video capabilities can be incorporated into a wide range of devices, including digital television, digital live broadcast systems, wireless broadcast systems, personal digital assistants (PDAs), laptop or desktop computers, tablet computers, e-book readers, digital Cameras, digital recording devices, digital media players, video game devices, video game consoles, cellular or satellite radio phones, video conferencing devices, video streaming devices, and the like.
  • Digital video devices implement video compression techniques such as MPEG-2, MPEG-4, ITU-TH.263, ITU-TH.264/MPEG-4 Part 10 Advanced Video Codec (AVC), ITU-TH.265 High
  • the video compression techniques defined in the standards defined by the Efficiency Video Coding and Decoding (HEVC) standard and those described in the extensions to the standards enable more efficient transmission and reception of digital video information.
  • Video devices may transmit, receive, encode, decode, and/or store digital video information more efficiently by implementing these video codec techniques.
  • the concept of a frame refers to a complete image, which can be played by a frame-by-frame image in a certain order and frame rate.
  • the frame rate reaches a certain speed, the interval between the two frames is less than the resolution limit of the human eye, and a short visual pause occurs, so that it appears to be dynamically appearing on the screen.
  • the basis for the compression of video files is the compression encoding of single-frame digital images.
  • the spatial structure is the same or similar in a frame of image. For example, there is a close correlation and similarity between the colors of sampling points in the same object or background.
  • one frame of image has substantially a large correlation with its previous or subsequent frame, and the difference in pixel values of the description information is small, and these are all parts that can be compressed.
  • there is not only spatial redundancy information in the video file but also a large amount of time redundant information, which is caused by the composition of the video.
  • the frame rate of video sampling is generally 25 frames/second to 30 frames/second, and in special cases, 60 frames/second may occur. That is to say, the sampling interval between two adjacent frames is at least 1/30 second to 1/25 second. In such a short period of time, there is basically a large amount of similar information in the sampled image, and there is a great correlation between the images.
  • visual redundancy refers to the use of the human eye to be sensitive to changes in luminance and relatively insensitive to changes in chrominance to properly compress the video bitstream.
  • visual redundancy In high-brightness areas, the sensitivity of human eye vision to brightness changes is declining, and it is more sensitive to the edge of the object, and the inner area is relatively insensitive; it is sensitive to the overall structure and relatively insensitive to internal detail changes.
  • video image information Because the final service object of video image information is our human population, we can make full use of these characteristics of the human eye to compress the original video image information to achieve better compression.
  • video image information also has redundancy in information entropy redundancy, structural redundancy, knowledge redundancy, importance redundancy, etc. information.
  • the purpose of video compression coding is to use various technical methods to remove redundant information in the video sequence to achieve the effect of reducing storage space and saving transmission bandwidth.
  • video compression processing technologies mainly include intra prediction, inter prediction, transform quantization, entropy coding, and deblocking filtering.
  • video compression coding standards There are four main types of mainstream compression coding: chroma sampling, predictive coding, transform coding, and quantization coding.
  • Chroma sampling This method makes full use of the visual psychological characteristics of the human eye, and tries to minimize the amount of data described by a single element from the underlying data representation.
  • Most of the television systems use luminance-chrominance-chrominance (YUV) color coding, which is widely adopted by European television systems.
  • the YUV color space includes a luminance signal Y and two color difference signals U and V, and the three components are independent of each other.
  • the YUV color mode is more flexible in representation, and the transmission takes up less bandwidth, which is superior to the traditional red, green and blue (RGB) color model.
  • the YUV 4:2:0 form indicates that the two chrominance components U and V are only half of the luminance Y component in both the horizontal direction and the vertical direction, that is, there are four luminance components Y among the four sampled pixels, and the chrominance component There is only one U and V.
  • the amount of data is further reduced, which is only about 33% of the original.
  • the use of human eye physiological visual characteristics to achieve video compression through this color sampling method is one of the widely used video data compression methods.
  • Predictive coding that is, using the data information of the previously encoded frame to predict the frame currently to be encoded.
  • a predicted value is obtained by prediction, which is not completely equivalent to the actual value, and there is a certain residual value. If the prediction is more suitable, the closer the predicted value is to the actual value, the smaller the residual value, so that the residual value can be encoded to greatly reduce the amount of data, and the residual value plus the predicted value is used when decoding at the decoding end.
  • Restoring and reconstructing the initial image is the basic idea of predictive coding. In the mainstream coding standard, predictive coding is divided into two basic types: intra prediction and inter prediction.
  • Transform coding instead of directly encoding the original spatial domain information, the information sample value is converted from the current domain to another artificial domain (usually called the transform domain) according to some form of transformation function. Compression coding is performed according to the distribution characteristics of the information in the transform domain.
  • the reason for transform coding is that video image data tends to have large data correlation in the spatial domain, resulting in the existence of a large amount of redundant information, and direct encoding requires a large amount of bits.
  • the data correlation is greatly reduced, so that the redundant information of the encoding is reduced, and the amount of data required for encoding is also greatly reduced, so that a higher compression ratio can be obtained, and a better compression effect can be achieved.
  • Typical transform codes include a Kalo (K-L) transform, a Fourier transform, and the like.
  • Integer Discrete Cosine Transform (DCT) is a transform coding method commonly used in many international standards.
  • Quantization coding The above-mentioned transform coding does not compress data itself.
  • the quantization process is a powerful means of compressing data, and it is also the main reason for the loss of data in lossy compression.
  • the process of quantification is the process of forcing a large dynamic input value to be forced into fewer output values. Because the range of quantized input values is large, more bit numbers are needed, and the range of output values after "forced planning" is small, so that only a small number of bits can be represented.
  • Each quantized input is normalized to a quantized output, that is, quantized into an order of magnitude, often referred to as a quantization level (usually specified by the encoder).
  • the encoder control module selects the coding mode adopted by the image block according to the local characteristics of different image blocks in the video frame. Performing frequency domain or spatial domain prediction on the intra-predictive-coded block, performing motion-compensated prediction on the inter-predictive-coded block, and performing residual quantization coefficients (also referred to as transform quantization coefficients) by transforming and quantizing the residuals of the prediction. Finally, the final code stream is generated by an entropy encoder. In order to avoid the accumulation of prediction errors, the intra or inter prediction reference signals are obtained by the decoding module at the encoding end.
  • the transformed and quantized residual coefficients are reconstructed by inverse quantization and inverse transform, and then added to the predicted reference signal to obtain a reconstructed image.
  • the loop filtering performs pixel correction on the reconstructed image to improve the encoding quality of the reconstructed image.
  • the present application provides a decoding method, an encoding method, a decoding device, and an encoding device, which can improve encoding efficiency.
  • a decoding method comprising: obtaining an AC component and a DC component residual of a transformed quantized coefficient of a target image block from a code stream; and inversely quantizing the AC component and the preset DC component And inverse transform, obtaining a transition residual, and adding the transition residual to the predicted value of the target image block to obtain a transition reconstructed image block of the target image block; reconstructing the pixel in the reference pixel region of the image block and the target image block according to the transition Similarity, determining a predicted value of a DC component of the transformed quantized coefficient of the target image block; determining an original DC component of the target image according to the predicted value of the DC component and the DC component residual; inversely quantizing the original DC component and the AC component of the target image And inverse transform, obtaining a residual signal of the target image, and decoding the target image according to the residual signal.
  • the reconstructed transform quantized coefficients may be obtained from the code stream, wherein the reconstructed transform quantized coefficients include a difference between the predicted values of the first DC component and the DC component of the target image block and the communication of the target image block AC component; performing inverse quantization and inverse transform on the transform transform quantized coefficients to obtain a transition reconstructed image block of the target image block, wherein the transition transform quantized coefficient includes a second DC component and an AC component, and the second DC component is a preset value; Reconstructing a similarity between a pixel of the image block and a pixel in a reference pixel region of the target image block, determining a predicted value of the DC component; a difference between the predicted value of the DC component and the DC component (ie, a DC component residual) and a DC component Predicting a value, determining a DC component (ie, an original DC component); performing inverse quantization and inverse transform on the initial transformed quantized coefficients to obtain a residual signal of the
  • an encoding method comprising: transforming and quantizing a residual signal of a target image block to obtain a DC component and an AC AC component of a transform quantized coefficient of the target image block; And the preset DC component is inversely quantized and inverse transformed to obtain a transition residual, and the transition residual is added to the predicted value of the target image block to obtain a transition reconstructed image block of the target image block; and the pixels of the reconstructed image block are reconstructed according to the transition Determining the similarity of the pixels in the reference pixel region of the target image block, determining the predicted value of the DC component of the transformed quantized coefficient of the target image block; determining the DC component residual of the target image based on the predicted value of the DC component and the original DC component of the target image block Poor; write the AC component and DC component residuals to the code stream.
  • the residual signal of the target image block may be transformed and quantized to obtain an initial transform quantized coefficient of the target image block, and the initial transform quantized coefficient includes a first DC component and an AC AC component;
  • the coefficients are inverse quantized and inverse transformed to obtain a transition reconstructed image block of the target image block, wherein the transition transform quantized coefficient includes a second DC component and an AC AC component, and the second DC component is a preset value; and the pixels of the reconstructed image block are reconstructed according to the transition a similarity of pixels in a reference pixel region of the target image block, determining a predicted value of the DC component (ie, the original DC component); writing the reconstructed transform quantized coefficient to the code stream, wherein the reconstructed transform quantized coefficient includes a DC component and a predicted value The difference (ie the DC component residual) and the AC component.
  • the transform quantized coefficients may be inverse quantized and inverse transformed with the DC component set to a preset value to obtain a transition reconstructed image block of the target image block, and the pixels of the image block are reconstructed according to the transition.
  • the similarity of the pixels in the reference pixel region of the target image block determines the predicted value of the DC component.
  • a DC component residual can be used instead of the DC component. Since the absolute value of the DC component residual is smaller than the DC component, the coding efficiency is improved.
  • the technical solution of the present application determines the predicted value of the DC component according to the similarity between the pixel of the transition reconstructed image block whose DC component is a preset value and the pixel in the reference pixel region, and improves the accuracy of the DC component prediction value, so that DC The component residual is smaller, and therefore, the coding efficiency is higher than the scheme in which the predicted value of the DC component of the reference pixel region is directly used as the predicted value of the DC component of the target image block.
  • determining the predicted value of the DC component of the transform quantized coefficient of the target image block according to the similarity between the transition reconstructed image block and the pixel in the reference pixel region of the target image block including: in the target image block Determining at least one line and a first set of pixel signals corresponding to each line and an adjacent second set of pixel signals, wherein the first set of pixel signals comprises a first pixel signal of the transition reconstructed image block, in a direction specified by the prediction mode,
  • the second set of pixel signals includes a second pixel signal of the reference pixel region, the first pixel signal is adjacent to the second pixel signal; and the offset of the first pixel signal is solved such that the first pixel signal of the at least one line is increased by an offset
  • the sum of the squares of the second reconstructed signal and the second step of the second pixel signal is the smallest, wherein the pixel signal used to represent the two steps in the transition reconstructed image block is increased in offset, and the offset is used to represent DC
  • the two-step degree can more accurately reflect the correlation or similarity between the pixel of the target image block and the reference image block, so that the DC component residual is made. It is smaller, which further improves the coding efficiency.
  • the first set of pixel signals and the second set of pixel signals corresponding to each of the at least one line satisfy one of the following two formulas and formulas:
  • ⁇ 1 is used to indicate a threshold value when the reconstructed signal of the first group of pixel signals has two steps
  • ⁇ 2 is used to indicate a threshold value when the reconstructed signal of the first group of pixel signals does not have two steps.
  • i k is the number of at least one line
  • j is the number of the pixel signal on each of the at least one line.
  • the offset is calculated according to the following formula:
  • ⁇ x is the offset
  • C represents a set of lines satisfying one of the two formulas
  • q is the number of the pixel signal on each of the at least one line. Since only the pixels on some lines with strong correlation are selected for calculating the predicted value of the DC component, the accuracy of the predicted value of the DC component is further improved, so that the DC component residual is smaller, thereby further improving the coding efficiency.
  • determining the predicted value of the DC component of the transform quantized coefficient of the target image block according to the similarity between the transition reconstructed image block and the pixel in the reference pixel region of the target image block including: in the target image block Determining, in the direction specified by the prediction mode, a plurality of pixel pairs and a first pixel signal corresponding to each pixel pair and an adjacent second pixel signal, wherein the first pixel signal is a pixel signal of the transition reconstructed image block, and the second pixel The signal is a pixel signal of the reference pixel region; the first pixel signal is increased by an offset to obtain the weight of the first pixel signal Constructing a signal, wherein the offset is used to represent a predicted value before quantization of the DC component; and the offset is calculated such that a sum of squared degrees of the reconstructed signal of the first pixel signal of the plurality of pixel pairs and the second pixel signal The smallest; the value of the offset obtained by the solution is quantized to obtain a predicted value
  • the predicted value of the DC component is calculated according to the following formula:
  • ⁇ x is the offset and n is the number of pixels in each row or column of the transition reconstructed image block
  • n the number of pixels in each row or column of the transition reconstructed image block
  • For the first pixel signal For the second pixel signal, versus Adjacent in the direction specified by the prediction mode. Since only the adjacent pixels between the transition reconstructed image block and the reference pixel region are selected for calculating the predicted value of the DC component, the computational complexity is reduced in the case of ensuring the accuracy of the DC component predicted value.
  • the first group of pixel signals and the second group of signals may be signals of the pixels themselves passing through the at least one line, or may be signals obtained by interpolating signals of pixels around the at least one line.
  • determining the predicted value of the DC component of the transform quantized coefficient of the target image block according to the similarity between the transition reconstructed image block and the pixel in the reference pixel region of the target image block including: acquiring the reference pixel region a first set of pixel signals located above the transition reconstructed image block, a second set of pixel signals on the left side of the transition reconstructed image block, a third set of pixel signals on the upper side of the transition reconstructed image block, and a fourth left side of the transition reconstructed image block
  • the group of pixel signals, the first group of pixel signals and the third group of pixel signals respectively comprise M rows of pixel signals, the second group of pixel signals and the fourth group of pixel signals respectively comprise H columns of pixel signals, M and H are positive integers; Comparing the difference between the average value of the group of pixel signals and the average value of the third group of pixel signals to obtain a first difference; calculating a difference between an average value of the second group of pixel signals and an average value
  • the computational complexity is reduced in the case of ensuring the accuracy of the DC component prediction value.
  • the average value is used to calculate the predicted value of the DC component, making the design of the encoder or decoder simple.
  • M is an integer greater than or equal to 2
  • N is an integer greater than or equal to 2.
  • M and H are multiple rows or columns, it is possible to make full use of the correlation of more signals around the target image block and the relationship between the target image block and the surrounding signals to predict the predicted value of the DC component, so that the DC component residual is more Small, the accuracy of the predicted value of the DC component is improved, thereby further improving the coding efficiency.
  • the target image block is a transform block. Since the transform and quantization and the inverse transform and the inverse quantization can be in units of transform blocks, the predictive value of the DC component is also determined in units of transform blocks, which can reduce the computational complexity of the single block and make the design of the encoder or decoder simple.
  • the preset DC component is zero. By setting the predicted value of the DC component to zero, the computational complexity can be reduced compared to setting it to other values.
  • the method further includes: determining a size of the DC component, and if the DC component is greater than a preset threshold, performing the foregoing method, if the DC component is less than a preset threshold, directly encoding according to the DC component and the AC component . Since the DC component can be adopted when the DC component is greater than the preset threshold, the improvement of the coding efficiency compensates for the performance loss caused by the algorithm complexity of predicting the DC residual, and improves the overall performance of the coding or decoding.
  • a decoding apparatus comprising means for performing the decoding method of the first aspect.
  • the decoding device includes: an entropy decoding module, configured to obtain, from the code stream, an AC AC component and a DC DC component residual of the transformed quantized coefficients of the target image block; and a first inverse quantization and inverse transform module, configured to the AC component and the preset DC points The quantity is inverse-quantized and inverse-transformed to obtain a transition residual, and the transition residual is added to the predicted value of the target image block to obtain a transition reconstructed image block of the target image block; and the prediction module is configured to reconstruct the image block and the target image block according to the transition a similarity of pixels in the reference pixel region, determining a predicted value of a DC component of the transform quantized coefficient of the target image block; and a second inverse quantization and inverse transform module for determining a predicted value and a DC component residual according to the DC component
  • the original DC component of the target image block is subjected to
  • an encoding device comprising means for performing the encoding method of the second aspect.
  • the encoding device includes: a transform and quantization module, configured to transform and quantize a residual signal of the target image block to obtain a DC component and an AC AC component of the transform quantized coefficient of the target image block; and an inverse quantization and inverse transform module, configured to: Performing inverse quantization and inverse transform on the AC component and the preset DC component to obtain a transition residual, and adding the transition residual to the predicted value of the target image block to obtain a transition reconstructed image block of the target image block; and a prediction module for Determining a predicted value of a DC component of a transform quantized coefficient of the target image block according to a similarity between a pixel of the transition reconstructed image block and a pixel in a reference pixel region of the target image block; and an entropy encoding module for predicting a value according to the DC component
  • the original DC component of the target image block determines the
  • a decoding apparatus includes a video decoder configured to: obtain an AC AC component and a DC DC component residual of a transform quantized coefficient of a target image block from a code stream; Performing inverse quantization and inverse transform on the AC component and the preset DC component to obtain a transition residual, and adding the transition residual to the predicted value of the target image block to obtain a transition reconstructed image block of the target image block; reconstructing the image block according to the transition Determining a predicted value of a DC component of a transform quantized coefficient of the target image block with a similarity of pixels in a reference pixel region of the target image block; determining an original DC component of the target image block according to the predicted value of the DC component and the DC component residual; The original DC component and the AC component of the target image block are inverse quantized and inverse transformed to obtain a residual signal of the target image block, and the target image block is decoded according to the residual signal.
  • an encoding apparatus includes a video decoder configured to perform transforming and quantizing a residual signal of a target image block to obtain a transform quantized coefficient of a target image block.
  • DC DC component and AC AC component inversely quantize and inverse transform the AC component and the preset DC component to obtain a transition residual, and add the transition residual to the predicted value of the target image block to obtain a transition of the target image block.
  • Reconstructing the image block determining a predicted value of the DC component of the transformed quantized coefficient of the target image block according to the similarity between the pixel of the transition reconstructed image block and the pixel in the reference pixel region of the target image block; and the predicted value and the target image according to the DC component
  • the original DC component of the block determines the DC component residual of the target image block; the AC component and the DC component residual are written to the code stream.
  • a computer readable storage medium storing instructions, when executed, causes one or more processors of a decoding device to: perform transform quantization of a target image block from a code stream The AC component of the coefficient and the residual of the DC component of the DC; inversely and inversely transform the AC component and the preset DC component to obtain a transition residual, and add the transition residual to the predicted value of the target image block to obtain the target image.
  • a transition reconstructed image block of the block determining a predicted value of a DC component of the transform quantized coefficient of the target image block according to a similarity between the transition reconstructed image block and a pixel in a reference pixel region of the target image block; and a predicted value according to the DC component and the DC
  • the component residual determines the original DC component of the target image block; inversely quantizes and inverse transforms the original DC component and the AC component of the target image block to obtain a residual signal of the target image block, and performs the target image block according to the residual signal. decoding.
  • a computer readable storage medium storing instructions, when executed, causes one or more processors of an encoding device to: transform a residual signal of a target image block and the amount
  • the DC component and the AC AC component of the transform quantized coefficient of the target image block are obtained, and the AC component and the preset DC component are inverse quantized and inverse transformed to obtain a transition residual, and the transition residual is matched with the target image block.
  • the predicted values are added to obtain a transition reconstructed image block of the target image block; and the predicted value of the DC component of the transformed quantized coefficient of the target image block is determined according to the similarity between the pixels of the transition reconstructed image block and the pixels in the reference pixel region of the target image block. Determining a DC component residual of the target image block based on the predicted value of the DC component and the original DC component of the target image block; writing the AC component and the DC component residual to the code stream.
  • FIG. 1 is a schematic block diagram of a video encoding system in accordance with an embodiment of the present invention
  • FIG. 2 is a schematic diagram of an apparatus for video encoding in accordance with an embodiment of the present invention
  • FIG. 3 is a schematic block diagram of another video codec system according to an embodiment of the present invention.
  • FIG. 4 is a schematic block diagram of a video encoder in accordance with an embodiment of the present invention.
  • FIG. 5 is a schematic block diagram of a video decoder according to an embodiment of the present invention.
  • FIG. 6 is a schematic block diagram of a video encoder in accordance with an embodiment of the present invention.
  • FIG. 7 is a schematic diagram of a video decoder in accordance with an embodiment of the present invention.
  • FIG. 8 is a schematic flowchart of an encoding method according to an embodiment of the present invention.
  • FIG. 9 is a schematic flowchart of a decoding method according to an embodiment of the present invention.
  • Figure 10 shows 35 prediction modes of HEVC
  • 11A and 11B are schematic diagrams of selecting a pixel signal based on a prediction mode according to an embodiment of the present invention.
  • 12A-12D are schematic diagrams of selecting pixel signals based on a direction prediction mode according to an embodiment of the present invention.
  • FIG. 13A and FIG. 13B are schematic diagrams showing pixel signals selected based on a prediction mode according to another embodiment of the present invention.
  • FIG. 14 is a schematic flow chart of a process of determining a predicted value of a DC component, in accordance with an embodiment of the present invention.
  • FIG. 15 is a schematic flowchart of a process of determining a predicted value of a DC component according to another embodiment of the present invention.
  • 16 is a schematic flowchart of a process of determining a predicted value of a DC component according to another embodiment of the present invention.
  • 17 is a schematic structural diagram of an embodiment of the present invention applicable to a television application
  • Figure 18 is a schematic block diagram of an embodiment of the present invention suitable for use in a mobile phone application.
  • FIG. 1 is a schematic block diagram of a video codec device or electronic device 50 that may incorporate a codec in accordance with an embodiment of the present invention.
  • 2 is a schematic diagram of an apparatus for video encoding in accordance with an embodiment of the present invention. The units in Figures 1 and 2 will be explained below.
  • the electronic device 50 can be, for example, a mobile terminal or user equipment of a wireless communication system. It should be understood that embodiments of the invention may be practiced in any electronic device or device that may require encoding and decoding, or encoding, or decoding of a video image.
  • Device 50 can include a housing 30 for incorporating and protecting the device.
  • Device 50 may also include display 32 in the form of a liquid crystal display.
  • the display may be any suitable display technology suitable for displaying images or video.
  • Device 50 may also include a keypad 34.
  • any suitable data or user interface mechanism may be utilized.
  • the user interface can be implemented as a virtual keyboard or data entry system as part of a touch sensitive display.
  • the device can include a microphone 36 or any suitable audio input, The audio input can be a digital or analog signal input.
  • the apparatus 50 may also include an audio output device, which in an embodiment of the invention may be any of the following: an earphone 38, a speaker, or an analog audio or digital audio output connection.
  • Device 50 may also include battery 40, and in other embodiments of the invention, the device may be powered by any suitable mobile energy device, such as a solar cell, fuel cell, or clock mechanism generator.
  • the device may also include an infrared port 42 for short-range line of sight communication with other devices.
  • device 50 may also include any suitable short range communication solution, such as a Bluetooth wireless connection or a USB/Firewire wired connection.
  • Device 50 may include a controller 56 or processor for controlling device 50.
  • Controller 56 may be coupled to memory 58, which may store data in the form of images and audio in an embodiment of the invention, and/or may also store instructions for execution on controller 56.
  • Controller 56 may also be coupled to codec circuitry 54 suitable for implementing encoding and decoding of audio and/or video data or assisted encoding and decoding by controller 56.
  • the apparatus 50 may also include a card reader 48 and a smart card 46 for providing user information and for providing authentication information for authenticating and authorizing users on the network, such as a Universal Integrated Circuit Card (UICC) and a UICC read. Take the device.
  • a card reader 48 and a smart card 46 for providing user information and for providing authentication information for authenticating and authorizing users on the network, such as a Universal Integrated Circuit Card (UICC) and a UICC read. Take the device.
  • UICC Universal Integrated Circuit Card
  • UICC Universal Integrated Circuit Card
  • Apparatus 50 may also include a radio interface circuit 52 coupled to the controller and adapted to generate, for example, a wireless communication signal for communicating with a cellular communication network, a wireless communication system, or a wireless local area network. Apparatus 50 may also include an antenna 44 coupled to radio interface circuitry 52 for transmitting radio frequency signals generated at radio interface circuitry 52 to other apparatus(s) and for receiving radio frequency signals from other apparatus(s).
  • a radio interface circuit 52 coupled to the controller and adapted to generate, for example, a wireless communication signal for communicating with a cellular communication network, a wireless communication system, or a wireless local area network.
  • Apparatus 50 may also include an antenna 44 coupled to radio interface circuitry 52 for transmitting radio frequency signals generated at radio interface circuitry 52 to other apparatus(s) and for receiving radio frequency signals from other apparatus(s).
  • device 50 includes a camera capable of recording or detecting a single frame, and codec 54 or controller receives the individual frames and processes them.
  • the device may receive video image data to be processed from another device prior to transmission and/or storage.
  • device 50 may receive images for encoding/decoding via a wireless or wired connection.
  • FIG. 3 is a schematic block diagram of another video codec system 10 in accordance with an embodiment of the present invention.
  • video codec system 10 includes source device 12 and destination device 14.
  • Source device 12 produces encoded video data.
  • source device 12 may be referred to as a video encoding device or a video encoding device.
  • Destination device 14 may decode the encoded video data produced by source device 12.
  • destination device 14 may be referred to as a video decoding device or a video decoding device.
  • Source device 12 and destination device 14 may be examples of video codec devices or video codec devices.
  • Source device 12 and destination device 14 may include a wide range of devices including desktop computers, mobile computing devices, notebook (eg, laptop) computers, tablet computers, set top boxes, smart phones, etc., televisions, cameras, display devices , digital media player, video game console, on-board computer, or the like.
  • Channel 16 may include one or more media and/or devices capable of moving encoded video data from source device 12 to destination device 14.
  • channel 16 may include one or more communication media that enable source device 12 to transmit encoded video data directly to destination device 14 in real time.
  • source device 12 may modulate the encoded video data in accordance with a communication standard (eg, a wireless communication protocol) and may transmit the modulated video data to destination device 14.
  • the one or more communication media may include wireless and/or wired communication media, such as a radio frequency (RF) spectrum or one or more physical transmission lines.
  • RF radio frequency
  • One or more communication media may form part of a packet-based network (eg, a local area network, a wide area network, or a global network (eg, the Internet)).
  • the one or more communication media may include routers, switches, base stations, or other devices that facilitate communication from the source device 12 to the destination device 14.
  • channel 16 can include a storage medium that stores encoded video data generated by source device 12.
  • destination device 14 can access the storage medium via disk access or card access.
  • the storage medium may include a variety of locally accessible data storage media, such as Blu-ray Disc, DVD, CD-ROM, flash memory, or other suitable digital storage medium for storing encoded video data.
  • channel 16 can include a file server or another intermediate storage device that stores encoded video data generated by source device 12.
  • destination device 14 may access the encoded video data stored at a file server or other intermediate storage device via streaming or download.
  • the file server may be a server type capable of storing encoded video data and transmitting the encoded video data to the destination device 14.
  • the instance file server includes a web server (eg, for a website), a file transfer protocol (FTP) server, a network attached storage (NAS) device, and a local disk drive.
  • FTP file transfer protocol
  • NAS network attached storage
  • Destination device 14 can access the encoded video data via a standard data connection (e.g., an internet connection).
  • a standard data connection e.g., an internet connection.
  • An instance type of a data connection includes a wireless channel (eg, a Wi-Fi connection), a wired connection (eg, DSL, cable modem, etc.), or both, suitable for accessing encoded video data stored on a file server. combination.
  • the transmission of the encoded video data from the file server may be streaming, downloading, or a combination of both.
  • the technology of the present invention is not limited to a wireless application scenario.
  • the technology can be applied to video codecs supporting multiple multimedia applications such as aerial television broadcasting, cable television transmission, satellite television transmission, and streaming video transmission ( For example, via the Internet), encoding of video data stored on a data storage medium, decoding of video data stored on a data storage medium, or other application.
  • video codec system 10 may be configured to support one-way or two-way video transmission to support applications such as video streaming, video playback, video broadcasting, and/or video telephony.
  • source device 12 includes video source 18, video encoder 20, and output interface 22.
  • output interface 22 can include a modulator/demodulator (modem) and/or a transmitter.
  • Video source 18 may include a video capture device (eg, a video camera), a video archive containing previously captured video data, a video input interface to receive video data from a video content provider, and/or a computer for generating video data.
  • Video encoder 20 may encode video data from video source 18.
  • source device 12 transmits the encoded video data directly to destination device 14 via output interface 22.
  • the encoded video data may also be stored on a storage medium or file server for later access by the destination device 14 for decoding and/or playback.
  • destination device 14 includes an input interface 28, a video decoder 30, and a display device 32.
  • input interface 28 includes a receiver and/or a modem.
  • Input interface 28 can receive the encoded video data via channel 16.
  • Display device 32 may be integral with destination device 14 or may be external to destination device 14. In general, display device 32 displays the decoded video data.
  • Display device 32 may include a variety of display devices such as liquid crystal displays (LCDs), plasma displays, organic light emitting diode (OLED) displays, or other types of display devices.
  • LCDs liquid crystal displays
  • OLED organic light emitting diode
  • Video encoder 20 and video decoder 30 may operate in accordance with a video compression standard (eg, the High Efficiency Video Codec H.265 standard) and may conform to the HEVC Test Model (HM).
  • a video compression standard eg, the High Efficiency Video Codec H.265 standard
  • HM HEVC Test Model
  • ITU-T.265(V3)(04/2015) was released on April 29, 2015 and is available for download from http://handle.itu.int/11.1002/1000/12455 The entire content is incorporated herein by reference.
  • video encoder 20 and video decoder 30 may operate in accordance with other proprietary or industry standards, including ITU-TH.261, ISO/IEC MPEG-1 Visual, ITU-TH.262, or ISO/IEC MPEG-2 Visual, ITU-TH.263, ISO/IEC MPEG-4 Visual, ITU-TH.264 (also known as ISO/IEC MPEG-4 AVC), includes scalable video codec (SVC) and multi-view video codec (MVC) extensions.
  • SVC scalable video codec
  • MVC multi-view video codec
  • FIG. 3 is merely an example and the techniques of the present invention are applicable to video codec applications (eg, single-sided video encoding or video decoding) that do not necessarily include any data communication between the encoding device and the decoding device.
  • data is retrieved from local memory, data is streamed over a network, or manipulated in a similar manner.
  • the encoding device may encode the data and store the data to a memory, and/or the decoding device may retrieve the data from the memory and decode the data.
  • encoding and decoding are performed by a plurality of devices that only encode data to and/or retrieve data from the memory and decode the data by not communicating with each other.
  • Video encoder 20 and video decoder 30 may each be implemented as any of a variety of suitable circuits, such as one or more microprocessors, digital signal processors (DSPs), application specific integrated circuits (ASICs), field programmable Gate array (FPGA), discrete logic, hardware, or any combination thereof. If the technology is implemented partially or wholly in software, the device may store the instructions of the software in a suitable non-transitory computer readable storage medium, and the instructions in the hardware may be executed using one or more processors to perform the techniques of the present invention. . Any of the foregoing (including hardware, software, a combination of hardware and software, etc.) can be considered as one or more processors. Each of video encoder 20 and video decoder 30 may be included in one or more encoders or decoders, any of which may be integrated into a combined encoder/decoder (codec) in other devices Part of the (CODEC).
  • codec combined encoder/decoder
  • the invention may generally refer to video encoder 20 "signaling" certain information to another device (e.g., video decoder 30).
  • the term “signaling” may generally refer to a syntax element and/or to convey the communication of encoded video data. This communication can occur in real time or near real time. Alternatively, this communication may occur over a time span, such as may occur when encoding the encoded data to a computer readable storage medium at the time of encoding, the syntax elements being stored by the decoding device and then by the decoding device Retrieve at any time.
  • Video encoder 20 encodes the video data.
  • Video data may include one or more pictures.
  • Video encoder 20 may generate a code stream that contains encoded information for the video data in the form of a bitstream.
  • the encoded information may include encoded picture data and associated data.
  • Associated data can include sequence parameter sets (SPS), picture parameter sets (PPS), and other syntax structures.
  • SPS sequence parameter sets
  • PPS picture parameter sets
  • An SPS can contain parameters that are applied to zero or more sequences.
  • the PPS can contain parameters that are applied to zero or more pictures.
  • a grammatical structure refers to a collection of zero or more syntax elements arranged in a specified order in a code stream.
  • video encoder 20 may partition the picture into a raster of coded tree blocks (CTB).
  • CTB coded tree blocks
  • a CTB may be referred to as a "tree block,” a "maximum coding unit” (LCU), or a "coding tree unit.”
  • the CTB is not limited to a particular size and may include one or more coding units (CUs).
  • Each CTB can be associated with a block of pixels of equal size within the picture.
  • Each pixel can correspond to one luminance (luminance or luma) sample and two chrominance or chroma samples.
  • each CTB can be associated with one luma sample block and two chroma sample blocks.
  • the CTB of a picture can be divided into one or more stripes.
  • each stripe contains an integer number of CTBs.
  • video encoder 20 may generate coded information for each strip of the picture, i.e., the CTB within the encoded strip.
  • video encoder 20 may recursively perform quadtree partitioning on the block of pixels associated with the CTB to partition the block of pixels into decreasing blocks of pixels. Smaller blocks of pixels can be associated with the CU.
  • FIG. 4 is a schematic block diagram of a video encoder 20 according to an embodiment of the present invention, including an encoding end prediction module 201, a transform quantization module 202, an entropy encoding module 203, a code reconstruction module 204, and an encoding end filtering module 205.
  • FIG. A schematic block diagram of the video decoder 30 of the embodiment of the present invention includes a decoding end prediction module 206, an inverse transform inverse quantization module 207, an entropy decoding module 208, a decoding reconstruction module 209, and a decoding filtering module 210. specific:
  • Video encoder 20 may generate one or more prediction units (PUs) that each no longer partition the CU (ie, the smallest sized CU). Each PU of a CU may be associated with a different block of pixels within a block of pixels of the CU. Video encoder 20 may generate predictive pixel blocks for each PU of the CU. Video encoder 20 may use intra prediction or inter prediction to generate predictive pixel blocks for the PU. If video encoder 20 uses intra prediction to generate a predictive pixel block of a PU, video encoder 20 may generate a decoded pixel based on a picture (or referred to as an image) associated with the PU and located within the same video frame.
  • PUs prediction units
  • the predictive pixel block of the PU If video encoder 20 uses inter prediction to generate a predictive pixel block for a PU, video encoder 20 may be based on one or more different from the picture associated with the PU, ie, different than the video frame in which the PU is located. The decoded pixels of the picture are used to generate a predictive pixel block of the PU. Video encoder 20 may generate residual pixel blocks of the CU based on the predictive pixel blocks of the PU of the CU. The residual pixel block of the CU may indicate a difference (or referred to as a residual signal or residual data) between the sampled value in the predictive pixel block of the PU of the CU and the corresponding sampled value in the initial block of pixels of the CU.
  • Transform quantization module 202 is operative to process the predicted residual signal or residual data.
  • Video encoder 20 may perform recursive quadtree partitioning on the residual pixel blocks of the CU to partition the residual pixel blocks of the CU into one or more smaller residual pixel blocks associated with the transform units (TUs) of the CU. Because the pixels in the pixel block associated with the TU each correspond to one luma sample and two chroma samples, each TU can be associated with one luma residual sample block and two chroma residual sample blocks.
  • Video encoder 20 may apply one or more transforms to the residual sample block associated with the TU to generate a coefficient block (ie, a block of coefficients).
  • the transform can be a discrete cosine transform, ie a DCT transform, or a variant thereof, such as a discrete sine transform, ie a DST transform.
  • a transform matrix of DCT a two-dimensional transform is calculated by applying a one-dimensional transform in the horizontal and vertical directions to obtain a coefficient block (also referred to as a transform coefficient).
  • Video encoder 20 may perform a quantization procedure for each of the transform coefficients. Quantization generally refers to the quantization of transform coefficients to reduce the amount of data used to represent the coefficients, which can be understood as the process of mapping an infinite number of data into a finite number of data to provide further compression.
  • the transform quantized coefficients After transforming and quantizing, the transform quantized coefficients are obtained, and the transform quantized coefficients include a DC component and an AC component.
  • the process of obtaining the DC component and the AC component is not the focus of the present invention, and the expansion process is not described here.
  • the specific solution process can refer to the ITU. -TH.265 (V3) (04/2015).
  • the process performed by inverse transform inverse quantization module 207 includes the inverse of transform quantization module 202.
  • Video encoder 20 may generate a set of syntax elements that represent coefficients in the quantized coefficient block.
  • Video encoder 20 may apply an entropy encoding operation (eg, a context adaptive binary arithmetic coding (CABAC) operation) to some or all of the above syntax elements by entropy encoding module 203.
  • CABAC context adaptive binary arithmetic coding
  • video encoder 20 may binarize the syntax elements to form a binary sequence that includes one or more bits (referred to as "binary").
  • Video encoder 20 may encode a portion of the binary using regular encoding, and may use bypass encoding to encode other portions of the binary.
  • video encoder 20 may apply inverse quantization and inverse transform to the transformed coefficient block by code reconstruction module 204 to reconstruct the residual sample block from the transformed coefficient block.
  • Video encoder 20 may add the reconstructed residual sample block to a corresponding sample block of one or more predictive sample blocks to produce a reconstructed sample block.
  • video encoder 20 may reconstruct the block of pixels associated with the TU. The pixel block of each TU of the CU is reconstructed in this way until the entire pixel block reconstruction of the CU is completed.
  • Video encoder 20 may further include a DC component prediction module 206 for determining a predicted value of the DC component, in particular, DC component prediction module 206 may be based on transform quantization coefficients output by transform quantization module 202, in accordance with an embodiment of the present invention.
  • DC component prediction module 206 may be based on transform quantization coefficients output by transform quantization module 202, in accordance with an embodiment of the present invention. Construct a new set of transform quantized coefficients, DCs in the constructed new transform quantized coefficients The component is forced to 0, and the AC component is the AC component output by the transform quantization module 202.
  • the inverse transform and inverse quantization module 207 performs inverse quantization and inverse processing on the new transform quantized coefficient, and obtains the inverse transform and the inverse quantization process.
  • the reconstructed residual sample block, and the reconstructed residual sample block may be added to a corresponding sample block of one or more predictive sample blocks to produce a reconstructed sample block (hereinafter also referred to as a transition reconstructed image block).
  • the predicted value of the DC component is determined according to the similarity or correlation of the pixels of the transition reconstructed image block with the pixels of the reference pixel region.
  • a residual signal of the DC component is obtained from the predicted value of the DC component and the DC component output by the transform quantization module 202.
  • the residual signal and the AC component of the DC component are written to the code stream by the entropy encoding module 203.
  • video encoder 20 After video encoder 20 reconstructs the block of pixels of the CU, video encoder 20 performs a deblocking filtering operation through encoding end filtering module 205 to reduce the blockiness of the block of pixels associated with the CU. After video encoder 20 performs the deblocking filtering operation, video encoder 20 may use sample adaptive offset (SAO) to modify the reconstructed block of pixels of the CTB of the picture. After performing these operations, video encoder 20 may store the reconstructed blocks of pixels of the CU in a decoded picture buffer for use in generating predictive blocks of pixels for other CUs.
  • SAO sample adaptive offset
  • Video decoder 30 can receive the code stream.
  • the code stream contains the encoded information of the video data encoded by the video encoder 20 in the form of a bit stream.
  • Video decoder 30 parses the code stream through entropy decoding module 208 to extract syntax elements from the code stream.
  • video decoder 30 may perform regular decoding on partial bins and may perform bypass decoding on bins of other portions, and the bins in the code stream have mapping relationships with syntax elements, through parsing The binary gets the syntax element.
  • Video decoder 30, through decoding reconstruction module 209, may reconstruct a picture of the video data based on the syntax elements extracted from the code stream.
  • the process of reconstructing video data based on syntax elements is generally reciprocal to the process performed by video encoder 20 to generate syntax elements.
  • video decoder 30 may generate a predictive pixel block of a PU of a CU based on syntax elements associated with the CU.
  • video decoder 30 may inverse quantize the coefficient blocks associated with the TUs of the CU.
  • Video decoder 30 may perform an inverse transform on the inverse quantized coefficient block to reconstruct a residual pixel block associated with the TU of the CU.
  • Video decoder 30 may reconstruct a block of pixels of the CU based on the predictive pixel block and the residual pixel block.
  • the video decoder 30 may further include a DC component prediction module 211 for determining a predicted value of the DC component, in particular, the DC component prediction module 211 may be based on the transform coefficients obtained from the code stream.
  • the DC component of the reconstructed another transform coefficient is forced to 0, and the AC component is the AC component output by the entropy decoding module 208, and the weight is inversed by the inverse quantization and inverse transform module.
  • Another transform module of the structure performs an inverse quantization and inverse transform process, reconstructs a residual sample block obtained by an inverse transform and an inverse quantization process, and can add the reconstructed residual sample block to a corresponding sample of one or more predictive sample blocks Block to generate a reconstructed sample block (hereinafter also referred to as a transition reconstructed image block).
  • the predicted value of the DC component is determined according to the similarity or correlation of the pixels of the transition reconstructed image block with the pixels of the reference pixel region.
  • the reconstructed DC component is obtained from the predicted value of the DC component and the residual signal of the DC component output by the entropy decoding module 208.
  • video decoder 30 After video decoder 30 reconstructs the block of pixels of the CU, video decoder 30 performs a deblocking filtering operation through decoding filter module 210 to reduce the blockiness of the block of pixels associated with the CU. Additionally, video decoder 30 may perform the same SAO operations as video encoder 20 based on one or more SAO syntax elements. After video decoder 30 performs these operations, video decoder 30 may store the block of pixels of the CU in a decoded picture buffer. The decoded picture buffer can provide reference pictures for subsequent motion compensation, intra prediction, and presentation by the display device.
  • FIG. 6 is a block diagram illustrating an example video encoder 20 that is configured to implement the techniques of the present invention. It should be understood that FIG. 6 is exemplary and should not be considered as limiting the techniques as broadly exemplified and described herein. As shown in FIG.
  • the video encoder 20 includes a prediction processing unit 100, a residual generation unit 102, a transform processing unit 104, a quantization processing unit 106, Inverse quantization processing unit 108, inverse transform processing unit 110, reconstruction unit 112, filter unit 113, decoded picture buffer 114, inverse quantization processing unit 131, inverse transform processing unit 132, reconstruction unit 133, DC within DC prediction module 130
  • the entropy encoding unit 116 includes a regular CABAC codec engine and a bypass codec engine.
  • the prediction processing unit 100 includes an inter prediction processing unit 121 and an intra prediction processing unit 126.
  • the inter prediction processing unit 121 includes a motion estimation unit and a motion compensation unit.
  • video encoder 20 may include more, fewer, or different functional components.
  • Video encoder 20 receives the video data.
  • video encoder 20 may encode each strip of each picture of the video data.
  • video encoder 20 may encode each CTB in the strip.
  • prediction processing unit 100 may perform quadtree partitioning on the pixel blocks associated with the CTB to divide the block of pixels into decreasing blocks of pixels. For example, prediction processing unit 100 may partition a block of pixels of a CTB into four equally sized sub-blocks, split one or more of the sub-blocks into four equally sized sub-sub-blocks, and the like.
  • Video encoder 20 may encode the CU of the CTB in the picture to generate coded information for the CU.
  • Video encoder 20 may encode the CU of the CTB according to the fold scan order. In other words, video encoder 20 may encode the CU by the upper left CU, the upper right CU, the lower left CU, and then the lower right CU.
  • video encoder 20 may encode the CU associated with the sub-block of the pixel block of the partitioned CU according to the fold scan order.
  • prediction processing unit 100 can partition the pixel blocks of the CU in one or more PUs of the CU.
  • Video encoder 20 and video decoder 30 can support a variety of PU sizes. Assuming that the size of a particular CU is 2N ⁇ 2N, video encoder 20 and video decoder 30 may support a PU size of 2N ⁇ 2N or N ⁇ N for intra prediction, and support 2N ⁇ 2N, 2N ⁇ N, N ⁇ 2N, N x N or similarly sized symmetric PUs for inter prediction. Video encoder 20 and video decoder 30 may also support asymmetric PUs of 2N x nU, 2N x nD, nL x 2N, and nR x 2N for inter prediction.
  • the inter prediction processing unit 121 may generate predictive data of the PU by performing inter prediction on each PU of the CU.
  • the predictive data of the PU may include motion information corresponding to the predictive pixel block of the PU and the PU.
  • the strip can be an I strip, a P strip or a B strip.
  • the inter prediction unit 121 may perform different operations on the PU of the CU depending on whether the PU is in an I slice, a P slice, or a B slice. In the I slice, all PUs perform intra prediction.
  • motion estimation unit 122 may search for a reference picture in a list of reference pictures (eg, "List 0") to find a reference block for the PU.
  • the reference block of the PU may be the pixel block that most closely corresponds to the pixel block of the PU.
  • Motion estimation unit 122 may generate a reference picture index that indicates a reference picture of the PU-containing reference block in list 0, and a motion vector that indicates a spatial displacement between the pixel block of the PU and the reference block.
  • the motion estimation unit 122 may output the reference picture index and the motion vector as motion information of the PU.
  • Motion compensation unit 124 may generate a predictive pixel block of the PU based on the reference block indicated by the motion information of the PU.
  • motion estimation unit 122 may perform uni-directional inter prediction or bi-directional inter prediction on the PU.
  • motion estimation unit 122 may search for a reference picture of a first reference picture list ("List 0") or a second reference picture list ("List 1") to find a reference block for the PU.
  • the motion estimation unit 122 may output the following as the motion information of the PU: a reference picture index indicating a position in the list 0 or the list 1 of the reference picture containing the reference block, a space between the pixel block indicating the PU and the reference block The motion vector of the displacement, and the prediction direction indicator indicating whether the reference picture is in list 0 or in list 1.
  • motion estimation unit 122 may search for reference pictures in list 0 to find reference blocks for the PU, and may also search for reference pictures in list 1 to find another reference block for the PU.
  • Motion estimation unit 122 may generate a reference picture index indicating the list 0 of the reference picture containing the reference block and the location in list 1. Additionally, motion estimation unit 122 may generate a motion vector that indicates a spatial displacement between the reference block and the pixel block of the PU.
  • PU motion information can include PU Reference picture index and motion vector.
  • Motion compensation unit 124 may generate a predictive pixel block of the PU based on the reference block indicated by the motion information of the PU.
  • Intra prediction processing unit 126 may generate predictive data for the PU by performing intra prediction on the PU.
  • the predictive data of the PU may include predictive pixel blocks of the PU and various syntax elements.
  • Intra prediction processing unit 126 may perform intra prediction on I slices, P slices, and PUs within B slices.
  • intra-prediction processing unit 126 may use multiple intra-prediction modes to generate multiple sets of predictive data for the PU.
  • intra-prediction processing unit 126 may spread samples of sample blocks from neighboring PUs across sample blocks of the PU in a direction associated with the intra-prediction mode. It is assumed that the coding order from left to right and from top to bottom is used for PU, CU and CTB, and the adjacent PU may be above the PU, at the upper right of the PU, at the upper left of the PU or to the left of the PU.
  • Intra prediction processing unit 126 may use a different number of intra prediction modes, for example, 33 directional intra prediction modes. In some examples, the number of intra prediction modes may depend on the size of the pixel block of the PU.
  • the prediction processing unit 100 may select the predictive data of the PU of the CU from among the predictive data generated by the inter prediction processing unit 121 for the PU or the predictive data generated by the intra prediction processing unit 126 for the PU. In some examples, prediction processing unit 100 selects predictive data for the PU of the CU based on the rate/distortion metric of the set of predictive data. For example, a Lagrangian cost function is used to select between an encoding mode and its parameter values, such as motion vectors, reference indices, and intra prediction directions.
  • a predictive pixel block that selects predictive data may be referred to herein as a selected predictive pixel block.
  • Residual generation unit 102 may generate a residual pixel block of the CU based on the pixel block of the CU and the selected predictive pixel block of the PU of the CU. For example, the residual generation unit 102 may generate a residual pixel block of the CU such that each sample in the residual pixel block has a value equal to a difference between: a sample in a pixel block of the CU, and a PU of the CU Corresponding samples in the predictive pixel block are selected.
  • the prediction processing unit 100 may perform quadtree partitioning to partition the residual pixel block of the CU into sub-blocks. Each residual pixel block that is no longer partitioned may be associated with a different TU of the CU. The size and location of the residual pixel block associated with the TU of the CU is not necessarily related to the size and location of the pixel block of the CU-based PU.
  • Transform processing unit 104 may generate a coefficient block for each TU of the CU by applying one or more transforms to the residual sample block associated with the TU. For example, transform processing unit 104 may apply a discrete cosine transform (DCT), a directional transform, or a conceptually similar transform to the residual sample block.
  • DCT discrete cosine transform
  • Quantization unit 106 may quantize the coefficients in the coefficient block. For example, an n-bit coefficient can be truncated to an m-bit coefficient during quantization, where n is greater than m. Quantization unit 106 may quantize the coefficient block associated with the TU of the CU based on a quantization parameter (QP) value associated with the CU. Video encoder 20 may adjust the degree of quantization applied to the coefficient block associated with the CU by adjusting the QP value associated with the CU.
  • QP quantization parameter
  • the inverse quantization unit 108 and the inverse transform processing unit 110 may apply the inverse quantization and the inverse transform to the transformed system, respectively.
  • the blocks are reconstructed from the coefficient block to reconstruct the residual sample block.
  • Reconstruction unit 112 may add samples of the reconstructed residual sample block to corresponding samples of one or more predictive sample blocks generated by prediction processing unit 100 to generate a reconstructed sample block associated with the TU. By reconstructing the sample block of each TU of the CU in this manner, video encoder 20 may reconstruct the block of pixels of the CU.
  • the inverse quantization processing unit 131 and the inverse transform processing unit 132 may apply an inverse quantization and an inverse transform, respectively, to the transformed coefficient block and the DC component in the coefficient block is set to 0 to reconstruct the residual sample block from the coefficient block.
  • Reconstruction unit 133 may add samples of the reconstructed residual sample block to corresponding samples (also referred to as prediction values) of one or more predictive sample blocks generated by prediction processing unit 100 to generate reconstructed TUs associated with the TUs A sample block (hereinafter also referred to as a transition reconstructed image block).
  • the transition reconstructed image block is different from the reconstructed sample block generated by the reconstruction unit 112.
  • the DC component prediction processing unit 134 may determine the predicted value of the DC component based on the similarity or correlation of the transition reconstructed image block obtained by the reconstruction unit 133 with the pixels of the parameter pixel region.
  • the DC residual generation unit 135 may generate a DC residual (residual) based on the DC component prediction value output by the DC component prediction processing unit 134 and the original DC component generated by the quantization processing unit 106. For example, the DC component prediction value can be subtracted from the original DC component to obtain a DC residual.
  • the inverse quantization processing unit 131 and the inverse transform processing unit 132 may also multiplex the functions of the inverse quantization processing unit 108 and the inverse transform processing unit 110, respectively, in which case a coefficient reconstruction unit may be additionally provided for
  • the quantization processing unit 106 acquires the DC component of the transform quantized coefficient, forces the DC component to zero, and inputs the DC component that is forced to zero to the inverse quantization processing unit 108 for inverse quantization.
  • the inverse quantization processing unit 108 is operative to inverse quantize the AC component and the DC component that is forced to zero, and perform an inverse transform by the inverse transform processing unit 110 to reconstruct the residual sample block from the coefficient block.
  • reconstruction unit 133 may add samples of the reconstructed residual sample block to corresponding samples (also referred to as prediction values) of one or more predictive sample blocks generated by prediction processing unit 100 to generate a reconstruction associated with the TU.
  • the subsequent sample block hereinafter also referred to as a transition reconstructed image block).
  • Filter unit 113 may perform a deblocking filtering operation to reduce blockiness of pixel blocks associated with the CU. Further, the filter unit 113 may apply the SAO offset determined by the prediction processing unit 100 to the reconstructed sample block to restore the pixel block. Filter unit 113 may generate encoding information for the SAO syntax elements of the CTB.
  • the decoded picture buffer 114 can store the reconstructed block of pixels.
  • Inter prediction unit 121 may perform inter prediction on PUs of other pictures using reference pictures containing the reconstructed pixel blocks.
  • intra-prediction processing unit 126 can use the reconstructed block of pixels in decoded picture buffer 114 to perform intra-prediction on other PUs in the same picture as the CU.
  • Entropy encoding unit 116 may receive data from other functional components of video encoder 20. For example, entropy encoding unit 116 may receive the residual and AC components of the DC component from DC residual generation unit 135 and may receive syntax elements from prediction processing unit 100. Entropy encoding unit 116 may perform one or more entropy encoding operations on the data to generate entropy encoded data. For example, entropy encoding unit 116 may perform context adaptive variable length codec (CAVLC) operations, CABAC operations, variable to variable (V2V) length codec operations, grammar-based context adaptive binary arithmetic coding on data.
  • CAVLC context adaptive variable length codec
  • entropy encoding unit 116 may encode regular CABAC codec bins of syntax elements using regular CABAC engine 118, and may encode pass-through codec bins using bypass codec engine 120.
  • FIG. 7 is a block diagram illustrating an example video decoder 30 that is configured to implement the techniques of the present invention. It should be understood that FIG. 7 is exemplary and should not be considered as limiting the techniques as broadly exemplified and described herein.
  • the video decoder 30 includes an entropy decoding unit 150, a prediction processing unit 152, an inverse quantization unit 154, an inverse transform processing unit 156, and a heavy
  • the unit 158, the filter unit 159, the inverse quantization processing unit 171, the inverse transform processing unit 172, the reconstruction unit 173, the DC component prediction processing unit 174, the DC component generation unit 175, and the decoded picture buffer 160 in the DC prediction module 170 are constructed.
  • the prediction processing unit 152 includes an inter prediction processing unit 162 and an intra prediction processing unit 164.
  • the entropy decoding unit 150 includes a regular CABAC codec engine and a bypass codec engine.
  • video decoder 30 may include more, fewer, or different functional components.
  • Video decoder 30 can receive the code stream.
  • Entropy decoding unit 150 may parse the code stream to extract syntax elements from the code stream. As part of parsing the code stream, entropy decoding unit 150 may parse the entropy encoded syntax elements in the code stream. In addition, the entropy decoding unit 150 may also extract varying quantized coefficients from the code stream, including: a residual of the DC component and an AC component.
  • the inverse quantization processing unit 171 and the inverse transform processing unit 172 may respectively apply inverse quantization and inverse transform to the transformed coefficient block and the DC component in the coefficient block is set to 0 to reconstruct the residual sample block from the coefficient block.
  • Reconstruction unit 173 may add samples of the reconstructed residual sample block to corresponding samples (also referred to as prediction values) of one or more predictive sample blocks generated by prediction processing unit 152 to generate reconstructed TUs associated with the TUs.
  • Sampling block also known as a transition reconstructed image block. The transition reconstructed image block is different from the reconstructed sample block generated by the reconstruction unit 158, that is, the reconstructed image block used for acquiring the DC component is not directly used to reconstruct the decoded image.
  • the DC component prediction processing unit 174 may determine the predicted value of the DC component based on the similarity or correlation of the transition reconstructed image block obtained by the reconstruction unit 173 with the pixels of the parameter pixel region.
  • the DC residual generation unit 175 may generate a DC component based on the DC component prediction value output by the DC component prediction processing unit 174 and the DC residual (residual) generated by the entropy decoding unit 150. For example, the DC component prediction value can be added to the DC residual to obtain a DC component.
  • the inverse quantization processing unit 171 and the inverse transform processing unit 172 may also multiplex the functions of the inverse quantization processing unit 154 and the inverse transform processing unit 156, respectively, in which case another coefficient reconstruction unit may be provided for
  • the DC component of the transformed quantized coefficient is forced to zero, and the DC component that is forced to zero is input to the inverse quantization processing unit 154 for inverse quantization.
  • the inverse quantization processing unit 156 is operative to further inverse quantize the AC component and the forced-zero DC component, and inverse transform by the inverse transform processing unit 156 to reconstruct the residual sample block from the coefficient block.
  • reconstruction unit 173 can add samples of the reconstructed residual sample block to corresponding samples (also referred to as prediction values) of one or more predictive sample blocks generated by prediction processing unit 152 to generate a reconstruction associated with the TU.
  • the subsequent sample block hereinafter also referred to as a transition reconstructed image block).
  • the prediction processing unit 152, the inverse quantization unit 154, the inverse transform processing unit 156, the reconstruction unit 158, and the filter unit 159 may decode the video data according to syntax elements such as a DC component and an AC component, that is, generate decoded video data.
  • the syntax elements may include a regular CABAC codec binary and a bypass codec binary.
  • Entropy decoding unit 150 may use a regular CABAC codec engine 166 to decode the regular CABAC codec bins, and may use the bypass codec engine 168 to decode the bypass codec bins.
  • intra prediction processing unit 164 may perform intra prediction to generate a predictive sample block for the PU.
  • Intra-prediction processing unit 164 may use an intra-prediction mode to generate a predictive pixel block of a PU based on a block of pixels of a spatially neighboring PU.
  • Intra prediction processing unit 164 may determine an intra prediction mode for the PU based on one or more syntax elements parsed from the code stream.
  • the inter prediction processing unit 162 may include a motion compensation unit that may construct the first reference picture list and the second reference picture list according to syntax elements parsed from the code stream. Furthermore, if the PU uses inter prediction coding, the entropy decoding unit 150 may parse the motion information of the PU. The motion compensation unit 162 can be based on the motion signal of the PU Information to determine one or more reference blocks of the PU. Motion compensation unit 162 can generate a predictive pixel block of the PU from one or more reference blocks of the PU.
  • video decoder 30 may perform a reconstruction operation on a CU that is no longer split. To perform a reconstruction operation on a CU that is no longer split, video decoder 30 may perform a reconstruction operation on each TU of the CU. By performing a reconstruction operation on each TU of the CU, video decoder 30 may reconstruct the residual pixel blocks associated with the CU.
  • inverse quantization unit 154 may inverse quantize (ie, dequantize) the coefficient block associated with the TU. Inverse quantization unit 154 may determine the degree of quantization using the QP value associated with the CU of the TU, and determine the degree of inverse quantization that the inverse quantization unit 154 will apply.
  • inverse transform processing unit 156 may apply one or more inverse transforms to the coefficient block to generate a residual sample block associated with the TU.
  • inverse transform processing unit 156 may map inverse DCT, inverse integer transform, Karhunen-Loeve transform (KLT), inverse rotation transform, inverse directional transform, or other transform to the encoding end. The inverse transform is applied to the coefficient block.
  • Reconstruction unit 158 may use the residual pixel block associated with the TU of the CU and the predictive pixel block of the PU of the CU (ie, intra-prediction data or inter-prediction data) to reconstruct the block of pixels of the CU, where applicable.
  • reconstruction unit 158 can add samples of the residual pixel block to corresponding samples of the predictive pixel block to reconstruct the pixel block of the CU.
  • Filter unit 159 may perform a deblocking filtering operation to reduce the blockiness of the block of pixels associated with the CU of the CTB. Additionally, filter unit 159 can modify the pixel values of the CTB based on the SAO syntax elements parsed from the code stream. For example, filter unit 159 can determine the correction value based on the SAO syntax element of the CTB and add the determined correction value to the sample value in the reconstructed pixel block of the CTB. By modifying some or all of the pixel values of the CTB of the picture, the filter unit 159 can modify the reconstructed picture of the video data according to the SAO syntax element.
  • Video decoder 30 may store the block of pixels of the CU in decoded picture buffer 160.
  • the decoded picture buffer 160 may provide reference pictures for subsequent motion compensation, intra prediction, and presentation by a display device (eg, display device 32 of FIG. 3).
  • video decoder 30 may perform intra-prediction operations or inter-prediction operations on PUs of other CUs according to the blocks of pixels in decoded picture buffer 160.
  • the transform quantized coefficient of the write code stream is the residual and AC component of the DC component instead of being written like the conventional video coding technique.
  • the transform quantized coefficients of the incoming stream are the DC component and the AC component. Since the absolute value of the residual of the DC component is usually smaller than the DC component, the smaller the value, the higher the coding efficiency, thus improving the coding efficiency.
  • Figure 8 is a schematic illustration of an encoding process in accordance with one embodiment of the present invention.
  • each image block is called a Coding Tree Unit (CTU).
  • the CTU can be further divided into a plurality of sub-blocks, each of which is a Coding Unit (CU) that can independently determine its coding parameters.
  • the division of the CTU adopts the structure of a quadtree.
  • each CU will independently select its optimal coding parameters, for example, prediction mode selection, division of a prediction unit (PU), and the like.
  • the prediction 810 Perform prediction on the encoder to obtain a predicted value of the predicted block, and subtract the predicted value of the predicted block from the original value to obtain a residual signal of the predicted block.
  • the prediction can be intra prediction.
  • the original value refers to the original pixel value of the image block to be encoded, including the target image block.
  • the residual signal of the target image block is transformed and quantized to obtain an initial transformed quantized coefficient of the target image block, and the initial transformed quantized coefficient includes a DC component and an AC component.
  • the target image block of this embodiment may be a transform block or correspond to a transform block.
  • a prediction block can be divided into multiple transform blocks, and transform blocks belonging to the same prediction block use the same prediction mode.
  • Embodiments of the present invention may transform and quantize residual signals in units of transform blocks. The residual signal of each transform block can be obtained from the residual signal of the prediction block.
  • 810 to 830 are similar to conventional video coding, and are not described herein again.
  • inverse transform and inverse transform are performed on the transform transform quantized coefficients to obtain a transition residual, and the transition residual is added to the predicted value to obtain a transition reconstructed image block of the target image block, wherein the transition transform quantized coefficient includes the second DC component and 820 is obtained.
  • the AC component, the second DC component is a preset value (for example, the preset value is 0).
  • the DC component of the transition transform quantized coefficient is set to zero, that is, the transform quantized coefficient includes an AC component parsed from the code stream and a DC component having a value of 0, and the transition transform quantized coefficient is performed.
  • Inverse quantization and inverse transform the reconstruction residual of the target image block when the DC component is zero is obtained, and the reconstructed residual is added to the predicted value of the target image block to obtain a reconstructed value, thereby obtaining the target image block when the DC component is zero.
  • the predicted value of the DC component is determined according to the similarity of the pixels of the reconstructed image block to the pixels in the reference pixel region.
  • a specific method of calculating the predicted value DC pred of the DC component of the target image block refer to the description in the embodiment of FIG. 14, and details are not described herein again.
  • the DC component of the target image block is subtracted from the predicted value of the DC component to obtain a residual ⁇ DC of the DC component.
  • the reconstructed transform quantized coefficients are written to the code stream, wherein the reconstructed transform quantized coefficients include the difference between the DC component and the predicted value and the AC component obtained by 820.
  • the reconstructed transform quantized coefficients may be entropy encoded, and the encoded reconstructed transform quantized coefficients may be written into the code stream.
  • information such as the division manner of the CU in the CTU, the coding mode of the CU, the division mode of the PU in the CU, and the prediction mode selection are also subjected to entropy coding to the code stream.
  • the transform quantized coefficients may be inverse quantized and inverse transformed with the DC component set to a preset value to obtain a transition reconstructed image block of the target image block, and the pixels of the image block are reconstructed according to the transition.
  • the similarity of the pixels in the reference pixel region of the target image block determines the predicted value of the DC component.
  • the DC component residual can be used instead of the DC component. Since the absolute value of the DC component residual is smaller than the DC component, the coding efficiency is improved.
  • the embodiment of the present invention uses the similarity between the pixel of the transition reconstructed image block and the pixel in the reference pixel region with the DC component as a preset value. Determining the predicted value of the DC component improves the accuracy of the DC component prediction value, so that the DC component residual is smaller, and therefore, the coding efficiency is higher.
  • FIG. 9 is a schematic diagram of a decoding process in accordance with another embodiment of the present invention.
  • the processing method of the decoding end corresponds to the encoding end, and after the decoding end obtains a CTU code stream, the entropy decoding obtains the division manner of the CU in the CTU and the division mode, the prediction mode, and the transformation quantization coefficient of the PU in each CU, and then Decode in order. After all CUs in the CTU are decoded, the subsequent CTUs can continue to be decoded.
  • reconstructed transform quantized coefficients are obtained from a code stream, wherein the reconstructed transform quantized coefficients include a difference between a first DC component of the target image block and a predicted value of the DC component and an AC component of the target image block.
  • the decoding end can receive the code stream sent by the encoding end, and perform entropy decoding on the code stream to obtain prediction mode information and transform quantization coefficients of the target image block.
  • the prediction mode information may include a prediction mode of each prediction block, for example, in HEVC, the prediction mode may be one of a DC prediction mode, a planar prediction mode, and 33 angle (or direction) prediction modes.
  • the transform quantized coefficients include ⁇ DC and AC components.
  • a plurality of transform blocks may be divided correspondingly, and transform blocks belonging to the same prediction block adopt the same prediction mode.
  • the target image block of this embodiment may be a transform block or correspond to a transform block.
  • the transition transform quantized coefficients are inverse quantized and inverse transformed to obtain a transition reconstructed image block of the target image block, wherein the transition transform quantized coefficients include a second DC component and an AC component, and the second DC component is a preset value.
  • the DC component of the transform quantized coefficient is set to zero, that is, the transform quantized coefficient includes an AC component parsed from the code stream and a DC component having a value of 0, and inversely quantizes the transform quantized coefficient.
  • the predicted value of the DC component is determined according to the similarity of the pixels of the transition reconstructed image block to the pixels in the reference pixel region of the target image block.
  • a specific method of calculating the predicted value DC pred of the DC component of the target image block is described in the following description of the embodiment of FIG.
  • the DC component is determined based on the difference between the predicted values of the DC component and the DC component and the predicted value of the DC component.
  • the predicted value DC pred of the DC component obtained in the above step and the ⁇ DC parsed from the code stream may be added to obtain a quantized DC component value of the target image block.
  • the initial transform quantized coefficients are inverse quantized and inverse transformed to obtain a residual signal of the target image block, wherein the initial transform quantized coefficients include a DC component and an AC component.
  • the DC component value of the target image block obtained in the above step is inversely quantized and inverse-transformed with the AC component value of the target image block parsed from the code stream, thereby obtaining a reconstructed residual signal of the target image block.
  • a reconstructed image block is generated based on the predicted value and the reconstructed residual signal.
  • the transform quantized coefficients may be inverse quantized and inverse transformed with the DC component set to a preset value to obtain a transition reconstructed image block of the target image block, and the pixels of the image block are reconstructed according to the transition.
  • the similarity of the pixels in the reference pixel region of the target image block determines the predicted value of the DC component.
  • a DC component residual can be used instead of the DC component. Since the absolute value of the DC component residual is smaller than the DC component, the coding efficiency is improved.
  • the embodiment of the present invention uses the similarity between the pixel of the transition reconstructed image block and the pixel in the reference pixel region with the DC component as a preset value. Determining the predicted value of the DC component improves the accuracy of the DC component prediction value, so that the DC component residual is smaller, and therefore, the coding efficiency is higher.
  • a method of determining a predicted value of a DC component of a target image block will be described in detail below with reference to FIGS. 10 through 13B.
  • Figure 10 shows the 35 prediction modes of HEVC.
  • the intra prediction modes in HEVC include a direct current (DC) prediction mode (index 1), a Planar prediction mode (index 0), and an angle prediction mode (index 2 to 34).
  • DC direct current
  • Planar prediction mode index 0
  • angle prediction mode index 2 to 34
  • the current image block is predicted using the information of the reconstructed block of its left and upper reference pixel regions.
  • the DC prediction mode is adopted, the average value of the reference pixels is used as the predicted value for all the pixels in the current image block.
  • the planar prediction mode the reference pixels are used for bilinear interpolation to obtain the predicted values of all pixels in the current image block.
  • the pixel prediction in the current image block is projected to the reference pixel region along a specific angle by using the texture of the current image block and the texture height of the adjacent reconstructed block, and the reference pixel region corresponds to
  • the 1/32 precision pixel value is used as the predicted value of the pixel in the current image block.
  • the pixel value of the reference pixel area 1/32 precision can be interpolated using two adjacent reference pixels. If the corresponding reference pixel of the projection is exactly the entire pixel of the reference area, the corresponding reference pixel value can be directly copied.
  • the embodiment of the present invention may select the pixel signal according to the direction specified by the prediction mode, and determine the predicted value of the DC component according to the texture correlation of the pixel signal. Three methods of determining the predicted value of the DC component are described in detail below.
  • FIG. 14 is a schematic flow chart of a process of determining a predicted value of a DC component, in accordance with an embodiment of the present invention.
  • 850 of FIG. 8 and step 930 of FIG. 9 may include the following steps:
  • the intra prediction mode pred mode of the target image block is obtained from the code stream, and four phases are respectively taken in the reference pixel region around the transition reconstructed image block and the transition reconstructed image block along the direction specified by the intra prediction mode.
  • Adjacent pixel signal Adjacent pixel signal.
  • the kth line is not a real existing line, but is convenient for describing an imaginary ray that passes through the four adjacent pixels and is consistent with the direction indicated by the intra prediction mode, and The same direction in the intra prediction mode means that the directions are substantially parallel or parallel. It should be understood that embodiments of the present invention may refer to a line passing through a pixel when the line passes through, and may also refer to a virtual pixel that the line is interpolated through the actual pixel.
  • FIG. 11A is a schematic diagram of selecting a pixel signal based on a direction prediction mode according to an embodiment of the present invention.
  • FIG. 11B is a schematic diagram of selecting a pixel signal based on a DC prediction mode and a planar prediction mode.
  • a set of pixel signals is selected in the reference pixel region on the kth line, including: Four signals, selecting a set of pixel signals in the transition reconstructed image block, including: Four signals.
  • the subscript i k represents the kth line along the prediction direction
  • j represents the jth pixel signal on the line.
  • Adjacent several other signals are arranged in order of number.
  • the pixel signal is selected in accordance with the prediction direction; and for the DC prediction mode and the planar prediction mode, only the correlation of adjacent pixels in the horizontal direction and the vertical direction can be considered, and thus the pixel signals are selected in the horizontal direction and the vertical direction.
  • the direction specified by the direction prediction mode may refer to a prediction direction
  • the direction specified by the DC prediction mode and the plane prediction mode may refer to a horizontal direction and a vertical direction
  • k is an integer between 1 and 2n-1.
  • the embodiment of the present invention is described by taking four signals in each of the reference pixel region and the transition reconstructed image block as an example.
  • the embodiment of the present invention is not limited thereto, and the number of selected signals may be Is other values, for example, values greater than 4 or equal to 3.
  • the second order of the selected two sets of pixel signals on each line is calculated as follows:
  • the i kth line has strong texture characteristics in the prediction direction according to the size of the two steps, and when the two step values are small (for example, less than a certain preset value), the signal is in the direction. It is consistent, that is, it has strong texture characteristics.
  • This embodiment is described by taking the above method of obtaining two steps as an example, but the embodiment of the present invention is not limited thereto, and any other method for obtaining two steps can be used in the embodiment of the present invention.
  • the following formula (2) is used to determine whether the i kth line has strong texture characteristics in the prediction direction.
  • the following formula (3) is used.
  • the embodiment of the present invention determines the texture characteristics by using two steps.
  • the embodiment of the present invention is not limited thereto, and other methods may be used to determine whether the selected signal on each line has strong texture characteristics.
  • a step or the like may be used to determine whether the selected signal on each line has strong texture characteristics.
  • the way to determine the texture characteristics can also be a combination of multiple ways. For example, for some signals selected on the line, two steps are used, and for other signals selected, a step is used.
  • the prediction modes in other directions are weighted averaged by two adjacent pixels to obtain the above pixel signals.
  • the two sets of pixel signals may be signals of the pixels themselves passing through the line, or may be signals obtained by interpolating signals of pixels on both sides of the line.
  • 12A-12D are schematic diagrams of selecting pixel signals based on a direction prediction mode, in accordance with an embodiment of the present invention.
  • a block of the left and lower left corners of the transition reconstructed image block may be used as a reference, and a ray is taken along the prediction direction (indicating the prediction direction in FIG. 10) starting from x i,1
  • the interpolated pixel point y' i,j is obtained by weighted averaging of two adjacent original pixel points.
  • Prediction modes 4-9 are similar and will not be described here.
  • a block of the left and upper left corners of the transition reconstructed image block is used as a reference, and a ray is taken along the prediction direction (the prediction direction indicated in FIG. 10) starting from x i,1 .
  • the interpolated pixel point y' i,j is obtained by weighted averaging of two adjacent original pixel points.
  • Prediction mode 12-17 is similar, and will not be described here.
  • the block on the upper side and the upper left corner of the transition reconstructed image block is used as a reference, and a ray is made along the prediction direction starting from x 1, i , and the pixel point y' i, j is interpolated . It is obtained by weighted averaging of two adjacent original pixel points. Modes 20-25 are similar, and will not be described here.
  • a block of the upper and upper right corners of the transition reconstructed image block may be used as a reference, and a ray is made along the prediction direction starting from x 1, i , and the pixel point y' i is interpolated.
  • j is obtained by weighted averaging of two adjacent original pixel points.
  • Modes 28-33 are similar and will not be described here.
  • interpolation point is not limited to the point between the upper pixel point and the lower side pixel point on the ray, and the point between the left side pixel point and the right side pixel point on the ray may be selected. Accordingly, interpolation can be performed according to the pixel point on the left side and the pixel point on the right side to obtain y' i,j .
  • the interior of the transition reconstructed image block is selected. Adjacent pixels; if 4 adjacent pixels cannot be found, only the pixels of the reference block or reference pixel area are selected.
  • the signal inside the transition reconstructed image block is increased by an offset ⁇ x to obtain a reconstructed signal having a DC component.
  • the offset represents the DC predicted value before quantization of the reconstructed image block.
  • q is the number of the pixel signal on each line in the above set C.
  • the two-step degree can more accurately reflect the correlation or similarity between the pixel of the target image block and the reference image block, so that the DC component residual is made. It is smaller, which further improves the coding efficiency.
  • the two pixel signals adjacent to the transition reconstructed image block and the reference region may be directly selected according to the direction specified by the prediction mode to form a plurality of pixel pairs, and according to formula (13) And formula (14) to determine the DC component prediction value.
  • the two-step degree can more accurately reflect the correlation or similarity between the pixel of the target image block and the reference image block, the accuracy of the predicted value of the DC component obtained based on the two-step degree is higher, so that the DC component residual is made. It is smaller, which further improves the coding efficiency.
  • the accuracy of the predicted value of the DC component is further improved, so that the DC component residual is smaller, thereby further improving. Coding efficiency.
  • 15 is a schematic flow chart of a process of determining a predicted value of a DC component, in accordance with another embodiment of the present invention.
  • step 850 of FIG. 8 and step 930 of FIG. 9 may include the following steps:
  • the transition reconstructed image block takes a signal Reference area signal
  • the pixel signal can be selected according to the prediction direction; and for the DC prediction plane prediction mode, only the correlation of the pixels in the vertical and horizontal directions can be considered, that is, the pixel signals are selected in the horizontal direction and the vertical direction.
  • the pixel signal can be selected according to the prediction direction; and for the DC prediction plane prediction mode, only the correlation of the pixels in the vertical and horizontal directions can be considered, that is, the pixel signals are selected in the horizontal direction and the vertical direction. Similar to the process when the above set C is not empty, for the signal in the transition reconstructed image block Now add an offset ⁇ x.
  • FIG. 14 can be combined with the embodiment of FIG.
  • the line having the texture direction characteristic is not found by 1420 of FIG. 14, that is, the set C is empty
  • the two pixels adjacent to the reference area of the transition reconstructed image block are selected according to the scheme of FIG. 15 to determine the predicted value of the DC component. .
  • 16 is a schematic flow chart of a process of determining a predicted value of a DC component, in accordance with another embodiment of the present invention.
  • step 850 of FIG. 8 and step 930 of FIG. 9 may include the following steps:
  • M and H may be greater than or equal to 2 and less than the size N of the transition reconstructed image block.
  • M and H can be equal or unequal.
  • the predicted value of the DC component after the target image block is quantized m is the number of target image blocks.
  • the embodiments of the present invention can be applied to various electronic devices. Illustratively, an example in which the embodiments of the present invention are applied to a television device and a mobile phone device is given below.
  • the computational complexity is reduced in the case of ensuring the accuracy of the DC component prediction value.
  • the average value is used to calculate the predicted value of the DC component, making the design of the encoder or decoder simple.
  • M and H are multiple rows or columns, it is possible to make full use of the correlation of more signals around the target image block and the relationship between the target image block and the surrounding signals to predict the predicted value of the DC component, so that the DC component residual is more Small, the accuracy of the predicted value of the DC component is improved, thereby further improving the coding efficiency.
  • the method further includes: before performing the DC prediction, first determining the size of the DC value of the coded image, and if the DC value is greater than the preset threshold, performing the foregoing method of the present invention, if the DC value is less than a preset threshold, Then, the conventional method is used for processing, that is, the DC prediction method is not used, but the target image is directly
  • the present invention can add a syntax element, such as a flag, to the decoder to obtain the DC component and the AC component of the target image block, such as DC_pred_present, in a parameter such as slice header, PPS, SPS, and the like.
  • Flag when it is 1, indicates that the DC is acquired in a predictive manner. When it is zero, it indicates that the DC is obtained by directly parsing the code stream, or vice versa.
  • FIG. 17 is a schematic block diagram of an embodiment of the present invention suitable for use in a television set.
  • the television device 1700 includes an antenna 1701, a tuner 1702, a demultiplexer 1703, a decoder 1704, a video signal processor 1705, a display unit 1706, an audio signal processor 1707, a speaker 1708, an external interface 1709, a controller 1710, User interface 1711 and bus 1712.
  • the tuner 1702 extracts a signal of a desired channel from a broadcast signal received via the antenna 1701, and demodulates the extracted signal.
  • the tuner 1702 then outputs the encoded bit stream obtained by the demodulation to the demultiplexer 1703. That is, the tuner 1702 functions as a transmitting device in the television device 1700 that receives the encoded stream of the encoded image.
  • the demultiplexer 1703 separates the video stream and the audio stream of the program to be viewed from the encoded bit stream, and outputs the separated stream to the decoder 1704.
  • the demultiplexer 1703 also extracts auxiliary data, such as an electronic program guide, from the encoded bitstream, and provides the extracted data to the controller 1710. If the encoded bit stream is scrambled, the demultiplexer 703 can descramble the encoded bit stream.
  • the decoder 1704 decodes the video stream and the audio stream input from the demultiplexer 1703. The decoder 1704 then outputs the video data generated by the decoding to the video signal processor 1705. The decoder 1704 also outputs the audio data generated by the decoding to the audio signal processor 1707.
  • the video signal processor 1705 reproduces the video data input from the decoder 1704 and displays the video data on the display unit 1706.
  • the video signal processor 1705 can also display an application screen provided via the network on the display unit 1706. Additionally, video signal processor 1705 can perform additional processing, such as noise removal, on the video data in accordance with the settings.
  • the video signal processor 1705 can also generate an image of a GUI (Graphical User Interface) and superimpose the generated image on the output image.
  • GUI Graphic User Interface
  • the display unit 1706 is driven by a drive signal supplied from the video signal processor 1705, and displays a video or image on a video screen of a display device such as a liquid crystal display, a plasma display, or an OELD (with an organic electroluminescence display).
  • a display device such as a liquid crystal display, a plasma display, or an OELD (with an organic electroluminescence display).
  • the audio signal processor 1707 performs reproduction processing, for example, digital-to-analog conversion and amplification, on the audio data input from the decoder 1704, and outputs audio through the speaker 1708. Additionally, the audio signal processor 1707 can perform additional processing on the audio data, such as noise removal.
  • the external interface 1709 is an interface for connecting the television device 1700 with an external device or network.
  • a video stream or audio stream received via external interface 1709 can be decoded by decoder 1704. That is, the external interface 1709 is also used as a transmitting device in the television device 1700 that receives the encoded stream of the encoded image.
  • the controller 1710 includes a processor and a memory.
  • the memory stores programs to be executed by the processor, program data, auxiliary data, data acquired via the network, and the like. For example, when the television device 900 is booted, the program stored in the memory is read and executed by the processor.
  • the processor controls the operation of the television device 1700 based on control signals input from the user interface 1711.
  • User interface 1711 is coupled to controller 1710.
  • the user interface 1711 includes buttons and switches for the user to operate the television device 1700 and a receiving unit for receiving remote control signals.
  • the user interface 1711 detects an operation performed by the user via these components, generates a control signal, and outputs the generated control signal to the controller 1710.
  • the bus 1712 includes a tuner 1702, a demultiplexer 1703, a decoder 1704, a video signal processor 1705, The audio signal processor 1707, the external interface 1709, and the controller 1710 are connected to each other.
  • the decoder 1704 has the functions of the video decoding device according to the above embodiment.
  • FIG 18 is a schematic block diagram of an embodiment of the present invention suitable for use in a mobile phone application.
  • the mobile telephone device 1720 includes an antenna 1721, a communication unit 1722, an audio codec 1723, a speaker 1724, a microphone 1725, a camera unit 1726, an image processor 1727, a demultiplexer 1728, a recording/reproducing unit 1729, and a display unit 1730.
  • the antenna 1721 is connected to the communication unit 1722.
  • the speaker 1724 and the microphone 1725 are connected to an audio codec 1723.
  • the operation unit 932 is connected to the controller 1731.
  • the bus 1733 connects the communication unit 1722, the audio codec 1723, the camera unit 1726, the image processor 1727, the demultiplexer 1728, the recording/reproducing unit 1729, the display unit 1730, and the controller 1731 to each other.
  • the mobile telephone device 1720 performs operations in various operational modes, such as transmission/reception of audio signals, transmission/reception of email and image data, photographing of images, recording of data, etc., the various modes of operation including voice calls Mode, data communication mode, imaging mode, and video telephony mode.
  • the analog audio signal generated by the microphone 1725 is supplied to the audio codec 1723.
  • the audio codec 1723 converts the analog audio signal into audio data, performs analog-to-digital conversion on the converted audio data, and compresses the audio data.
  • the audio codec 1723 then outputs the audio data obtained as a result of the compression to the communication unit 1722.
  • Communication unit 1722 encodes and modulates the audio data to produce a signal to be transmitted.
  • the communication unit 1722 then transmits the generated signal to be transmitted to the base station via the antenna 1721.
  • the communication unit 1722 also amplifies the radio signal received via the antenna 1721 and performs frequency conversion on the radio signal received via the antenna 1721 to obtain the received signal.
  • the communication unit 1722 then demodulates and decodes the received signal to generate audio data, and outputs the generated audio data to the audio codec 1723.
  • the audio codec 1723 decompresses the audio data and performs digital to analog conversion on the audio data to generate an analog audio signal.
  • the audio codec 1723 then supplies the generated audio signal to the speaker 1724 to output audio from the speaker 1724.
  • the controller 1731 In the data communication mode, for example, the controller 1731 generates text data to be included in the email in accordance with an operation by the user via the operation unit 1732.
  • the controller 1731 also displays text on the display unit 1730.
  • the controller 1731 also generates email data in response to an instruction from the user for transmission via the operation unit 1732, and outputs the generated email data to the communication unit 1722.
  • the communication unit 17922 encodes and modulates the email data to generate a signal to be transmitted.
  • the communication unit 1722 then transmits the generated signal to be transmitted to the base station via the antenna 1721.
  • the communication unit 1722 also amplifies the radio signal received via the antenna 1721 and performs frequency conversion on the radio signal received via the antenna 1721 to obtain the received signal.
  • the communication unit 1722 then demodulates and decodes the received signal to recover the email data, and outputs the restored email data to the controller 1731.
  • the controller 1731 displays the content of the email on the display unit 1730, and stores the email data in the storage medium of the recording/reproducing unit 1729.
  • the recording/reproducing unit 1729 includes a readable/writable storage medium.
  • the storage medium may be an internal storage medium, or may be an externally mounted storage medium such as a hard disk, a magnetic disk, a magneto-optical disk, a USB (Universal Serial Bus) memory, or a memory card.
  • USB Universal Serial Bus
  • the camera unit 1726 images the subject to generate image data, and outputs the generated image data to the image processor 1727.
  • the image processor 1727 encodes the image data input from the camera unit 1726, and stores the encoded stream in the storage medium of the storage/reproduction unit 1729.
  • the demultiplexer 1728 In the video telephony mode, the demultiplexer 1728 multiplexes the video stream encoded by the image processor 1727 and the audio stream input from the audio codec 1723, and outputs the multiplexed stream to the communication unit 1722. .
  • Communication unit 1722 encodes and modulates the multiplexed stream to produce a signal to be transmitted.
  • the communication unit 1722 then transmits the generated signal to be transmitted to the base station via the antenna 1721.
  • the communication unit 1722 also amplifies the radio signal received via the antenna 1721 and performs frequency conversion on the radio signal received via the antenna 1721 to obtain the received signal.
  • the signal to be transmitted and the received signal may comprise an encoded bit stream.
  • the communication unit 1722 then demodulates and decodes the received signal to recover the stream, and outputs the recovered stream to the demultiplexer 1728.
  • the demultiplexer 1728 separates the video stream and the audio stream from the input stream, outputs the video stream to the image processor 1727, and outputs the audio stream to the audio codec 1723.
  • Image processor 1727 decodes the video stream to produce video data.
  • the video data is supplied to the display unit 1730, and a series of images are displayed by the display unit 1730.
  • the audio codec 1723 decompresses the audio stream and performs digital to analog conversion on the audio stream to produce an analog audio signal.
  • the audio codec 1723 then supplies the generated audio signal to the speaker 1724 to output audio from the speaker 1724.
  • the image processor 1727 has the functions of the video encoding device and the video decoding device according to the above-described embodiments.
  • the functions described may be implemented in hardware, software, firmware, or any combination thereof. If implemented in software, the functions may be stored on or transmitted as one or more instructions or code via a computer-readable medium and executed by a hardware-based processing unit.
  • the computer readable medium can comprise a computer readable storage medium (which corresponds to a tangible medium such as a data storage medium) or a communication medium comprising, for example, any medium that facilitates transfer of the computer program from one place to another in accordance with a communication protocol. .
  • computer readable media generally may correspond to (1) a non-transitory tangible computer readable storage medium, or (2) a communication medium such as a signal or carrier wave.
  • Data storage media may be any available media that can be accessed by one or more computers or one or more processors to retrieve instructions, code and/or data structures for use in carrying out the techniques described herein.
  • the computer program product can comprise a computer readable medium.
  • certain computer-readable storage media may comprise RAM, ROM, EEPROM, CD-ROM or other optical disk storage, disk storage or other magnetic storage device, flash memory, or may be used to store instructions or data structures. Any other medium in the form of the desired program code and accessible by the computer. Also, any connection is properly termed a computer-readable medium. For example, if you use coaxial cable, fiber optic cable, twisted pair, digital subscriber line (DSL), or wireless technology (eg, infrared, radio, and microwave) to send commands from a website, server, or other remote source, coaxial cable , fiber optic cable, twisted pair, DSL, or wireless technologies (eg, infrared, radio, and microwave) are included in the definition of the media.
  • coaxial cable, fiber optic cable, twisted pair, digital subscriber line (DSL), or wireless technology eg, infrared, radio, and microwave
  • a magnetic disk and an optical disk include a compact disk (CD), a laser disk, an optical disk, a digital video disk (DVD), a flexible disk, and a Blu-ray disk, wherein the disk usually reproduces data magnetically, and the disk passes the laser Optically copy data. Combinations of the above should also be included within the scope of computer readable media.
  • processors such as one or more digital signal processors (DSPs), general purpose microprocessors, application specific integrated circuits (ASICs), field programmable logic arrays (FPGAs), or other equivalent integrated or discrete logic circuits
  • DSPs digital signal processors
  • ASICs application specific integrated circuits
  • FPGAs field programmable logic arrays
  • processors may refer to any of the foregoing structures or any other structure suitable for implementing the techniques described herein.
  • the functionality described herein may be provided within dedicated hardware and/or software modules configured for encoding and decoding, or incorporated in a combined codec.
  • the techniques can be fully implemented in one or more circuits or logic elements.
  • the techniques of the present invention can be broadly implemented by a variety of devices or devices, including a wireless handset, an integrated circuit (IC), or a collection of ICs (eg, a chipset).
  • IC integrated circuit
  • Various components, modules or units are described in this disclosure to emphasize functional aspects of the apparatus configured to perform the disclosed techniques, but are not necessarily required to be implemented by different hardware units. Rather, as described above, various units may be combined in a codec hardware unit or combined with suitable software and/or by a collection of interoperable hardware units (including one or more processors as described above). Or firmware to provide.
  • system and “network” are used interchangeably herein. It should be understood that the term “and/or” herein is merely an association relationship describing an associated object, indicating that there may be three relationships, for example, A and/or B, which may indicate that A exists separately, and A and B exist simultaneously. There are three cases of B alone. In addition, the character "/" in this article generally indicates that the contextual object is an "or" relationship.
  • B corresponding to A means that B is associated with A, and B can be determined from A.
  • determining B from A does not mean that B is only determined based on A, and that B can also be determined based on A and/or other information.
  • the disclosed systems, devices, and methods may be implemented in other manners.
  • the device embodiments described above are merely illustrative.
  • the division of the unit is only a logical function division.
  • there may be another division manner for example, multiple units or components may be combined or Can be integrated into another system, or some features can be ignored or not executed.
  • the mutual coupling or direct coupling or communication connection shown or discussed may be an indirect coupling or communication connection through some interface, device or unit, and may be in an electrical, mechanical or other form.
  • the units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, may be located in one place, or may be distributed to multiple network units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution of the embodiment.
  • each functional unit in each embodiment of the present invention may be integrated into one processing unit, or each unit may exist physically separately, or two or more units may be integrated into one unit.
  • the disclosed systems, devices, and methods may be implemented in other manners.
  • the device embodiments described above are merely illustrative.
  • the division of the unit is only a logical function division.
  • there may be another division manner for example, multiple units or components may be combined or Can be integrated into another system, or some features can be ignored or not executed.
  • the mutual coupling or direct coupling or communication connection shown or discussed may be an indirect coupling or communication connection through some interface, device or unit, and may be in an electrical, mechanical or other form.
  • the units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, may be located in one place, or may be distributed to multiple network units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution of the embodiment.
  • each functional unit in each embodiment of the present invention may be integrated into one processing unit, or each unit may exist physically separately, or two or more units may be integrated into one unit.
  • the above integrated unit can be implemented in the form of hardware or in the form of a software functional unit.
  • the integrated unit if implemented in the form of a software functional unit and sold or used as a standalone product, may be stored in a computer readable storage medium.
  • the technical solution of the present invention which is essential or contributes to the prior art, or all or part of the technical solution, may be embodied in the form of a software product stored in a storage medium.
  • a number of instructions are included to cause a computer device (which may be a personal computer, server, or network device, etc.) to perform all or part of the steps of the methods described in various embodiments of the present invention.
  • the foregoing storage medium includes: a U disk, a mobile hard disk, a read-only memory (ROM), a random access memory (RAM), a magnetic disk, or an optical disk, and the like. .

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)
  • Compression, Expansion, Code Conversion, And Decoders (AREA)

Abstract

一种解码方法、编码方法、解码设备和编码设备。解码方法包括:从码流中获取目标图像块的变换量化系数的交流AC分量及直流DC分量残差;对AC分量以及预设的DC分量进行逆量化和逆变换,得到目标图像块的过渡重建图像块;根据过渡重建图像块与目标图像块的参考像素区域中的像素的相似性,确定目标图像块的变换量化系数的DC分量的预测值;根据DC分量的预测值和DC分量残差确定目标图像块的原始DC分量;对目标图像块的原始DC分量以及AC分量进行逆量化和逆变换,得到目标图像块的残差信号,并根据所残差信号对目标图像块进行解码。

Description

解码方法、编码方法、解码设备和编码设备 技术领域
本发明实施例涉及视频编解码及压缩领域,更具体地,涉及一种解码方法、编码方法、解码设备和编码设备。
背景技术
数字视频能力可并入到大范围的装置中,包含数字电视、数字直播系统、无线广播系统、个人数字助理(PDA)、膝上型或桌上型计算机、平板计算机、电子书阅读器、数码相机、数字记录装置、数字媒体播放器、视频游戏装置、视频游戏控制台、蜂窝式或卫星无线电电话、视频会议装置、视频流装置等等。数字视频装置实施视频压缩技术,例如由MPEG-2、MPEG-4、ITU-TH.263、ITU-TH.264/MPEG-4第10部分高级视频编解码(AVC)、ITU-TH.265高效率视频编解码(HEVC)标准定义的标准和所述标准的扩展部分中所描述的那些视频压缩技术,从而更高效地发射及接收数字视频信息。视频装置可通过实施这些视频编解码技术来更高效地发射、接收、编码、解码和/或存储数字视频信息。
在视频编解码领域中,帧的概念是指一副完整的图像,由一帧一帧的图像按照一定的次序和帧速率组成视频格式后即可播放。当帧速率达到一定的速度后,两帧之间的间隔时间小于人眼的分辨极限,会出现短暂的视觉停留,方才能看似动态的出现在屏幕上。视频文件能够实现压缩的基础是单帧数字图像的压缩编码,数字化后的图像中存在很多的重复表示信息,称之为冗余信息。一帧图像中往往存有许多空间结构相同或相似的地方,例如同一物体或背景中的采样点颜色之间大都存在着紧密关联性和相似性。在多帧图像组中,一帧图像和其前一帧或后一帧基本上都有巨大的相关性,描述信息的像素值差别很小,这些都是可以被压缩的部分。同样的道理,视频文件中不但存在着空间冗余信息,而且包含大量的时间冗余信息,这是由视频的组成结构导致的。例如,视频采样的帧速率一般为25帧/秒至30帧/秒,特殊情况中会出现60帧/秒的可能。也就是说,相邻两帧间的采样时间间隔至少为1/30秒到1/25秒。在这么短的时间内,采样得到的图像画面中基本上都存在大量的相似信息,画面之间存在巨大关联性。但是在原始的数字视频录入系统中是各自独立的记录,没有考虑和利用到这些连贯相似特性,这就造成了相当巨大的重复多余的数据量。另外,已通过研究表明,从人眼的视觉敏感度这一心理特性的角度出发,视频信息中也是存在可以用来压缩的部分,即视觉冗余。所谓视觉冗余,是指利用人眼对亮度变化比较敏感,而对色度的变化相对不敏感的特性来适当的压缩视频比特流。在高亮度的区域,人眼视觉对亮度变化的敏感度呈现下降趋势,转而对物体的边缘处较为敏感,内部区域相对不敏感;对整体结构较为敏感,对内部细节变换相对不敏感。因为视频图像信息的最终服务对象是我们人类群体,所以可以充分利用人眼的这些特性对原有的视频图像信息进行压缩处理,达到更佳的压缩效果。除了上面提到的空间冗余、时间冗余和视觉冗余外,视频图像信息中还会存在信息熵冗余、结构冗余、知识冗余、重要性冗余等等这一系列的冗余信息。视频压缩编码的目的就是使用各种技术方法将视频序列中的冗余信息去除掉,以达到减小存储空间和节省传输带宽的效果。
就目前的技术发展现状而言,视频压缩处理技术主要包括帧内预测、帧间预测、变换量化、熵编码以及消块滤波处理等。在国际通用范围内,存在的视频压缩编码标准中 主流的压缩编码方式主要有四种:色度抽样、预测编码、变换编码和量化编码。
色度抽样:此方式充分利用了人眼的视觉心理特点,从底层的数据表示中就开始设法最大限度的缩减单个元素描述的数据量。在电视系统中多数采用的是亮度-色度-色度(YUV)颜色编码,它是欧洲电视系统广泛采用的标准。YUV颜色空间中包括一个亮度信号Y和两个色差信号U和V,三个分量彼此独立。YUV颜色模式彼此分开的表示方式更加灵活,传输占用带宽少,比传统的红绿蓝(RGB)色彩模型更具优势。例如,YUV 4:2:0形式表示两色度分量U和V在水平方向和垂直方向上都只有亮度Y分量的一半,即4个采样像素点中存在4个亮度分量Y,而色度分量U和V则只有一个。这样表示时,数据量进一步缩小,仅为原始的33%左右。利用人眼生理视觉特性,通过这种色度抽样的方式实现视频压缩的目的,是目前广泛采用的视频数据压缩方式之一。
预测编码:即利用之前已编码帧的数据信息来预测当前将要编码的帧。通过预测得到一个预测值,它不完全等同与实际值,之间存在着一定的残差值。如果预测越适合,则预测值就会越接近实际值,残差值就越小,这样对残差值进行编码就能大大减小数据量,在解码端解码时运用残差值加上预测值还原重构出初始图像,这就是预测编码的基本思想方法。在主流编码标准中预测编码分为帧内预测和帧间预测两种基本类型。
变换编码:是不直接对原本的空间域信息进行编码,而是将信息采样值按照某种形式的变换函数,从当前域转换到另外一种人为定义域中(通常称为变换域),再根据信息在变换域的分布特性进行压缩编码。变换编码的原因是:视频图像数据往往在空间域的数据相关性大,导致大量冗余信息的存在,直接编码需要很大的比特量。而在变换域中数据相关性则大大减少,使得编码的冗余信息减少,编码所需的数据量也随之大大减少,这样就能够得到较高的压缩比,而且可以实现较好的压缩效果。典型的变换编码有卡洛(K-L)变换、傅立叶变换等。整数离散余弦变换(DCT)是许多国际标准中都普遍采用的变换编码方式。
量化编码:上述提到的变换编码其实本身并不压缩数据,量化过程才是压缩数据的有力手段,也是有损压缩中数据“损失”的主要原因。量化的过程就是将动态范围较大的输入值强行规划成较少的输出值的过程。因为量化输入值范围较大,需要较多的比特数表示,而“强行规划”后的输出值范围较小,从而只需要少量的比特数即可表示。每个量化输入被归一化到一个量化输出,即量化到某个数量级中,这些数量级通常被称之为量化级(通常由编码器规定)。
在基于混合编码架构的编码算法中,上述压缩编码方式被混合使用,编码器控制模块根据视频帧中不同图像块的局部特性,选择该图像块所采用的编码模式。对帧内预测编码的块进行频域或空域预测,对帧间预测编码的块进行运动补偿预测,预测的残差再通过变换和量化处理形成残差系数(也可以称为变换量化系数),最后通过熵编码器生成最终的码流。为避免预测误差的累积,帧内或帧间预测的参考信号是通过编码端的解码模块得到。变换和量化后的残差系数经过逆量化和逆变换重建残差信号,再与预测的参考信号相加得到重建的图像。环路滤波会对重建后的图像进行像素修正,提高重建图像的编码质量。
然而,随着人们对高清视频的要求的提高,对图像的编码效率的要求也越来越高,因此,如何进一步提高编码效率成了亟待解决的问题。
发明内容
本申请提供了一种解码方法、编码方法、解码设备和编码设备,能够提高编码效率。
根据本发明的第一方面,提供了一种解码方法,包括:从码流中获取目标图像块的变换量化系数的AC分量及DC分量残差;对AC分量以及预设的DC分量进行逆量化和逆变换,得到过渡残差,并将过渡残差与目标图像块的预测值相加得到目标图像块的过渡重建图像块;根据过渡重建图像块与目标图像块的参考像素区域中的像素的相似性,确定目标图像块的变换量化系数的DC分量的预测值;根据DC分量的预测值和DC分量残差确定目标图像的原始DC分量;对目标图像的原始DC分量以及AC分量进行逆量化和逆变换,得到目标图像的残差信号,并根据所残差信号对目标图像进行解码。
例如,在解码端,可以从码流中获取重构变换量化系数,其中重构变换量化系数包括目标图像块的第一直流DC分量与DC分量的预测值的差值和目标图像块的交流AC分量;对过渡变换量化系数进行逆量化和逆变换,得到目标图像块的过渡重建图像块,其中过渡变换量化系数包括第二DC分量和AC分量,第二DC分量为预设值;根据过渡重建图像块的像素与目标图像块的参考像素区域中的像素的相似性,确定DC分量的预测值;根据DC分量与DC分量的预测值的差值(即DC分量残差)以及DC分量的预测值,确定DC分量(即原始DC分量);对初始变换量化系数进行逆量化和逆变换,得到目标图像块的残差信号,其中初始变换量化系数包括DC分量和AC分量。
根据本发明的第二方面,提供了一种编码方法,包括:对目标图像块的残差信号进行变换和量化,得到目标图像块的变换量化系数的直流DC分量和交流AC分量;对AC分量以及预设的DC分量进行逆量化和逆变换,得到过渡残差,并将过渡残差与目标图像块的预测值相加得到目标图像块的过渡重建图像块;根据过渡重建图像块的像素与目标图像块的参考像素区域中的像素的相似性,确定目标图像块的变换量化系数的DC分量的预测值;根据DC分量的预测值和目标图像块的原始DC分量确定目标图像的DC分量残差;将AC分量和DC分量残差写入码流。
例如,在编码端,可以对目标图像块的残差信号进行变换和量化,得到目标图像块的初始变换量化系数,初始变换量化系数包括第一直流DC分量和交流AC分量;对过渡变换量化系数进行逆量化和逆变换,得到目标图像块的过渡重建图像块,其中过渡变换量化系数包括第二DC分量和交流AC分量,第二DC分量为预设值;根据过渡重建图像块的像素与目标图像块的参考像素区域中的像素的相似性,确定DC分量(即原始DC分量)的预测值;将重构变换量化系数写入码流,其中重构变换量化系数包括DC分量与预测值的差值(即DC分量残差)和AC分量。
基于本申请的技术方案,可以在将DC分量设置为预设值的情况下对变换量化系数进行逆量化和逆变换,得到目标图像块的过渡重建图像块,并且根据过渡重建图像块的像素与目标图像块的参考像素区域中的像素的相似性,确定DC分量的预测值。在对目标图像块进行编码或解码时,可以使用DC分量残差来替代DC分量。由于DC分量残差的绝对值小于DC分量,因此提高了编码效率。
另外,本申请的技术方案根据DC分量为预设值的过渡重建图像块的像素与参考像素区域中的像素的相似性来确定DC分量的预测值,提高了DC分量预测值的精度,使得DC分量残差更小,因此,与直接使用参考像素区域的DC分量的预测值作为目标图像块的DC分量的预测值的方案相比,编码效率更高。
在某些实现方式中,上述根据过渡重建图像块与目标图像块的参考像素区域中的像素的相似性,确定目标图像块的变换量化系数的DC分量的预测值,包括:在目标图像块的预测模式指定的方向上,确定至少一条线以及对应于每条线的第一组像素信号和相邻的第二组像素信号,其中第一组像素信号包括过渡重建图像块的第一像素信号,第二组像素信号包括参考像素区域的第二像素信号,第一像素信号与第二像素信号相邻;求解第一像素信号的偏移量,使得至少一条线中第一像素信号增加偏移量后的重建信号的二阶梯度与第二像素信号的二阶梯度的平方和最小,其中过渡重建图像块中用于表示二阶梯度的像素信号增加偏移量,该偏移量用来表示DC分量的量化前的预测值;将求解得到的偏移量的值进行量化得到DC分量的预测值。由于二阶梯度能够更精确地体现目标图像块的像素与参考图像块之间的相关性或相似性,因此,基于二阶梯度得到的DC分量的预测值的精度更高,使得DC分量残差更小,从而进一步提高了编码效率。
在某些实现方式中,对应于至少一条线中的每条线的第一组像素信号和第二组像素信号满足如下两个公式和公式之一:
Figure PCTCN2017071602-appb-000001
其中,λ1用于表示第一组像素信号的重建信号存在二阶梯度时的阈值,λ2用于表示第一组像素信号的重建信号不存在二阶梯度时的阈值,
Figure PCTCN2017071602-appb-000002
为第一组像素信号的二阶梯度,
Figure PCTCN2017071602-appb-000003
为第二组像素信号的二阶梯度,ik为至少一条线的编号,j为至少一条线中的每条线上的像素信号的编号,
其中,偏移量根据如下公式来计算:
Figure PCTCN2017071602-appb-000004
其中,δx为偏移量,
Figure PCTCN2017071602-appb-000005
为第一像素信号的重建信号的二阶梯度,
Figure PCTCN2017071602-appb-000006
为第二像素信号的二阶梯度,C表示满足公两个公式之一线的集合,q为至少一条线中的每条线上的像素信号的编号。由于只选取相关性较强的某些线上的像素用于计算DC分量的预测值,因此,DC分量的预测值的精度得到进一步提高,使得DC分量残差更小,从而进一步提高了编码效率。
在某些实现方式中,
Figure PCTCN2017071602-appb-000007
在某些实现方式中,
Figure PCTCN2017071602-appb-000008
在某些实现方式中,上述根据过渡重建图像块与目标图像块的参考像素区域中的像素的相似性,确定目标图像块的变换量化系数的DC分量的预测值,包括:在目标图像块的预测模式指定的方向上,确定多个像素对以及对应于每个像素对的第一像素信号和相邻的第二像素信号,其中第一像素信号为过渡重建图像块的像素信号,第二像素信号为参考像素区域的像素信号;将第一像素信号增加一个偏移量,得到第一像素信号的重 建信号,其中该偏移量用来表示DC分量的量化前的预测值;求解偏移量,使得多个像素对的第一像素信号的重建信号与第二像素信号的一阶梯度的平方和最小;将求解得到的偏移量的值进行量化得到DC分量的预测值。
在某些实现方式中,DC分量的预测值根据下列公式来计算:
Figure PCTCN2017071602-appb-000009
其中δx为偏移量,n为过渡重建图像块的每一行或列的像素的个数,
Figure PCTCN2017071602-appb-000010
Figure PCTCN2017071602-appb-000011
为第一像素信号,
Figure PCTCN2017071602-appb-000012
为第二像素信号,
Figure PCTCN2017071602-appb-000013
Figure PCTCN2017071602-appb-000014
在预测模式指定的方向上相邻。由于只选取过渡重建图像块和参考像素区域之间的相邻像素用于计算DC分量的预测值,因此,在保证DC分量预测值的精度的情况下,降低了计算复杂度。
在某些实现中,上述第一组像素信号和第二组信号可以是上述至少一条线穿过的像素本身的信号,也可以是上述至少一条线周围的像素的信号进行插值得到的信号
在某些实现方式中,上述根据过渡重建图像块与目标图像块的参考像素区域中的像素的相似性,确定目标图像块的变换量化系数的DC分量的预测值,包括:获取参考像素区域中位于过渡重建图像块上方的第一组像素信号、过渡重建图像块左侧的第二组像素信号、过渡重建图像块内部上侧的第三组像素信号和过渡重建图像块内部左侧的第四组像素信号,第一组像素信号和第三组像素信号分别包括M行像素信号,第二组像素信号和第四组像素信号分别包括H列像素信号,M和H为正整数;计算第一组像素信号的平均值与第三组像素信号的平均值之差,得到第一差值;计算第二组像素信号的平均值与第四组像素信号的平均值之差,得到第二差值;将第一差值与第二差值的平均值进行量化得到DC分量的预测值。由于选取多行或多列过渡重建图像块和参考像素区域的像素用于计算DC分量的预测值,因此,在保证DC分量预测值的精度的情况下,降低了计算复杂度。另外,采用平均值来计算DC分量的预测值,使得编码器或解码器的设计简单。
在某些实现方式中,M为大于或等于2的整数,N为大于或等于2的整数。另外,当M和H为多行或多列时,能够充分利用目标图像块周围更多信号的相关性以及目标图像块与周围信号的关系来预测DC分量的预测值,使得DC分量残差更小,提高了DC分量的预测值的精度,从而进一步提高了编码效率。
在某些实现方式中,目标图像块为变换块。由于变换和量化以及逆变换和逆量化可以以变换块为单位,因此,在确定DC分量的预测值时也以变换块为单位,能够降低计算的复杂单,并使得编码器或解码器的设计简单。
在某些实现方式中,预设的DC分量为零。通过将DC分量的预测值设置为零,与设置为其它数值相比,能够降低计算的复杂度。
在某些实施例中,上述方法还包括:确定DC分量的大小,如果DC分量大于预设阈值,则执行上述方法,如果DC分量小于预设的阈值,则直接根据DC分量和AC分量进行编码。由于可以在DC分量大于预设阈值时才采用根据DC分量,使得编码效率的提高弥补了预测DC残差的算法复杂度带来的性能损失,提高了编码或译码的整体性能。
根据本发明的第三方面,一种解码设备,包括用于执行第一方面的解码方法的模块。解码设备包括:熵解码模块,用于从码流中获取目标图像块的变换量化系数的交流AC分量及直流DC分量残差;第一逆量化和逆变换模块,用于对AC分量以及预设的DC分 量进行逆量化和逆变换,得到过渡残差,并且过渡残差与目标图像块的预测值相加得到目标图像块的过渡重建图像块;预测模块,用于根据过渡重建图像块与目标图像块的参考像素区域中的像素的相似性,确定目标图像块的变换量化系数的DC分量的预测值;第二逆量化和逆变换模块,用于对根据DC分量的预测值和DC分量残差确定目标图像块的原始DC分量,对目标图像块的原始DC分量以及AC分量进行逆量化和逆变换,得到目标图像块的残差信号,并根据所残差信号对目标图像块进行解码。
根据本发明的第四方面,一种编码设备,包括用于执行第二方面的编码方法的模块。编码设备包括:变换和量化模块,用于对目标图像块的残差信号进行变换和量化,得到目标图像块的变换量化系数的直流DC分量和交流AC分量;逆量化和逆变换模块,用于对AC分量以及预设的DC分量进行逆量化和逆变换,得到过渡残差,并将过渡残差与目标图像块的预测值相加得到目标图像块的过渡重建图像块;预测模块,用于根据过渡重建图像块的像素与目标图像块的参考像素区域中的像素的相似性,确定目标图像块的变换量化系数的DC分量的预测值;熵编码模块,用于根据DC分量的预测值和目标图像块的原始DC分量确定目标图像块的DC分量残差。
根据本发明的第五方面,一种解码设备,解码设备包括经配置以进行以下操作的视频解码器:从码流中获取目标图像块的变换量化系数的交流AC分量及直流DC分量残差;对AC分量以及预设的DC分量进行逆量化和逆变换,得到过渡残差,并将过渡残差与目标图像块的预测值相加得到目标图像块的过渡重建图像块;根据过渡重建图像块与目标图像块的参考像素区域中的像素的相似性,确定目标图像块的变换量化系数的DC分量的预测值;根据DC分量的预测值和DC分量残差确定目标图像块的原始DC分量;对目标图像块的原始DC分量以及AC分量进行逆量化和逆变换,得到目标图像块的残差信号,并根据所残差信号对目标图像块进行解码。
根据本发明的第六方面,一种用于编码设备,编码设备包括经配置以进行以下操作的视频解码器:对目标图像块的残差信号进行变换和量化,得到目标图像块的变换量化系数的直流DC分量和交流AC分量;对AC分量以及预设的DC分量进行逆量化和逆变换,得到过渡残差,并将过渡残差与目标图像块的预测值相加得到目标图像块的过渡重建图像块;根据过渡重建图像块的像素与目标图像块的参考像素区域中的像素的相似性,确定目标图像块的变换量化系数的DC分量的预测值;根据DC分量的预测值和目标图像块的原始DC分量确定目标图像块的DC分量残差;将AC分量和DC分量残差写入码流。
根据本发明的第七方面,一种存储有指令的计算机可读存储媒体,指令在被执行时使解码设备的一个或多个处理器进行以下操作:从码流中获取目标图像块的变换量化系数的交流AC分量及直流DC分量残差;对AC分量以及预设的DC分量进行逆量化和逆变换,得到过渡残差,并将过渡残差与目标图像块的预测值相加得到目标图像块的过渡重建图像块;根据过渡重建图像块与目标图像块的参考像素区域中的像素的相似性,确定目标图像块的变换量化系数的DC分量的预测值;根据DC分量的预测值和DC分量残差确定目标图像块的原始DC分量;对目标图像块的原始DC分量以及AC分量进行逆量化和逆变换,得到目标图像块的残差信号,并根据所残差信号对目标图像块进行解码。
根据本发明的第八方面,一种存储有指令的计算机可读存储媒体,指令在被执行时使编码设备的一个或多个处理器进行以下操作:对目标图像块的残差信号进行变换和量 化,得到目标图像块的变换量化系数的直流DC分量和交流AC分量;对AC分量以及预设的DC分量进行逆量化和逆变换,得到过渡残差,并将过渡残差与目标图像块的预测值相加得到目标图像块的过渡重建图像块;根据过渡重建图像块的像素与目标图像块的参考像素区域中的像素的相似性,确定目标图像块的变换量化系数的DC分量的预测值;根据DC分量的预测值和目标图像块的原始DC分量确定目标图像块的DC分量残差;将AC分量和DC分量残差写入码流。
附图说明
图1是根据本发明实施例的视频编码系统的示意性框图;
图2是根据本发明实施例的用于视频编码的示意性装置图;
图3是根据本发明实施例的另一视频编解码系统的示意性框图;
图4是根据本发明实施例的视频编码器的示意性框图;
图5是根据本发明实施例的视频解码器的示意性框图;
图6是根据本发明实施例的视频编码器的示意性装置图;
图7是根据本发明实施例的视频解码器的示意性装置图;
图8是根据本发明实施例的编码方法的示意性流程图;
图9是根据本发明实施例的解码方法的示意性流程图;
图10示出了HEVC的35种预测模式;
图11A和图11B是本发明实施例的基于预测模式选择像素信号的示意图;
图12A-12D是本发明的实施例的基于方向预测模式选取像素信号的示意图;
图13A和图13B为本发明的另一实施例的基于预测模式选取像素信号的示意图;
图14是根据本发明的实施例的确定DC分量的预测值的过程的示意性流程图;
图15是根据本发明的另一实施例的确定DC分量的预测值的过程的示意性流程图;
图16是根据本发明的另一实施例的确定DC分量的预测值的过程的示意性流程图;
图17是本发明实施例适用于电视机应用的示意性结构图;
图18是本发明实施例适用于移动电话应用的示意性结构图。
具体实施方式
下面将结合附图,对本申请中的技术方案进行描述。。
图1是视频编解码装置或电子设备50的示意性框图,该装置或者电子设备可以并入根据本发明的实施例的编码解码器。图2是根据本发明实施例的用于视频编码的示意性装置图。下面将说明图1和图2中的单元。
电子设备50可以例如是无线通信系统的移动终端或者用户设备。应理解,可以在可能需要对视频图像进行编码和解码,或者编码,或者解码的任何电子设备或者装置内实施本发明的实施例。
装置50可以包括用于并入和保护设备的壳30。装置50还可以包括形式为液晶显示器的显示器32。在本发明的其它实施例中,显示器可以是适合于显示图像或者视频的任何适当的显示器技术。装置50还可以包括小键盘34。在本发明的其它实施例中,可以运用任何适当的数据或者用户接口机制。例如,可以实施用户接口为虚拟键盘或者数据录入系统作为触敏显示器的一部分。装置可以包括麦克风36或者任何适当的音频输入,该 音频输入可以是数字或者模拟信号输入。装置50还可以包括如下音频输出设备,该音频输出设备在本发明的实施例中可以是以下各项中的任何一项:耳机38、扬声器或者模拟音频或者数字音频输出连接。装置50也可以包括电池40,在本发明的其它实施例中,设备可以由任何适当的移动能量设备,比如太阳能电池、燃料电池或者时钟机构生成器供电。装置还可以包括用于与其它设备的近程视线通信的红外线端口42。在其它实施例中,装置50还可以包括任何适当的近程通信解决方案,比如蓝牙无线连接或者USB/火线有线连接。
装置50可以包括用于控制装置50的控制器56或者处理器。控制器56可以连接到存储器58,该存储器在本发明的实施例中可以存储形式为图像的数据和音频的数据,和/或也可以存储用于在控制器56上实施的指令。控制器56还可以连接到适合于实现音频和/或视频数据的编码和解码或者由控制器56实现的辅助编码和解码的编码解码器电路54。
装置50还可以包括用于提供用户信息并且适合于提供用于在网络认证和授权用户的认证信息的读卡器48和智能卡46,例如通用集成电路卡(Universal Integrated Circuit Card,UICC)和UICC读取器。
装置50还可以包括无线电接口电路52,该无线电接口电路连接到控制器并且适合于生成例如用于与蜂窝通信网络、无线通信系统或者无线局域网通信的无线通信信号。装置50还可以包括天线44,该天线连接到无线电接口电路52用于向其它(多个)装置发送在无线电接口电路52生成的射频信号并且用于从其它(多个)装置接收射频信号。
在本发明的一些实施例中,装置50包括能够记录或者检测单帧的相机,编码解码器54或者控制器接收到这些单帧并对它们进行处理。在本发明的一些实施例中,装置可以在传输和/或存储之前从另一设备接收待处理的视频图像数据。在本发明的一些实施例中,装置50可以通过无线或者有线连接接收图像用于编码/解码。
图3是根据本发明实施例的另一视频编解码系统10的示意性框图。如图3所示,视频编解码系统10包含源装置12及目的地装置14。源装置12产生经编码视频数据。因此,源装置12可被称作视频编码装置或视频编码设备。目的地装置14可解码由源装置12产生的经编码视频数据。因此,目的地装置14可被称作视频解码装置或视频解码设备。源装置12及目的地装置14可为视频编解码装置或视频编解码设备的实例。源装置12及目的地装置14可包括广泛范围的装置,包含台式计算机、移动计算装置、笔记本(例如,膝上型)计算机、平板计算机、机顶盒、智能电话等手持机、电视、相机、显示装置、数字媒体播放器、视频游戏控制台、车载计算机,或其类似者。
目的地装置14可经由信道16接收来自源装置12的编码后的视频数据。信道16可包括能够将经编码视频数据从源装置12移动到目的地装置14的一个或多个媒体及/或装置。在一个实例中,信道16可包括使源装置12能够实时地将编码后的视频数据直接发射到目的地装置14的一个或多个通信媒体。在此实例中,源装置12可根据通信标准(例如,无线通信协议)来调制编码后的视频数据,且可将调制后的视频数据发射到目的地装置14。一个或多个通信媒体可包含无线及/或有线通信媒体,例如射频(RF)频谱或一根或多根物理传输线。一个或多个通信媒体可形成基于包的网络(例如,局域网、广域网或全球网络(例如,因特网))的部分。一个或多个通信媒体可包含路由器、交换器、基站,或促进从源装置12到目的地装置14的通信的其它设备。
在另一实例中,信道16可包含存储由源装置12产生的编码后的视频数据的存储媒体。在此实例中,目的地装置14可经由磁盘存取或卡存取来存取存储媒体。存储媒体可包含多种本地存取式数据存储媒体,例如蓝光光盘、DVD、CD-ROM、快闪存储器,或用于存储经编码视频数据的其它合适数字存储媒体。
在另一实例中,信道16可包含文件服务器或存储由源装置12产生的编码后的视频数据的另一中间存储装置。在此实例中,目的地装置14可经由流式传输或下载来存取存储于文件服务器或其它中间存储装置处的编码后的视频数据。文件服务器可以是能够存储编码后的视频数据且将编码后的视频数据发射到目的地装置14的服务器类型。实例文件服务器包含web服务器(例如,用于网站)、文件传送协议(FTP)服务器、网络附加存储(NAS)装置,及本地磁盘驱动器。
目的地装置14可经由标准数据连接(例如,因特网连接)来存取编码后的视频数据。数据连接的实例类型包含适合于存取存储于文件服务器上的编码后的视频数据的无线信道(例如,Wi-Fi连接)、有线连接(例如,DSL、缆线调制解调器等),或两者的组合。编码后的视频数据从文件服务器的发射可为流式传输、下载传输或两者的组合。
本发明的技术不限于无线应用场景,示例性的,可将技术应用于支持以下应用等多种多媒体应用的视频编解码:空中电视广播、有线电视发射、卫星电视发射、流式传输视频发射(例如,经由因特网)、存储于数据存储媒体上的视频数据的编码、存储于数据存储媒体上的视频数据的解码,或其它应用。在一些实例中,视频编解码系统10可经配置以支持单向或双向视频发射,以支持例如视频流式传输、视频播放、视频广播及/或视频电话等应用。
在图3的实例中,源装置12包含视频源18、视频编码器20及输出接口22。在一些实例中,输出接口22可包含调制器/解调器(调制解调器)及/或发射器。视频源18可包含视频俘获装置(例如,视频相机)、含有先前俘获的视频数据的视频存档、用以从视频内容提供者接收视频数据的视频输入接口,及/或用于产生视频数据的计算机图形系统,或上述视频数据源的组合。
视频编码器20可编码来自视频源18的视频数据。在一些实例中,源装置12经由输出接口22将编码后的视频数据直接发射到目的地装置14。编码后的视频数据还可存储于存储媒体或文件服务器上以供目的地装置14稍后存取以用于解码及/或播放。
在图3的实例中,目的地装置14包含输入接口28、视频解码器30及显示装置32。在一些实例中,输入接口28包含接收器及/或调制解调器。输入接口28可经由信道16接收编码后的视频数据。显示装置32可与目的地装置14整合或可在目的地装置14外部。一般来说,显示装置32显示解码后的视频数据。显示装置32可包括多种显示装置,例如液晶显示器(LCD)、等离子体显示器、有机发光二极管(OLED)显示器或其它类型的显示装置。
视频编码器20及视频解码器30可根据视频压缩标准(例如,高效率视频编解码H.265标准)而操作,且可遵照HEVC测试模型(HM)。H.265标准的文本描述ITU-TH.265(V3)(04/2015)于2015年4月29号发布,可从http://handle.itu.int/11.1002/1000/12455下载,文件的全部内容以引用的方式并入本文中。
或者,视频编码器20及视频解码器30可根据其它专属或行业标准而操作,标准包含ITU-TH.261、ISO/IECMPEG-1Visual、ITU-TH.262或ISO/IECMPEG-2Visual、 ITU-TH.263、ISO/IECMPEG-4Visual,ITU-TH.264(还称为ISO/IECMPEG-4AVC),包含可分级视频编解码(SVC)及多视图视频编解码(MVC)扩展。应理解,本发明的技术不限于任何特定编解码标准或技术。
此外,图3仅为实例且本发明的技术可应用于未必包含编码装置与解码装置之间的任何数据通信的视频编解码应用(例如,单侧的视频编码或视频解码)。在其它实例中,从本地存储器检索数据,经由网络流式传输数据,或以类似方式操作数据。编码装置可编码数据且将数据存储到存储器,及/或解码装置可从存储器检索数据且解码数据。在许多实例中,通过彼此不进行通信而仅编码数据到存储器及/或从存储器检索数据及解码数据的多个装置执行编码及解码。
视频编码器20及视频解码器30各自可实施为多种合适电路中的任一者,例如一个或多个微处理器、数字信号处理器(DSP)、专用集成电路(ASIC)、现场可编程门阵列(FPGA)、离散逻辑、硬件或其任何组合。如果技术部分地或者全部以软件实施,则装置可将软件的指令存储于合适的非瞬时计算机可读存储媒体中,且可使用一个或多个处理器执行硬件中的指令以执行本发明的技术。可将前述各者中的任一者(包含硬件、软件、硬件与软件的组合等)视为一个或多个处理器。视频编码器20及视频解码器30中的每一者可包含于一个或多个编码器或解码器中,其中的任一者可整合为其它装置中的组合式编码器/解码器(编解码器(CODEC))的部分。
本发明大体上可指代视频编码器20将某一信息“用信号发送”到另一装置(例如,视频解码器30)。术语“用信号发送”大体上可指代语法元素及/或表示编码后的视频数据的传达。此传达可实时或近实时地发生。或者,此通信可在一时间跨度上发生,例如可在编码时以编码后得到的二进制数据将语法元素存储到计算机可读存储媒体时发生,语法元素在存储到此媒体之后接着可由解码装置在任何时间检索。
视频编码器20编码视频数据。视频数据可包括一个或多个图片。视频编码器20可产生码流,码流以比特流的形式包含了视频数据的编码信息。编码信息可以包含编码图片数据及相关联数据。相关联数据可包含序列参数集(SPS)、图片参数集(PPS)及其它语法结构。SPS可含有应用于零个或多个序列的参数。PPS可含有应用于零个或多个图片的参数。语法结构是指码流中以指定次序排列的零个或多个语法元素的集合。
产生图片的编码信息,视频编码器20可将图片分割成编码树型块(CTB)的栅格。在一些例子中,CTB可被称作“树型块”、“最大编码单元”(LCU)或“编码树型单元”。CTB不限于特定大小且可包含一个或多个编码单元(CU)。每一个CTB可以与图片内的具有相等大小的像素块相关联。每一像素可对应一个亮度(luminance或luma)采样及两个色度(chrominance或chroma)采样。因此,每一个CTB可与一个亮度采样块及两个色度采样块相关联。图片的CTB可分成一个或多个条带。在一些实例中,每一个条带包含整数个数的CTB。作为编码图片的一部分,视频编码器20可产生图片的每一条带的编码信息,即编码条带内的CTB。为了编码CTB,视频编码器20可对与CTB相关联的像素块递归地执行四叉树分割,以将像素块分割成递减的像素块。较小的像素块可以和CU相关联。
图4是根据本发明实施例的视频编码器20的示意性框图,包括编码端预测模块201、变换量化模块202、熵编码模块203、编码重建模块204、编码端滤波模块205,图5是根据本发明实施例的视频解码器30的示意性框图,包括解码端预测模块206、逆变换逆量化模块207、熵解码模块208、解码重建模块209、解码滤波模块210。具体的:
编码端预测模块201和解码端预测模块206用于产生预测数据。视频编码器20可产生每一不再分割CU(即,最小尺寸的CU)的一个或多个预测单元(PU)。CU的每一个PU可与CU的像素块内的不同像素块相关联。视频编码器20可针对CU的每一个PU产生预测性像素块。视频编码器20可使用帧内预测或帧间预测来产生PU的预测性像素块。如果视频编码器20使用帧内预测来产生PU的预测性像素块,则视频编码器20可基于与PU相关联且位于同一视频帧内的图片(或称为图像)的解码后的像素来产生PU的预测性像素块。如果视频编码器20使用帧间预测来产生PU的预测性像素块,则视频编码器20可基于不同于与PU相关联的图片,即不同于所述PU所在的视频帧,的一个或多个图片的解码后的像素来产生PU的预测性像素块。视频编码器20可基于CU的PU的预测性像素块来产生CU的残余像素块。CU的残余像素块可指示CU的PU的预测性像素块中的采样值与CU的初始像素块中的对应采样值之间的差(或称为残差信号或残差数据)。
变换量化模块202用于对经过预测的残差信号或残差数据进行处理。视频编码器20可对CU的残余像素块执行递归四叉树分割以将CU的残余像素块分割成与CU的变换单元(TU)相关联的一个或多个较小残余像素块。因为与TU相关联的像素块中的像素各自对应一个亮度采样及两个色度采样,所以每一个TU可与一个亮度的残余采样块及两个色度的残余采样块相关联。视频编码器20可将一个或多个变换应用于与TU相关联的残余采样块以产生系数块(即,系数的块)。变换可以是离散余弦变换,即DCT变换,或者它的变体,如离散正弦变换,即DST变换。采用DCT的变换矩阵,通过在水平和竖直方向应用一维变换计算二维变换,获得系数块(也称为变换系数)。视频编码器20可对系数块中的每一个变换系数执行量化程序。量化一般指变换系数经量化以减少用以表示系数的数据量,可以理解为将无限个数据映射成有限个数据,从而提供进一步压缩的过程。经过变换和量化后得到变换量化系数,该变换量化系数包括DC分量和AC分量,DC分量及AC分量的求取过程非本发明的重点,在此不做展开说明,具体的求解过程可以参照ITU-TH.265(V3)(04/2015)。逆变换逆量化模块207执行的过程包括变换量化模块202的逆过程。
视频编码器20可产生表示量化后系数块中的系数的语法元素的集合。视频编码器20通过熵编码模块203可将熵编码操作(例如,上下文自适应二进制算术译码(CABAC)操作)应用于上述语法元素中的部分或者全部。为将CABAC编码应用于语法元素,视频编码器20可将语法元素二进制化以形成包括一个或多个位(称作“二进位”)的二进制序列。视频编码器20可使用规则(regular)编码来编码二进位中的一部分,且可使用旁通(bypass)编码来编码二进位中的其它部分。
除熵编码系数块的语法元素外,视频编码器20通过编码重建模块204,可将逆量化及逆变换应用于变换后的系数块,以从变换后的系数块重建残余采样块。视频编码器20可将重建后的残余采样块加到一个或多个预测性采样块的对应采样块,以产生重建后的采样块。通过重建每一色彩分量的采样块,视频编码器20可重建与TU相关联的像素块。以此方式重建CU的每一TU的像素块,直到CU的整个像素块重建完成。
根据本发明的实施例,视频编码器20还可以进一步包括DC分量预测模块206,用于确定DC分量的预测值,具体而言,DC分量预测模块206可以根据变换量化模块202输出的变换量化系数构建一组新的变换量化系数,该构建的新的变换量化系数中的DC 分量被强制为0,AC分量为变换量化模块202输出的AC分量,通过逆变换及逆量化模块207对该新的变换量化系数进行逆量化及逆变处理,通过逆变换和逆量化过程得到的重建后的残余采样块,并且可将重建后的残余采样块加到一个或多个预测性采样块的对应采样块,以产生重建后的采样块(下文中也称为过渡重建图像块)。最后,根据过渡重建图像块的像素与参考像素区域的像素的相似性或相关性,确定DC分量的预测值。最后根据DC分量的预测值和变换量化模块202输出的DC分量得到DC分量的残差信号。DC分量的残差信号和AC分量由熵编码模块203写入码流。
在视频编码器20重建构CU的像素块之后,视频编码器20通过编码端滤波模块205,执行消块滤波操作以减少与CU相关联的像素块的块效应。在视频编码器20执行消块滤波操作之后,视频编码器20可使用采样自适应偏移(SAO)来修改图片的CTB的重建后的像素块。在执行这些操作之后,视频编码器20可将CU的重建后的像素块存储于解码图片缓冲器中以用于产生其它CU的预测性像素块。
视频解码器30可接收码流。码流以比特流的形式包含了由视频编码器20编码的视频数据的编码信息。视频解码器30通过熵解码模块208,解析码流以从码流提取语法元素。当视频解码器30执行CABAC解码时,视频解码器30可对部分二进位执行规则解码且可对其它部分的二进位执行旁通解码,码流中的二进位与语法元素具有映射关系,通过解析二进位获得语法元素。
视频解码器30通过解码重建模块209,可基于从码流提取的语法元素来重建视频数据的图片。基于语法元素来重建视频数据的过程大体上与由视频编码器20执行以产生语法元素的过程互逆。举例来说,视频解码器30可基于与CU相关联的语法元素来产生CU的PU的预测性像素块。另外,视频解码器30可逆量化与CU的TU相关联的系数块。视频解码器30可对逆量化后的系数块执行逆变换以重建与CU的TU相关联的残余像素块。视频解码器30可基于预测性像素块及残余像素块来重建CU的像素块。
根据本发明的实施例,视频解码器30还可以进一步包括DC分量预测模块211,用于确定DC分量的预测值,具体而言,DC分量预测模块211可根据从码流中获取的变换系数的来重构另一变换量化系数,该重构的另一换量化系数中的DC分量被强制为0,AC分量为熵解码模块208输出的AC分量,并通过逆量化及逆变换模块对该重构的另一变换模块执行逆量化和逆变换过程,通过逆变换和逆量化过程得到的重建残余采样块,并且可将重建后的残余采样块加到一个或多个预测性采样块的对应采样块,以产生重建后的采样块(下文中也称为过渡重建图像块)。最后,根据过渡重建图像块的像素与参考像素区域的像素的相似性或相关性,确定DC分量的预测值。最后根据DC分量的预测值和熵解码模块208输出的DC分量的残差信号得到重建后的DC分量。
在视频解码器30重建CU的像素块之后,视频解码器30通过解码滤波模块210,执行消块滤波操作以减少与CU相关联的像素块的块效应。另外,基于一个或多个SAO语法元素,视频解码器30可执行与视频编码器20相同的SAO操作。在视频解码器30执行这些操作之后,视频解码器30可将CU的像素块存储于解码图片缓冲器中。解码图片缓冲器可提供用于后续运动补偿、帧内预测及显示装置呈现的参考图片。
图6为说明经配置以实施本发明的技术的实例视频编码器20的框图。应理解,图6是示例性的而不应视为限制如本发明广泛例证并描述的技术。如图6所示,视频编码器20包含预测处理单元100、残余产生单元102、变换处理单元104、量化处理单元106、 逆量化处理单元108、逆变换处理单元110、重建单元112、滤波器单元113、解码图片缓冲器114、DC预测模块130内的逆量化处理单元131、逆变换处理单元132、重建单元133、DC分量预测处理单元134、残差产生单元135及熵编码单元116。熵编码单元116包含规则CABAC编解码引擎及旁通编解码引擎。预测处理单元100包含帧间预测处理单元121及帧内预测处理单元126。帧间预测处理单元121包含运动估计单元及运动补偿单元。在其它实例中,视频编码器20可包含更多、更少或不同的功能组件。
视频编码器20接收视频数据。为编码视频数据,视频编码器20可编码视频数据的每一图片的每一条带。作为编码条带的一部分,视频编码器20可编码条带中的每一CTB。作为编码CTB的一部分,预测处理单元100可对与CTB相关联的像素块执行四叉树分割,以将像素块分成递减的像素块。举例来说,预测处理单元100可将CTB的像素块分割成四个相等大小的子块,将子块中的一个或多个分割成四个相等大小的子子块,等等。
视频编码器20可编码图片中的CTB的CU以产生CU的编码信息。视频编码器20可根据折形扫描次序来编码CTB的CU。换句话说,视频编码器20可按左上CU、右上CU、左下CU及接着右下CU来编码CU。当视频编码器20编码分割后的CU时,视频编码器20可根据折形扫描次序来编码与分割后的CU的像素块的子块相关联的CU。
此外,预测处理单元100可在CU的一个或多个PU中分割CU的像素块。视频编码器20及视频解码器30可支持各种PU大小。假定特定CU的大小为2N×2N,视频编码器20及视频解码器30可支持2N×2N或N×N的PU大小以用于帧内预测,且支持2N×2N、2N×N、N×2N、N×N或类似大小的对称PU以用于帧间预测。视频编码器20及视频解码器30还可支持2N×nU、2N×nD、nL×2N及nR×2N的不对称PU以用于帧间预测。
帧间预测处理单元121可通过对CU的每一PU执行帧间预测而产生PU的预测性数据。PU的预测性数据可包含对应于PU的预测性像素块及PU的运动信息。条带可为I条带、P条带或B条带。帧间预测单元121可根据PU是在I条带、P条带还是B条带中而对CU的PU执行不同操作。在I条带中,所有PU执行帧内预测。
如果PU在P条带中,则运动估计单元122可搜索参考图片的列表(例如,“列表0”)中的参考图片以查找PU的参考块。PU的参考块可为最紧密地对应于PU的像素块的像素块。运动估计单元122可产生指示列表0中的含有PU的参考块的参考图片的参考图片索引,及指示PU的像素块与参考块之间的空间位移的运动向量。运动估计单元122可将参考图片索引及运动向量作为PU的运动信息而输出。运动补偿单元124可基于由PU的运动信息指示的参考块来产生PU的预测性像素块。
如果PU在B条带中,则运动估计单元122可对PU执行单向帧间预测或双向帧间预测。为对PU执行单向帧间预测,运动估计单元122可搜索第一参考图片列表(“列表0”)或第二参考图片列表(“列表1”)的参考图片以查找PU的参考块。运动估计单元122可将以下各者作为PU的运动信息而输出:指示含有参考块的参考图片的列表0或列表1中的位置的参考图片索引、指示PU的像素块与参考块之间的空间位移的运动向量,及指示参考图片是在列表0中还是在列表1中的预测方向指示符。为对PU执行双向帧间预测,运动估计单元122可搜索列表0中的参考图片以查找PU的参考块,且还可搜索列表1中的参考图片以查找PU的另一参考块。运动估计单元122可产生指示含有参考块的参考图片的列表0及列表1中的位置的参考图片索引。另外,运动估计单元122可产生指示参考块与PU的像素块之间的空间位移的运动向量。PU的运动信息可包含PU的 参考图片索引及运动向量。运动补偿单元124可基于由PU的运动信息指示的参考块来产生PU的预测性像素块。
帧内预测处理单元126可通过对PU执行帧内预测而产生PU的预测性数据。PU的预测性数据可包含PU的预测性像素块及各种语法元素。帧内预测处理单元126可对I条带、P条带及B条带内的PU执行帧内预测。
为对PU执行帧内预测,帧内预测处理单元126可使用多个帧内预测模式来产生PU的预测性数据的多个集合。为使用帧内预测模式来产生PU的预测性数据的集合,帧内预测处理单元126可在与帧内预测模式相关联的方向上跨越PU的采样块扩展来自相邻PU的采样块的采样。假定从左向右、从上而下的编码次序用于PU、CU及CTB,相邻PU可在PU的上方,在PU的右上方,在PU的左上方或在PU的左方。帧内预测处理单元126可使用包含不同数目的帧内预测模式,例如,33个方向性帧内预测模式。在一些实例中,帧内预测模式的数目可取决于PU的像素块的大小。
预测处理单元100可从通过帧间预测处理单元121针对PU而产生的预测性数据或通过帧内预测处理单元126针对PU而产生的预测性数据当中选择CU的PU的预测性数据。在一些实例中,预测处理单元100基于预测性数据的集合的速率/失真量度来选择CU的PU的预测性数据。例如,使用拉格朗日代价函数来在编码模式及其参数值(比如运动矢量、参考索引和帧内预测方向)之间进行选择。这一种类的代价函数使用加权因子lambda将由于有损编码方法所致的实际的或者估计的图像失真与为了表示在图像区域中的像素值而需要的实际的或估计的信息量联系在一起:C=D+lambda×R,其中C是待最小化的拉格朗日代价,D是具有模式及其参数的图像失真(例如均方误差),R是为了在解码器中重构图像块而需要的比特数(例如包括用于表示候选运动矢量的数据量)。一般的,代价最小的编码模式被选定作为实际编码模式。选定预测性数据的预测性像素块在本文中可被称作选定预测性像素块。
残余产生单元102可基于CU的像素块及CU的PU的选定预测性像素块来产生CU的残余像素块。举例来说,残余产生单元102可产生CU的残余像素块,使得残余像素块中的每一采样具有等于以下两者之间的差的值:CU的像素块中的采样,及CU的PU的选定预测性像素块中的对应采样。
预测处理单元100可执行四叉树分割以将CU的残余像素块分割成子块。每一不再划分的残余像素块可与CU的不同TU相关联。与CU的TU相关联的残余像素块的大小及位置与基于CU的PU的像素块的大小及位置没有必然联系。
因为TU的残余像素块的像素可对应一个亮度采样及两个色度采样,所以每一个TU可与一个亮度采样块及两个色度采样块相关联。变换处理单元104可通过将一个或多个变换应用于与TU相关联的残余采样块而产生CU的每一个TU的系数块。举例来说,变换处理单元104可将离散余弦变换(DCT)、方向性变换或概念上类似的变换应用于残余采样块。
量化单元106可量化系数块中的系数。举例来说,n位系数可在量化期间舍位到m位系数,其中n大于m。量化单元106可基于与CU相关联的量化参数(QP)值来量化与CU的TU相关联的系数块。视频编码器20可通过调整与CU相关联的QP值来调整应用于与CU相关联的系数块的量化程度。
逆量化单元108及逆变换处理单元110可分别将逆量化及逆变换应用于变换后的系 数块以从系数块重建残余采样块。重建单元112可将重建后的残余采样块的采样加到预测处理单元100产生的一个或多个预测性采样块的对应采样,以产生与TU相关联的重建后的采样块。通过此方式重建CU的每一个TU的采样块,视频编码器20可重建CU的像素块。
逆量化处理单元131及逆变换处理单元132可分别将逆量化及逆变换应用于变换后的系数块且该系数块中的DC分量被设置为0,以从该系数块重建残余采样块。重建单元133可将重建后的残余采样块的采样加到预测处理单元100产生的一个或多个预测性采样块的对应采样(也称为预测值),以产生与TU相关联的重建后的采样块(下文中也称为过渡重建图像块)。该过渡重建图像块不同于重建单元112产生的重建后的采样块。DC分量预测处理单元134可根据重建单元133得到的过渡重建图像块与参数像素区域的像素的相似性或相关性,确定DC分量的预测值。DC残差产生单元135可基于DC分量预测处理单元134输出的DC分量预测值与量化处理单元106产生的原始DC分量产生DC残差(残余)。举例来说,可以将DC分量预测值与原始DC分量相减得到DC残差。
应理解,逆量化处理单元131和逆变换处理单元132也可以分别复用逆量化处理单元108和逆变换处理单元110的功能,在这种情况下,可以另外设置系数重构单元,用于从量化处理单元106获取变换量化系数的DC分量,将DC分量强制为零,并将强制为零的DC分量输入到逆量化处理单元108进行逆量化。逆量化处理单元108用于进一步将AC分量和强制为零的DC分量进行逆量化,并由逆变换处理单元110进行逆变换,以从该系数块重建残余采样块。同样,重建单元133可将重建后的残余采样块的采样加到预测处理单元100产生的一个或多个预测性采样块的对应采样(也称为预测值),以产生与TU相关联的重建后的采样块(下文中也称为过渡重建图像块)。
滤波器单元113可执行消块滤波操作以减少与CU相关联的像素块的块效应。此外,滤波器单元113可将由预测处理单元100确定的SAO偏移应用于重建后的采样块以恢复像素块。滤波器单元113可产生CTB的SAO语法元素的编码信息。
解码图片缓冲器114可存储重建后的像素块。帧间预测单元121可使用含有重建后的像素块的参考图片来对其它图片的PU执行帧间预测。另外,帧内预测处理单元126可使用解码图片缓冲器114中的重建后的像素块来对在与CU相同的图片中的其它PU执行帧内预测。
熵编码单元116可接收来自视频编码器20的其它功能组件的数据。举例来说,熵编码单元116可接收来自DC残差产生单元135的DC分量的残差以及AC分量且可接收来自预测处理单元100的语法元素。熵编码单元116可对数据执行一个或多个熵编码操作以产生熵编码后的数据。举例来说,熵编码单元116可对数据执行上下文自适应可变长度编解码(CAVLC)操作、CABAC操作、可变到可变(V2V)长度编解码操作、基于语法的上下文自适应二进制算术编解码(SBAC)操作、机率区间分割熵(PIPE)编解码操作,或其它类型的熵编码操作。在一特定实例中,熵编码单元116可使用规则CABAC引擎118来编码语法元素的经规则CABAC编解码二进位,且可使用旁通编解码引擎120来编码经旁通编解码二进位。
图7为说明经配置以实施本发明的技术的实例视频解码器30的框图。应理解,图7是示例性的而不应视为限制如本发明广泛例证并描述的技术。如图7所示,视频解码器30包含熵解码单元150、预测处理单元152、逆量化单元154、逆变换处理单元156、重 建单元158、滤波器单元159、DC预测模块170内的逆量化处理单元171、逆变换处理单元172、重建单元173、DC分量预测处理单元174、DC分量产生单元175以及解码图片缓冲器160。预测处理单元152包含帧间预测处理单元162以及帧内预测处理单元164。熵解码单元150包含规则CABAC编解码引擎及旁通编解码引擎。在其它实例中,视频解码器30可包含更多、更少或不同的功能组件。
视频解码器30可接收码流。熵解码单元150可解析码流以从码流提取语法元素。作为解析码流的一部分,熵解码单元150可解析码流中的经熵编码后的语法元素。另外,熵解码单元150还可以从码流中提取变化量化系数,包括:DC分量的残差和AC分量。
逆量化处理单元171及逆变换处理单元172可分别将逆量化及逆变换应用于变换后的系数块且该系数块中的DC分量被设置为0,以从该系数块重建残余采样块。重建单元173可将重建后的残余采样块的采样加到预测处理单元152产生的一个或多个预测性采样块的对应采样(也称为预测值),以产生与TU相关联的重建后的采样块(也称为过渡重建图像块)。该过渡重建图像块不同于重建单元158产生的重建后的采样块,即该过渡重建图像块为获取DC分量所使用的重建图像块,不能直接用来重建解码图像。DC分量预测处理单元174可根据重建单元173得到的过渡重建图像块与参数像素区域的像素的相似性或相关性,确定DC分量的预测值。DC残差产生单元175可基于DC分量预测处理单元174输出的DC分量预测值与熵解码单元150产生的DC残差(残余)产生DC分量。举例来说,可以将DC分量预测值与在DC残差相加得到DC分量。
应理解,逆量化处理单元171和逆变换处理单元172也可以分别复用逆量化处理单元154和逆变换处理单元156的功能,在这种情况下,可以设置另一系数重构单元,用于将变换量化系数的DC分量强制为零,并将强制为零的DC分量输入到逆量化处理单元154进行逆量化。逆量化处理单元156用于进一步将AC分量和强制为零的DC分量进行逆量化,并由逆变换处理单元156进行逆变换,以从该系数块重建残余采样块。同样,重建单元173可将重建后的残余采样块的采样加到预测处理单元152产生的一个或多个预测性采样块的对应采样(也称为预测值),以产生与TU相关联的重建后的采样块(下文中也称为过渡重建图像块)。
预测处理单元152、逆量化单元154、逆变换处理单元156、重建单元158及滤波器单元159可根据根据DC分量和AC分量等语法元素来解码视频数据,即产生解码后的视频数据。
语法元素可包含经规则CABAC编解码二进位及经旁通编解码二进位。熵解码单元150可使用规则CABAC编解码引擎166来解码经规则CABAC编解码二进位,且可使用旁通编解码引擎168来解码经旁通编解码二进位。
如果PU使用帧内预测编码,则帧内预测处理单元164可执行帧内预测以产生PU的预测性采样块。帧内预测处理单元164可使用帧内预测模式以基于空间相邻PU的像素块来产生PU的预测性像素块。帧内预测处理单元164可根据从码流解析的一个或多个语法元素来确定PU的帧内预测模式。
帧间预测处理单元162可以包括运动补偿单元,运动补偿单元可根据从码流解析的语法元素来构造第一参考图片列表及第二参考图片列表。此外,如果PU使用帧间预测编码,则熵解码单元150可解析PU的运动信息。运动补偿单元162可根据PU的运动信 息来确定PU的一个或多个参考块。运动补偿单元162可根据PU的一个或多个参考块来产生PU的预测性像素块。
另外,视频解码器30可对不再分割的CU执行重建操作。为对不再分割的CU执行重建操作,视频解码器30可对CU的每一TU执行重建操作。通过对CU的每一TU执行重建操作,视频解码器30可重建与CU相关联的残余像素块。
作为对CU的TU执行重建操作的一部分,逆量化单元154可逆量化(即,解量化)与TU相关联的系数块。逆量化单元154可使用与TU的CU相关联的QP值来确定量化程度,且与确定逆量化单元154将应用的逆量化程度相同。
在逆量化单元154逆量化系数块中的DC分量和AC分量之后,逆变换处理单元156可将一个或多个逆变换应用于系数块,以便产生与TU相关联的残余采样块。举例来说,逆变换处理单元156可将逆DCT、逆整数变换、逆卡忽南-拉维(Karhunen-Loeve)变换(KLT)、逆旋转变换、逆方向性变换或其它与编码端的变换对应的逆变换应用于系数块。
重建单元158可在适用时使用与CU的TU相关联的残余像素块及CU的PU的预测性像素块(即,帧内预测数据或帧间预测数据)以重建CU的像素块。特定来说,重建单元158可将残余像素块的采样加到预测性像素块的对应采样以重建CU的像素块。
滤波器单元159可执行消块滤波操作以减少与CTB的CU相关联的像素块的块效应。另外,滤波器单元159可根据从码流解析的SAO语法元素来修改CTB的像素值。举例来说,滤波器单元159可根据CTB的SAO语法元素来确定修正值,且将所确定的修正值加到CTB的重建后的像素块中的采样值。通过修改图片的CTB的部分或全部像素值,滤波器单元159可根据SAO语法元素来修正视频数据的重建图片。
视频解码器30可将CU的像素块存储于解码图片缓冲器160中。解码图片缓冲器160可提供参考图片以用于后续运动补偿、帧内预测及显示装置(例如,图3的显示装置32)呈现。举例来说,视频解码器30可根据解码图片缓冲器160中的像素块来对其它CU的PU执行帧内预测操作或帧间预测操作。
下面结合图8和图9,以HEVC为例详细说明本发明实施例的编码过程和解码过程。应理解,在诸如H.264之类的视频编码技术中,也可以采用类似的方法。
如前所述,本发明的实施例与常规视频编码技术的主要不同之外在于,写入码流的变换量化系数为DC分量的残差和AC分量,而非像常规视频编码技术那样,写入码流的变换量化系数为DC分量和AC分量。由于DC分量的残差的绝对值通常小于DC分量,而值越小编码效率越高,因此提高了编码效率。
图8是根据本发明的一个实施例的编码过程的示意图。
在编码器进行编码时,待编码图像首先被划分为互不重叠的图像块。每个图像块被称为一个编码树单元CTU(Coding Tree Unit)。CTU可被进一步划分为多个子块,每个子块即是一个可独立确定其编码参数的编码块(Coding Unit,CU)。CTU的划分采用四叉树的结构。在压缩编码过程中,每个CU将独立地选择其最佳的编码参数,例如,预测模式选择、预测块(Prediction Unit,PU)的划分等。当一个CTU中的所有CU被按次序编码压缩结束后,才能继续编码后续CTU。
810,在编码器进行预测得到预测块的预测值,并将预测块的预测值与原始值相减,得到预测块的残差信号。例如,该预测可以是帧内预测。其中,所述的原始值指的是待编码图像块,包括目标图像块,的原始的像素值。
820,对目标图像块的残差信号进行变换和量化,得到目标图像块的变换量化系数的直流DC分量和交流AC分量。
例如,对目标图像块的残差信号进行变换和量化,得到目标图像块的初始变换量化系数,初始变换量化系数包括DC分量和AC分量。
本实施例的目标图像块可以为变换块或对应于变换块。一个预测块可以划分为多个变换块,属于同一个预测块的变换块采用相同的预测模式。本发明的实施例可以以变换块为单位对残差信号进行变换和量化。每个变换块的残差信号可以从预测块的残差信号中获得。
830,对初始变换量化系数进行逆量化和逆变换得到目标图像块的重建残差信号,然后将图标图像块的重建残差与目标像块的预测值相加得到重建值,从而得到目标图像块的重建图像块,以供后续预测编码作为参考信息。
应理解,810至830与常规视频编码类似,在此不再赘述。
840,对AC分量以及预设的DC分量进行逆量化和逆变换,得到过渡残差,并将过渡残差与目标图像块的预测值相加得到目标图像块的过渡重建图像块。
例如,对过渡变换量化系数进行逆量化和逆变换,得到过渡残差,过渡残差与预测值相加得到目标图像块的过渡重建图像块,其中过渡变换量化系数包括第二DC分量和820得到的AC分量,第二DC分量为预设值(例如,预设值为0)。采用第二DC分量设置为0的方案的好处在于能够降低计算的复杂度。
针对每个目标图像块,将过渡变换量化系数的DC分量设置为零,即该变换量化系数包括从码流中解析出的AC分量和取值为0的DC分量,并对过渡变换量化系数进行逆量化和逆变换,得到DC分量为零时目标图像块的重建残差,并将该重建残差与目标像块的预测值相加得到重建值,从而得到目标图像块在DC分量为零时的过渡重建图像块,记为BDC=0。应理解,840得到的过渡重建图像块不同于830得到的重建图像块。
850,根据过渡重建图像块的像素与目标图像块的参考像素区域中的像素的相似性,确定目标图像块的变换量化系数的DC分量的预测值。
例如,根据重过渡建图像块的像素与参考像素区域中的像素的相似性,确定DC分量的预测值。计算目标图像块的DC分量的预测值DCpred的具体方法参见图14的实施例中的描述,在此不再赘述。
860,根据DC分量的预测值和目标图像块的原始DC分量确定目标图像块的DC分量残差。
例如,将目标图像块的DC分量与DC分量的预测值相减得到DC分量的残差ΔDC。
870,将AC分量和DC分量残差写入码流。
例如,将重构变换量化系数写入码流,其中重构变换量化系数包括DC分量与预测值的差值和820得到的AC分量。可以对重构变换量化系数进行熵编码,并将编码后的重构变换量化系数写入码流。另外,CTU中CU的划分方式、CU的编码模式、CU中PU的划分模式、预测模式选择等信息也经过熵编码写入码流。
基于本申请的技术方案,可以在将DC分量设置为预设值的情况下对变换量化系数进行逆量化和逆变换,得到目标图像块的过渡重建图像块,并且根据过渡重建图像块的像素与目标图像块的参考像素区域中的像素的相似性,确定DC分量的预测值。在对目 标图像块进行编码或解码时,可以使用DC分量残差来替代DC分量。由于DC分量残差的绝对值小于DC分量,因此提高了编码效率。
另外,虽然也可以直接采用参考像素区域的DC分量作为DC分量的预测值,但本发明实施例根据DC分量为预设值的过渡重建图像块的像素与参考像素区域中的像素的相似性来确定DC分量的预测值,提高了DC分量预测值的精度,使得DC分量残差更小,因此,编码效率更高。
图9是根据本发明的另一实施例的解码过程的示意图。
解码端的处理方法与编码端相对应,解码端得到一个CTU的码流后,熵解码得到该CTU中CU的划分方式和其中各个CU中PU的划分模式,预测模式以及变换量化系数等信息,然后按照顺序依次解码。当CTU中所有CU均解码完后,才能继续解码后续的CTU。
910,从码流中获取目标图像块的变换量化系数的交流AC分量及直流DC分量残差。
例如,从码流中获取重构变换量化系数,其中重构变换量化系数包括目标图像块的第一直流DC分量与DC分量的预测值的差值和目标图像块的交流AC分量。解码端可以接收编码端发送的码流,并对码流进行熵解码得到目标图像块的预测模式信息和变换量化系数。预测模式信息可以包括每个预测块的预测模式,例如,在HEVC中,预测模式可以是DC预测模式、平面预测模式和33种角度(或方向)预测模式中的一种。变换量化系数包括ΔDC和AC分量。对应于每个预测块,可以相应地划分多个变换块,属于同一个预测块的变换块采用相同的预测模式。本实施例的目标图像块可以为变换块或对应于变换块。
920,对AC分量以及预设的DC分量进行逆量化和逆变换,得到过渡残差,并将过渡残差与目标图像块的预测值相加得到目标图像块的过渡重建图像块。
例如,对过渡变换量化系数进行逆量化和逆变换,得到目标图像块的过渡重建图像块,其中过渡变换量化系数包括第二DC分量和AC分量,第二DC分量为预设值。针对每个目标图像块,将变换量化系数的DC分量设置为零,即变换量化系数包括从码流中解析出的AC分量和取值为0的DC分量,并对该变换量化系数进行逆量化和逆变换,得到DC分量为零时的目标像块的重建残差,并将该重建残差与目标像块的预测值相加获得目标图像块在DC分量为零时的过渡重建图像块,记为BDC=0
930,根据过渡重建图像块与目标图像块的参考像素区域中的像素的相似性,确定目标图像块的变换量化系数的DC分量的预测值。
例如,根据过渡重建图像块的像素与目标图像块的参考像素区域中的像素的相似性,确定DC分量的预测值。计算目标图像块的DC分量的预测值DCpred的具体方法参见下面图14的实施例的描述。
940,根据DC分量的预测值和DC分量残差确定目标图像块的原始DC分量。
例如,根据DC分量与DC分量的预测值的差值以及DC分量的预测值,确定DC分量。可以将上述步骤中获得的DC分量的预测值DCpred和从码流中解析出来的ΔDC相加,得到目标图像块的量化后的DC分量值。
950,对目标图像块的原始DC分量以及AC分量进行逆量化和逆变换,得到目标图像块的残差信号。
例如,对初始变换量化系数进行逆量化和逆变换,得到目标图像块的残差信号,其中初始变换量化系数包括DC分量和AC分量。将上述步骤中获得的目标图像块的DC分量值,与从码流中解析出来的目标图像块的AC分量值,进行逆量化和逆变换,从而得到目标图像块的重建残差信号。
960,根据所残差信号对目标图像块进行解码。
例如,根据预测值和重建残差信号生成重建图像块。
应理解,根据预测值和重建残差信号生成重建图像块的过程与常规视频编码类似,在此不再赘述。
基于本申请的技术方案,可以在将DC分量设置为预设值的情况下对变换量化系数进行逆量化和逆变换,得到目标图像块的过渡重建图像块,并且根据过渡重建图像块的像素与目标图像块的参考像素区域中的像素的相似性,确定DC分量的预测值。在对目标图像块进行编码或解码时,可以使用DC分量残差来替代DC分量。由于DC分量残差的绝对值小于DC分量,因此提高了编码效率。
另外,虽然也可以直接采用参考像素区域的DC分量作为DC分量的预测值,但本发明实施例根据DC分量为预设值的过渡重建图像块的像素与参考像素区域中的像素的相似性来确定DC分量的预测值,提高了DC分量预测值的精度,使得DC分量残差更小,因此,编码效率更高。
下面结合图10至图13B详细描述确定目标图像块的DC分量的预测值的方法。
图10示出了HEVC的35种预测模式。参见图10,HEVC中的帧内预测模式包括直流(DC)预测模式(索引为1),平面(Planar)预测模式(索引为0)和角度预测模式(索引为2至34)。在进行帧内预测时,可以从上述35种预测模式中选择一种。当前图像块使用其左边和上边参考像素区域的已重建块的信息进行预测。采用DC预测模式时,当前图像块中所有像素使用参考像素的平均值作为预测值。采用平面预测模式时,使用参考像素进行双线性插值得到当前图像块中所有像素的预测值。采用角度预测式时,利用当前图像块所包含的纹理与邻近已重建块的纹理高度相关这一特性,当前图像块中的像素预测将沿一特定角度投射到参考像素区域,参考像素区域内对应的1/32精度的像素值作为当前图像块中的像素的预测值。参考像素区域1/32精度的像素值可以使用临近的两个参考像素进行内插。若投射对应的参考像素恰好是参考区域的整像素,则可以直接复制对应的参考像素值。
本发明实施例在确定DC分量的预测值时,可以根据预测模式指定的方向选择像素信号,并且根据像素信号的纹理相关性来确定DC分量的预测值。下面详细说明三种确定DC分量的预测值的方法。
图14是根据本发明的实施例的确定DC分量的预测值的过程的示意性流程图。
具体来说,图8的850和图9的步骤930可以包括如下步骤:
1410,根据预测模式指示的方向,在过渡重建图像块和参数像素区中沿第k条线选取多个相邻的像素信号;
例如,从码流中获取目标图像块的帧内预测模式predmode,沿着该帧内预测模式所指定的方向,分别在过渡重建图像块和过渡重建图像块周围的参考像素区域取四个相邻的像素信号。其中,所述的第k条线并非真实存在的线,而是便于描述假想的穿过上述的四个相邻的像素且与所述帧内预测模式所指的方向一致的射线,与所述帧内预测模式 所指的方向一致是指的方向大致平行或者平行。应理解,本发明的实施例在提到线穿过的像素时可以指线穿过实际像素,也可以指线穿过实际像素经插值得到的虚拟像素。
图11A是本发明实施例的基于方向预测模式选择像素信号的示意图。图11B是基于DC预测模式和平面预测模式选择像素信号的示意图。如图11A所示,在第k条线上在参考像素区域选取一组像素信号,包括:
Figure PCTCN2017071602-appb-000015
四个信号,在过渡重建图像块中选取一组像素信号,包括:
Figure PCTCN2017071602-appb-000016
四个信号。其中下标ik表示沿着预测方向的第k条线,j表示该线上的第j像素信号。
Figure PCTCN2017071602-appb-000017
Figure PCTCN2017071602-appb-000018
相邻,其它几个信号按编号的顺序依次排列。对于方向预测模式,按照预测方向选取像素信号;而对于DC预测模式和平面预测模式,可以只考虑水平方向和垂直方向的相邻像素的相关性,因此在水平方向和垂直方向上选择像素信号。在本发明的实施例中,方向预测模式指定的方向可以指预测方向,而DC预测模式和平面预测模式指定的方向可以指水平方向和垂直方向,k为1与2n-1之间的整数。
应理解,为了描述方便,本发明的实施例以在参考像素区和过渡重建图像块各选取四个信号为例进行说明,本发明的实施例并不限于此,所选取的信号的个数可以是其它值,例如,大于4或等于3的值。
1420,计算第k条线上选取的两组像素信号的二阶梯度,根据两组像素的二阶度确定具有纹理方向性的线的集合。
每条线上的选择的两组像素信号的二阶度的计算方法如下:
Figure PCTCN2017071602-appb-000019
可以根据二阶梯度的大小判断第ik条线在预测方向上是否具有较强的纹理特性,当二阶梯度值较小(例如,小于某个预设值)时,表明信号在该方向上是一致的,也即具有较强的纹理特性。本实施例是以上述求二阶梯度的方法为例进行描述,但本发明的实施例并不限于此,其它任何求二阶梯度的方法均可以用于本发明的实施例。
在具体实现中,使用下列公式(2)来判断第ik条线在预测方向上是否具有较强的纹理特性。特别地,当过渡重建图像块中找不到用于计算二阶梯度的信号,并且在参考区域能够找到用于计算二阶梯度的信号时,例如,使用下列公式(3)。
Figure PCTCN2017071602-appb-000020
Figure PCTCN2017071602-appb-000021
其中λ1和λ2的取值,可以根据实验经验使用λ1=40,λ2=20。若第ik条线满足公式(2)或者(3),则将该条线加入到集合C中,集合C表示具有纹理方向性的线的集合。
应理解,本发明的实施例是以二阶梯度来判断纹理特性,本发明的实施例并不限于此,也可以采用其它方法来判断每条线上的选取的信号是否具有较强的纹理特性,例如,一阶梯度等。当然,判断纹理特性的方式也可以是多种方式的结合,例如,对于有些线上选取的信号采用二阶梯度,对于另一些上选择的信号采用一阶梯度。
参见图10,除方向2、10、18、26和34之外,其他方向上的预测模式均用邻近的两个像素做加权平均来获得上述像素信号
Figure PCTCN2017071602-appb-000022
Figure PCTCN2017071602-appb-000023
换句话 说,上述两组像素信号可以是上述线穿过的像素本身的信号,也可以是上述线两侧的像素的信号进行插值得到的信号。
下面详细描述不同预测模式下获得上述
Figure PCTCN2017071602-appb-000024
Figure PCTCN2017071602-appb-000025
信号的方法。为了描述方便,以下图中的角标表示的是信号的水平和竖直方向的索引,与上述角标有所不同。图12A-12D是本发明的实施例的基于方向预测模式选取像素信号的示意图。
参见图12A,对于预测模式3-9的插值,可以使用过渡重建图像块左边和左下角的块作为参考,以xi,1为起点沿着预测方向(图10中指示预测方向)作一条射线,插值像素点y'i,j由相邻的两个原始像素点加权平均得到。预测模式4-9与之类似,在此不再赘述。
Figure PCTCN2017071602-appb-000026
其中
Figure PCTCN2017071602-appb-000027
Figure PCTCN2017071602-appb-000028
为插值点距离上侧像素点的距离,
Figure PCTCN2017071602-appb-000029
为插值点距离下侧像素点的距离。其余的插值点以此类推。
参见图12B,对于预测模式11-17的插值,使用过渡重建图像块左边和左上角的块作为参考,以xi,1为起点沿着预测方向(图10中指示的预测方向)作一条射线,插值像素点y′i,j由相邻的两个原始像素点加权平均得到。预测模式12-17与之类似,在此不再赘述。
Figure PCTCN2017071602-appb-000030
其中
Figure PCTCN2017071602-appb-000031
Figure PCTCN2017071602-appb-000032
为插值点距离上侧像素点的距离,
Figure PCTCN2017071602-appb-000033
为插值点距离下侧像素点的距离。其余点的插值以此类推。
参见图12C,对于预测模式19-25的插值,使用过渡重建图像块上边和左上角的块作为参考,以x1,i为起点沿着预测方向作一条射线,插值像素点y′i,j由相邻的两个原始像素点加权平均得到。模式20-25与之类似,在此不再赘述。
Figure PCTCN2017071602-appb-000034
其中
Figure PCTCN2017071602-appb-000035
Figure PCTCN2017071602-appb-000036
为插值点距离左侧像素点的距离,
Figure PCTCN2017071602-appb-000037
为插值点距离右侧像素点的距离。其余的插值点以此类推。
参见图12D,对于方向预测模式27-33的插值,可以使用过渡重建图像块上边和右上角的块作为参考,以x1,i为起点沿着预测方向作一条射线,插值像素点y′i,j由相邻的两个原始像素点加权平均得到。模式28-33与之类似,在此不再赘述。
Figure PCTCN2017071602-appb-000038
其中
Figure PCTCN2017071602-appb-000039
Figure PCTCN2017071602-appb-000040
为插值点距离左侧像素点的距离,
Figure PCTCN2017071602-appb-000041
为插值点距离右侧像素点的距离。其余的插值点以此类推。
应理解,插值点的选择并不限于该射线上位于上侧像素点与下侧像素点之间的点,也可以选择位于该射线上位于左侧像素点与右侧像素点之间的点,相应地,可以根据左侧的像素点和右侧的像素点进行插值得到y'i,j
对于过渡重建图像块内部的像素点,在图12A至图12D中的射线反向延长线上如果能找到4个相邻像素点或通过插值得到的像素点,则选取过渡重建图像块内部的4个相邻像素点;如果不能找到4个相邻像素点,则只选取参考块或参考像素区域的像素点。
1430,求解第一像素信号的偏移量,使得集合C中的线上的第一像素信号增加偏移量后的重建信号的二阶梯度与第二像素信号的二阶梯度的平方和最小,将过渡重建图像块内部的信号增加偏移量δx,得到具有DC分量的重建信号。该偏移量表示过渡重建图像块的量化前的DC预测值。当偏移量δx是实际的量化前的DC分量时,
Figure PCTCN2017071602-appb-000042
即为实际的重建信号。
根据图像信号在局部区域具有较强相关性或相似性的特性,即
Figure PCTCN2017071602-appb-000043
Figure PCTCN2017071602-appb-000044
具有较强的相关性,来计算过渡重建图像块的DC预测值。相关性或相似性使用二阶梯度来衡量,具体见下例公式(11),该问题转化为求δx,使得该方向上的信号的二阶梯度的平方和最小。该凸优化(Convex optimization)问题是一个二次函数,通过计算公式(12)即可求得δx的值。
Figure PCTCN2017071602-appb-000045
Figure PCTCN2017071602-appb-000046
Figure PCTCN2017071602-appb-000047
Figure PCTCN2017071602-appb-000048
Figure PCTCN2017071602-appb-000049
其中q={1,2},k∈C。q为上述集合C中的每条线上的像素信号的编号。
由于二阶梯度能够更精确地体现目标图像块的像素与参考图像块之间的相关性或相似性,因此,基于二阶梯度得到的DC分量的预测值的精度更高,使得DC分量残差更小,从而进一步提高了编码效率。
1440,对求解得到的偏移量的值进行量化,得到量化后的DC分量的预测值。
例如,量化后的DC分量的预测值为DCpred=Q(m·δx),其中,Q(·)是量化操作,m为目标图像块中像素的个数。
应理解,作为图14的实施例的替代实施例,也可以直接按照预测模式指定的方向选取过渡重建图像块和参考区域相邻的两个像素信号组成多个像素对,并根据公式(13)和公式(14)来确定DC分量预测值。具体过程参见方法一中的描述,在此不赘述。
由于二阶梯度能够更精确地体现目标图像块的像素与参考图像块之间的相关性或相似性,因此,基于二阶梯度得到的DC分量的预测值的精度更高,使得DC分量残差更小,从而进一步提高了编码效率。另外,由于只选取相关性较强的某些线上的像素用于计算DC分量的预测值,因此,DC分量的预测值的精度得到进一步提高,使得DC分量残差更小,从而进一步提高了编码效率。
图15是根据本发明的另一实施例的确定DC分量的预测值的过程的示意性流程图。
具体来说,图8的步骤850和图9的步骤930可以包括如下步骤:
1510,根据预测模式指示的方向,在过渡重建图像块和参数像素区中确定多个像素对以及对应于每个像素对的第一像素信号和相邻的第二像素信号,第一像素信号为过渡重建图像块的像素信号,第二像素信号为参考像素区域的像素信号。
图13A和图13B为根据本发明的另一实施例的选取像素信号的示意图。如图13A和图13B所示,过渡重建图像块取信号
Figure PCTCN2017071602-appb-000050
参考区域取信号
Figure PCTCN2017071602-appb-000051
对于方向预测模式,可以按照预测方向选取像素信号;而对于DC预测平面预测模式,可以只考虑垂直和水平方向的像素的相关性,即按照水平方向和垂直方向选取像素信号。
1520,求解第一像素信号的偏移量,使得多个像素对的第一像素信号增加偏移后的重建信号与第二像素信号的一阶梯度的平方和最小,其中其中该偏移量用来表示DC分量的量化前的预测值。
对于方向预测模式,可以按照预测方向选取像素信号;而对于DC预测平面预测模式,可以只考虑垂直和水平方向的像素的相关性,即按照水平方向和垂直方向选取像素信号。同上述集合C不为空时的过程类似,对于过渡重建图像块内的信号
Figure PCTCN2017071602-appb-000052
现将其加上一个偏移量δx。
根据图像信号在局部区域具有较强相关性的特性,即
Figure PCTCN2017071602-appb-000053
Figure PCTCN2017071602-appb-000054
具有较强的相关性,来计算目标图像块的DC分量的预测值。相关性可以使用一阶梯度来衡量,具体见公式(13),该问题转化为求δx,使得该方向上的信号的一阶梯度平方和最小。该凸优化问题是一个二次函数,仅需要计算公式(14)即可求得δx的值。
Figure PCTCN2017071602-appb-000055
Figure PCTCN2017071602-appb-000056
1530,将求解得到的偏移量的值进行量化得到DC分量的预测值。
应理解,可以将图14的实施例与图15的实施例相结合。例如,如果通过图14的1420没有找到具有纹理方向特性的线,即集合C为空,则按照图15的方案选取过渡重建图像块与参考区域相邻的两个像素来确定DC分量的预测值。
由于只选取过渡重建图像块和参考像素区域之间的相邻像素用于计算DC分量的预测值,因此,在保证DC分量预测值的精度的情况下,降低了计算复杂度。
图16是根据本发明的另一实施例的确定DC分量的预测值的过程的示意性流程图。
具体来说,图8的步骤850和图9的步骤930可以包括如下步骤:
1610,选取过渡重建图像块的上方参考像素区域中的M行像素,和过渡重建图像块的左侧的参考像素区域中的H列像素,分别计算M行像素和H列像素的平均值Ave(UpM)和Ave(LeftH)。
1620,选取过渡重建图像块内部上侧的M行像素和过渡重建图像块内部左侧的H列像素,分别计算两部分像素的平均值Ave(CurM)和Ave(CurH)。
其中M和H可以大于或等于2,且小于过渡重建图像块的大小N。M和H可以相等,也可以不等。
1630,计算过渡重建图像块上方的参考像素区域中的M行像素的均值与过渡重建图像块内部上侧的M行像素的均值的差,得到第一差值
Figure PCTCN2017071602-appb-000057
其中
Figure PCTCN2017071602-appb-000058
1640,计算过渡重建图像块左侧的参考像素区域中的H列像素的均值与过渡重建图像块内部左侧的H列像素的均值的差,得到第二差值
Figure PCTCN2017071602-appb-000059
其中
Figure PCTCN2017071602-appb-000060
1650,将
Figure PCTCN2017071602-appb-000061
Figure PCTCN2017071602-appb-000062
的均值
Figure PCTCN2017071602-appb-000063
作为目标图像块的DC分量量化前的预测值,则目标图像块量化后的DC分量的预测值为
Figure PCTCN2017071602-appb-000064
m为目标图像块的个数。
本发明实施例可以应用于各种电子装置中,示例性的,下面给出本发明实施例应用于电视设备和移动电话设备的例子。
由于选取多行或多列过渡重建图像块和参考像素区域的像素用于计算DC分量的预测值,因此,在保证DC分量预测值的精度的情况下,降低了计算复杂度。采用平均值来计算DC分量的预测值,使得编码器或解码器的设计简单。另外,当M和H为多行或多列时,能够充分利用目标图像块周围更多信号的相关性以及目标图像块与周围信号的关系来预测DC分量的预测值,使得DC分量残差更小,提高了DC分量的预测值的精度,从而进一步提高了编码效率。
此外,本法明还包括在实施DC预测之前,首先对待编码图像的DC值的大小进行判断,如果DC值大于预设阈值,则执行本发明上述的方法,如果DC值小于预设的阈值,则采用常规的方法进行处理,即不采用上述的DC预测方法,而是直接对目标图像 块的DC分量及AC分量进行编码。而对应于解码端,本发明可以通过在slice header,PPS,SPS,等参数中增加语法元素,比如一个flag来指示解码端采用何种方式来获取目标图像块的DC分量及AC分量,比如DC_pred_present flag,当其为1时表示,DC采用预测的方式获取,当其为零时表示DC采用直接解析码流的方式获取,或者也可以相反。
图17是本发明实施例适用于电视机应用的示意性结构图。电视设备1700包括天线1701、调谐器1702、多路解复用器1703、解码器1704、视频信号处理器1705、显示单元1706、音频信号处理器1707、扬声器1708、外部接口1709、控制器1710、用户接口1711和总线1712。
调谐器1702从经天线1701接收到的广播信号提取期望频道的信号,并且解调提取的信号。调谐器1702随后将通过解调获得的编码比特流输出到多路解复用器1703。也就是说,调谐器1702在接收编码图像的编码流的电视设备1700中用作发送装置。
多路解复用器1703从编码比特流分离将要观看的节目的视频流和音频流,并且将分离的流输出到解码器1704。多路解复用器1703还从编码比特流提取辅助数据,例如,电子节目指南,并且将提取的数据提供给控制器1710。如果编码比特流被加扰,则多路解复用器703可对编码比特流进行解扰。
解码器1704对从多路解复用器1703输入的视频流和音频流进行解码。解码器1704随后将通过解码产生的视频数据输出到视频信号处理器1705。解码器1704还将通过解码产生的音频数据输出到音频信号处理器1707。
视频信号处理器1705再现从解码器1704输入的视频数据,并且在显示单元1706上显示视频数据。视频信号处理器1705还可在显示单元1706上显示经网络提供的应用画面。另外,视频信号处理器1705可根据设置对视频数据执行额外的处理,例如,噪声去除。视频信号处理器1705还可产生GUI(图形用户界面)的图像并且将产生的图像叠加在输出图像上。
显示单元1706由从视频信号处理器1705提供的驱动信号驱动,并且在显示装置,例如,液晶显示器、等离子体显示器或OELD(有机场致发光显示器)的视频屏幕上显示视频或图像。
音频信号处理器1707对从解码器1704输入的音频数据执行再现处理,例如,数模转换和放大,并且通过扬声器1708输出音频。另外,音频信号处理器1707可对音频数据执行额外的处理,例如,噪声去除。
外部接口1709是用于连接电视设备1700与外部装置或网络的接口。例如,经外部接口1709接收的视频流或音频流可由解码器1704解码。也就是说,外部接口1709也在接收编码图像的编码流的电视设备1700中用作发送装置。
控制器1710包括处理器和存储器。存储器存储将要由处理器执行的程序、节目数据、辅助数据、经网络获取的数据等。例如,当电视设备900启动时,存储在存储器中的程序由处理器读取并且执行。处理器根据从用户接口1711输入的控制信号控制电视设备1700的操作。
用户接口1711连接到控制器1710。例如,用户接口1711包括用于使用户操作电视设备1700的按钮和开关以及用于接收遥控信号的接收单元。用户接口1711检测由用户经这些部件执行的操作,产生控制信号,并且将产生的控制信号输出到控制器1710。
总线1712将调谐器1702、多路解复用器1703、解码器1704、视频信号处理器1705、 音频信号处理器1707、外部接口1709和控制器1710彼此连接。
在具有这种结构的电视设备1700中,解码器1704具有根据上述实施例的视频解码装置的功能。
图18是本发明实施例适用于移动电话应用的示意性结构图。移动电话装置1720包括天线1721、通信单元1722、音频编解码器1723、扬声器1724、麦克风1725、相机单元1726、图像处理器1727、多路解复用器1728、记录/再现单元1729、显示单元1730、控制器1731、操作单元1732和总线1733。
天线1721连接到通信单元1722。扬声器1724和麦克风1725连接到音频编码解码器1723。操作单元932连接到控制器1731。总线1733将通信单元1722、音频编解码器1723、相机单元1726、图像处理器1727、多路解复用器1728、记录/再现单元1729、显示单元1730和控制器1731彼此连接。
移动电话装置1720在各种操作模式下执行操作,例如,音频信号的发送/接收、电子邮件和图像数据的发送/接收、图像的拍摄、数据的记录等,所述各种操作模式包括语音呼叫模式、数据通信模式、成像模式和视频电话模式。
在语音呼叫模式下,由麦克风1725产生的模拟音频信号被提供给音频编解码器1723。音频编解码器1723将模拟音频信号转换成音频数据,对转换的音频数据执行模数转换,并且压缩音频数据。音频编码解码器1723随后将作为压缩结果得到的音频数据输出到通信单元1722。通信单元1722对音频数据进行编码和调制以产生待发送的信号。通信单元1722随后经天线1721将产生的待发送的信号发送给基站。通信单元1722还放大经天线1721接收到的无线电信号并且对经天线1721接收到的无线电信号执行频率转换以获得接收到的信号。通信单元1722随后对接收到的信号进行解调和解码以产生音频数据,并且将产生的音频数据输出到音频编解码器1723。音频编解码器1723解压缩音频数据并且对音频数据执行数模转换以产生模拟音频信号。音频编码解码器1723随后将产生的音频信号提供给扬声器1724以从扬声器1724输出音频。
在数据通信模式下,例如,控制器1731根据由用户经操作单元1732的操作产生将要被包括在电子邮件中的文本数据。控制器1731还在显示单元1730上显示文本。控制器1731还响应于经操作单元1732来自用户的用于发送的指令产生电子邮件数据,并且将产生的电子邮件数据输出到通信单元1722。通信单元17922对电子邮件数据进行编码和调制以产生待发送的信号。通信单元1722随后经天线1721将产生的待发送的信号发送给基站。通信单元1722还放大经天线1721接收到的无线电信号并且对经天线1721接收到的无线电信号执行频率转换以获得接收到的信号。通信单元1722随后对接收到的信号进行解调和解码以恢复电子邮件数据,并且将恢复的电子邮件数据输出到控制器1731。控制器1731在显示单元1730上显示电子邮件的内容,并且将电子邮件数据存储在记录/再现单元1729的存储介质中。
记录/再现单元1729包括可读/可写存储介质。例如,存储介质可以是内部存储介质,或者可以是在外部安装的存储介质,例如,硬盘、磁盘、磁光盘、USB(通用串行总线)存储器或存储卡。
在成像模式下,相机单元1726对对象成像以产生图像数据,并且将产生的图像数据输出到图像处理器1727。图像处理器1727对从相机单元1726输入的图像数据进行编码,并且将编码流存储在存储/再现单元1729的存储介质中。
在视频电话模式下,多路解复用器1728多路复用由图像处理器1727编码的视频流和从音频编码解码器1723输入的音频流,并且将多路复用流输出到通信单元1722。通信单元1722对多路复用流进行编码和调制以产生待发送的信号。通信单元1722随后经天线1721将产生的待发送的信号发送给基站。通信单元1722还放大经天线1721接收到的无线电信号并且对经天线1721接收到的无线电信号执行频率转换以获得接收到的信号。待发送的信号和接收到的信号可包括编码比特流。通信单元1722随后对接收到的信号进行解调和解码以恢复流,并且将恢复的流输出到多路解复用器1728。多路解复用器1728从输入流分离视频流和音频流,将视频流输出到图像处理器1727并且将音频流输出到音频编解码器1723。图像处理器1727对视频流进行解码以产生视频数据。视频数据被提供给显示单元1730,并且一系列图像由显示单元1730显示。音频编解码器1723解压缩音频流并且对音频流执行数模转换以产生模拟音频信号。音频编解码器1723随后将产生的音频信号提供给扬声器1724以从扬声器1724输出音频。
在具有这种结构的移动电话装置1720中,图像处理器1727具有根据上述实施例的视频编码装置和视频解码装置的功能。
在一个或多个实例中,所描述的功能可以硬件、软件、固件或其任何组合来实施。如果以软件实施,则功能可作为一个或多个指令或代码而存储于计算机可读媒体上或经由计算机可读媒体而发送,且通过基于硬件的处理单元执行。计算机可读媒体可包含计算机可读存储媒体(其对应于例如数据存储媒体等有形媒体)或通信媒体,通信媒体包含(例如)根据通信协议促进计算机程序从一处传送到另一处的任何媒体。以此方式,计算机可读媒体大体上可对应于(1)非瞬时的有形计算机可读存储媒体,或(2)例如信号或载波等通信媒体。数据存储媒体可为可由一个或多个计算机或一个或多个处理器存取以检索指令、代码及/或数据结构以用于实施本发明中所描述的技术的任何可用媒体。计算机程序产品可包含计算机可读媒体。
通过实例而非限制,某些计算机可读存储媒体可包括RAM、ROM、EEPROM、CD-ROM或其它光盘存储器、磁盘存储器或其它磁性存储装置、快闪存储器,或可用以存储呈指令或数据结构的形式的所要程序代码且可由计算机存取的任何其它媒体。而且,任何连接可适当地称为计算机可读媒体。举例来说,如果使用同轴电缆、光缆、双绞线、数字用户线(DSL)或无线技术(例如,红外线、无线电及微波)而从网站、服务器或其它远程源发送指令,则同轴电缆、光缆、双绞线、DSL或无线技术(例如,红外线、无线电及微波)包含于媒体的定义中。然而,应理解,计算机可读存储媒体及数据存储媒体不包含连接、载波、信号或其它瞬时媒体,而是有关非瞬时有形存储媒体。如本文中所使用,磁盘及光盘包含压缩光盘(CD)、激光光盘、光学光盘、数字影音光盘(DVD)、软性磁盘及蓝光光盘,其中磁盘通常以磁性方式复制数据,而光盘通过激光以光学方式复制数据。以上各物的组合还应包含于计算机可读媒体的范围内。
可由例如一个或多个数字信号处理器(DSP)、通用微处理器、专用集成电路(ASIC)、现场可编程逻辑阵列(FPGA)或其它等效集成或离散逻辑电路等一个或多个处理器来执行指令。因此,如本文中所使用的术语“处理器”可指代前述结构或适于实施本文中所描述的技术的任何其它结构中的任一者。另外,在一些方面中,可将本文中所描述的功能性提供于经配置以用于编码及解码的专用硬件及/或软件模块内,或并入于组合式编解码器中。而且,所述技术可完全实施于一个或多个电路或逻辑元件中。
本发明的技术可以广泛地由多种装置或设备来实施,所述装置或设备包含无线手持机、集成电路(IC)或IC集合(例如,芯片组)。在本发明中描述各种组件、模块或单元以强调经配置以执行所揭示技术的装置的功能方面,但未必要求通过不同硬件单元来实现。确切地说,如上文所描述,各种单元可组合于编解码器硬件单元中,或通过交互操作性硬件单元(包含如上文所描述的一个或多个处理器)的集合结合合适软件及/或固件来提供。
应理解,说明书通篇中提到的“一个实施例”或“一实施例”意味着与实施例有关的特定特征、结构或特性包括在本发明的至少一个实施例中。因此,在整个说明书各处出现的“在一个实施例中”或“在一实施例中”未必一定指相同的实施例。此外,这些特定的特征、结构或特性可以任意适合的方式结合在一个或多个实施例中。
在本发明的各种实施例中,应理解,上述各过程的序号的大小并不意味着执行顺序的先后,各过程的执行顺序应以其功能和内在逻辑确定,而不应对本发明实施例的实施过程构成任何限定。
另外,本文中术语“系统”和“网络”在本文中常可互换使用。应理解,本文中术语“和/或”,仅仅是一种描述关联对象的关联关系,表示可以存在三种关系,例如,A和/或B,可以表示:单独存在A,同时存在A和B,单独存在B这三种情况。另外,本文中字符“/”,一般表示前后关联对象是一种“或”的关系。
在本申请所提供的实施例中,应理解,“与A相应的B”表示B与A相关联,根据A可以确定B。但还应理解,根据A确定B并不意味着仅仅根据A确定B,还可以根据A和/或其它信息确定B。
本领域普通技术人员可以意识到,结合本文中所公开的实施例描述的各示例的单元及算法步骤,能够以电子硬件、计算机软件或者二者的结合来实现,为了清楚地说明硬件和软件的可互换性,在上述说明中已经按照功能一般性地描述了各示例的组成及步骤。这些功能究竟以硬件还是软件方式来执行,取决于技术方案的特定应用和设计约束条件。专业技术人员可以对每个特定的应用来使用不同方法来实现所描述的功能,但是这种实现不应认为超出本发明的范围。
所属领域的技术人员可以清楚地了解到,为描述的方便和简洁,上述描述的系统、装置和单元的具体工作过程,可以参考前述方法实施例中的对应过程,在此不再赘述。
在本申请所提供的几个实施例中,应该理解到,所揭露的系统、装置和方法,可以通过其它的方式实现。例如,以上所描述的装置实施例仅仅是示意性的,例如,所述单元的划分,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分方式,例如多个单元或组件可以结合或者可以集成到另一个系统,或一些特征可以忽略,或不执行。另一点,所显示或讨论的相互之间的耦合或直接耦合或通信连接可以是通过一些接口,装置或单元的间接耦合或通信连接,可以是电性,机械或其它的形式。
所述作为分离部件说明的单元可以是或者也可以不是物理上分开的,作为单元显示的部件可以是或者也可以不是物理单元,即可以位于一个地方,或者也可以分布到多个网络单元上。可以根据实际的需要选择其中的部分或者全部单元来实现本实施例方案的目的。
另外,在本发明各个实施例中的各功能单元可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中。
以上所述,仅为本发明的具体实施方式,但本发明的保护范围并不局限于此,任何熟悉本技术领域的技术人员在本发明揭露的技术范围内,可轻易想到变化或替换,都应涵盖在本发明的保护范围之内。因此,本发明的保护范围应以所述权利要求的保护范围为准。
本领域普通技术人员可以意识到,结合本文中所公开的实施例描述的各示例的单元及算法步骤,能够以电子硬件、计算机软件或者二者的结合来实现,为了清楚地说明硬件和软件的可互换性,在上述说明中已经按照功能一般性地描述了各示例的组成及步骤。这些功能究竟以硬件还是软件方式来执行,取决于技术方案的特定应用和设计约束条件。专业技术人员可以对每个特定的应用来使用不同方法来实现所描述的功能,但是这种实现不应认为超出本发明的范围。
所属领域的技术人员可以清楚地了解到,为描述的方便和简洁,上述描述的系统、装置和单元的具体工作过程,可以参考前述方法实施例中的对应过程,在此不再赘述。
在本申请所提供的几个实施例中,应该理解到,所揭露的系统、装置和方法,可以通过其它的方式实现。例如,以上所描述的装置实施例仅仅是示意性的,例如,所述单元的划分,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分方式,例如多个单元或组件可以结合或者可以集成到另一个系统,或一些特征可以忽略,或不执行。另一点,所显示或讨论的相互之间的耦合或直接耦合或通信连接可以是通过一些接口,装置或单元的间接耦合或通信连接,可以是电性,机械或其它的形式。
所述作为分离部件说明的单元可以是或者也可以不是物理上分开的,作为单元显示的部件可以是或者也可以不是物理单元,即可以位于一个地方,或者也可以分布到多个网络单元上。可以根据实际的需要选择其中的部分或者全部单元来实现本实施例方案的目的。
另外,在本发明各个实施例中的各功能单元可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中。上述集成的单元既可以采用硬件的形式实现,也可以采用软件功能单元的形式实现。
所述集成的单元如果以软件功能单元的形式实现并作为独立的产品销售或使用时,可以存储在一个计算机可读取存储介质中。基于这样的理解,本发明的技术方案本质上或者说对现有技术做出贡献的部分或者该技术方案的全部或部分可以以软件产品的形式体现出来,该计算机软件产品存储在一个存储介质中,包括若干指令用以使得一台计算机设备(可以是个人计算机,服务器,或者网络设备等)执行本发明各个实施例所述方法的全部或部分步骤。而前述的存储介质包括:U盘、移动硬盘、只读存储器(ROM,Read-Only Memory)、随机存取存储器(RAM,Random Access Memory)、磁碟或者光盘等各种可以存储程序代码的介质。
以上所述,仅为本发明的具体实施方式,但本发明的保护范围并不局限于此,任何熟悉本技术领域的技术人员在本发明揭露的技术范围内,可轻易想到变化或替换,都应涵盖在本发明的保护范围之内。因此,本发明的保护范围应以所述权利要求的保护范围为准。

Claims (26)

  1. 一种解码方法,其特征在于,所述解码方法包括:
    从码流中获取目标图像块的变换量化系数的交流AC分量及直流DC分量残差;
    对所述AC分量以及预设的DC分量进行逆量化和逆变换,得到过渡残差,并将所述过渡残差与目标图像块的预测值相加得到所述目标图像块的过渡重建图像块;
    根据所述过渡重建图像块与所述的目标图像块的参考像素区域中的像素的相似性,确定所述目标图像块的变换量化系数的DC分量的预测值;
    根据所述DC分量的预测值和所述DC分量残差确定所述目标图像的原始DC分量;
    对所述目标图像的原始DC分量以及AC分量进行逆量化和逆变换,得到所述目标图像的残差信号,并根据所残差信号对所述目标图像进行解码。
  2. 根据权利要求1所述的解码方法,其特征在于,所述根据所述过渡重建图像块与所述的目标图像块的参考像素区域中的像素的相似性,确定所述目标图像块的变换量化系数的DC分量的预测值,包括:
    在所述目标图像块的预测模式指定的方向上,确定至少一条线以及对应于每条线的第一组像素信号和相邻的第二组像素信号,其中所述第一组像素信号包括所述过渡重建图像块的第一像素信号,所述第二组像素信号包括所述参考像素区域的第二像素信号,所述第一像素信号与所述第二像素信号相邻;
    求解所述第一像素信号的偏移量,使得所述至少一条线中所述第一像素信号增加所述偏移量后的重建信号的二阶梯度与所述第二像素信号的二阶梯度的平方和最小,其中所述过渡重建图像块中用于表示所述二阶梯度的像素信号增加所述偏移量,该偏移量用来表示所述DC分量的量化前的预测值;
    将求解得到的偏移量的值进行量化得到所述DC分量的预测值。
  3. 根据权利要求2所述的解码方法,其特征在于,对应于所述至少一条线中的每条线的第一组像素信号和第二组像素信号满足如下两个公式和公式之一:
    Figure PCTCN2017071602-appb-100001
    Figure PCTCN2017071602-appb-100002
    其中,λ1用于表示所述第一组像素信号的重建信号存在二阶梯度时的阈值,λ2用于表示所述第一组像素信号的重建信号不存在二阶梯度时的阈值,
    Figure PCTCN2017071602-appb-100003
    为所述第一组像素信号的二阶梯度,
    Figure PCTCN2017071602-appb-100004
    为所述第二组像素信号的二阶梯度,ik为所述至少一条线的编号,j为所述至少一条线中的每条线上的像素信号的编号,
    其中,所述偏移量根据如下公式来计算:
    Figure PCTCN2017071602-appb-100005
    Figure PCTCN2017071602-appb-100006
    其中,δx为所述偏移量,
    Figure PCTCN2017071602-appb-100007
    为所述第一像素信号的重建信号的二阶梯度,
    Figure PCTCN2017071602-appb-100008
    为所述第二像素信号的二阶梯度,C表示满足公所述两个公式之一线的集合,q为所述至少一条线中的每条线上的像素信号的编号。
  4. 根据权利要求1所述的解码方法,其特征在于,所述根据所述根据所述过渡重 建图像块与所述的目标图像块的参考像素区域中的像素的相似性,确定所述目标图像块的变换量化系数的DC分量的预测值,包括:
    在所述目标图像块的预测模式指定的方向上,确定多个像素对以及对应于每个像素对的第一像素信号和相邻的第二像素信号,其中所述第一像素信号为所述过渡重建图像块的像素信号,所述第二像素信号为所述参考像素区域的像素信号;
    求解所述第一像素信号的偏移量,使得所述多个像素对的第一像素信号增加所述偏移后的重建信号与第二像素信号的一阶梯度的平方和最小,其中其中所述偏移量用来表示所述DC分量的量化前的预测值;
    将求解得到的偏移量的值进行量化得到所述DC分量的预测值。
  5. 根据权利要求4所述的解码方法,其特征在于,所述DC分量的预测值根据如下公式来计算:
    Figure PCTCN2017071602-appb-100009
    其中δx为所述偏移量,n为所述过渡重建图像块的每一行或列的像素的个数,
    Figure PCTCN2017071602-appb-100010
    为所述第一像素信号,
    Figure PCTCN2017071602-appb-100011
    为所述第二像素信号,
    Figure PCTCN2017071602-appb-100012
    Figure PCTCN2017071602-appb-100013
    在所述预测模式指定的方向上相邻。
  6. 根据权利要求1所述的解码方法,其特征在于,所述根据所述过渡重建图像块与所述的目标图像块的参考像素区域中的像素的相似性,确定所述目标图像块的变换量化系数的DC分量的预测值,包括:
    获取所述参考像素区域中位于所述过渡重建图像块上方的第一组像素信号、所述过渡重建图像块左侧的第二组像素信号、所述过渡重建图像块内部上侧的第三组像素信号和所述过渡重建图像块内部左侧的第四组像素信号,所述第一组像素信号和所述第三组像素信号分别包括M行像素信号,所述第二组像素信号和所述第四组像素信号分别包括H列像素信号,所述M和H为正整数;
    计算所述第一组像素信号的平均值与所述第三组像素信号的平均值之差,得到第一差值;
    计算所述第二组像素信号的平均值与所述第四组像素信号的平均值之差,得到第二差值;
    将所述第一差值与第二差值的平均值进行量化得到所述DC分量的预测值。
  7. 根据权利要求所述的解码方法,其特征在于,所述M为大于或等于2的整数,所述N为大于或等于2的整数。
  8. 一种编码方法,其特征在于,包括:
    对目标图像块的残差信号进行变换和量化,得到所述目标图像块的变换量化系数的直流DC分量和交流AC分量;
    对所述AC分量以及预设的DC分量进行逆量化和逆变换,得到过渡残差,并将所述过渡残差与目标图像块的预测值相加得到所述目标图像块的过渡重建图像块;
    根据所述过渡重建图像块的像素与所述目标图像块的参考像素区域中的像素的相似性,确定所述目标图像块的变换量化系数的DC分量的预测值;
    根据所述DC分量的预测值和所述目标图像块的原始DC分量确定所述目标图像的DC分量残差;
    将所述AC分量和所述DC分量残差写入码流。
  9. 根据权利要求8所述的编码方法,其特征在于,所述根据所述过渡重建图像块的像素与所述目标图像块的参考像素区域中的像素的相似性,确定所述目标图像块的变换量化系数的DC分量的预测值,包括:
    在所述目标图像块的预测模式指定的方向上,确定至少一条线以及对应于每条线的第一组像素信号和相邻的第二组像素信号,其中所述第一组像素信号包括所述过渡重建图像块的第一像素信号,所述第二组像素信号包括所述参考像素区域的第二像素信号,所述第一像素信号与所述第二像素信号相邻;
    求解所述第一像素信号的偏移量,使得所述至少一条线中所述第一像素信号增加所述偏移量后的重建信号的二阶梯度与所述第二像素信号的二阶梯度的平方和最小,其中所述过渡重建图像块中用于表示所述二阶梯度的像素信号增加所述偏移量,该偏移量用来表示所述DC分量的量化前的预测值;
    将求解得到的偏移量的值进行量化得到所述DC分量的预测值。
  10. 根据权利要求9所述的编码方法,其特征在于,对应于所述至少一条线中的每条线的第一组像素信号和第二组像素信号满足如下两个公式和公式之一:
    Figure PCTCN2017071602-appb-100014
    Figure PCTCN2017071602-appb-100015
    其中,λ1用于表示所述第一组像素信号的重建信号存在二阶梯度时的阈值,λ2用于表示所述第一组像素信号的重建信号不存在二阶梯度时的阈值,
    Figure PCTCN2017071602-appb-100016
    为所述第一组像素信号的二阶梯度,
    Figure PCTCN2017071602-appb-100017
    为所述第二组像素信号的二阶梯度,ik为所述至少一条线的编号,j为所述至少一条线中的每条线上的像素信号的编号,
    其中,所述偏移量根据如下公式来计算:
    Figure PCTCN2017071602-appb-100018
    Figure PCTCN2017071602-appb-100019
    其中,δx为所述偏移量,
    Figure PCTCN2017071602-appb-100020
    为所述第一像素信号的重建信号的二阶梯度,
    Figure PCTCN2017071602-appb-100021
    为所述第二像素信号的二阶梯度,C表示满足公所述两个公式之一线的集合,q为所述至少一条线中的每条线上的像素信号的编号。
  11. 根据权利要求8所述的编码方法,其特征在于,所述根据所述过渡重建图像块的像素与所述目标图像块的参考像素区域中的像素的相似性,确定所述目标图像块的变换量化系数的DC分量的预测值,包括:
    在所述目标图像块的预测模式指定的方向上,确定多个像素对以及对应于每个像素对的第一像素信号和相邻的第二像素信号,其中所述第一像素信号为所述过渡重建图像块的像素信号,所述第二像素信号为所述参考像素区域的像素信号;
    求解所述第一像素信号的偏移量,使得所述多个像素对的第一像素信号增加所述偏移后的重建信号与第二像素信号的一阶梯度的平方和最小,其中该偏移量用来表示所述DC分量的量化前的预测值;
    将求解得到的偏移量的值进行量化得到所述DC分量的预测值。
  12. 根据权利要求11所述的编码方法,其特征在于,所述DC分量的预测值根据如下公式来计算:
    Figure PCTCN2017071602-appb-100022
    其中δx为所述偏移量,n为所述过渡重建图像块的每一行或列的像素的个数,
    Figure PCTCN2017071602-appb-100023
    为所述第一像素信号,
    Figure PCTCN2017071602-appb-100024
    为所述第二像素信号,
    Figure PCTCN2017071602-appb-100025
    Figure PCTCN2017071602-appb-100026
    在所述预测模式指定的方向上相邻。
  13. 根据权利要求8所述的编码方法,其特征在于,所述根据所述过渡重建图像块的像素与所述目标图像块的参考像素区域中的像素的相似性,确定所述目标图像块的变换量化系数的DC分量的预测值,包括:
    获取所述参考像素区域中位于所述过渡重建图像块上方的第一组像素信号、所述过渡重建图像块左侧的第二组像素信号、所述过渡重建图像块内部上侧的第三组像素信号和所述过渡重建图像块内部左侧的第四组像素信号,所述第一组像素信号和所述第三组像素信号分别包括M行像素信号,所述第二组像素信号和所述第四组像素信号分别包括H列像素信号,所述M和H为正整数;
    计算所述第一组像素信号的平均值与所述第三组像素信号的平均值之差,得到第一差值;
    计算所述第二组像素信号的平均值与所述第四组像素信号的平均值之差,得到第二差值;
    将所述第一差值与第二差值的平均值进行量化得到所述DC分量的预测值。
  14. 根据权利要求13所述的编码方法,其特征在于,所述M为大于或等于2的整数,所述N为大于或等于2的整数。
  15. 一种解码设备,其特征在于,包括:
    熵解码模块,用于从码流中获取目标图像块的变换量化系数的交流AC分量及直流DC分量残差;
    第一逆量化和逆变换模块,用于对所述AC分量以及预设的DC分量进行逆量化和逆变换,得到过渡残差,并将所述过渡残差与目标图像块的预测值相加得到所述目标图像块的过渡重建图像块;
    得到所述目标图像块的过渡重建图像块;
    预测模块,用于根据所述过渡重建图像块与所述的目标图像块的参考像素区域中的像素的相似性,确定所述目标图像块的变换量化系数的DC分量的预测值;
    第二逆量化和逆变换模块,用于对根据所述DC分量的预测值和所述DC分量残差确定所述目标图像的原始DC分量,对所述目标图像的原始DC分量以及AC分量进行逆量化和逆变换,得到所述目标图像的残差信号,并根据所残差信号对所述目标图像进行解码。
  16. 根据权利要求15所述的解码设备,其特征在于,所述预测模块用于在所述目标图像块的预测模式指定的方向上,确定至少一条线以及对应于每条线的第一组像素信号和相邻的第二组像素信号,其中所述第一组像素信号包括所述过渡重建图像块的第一像素信号,所述第二组像素信号包括所述参考像素区域的第二像素信号,所述第一像素信号与所述第二像素信号相邻;求解所述第一像素信号的偏移量,使得所述至少一条线 中所述第一像素信号增加所述偏移量后的重建信号的二阶梯度与所述第二像素信号的二阶梯度的平方和最小,其中所述过渡重建图像块中用于表示所述二阶梯度的像素信号增加所述偏移量,该偏移量用来表示所述DC分量的量化前的预测值;将求解得到的偏移量的值进行量化得到所述DC分量的预测值。
  17. 根据权利要求16所述的解码设备,其特征在于,对应于所述至少一条线中的每条线的第一组像素信号和第二组像素信号满足如下两个公式和公式之一:
    Figure PCTCN2017071602-appb-100027
    Figure PCTCN2017071602-appb-100028
    其中,λ1用于表示所述第一组像素信号的重建信号存在二阶梯度时的阈值,λ2用于表示所述第一组像素信号的重建信号不存在二阶梯度时的阈值,
    Figure PCTCN2017071602-appb-100029
    为所述第一组像素信号的二阶梯度,
    Figure PCTCN2017071602-appb-100030
    为所述第二组像素信号的二阶梯度,ik为所述至少一条线的编号,j为所述至少一条线中的每条线上的像素信号的编号,
    其中,所述偏移量根据如下公式来计算:
    Figure PCTCN2017071602-appb-100031
    Figure PCTCN2017071602-appb-100032
    其中,δx为所述偏移量,
    Figure PCTCN2017071602-appb-100033
    为所述第一像素信号的重建信号的二阶梯度,
    Figure PCTCN2017071602-appb-100034
    为所述第二像素信号的二阶梯度,C表示满足公所述两个公式之一线的集合,q为所述至少一条线中的每条线上的像素信号的编号。
  18. 根据权利要求15所述的解码设备,其特征在于,所述预测模块在所述目标图像块的预测模式指定的方向上,确定多个像素对以及对应于每个像素对的第一像素信号和相邻的第二像素信号,其中所述第一像素信号为所述过渡重建图像块的像素信号,所述第二像素信号为所述参考像素区域的像素信号;求解所述第一像素信号的偏移量,使得所述多个像素对的第一像素信号增加所述偏移后的重建信号与第二像素信号的一阶梯度的平方和最小,其中该偏移量用来表示所述DC分量的量化前的预测值;将求解得到的偏移量的值进行量化得到所述DC分量的预测值。
  19. 根据权利要求18所述的解码设备,其特征在于,所述DC分量的预测值根据如下公式来计算:
    Figure PCTCN2017071602-appb-100035
    其中δx为所述偏移量,n为所述过渡重建图像块的每一行或列的像素的个数,
    Figure PCTCN2017071602-appb-100036
    为所述第一像素信号,
    Figure PCTCN2017071602-appb-100037
    为所述第二像素信号,
    Figure PCTCN2017071602-appb-100038
    Figure PCTCN2017071602-appb-100039
    在所述预测模式指定的方向上相邻。
  20. 根据权利要求15所述的解码设备,其特征在于,所述解码设备获取所述参考像素区域中位于所述过渡重建图像块上方的第一组像素信号、所述过渡重建图像块左侧的第二组像素信号、所述过渡重建图像块内部上侧的第三组像素信号和所述过渡重建图像块内部左侧的第四组像素信号,所述第一组像素信号和所述第三组像素信号分别包括M行像素信号,所述第二组像素信号和所述第四组像素信号分别包括H列像素信号,所 述M和H为正整数;计算所述第一组像素信号的平均值与所述第三组像素信号的平均值之差,得到第一差值;计算所述第二组像素信号的平均值与所述第四组像素信号的平均值之差,得到第二差值;将所述第一差值与第二差值的平均值进行量化得到所述DC分量的预测值。
  21. 一种编码设备,其特征在于,包括:
    变换和量化模块,用于对目标图像块的残差信号进行变换和量化,得到所述目标图像块的变换量化系数的直流DC分量和交流AC分量;
    逆量化和逆变换模块,用于对所述AC分量以及预设的DC分量进行逆量化和逆变换,得到过渡残差,并将所述过渡残差与目标图像块的预测值相加得到所述目标图像块的过渡重建图像块;
    预测模块,用于根据所述过渡重建图像块的像素与所述目标图像块的参考像素区域中的像素的相似性,确定所述目标图像块的变换量化系数的DC分量的预测值;
    熵编码模块,用于根据所述DC分量的预测值和所述目标图像块的原始DC分量确定所述目标图像的DC分量残差。
  22. 根据权利要求21所述的编码设备,其特征在于,所述预测模块在所述目标图像块的预测模式指定的方向上,确定至少一条线以及对应于每条线的第一组像素信号和相邻的第二组像素信号,其中所述第一组像素信号包括所述过渡重建图像块的第一像素信号,所述第二组像素信号包括所述参考像素区域的第二像素信号,所述第一像素信号与所述第二像素信号相邻;求解所述第一像素信号的偏移量,使得所述至少一条线中所述第一像素信号增加所述偏移量后的重建信号的二阶梯度与所述第二像素信号的二阶梯度的平方和最小,其中所述过渡重建图像块中用于表示所述二阶梯度的像素信号增加所述偏移量,该偏移量用来表示所述DC分量的量化前的预测值;将求解得到的偏移量的值进行量化得到所述DC分量的预测值。
  23. 根据权利要求22所述的编码设备,其特征在于,对应于所述至少一条线中的每条线的第一组像素信号和第二组像素信号满足如下两个公式和公式之一:
    Figure PCTCN2017071602-appb-100040
    Figure PCTCN2017071602-appb-100041
    其中,λ1用于表示所述第一组像素信号的重建信号存在二阶梯度时的阈值,λ2用于表示所述第一组像素信号的重建信号不存在二阶梯度时的阈值,
    Figure PCTCN2017071602-appb-100042
    为所述第一组像素信号的二阶梯度,
    Figure PCTCN2017071602-appb-100043
    为所述第二组像素信号的二阶梯度,ik为所述至少一条线的编号,j为所述至少一条线中的每条线上的像素信号的编号,
    其中,所述偏移量根据如下公式来计算:
    Figure PCTCN2017071602-appb-100044
    Figure PCTCN2017071602-appb-100045
    其中,δx为所述偏移量,
    Figure PCTCN2017071602-appb-100046
    为所述第一像素信号的重建信号的二阶梯度,
    Figure PCTCN2017071602-appb-100047
    为所述第二像素信号的二阶梯度,C表示满足公所述两个公式之一线的集合,q为所述至少一条线中的每条线上的像素信号的编号。
  24. 根据权利要求21所述的编码设备,其特征在于,所述预测模块在所述目标图像块的预测模式指定的方向上,确定多个像素对以及对应于每个像素对的第一像素信号和相邻的第二像素信号,其中所述第一像素信号为所述过渡重建图像块的像素信号,所述第二像素信号为所述参考像素区域的像素信号;求解所述第一像素信号的偏移量,使得所述多个像素对的第一像素信号增加所述偏移后的重建信号与第二像素信号的一阶梯度的平方和最小,其中该偏移量用来表示所述DC分量的量化前的预测值;将求解得到的偏移量的值进行量化得到所述DC分量的预测值。
  25. 根据权利要求24所述的编码设备,其特征在于,所述DC分量的预测值根据如下公式来计算:
    Figure PCTCN2017071602-appb-100048
    其中δx为所述偏移量,n为所述过渡重建图像块的每一行或列的像素的个数,
    Figure PCTCN2017071602-appb-100049
    为所述第一像素信号,
    Figure PCTCN2017071602-appb-100050
    为所述第二像素信号,
    Figure PCTCN2017071602-appb-100051
    Figure PCTCN2017071602-appb-100052
    在所述预测模式指定的方向上相邻。
  26. 根据权利要求22所述的编码设备,其特征在于,所述预测模块获取所述参考像素区域中位于所述过渡重建图像块上方的第一组像素信号、所述过渡重建图像块左侧的第二组像素信号、所述过渡重建图像块内部上侧的第三组像素信号和所述过渡重建图像块内部左侧的第四组像素信号,所述第一组像素信号和所述第三组像素信号分别包括M行像素信号,所述第二组像素信号和所述第四组像素信号分别包括H列像素信号,所述M和H为正整数;计算所述第一组像素信号的平均值与所述第三组像素信号的平均值之差,得到第一差值;计算所述第二组像素信号的平均值与所述第四组像素信号的平均值之差,得到第二差值;将所述第一差值与第二差值的平均值进行量化得到所述DC分量的预测值。
PCT/CN2017/071602 2016-01-25 2017-01-18 解码方法、编码方法、解码设备和编码设备 WO2017129023A1 (zh)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201610050028.5 2016-01-25
CN201610050028.5A CN106998470B (zh) 2016-01-25 2016-01-25 解码方法、编码方法、解码设备和编码设备

Publications (1)

Publication Number Publication Date
WO2017129023A1 true WO2017129023A1 (zh) 2017-08-03

Family

ID=59398967

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2017/071602 WO2017129023A1 (zh) 2016-01-25 2017-01-18 解码方法、编码方法、解码设备和编码设备

Country Status (2)

Country Link
CN (1) CN106998470B (zh)
WO (1) WO2017129023A1 (zh)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
TWI692739B (zh) * 2018-10-17 2020-05-01 美商耐能股份有限公司 影像深度解碼器及計算機裝置
CN113068028A (zh) * 2018-08-09 2021-07-02 Oppo广东移动通信有限公司 视频图像分量的预测方法、装置及计算机存储介质
CN113347425A (zh) * 2019-01-16 2021-09-03 Oppo广东移动通信有限公司 信息处理方法及装置、设备、存储介质

Families Citing this family (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109063091B (zh) * 2018-07-26 2021-06-15 成都大学 混合编码的数据迁移方法、数据迁移装置和存储介质
CN116828192A (zh) * 2018-12-15 2023-09-29 华为技术有限公司 图像重建方法和装置
CN110210283B (zh) * 2019-04-09 2021-05-14 深圳市梦网视讯有限公司 一种基于换装应用的图像处理方法和系统
CN112449191B (zh) * 2019-08-27 2024-05-17 华为技术有限公司 压缩多个图像的方法、解压缩图像的方法和装置
CN112913242B (zh) * 2020-07-24 2023-02-28 深圳市大疆创新科技有限公司 编码方法和编码装置
CN113170161A (zh) * 2020-08-31 2021-07-23 深圳市大疆创新科技有限公司 图像编码方法、图像解码方法、装置和存储介质
CN113068037B (zh) * 2021-03-17 2022-12-06 上海哔哩哔哩科技有限公司 用于样点自适应补偿的方法及装置、设备和介质
CN114119789B (zh) * 2022-01-27 2022-05-03 电子科技大学 基于在线学习的轻量化hevc色度图像质量增强方法

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1859574A (zh) * 2005-07-18 2006-11-08 华为技术有限公司 一种确定增强层帧内预测模式的方法和编解码设备
WO2011088592A1 (en) * 2010-01-22 2011-07-28 Thomson Licensing Method and device for encoding an image block of an image and corresponding decoding method and device
CN102595122A (zh) * 2011-01-14 2012-07-18 华为技术有限公司 预测模式的编解码方法、编解码设备及网络系统
CN104333761A (zh) * 2014-11-20 2015-02-04 富春通信股份有限公司 一种hevc基本单元级码率分配方法

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1859574A (zh) * 2005-07-18 2006-11-08 华为技术有限公司 一种确定增强层帧内预测模式的方法和编解码设备
WO2011088592A1 (en) * 2010-01-22 2011-07-28 Thomson Licensing Method and device for encoding an image block of an image and corresponding decoding method and device
CN102595122A (zh) * 2011-01-14 2012-07-18 华为技术有限公司 预测模式的编解码方法、编解码设备及网络系统
CN104333761A (zh) * 2014-11-20 2015-02-04 富春通信股份有限公司 一种hevc基本单元级码率分配方法

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113068028A (zh) * 2018-08-09 2021-07-02 Oppo广东移动通信有限公司 视频图像分量的预测方法、装置及计算机存储介质
US11503312B2 (en) 2018-08-09 2022-11-15 Guangdong Oppo Mobile Telecommunications Corp., Ltd. Method and devices for video predicting a first color component from a second component
CN113068028B (zh) * 2018-08-09 2023-05-26 Oppo广东移动通信有限公司 视频图像分量的预测方法、装置及计算机存储介质
TWI692739B (zh) * 2018-10-17 2020-05-01 美商耐能股份有限公司 影像深度解碼器及計算機裝置
CN113347425A (zh) * 2019-01-16 2021-09-03 Oppo广东移动通信有限公司 信息处理方法及装置、设备、存储介质
CN113347425B (zh) * 2019-01-16 2023-05-26 Oppo广东移动通信有限公司 信息处理方法及装置、设备、存储介质

Also Published As

Publication number Publication date
CN106998470A (zh) 2017-08-01
CN106998470B (zh) 2020-03-20

Similar Documents

Publication Publication Date Title
US11245932B2 (en) Encoding method and apparatus and decoding method and apparatus
KR102431537B1 (ko) 루마 및 크로마 성분에 대한 ibc 전용 버퍼 및 디폴트 값 리프레싱을 사용하는 인코더, 디코더 및 대응하는 방법들
WO2017129023A1 (zh) 解码方法、编码方法、解码设备和编码设备
US11025945B2 (en) History-based image coding method and apparatus
KR102578820B1 (ko) 히스토리 기반 움직임 정보를 이용한 영상 코딩 방법 및 그 장치
US20240048686A1 (en) Inter prediction method and apparatus
TWI688262B (zh) 用於視訊寫碼之重疊運動補償
WO2017071480A1 (zh) 参考帧编解码的方法与装置
CN113748677A (zh) 编码器、解码器及对应的帧内预测方法
TW201931853A (zh) 具有聯合像素/變換為基礎之量化之視頻寫碼之量化參數控制
CN113170202B (zh) 用于构建应用多假设预测的块的mpm列表的编码器、解码器及相应方法
JP2022524357A (ja) エンコーダ、デコーダ、及び対応するインター予測方法
US20220295071A1 (en) Video encoding method, video decoding method, and corresponding apparatus
CN113545063A (zh) 使用线性模型进行帧内预测的方法及装置
TW202044833A (zh) 使用不同色度格式之三角預測單元模式中之視訊寫碼
US20190313092A1 (en) Chrominance Prediction Method and Apparatus
WO2019128716A1 (zh) 图像的预测方法、装置及编解码器
JP2023126221A (ja) インター予測のためのエンコーダ、デコーダ、および対応する方法
WO2019109955A1 (zh) 帧间预测方法、装置及终端设备
CN116783893A (zh) 块内复制暂存帧缓冲区
TW202133619A (zh) 用於合併估計區域的基於歷史的運動向量預測約束
CN113615178A (zh) 视频译码中的色度帧内预测
CN110832859B (zh) 一种基于模板匹配的解码方法及装置
CN113228632A (zh) 用于局部亮度补偿的编码器、解码器、以及对应方法
RU2801326C2 (ru) Кодер, декодер и соответствующие способы, использующие выделенный буфер ibc, и значение по умолчанию, обновляющее компонент яркости и цветности

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17743638

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 17743638

Country of ref document: EP

Kind code of ref document: A1