CN115866263A - Video decoding method, video encoding method and related equipment - Google Patents

Video decoding method, video encoding method and related equipment Download PDF

Info

Publication number
CN115866263A
CN115866263A CN202111122423.7A CN202111122423A CN115866263A CN 115866263 A CN115866263 A CN 115866263A CN 202111122423 A CN202111122423 A CN 202111122423A CN 115866263 A CN115866263 A CN 115866263A
Authority
CN
China
Prior art keywords
value
interval range
texture direction
coding unit
pixel
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202111122423.7A
Other languages
Chinese (zh)
Inventor
周川
张晋荣
吕卓逸
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Vivo Mobile Communication Co Ltd
Original Assignee
Vivo Mobile Communication Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Vivo Mobile Communication Co Ltd filed Critical Vivo Mobile Communication Co Ltd
Priority to CN202111122423.7A priority Critical patent/CN115866263A/en
Publication of CN115866263A publication Critical patent/CN115866263A/en
Pending legal-status Critical Current

Links

Images

Landscapes

  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

The application discloses a video decoding method, a video coding method and related equipment, and belongs to the technical field of video processing. The video decoding method of the embodiment of the application comprises the following steps: the method comprises the steps that when a decoding end determines that target information does not meet a first preset condition, a first pixel predicted value of each pixel point in an encoding unit to be decoded is obtained by using a position-dependent prediction combination (PDPC) method, and the target information comprises information used for determining whether the PDPC method is used for obtaining the pixel predicted value; and the decoding end decodes the coding unit according to the first pixel prediction value.

Description

Video decoding method, video encoding method and related equipment
Technical Field
The present application belongs to the field of video processing technologies, and in particular, to a video decoding method, a video encoding method, and related devices.
Background
With the development of video coding and decoding technology, in the process of video decoding, position Dependent Prediction Combination (PDPC) calculation is usually performed for all modes of intra Prediction. For example, firstly, intra-frame prediction is carried out on a coding block to be decoded by using an intra-frame prediction algorithm to obtain a pixel prediction value, then, the pixel prediction value obtained by the intra-frame prediction is optimized by using the PDPC, in the optimization process, a second reference pixel point which is on the same straight line with a first reference pixel point in the appointed prediction direction needs to be obtained, and then, the two reference pixel points are combined to derive the pixel prediction value of each pixel point in the current coding block. Because the texture change inside the coding block is not in linear distribution, an accurate second reference pixel point cannot be obtained, and the accuracy of the output pixel predicted value is low.
Disclosure of Invention
The embodiment of the application provides a video decoding method, a video coding method and related equipment, which can improve the accuracy of a pixel prediction value.
In a first aspect, a video decoding method is provided, including:
the method comprises the steps that when a decoding end determines that target information does not meet a first preset condition, a first pixel prediction value of each pixel point in an encoding unit to be decoded is obtained by using a position-dependent prediction combination (PDPC) method, and the target information comprises information used for determining whether the PDPC method is used for obtaining the pixel prediction value;
and the decoding end decodes the coding unit according to the first pixel prediction value.
In a second aspect, a video encoding method is provided, including:
the encoding end determines an encoding unit of a video to be encoded;
the coding end codes the coding unit to obtain a target code stream;
the target code stream comprises offset information, the offset information is used for indicating an offset value, the offset value is used for determining a threshold value of a target interval range in a first preset condition, and the first preset condition is a trigger condition for obtaining a pixel predicted value without using a position-dependent predictive combination PDPC method.
In a third aspect, there is provided a video decoding apparatus, including:
the device comprises an acquisition module, a decoding module and a decoding module, wherein the acquisition module is used for acquiring a first pixel predicted value of each pixel point in a coding unit to be decoded by using a position-dependent prediction combination (PDPC) method under the condition that target information does not meet a first preset condition, and the target information comprises information for determining whether the PDPC method is used for acquiring the pixel predicted value;
a first determining module to decode the coding unit according to the first pixel prediction value.
In a fourth aspect, there is provided a video encoding apparatus comprising:
the second determining module is used for determining the coding unit of the video to be coded;
the coding module is used for coding the coding unit to obtain a target code stream;
the target code stream comprises offset information, the offset information is used for indicating an offset value, the offset value is used for determining a threshold value of a target interval range in a first preset condition, and the first preset condition is a trigger condition for obtaining a pixel predicted value without using a position-dependent prediction combination PDPC method.
In a fifth aspect, an electronic device is provided, which comprises a processor, a memory, and a program or instructions stored on the memory and executable on the processor, which when executed by the processor, implements the steps of the method according to the first aspect, or implements the steps of the method according to the second aspect.
In a sixth aspect, an electronic device is provided, comprising a processor and a communication interface, wherein,
the processor is configured to perform the following operations:
under the condition that target information does not meet a first preset condition, acquiring a first pixel predicted value of each pixel point in a coding unit to be decoded by using a position-dependent prediction combination (PDPC) method, wherein the target information comprises information for determining whether the PDPC method is used for acquiring the pixel predicted value;
decoding the coding unit according to the first pixel prediction value.
Alternatively, the processor is configured to perform the following operations:
determining a coding unit of a video to be coded;
coding the coding unit to obtain a target code stream;
the target code stream comprises offset information, the offset information is used for indicating an offset value, the offset value is used for determining a threshold value of a target interval range in a first preset condition, and the first preset condition is a trigger condition for obtaining a pixel predicted value without using a position-dependent prediction combination PDPC method.
In a seventh aspect, there is provided a readable storage medium on which a program or instructions are stored, which program or instructions, when executed by a processor, implement the steps of the method according to the first aspect or implement the steps of the method according to the second aspect.
In an eighth aspect, embodiments of the present application provide a chip, where the chip includes a processor and a communication interface, where the communication interface is coupled to the processor, and the processor is configured to execute a program or instructions to implement the steps of the method according to the first aspect, or to implement the steps of the method according to the second aspect.
In a ninth aspect, there is provided a computer program/program product stored on a non-transitory storage medium, the computer program/program product being executable by at least one processor to implement a method as described in the first aspect, or to implement a method as described in the second aspect.
According to the method and the device, the scene that the PDPC is limited to be used by setting the first preset condition is set, and the pixel predicted value of each pixel point in the coding unit can be obtained by using the PDPC method only for the coding unit which does not meet the first preset condition. Thus, the inaccuracy of the pixel prediction value output by the coding unit which is not suitable for PDPC can be reduced, and the accuracy of the whole pixel prediction value is improved. In addition, the PDPC method calculation is not performed for the coding unit not applicable to PDPC, so that the redundant calculation at the decoding side can be reduced.
Drawings
Fig. 1 is a block diagram of a network system to which an embodiment of the present application is applicable;
FIG. 2 is an exemplary diagram of a reference pixel definition during decoding according to an embodiment of the present application;
FIG. 3 is an exemplary diagram of the definition of angular mode values for different gradient directions according to an embodiment of the present application;
FIG. 4 is a diagram illustrating an example of PDPC prediction of non-vertical and non-horizontal angular prediction modes during decoding according to an embodiment of the present application;
fig. 5 is a flowchart of a video decoding method according to an embodiment of the present application;
fig. 6 is a flowchart of a video encoding method according to an embodiment of the present application;
fig. 7 is a block diagram of a video decoding apparatus according to an embodiment of the present application;
fig. 8 is a block diagram of a video encoding apparatus according to an embodiment of the present application;
fig. 9 is a block diagram of a communication device according to an embodiment of the present application;
fig. 10 is a block diagram of an electronic device according to an embodiment of the present application.
Detailed Description
The technical solutions in the embodiments of the present application will be described below clearly with reference to the drawings in the embodiments of the present application, and it is obvious that the described embodiments are some, but not all, embodiments of the present application. All other embodiments obtained by a person of ordinary skill in the art based on the embodiments in the present application are within the scope of protection of the present application.
The terms first, second and the like in the description and in the claims of the present application are used for distinguishing between similar elements and not necessarily for describing a particular sequential or chronological order. It is to be understood that the terms so used are interchangeable under appropriate circumstances such that the embodiments of the application are capable of operation in other sequences than those illustrated or otherwise described herein, and that the terms "first" and "second" used herein generally refer to a class and do not limit the number of objects, for example, a first object can be one or more. In addition, "and/or" in the specification and the claims means at least one of connected objects, and a character "/" generally means that a preceding and succeeding related objects are in an "or" relationship.
It is noted that the techniques described in the embodiments of the present application are not limited to Long Term Evolution (LTE)/LTE Evolution (LTE-Advanced) systems, but may also be used in other wireless communication systems, such as Code Division Multiple Access (CDMA), time Division Multiple Access (TDMA), frequency Division Multiple Access (FDMA), orthogonal Frequency Division Multiple Access (OFDMA), single-carrier Frequency-Division Multiple Access (SC-FDMA), and other systems. The terms "system" and "network" in the embodiments of the present application are often used interchangeably, and the described techniques can be used for both the above-mentioned systems and radio technologies, as well as for other systems and radio technologies. The following description describes a New Radio (NR) system for purposes of example, and using NR terminology in much of the description below, these techniques may also be applied to applications other than NR system applications, such as 6th generation (6 g) communication systems.
Fig. 1 shows a block diagram of a wireless communication system to which embodiments of the present application are applicable. The wireless communication system includes a terminal 11 and a network-side device 12. Wherein, the terminal 11 may be a Mobile phone, a Tablet Personal Computer (Tablet Personal Computer), a Laptop Computer (Laptop Computer) or a notebook Computer, a Personal Digital Assistant (PDA), a palmtop Computer, a netbook, an ultra-Mobile Personal Computer (UMPC), a Mobile Internet Device (MID), an Augmented Reality (AR)/Virtual Reality (VR) Device, an extended reality (XR), a robot, a Wearable Device (Wearable Device), a vehicle-mounted Device (VUE), a pedestrian terminal (PUE), a smart home (home equipment with wireless communication function, such as a refrigerator, a television, a washing machine, or furniture), and other terminal side equipment, and the Wearable equipment includes: smart watch, smart bracelet, smart earphone, smart glasses, smart jewelry (smart bracelet, smart ring, smart necklace, smart anklet, etc.), smart wristband, smart garment, game console, etc. It should be noted that the embodiment of the present application does not limit the specific type of the terminal 11. The network-side device 12 may be a Base Station or a core network device, where the Base Station may be referred to as a node B, an evolved node B, an access Point, a Base Transceiver Station (BTS), a radio Base Station, a radio Transceiver, a Basic Service Set (BSS), an Extended Service Set (ESS), a node B, an evolved node B (eNB), a home node B, a WLAN access Point, a WiFi node, a Transmit Receive Point (TRP), or some other suitable terminology in the field, as long as the same technical effect is achieved, the Base Station is not limited to a specific technical vocabulary, and it should be noted that in the embodiment of the present application, only the Base Station in the NR system is taken as an example, but the specific type of the Base Station is not limited.
For ease of understanding, some of the contents of the embodiments of the present application are described below:
uncompressed digital video in real life requires a large bit rate, for example, a video in a 4. However, these videos contain a lot of redundant information, and there is spatial redundancy in each frame of picture, and there is temporal redundancy between frames. The purpose of video encoding and decoding is to reduce redundancy of an input video signal by compressing video data.
In a two-dimensional video encoder framework, a video sequence is first divided into many frame images, and then each frame image is divided into a plurality of blocks called Coding Tree Units (CTUs), and then the CTUs are split into Coding Units (CUs). And then, predicting the CU image blocks, wherein the prediction of the image blocks is divided into an intra-frame prediction strategy and an inter-frame prediction strategy, the inter-frame prediction is that the current image block refers to the information of the front frame and/or the rear frame for encoding, the intra-frame encoding is that the current image block does not refer to the information of other frames, only refers to the information of the encoded image block of the current frame for prediction, and different prediction encoding strategies are selected according to different conditions during specific encoding. And performing prediction coding on the intra-frame prediction and inter-frame prediction current image blocks, subtracting the obtained prediction pixels from the original pixels to obtain residual image blocks, and performing transformation, quantization, inverse quantization and inverse transformation on the obtained residual image blocks, and then performing rate distortion optimization to calculate the prediction method with the minimum cost. And then entropy coding is carried out on the data of the predictive coding method and the quantized residual image block data together to write the data into a code stream, meanwhile, the residual image block after inverse transformation and the image block after predictive coding are added to form a reconstructed image block, loop filtering operation is carried out on the reconstructed image block, and information generated during loop filtering is also entropy coded to write into the code stream.
In the Intra-frame prediction Mode, a Decoder-side Intra Mode Derivation (DIMD) method is used for firstly calculating and obtaining histograms of all texture directions of reference pixels on the left and above of a current coding block, calculating amplitudes of all texture directions, respectively adding the amplitudes of different texture directions to be used as accumulated amplitudes of the texture directions, and then using angle directions corresponding to two texture directions with the largest accumulated amplitudes in the histograms of the texture directions as pixel prediction values of the current coding block. The specific process is as follows:
1. and solving gradient values of the reference pixels on the left and the upper sides by using a sobel algorithm, and calculating corresponding angle predicted values according to the gradient values.
Optionally, as shown in fig. 2, the small circle points to the left and above the current block in the figure are reference pixels, and the small dashed line blocks are regions where sobel algorithm calculation is performed, i.e. reference pixel blocks.
The Soble operator comprises two groups of 3x3 matrixes which are respectively a transverse matrix and a longitudinal matrix, and the transverse matrix and the longitudinal matrix are subjected to plane convolution with a reference pixel block to respectively obtain a transverse brightness difference approximate value and a longitudinal brightness difference approximate value. The image is represented by a, and Gx and Gy represent the gray values of the image detected by the transverse and longitudinal edges, respectively, wherein the calculation formula is as follows:
Figure BDA0003277729600000061
/>
Figure BDA0003277729600000062
the gradient direction is then calculated using the following formula:
Figure BDA0003277729600000063
wherein Θ represents the gradient direction calculated by the sobel algorithm. And calculating the amplitude corresponding to the gradient direction: | G (x, y) | = | Gx | Gy |.
2. And finding out a corresponding angle mode according to the gradient direction, and accumulating and recording the amplitude corresponding to the angle mode. The angle mode with the largest accumulated amplitude is found and recorded as dimd1, and the angle mode with the next largest accumulated amplitude is recorded as dimd2. The angular mode values for different gradient directions are shown in fig. 3.
The intra-frame prediction mode comprises a DC mode, a Planar mode and an angle mode, universal Video Coding (Versatile Video Coding), and a PDPC method is introduced to improve the Coding efficiency of 67 intra-frame prediction modes in addition to 67 intra-frame prediction modes used in VVC. If the current coding block uses the PDPC mode, after obtaining the pixel predicted values under 67 intra-frame prediction modes, updating the predicted values by adopting the PDPC method corresponding to each mode.
For the DC mode and Planar mode:
Pred[x,y]=val+((wL*(left[y]-val)+wT*(top[x]-val)+32)>>6);
where, val is the result after prediction of DC and planar modes, top [ x ] is the upper reference pixel corresponding to this prediction point, left [ y ] is the left reference pixel corresponding to this prediction point, wL =32> > min (31, ((x < < 1) > > scale)), where x is the x coordinate of this prediction point, scale = ((Log 2 (width) -2+ Log2 (height) -2+2) > > 2); wherein, width is the width of the current cu block, and height is the height of the current cu block.
For the angle prediction mode:
for the vertical direction: pred [ x, y ] = val + ((wL: (left-topLeft) + 32) > > 6); wherein val represents a value after the current angle prediction, left represents a reference pixel on the left side of the current sample, and toplex represents a first reference pixel on the top of the current sample;
for the horizontal direction: pred [ x, y ] = val + ((wT × (top-topLeft) + 32) > > 6); where val represents the predicted value of the current angle, wT =32> > (2*y > > scale), top represents the reference pixel at the top of the current sample, and toplet represents the first reference pixel to the left of the current sample;
for non-vertical and non-horizontal directions: pred [ x, y ] = val + ((wL × (left-val) + 32) > > 6);
scale=min(2,log2(sideSize)–log2(3*absInvAngle-2)-8);
sideSize=isModeVerheight:width;
absInvAngle=invAngTable[absAngMode];
invAngTable[32]={
0,16384,8192,5461,4096,2731,2048,1638,1365,1170,1024,910,819,712,630,565,512,468,420,364,321,287,256,224,191,161,128,96,64,48,32,16};
absAngMode=abs(intraPredAngleMode);
intraPredAngleMode=isModeVerpredMode-VER_IDx:-(predMode-HOR_IDx);
isModeVer=(predMode>=DIA_IDx);
predMode refers to the angular mode of the current prediction, DIA _ IDx value is 34, VER_IDx value is 50, HOR_IDx value is 18, and isModeVer indicates whether the current mode is a mode that uses an above reference pixel for prediction.
The PDPC principle for the angular prediction modes in the non-vertical and non-horizontal directions is as follows:
when the prediction direction is greater than 50 (i.e. the vertical direction), a pixel point is found in the left reference pixel along the current prediction direction, the weight of the pixel point is determined according to the position of the pixel point above the distance, and the closer the pixel point is to the upper reference pixel, the higher the weight is. When the prediction direction is smaller than 18 (i.e. horizontal direction), a pixel point is found in the upper reference pixel along the current prediction direction, the weight of the pixel point is determined according to the position of the pixel point away from the left reference pixel, and the weight of the pixel point closer to the left reference pixel is larger.
As shown in fig. 4, taking the intra prediction mode value as 66 as an example: pred (x, y) is a pixel point to be predicted, and because the current prediction direction is the oblique upward direction, after the angle prediction mode, pred (x, y) can adopt a left reference pixel R (-1,y) as a prediction pixel to obtain a prediction pixel block; and after PDPC calculation, fitting again by using R (x-1) as a prediction point to update the prediction pixel block. pred (x, y) is fitted with different weights depending on the distance R (-1,y) and the R (x, -1) position.
The PDPC method of the angle prediction mode in the non-vertical direction and the non-horizontal direction can better improve the coding performance for the fact that the distribution of the texture inside the current coding block is straight and the texture pixels are uniformly changed. However, the trend of the texture information inside the current coding block is not considered, if the trend of the internal texture is a curve, the PDPC cannot accurately find the second reference pixel point, and if the PDPC is still adopted, the complexity of coding and decoding is increased, which causes the waste of computing resources.
The video decoding method provided by the embodiments of the present application is described in detail below with reference to the accompanying drawings by using some embodiments and application scenarios thereof.
Referring to fig. 5, fig. 5 is a flowchart of a video decoding method according to an embodiment of the present application, and as shown in fig. 5, the method includes the following steps:
step 501, when a decoding end determines that target information does not meet a first preset condition, a position-dependent prediction combination (PDPC) method is used for obtaining a first pixel prediction value of each pixel point in a coding unit to be decoded, wherein the target information comprises information used for determining whether the PDPC method is used for obtaining the pixel prediction value;
in this embodiment of the present application, the decoding end and the encoding end may be understood as terminals in the above embodiments, where the decoding end and the encoding end may be the same terminal or different terminals, and are not further limited herein. The estimated area may be determined based on the height value and other encoding information, which may include the number of columns of the encoding unit for performing the PDPC.
Alternatively, the first preset condition may be a trigger condition for obtaining the pixel prediction value without using the PDPC method. The first preset condition may specifically be determined according to the target information, the first preset condition may include one or more conditions, and when the first preset condition includes multiple conditions, and one of the conditions is not satisfied, the target information may be considered not to satisfy the first preset condition.
It should be understood that, in the case that the first preset condition is not satisfied, the coding unit to be decoded may be applied to the PDPC method, that is, the PDPC method may be used to obtain the first pixel prediction value of each pixel point in the coding unit.
The step of acquiring, by the decoding end, the first pixel prediction value of each pixel point in the coding unit to be decoded by using the PDPC method may be understood as: the decoding end firstly obtains a second pixel predicted value of each pixel point in the coding unit by using an intra-frame prediction algorithm, and then optimizes and corrects the second pixel predicted value by adopting a PDPC method to obtain the first pixel predicted value.
Step 502, the decoding end decodes the coding unit according to the first pixel prediction value.
In this embodiment, the decoding, by the decoding end, the encoding unit according to the first pixel prediction value may include determining a reconstruction value of the encoding unit and a subsequent decoding process based on the first pixel prediction value and a target residual value, which may refer to related technologies and will not be further described herein.
Optionally, the target residual value is obtained by performing inverse quantization and inverse transformation on the basis of target residual information of the coding unit, where the target residual information is obtained from a target code stream corresponding to the coding unit.
According to the method and the device, the scene that the PDPC is limited to be used by setting the first preset condition is set, and the pixel predicted value of each pixel point in the coding unit can be obtained by using the PDPC method only for the coding unit which does not meet the first preset condition. Thus, the inaccuracy of the pixel prediction value output by the coding unit which is not suitable for PDPC can be reduced, and the accuracy of the whole pixel prediction value is improved. In addition, the PDPC method calculation is not performed for the coding unit not applicable to PDPC, so that the redundant calculation at the decoding side can be reduced.
Optionally, in some embodiments, the target information includes at least one of first information, decoder Intra Mode Derivation (DIMD) identification information, template-based Intra Mode Derivation (TIMD) identification information, a height value of the coding unit, and a predicted area of the coding unit for performing the PDPC method, the first information includes at least two of an Intra prediction Mode value, a first texture direction value, and a second texture direction value, and the first texture direction value and the second texture direction value are texture direction values of reference pixels of the coding unit.
In this embodiment, the first texture direction value and the second texture direction value may be understood as texture direction values pointing to upper and left reference pixels of a coding unit to be decoded, where the first texture direction value may be an upward-pointing texture direction value or a left-pointing texture direction value, the second texture direction value may be a left-pointing texture direction value or an upward-pointing texture direction value, and the first texture direction value and the second texture direction value are different.
It should be understood that the texture direction values mentioned in the embodiments of the present application may be understood as values corresponding to a certain texture direction, and the values corresponding to each texture direction may be as shown in fig. 4.
Optionally, in some embodiments, the first preset condition comprises at least one of:
the intra-frame prediction mode value is greater than or equal to a first preset value, and the first texture direction value or the second texture direction value is located in a first interval range;
the intra-frame prediction mode value is smaller than or equal to a second preset value, and the first texture direction value or the second texture direction value is located in a second interval range;
one of the first texture direction value and the second texture direction value is positioned in a third interval range, and the other texture direction value is positioned in a fourth interval range;
the intra prediction mode value is within a fifth interval range, and the first texture direction value or the second texture direction value is within a sixth interval range;
the intra prediction mode value is in a seventh interval range, and the first texture direction value or the second texture direction value is in an eighth interval range;
the TIMD identification information is a first preset value;
the DIMD identification information is a second preset value;
the height value is in a ninth interval range;
the coding unit is used for executing the estimated area of the PDPC and is positioned in the tenth interval range;
the first preset value is larger than the second preset value, and the minimum threshold of the fifth interval range is larger than the maximum threshold of the seventh interval range.
In the embodiment of the present application, the first preset value, the second preset value, the first interval range, the second interval range, the third interval range, the fourth interval range, the fifth interval range, the sixth interval range, the seventh interval range, the eighth interval range, the ninth interval range, and the tenth interval range may be agreed by a protocol, or may be indicated by an encoding end. For example, the target code stream may carry indication information to indicate at least one of a first preset value, a second preset value, a first interval range, a second interval range, a third interval range, a fourth interval range, a fifth interval range, a sixth interval range, a seventh interval range, an eighth interval range, a ninth interval range, and a tenth interval range.
It should be noted that the first preset value and the second preset value may be set according to actual needs, for example, in an embodiment, the first preset value and the second preset value are 1. The time index 1 of the TIMD indicates that the TIMD method is adopted when the coding block performs intra prediction, the time index 0 of the TIMD method indicates that the TIMD method is not adopted, the time index 1 of the DIMD method indicates that the DIMD method is adopted when the coding block performs intra prediction, and the time index 0 of the DIMD method indicates that the DIMD method is not adopted.
Further, in some embodiments, the method further comprises:
the decoding end determines a target interval range according to a target object, where the target object includes the intra-frame prediction mode value, or the target object includes the intra-frame prediction mode value and a target offset value, the target offset value is used to indicate a degree of deviation of the intra-frame prediction mode value from a texture direction value, and the target interval range includes at least one of the first interval range, the second interval range, the third interval range, the fourth interval range, the fifth interval range, the sixth interval range, the seventh interval range, and the eighth interval range.
The target offset value may include offset values corresponding to respective interval ranges, where the offset values corresponding to the interval ranges may be the same or different, or may be partially the same or different. And are not further limited herein. This will be explained in detail in the following examples.
Optionally, the minimum threshold of the first interval range may be a third preset value or cur _ mode-a + B1, and the maximum threshold may be a fourth preset value; the minimum threshold of the second interval range may be a fifth preset value, and the maximum threshold may be a sixth preset value or cur _ mode + a + B1; wherein cur _ mode is the intra prediction mode value, a is the first preset value, and B1 is a first offset value.
In the embodiment of the present application, the first preset value, the second preset value, the third preset value, the fourth preset value, the fifth preset value, and the sixth preset value may be set according to actual needs, for example, in some embodiments, the first preset value is 50, the second preset value is 18, the third preset value is 26, the fourth preset value and the fifth preset value are 34, and the sixth preset value is 42. It should be understood that in other embodiments, a may be replaced by other threshold values, and a in the minimum threshold of the first interval range and a in the minimum threshold of the second interval range may be the same or different, for example, cur _ mode-a + B1 may be replaced by cur _ mode-A1+ B1, cur _ mode-A2+ B1, A1 and A2 represent preset threshold values, and A1 and A2 may be the same or different.
Alternatively, B1 above indicates the degree of deviation of the intra prediction mode value from the texture direction value. Of course, in other embodiments, a and B1 may be defined as a whole, and in this case, the whole of a and B1 may be understood as the deviation degree of the intra prediction mode value from the texture direction value.
Optionally, the minimum threshold of the third interval range is a seventh preset value, and the maximum threshold is a sum of the seventh preset value and a second offset value; the minimum threshold of the fourth interval range is a difference value between an eighth preset value and a third deviation value, and the maximum threshold is the eighth preset value. Here, the second offset value may be understood as a degree of deviation of the intra prediction mode value from the texture direction value.
In this embodiment of the application, the sizes of the seventh preset value and the eighth preset value may be set according to actual needs, for example, in some embodiments, the seventh preset value may be 50, and the eighth preset value is 18. The third offset value may be understood as a deviation degree of the intra prediction mode value from the texture direction value.
Optionally, a minimum threshold of the fifth interval range and the eighth interval range is a ninth preset value, and a maximum threshold is a sum of the ninth preset value and a fourth offset value; the minimum value of the sixth interval range and the seventh interval range is a difference value between a tenth preset value and a fifth offset value, and the maximum value of the sixth interval range and the seventh interval range is the tenth preset value.
In the embodiment of the present application, the sizes of the ninth preset value and the tenth preset value may be set according to actual needs, for example, in some embodiments, the ninth preset value may be 50, and the tenth preset value is 18. It should be noted that. It should be understood that, in other embodiments, the fifth interval range and the eighth interval range may be different ranges, and likewise, the sixth interval range and the seventh interval range may be different ranges. The fourth offset value and the fifth offset value may be understood as a degree of deviation of the intra prediction mode value from the texture direction value, and the fourth offset value and the fifth offset value may be the same or different in magnitude.
It should be noted that, when the first preset condition includes a condition that the target code stream corresponding to the encoding unit includes the TIMD identification information, if the target code stream corresponding to the encoding unit includes the TIMD identification information, it indicates that the encoding unit is not suitable for executing the PDPC. The first preset condition includes that the TIMD flag information is a first preset value, and when the TIMD flag information is the first preset value, it indicates that the coding unit is not suitable for performing PDPC, where the first preset value may be understood as a flag value indicating intra-frame prediction using the TIMD method, and is, for example, 1. The first preset condition includes that the dim flag information is a second preset value, which may be understood as a flag value indicating intra prediction using the dim method, for example, 1, when the dim flag information is the second preset value, the flag value indicates that the coding unit is not suitable for performing PDPC.
It should be understood that the target offset value may be indicated by a protocol convention or an encoding end, and in some embodiments, in the case of indication by the encoding end, the decoding end needs to first obtain the target offset value. For example, in a case where the target object includes the target offset value, the method further includes:
and the decoding end acquires offset information from the target code stream corresponding to the coding unit, wherein the offset information is used for indicating the target offset value.
Optionally, the offset information may specifically be an offset value, an index value of the offset value in the offset value set, and an identifier of the offset value. Wherein the index value indicates the position of the offset value in the same offset value set preset by the encoding end and the decoding end; the identification of the offset value may indicate image type information, e.g. indicating a first image type when identified as a first value identification, using an offset value of 1; the flag indicates a second picture type when the flag is a second value flag, and an offset value of 2 is used.
It should be noted that the preset offset value may be transmitted to the decoding end by the encoding end through the target code stream, or the offset value that needs to be adopted by the video to be encoded may be determined by the encoding end and then transmitted to the decoding end through the target code stream. Specifically, the target code stream may additionally carry offset information, and the offset information indicates an offset value used by the decoding end.
In some embodiments, the encoding end may determine the offset information according to image information of the video to be encoded, the image information including at least one of an image type and an image content. The image type may include a screen content type and a natural image type, and if the screen content type is the offset value, the offset value is a first value; if it is a natural image type, the offset value is a second value.
It should be noted that, when a plurality of offset values need to be indicated, a specific value of the set of offset values may be directly indicated, or a position index of the set of offset values in the offset value set may be indicated.
Optionally, when the first preset condition is satisfied, the second pixel prediction value may not be corrected, or the second pixel prediction value may be corrected by using another optimization method. It is to be understood that the reconstructed value of the coding unit may be determined directly based on the second pixel prediction value without modifying the second pixel prediction value. For example, in some embodiments, the method further comprises:
and under the condition that the first preset condition is met, the decoding end decodes the coding unit according to a second pixel prediction value, wherein the second pixel prediction value is a pixel prediction value obtained by performing intra-frame prediction on the coding unit based on an intra-frame prediction mode corresponding to the intra-frame prediction mode value.
In this embodiment of the application, the target information meeting the first preset condition may be understood as all conditions included in the first preset condition being met. For example, when the first preset condition includes a condition a and a condition B, if the target information does not satisfy any one of the conditions a, or if the target information does not satisfy all the conditions in the condition a and does not satisfy any one of the conditions B, it indicates that the target information satisfies the first preset condition, otherwise, the target information satisfies the first preset condition. Wherein condition a comprises at least one of: the TIMD identification information is a first preset value; the DIMD identification information is a second preset value; the height value is in a ninth interval range; the coding unit is used for executing the estimated area of the PDPC and is positioned in the tenth interval range. Condition B includes at least one of: the intra-frame prediction mode value is greater than or equal to a first preset value, and the first texture direction value or the second texture direction value is located in a first interval range; the intra-frame prediction mode value is smaller than or equal to a second preset value, and the first texture direction value or the second texture direction value is located in a second interval range; one of the first texture direction value and the second texture direction value is located in a third interval range, and the other texture direction value is located in a fourth interval range; the intra prediction mode value is within a fifth interval range, and the first texture direction value or the second texture direction value is within a sixth interval range; the intra prediction mode value is located in a seventh interval range, and the first texture direction value or the second texture direction value is located in an eighth interval range.
Optionally, the determining manner of the first texture direction value and the second texture direction value may be set according to actual needs, for example, in some embodiments, the method further includes:
the decoding end determines target pixel points, wherein the target pixel points comprise pixel points positioned on the left side of the coding unit and pixel points positioned above the coding unit;
and the decoding end performs texture analysis on the target pixel point by using a preset texture analysis mode to obtain the first texture direction value and the second texture direction value.
In the embodiment of the present application, in the intra prediction process, a first texture direction value and a second texture direction value of the target reference pixel point may be derived by using a preset texture analysis method. The target reference pixel points may be N1 rows and N2 columns of reference pixel points on the left side and the N1 row above the coding unit to be decoded, and the sizes of N1 and N2 may be set according to actual needs, for example, N1 and N2 are integers greater than or equal to 3. The reference pixel point may be adjacent to or not adjacent to the coding unit to be decoded. The predetermined texture analysis may be derived using DIMD.
In some embodiments, the method further comprises:
the decoding end determines a first reference pixel point and a second reference pixel point, wherein the first reference pixel point is a pixel point positioned on the left side of the coding unit, and the second reference pixel point is a pixel point positioned above the coding unit;
and the decoding end obtains a first texture direction value corresponding to the first reference pixel point and a second texture direction value corresponding to the second reference pixel point by using a preset texture analysis mode.
It should be understood that, in the embodiment of the present application, a preset texture analysis mode may be first used to analyze a first reference pixel point, so as to obtain a first texture direction value corresponding to the first reference pixel point; and then analyzing the second reference pixel point to obtain a second texture direction value corresponding to the second reference pixel point.
Optionally, the preset texture analysis manner may be set according to actual needs, for example, in some embodiments, the preset texture analysis manner is derived by using a DIMD. The first reference pixel points may be left N2 rows of reference pixel points adjacent to the coding unit to be decoded, the second reference pixel points may be upper N1 rows of reference pixel points adjacent to the coding unit to be decoded, and the sizes of N1 and N2 may be set according to actual needs, for example, N1 and N2 are integers greater than or equal to 3. The first reference pixel point and the second reference pixel point may be adjacent to or not adjacent to the coding unit to be decoded.
In the embodiment of the present application, the first texture direction value is a texture direction value indicating the left side, and therefore the first texture direction value is within a certain interval range, for example, 2 to 34; the second texture direction value is a texture direction value indicating an upper direction, and thus the second texture direction value is within a certain interval range, for example, 34 to 66. For this reason, the definition of the first preset condition described above can be simplified. For example, the first preset condition includes at least one of:
the intra-frame prediction mode value is greater than or equal to a first preset value, and the first texture direction value is located in a first interval range;
the intra-frame prediction mode value is smaller than or equal to a second preset value, and the second texture direction value is located in a second interval range;
the first texture direction value is located in a third interval range, and the second texture direction value is located in a fourth interval range;
the intra prediction mode value is within a fifth interval range, and the first texture direction value is within a sixth interval range;
the intra-frame prediction mode value is located in a seventh interval range, and the second texture direction value is located in an eighth interval range;
the TIMD identification information is a first preset value;
the DIMD identification information is a second preset value;
the height value is in a ninth interval range;
the coding unit is used for executing the estimated area of the PDPC and is positioned in the tenth interval range;
the first preset value is larger than the second preset value, and the minimum threshold of the fifth interval range is larger than the maximum threshold of the seventh interval range.
For a better understanding of the present application, the following detailed description is given by way of some specific examples.
The first embodiment is as follows: the decoding end obtains the pixel predicted value of each pixel point in the coding unit to be decoded according to the following steps:
the method comprises the following steps of firstly, obtaining intra-frame prediction information of a coding unit to be decoded from a code stream, wherein the intra-frame prediction information comprises: an intra prediction mode value of a coding unit to be decoded;
secondly, deriving a first texture direction value DIMD1 and a second texture direction value DIMD2 of adjacent reference pixel points of a coding unit to be decoded by using the DIMD;
it should be noted that the adjacent reference pixel points may be reference pixel points in three rows above and three columns on the left adjacent to the coding unit to be decoded, or reference pixel points in three rows above and three columns on the left above.
Thirdly, judging whether the coding unit to be decoded uses PDPC to obtain the pixel prediction value of each pixel point in the coding unit according to the intra-frame prediction mode value cur _ mode of the coding unit to be decoded and the first texture direction value dimd1 and the second texture direction value dimd2 derived in the second step; specifically, when one of the first preset conditions is met, the PDPC is not used to obtain the pixel prediction value of each pixel point in the coding unit to be decoded, otherwise, the PDPC is used to obtain the pixel prediction value of each pixel point in the coding unit to be decoded;
the first preset condition includes:
condition 1. If cur _ mode is greater than 50 and dimd1 or dimd2 is in the interval greater than 26, less than 34;
condition 2. If cur _ mode is less than 18 and dimd1 or dimd2 is in the interval greater than 34, less than 42;
it should be noted that the thresholds 50 and 18, and the intervals are greater than 26, less than 34 and greater than 34, and less than 42, may be other values, and are not limited herein.
And fourthly, obtaining residual error information of the coding unit to be decoded from the code stream, carrying out inverse quantization and inverse transformation on the residual error information to obtain a residual error value, and adding the residual error value and the predicted value to obtain a reconstruction value of the coding unit to be decoded.
The present embodiment may be used in a luminance block or a chrominance block, or both the luminance block and the chrominance block, which is not limited herein.
It should be noted that different first preset conditions may also be set for different embodiments, for example, in some embodiments, the above condition 1 and condition 2 may be replaced by condition 3 and condition 4, or the first preset condition further includes condition 3 and condition 4. Wherein the content of the first and second substances,
condition 3. If cur _ mode is greater than 50 and either dimd1 or dimd2 is in an interval greater than ((cur _ mode-50+ offset)), less than 34;
condition 4. If cur _ mode is less than 18 and dimd1 or dimd2 is in an interval greater than 50, less than (cur _ mode +50+ offset);
it should be noted that the threshold values 50 and 18, the range of which is less than 34 and greater than 50, may be other values, and are not limited herein. Offset can take any positive, negative, 0, etc.
Further, in some embodiments, the above conditions 1 and 2 may be replaced with conditions 5 and 6, or the first preset condition further includes conditions 5 and 6. Wherein the content of the first and second substances,
condition 5. If dimd1 is greater than 50 and less than (50 + offset) and dimd2 is greater than (18-offset) and less than 18;
condition 6. If dimd1 is greater than (18-offset) and less than 18, and dimd2 is greater than 50 and less than (50 + offset);
it should be noted that the thresholds 50 and 18 may be other values, and are not limited herein. Offset can take any positive, negative, 0, etc.
Further, in some embodiments, the above conditions 1 and 2 may be replaced with conditions 7 and 8, or the first preset condition further includes conditions 7 and 8. Wherein the content of the first and second substances,
condition 7. If cur _ mode is in an interval greater than 50, less than (50 + offset), and either dimd1 or dimd2 is in an interval greater than (18-offset), less than 18;
conditional 8. If cur _ mode is greater than (18-offset), less than 18, and either dimd1 or dimd2 is in the interval greater than 50, less than (50 + offset);
it should be noted that the thresholds 50 and 18 may be other values, and are not limited herein. Offset can take any positive, negative, 0, etc.
Further, in some embodiments, at least one of TIMD identification information, a height value of the encoding unit, and an estimated area of the encoding unit for performing PDPC may be added as a judgment condition of the first preset condition. For example, in some embodiments, the first preset condition may further include:
the TIMD identification information is 1;
the DIMD identification information is 1;
the height value is less than or equal to 16;
the coding unit is configured to perform the prediction of the area of the PDPC to be less than or equal to 16 x 12.
It should be noted that the thresholds 16 and 12 may be other values, and are not limited herein.
Example two: the decoding end obtains the pixel prediction value of each pixel point in the coding unit to be decoded according to the following steps:
the method comprises the following steps of firstly, obtaining intra-frame prediction information of a coding unit to be decoded from a code stream, wherein the intra-frame prediction information comprises: an intra prediction mode value of a coding unit to be decoded;
secondly, deriving a texture direction value DIMD _ top of the upper reference pixel by using the DIMD, and deriving a texture direction value DIMD _ left of the left reference pixel by using the DIMD;
the upper reference pixels may be reference pixels of three or more rows above and adjacent to the coding unit to be decoded, and the left reference pixels may be reference pixels of three or more rows on the left.
Thirdly, judging whether the coding unit to be decoded obtains the pixel prediction value of each pixel point in the coding unit by using a PDPC method or not according to the intra-frame prediction mode value cur _ mode of the coding unit to be decoded, the texture direction value dimd _ top of the upper reference pixel and the texture direction value dimd _ left of the left reference pixel which are derived in the second step; specifically, when one of the following first preset conditions is met, the PDPC is not used for obtaining the pixel prediction value of each pixel point in the coding unit to be decoded, otherwise, the PDPC is used for obtaining the pixel prediction value of each pixel point in the coding unit to be decoded;
condition 1. If cur _ mode is greater than 50 and dimd _ left is in the interval greater than 26, less than 34;
condition 2. If cur _ mode is less than 18 and dimd _ top is in the interval greater than 34, less than 42;
it should be noted that the thresholds 50 and 18, and the intervals are greater than 26, less than 34 and greater than 34, and less than 42, may be other values, and are not limited herein.
And fourthly, obtaining residual error information of the coding unit to be decoded from the code stream, carrying out inverse quantization and inverse transformation on the residual error information to obtain a residual error value, and adding the residual error value and the predicted value to obtain a reconstruction value of the coding unit to be decoded.
It should be noted that, this embodiment may be used in a luminance block or a chrominance block, or both the luminance block and the chrominance block, which is not limited herein.
It should be noted that different first preset conditions may also be set for different embodiments, for example, in some embodiments, the above condition 1 and condition 2 may be replaced by condition 3 and condition 4, or the first preset condition further includes condition 3 and condition 4. Wherein the content of the first and second substances,
condition 3. If cur _ mode is greater than 50, and dimd _ left is in an interval greater than (cur _ mode-50+ offset), less than 34;
condition 4. If cur _ mode is less than 18 and dimd _ top is in the interval greater than 50, less than (cur _ mode +50+ offset);
it should be noted that the threshold values 50 and 18, which range from less than 34 to greater than 50, may be other values, and are not limited herein. Offset can take any positive, negative, 0, etc.
Further, in some embodiments, the above conditions 1 and 2 may be replaced with the condition 5, or the first preset condition further includes the condition 5. Wherein, the first and the second end of the pipe are connected with each other,
condition 5. If dimd _ top is greater than 50 and less than (50 + offset), and dimd _ left is greater than (18-offset) and less than 18.
It should be noted that the thresholds 50 and 18 may be other values, and are not limited herein. Offset can take any of positive, negative, 0, etc.
Further, in some embodiments, the above conditions 1 and 2 may be replaced with conditions 6 and 7, or the first preset condition further includes conditions 6 and 7. Wherein the content of the first and second substances,
condition 6. If cur _ mode is in an interval greater than 50, less than (50 + offset), and dimd _ left is in an interval greater than (18-offset), less than 18;
condition 7. If cur _ mode is greater than (18-offset), less than 18, and dimd _ top is in the interval greater than 50, less than (50 + offset);
it should be noted that the thresholds 50 and 18 may be other values, and are not limited herein. Offset can take any positive, negative, 0, etc.
Further, in some embodiments, at least one of TIMD identification information, a height value of the encoding unit, and an estimated area of the encoding unit for performing PDPC may be added as a judgment condition of the first preset condition. For example, in some embodiments, the first preset condition may further include:
the TIMD identification information is 1;
the DIMD identification information is 1;
said height value is less than or equal to 16;
the coding unit is configured to perform the prediction of the area of the PDPC to be less than or equal to 16 x 12.
It should be noted that the thresholds 16 and 12 may be other values, and are not limited herein.
Example three: the decoding end obtains the pixel prediction value of each pixel point in the coding unit to be decoded according to the following steps:
the method comprises the steps of firstly, acquiring relevant information for determining whether a PDPC method is used for acquiring a pixel predicted value from a code stream, wherein the relevant information comprises TIMD identification information for example;
secondly, judging whether the coding unit to be decoded uses a PDPC method to obtain a pixel prediction value of each pixel point in the coding unit according to the TIMD identification information; specifically, when the condition 1 is satisfied, the PDPC is not used to obtain the pixel prediction value of each pixel point in the coding unit to be decoded, otherwise, the PDPC method is used to obtain the pixel prediction value of each pixel point in the coding unit to be decoded. Wherein the condition 1 includes that TIMD identification information is 1.
And thirdly, obtaining residual error information of the coding unit to be decoded from the code stream, carrying out inverse quantization and inverse transformation on the residual error information to obtain a residual error value, and adding the residual error value and the pixel predicted value to obtain a reconstructed value of the coding unit to be decoded.
Further, the TIMD identification information may be replaced with DIMD identification information, or the related information further includes DIMD identification information, and the condition 1 further includes that the DIMD identification information is 1.
Example four: the decoding end obtains the pixel prediction value of each pixel point in the coding unit to be decoded according to the following steps:
the method comprises the steps of firstly, acquiring relevant information for determining whether a pixel predicted value is acquired by using a PDPC method from a code stream, wherein the relevant information comprises a height value of an encoding unit to be decoded;
secondly, judging whether the coding unit to be decoded uses a PDPC method to obtain a pixel prediction value of each pixel point in the coding unit according to the height value of the coding unit; specifically, when the condition 1 is satisfied, the PDPC is not used to obtain the pixel prediction value of each pixel point in the coding unit to be decoded, otherwise, the PDPC method is used to obtain the pixel prediction value of each pixel point in the coding unit to be decoded. Wherein condition 1 comprises: the height value is less than or equal to 16.
Further, the related information may include an estimated area of the encoding unit for executing the PDPC method, and the condition further includes: the coding unit is used for performing prediction of the PDPC, wherein the prediction area is less than or equal to 16 × 12;
alternatively, the height value of the coding unit may be replaced by an estimated area of the coding unit for performing the PDPC method, and the condition 1 is that the estimated area of the coding unit for performing the PDPC is less than or equal to 16 × 12.
It should be noted that 16 may represent a height value of the coding unit, 12 may represent a value calculated based on other coding information, and the other coding information may include the number of columns of the coding unit for performing PDPC, and the thresholds 16 and 12 may be other numerical values, which is not limited herein.
And thirdly, obtaining residual error information of the coding unit to be decoded from the code stream, carrying out inverse quantization and inverse transformation on the residual error information to obtain a residual error value, and adding the residual error value and the pixel predicted value to obtain a reconstructed value of the coding unit to be decoded.
Referring to fig. 6, fig. 6 is a flowchart of a video encoding method according to an embodiment of the present application, as shown in fig. 6, including the following steps:
601, a coding end determines a coding unit of a video to be coded;
step 602, the coding end codes the coding unit to obtain a target code stream;
the target code stream comprises offset information, the offset information is used for indicating an offset value, the offset value is used for determining a threshold value of a target interval range in a first preset condition, and the first preset condition is a trigger condition for obtaining a pixel predicted value without using a position-dependent prediction combination PDPC method.
Optionally, before determining the coding unit of the video to be coded, the method further includes:
and the encoding end determines the offset information according to the image information of the video to be encoded, wherein the image information comprises at least one of image type and image content.
Optionally, the target information includes at least one of first information, decoder intra mode derivation DIMD identification information, template-based intra mode derivation TIMD identification information, a height value of the coding unit, and a predicted area of the coding unit for performing the PDPC method, the first information includes at least two of an intra prediction mode value, a first texture direction value, and a second texture direction value, the first texture direction value and the second texture direction value being texture direction values of reference pixels of the coding unit.
Optionally, the first preset condition comprises at least one of:
the intra-frame prediction mode value is greater than or equal to a first preset value, and the first texture direction value or the second texture direction value is located in a first interval range;
the intra-frame prediction mode value is smaller than or equal to a second preset value, and the first texture direction value or the second texture direction value is located in a second interval range;
one of the first texture direction value and the second texture direction value is located in a third interval range, and the other texture direction value is located in a fourth interval range;
the intra prediction mode value is within a fifth interval range, and the first texture direction value or the second texture direction value is within a sixth interval range;
the intra prediction mode value is in a seventh interval range, and the first texture direction value or the second texture direction value is in an eighth interval range;
the TIMD identification information is a first preset value;
the DIMD identification information is a second preset value;
the height value is in a ninth interval range;
the coding unit is used for executing the estimated area of the PDPC and is positioned in the tenth interval range;
the first preset value is larger than the second preset value, and the minimum threshold of the fifth interval range is larger than the maximum threshold of the seventh interval range.
It should be noted that, this embodiment is used as an implementation of the encoding end corresponding to the embodiment shown in fig. 5, and specific implementations thereof may refer to relevant descriptions of the embodiment shown in fig. 5 and achieve the same beneficial effects, and are not described herein again to avoid repeated descriptions.
It should be noted that, in the video decoding method provided in the embodiment of the present application, the execution main body may be a video decoding apparatus, or a control module in the video decoding apparatus for executing the video decoding method. In the embodiment of the present application, a video decoding apparatus executing a video decoding method is taken as an example to describe the video decoding apparatus provided in the embodiment of the present application.
Referring to fig. 7, fig. 7 is a structural diagram of a video decoding apparatus according to an embodiment of the present application, and as shown in fig. 7, a video decoding apparatus 700 includes:
an obtaining module 701, configured to obtain a first pixel prediction value of each pixel point in an encoding unit to be decoded by using a position-dependent prediction combination PDPC method when it is determined that target information does not satisfy a first preset condition, where the target information includes information used to determine whether the PDPC method is used to obtain the pixel prediction value;
a first determining module 702, configured to decode the coding unit according to the first pixel prediction value.
Optionally, the target information includes at least one of first information, decoder intra mode derivation DIMD identification information, template-based intra mode derivation TIMD identification information, a height value of the coding unit, and a prediction area used by the coding unit to perform the PDPC method, the first information includes at least two of an intra prediction mode value, a first texture direction value, and a second texture direction value, the first texture direction value and the second texture direction value being texture direction values of reference pixels of the coding unit.
Optionally, the first preset condition comprises at least one of:
the intra-frame prediction mode value is greater than or equal to a first preset value, and the first texture direction value or the second texture direction value is located in a first interval range;
the intra-frame prediction mode value is smaller than or equal to a second preset value, and the first texture direction value or the second texture direction value is located in a second interval range;
one of the first texture direction value and the second texture direction value is located in a third interval range, and the other texture direction value is located in a fourth interval range;
the intra prediction mode value is located in a fifth interval range, and the first texture direction value or the second texture direction value is located in a sixth interval range;
the intra prediction mode value is in a seventh interval range, and the first texture direction value or the second texture direction value is in an eighth interval range;
the TIMD identification information is a first preset value;
the DIMD identification information is a second preset value;
the height value is in a ninth interval range;
the coding unit is used for executing the estimated area of the PDPC and is positioned in the tenth interval range;
the first preset value is larger than the second preset value, and the minimum threshold of the fifth interval range is larger than the maximum threshold of the seventh interval range.
Optionally, the video decoding apparatus 700 further includes:
a third determining module, configured to determine a target interval range according to a target object, where the target object includes the intra-prediction mode value, or the target object includes the intra-prediction mode value and a target offset value, where the target offset value is used to indicate a degree of deviation of the intra-prediction mode value from a texture direction value, and the target interval range includes at least one of the first interval range, the second interval range, the third interval range, the fourth interval range, the fifth interval range, the sixth interval range, the seventh interval range, and the eighth interval range.
Optionally, in a case that the target object includes the target offset value, the obtaining module 702 is further configured to: and acquiring offset information from a target code stream corresponding to the coding unit, wherein the offset information is used for indicating the target offset value.
Optionally, the decoding module 702 is further configured to: and decoding the coding unit according to a second pixel prediction value under the condition that the first preset condition is met, wherein the second pixel prediction value is a pixel prediction value obtained by performing intra-frame prediction on the coding unit based on an intra-frame prediction mode corresponding to the intra-frame prediction mode value.
Optionally, the video decoding apparatus 700 further includes:
a third determining module, configured to determine a target pixel by the decoding end, where the target pixel includes a pixel located on the left side of the encoding unit and a pixel located above the encoding unit; and the decoding end performs texture analysis on the target pixel point by using a preset texture analysis mode to obtain the first texture direction value and the second texture direction value.
Optionally, the video decoding apparatus 700 further includes:
a third determining module, configured to determine a first reference pixel and a second reference pixel, where the first reference pixel is a pixel located on the left side of the encoding unit, and the second reference pixel is a pixel located above the encoding unit; and obtaining a first texture direction value corresponding to the first reference pixel point and a second texture direction value corresponding to the second reference pixel point by using a preset texture analysis mode.
It should be noted that, in the video encoding method provided in the embodiment of the present application, the execution subject may be a video encoding apparatus, or a control module in the video encoding apparatus for executing the video encoding method. In the embodiment of the present application, a video encoding apparatus executing a video encoding method is taken as an example to describe the video encoding apparatus provided in the embodiment of the present application.
Referring to fig. 8, fig. 8 is a block diagram of a video encoding apparatus according to an embodiment of the present disclosure, and as shown in fig. 8, the video encoding apparatus 800 includes:
a second determining module 801, configured to determine an encoding unit of a video to be encoded;
an encoding module 802, configured to encode the encoding unit to obtain a target code stream;
the target code stream comprises offset information, the offset information is used for indicating an offset value, the offset value is used for determining a threshold value of a target interval range in a first preset condition, and the first preset condition is a trigger condition for obtaining a pixel predicted value without using a position-dependent prediction combination PDPC method.
Optionally, the second determining module 801 is further configured to determine the offset information according to image information of the video to be encoded, where the image information includes at least one of an image type and an image content.
Optionally, the target information includes at least one of first information, decoder intra mode derivation DIMD identification information, template-based intra mode derivation TIMD identification information, a height value of the coding unit, and a predicted area of the coding unit for performing the PDPC method, the first information includes at least two of an intra prediction mode value, a first texture direction value, and a second texture direction value, the first texture direction value and the second texture direction value being texture direction values of reference pixels of the coding unit.
Optionally, the first preset condition comprises at least one of:
the intra-frame prediction mode value is greater than or equal to a first preset value, and the first texture direction value or the second texture direction value is located in a first interval range;
the intra-frame prediction mode value is smaller than or equal to a second preset value, and the first texture direction value or the second texture direction value is located in a second interval range;
one of the first texture direction value and the second texture direction value is positioned in a third interval range, and the other texture direction value is positioned in a fourth interval range;
the intra prediction mode value is within a fifth interval range, and the first texture direction value or the second texture direction value is within a sixth interval range;
the intra prediction mode value is in a seventh interval range, and the first texture direction value or the second texture direction value is in an eighth interval range;
the TIMD identification information is a first preset value;
the DIMD identification information is a second preset value;
the height value is in a ninth interval range;
the coding unit is used for executing the estimated area of the PDPC and is positioned in the tenth interval range;
the first preset value is larger than the second preset value, and the minimum threshold of the fifth interval range is larger than the maximum threshold of the seventh interval range.
The video decoding apparatus and the video encoding apparatus in the embodiments of the present application may be apparatuses, apparatuses or electronic devices having an operating system, or may be components, integrated circuits, or chips in a terminal. The device can be a mobile terminal or a non-mobile terminal. For example, the mobile terminal may include, but is not limited to, the above-listed type of terminal 11, and the non-mobile terminal may be a server, a Network Attached Storage (NAS), a Personal Computer (PC), a Television (TV), a teller machine, a kiosk, or the like, and the embodiments of the present application are not limited in particular.
The video decoding apparatus and the video encoding apparatus provided in the embodiment of the present application can implement each process implemented by the method embodiments of fig. 5 to fig. 6, and achieve the same technical effect, and are not described herein again to avoid repetition.
Optionally, as shown in fig. 9, an embodiment of the present application further provides a communication device 900, which includes a processor 901, a memory 902, and a program or an instruction stored in the memory 902 and executable on the processor 901, for example, when the communication device 900 is a decoding end, the program or the instruction is executed by the processor 901 to implement the processes of the video decoding method embodiment, and the same technical effect can be achieved. When the communication device 900 is an encoding end, the program or the instruction is executed by the processor 901 to implement the processes of the above-described video encoding method embodiment, and the same technical effect can be achieved.
An embodiment of the present application further provides an electronic device, which includes a processor and a communication interface, where the processor is configured to perform the following operations:
under the condition that target information does not meet a first preset condition, acquiring a first pixel predicted value of each pixel point in a coding unit to be decoded by using a position-dependent prediction combination (PDPC) method, wherein the target information comprises information for determining whether the PDPC method is used for acquiring the pixel predicted value;
decoding the coding unit according to the first pixel prediction value.
Alternatively, the processor is configured to perform the following operations:
determining a coding unit of a video to be coded;
coding the coding unit to obtain a target code stream;
the target code stream comprises offset information, the offset information is used for indicating an offset value, the offset value is used for determining a threshold value of a target interval range in a first preset condition, and the first preset condition is a trigger condition for obtaining a pixel predicted value without using a position-dependent predictive combination PDPC method.
All implementation processes and implementation manners of the method embodiment can be applied to the electronic equipment embodiment, and the same technical effect can be achieved. Specifically, fig. 10 is a schematic hardware structure diagram of an electronic device for implementing various embodiments of the present application.
The electronic device 1000 includes, but is not limited to: a radio frequency unit 1001, a network module 1002, an audio output unit 1003, an input unit 1004, a sensor 1005, a display unit 1006, a user input unit 1007, an interface unit 1008, a memory 1009, a processor 1010, and the like.
Those skilled in the art will appreciate that the electronic device 1000 may further comprise a power source (e.g., a battery) for supplying power to various components, and the power source may be logically connected to the processor 1010 through a power management system, so as to implement functions of managing charging, discharging, and power consumption through the power management system. The electronic device structure shown in fig. 10 does not constitute a limitation of the electronic device, and the electronic device may include more or less components than those shown, or combine some components, or arrange different components, and thus, the description is not repeated here.
It should be understood that, in the embodiment of the present application, the input Unit 1004 may include a Graphics Processing Unit (GPU) and a microphone, and the Graphics Processing Unit processes image data of still pictures or videos obtained by an image capturing device (such as a camera) in a video capturing mode or an image capturing mode. The display unit 1006 may include a display panel, which may be configured in the form of a liquid crystal display, an organic light emitting diode, or the like. The user input unit 1007 includes a touch panel and other input devices. Touch panels, also known as touch screens. The touch panel may include two parts of a touch detection device and a touch controller. Other input devices may include, but are not limited to, a physical keyboard, function keys (e.g., volume control keys, switch keys, etc.), a trackball, a mouse, and a joystick, which are not described in detail herein.
In this embodiment, after receiving downlink data from a network side device, the radio frequency unit 1001 processes the downlink data to the processor 1010; in addition, the uplink data is sent to the network side equipment. In general, radio frequency unit 1001 includes, but is not limited to, an antenna, at least one amplifier, a transceiver, a coupler, a low noise amplifier, a duplexer, and the like.
The memory 1009 may be used to store software programs or instructions and various data. The memory 109 may mainly include a stored program or instruction area and a stored data area, wherein the stored program or instruction area may store an operating system, an application program or instruction (such as a sound playing function, an image playing function, etc.) required for at least one function, and the like. Further, the Memory 1009 may include a high-speed random access Memory and may also include a non-transitory Memory, wherein the non-transitory Memory may be a Read-Only Memory (ROM), a Programmable Read-Only Memory (PROM), an Erasable Programmable Read-Only Memory (EPROM), an Electrically Erasable Programmable Read-Only Memory (EEPROM), or a flash Memory. Such as at least one magnetic disk storage device, flash memory device, or other non-transitory solid state storage device.
Processor 1010 may include one or more processing units; alternatively, processor 1010 may integrate an application processor that handles primarily the operating system, user interface, and application programs or instructions, and a modem processor that handles primarily wireless communications, such as a baseband processor. It will be appreciated that the modem processor described above may not be integrated into processor 1010.
Processor 1010 is configured to perform the following operations:
under the condition that target information does not meet a first preset condition, acquiring a first pixel predicted value of each pixel point in a coding unit to be decoded by using a position-dependent prediction combination (PDPC) method, wherein the target information comprises information for determining whether the PDPC method is used for acquiring the pixel predicted value;
decoding the coding unit according to the first pixel prediction value.
Alternatively, processor 1010 is configured to perform the following operations:
determining a coding unit of a video to be coded;
coding the coding unit to obtain a target code stream;
the target code stream comprises offset information, the offset information is used for indicating an offset value, the offset value is used for determining a threshold value of a target interval range in a first preset condition, and the first preset condition is a trigger condition for obtaining a pixel predicted value without using a position-dependent prediction combination PDPC method.
The processor 1010 in the embodiment of the present application can implement the steps in the video decoding method and the video encoding method, and can achieve the same effect.
An embodiment of the present application further provides a readable storage medium, where a program or an instruction is stored on the readable storage medium, and when the program or the instruction is executed by a processor, the program or the instruction implements each process of the above-mentioned embodiment of the video decoding method or the video encoding method, and can achieve the same technical effect, and in order to avoid repetition, details are not repeated here.
The processor is the processor in the electronic device described in the above embodiment. The readable storage medium includes a computer readable storage medium, such as a Read-Only Memory (ROM), a Random Access Memory (RAM), a magnetic disk or an optical disk, and so on.
The embodiment of the present application further provides a chip, where the chip includes a processor and a communication interface, the communication interface is coupled to the processor, and the processor is configured to run a program or an instruction to implement each process of the above-mentioned embodiment of the video decoding method or the video encoding method, and can achieve the same technical effect, and in order to avoid repetition, details are not repeated here.
It should be understood that the chips mentioned in the embodiments of the present application may also be referred to as system-on-chip, system-on-chip or system-on-chip, etc.
An embodiment of the present application further provides a program product, where the program product is stored in a non-transitory storage medium, and the program product is executed by at least one processor to implement each process of the above-mentioned video decoding method or video encoding method embodiment, and can achieve the same technical effect, and in order to avoid repetition, the details are not repeated here.
It should be noted that, in this document, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising a … …" does not exclude the presence of another identical element in a process, method, article, or apparatus that comprises the element. Further, it should be noted that the scope of the methods and apparatus of the embodiments of the present application is not limited to performing the functions in the order illustrated or discussed, but may include performing the functions in a substantially simultaneous manner or in a reverse order based on the functions involved, e.g., the methods described may be performed in an order different than that described, and various steps may be added, omitted, or combined. Additionally, features described with reference to certain examples may be combined in other examples.
Through the above description of the embodiments, those skilled in the art will clearly understand that the method of the above embodiments can be implemented by software plus a necessary general hardware platform, and certainly can also be implemented by hardware, but in many cases, the former is a better implementation manner. Based on such understanding, the technical solutions of the present application may be embodied in the form of a computer software product, which is stored in a storage medium (such as ROM/RAM, magnetic disk, optical disk) and includes instructions for enabling a terminal (such as a mobile phone, a computer, a server, an air conditioner, or a base station) to execute the method according to the embodiments of the present application.
While the present embodiments have been described with reference to the accompanying drawings, it is to be understood that the invention is not limited to the precise embodiments described above, which are meant to be illustrative and not restrictive, and that various changes may be made therein by those skilled in the art without departing from the spirit and scope of the invention as defined by the appended claims.

Claims (17)

1. A video decoding method, comprising:
the method comprises the steps that when a decoding end determines that target information does not meet a first preset condition, a first pixel predicted value of each pixel point in an encoding unit to be decoded is obtained by using a position-dependent prediction combination (PDPC) method, and the target information comprises information used for determining whether the PDPC method is used for obtaining the pixel predicted value;
under the condition that the first preset condition is not satisfied, the decoding end obtains a first pixel prediction value of each pixel point in the coding unit to be decoded by using a PDPC method;
and the decoding end decodes the coding unit according to the first pixel prediction value.
2. The method of claim 1, wherein the target information comprises at least one of first information, decoder intra mode derived DIMD identification information, template-based intra mode derived TIMD identification information, a height value of the coding unit, and a predicted area of the coding unit for performing the PDPC method, the first information comprises at least two of an intra prediction mode value, a first texture direction value, and a second texture direction value, the first texture direction value and the second texture direction value being texture direction values of reference pixels of the coding unit.
3. The method according to claim 2, characterized in that said first preset condition comprises at least one of:
the intra-frame prediction mode value is greater than or equal to a first preset value, and the first texture direction value or the second texture direction value is located in a first interval range;
the intra-frame prediction mode value is smaller than or equal to a second preset value, and the first texture direction value or the second texture direction value is located in a second interval range;
one of the first texture direction value and the second texture direction value is located in a third interval range, and the other texture direction value is located in a fourth interval range;
the intra prediction mode value is located in a fifth interval range, and the first texture direction value or the second texture direction value is located in a sixth interval range;
the intra prediction mode value is in a seventh interval range, and the first texture direction value or the second texture direction value is in an eighth interval range;
the TIMD identification information is a first preset value;
the DIMD identification information is a second preset value;
the height value is in a ninth interval range;
the coding unit is used for executing the estimated area of the PDPC and is positioned in the tenth interval range;
the first preset value is larger than the second preset value, and the minimum threshold of the fifth interval range is larger than the maximum threshold of the seventh interval range.
4. The method of claim 3, further comprising:
the decoding end determines a target interval range according to a target object, wherein the target object includes the intra-frame prediction mode value, or the target object includes the intra-frame prediction mode value and a target offset value, the target offset value is used for indicating a deviation degree of the intra-frame prediction mode value from a texture direction value, and the target interval range includes at least one of the first interval range, the second interval range, the third interval range, the fourth interval range, the fifth interval range, the sixth interval range, the seventh interval range and the eighth interval range.
5. The method of claim 4, wherein if the target object comprises the target offset value, the method further comprises:
and the decoding end acquires offset information from the target code stream corresponding to the coding unit, wherein the offset information is used for indicating the target offset value.
6. The method according to any one of claims 1 to 5, further comprising:
and under the condition that the first preset condition is met, the decoding end decodes the coding unit according to a second pixel prediction value, wherein the second pixel prediction value is a pixel prediction value obtained by performing intra-frame prediction on the coding unit based on an intra-frame prediction mode corresponding to the intra-frame prediction mode value.
7. The method according to any one of claims 2 to 5, further comprising:
the decoding end determines target pixel points, wherein the target pixel points comprise pixel points positioned on the left side of the coding unit and pixel points positioned above the coding unit;
and the decoding end performs texture analysis on the target pixel point by using a preset texture analysis mode to obtain the first texture direction value and the second texture direction value.
8. The method according to any one of claims 2 to 5, further comprising:
the decoding end determines a first reference pixel point and a second reference pixel point, wherein the first reference pixel point is a pixel point positioned on the left side of the coding unit, and the second reference pixel point is a pixel point positioned above the coding unit;
and the decoding end obtains the first texture direction value corresponding to the first reference pixel point and the second texture direction value corresponding to the second reference pixel point by using a preset texture analysis mode.
9. A video encoding method, comprising:
the encoding end determines an encoding unit of a video to be encoded;
the coding end codes the coding unit to obtain a target code stream;
the target code stream comprises offset information, the offset information is used for indicating an offset value, the offset value is used for determining a threshold value of a target interval range in a first preset condition, and the first preset condition is a trigger condition for obtaining a pixel predicted value without using a position-dependent prediction combination PDPC method.
10. The method of claim 9, wherein prior to determining the coding unit of the video to be encoded, the method further comprises:
and the encoding end determines the offset information according to the image information of the video to be encoded, wherein the image information comprises at least one of image type and image content.
11. A video decoding apparatus, comprising:
the decoding device comprises an acquisition module, a decoding module and a decoding module, wherein the acquisition module is used for acquiring a first pixel predicted value of each pixel point in a coding unit to be decoded by using a position-dependent prediction combination (PDPC) method under the condition that target information does not meet a first preset condition, and the target information comprises information used for determining whether the PDPC method is used for acquiring the pixel predicted value;
a first determining module to decode the coding unit according to the first pixel prediction value.
12. The apparatus of claim 11, wherein the target information comprises at least one of first information, decoder intra mode derived DIMD identification information, template-based intra mode derived TIMD identification information, a height value of the coding unit, and a predicted area of the coding unit for performing the PDPC method, wherein the first information comprises at least two of an intra prediction mode value, a first texture direction value, and a second texture direction value, and wherein the first texture direction value and the second texture direction value are texture direction values of a reference pixel of the coding unit.
13. The apparatus of claim 12, wherein the first preset condition comprises at least one of:
the intra-frame prediction mode value is greater than or equal to a first preset value, and the first texture direction value or the second texture direction value is located in a first interval range;
the intra-frame prediction mode value is smaller than or equal to a second preset value, and the first texture direction value or the second texture direction value is located in a second interval range;
one of the first texture direction value and the second texture direction value is positioned in a third interval range, and the other texture direction value is positioned in a fourth interval range;
the intra prediction mode value is within a fifth interval range, and the first texture direction value or the second texture direction value is within a sixth interval range;
the intra prediction mode value is located in a seventh interval range, and the first texture direction value or the second texture direction value is located in an eighth interval range;
the TIMD identification information is a first preset value;
the DIMD identification information is a second preset value;
the height value is in a ninth interval range;
the coding unit is used for executing the estimated area of the PDPC and is positioned in the tenth interval range;
the first preset value is larger than the second preset value, and the minimum threshold of the fifth interval range is larger than the maximum threshold of the seventh interval range.
14. A video encoding apparatus, comprising:
the second determining module is used for determining the coding unit of the video to be coded;
the coding module is used for coding the coding unit to obtain a target code stream;
the target code stream comprises offset information, the offset information is used for indicating an offset value, the offset value is used for determining a threshold value of a target interval range in a first preset condition, and the first preset condition is a trigger condition for obtaining a pixel predicted value without using a position-dependent prediction combination PDPC method.
15. The apparatus of claim 14, wherein the second determining module is further configured to determine the offset information according to image information of the video to be encoded, and the image information comprises at least one of an image type and an image content.
16. An electronic device, comprising: memory, processor and program stored on the memory and executable on the processor, the program implementing the steps in the video decoding method according to any one of claims 1 to 8 when executed by the processor or implementing the steps in the video encoding method according to any one of claims 9 to 10 when executed by the processor.
17. A readable storage medium, characterized in that it stores thereon a program or instructions which, when executed by a processor, implements the steps in the video decoding method according to any one of claims 1 to 8, or which, when executed by said processor, implements the steps in the video encoding method according to any one of claims 9 to 10.
CN202111122423.7A 2021-09-24 2021-09-24 Video decoding method, video encoding method and related equipment Pending CN115866263A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111122423.7A CN115866263A (en) 2021-09-24 2021-09-24 Video decoding method, video encoding method and related equipment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111122423.7A CN115866263A (en) 2021-09-24 2021-09-24 Video decoding method, video encoding method and related equipment

Publications (1)

Publication Number Publication Date
CN115866263A true CN115866263A (en) 2023-03-28

Family

ID=85652570

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111122423.7A Pending CN115866263A (en) 2021-09-24 2021-09-24 Video decoding method, video encoding method and related equipment

Country Status (1)

Country Link
CN (1) CN115866263A (en)

Similar Documents

Publication Publication Date Title
US20200260117A1 (en) Methods and Apparatuses for Coding and Decoding Depth Map
US20160044329A1 (en) Image Predictive Coding Method and Image Encoder
JP6334006B2 (en) System and method for high content adaptive quality restoration filtering for video coding
US10034021B2 (en) Methods and apparatuses for coding and decoding depth map
WO2016050051A1 (en) Image prediction method and relevant device
EP3550841A1 (en) Video encoding and decoding with improved error resilience
US10477249B2 (en) Video processing for masking coding artifacts using dynamic noise maps
CN104067619A (en) Video decoder, video encoder, video decoding method, and video encoding method
CN1981533A (en) Method and system for performing deblocking filtering
CN109587491A (en) A kind of intra-frame prediction method, device and storage medium
JP2018533273A5 (en)
US9712828B2 (en) Foreground motion detection in compressed video data
CN111654696B (en) Intra-frame multi-reference-line prediction method and device, storage medium and terminal
CN103530895A (en) Method and device for detecting mosaic
CN105100799A (en) Method for reducing intraframe coding time delay in HEVC encoder
CN115379214B (en) Image processing method, intelligent terminal and storage medium
CN116567246A (en) AVC coding method and device
CN115866263A (en) Video decoding method, video encoding method and related equipment
US9769489B2 (en) Error tracking and mitigation for motion compensation-based video compression
CN112738522A (en) Video coding method and device
WO2023198144A1 (en) Inter-frame prediction method and terminal
CN110933419B (en) Method and equipment for determining motion vector and boundary strength
CN115883833A (en) Intra-frame prediction method and device
CN116320428A (en) Inter-frame prediction method and terminal
CN115474051A (en) Point cloud encoding method, point cloud decoding method and terminal

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination