WO2023123284A1 - Procédé de décodage, procédé de codage, décodeur, codeur et support de stockage - Google Patents

Procédé de décodage, procédé de codage, décodeur, codeur et support de stockage Download PDF

Info

Publication number
WO2023123284A1
WO2023123284A1 PCT/CN2021/143372 CN2021143372W WO2023123284A1 WO 2023123284 A1 WO2023123284 A1 WO 2023123284A1 CN 2021143372 W CN2021143372 W CN 2021143372W WO 2023123284 A1 WO2023123284 A1 WO 2023123284A1
Authority
WO
WIPO (PCT)
Prior art keywords
value
attribute
color component
prediction
component
Prior art date
Application number
PCT/CN2021/143372
Other languages
English (en)
Chinese (zh)
Inventor
魏红莲
Original Assignee
Oppo广东移动通信有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Oppo广东移动通信有限公司 filed Critical Oppo广东移动通信有限公司
Priority to PCT/CN2021/143372 priority Critical patent/WO2023123284A1/fr
Publication of WO2023123284A1 publication Critical patent/WO2023123284A1/fr

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/186Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/597Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding

Definitions

  • the embodiment of the present application relates to a video encoding technology, and relates to but not limited to a decoding method, an encoding method, a decoder, an encoder, and a storage medium.
  • attribute information coding is mainly aimed at the coding of color information (that is, color components), so as to transform color information from the spatial domain to the frequency domain, and convert it into Luminance-Chrominance (YUV) that is more in line with the visual characteristics of the human eye. color space, and then perform attribute encoding on the preprocessed attribute information to obtain a quantized residual value, and input the quantized residual value into the attribute entropy encoder to form a binary attribute code stream.
  • color information that is, color components
  • YUV Luminance-Chrominance
  • the point cloud compression reference platform (Point Cloud Reference Model, PCRM) conducts secondary prediction of attribute residual values based on cross-component attribute residuals, and further eliminates the correlation between the three color components. redundancy.
  • PCRM Point Cloud Reference Model
  • the secondary prediction algorithm of the residual value of the attribute uses the R component to predict the G component, and uses the sum of the R component and the G component to predict the B component.
  • this prediction method is not accurate.
  • Embodiments of the present application provide a decoding method, an encoding method, a decoder, an encoder, and a storage medium, which can improve the accuracy of encoding and decoding during attribute encoding.
  • the embodiment of the present application provides a decoding method, which is applied to a decoder, and the method includes:
  • the attribute residual value, the attribute prediction value and the initial cross-component attribute residual prediction value decode and reconstruct the color component of the current point to obtain the attribute reconstruction value of the color component.
  • the embodiment of the present application also provides an encoding method, which is applied to an encoder, and the method includes:
  • the color component of the current point is predicted twice to obtain the attribute residual value of the color component.
  • the embodiment of the present application provides a decoder, including:
  • the parsing part is configured to parse the code stream and determine the attribute residual value of the color component corresponding to the current point;
  • the first acquisition part is configured to acquire the attribute prediction value of the color component corresponding to the current point
  • the first determining part is configured to determine a prediction mode of the color component of the current point; wherein the prediction mode is determined based on the degree of difference between the color components;
  • the decoding part is configured to decode and reconstruct the color component of the current point based on the prediction mode, the attribute residual value, the attribute prediction value and the initial cross-component attribute residual prediction value, to obtain the color component Property reconstruction value.
  • an encoder including:
  • the second acquisition part is configured to acquire the attribute prediction value of the color component corresponding to the current point in the point cloud;
  • the second determination part is configured to determine a standard value of the color component corresponding to the current point based on the attribute prediction value; determine at least two color components corresponding to the current point, and at least the standard value two degrees of difference; based on the at least two degrees of difference, determining a prediction mode of the color component;
  • the prediction part is configured to perform secondary prediction on the color component of the current point based on the prediction mode, the attribute prediction value and the initial cross-component attribute residual prediction value, to obtain the attribute residual value of the color component.
  • the embodiment of the present application also provides a decoder, including:
  • the first memory stores a computer program that can run on the first processor, and the first processor implements the decoding method of the decoder when executing the program.
  • the embodiment of the present application also provides an encoder, including:
  • the second memory stores a computer program that can run on the second processor, and the second processor executes the program and the encoding method of the encoder.
  • the embodiment of the present application provides a storage medium on which a computer program is stored, and when the computer program is executed by the first processor, the decoding method of the claim decoder is realized; or, the computer program is executed by the second processor When executed, the encoding method of the claim encoder is realized.
  • the embodiment of the present application provides a decoding method, an encoding method, a decoder, an encoder, and a storage medium.
  • determine the attribute residual value of the color component corresponding to the current point obtain the attribute of the color component corresponding to the current point Prediction value; determine the prediction mode of the color component of the current point; where the prediction mode is determined based on the degree of difference between the color components; based on the prediction mode, attribute residual value, attribute prediction value and initial cross-component attribute residual prediction Value, decode and reconstruct the color component of the current point to obtain the attribute reconstruction value of the color component. Since the adaptive selection of the prediction mode can be performed based on the degree of difference between the color components, when the decoding and reconstruction of the color component of the current point is performed according to the prediction mode, the accuracy of the attribute reconstruction value of the obtained color component higher.
  • Figure 1A is an exemplary three-dimensional point cloud image provided by the embodiment of the present application.
  • FIG. 1B is a partially enlarged view of an exemplary three-dimensional point cloud image provided by an embodiment of the present application
  • Figure 2A- Figure 2F is an exemplary point cloud image at different angles provided by the embodiment of the present application.
  • Fig. 2G is an exemplary illustration of the data storage format corresponding to Fig. 2A-Fig. 2F provided by the embodiment of the present application;
  • FIG. 3 is a schematic diagram of the composition and structure of an exemplary video codec network architecture provided by an embodiment of the present application
  • FIG. 4 is a structural diagram of an exemplary video coding system provided by an embodiment of the present application.
  • FIG. 5 is a structural diagram of an exemplary video decoding system provided by an embodiment of the present application.
  • Fig. 6A is an exemplary bounding box illustration provided by the embodiment of the present application.
  • 6B-6G are exemplary schematic diagrams of iterative octree division of bounding boxes provided by the embodiment of the present application.
  • FIG. 7A is a schematic diagram 1 of an exemplary encoding sequence of Morton codes in two-dimensional space provided by the embodiment of the present application;
  • FIG. 7B is a second schematic diagram of the coding sequence of an exemplary Morton code in a two-dimensional space provided by the embodiment of the present application.
  • FIG. 7C is a schematic diagram 3 of an exemplary Morton code encoding sequence in two-dimensional space provided by the embodiment of the present application.
  • FIG. 8 is a schematic diagram of an exemplary Morton code encoding sequence in three-dimensional space provided by an embodiment of the present application.
  • FIG. 9 is a schematic diagram of wavelengths of exemplary different color components provided by the embodiment of the present application.
  • FIG. 10 is a flow chart of a decoding method further provided in the embodiment of the present application.
  • FIG. 11 is a flow chart of an encoding method further provided in the embodiment of the present application.
  • FIG. 12 is a first schematic structural diagram of a decoder provided by an embodiment of the present application.
  • FIG. 13 is a second schematic structural diagram of a decoder provided in an embodiment of the present application.
  • FIG. 14 is a first schematic structural diagram of an encoder provided by an embodiment of the present application.
  • FIG. 15 is a second structural schematic diagram of an encoder provided by an embodiment of the present application.
  • point cloud compression Point Cloud Compression, PCC
  • Geometry-based Point Cloud Compression G-PCC
  • video-based point cloud compression Video Point Cloud Compression, V-PCC
  • level of detail Level of Detail, LOD
  • area adaptive analysis Region Adaptive Hierarchal Transform RAHT
  • slice bounding box
  • octree triangle soup, trisoup, block, vertex , Root Node (RootNode)
  • MPEG Moving Picture Experts Group
  • ISO International Standardization Organization
  • ISO International Electrotechnical Commission
  • audio and video coding standard AVS.
  • Point Cloud is a three-dimensional representation of the surface of an object.
  • the point cloud (data) on the surface of an object can be collected through acquisition devices such as photoelectric radar, lidar, laser scanner, and multi-view camera.
  • a point cloud is a set of discrete point sets randomly distributed in space that express the spatial structure and surface properties of a 3D object or scene.
  • Figure 1A shows a 3D point cloud image
  • Figure 1B shows a partial enlarged view of a 3D point cloud image. It can be seen that the point cloud surface is composed of densely distributed points.
  • the two-dimensional image has information expression at each pixel, and the distribution is regular, so there is no need to additionally record its position information; however, the distribution of points in the point cloud in three-dimensional space is random and irregular, so it is necessary to record each
  • the position of the point in space can completely express a point cloud.
  • each position in the acquisition process has corresponding attribute information, usually RGB color value, and the color value reflects the color of the object; for point cloud, the attribute information corresponding to each point is in addition to color information. , and the more common one is the reflectance value, which reflects the surface material of the object. Therefore, the points in the point cloud can include point location information and point attribute information.
  • the position information of the point may be three-dimensional coordinate information (x, y, z) of the point.
  • the location information of a point may also be referred to as geometric information of a point.
  • the point attribute information may include color information (three-dimensional color information) and/or reflectance (one-dimensional reflectance information r) and the like.
  • color information may be information on any color space.
  • color information may be RGB information. Wherein, R represents red (Red, R), G represents green (Green, G), and B represents blue (Blue, B).
  • the color information may be luminance chrominance (YCbCr, YUV) information. Among them, Y represents brightness (Luma), Cb(U) represents blue color difference, and Cr(V) represents red color difference.
  • the points in the point cloud can include the three-dimensional coordinate information of the point and the reflectance value of the point.
  • the points in the point cloud may include the three-dimensional coordinate information of the point and the three-dimensional color information of the point.
  • the point cloud is obtained by combining the principles of laser measurement and photogrammetry.
  • the points in the point cloud can include the three-dimensional coordinate information of the point, the reflectance value of the point and the three-dimensional color information of the point.
  • Figure 2A- Figure 2G shows a point cloud image and its corresponding data storage format, wherein Figure 2A- Figure 2F provides six viewing angles of the point cloud image;
  • Figure 2G consists of the file header information part and data
  • the header information includes the data format, data representation type, total point cloud points, and the content represented by the point cloud.
  • the point cloud is in ".ply" format, represented by ASCII code, the total number of points is 207242, and each point has three-dimensional coordinate information (x, y, z) and three-dimensional color information RGB (Red, Green, Blue).
  • Point clouds can be divided into the following ways:
  • Static point cloud that is, the object is stationary, and the device that acquires the point cloud is also stationary;
  • Dynamic point cloud the object is moving, but the device for obtaining the point cloud is still;
  • Dynamically obtain point cloud The device for obtaining point cloud is in motion.
  • point cloud For example, according to the purpose of point cloud, it is divided into two categories:
  • Category 1 Machine perception point cloud, which can be used in scenarios such as autonomous navigation systems, real-time inspection systems, geographic information systems, visual sorting robots, and emergency rescue robots;
  • Category 2 Human eyes perceive point clouds, which can be used in point cloud application scenarios such as digital cultural heritage, free viewpoint broadcasting, 3D immersive communication, and 3D immersive interaction.
  • Point cloud can flexibly and conveniently express the spatial structure and surface properties of three-dimensional objects or scenes, and because point cloud is obtained by directly sampling real objects, it can provide a strong sense of reality under the premise of ensuring accuracy, so it is widely used.
  • the collection of point clouds mainly has the following methods: computer generation, 3D laser scanning, 3D photogrammetry, etc.
  • Computers can generate point clouds of virtual three-dimensional objects and scenes; 3D laser scanning can obtain point clouds of static real-world three-dimensional objects or scenes, and can obtain millions of point clouds per second; 3D photogrammetry can obtain dynamic real-world three-dimensional objects or scenes The point cloud of tens of millions of points can be obtained per second.
  • These technologies reduce the cost and time period of point cloud data acquisition, and improve the accuracy of the data.
  • the transformation of point cloud data acquisition methods has made it possible to acquire a large amount of point cloud data. With the growth of application requirements, the processing of massive 3D point cloud data encounters the bottleneck of storage space and transmission bandwidth limitations.
  • the number of points in each frame of point cloud is 700,000, and each point has coordinate information xyz (float) and color information RGB (uchar)
  • the data volume of 10s is about 1280 ⁇ 720 ⁇ 12bit ⁇ 24fps ⁇ 10s ⁇ 0.33GB
  • point cloud compression has become a key issue to promote the development of the point cloud industry.
  • point cloud is a collection of massive points
  • storing point cloud will not only consume a lot of memory, but also is not conducive to transmission, and there is not such a large bandwidth to support the direct transmission of point cloud at the network layer without compression, so , need to compress the point cloud.
  • the point cloud coding framework that can compress the point cloud can be the G-PCC codec framework or the V-PCC codec framework provided by MPEG, or the AVS-PCC codec framework provided by AVS.
  • the G-PCC codec framework can be used to compress the first type of static point cloud and the third type of dynamically acquired point cloud
  • the V-PCC codec framework can be used to compress the second type of dynamic point cloud.
  • the G-PCC codec framework is also called point cloud codec TMC13
  • the V-PCC codec framework is also called point cloud codec TMC2.
  • the embodiment of the present application provides a network architecture of a video encoding and decoding system including a decoding method and an encoding method.
  • FIG. It includes one or more electronic devices 13 to 1N and a communication network 01 , where the electronic devices 13 to 1N can perform video interaction through the communication network 01 .
  • the electronic device may be various types of devices with video codec functions during implementation, for example, the electronic device may include a mobile phone, a tablet computer, a personal computer, a personal digital assistant, a navigator, a digital phone, a video phone, TV sets, sensing devices, servers, etc., are not limited in this embodiment of the application.
  • the decoder or encoder in the embodiment of the present application may be the above-mentioned electronic device.
  • the electronic device in the embodiment of the present application has a video encoding and decoding function, and generally includes a video encoder (ie, an encoder) and a video decoder (ie, a decoder).
  • a video encoder ie, an encoder
  • a video decoder ie, a decoder
  • the codec framework is described in the AVS-PCC codec framework.
  • Point cloud compression generally uses point cloud geometric information and attribute information to compress separately. Compression of cloud attributes; at the decoding end, the geometric information of the point cloud is first decoded in the geometric decoder, and then the decoded geometric information is input into the attribute decoder as additional information to assist in the compression of the point cloud attributes.
  • the entire codec consists of preprocessing/postprocessing, geometric encoding/decoding, and attribute encoding/decoding.
  • the present application provides a video encoding system, as shown in FIG. 4 , the framework of the point cloud compression reference platform PCRM of AVS point cloud.
  • the video encoding system 11 includes a geometric encoder: a coordinate translation unit 111, a coordinate quantization unit 112, and an octave Tree construction unit 113 , geometric entropy encoder 114 , geometric reconstruction unit 115 .
  • Attribute encoder attribute recoloring unit 116 , color space transformation unit 117 , first attribute prediction unit 118 , quantization unit 119 and attribute entropy encoder 1110 .
  • the original geometric information is firstly preprocessed, the geometric origin is normalized to the minimum value position in the point cloud space by the coordinate translation unit 111, and the geometric information is converted from the floating point by the coordinate quantization unit 112 Points are converted into plastics, which is convenient for subsequent regularization processing; then the regularized geometric information is geometrically coded, and the point cloud space is recursively divided by the octree structure through the octree construction unit 113, and each time the current node is divided into Eight sub-blocks of the same size, and judge the occupancy code word of each sub-block.
  • the sub-block When the sub-block does not contain a point, it is recorded as empty, otherwise it is recorded as non-empty, and the occupancy code of all blocks is recorded in the last layer of recursive division.
  • word information, and carry out geometric encoding the geometric information expressed by the octree structure is input into the geometric entropy encoder 114 on the one hand to form a geometric code stream; on the other hand, the geometric reconstruction unit 115 in the encoder performs geometric reconstruction processing.
  • the geometric information of is fed into the attribute encoder as additional information.
  • the original attribute information is firstly preprocessed. Since the geometric information has changed after the geometric encoding, the attribute recoloring unit 116 is used to reassign the attribute value for each point after the geometric encoding to realize the attribute Recolor.
  • the attribute information to be processed is color information
  • the original color information needs to be transformed into a YUV color space that is more in line with the visual characteristics of the human eye through the color space transformation unit 117;
  • Unit 118 performs attribute encoding or prediction on the preprocessed attribute information.
  • the attribute prediction first needs to reorder the point cloud, and the reordering method is Morton code. Therefore, the traversal order of attribute encoding or attribute prediction is Morton order.
  • the attribute prediction method in PCRM is a single-point prediction based on Morton order, that is, according to the Morton order, one point is traced forward from the current point to be encoded (current point), and the neighbor point found is the prediction reference point of the current point to be encoded. Then the attribute reconstruction value of the predicted reference point is used as the attribute prediction value, and the attribute residual value is the difference between the attribute value of the current point to be encoded and the attribute prediction value; finally, the residual is quantized by the quantization unit 119, and the quantized residual The difference is input to the attribute entropy encoder 1110 to form an attribute code stream.
  • the present application provides a video decoding system, as shown in Figure 5, the framework of the point cloud compression reference platform PCRM of AVS point cloud, the video decoding system 12 includes a geometric decoder: a geometric entropy decoder 121, an octree reconstruction unit 122 .
  • the coordinate inverse quantization unit 123 and the coordinate inverse translation unit 124 .
  • Attribute decoder attribute entropy decoder 125 , inverse quantization unit 126 , second attribute prediction unit 127 and color space inverse transform unit 128 .
  • the geometry and attributes are also decoded separately.
  • the geometric decoding part first, the geometric code stream is entropy decoded by the geometric entropy decoder 121 to obtain the geometric information of each node, and then the octree structure is constructed by the octree reconstruction unit 122 in the same manner as the geometric encoding, Combining the decoded geometry to reconstruct the geometric information expressed through the octree structure after the coordinate transformation, on the one hand, the information is subjected to coordinate inverse quantization by the coordinate inverse quantization unit 123 and detranslated by the coordinate inverse translation unit 124 to obtain the decoded geometric information. On the other hand, it is input into the attribute decoder as additional information.
  • the Morton sequence is constructed in the same way as the encoding end, and the attribute code stream is first entropy decoded by the attribute entropy decoder 125 to obtain the quantized residual information; then the inverse quantization is performed by the inverse quantization unit 126, Obtain the point cloud residual value; similarly, in the same manner as the attribute encoding, obtain the attribute prediction value of the point to be decoded by the second attribute prediction unit 127, and then add the attribute prediction value and the attribute residual value to restore The YUV attribute value of the current point to be decoded is obtained; finally, the decoded attribute information is obtained through the color space inverse transformation of the color space inverse transformation unit 128.
  • the recursive octree structure is used to express the point in the point cloud as the center of the cube in a regular way, as shown in Figure 6A.
  • the boundary values of the point cloud in the x, y, and z directions are formula (1) - formula (6):
  • x min min(x 0 ,x 1 ,...,x K-1 ) (1)
  • x man max(x 0 ,x 1 ,...,x K-1 ) (4)
  • the origin (x origin , y origin , z origin ) of the bounding box can be calculated as formula (7) - formula (9):
  • the bounding box is first divided into octrees, and each time eight sub-blocks are obtained, and then the non-empty blocks (blocks containing points) in the sub-blocks are divided again
  • the octree division of so recursively divided until a certain depth, the non-empty sub-block of the final size is called voxel, each voxel contains one or more points, and the geometric positions of these points are normalized to the center point of the voxel , the attribute value of the center point is the average value of the attribute values of all points in the voxel, and finally the image shown in Figure 6G is obtained.
  • regularizing the point cloud into blocks in space is beneficial to the description of the relationship between points in the point cloud, and then can express a specific encoding sequence, and encode each voxel in a certain order, that is, the encoded voxel represents The points (or "nodes") of , a commonly used coding order is the cross-separated Morton order.
  • a commonly used coding order is the cross-separated Morton order.
  • the encoding sequence of the Morton code in the two-dimensional space in Fig. 7A-Fig. 7C is used, taking a block of size 8*8 as an example, where the order of the arrows represents the encoding order of the points under the Morton order. As shown in FIG.
  • FIG. 7A the "z"-shaped Morton coding order of 2*2 pixels in the block is shown.
  • Figure 7B shows the "z”-shaped Morton coding sequence between four 2*2 blocks
  • Figure 7C shows the "z”-shaped Morton coding sequence between four 4*4 blocks, which form the entire 8*8 block Morton coding sequence.
  • the Morton coding sequence extended to the three-dimensional space is shown in Figure 8, which shows 16 nodes, and the Morton coding sequence between each "z” and “z” inside each "z” is First encode along the x-axis, then along the y-axis, and finally along the z-axis.
  • the attribute residual quadratic prediction algorithm currently used by PCRM uses the R component to predict the G component, and uses the sum of the R component and the G component to predict the B component, which can effectively eliminate the correlation between the three components. Thereby improving the coding efficiency.
  • the wavelength distribution range of red light is 620-750nm
  • the wavelength distribution range of green light is 495-570nm
  • the wavelength distribution range of blue light is 450-475nm, as shown in Figure 9, according to the From the point of view of the coincidence degree, it is obvious that it is not very reasonable to use the R component to predict the other two components, and the prediction is not accurate and fixed.
  • An embodiment of the present application provides a decoding method, which is applied to a video decoding device, that is, a decoder.
  • the function realized by this method can be realized by calling the program code by the first processor in the video decoding device.
  • the program code can be stored in the computer storage medium.
  • the video decoding device includes at least the first processor and the first memory medium. Wherein, the current decoding point and the current encoding point are both represented by the current point below.
  • FIG. 10 is a schematic diagram of an implementation flow of a decoding method according to an embodiment of the present application. The method includes:
  • the decoder can analyze the attribute residual value of the color component corresponding to the current point from the code stream.
  • the attribute residual value analyzed in the code stream is a quantized residual value.
  • the color component at the current point may include: a first color component, a second color component, and a third color component.
  • the first color component, the second color component and the third color component may be three RGB components respectively.
  • the points in the point cloud all have attribute information of RGB three components.
  • the decoder can parse out the attribute residual values corresponding to each color component of the current point.
  • the decoder needs to perform decoding processing on each color component of the current point.
  • the encoding can be based on the single-point prediction of the Morton order, that is, one point is traced forward from the current point according to the Morton order, the found point is the prediction reference point of the current point, and then the prediction reference point
  • the attribute reconstruction value of is used as the attribute prediction value.
  • the decoder predicts the attribute of the current point, it can obtain the attribute reconstruction value of the prediction reference point of the current point, that is, obtain the attribute prediction value of the color component corresponding to the current point.
  • the decoder can perform secondary attribute prediction to eliminate redundancy after decoding the attribute residual value. Therefore, in the process of secondary attribute prediction, the decoder can determine the prediction mode of the color component of the current point during the secondary attribute prediction, and then perform secondary attribute prediction according to the prediction mode, and combine the attribute residual value and attribute prediction value and the initial cross-component attribute residual prediction value, decode and reconstruct the color component of the current point, and obtain the attribute reconstruction value of the color component.
  • the prediction mode represents the encoding and decoding sequence of the color components or represents the prediction form of the color components (for example, which color component is encoded and decoded first). Wherein, the prediction mode is determined based on the degree of difference between the color components.
  • the prediction mode includes: the first color component to be predicted; the first color component to be predicted is the first coded color among the first color component, the second color component and the third color component portion.
  • the decoder can adaptively decide how to decode and reconstruct the color component by first encoding the first color component to be predicted. Considering the adaptive prediction form of the color component, the secondary attribute prediction can be improved. the accuracy.
  • S1031 and S1032 may be used to determine the prediction mode, and S1033-1035 may also be used to implement, which is not limited in this embodiment of the present application.
  • the prediction mode flag is transmitted in the code stream, so that when decoding the current point, the decoder can simultaneously parse out the current point from the code stream
  • the prediction mode flag of the color component when performing secondary attribute prediction wherein the prediction mode flag indicates which color component is coded first. Therefore, the decoder can determine the prediction mode of the color component at the current point according to the prediction mode flag bit.
  • the prediction mode flag can indicate which color component is coded first.
  • the prediction mode flag can be represented in digital form. For example, 0 can be used to represent the first color component, 1 can be used to represent the second color component, and 2 can be used to represent the third color component. This embodiment of the present application does not limit it.
  • the prediction mode flag is determined based on the degree of difference between color components during encoding.
  • the decoder can determine the prediction mode through the indication of the prediction mode flag, without performing calculation of the prediction mode, which can save decoding time and improve decoding efficiency.
  • the decoder can also determine the prediction mode in a way that does not require the prediction mode flag bit.
  • the decoder can determine a standard value for the color component according to the attribute prediction value corresponding to the current point, and then compare the attribute prediction value of each color component of the current point with the standard value to determine at least two color components. degree of difference, and then determine the prediction mode based on at least two degrees of difference.
  • the property prediction value includes: a first property prediction value of the first color component, a second property prediction value of the second color component, and a third property prediction value of the third color component.
  • the decoder determines the standard value of the color component corresponding to the current point based on the attribute prediction value, including at least one of the following:
  • the median value of the first attribute predicted value, the second attribute predicted value and the third attribute predicted value is determined as a standard value.
  • the standard value can also be obtained according to other mathematical processing means of the predicted value of the first attribute, the predicted value of the second attribute and the predicted value of the third attribute, which is not limited in this embodiment of the present application.
  • the decoder after the decoder determines the standard value, it can compare the attribute prediction value of each color component of the current point with the standard value to obtain three color difference degrees, or it can compare the attribute prediction value of each color component in the current point The attribute prediction values of any two color components are compared with the standard values respectively, and two difference degrees corresponding to any two color components are obtained.
  • the decoder may determine the minimum degree of difference among at least two degrees of difference; and determine the color component corresponding to the minimum degree of difference as the prediction mode.
  • the decoder determines a sorting result of at least two different degrees based on the at least two different degrees; and determines a prediction mode according to the sorting result.
  • the prediction mode indicates that the first predicted color component to be predicted is firstly decoded in the second attribute prediction of the current point, and then the second attribute prediction is performed on other color components.
  • the decoder determines the sorting results of the at least two difference degrees; according to the sorting results, the prediction mode corresponding to the color component with the smallest difference can be determined. Moreover, when the secondary attribute prediction of the color components of the current point is performed, the decoding order of the color components can be determined from the order of the difference degrees from small to large according to the sorting results.
  • the decoder calculates the difference between the R component and the G component and the standard value (that is, the difference diff R and diff G ), and compares the two differences, and if diff R ⁇ diff G , then determine that the R component is the prediction mode for secondary prediction of attribute residuals, otherwise use the G component as the prediction mode for secondary prediction of attribute residuals.
  • the decoder calculates the differences (diff R , diff G, and diff B ) between the R component, the G component, and the B component and the standard value, and compares them. If diff R in diff R , diff G , and diff B If it is the smallest, determine the R component as the prediction mode for secondary prediction of the attribute residual; if the diff G is the smallest, use the G component as the prediction mode for the secondary prediction of the attribute residual; if the diff B is the smallest, use the B component as the prediction mode Quadratic prediction of attribute residuals.
  • the decoder determines the prediction mode of the current point, it first decodes the first color component to be predicted, and then decodes the other color components (the second color component to be predicted) based on the first color component to be predicted. component and the third color component to be predicted) for decoding and reconstruction (that is, secondary attribute prediction), or based on the first color component to be predicted, the second color component to be predicted is decoded and reconstructed, and the second color component to be predicted is used The component decodes and reconstructs the third color component to be predicted.
  • the second to-be-predicted color component and the third to-be-predicted color component are color components other than the first to-be-predicted color component among the first, second, and third color components.
  • the implementation of S104 may include:
  • the prediction mode based on the attribute residual value of the first color component to be predicted, determine the reconstruction value of the first attribute residual of the first color component to be predicted; wherein, the first color component to be predicted is the prediction mode The color component of the representation.
  • the decoder may dequantize the attribute residual value of the first to-be-predicted color component according to the prediction mode, and determine the first reconstructed attribute value of the first to-be-predicted color component. Since each point can have three color components, when performing secondary attribute prediction on the current point, each color component needs to be processed to complete the decoding and reconstruction of the current point.
  • the attribute reconstruction value includes: the first attribute reconstruction value corresponding to the first to-be-predicted color component, the second attribute reconstruction value corresponding to the second to-be-predicted color component, and the third to-be-predicted color component The third attribute reconstruction value corresponding to the predicted color component.
  • the decoder since the first color component to be predicted is the color component corresponding to the prediction mode, the decoder first decodes and reconstructs the first color component to be predicted, and then based on the first color component to be predicted component continues to decode other color components.
  • the decoder can add the initial cross-component attribute residual prediction value, the attribute prediction value corresponding to the first to-be-predicted color component, and the first attribute residual reconstruction value to obtain the first to-be-predicted color component A property reconstruction value.
  • the second cross-component attribute residual prediction value is the sum of the first attribute residual reconstruction value and the initial cross-component attribute residual prediction value, or is the first attribute residual Multiplier or divisor of the difference reconstruction value.
  • the third cross-component attribute residual prediction value is the sum of the first attribute residual reconstruction value and the initial cross-component attribute residual prediction value, or is the first attribute residual Multiplier or divisor of the difference reconstruction value.
  • the initial cross-component attribute residual prediction value may be 0, which is not limited in this embodiment of the present application.
  • the decoder uses the cross-component attribute residual prediction value to perform secondary attribute prediction, but the cross-component attribute residual prediction values corresponding to different color components of a point are related, but may not be the same .
  • the first color component to be predicted uses the initial cross-component attribute residual prediction value to perform secondary attribute prediction
  • the second color component to be predicted uses the second cross-component attribute residual prediction value to perform
  • the third color component to be predicted uses the third cross-component attribute residual prediction value for secondary attribute prediction.
  • the decoder decodes and reconstructs other color components based on the second cross-component attribute residual prediction value, and the realization of obtaining the second attribute reconstruction value and the third attribute reconstruction value may include:
  • decode and reconstruct the second color component to be predicted to obtain the second attribute reconstruction value; based on the second attribute residual reconstruction value, determine the third cross-component attribute residual difference prediction value; based on the third cross-component attribute residual prediction value, decode and reconstruct the third to-be-predicted color component to obtain the third attribute reconstruction value.
  • the third cross-component attribute residual prediction value is the sum of the second attribute residual reconstruction value and the second cross-component attribute residual prediction value, or is the second attribute The multiplier or divisor of the residual reconstruction value.
  • the first color component to be predicted is the G component
  • the second color component to be predicted is the R component
  • the third color component to be predicted is the B component. Then, when the decoder performs secondary attribute prediction on the G component, it uses the initial cross-component attribute residual prediction value to obtain the first attribute reconstruction value of the G component.
  • the first attribute residual reconstruction value of the G component can be used to update the initial cross-component attribute residual prediction value to obtain the second cross-component attribute residual value difference prediction value
  • the first attribute residual reconstruction value of the G component can be used to update the initial cross-component attribute residual prediction value to obtain the third cross-component attribute residual prediction value
  • the first attribute residual prediction value of the R component can also be used
  • the reconstruction value of the residual of the two attributes is updated, and the predicted value of the second cross-component attribute residual is updated to obtain the third predicted value of the cross-component attribute residual.
  • the first attribute residual reconstruction value of the G component can be used to update the initial cross-component attribute residual prediction value
  • the second cross-component attribute residual prediction value can be the first attribute residual reconstruction value + Obtain the second cross-component attribute residual prediction value for the initial cross-component attribute residual prediction value; or, use twice the first attribute residual reconstruction value as the second cross-component attribute residual prediction value.
  • the first attribute residual reconstruction value of the G component can be used to update the initial cross-component attribute residual prediction value, and the third cross-component attribute residual prediction value can be obtained, which can be the first attribute residual reconstruction value + the initial
  • the cross-component attribute residual predictors yield the third cross-component attribute residual predictors; alternatively, half the reconstructed values of the first attribute residuals are used as the third cross-component attribute residual predictors.
  • the adaptive selection of the prediction mode can be determined based on the prediction mode flag bit of the color component, when the decoding and reconstruction of the color component at the current point is performed according to the prediction mode, the properties of the obtained color component The accuracy of the reconstructed values is high.
  • the mean mean of the color three components of the attribute prediction value of the current point that is, the standard value (also can be the median, maximum value, minimum value, etc.);
  • the second cross-component attribute residual prediction value or the third cross-component attribute residual prediction value is set as a multiple or divisor of the attribute residual reconstruction value of the secondary prediction component. For example, when performing secondary prediction of attribute residuals on the R component, set the cross-component attribute residual prediction value to the attribute residual reconstruction value of the G component divided by 2; The predicted value of the component attribute residual is set as the reconstructed value of the attribute residual of the G component divided by 4.
  • the adaptive selection of the prediction mode can be performed based on the degree of difference between the color components, when the decoding and reconstruction of the color component of the current point is performed according to the prediction mode, the properties of the obtained color components The accuracy of the reconstructed values is high.
  • An embodiment of the present application provides an encoding method, which is applied to a video encoding device, that is, an encoder.
  • the function realized by this method can be realized by calling the program code by the second processor in the video encoding device, and of course the program code can be stored in the computer storage medium.
  • the video encoding device includes at least the second processor and the second storage medium.
  • Fig. 11 is a schematic diagram of an implementation process of encoding according to an embodiment of the present application. The method includes:
  • S203 Determine at least two degrees of difference between at least two color components corresponding to the current point and a standard value.
  • an encoding method is proposed.
  • the encoder can obtain the attribute of the color component corresponding to the current point in the point cloud after the attribute prediction is performed by the first attribute unit. Predicted value, in order to eliminate redundancy at this time, you can continue to perform secondary attribute prediction on the color component of the current point.
  • the encoder can first determine a standard value for the color component according to the attribute prediction value of each color component, and then determine the difference between at least two color components of the current point and the standard value, that is, at least Two degrees of difference, and then determine the prediction mode based on at least two degrees of difference.
  • the standard value can be directly obtained from the attribute prediction value of the color component, or it can be determined based on the attribute prediction value of at least two color components combined with the attribute information of the color component to obtain the initial attribute residual reconstruction value.
  • This application implements Examples are not limited.
  • At least two degrees of difference can be obtained by directly comparing attribute prediction values of at least two color components.
  • the property prediction value includes: a first property prediction value of the first color component, a second property prediction value of the second color component, and a third property prediction value of the third color component.
  • the encoder determines the standard value of the color component corresponding to the current point based on the attribute prediction value, including at least one of the following:
  • the median value of the first attribute predicted value, the second attribute predicted value and the third attribute predicted value is determined as a standard value.
  • the standard value can also be obtained according to other mathematical processing means of the predicted value of the first attribute, the predicted value of the second attribute and the predicted value of the third attribute, which is not limited in this embodiment of the present application.
  • the encoder after the encoder determines the standard value, it can compare the attribute prediction value of each color component of the current point with the standard value to obtain three color difference degrees, or it can use the The attribute prediction values of any two color components are compared with the standard values respectively to obtain two degrees of difference corresponding to any two color components, which is not limited in this embodiment of the present application.
  • the encoder determines the initial attribute residual reconstruction value obtained according to attribute prediction values of at least two color components combined with attribute information of the color components.
  • the encoder determines the standard value of the color component corresponding to the current point based on the attribute prediction value as follows:
  • the encoder determines the initial residual value corresponding to the current point based on the predicted attribute value and the attribute information of the current point; quantizes and dequantizes the initial residual value to obtain the reconstruction value of the initial attribute residual.
  • the encoder determines the initial attribute residual reconstruction value of each color component corresponding to the current point, and based on the initial attribute residual of each color component Reconstruct the value to determine the standard value of the color component corresponding to the current point.
  • the initial attribute residual reconstruction value includes: the first initial attribute residual reconstruction value of the first color component, the second initial attribute residual reconstruction value of the second color component, and the third color component The residual reconstruction value of the third genus initial attribute;
  • the encoder determines the standard value of the color component corresponding to the current point based on the initial attribute residual reconstruction value, including at least one of the following:
  • the encoder can compare the initial attribute residual reconstruction value of each color component of the current point with the standard value to obtain three color difference degrees, or the current The initial attribute residual reconstruction values of any two color components in the point are compared with the standard values to obtain two degrees of difference corresponding to any two color components, which is not limited in this embodiment of the present application.
  • the standard value can also be obtained according to other mathematical processing methods of the first initial attribute residual reconstruction value, the second initial attribute residual reconstruction value, and the third initial attribute residual reconstruction value, which is not limited in this embodiment of the present application.
  • the encoder may determine the minimum degree of difference among at least two degrees of difference; and determine the color component corresponding to the minimum degree of difference as the prediction mode.
  • the encoder determines a sorting result of the at least two difference degrees based on the at least two difference degrees; and determines a prediction mode according to the sorting result.
  • the prediction mode indicates that the first predicted color component to be predicted is firstly decoded in the second attribute prediction of the current point, and then the second attribute prediction is performed on other color components.
  • the prediction mode represents the encoding and decoding sequence of the color components or represents the prediction form of the color components (for example, which color component is encoded and decoded first).
  • the prediction mode includes: the first color component to be predicted; the first color component to be predicted is the first coded color among the first color component, the second color component and the third color component portion.
  • the encoder determines the sorting results of the at least two difference degrees; according to the sorting results, the prediction mode corresponding to the color component with the smallest difference can be determined. Moreover, when the secondary attribute prediction of the color component of the current point is performed, the coding sequence of the color components can be determined from the order of the degree of difference from small to large according to the sorting result.
  • the encoder calculates the difference between the R component and the G component and the standard value (ie, the difference diff R and diff G ), and compares the two differences, and if diff R ⁇ diff G , then determine that the R component is the prediction mode for secondary prediction of attribute residuals, otherwise use the G component as the prediction mode for secondary prediction of attribute residuals.
  • the standard value ie, the difference diff R and diff G
  • the encoder calculates the differences (diff R , diff G and diff B ) between the R component, the G component, and the B component and the standard value, and compares them. If diff R among diff R , diff G and diff B If it is the smallest, determine the R component as the prediction mode for secondary prediction of the attribute residual; if the diff G is the smallest, use the G component as the prediction mode for the secondary prediction of the attribute residual; if the diff B is the smallest, use the B component as the prediction mode Quadratic prediction of attribute residuals.
  • the encoder when the encoder determines the prediction mode of the color component based on at least two differences, it can generate a prediction mode flag and write it into the code stream; the prediction mode flag represents the prediction mode.
  • the encoder does not need to generate the prediction mode flag, and directly performs the same determination process of the prediction mode on the decoder side. .
  • the encoder determines the prediction mode of the color component of the current point during the secondary attribute prediction, and then performs secondary attribute prediction according to the prediction mode, and combines the predicted value of the attribute and the initial cross-component attribute residual prediction value, perform secondary prediction on the color component of the current point, and obtain the attribute residual value of the color component.
  • the encoder can perform secondary attribute prediction according to the prediction mode, and combine the attribute information of each color component, the attribute prediction value and the initial cross-component attribute residual prediction value to carry out the color component of the current point Quadratic prediction to obtain the attribute residual value of the color component.
  • the encoder determines the prediction mode of the current point, it first performs second prediction on the first color component to be predicted, and then performs second prediction on the other color components (the second color component to be predicted) based on the first color component to be predicted. predicted color component and the third color component to be predicted) for secondary prediction, or secondly predict the second color component to be predicted based on the first color component to be predicted, and use the second color component to be predicted to predict the second color component Three to-be-predicted color components perform secondary prediction.
  • the second to-be-predicted color component and the third to-be-predicted color component are color components other than the first to-be-predicted color component among the first, second, and third color components.
  • the encoder combines the first attribute information of the first color component to be predicted with the attribute prediction value of the first color component to be predicted and the initial cross-component attribute residual prediction value Quantize after subtraction in turn to achieve secondary prediction, and obtain the attribute residual value of the first color component to be predicted; after inverse quantization of the attribute residual value of the first color component to be predicted, the first color to be predicted is obtained
  • the first attribute residual reconstruction value of the component based on the first attribute residual reconstruction value, the second cross-component attribute residual prediction value is determined; based on the second cross-component attribute residual prediction value, the other color components are
  • the second prediction is to obtain the attribute residual value of the second to-be-predicted color component and the attribute residual value of the third to-be-predicted color component; wherein, the second to-be-predicted color component and the third to-be-predicted color component are the Other color components except the first to-be-predicted color component among the first color component, the second color component
  • the encoder can use the first attribute information of the first color component to be predicted according to the prediction mode, subtract the attribute prediction value of the first color component to be predicted, and then subtract the initial cross-component attribute
  • the residual prediction value is quantized to obtain the attribute residual value of the first color component to be predicted. Since each point can have three color components, when performing secondary attribute prediction on the current point, each color component needs to be processed to complete the coding part of the current point's secondary attribute prediction.
  • the encoder performs secondary prediction on other color components based on the second cross-component attribute residual prediction value to obtain the attribute residual value of the second to-be-predicted color component and the third to-be-predicted color component.
  • the implementation of predicting the attribute residual value of the color component is:
  • the encoder performs secondary prediction on the second color component to be predicted based on the second cross-component attribute residual prediction value to obtain the attribute residual value of the second color component to be predicted; for the second color component to be predicted After the attribute residual value is dequantized, the second attribute residual reconstruction value of the second color component to be predicted is obtained; based on the second attribute residual reconstruction value, the third cross-component attribute residual prediction value is determined; based on The third cross-component attribute residual prediction value performs secondary prediction on the third to-be-predicted color component to obtain the attribute residual value of the third to-be-predicted color component.
  • the second cross-component attribute residual prediction value is the sum of the first attribute residual reconstruction value and the initial cross-component attribute residual prediction value, or is the first attribute residual Multiplier or divisor of the difference reconstruction value.
  • the initial cross-component attribute residual prediction value may be 0, which is not limited in this embodiment of the present application.
  • the encoder uses the cross-component attribute residual prediction value to perform secondary attribute prediction, but the cross-component attribute residual prediction values corresponding to different color components of a point are related, but may not be the same .
  • the first color component to be predicted uses the initial cross-component attribute residual prediction value to perform secondary attribute prediction
  • the second color component to be predicted uses the second cross-component attribute residual prediction value to perform
  • the third color component to be predicted uses the third cross-component attribute residual prediction value for secondary attribute prediction.
  • the encoder After the encoder dequantizes the attribute residual value of the first color component to be predicted, and obtains the first attribute residual reconstruction value of the first color component to be predicted, the encoder Based on the first attribute residual reconstruction value, determine the second cross-component attribute residual prediction value and the third cross-component attribute residual prediction value; based on the second cross-component attribute residual prediction value, determine the second cross-component attribute residual prediction value The predicted color component is predicted twice to obtain the attribute residual value of the second color component to be predicted; based on the third cross-component attribute residual prediction value, the third color component to be predicted is predicted twice to obtain the third The attribute residual value of a color component to be predicted.
  • the third cross-component attribute residual prediction value is the sum of the second attribute residual reconstruction value and the second cross-component attribute residual prediction value, or is the second attribute The multiplier or divisor of the residual reconstruction value.
  • the first color component to be predicted is the G component
  • the second color component to be predicted is the R component
  • the third color component to be predicted is the B component. Then, when the decoder performs secondary attribute prediction on the G component, it uses the initial cross-component attribute residual prediction value to obtain the first attribute reconstruction value of the G component.
  • the first attribute residual reconstruction value of the G component can be used to update the initial cross-component attribute residual prediction value to obtain the second cross-component attribute residual value difference prediction value
  • the first attribute residual reconstruction value of the G component can be used to update the initial cross-component attribute residual prediction value to obtain the third cross-component attribute residual prediction value
  • the first attribute residual prediction value of the R component can also be used
  • the reconstruction value of the residual of the two attributes is updated, and the predicted value of the second cross-component attribute residual is updated to obtain the third predicted value of the cross-component attribute residual.
  • the first attribute residual reconstruction value of the G component can be used to update the initial cross-component attribute residual prediction value
  • the second cross-component attribute residual prediction value can be the first attribute residual reconstruction value + Obtain the second cross-component attribute residual prediction value for the initial cross-component attribute residual prediction value; or, use twice the first attribute residual reconstruction value as the second cross-component attribute residual prediction value.
  • the first attribute residual reconstruction value of the G component can be used to update the initial cross-component attribute residual prediction value, and the third cross-component attribute residual prediction value can be obtained, which can be the first attribute residual reconstruction value + the initial
  • the cross-component attribute residual predictors yield the third cross-component attribute residual predictors; alternatively, half the reconstructed values of the first attribute residuals are used as the third cross-component attribute residual predictors.
  • the encoder when the encoder obtains the attribute residual value of each color component, it writes the attribute residual value of the color component into the code stream.
  • the attribute residual value of the color component includes: the attribute residual value of the first color component to be predicted, the attribute residual value of the second color component to be predicted, and the third color component to be predicted Attribute residual value for the component.
  • the attribute residual value of the first color component to be predicted, the attribute residual value of the second color component to be predicted, and the attribute residual value of the third color component to be predicted are quantized attribute residual values.
  • the mean value mean of the initial attribute residual reconstruction value of the three color components that is, the standard value (also can be the median value, maximum value, minimum value, etc.), which is not limited by the embodiment of the present application;
  • step d) Set the initial cross-component attribute residual prediction value residualPrevComponent to an initial value of 0, and according to the attribute residual secondary prediction mode determined in step c), if the attribute residual reconstruction value of the G component is used to perform attribute residual secondary Prediction, then first perform the following operations on the G component;
  • the adaptive selection of the prediction mode can be performed based on the degree of difference between the color components, when the second prediction of the color component of the current point is performed according to the prediction mode, the obtained color component The accuracy of attribute reconstruction values is high.
  • the flag bit of the prediction mode can also be directly transmitted to the decoder, which improves the decoding efficiency of the decoder.
  • the embodiment of the present application is not limited;
  • step c) Set the initial cross-component attribute residual prediction value residualPrevComponent to an initial value of 0, and according to the attribute residual secondary prediction mode determined in step c), if the attribute residual reconstruction value of the G component is used to perform attribute residual secondary Prediction, then first perform the following operations on the G component;
  • the adaptive selection of the prediction mode can be performed based on the degree of difference between the color components, when the second prediction of the color component of the current point is performed according to the prediction mode, the obtained color component The accuracy of attribute reconstruction values is high.
  • the embodiment of the present application provides a decoder 1, including:
  • the parsing part 10 is configured to parse the code stream, and determine the attribute residual value of the color component corresponding to the current point;
  • the first acquiring part 11 is configured to acquire the attribute prediction value of the color component corresponding to the current point
  • the first determining part 12 is configured to determine the prediction mode of the color component of the current point; wherein, the prediction mode is determined based on the degree of difference between the color components;
  • the decoding part 13 is configured to decode and reconstruct the color component of the current point based on the prediction mode, the attribute residual value, the attribute prediction value and the initial cross-component attribute residual prediction value, to obtain the color component The property reconstruction value for .
  • the parsing part 10 is further configured to parse out the prediction mode flag when parsing the code stream;
  • the first determining part 12 is further configured to determine the prediction mode of the color component of the current point according to the prediction mode flag bit.
  • the first determining part 12 is further configured to determine the standard value of the color component corresponding to the current point based on the attribute prediction value; determine at least two color components corresponding to the current point A component, at least two degrees of difference from the standard value; based on the at least two degrees of difference, determine the prediction mode of the color component of the current point.
  • the first determination part 12 is further configured to determine the minimum difference degree among the at least two difference degrees; determine the color component corresponding to the minimum difference degree as the predicted model.
  • the first determining part 12 is further configured to determine a ranking result of the at least two degrees of difference; and determine the prediction mode according to the ranking result.
  • the property prediction value includes: a first property prediction value of the first color component, a second property prediction value of the second color component, and a third property prediction value of the third color component;
  • the first determining part 12 is further configured as at least one of the following:
  • the prediction mode includes: a first color component to be predicted; the first color component to be predicted is the first color component among the first color component, the second color component and the third color component The color components of the codec.
  • the decoding part 13 is further configured to determine the first color component of the first color component to be predicted based on the attribute residual value of the first color component to be predicted according to the prediction mode.
  • attribute residual reconstruction value wherein, the first color component to be predicted is the color component characterized by the prediction mode; based on the attribute prediction value, the first attribute residual reconstruction value and the initial span
  • the component attribute residual prediction value decodes and reconstructs the color component of the current point to obtain the attribute reconstruction value of the color component.
  • the decoding part 13 is further configured to dequantize the attribute residual value of the first color component to be predicted according to the prediction mode, and determine the first The first attribute residual reconstruction value of the color component to be predicted.
  • the attribute reconstruction value includes: a first attribute reconstruction value, a second attribute reconstruction value and a third attribute reconstruction value;
  • the decoding part 13 is further configured to add the initial cross-component attribute residual prediction value to the attribute prediction value corresponding to the first color component to be predicted and the first attribute residual reconstruction value to obtain the said first attribute reconstruction value of the first color component to be predicted;
  • the second color component to be predicted and the third color component to be predicted are the first color component, the second color component and the third color component except the first color component to be predicted other color components.
  • the decoding part 13 is further configured to decode and reconstruct the second color component to be predicted based on the second cross-component attribute residual prediction value to obtain the second attribute reconstruction value;
  • the decoding part 13 is further configured to use the initial cross-component attribute residual prediction value, the attribute prediction value corresponding to the first color component to be predicted, and the first After adding the attribute residual reconstruction values to obtain the first attribute reconstruction value of the first to-be-predicted color component, determine the second cross-component attribute residual based on the first attribute residual reconstruction value The predicted value and the third cross-component attribute residual prediction value; based on the second cross-component attribute residual prediction value, the second color component to be predicted is decoded and reconstructed to obtain the second attribute reconstruction value; based on the The third cross-component attribute residual prediction value is described, and the third color component to be predicted is decoded and reconstructed to obtain the third attribute reconstruction value.
  • the second cross-component attribute residual prediction value is the sum of the first cross-component attribute residual reconstruction value and the initial cross-component attribute residual prediction value, or, The multiple or divisor of the reconstruction value for the first attribute residual.
  • the third cross-component attribute residual prediction value is the sum of the second attribute residual reconstruction value and the second cross-component attribute residual prediction value, or , the multiple or divisor of the reconstruction value for the second attribute residual.
  • the third cross-component attribute residual prediction value is the sum of the first cross-component attribute residual reconstruction value and the initial cross-component attribute residual prediction value, or, The multiple or divisor of the reconstruction value for the first attribute residual.
  • the adaptive selection of the prediction mode can be performed based on the degree of difference between the color components, when the second prediction of the color component of the current point is performed according to the prediction mode, the obtained color component The accuracy of attribute reconstruction values is high.
  • the embodiment of this application also provides a decoder, including:
  • the first memory 14 stores a computer program that can run on the first processor 15, and the first processor 15 implements a decoding method corresponding to the decoder when executing the program.
  • the first processor 15 can be implemented by software, hardware, firmware or a combination thereof, and can use circuits, single or multiple application specific integrated circuits (ASIC), single or multiple general integrated circuits, single or multiple a microprocessor, a single or multiple programmable logic devices, or a combination of the aforementioned circuits or devices, or other suitable circuits or devices, so that the first processor 15 can perform the decoding on the decoder side in the aforementioned embodiments corresponding steps of the method.
  • ASIC application specific integrated circuits
  • the embodiment of the present application provides an encoder 2, as shown in Figure 14, including:
  • the second acquiring part 20 is configured to acquire the attribute prediction value of the color component corresponding to the current point in the point cloud;
  • the second determining part 21 is configured to determine the standard value of the color component corresponding to the current point based on the attribute prediction value; determine the difference between the at least two color components corresponding to the current point and the standard value at least two degrees of difference; based on the at least two degrees of difference, determining a prediction mode of the color component;
  • the prediction part 22 is configured to perform secondary prediction on the color component of the current point based on the prediction mode, the attribute prediction value and the initial cross-component attribute residual prediction value to obtain the attribute residual value of the color component.
  • the property prediction value includes: a first property prediction value of the first color component, a second property prediction value of the second color component, and a third property prediction value of the third color component;
  • the second determining part 21 is further configured as at least one of the following:
  • the prediction part 22 is further configured to determine the initial attribute residual reconstruction value corresponding to the current point based on the attribute prediction value and the attribute information of the current point; based on the initial attribute residual The difference reconstruction value determines the standard value of the color component corresponding to the current point.
  • the prediction part 22 is further configured to determine an initial residual value corresponding to the current point based on the predicted attribute value and the attribute information of the current point; Quantization and inverse quantization to obtain the residual reconstruction value of the initial attribute.
  • the initial attribute residual reconstruction value includes: the first initial attribute residual reconstruction value of the first color component, the second initial attribute residual reconstruction value of the second color component, and the third color The residual reconstruction value of the third attribute of the component's initial attribute;
  • the second determining part 21 is further configured as at least one of the following:
  • the encoder 2 further includes: a writing part 23;
  • the prediction part 22 is further configured to generate a prediction mode flag; the prediction mode flag represents the prediction mode;
  • the writing part 23 is further configured to write the prediction mode flag bit into the code stream.
  • the second determining part 21 is further configured to determine the minimum difference degree among the at least two difference degrees; determine the color component corresponding to the minimum difference degree as the predicted model.
  • the second determining part 21 is further configured to determine a ranking result of the at least two degrees of difference; and determine the prediction mode according to the ranking result.
  • the prediction part 22 is further configured to predict the first property information of the first color component to be predicted and the property prediction of the first color component to be predicted according to the prediction mode Value and the initial cross-component attribute residual prediction value are sequentially subtracted and then quantized to realize secondary prediction, and obtain the attribute residual value of the first color component to be predicted;
  • the prediction mode includes: the first color component to be predicted ;
  • the first color component to be predicted is the first coded color component among the first color component, the second color component and the third color component;
  • the first attribute residual reconstruction value of the first color component to be predicted is obtained; based on the first attribute residual reconstruction value, determine the second cross-component attribute residual prediction value; based on the second cross-component attribute residual prediction value, perform secondary prediction on other color components, and obtain the second attribute residual value of the color component to be predicted and the attribute residual value of the third color component to be predicted; the second color component to be predicted and the third color component to be predicted are the first color component, the second color component and the third color component Other color components in excluding the first color component to be predicted.
  • the prediction part 22 is further configured to perform secondary prediction on the second color component to be predicted based on the second cross-component attribute residual prediction value to obtain the second The attribute residual value of the color component to be predicted;
  • the second attribute residual reconstruction value of the second color component to be predicted is obtained;
  • a second prediction is performed on the third to-be-predicted color component to obtain an attribute residual value of the third to-be-predicted color component.
  • the prediction part 22 is further configured to obtain the first color to be predicted after dequantizing the attribute residual value of the first color component to be predicted.
  • the first attribute residual reconstruction value of the component based on the first attribute residual reconstruction value, the second cross-component attribute residual prediction value and the third cross-component attribute residual prediction value are determined; based on the The second cross-component attribute residual prediction value performs secondary prediction on the second to-be-predicted color component to obtain the attribute residual value of the second to-be-predicted color component; based on the third cross-component attribute residual prediction value, perform secondary prediction on the third color component to be predicted, and obtain the attribute residual value of the third color component to be predicted.
  • the second cross-component attribute residual prediction value is the sum of the first cross-component attribute residual reconstruction value and the initial cross-component attribute residual prediction value, or, The multiple or divisor of the reconstruction value for the first attribute residual.
  • the third cross-component attribute residual prediction value is the sum of the second attribute residual reconstruction value and the second cross-component attribute residual prediction value, or , the multiple or divisor of the reconstruction value for the second attribute residual.
  • the third cross-component attribute residual prediction value is the sum of the first cross-component attribute residual reconstruction value and the initial cross-component attribute residual prediction value, or, The multiple or divisor of the reconstruction value for the first attribute residual.
  • the encoder 2 further includes: a writing part 23;
  • the writing part 23 is further configured to write the property residual value of the color component into the code stream.
  • the attribute residual value of the color component includes: the attribute residual value of the first color component to be predicted, the attribute residual value of the second color component to be predicted, and the Describe the attribute residual value of the third color component to be predicted.
  • the adaptive selection of the prediction mode can be performed based on the degree of difference between the color components, when the second prediction of the color component of the current point is performed according to the prediction mode, the obtained color component The accuracy of attribute reconstruction values is high.
  • an encoder including:
  • the second memory 25 stores a computer program that can run on the second processor 24, and the second processor 24 executes the encoding method corresponding to the encoder when executing the program.
  • the embodiment of the present application provides a storage medium on which a computer program is stored.
  • the decoding method corresponding to the claim decoder is realized; or, the computer program is processed by the second processor.
  • the encoder is executed, the encoding method corresponding to the encoder of the claim is realized.
  • Each component in the embodiment of the present application may be integrated into one processing unit, each unit may exist separately physically, or two or more units may be integrated into one unit.
  • the above-mentioned integrated units can be implemented in the form of hardware or in the form of software function modules.
  • the integrated unit is implemented in the form of a software function module and is not sold or used as an independent product, it can be stored in a computer-readable storage medium.
  • the technical solution of this embodiment is essentially or It is said that the part that contributes to the prior art or the whole or part of the technical solution can be embodied in the form of a software product, the computer software product is stored in a storage medium, and includes several instructions to make a computer device (which can It is a personal computer, a server, or a network device, etc.) or a processor (processor) that executes all or part of the steps of the method described in this embodiment.
  • the aforementioned storage medium includes: magnetic random access memory (FRAM, ferromagnetic random access memory), read-only memory (ROM, Read Only Memory), programmable read-only memory (PROM, Programmable Read-Only Memory), erasable Programmable Read-Only Memory (EPROM, Erasable Programmable Read-Only Memory), Electrically Erasable Programmable Read-Only Memory (EEPROM, Electrically Erasable Programmable Read-Only Memory), Flash Memory (Flash Memory), Magnetic Surface Memory, Optical Disk , or compact disc read-only memory (CD-ROM, Compact Disc Read-Only Memory) and other media that can store program codes, the embodiments of the present disclosure are not limited.
  • FRAM magnetic random access memory
  • ROM read-only memory
  • PROM Read Only Memory
  • PROM programmable Read-Only Memory
  • EPROM Erasable Programmable Read-Only Memory
  • EEPROM Electrically Erasable Programmable Read-Only Memory
  • Flash Memory Flash Memory
  • Magnetic Surface Memory Optical
  • the embodiment of the present application provides a decoding method, an encoding method, a decoder, an encoder, and a storage medium.
  • determine the attribute residual value of the color component corresponding to the current point obtain the attribute of the color component corresponding to the current point Prediction value; determine the prediction mode of the color component of the current point; where the prediction mode is determined based on the degree of difference between the color components; based on the prediction mode, attribute residual value, attribute prediction value and initial cross-component attribute residual prediction Value, decode and reconstruct the color component of the current point to obtain the attribute reconstruction value of the color component. Since the adaptive selection of the prediction mode can be performed based on the degree of difference between the color components, when the decoding and reconstruction of the color component of the current point is performed according to the prediction mode, the accuracy of the attribute reconstruction value of the obtained color component higher.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)
  • Compression, Expansion, Code Conversion, And Decoders (AREA)

Abstract

Un mode de réalisation de la demande concerne un procédé de décodage, un procédé de codage, un décodeur, un codeur et un support de stockage. Au moyen de l'analyse d'un flux de données, déterminer une valeur d'erreur résiduelle d'attribut d'une composante de couleur correspondant à un point actuel ; acquérir une valeur de prédiction d'attribut d'une composante de couleur correspondant à un point actuel ; déterminer un modèle de prédiction d'une composante de couleur d'un point actuel ; le modèle de prédiction étant déterminé sur la base d'un degré de différence entre des composantes de couleur ; décoder et reconstruire la composante de couleur du point actuel sur la base du modèle de prédiction, de la valeur d'erreur résiduelle d'attribut, de la valeur de prédiction d'attribut et de la valeur de prédiction d'erreur résiduelle d'attribut de inter-composante initiale, ce qui permet d'obtenir une valeur de reconstruction d'attribut de la composante de couleur.
PCT/CN2021/143372 2021-12-30 2021-12-30 Procédé de décodage, procédé de codage, décodeur, codeur et support de stockage WO2023123284A1 (fr)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/CN2021/143372 WO2023123284A1 (fr) 2021-12-30 2021-12-30 Procédé de décodage, procédé de codage, décodeur, codeur et support de stockage

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2021/143372 WO2023123284A1 (fr) 2021-12-30 2021-12-30 Procédé de décodage, procédé de codage, décodeur, codeur et support de stockage

Publications (1)

Publication Number Publication Date
WO2023123284A1 true WO2023123284A1 (fr) 2023-07-06

Family

ID=86997081

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/143372 WO2023123284A1 (fr) 2021-12-30 2021-12-30 Procédé de décodage, procédé de codage, décodeur, codeur et support de stockage

Country Status (1)

Country Link
WO (1) WO2023123284A1 (fr)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111385555A (zh) * 2018-12-28 2020-07-07 上海天荷电子信息有限公司 原始和/或残差数据用分量间预测的数据压缩方法和装置
US20210385303A1 (en) * 2020-06-09 2021-12-09 Qualcomm Incorporated Attribute residual coding in g-pcc
CN113795870A (zh) * 2019-10-10 2021-12-14 腾讯美国有限责任公司 用于点云属性编解码的信道间预测和变换的技术和装置

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111385555A (zh) * 2018-12-28 2020-07-07 上海天荷电子信息有限公司 原始和/或残差数据用分量间预测的数据压缩方法和装置
CN113795870A (zh) * 2019-10-10 2021-12-14 腾讯美国有限责任公司 用于点云属性编解码的信道间预测和变换的技术和装置
US20210385303A1 (en) * 2020-06-09 2021-12-09 Qualcomm Incorporated Attribute residual coding in g-pcc

Similar Documents

Publication Publication Date Title
US11803986B2 (en) Point cloud data transmission device, point cloud data transmission method, point cloud data reception device, and point cloud data reception method
US11483363B2 (en) Point cloud data transmission device, point cloud data transmission method, point cloud data reception device, and point cloud data reception method
US11711535B2 (en) Video-based point cloud compression model to world signaling information
US11190803B2 (en) Point cloud coding using homography transform
EP4258671A1 (fr) Procédé de prédiction d'attribut de nuage de points, codeur, décodeur et support d'enregistrement
KR20230060534A (ko) 2차원 정규화 평면 투사에 기초한 포인트 클라우드 인코딩 및 디코딩 방법과 장치
WO2023123284A1 (fr) Procédé de décodage, procédé de codage, décodeur, codeur et support de stockage
WO2023015530A1 (fr) Procédés de codage et de décodage de nuage de points, codeur, décodeur et support de stockage lisible par ordinateur
CN116648915A (zh) 点云编解码方法、编码器、解码器以及计算机存储介质
WO2022257145A1 (fr) Procédé et appareil de prédiction d'attributs de nuage de points, et codec
WO2023097694A1 (fr) Procédé de décodage, procédé de codage, décodeur et codeur
WO2023023918A1 (fr) Procédé de décodage, procédé de codage, décodeur et codeur
WO2022257155A1 (fr) Procédé de décodage, procédé de codage, décodeur, codeur, dispositif de codage et dispositif de décodage
WO2022257143A1 (fr) Procédé et appareil de prédiction intra-trame, procédé et appareil de codage, procédé et appareil de décodage, codec, dispositif et support
WO2023240455A1 (fr) Procédé et appareil de codage de nuage de points, dispositif de codage, et support de stockage
WO2023173238A1 (fr) Procédé de codage, procédé de décodage, flux de code, codeur, décodeur et support d'enregistrement
WO2022257150A1 (fr) Procédés et appareil de codage et de décodage de nuage de points, codec de nuage de points et support de stockage
US20240185467A1 (en) Point cloud data transmission device, point cloud data transmission method, point cloud data reception device, and point cloud data reception method
WO2023173237A1 (fr) Procédé de codage, procédé de décodage, train de bits, codeur, décodeur et support d'enregistrement
WO2024077548A1 (fr) Procédé de décodage de nuage de points, procédé de codage de nuage de points, décodeur et codeur
US20230412837A1 (en) Point cloud data transmission method, point cloud data transmission device, point cloud data reception method, and point cloud data reception device
WO2023197337A1 (fr) Procédé et appareil de détermination d'indice, décodeur et codeur
WO2022217472A1 (fr) Procédés de codage et de décodage de nuage de points, codeur, décodeur et support de stockage lisible par ordinateur
WO2024065406A1 (fr) Procédés de codage et de décodage, train de bits, codeur, décodeur et support de stockage
WO2024007144A1 (fr) Procédé de codage, procédé de décodage, flux de code, codeurs, décodeurs et support de stockage

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21969641

Country of ref document: EP

Kind code of ref document: A1