WO2022258055A1 - 点云属性信息编码方法、解码方法、装置及相关设备 - Google Patents
点云属性信息编码方法、解码方法、装置及相关设备 Download PDFInfo
- Publication number
- WO2022258055A1 WO2022258055A1 PCT/CN2022/098193 CN2022098193W WO2022258055A1 WO 2022258055 A1 WO2022258055 A1 WO 2022258055A1 CN 2022098193 W CN2022098193 W CN 2022098193W WO 2022258055 A1 WO2022258055 A1 WO 2022258055A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- points
- attribute
- information
- encoded
- decoded
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 220
- 230000009466 transformation Effects 0.000 claims description 240
- 238000013139 quantization Methods 0.000 claims description 231
- 238000004590 computer program Methods 0.000 claims description 8
- 238000010586 diagram Methods 0.000 description 9
- 230000006870 function Effects 0.000 description 9
- 230000000694 effects Effects 0.000 description 8
- 238000006243 chemical reaction Methods 0.000 description 6
- 238000007726 management method Methods 0.000 description 3
- 230000003190 augmentative effect Effects 0.000 description 2
- 101150116295 CAT2 gene Proteins 0.000 description 1
- 101100326920 Caenorhabditis elegans ctl-1 gene Proteins 0.000 description 1
- 101100342039 Halobacterium salinarum (strain ATCC 29341 / DSM 671 / R1) kdpQ gene Proteins 0.000 description 1
- 241000699670 Mus sp. Species 0.000 description 1
- 101100126846 Neurospora crassa (strain ATCC 24698 / 74-OR23-1A / CBS 708.71 / DSM 1257 / FGSC 987) katG gene Proteins 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 238000007599 discharging Methods 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 239000011521 glass Substances 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 239000011159 matrix material Substances 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000007781 pre-processing Methods 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/597—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/124—Quantisation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T9/00—Image coding
- G06T9/001—Model-based coding, e.g. wire frame
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/12—Selection from among a plurality of transforms or standards, e.g. selection between discrete cosine transform [DCT] and sub-band transform or selection between H.263 and H.264
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/13—Adaptive entropy coding, e.g. adaptive variable length coding [AVLC] or context adaptive binary arithmetic coding [CABAC]
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/136—Incoming video signal characteristics or properties
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/18—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a set of transform coefficients
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/184—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being bits, e.g. of the compressed video stream
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/60—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
- H04N19/625—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding using discrete cosine transform [DCT]
Definitions
- the present application belongs to the technical field of point cloud processing, and in particular relates to a point cloud attribute information encoding method, decoding method, device and related equipment.
- attribute information coding is divided into attribute prediction coding and attribute transformation coding.
- attribute transformation coding directly transforms the original attribute information. For areas with large local transformation ranges, there are still many redundancies in the obtained transformation coefficients. information, resulting in low coding efficiency.
- Embodiments of the present application provide a method for encoding point cloud attribute information, a decoding method, a device, and related equipment, which can solve the problem of low encoding efficiency of existing point cloud attribute information.
- a method for encoding point cloud attribute information including:
- DCT discrete cosine transform
- Quantizing the transform coefficients of the K points to be encoded performing entropy encoding based on the quantized transform coefficients, and generating a binary code stream
- the first information includes the K points to be encoded
- the second information includes attribute prediction information of the K points to be encoded
- the first information includes the K points to be encoded
- the second information includes attribute reconstruction information of the N encoded points
- K is a positive integer
- N is an integer greater than 1.
- a method for decoding point cloud attribute information including:
- the third information includes the K points to be decoded, and the fourth information includes attribute prediction information of the K points to be decoded; or, the third information includes the K points to be decoded
- the fourth information includes attribute reconstruction information of the N decoded points, K is a positive integer, and N is an integer greater than 1.
- a device for encoding point cloud attribute information including:
- a first acquiring module configured to acquire first information
- a first determination module configured to determine whether to perform discrete cosine transform (DCT) transformation on the K points to be encoded based on the second information associated with the first information;
- DCT discrete cosine transform
- the first transformation module is configured to perform DCT transformation on the K points to be coded to obtain transformation coefficients of the K points to be coded when it is determined to perform DCT transformation on the K points to be coded;
- An encoding module configured to quantize the transformation coefficients of the K points to be encoded, perform entropy encoding based on the quantized transformation coefficients, and generate a binary code stream;
- the first information includes the K points to be encoded
- the second information includes attribute prediction information of the K points to be encoded
- the first information includes the K points to be encoded
- the second information includes attribute reconstruction information of the N encoded points
- K is a positive integer
- N is an integer greater than 1.
- a device for decoding point cloud attribute information including:
- the second obtaining module is used to obtain the third information
- the second determination module is configured to determine whether to perform inverse DCT transformation on the K points to be decoded based on fourth information associated with the third information;
- the second transformation module is configured to perform inverse DCT transformation on the K points to be decoded to obtain attribute residuals of the K points to be decoded when it is determined to perform inverse DCT transformation on the K points to be decoded information;
- a decoding module configured to obtain attribute reconstruction information of the K points to be decoded based on the attribute residual information and attribute prediction information of the K points to be decoded, so as to decode undecoded points in the point cloud to be decoded;
- the third information includes the K points to be decoded, and the fourth information includes attribute prediction information of the K points to be decoded; or, the third information includes the K points to be decoded
- the fourth information includes attribute reconstruction information of the N decoded points, K is a positive integer, and N is an integer greater than 1.
- a terminal includes a processor, a memory, and a program or instruction stored in the memory and operable on the processor.
- the program or instruction When the program or instruction is executed by the processor Realize the steps of the point cloud attribute information encoding method as described in the first aspect, or realize the steps of the point cloud attribute information decoding method as described in the second aspect.
- a terminal including a processor and a communication interface, wherein the processor is used for:
- DCT discrete cosine transform
- Quantizing the transform coefficients of the K points to be encoded performing entropy encoding based on the quantized transform coefficients, and generating a binary code stream
- the first information includes the K points to be encoded, and the second information includes attribute prediction information of the K points to be encoded; or, the first information includes the K points to be encoded
- the second information includes attribute reconstruction information of the N encoded points, K is a positive integer, and N is an integer greater than 1;
- the processor is used to:
- the third information includes the K points to be decoded, and the fourth information includes attribute prediction information of the K points to be decoded; or, the third information includes the K points to be decoded
- the fourth information includes attribute reconstruction information of the N decoded points, K is a positive integer, and N is an integer greater than 1.
- a readable storage medium is provided, and a program or an instruction is stored on the readable storage medium, and when the program or instruction is executed by a processor, the method for encoding point cloud attribute information as described in the first aspect is implemented. steps, or realize the steps of the method for decoding point cloud attribute information as described in the second aspect.
- a chip in an eighth aspect, includes a processor and a communication interface, the communication interface is coupled to the processor, and the processor is used to run programs or instructions to achieve the points described in the first aspect The steps of the method for encoding cloud attribute information, or the steps for realizing the method for decoding point cloud attribute information as described in the second aspect.
- a ninth aspect provides a computer program/program product, the computer program/program product is stored in a non-volatile storage medium, and the program/program product is executed by at least one processor to implement the first aspect
- a communication device configured to perform the steps of the point cloud attribute information encoding method as described in the first aspect, or configured to perform the steps of the point cloud attribute information decoding method as described in the second aspect step.
- the code point to be coded in the process of encoding the point cloud to be encoded, it is necessary to determine whether to perform DCT transformation on the point cloud to be encoded according to the attribute prediction information of the point to be encoded or the attribute reconstruction information of the encoded point.
- the code point to be coded is subjected to DCT transformation, and then the scattered distribution of attribute information in the space domain can be converted into a relatively concentrated distribution in the transformation domain, so that the signal energy is concentrated in a few coefficients, which is more convenient Quantization and coding, so as to remove attribute redundancy and achieve the purpose of improving attribute coding efficiency and reconstruction performance.
- Figure 1 is a frame diagram of the AVS codec
- Fig. 2 is the conversion flowchart of coding end
- Fig. 3 is a flow chart of a method for encoding point cloud attribute information provided by an embodiment of the present application
- Fig. 4 is a flow chart of another point cloud attribute information encoding method provided by the embodiment of the present application.
- Fig. 5 is a flow chart of a method for decoding point cloud attribute information provided by an embodiment of the present application
- FIG. 6 is a structural diagram of a device for encoding point cloud attribute information provided by an embodiment of the present application.
- FIG. 7 is a structural diagram of a device for decoding point cloud attribute information provided by an embodiment of the present application.
- FIG. 8 is a structural diagram of a communication device provided by an embodiment of the present application.
- FIG. 9 is a structural diagram of a terminal provided by an embodiment of the present application.
- first, second and the like in the specification and claims of the present application are used to distinguish similar objects, and are not used to describe a specific sequence or sequence. It is to be understood that the terms so used are interchangeable under appropriate circumstances such that the embodiments of the application are capable of operation in sequences other than those illustrated or described herein and that "first" and “second” distinguish objects. It is usually one category, and the number of objects is not limited. For example, there may be one or more first objects.
- “and/or” in the description and claims means at least one of the connected objects, and the character “/” generally means that the related objects are an "or” relationship.
- Both the encoder corresponding to the encoding method and the decoder corresponding to the decoding method in the embodiment of the present application may be a terminal, and the terminal may also be called a terminal device or a user equipment (User Equipment, UE), and the terminal may be a mobile phone, a tablet computer ( Tablet Personal Computer), laptop computer (Laptop Computer) or notebook computer, personal digital assistant (Personal Digital Assistant, PDA), palmtop computer, netbook, ultra-mobile personal computer (ultra-mobile personal computer, UMPC), mobile Internet device (Mobile Internet Device, MID), augmented reality (augmented reality, AR) / virtual reality (virtual reality, VR) equipment, robot, wearable device (Wearable Device) or vehicle-mounted equipment (VUE), pedestrian terminal (PUE ) and other terminal-side devices, and wearable devices include: smart watches, bracelets, earphones, glasses, etc. It should be noted that, the embodiment of the present application does not limit the specific type of the terminal.
- the geometric information of the point cloud and the attribute information corresponding to each point are encoded separately.
- FIG. 1 is a frame diagram of an AVS codec.
- coordinate transformation is first performed on the geometric information so that all point clouds are contained in a bounding box. Then quantize. This step of quantization mainly plays the role of scaling. Due to the rounding of quantization, the geometric information of some points is the same. It is determined whether to remove duplicate points according to the parameters. The process of quantization and removal of duplicate points belongs to preprocessing. process. Next, divide the bounding box (octree/quadtree/binary tree) according to the order of breadth-first traversal, and encode the placeholder code of each node.
- bounding box octree/quadtree/binary tree
- the bounding box is divided into sub-cubes in turn, and the sub-cubes that are not empty (including points in the point cloud) are continued to be divided until the leaf nodes obtained by the division are 1x1x1 units. Stop dividing the cube, and then encode the points contained in the leaf nodes, and finally complete the encoding of the geometric octree to generate a binary code stream.
- the decoding end obtains the placeholder code of each node through continuous analysis according to the order of breadth-first traversal, and divides the nodes in turn until the unit cube of 1x1x1 is obtained. Get the number of points contained in each leaf node, and finally restore the geometrically reconstructed point cloud information
- Attribute coding is mainly carried out for color, reflectance information, etc. First judge whether to perform color space conversion. If color space conversion is performed, the color information is converted from red, green and blue (RGB) color space to YUV (Y is brightness component, UV is chroma component) color space. Then, the reconstructed point cloud is recolored with the original point cloud, so that the unencoded attribute information corresponds to the reconstructed geometric information.
- RGB red, green and blue
- YUV brightness component
- UV chroma component
- attribute prediction process is as follows: first the point cloud is re-ranked and then differentially predicted. There are two reordering methods: Morton reordering and Hilbert reordering.
- the attribute transformation process is as follows: firstly, wavelet transform is performed on the point cloud attributes, and the transformation coefficients are quantized; secondly, the attribute reconstruction value is obtained through inverse quantization and inverse wavelet transformation; then, the difference between the original attribute and the attribute reconstruction value is calculated to obtain the attribute residual and its Quantization; finally, entropy coding is performed on the quantized transform coefficients and attribute residuals to generate a binary code stream.
- the decoding of attribute information is the reverse process of encoding, which will not be described in detail here.
- DCT discrete cosine Transform
- DST discrete sine transform
- the transformation process at the encoding end is as follows: First, the residual signal is obtained by subtracting the predicted signal from the original signal, and then the residual signal is transformed by a DCT to obtain a transformation coefficient, and then the low-frequency component of the transformation coefficient block is processed twice Transform to obtain secondary transformation coefficients with more concentrated energy distribution, and then perform quantization and entropy coding to obtain code streams.
- the quantized coefficients are dequantized, inversely transformed twice, and inversely transformed once to obtain the recovered residual, which is added to the predicted signal to obtain the reconstructed signal, and then loop filtering is performed to reduce distortion.
- the secondary transformation may not necessarily be carried out, as shown in the dotted line in FIG. 2 .
- attribute transformation coding directly transforms the original attribute information. For areas with large local transformation ranges, there are still many redundant information in the obtained transformation coefficients, resulting in low coding efficiency.
- FIG. 3 is a flow chart of a method for encoding point cloud attribute information provided by an embodiment of the present application.
- the method may be applied to terminals such as mobile phones, tablet computers, and computers.
- the method includes the following steps:
- Step 301 Acquire first information.
- the first information may include points to be encoded, or may include encoded points.
- the first information includes the K points to be encoded, and the step 301 includes:
- Sorting the point cloud to be encoded and acquiring K points to be encoded in the sorted point cloud to be encoded.
- the point cloud to be coded may be sorted. For example, in the case where the attribute information is mainly for color, first determine whether the point cloud to be encoded needs to be converted into a color space. If the color space conversion is performed, the color information of the point cloud to be encoded can be converted from RGB color space to YUV color Space, and then use the original point cloud to recolor the point cloud to be encoded, so that the attribute information of the point cloud to be encoded corresponds to the pre-reconstructed geometric information.
- the sorting of the point cloud to be coded can be realized based on Hilbert code or Morton code.
- the sorting of the point cloud to be encoded, and obtaining the K points to be encoded in the sorted point cloud to be encoded include:
- the K points to be encoded may be selected based on a certain order. For example, assuming that the point cloud to be encoded includes 9 points to be encoded, after calculating the Hilbert codes corresponding to the 9 points to be encoded, and sorting the 9 points to be encoded according to Hilbert, it can be Every 3 points to be coded as a group, first select the top 3 points as the K points to be coded, perform subsequent attribute prediction and other processes, and then select the three points after the first 3 points as a group The group performs attribute prediction and other processes, and then selects the last 3 points for subsequent attribute prediction and other processes.
- the selected K points to be encoded it is necessary to judge whether the points to be encoded in the K points to be encoded are repeated points with the encoded points; It is to directly calculate the attribute residual information between the point to be encoded and the point to be encoded, and the point to be encoded is not used as a point among the K points to be encoded; if the point to be encoded and the point to be encoded are not repeated points , then the attribute information prediction of the K points to be encoded can be performed on the encoded points to obtain the attribute prediction information of the K points to be encoded, and the specific attribute information prediction method will be described in subsequent embodiments.
- Step 302. Determine whether to perform DCT transformation on the K points to be encoded based on the second information associated with the first information.
- the first information includes the K points to be encoded
- the second information includes attribute prediction information of the K points to be encoded
- the first information includes the K points to be encoded
- the second information includes attribute reconstruction information of the N encoded points
- K is a positive integer
- N is a positive integer greater than 1.
- the first information is K points to be encoded
- the second information is the attribute prediction information of the K points to be encoded
- the first information is the first N encoded points of the K points to be encoded
- the second information is the attribute reconstruction information of the N encoded points
- step 302 it also includes:
- the target point to be encoded is any one of the K points to be encoded
- obtaining the S neighbors closest to the target Manhattan to be coded including:
- target points to be encoded are selected based on preset principles, for example, the target points to be encoded may be sequentially selected according to the order of the K points to be encoded. Assuming that the K points to be encoded are determined after the point cloud to be encoded is sorted according to the Hilbert code, the target points to be encoded may be sequentially determined based on the size of the Hilbert code corresponding to the K points to be encoded.
- the target to-be-encoded point For the target to-be-encoded point, first, according to the Hilbert 1 order, find the first sequence M points of the target to-be-encoded point, and then according to the Hilbert 2-order, respectively search for the first sequence N1 points and the last sequence of the target to-be-encoded point Sequence N2 points, among the M+N1+N2 points, select the S neighbor points closest to the target point to be encoded in Manhattan, and determine the initial attribute prediction information of the target point to be encoded based on the S neighbor points, and then according to The first weight corresponding to the target point to be encoded and the initial attribute prediction information determine the attribute prediction information of the target point to be encoded.
- M, N1 and N2 are all positive integers.
- the given preset search range is determined according to the correlation between the initial point number of the point cloud sequence and the volume of the bounding box of the input point cloud.
- the determining the initial attribute prediction information of the target point to be encoded based on the S neighbor points includes:
- the Manhattan distance between each neighbor point and the target point to be encoded is different, and the second weight is the reciprocal of the Manhattan distance between the current neighbor point and the target point to be encoded , the second weight corresponding to each neighbor point is also different.
- the initial attribute prediction information of the target point to be encoded may be the sum of the product of the attribute information of each of the S neighbor points and its corresponding second weight, and then the initial attribute prediction information of the target point to be encoded is calculated.
- the initial attribute prediction information of each of the K points to be encoded can be calculated, and according to the first weight corresponding to each point to be encoded And the initial attribute prediction information, and the attribute prediction information corresponding to each point to be encoded can be calculated.
- the initial attribute prediction information of the target point to be encoded p n is Pre n
- the initial attribute prediction information of K points to be encoded (p n ,p n+1 ,...,p n+K-1 ) are respectively denoted as Pre n ,Pre n+1 ,...,Pre n+K-1
- the attribute prediction information of the K points to be encoded can be expressed for:
- AttrPre i is the attribute prediction information of K points to be encoded
- Pre j is the initial attribute prediction information of the jth point among the K points.
- each point in the current K points to be encoded uses The initial attribute prediction information of the second point in the K points is used as their attribute prediction information.
- the sum of the first weights corresponding to the K nodes to be coded is 1.
- the sum of the 4 first weights respectively corresponding to the 4 points to be encoded is 1.
- Sequence N2 points within the target range determined based on M, N1 and N2, obtain the S neighbor points closest to the target point to be coded in Manhattan.
- the target point to be coded first, according to the Morton 1 order, search for the preorder M points of the target to be coded point, and then according to the Morton 2 order, respectively search for the preorder N1 points and the postorder N2 points, among these M+N1+N2 points, select the S neighbor points closest to the target point to be encoded in Manhattan, and determine the initial attribute prediction information of the target point to be encoded based on these S neighbor points, and then according to the target
- the first weight corresponding to the point to be encoded and the initial attribute prediction information determine the attribute prediction information of the target point to be encoded.
- M, N1 and N2 are all positive integers.
- the given preset search range is determined according to the correlation between the initial point number of the point cloud sequence and the volume of the bounding box of the input point cloud.
- attribute weighted prediction is performed on the S neighbors, and the weight is the reciprocal of the distance between the current neighbor point and the target point to be encoded in Manhattan.
- the initial attribute prediction information of the target to-be-encoded point then calculate the attribute prediction information of the target to-be-encoded point according to the first weight corresponding to the target to-be-encoded point and the initial attribute prediction information, and then calculate the attribute prediction of K to-be-encoded points accordingly information.
- the method of obtaining the attribute prediction information can refer to the above-mentioned specific description of obtaining S neighbor points based on the double Hilbert order and performing attribute prediction. Neighbor points, the specific way to obtain attribute prediction information will not be repeated here.
- the search range of the Hilbert 1 order or the Morton 1 order is the first preset range
- the search range of the Hilbert 2 order or the Morton 2 order is the second preset range
- the Hilbert 2 order or the Morton 2 order search range is the third preset range or the second preset range. It should be noted that the search ranges of the first preset range, the second preset range and the third preset range may be set artificially according to encoding requirements.
- the binary code stream includes a first attribute parameter and a second attribute parameter, the first attribute parameter is used to characterize the first preset range, and the second attribute parameter is used to characterize the second preset range; in the case where the post-order search range of the Hilbert 2 sequence or the Morton 2 sequence is a third preset range, the binary code stream also includes a third attribute parameter, and the third attribute parameter is used To characterize the third preset range.
- the above-mentioned first preset range, second preset range and third preset range can be written into the code stream as attribute information parameter sets.
- the pre-order search range of the Hilbert 2-order or Morton 2-order is the same as the post-order search range of the Hilbert 2-order or Morton 2-order, that is, both are the second preset range.
- the code stream it is only necessary to write the first attribute parameter representing the search range of Hilbert 1 order or Morton 1 order, and the preorder and Hill characterizing Hilbert 2 order or Morton 2 order
- step 302 includes:
- the target point to be encoded is any one of the K points to be encoded;
- the K points to be encoded can be arranged in a certain order. For example, when the point cloud to be encoded is sorted according to the Morton code, the K points to be encoded can be sorted based on the size of the corresponding Morton code , the first point among the K points to be coded may refer to the smallest point of the Morton code, or the largest point.
- each point to be encoded selects the L neighbor points with the closest Manhattan distance from the R neighbor points, and each point to be encoded determines the corresponding initial attribute prediction information based on the corresponding L neighbor points, and further based on each The first weight corresponding to the point to be encoded and the initial attribute prediction information determine the attribute prediction information corresponding to each point to be encoded. In this way, attribute prediction information of the K points to be encoded can also be calculated and obtained.
- the sum of the first weights corresponding to the K nodes to be coded is 1.
- the determination of the initial attribute prediction information of the target point to be encoded based on the L neighbor points includes:
- the Manhattan distance between each neighbor point and the target point to be encoded is different, and the second weight is the reciprocal of the Manhattan distance between the current neighbor point and the target point to be encoded , the second weight corresponding to each neighbor point is also different.
- the initial attribute prediction information of the target point to be encoded may be the sum of the product of the attribute information of each neighbor point in the L neighbor points and its corresponding second weight, and then the initial attribute prediction information of the target point to be encoded is obtained by calculating information.
- the step 302 may include:
- the attribute prediction information of the K points to be encoded is recorded as AttrPre i , and the value range of i is 1 to K; obtain the maximum attribute prediction value and the minimum attribute prediction in the attribute prediction information corresponding to the K points to be encoded Values are recorded as AttrPre max and AttrPre min respectively, and the first threshold is recorded as Threshold1, if
- ⁇ Threshold1 then perform DCT transformation on the K points to be encoded.
- the first information may also be the first N encoded points including the K points to be encoded, and the second information includes attribute reconstruction information of the N encoded points; in this case , the step 302 includes:
- N is an integer greater than 1, that is, at least the first two encoded points of the K points to be encoded need to be obtained. Understandably, the attribute reconstruction information of the coded points can be obtained.
- the attribute reconstruction information refers to the attribute information that will be assigned to the point cloud sequence, but the attribute reconstruction information is different from the original attribute information in the original point cloud.
- the residual information of the above attributes is obtained.
- AttRec i the attribute reconstruction information of N points to be encoded is recorded as AttRec i , and the value of i ranges from 1 to N; the maximum attribute reconstruction value and the minimum attribute reconstruction information in the attribute reconstruction information corresponding to the N points to be encoded are obtained.
- Attribute reconstruction values are recorded as AttrRec max and AttrRec min respectively, and the third threshold is recorded as Threshpld3. If
- Step 303 In a case where it is determined to perform DCT transformation on the K points to be encoded, perform DCT transformation on the K points to be encoded to obtain transformation coefficients of the K points to be encoded.
- DCT transformation may be performed on attribute residual information of the K points to be encoded. Furthermore, before performing DCT transformation on the K points to be encoded, the method also includes:
- attribute residual information of the K points to be encoded is acquired.
- the attribute residual information of each of the K points to be encoded may be the difference between the original attribute information of the point to be encoded and the attribute prediction information.
- performing DCT transformation on the K points to be encoded to obtain transformation coefficients of the K points to be encoded includes:
- the attribute residual information of the K points to be encoded is denoted as AttrRes
- the DCT transformation is performed on the attribute residual information of the K points, where K represents the order of the DCT transformation.
- T is the K-order transformation matrix
- the transformation coefficient can be obtained from the following transformation formula:
- Step 304 Quantize the transform coefficients of the K points to be encoded, and perform entropy encoding based on the quantized transform coefficients to generate a binary code stream.
- the transformation coefficients of the K points to be encoded are obtained, the transformation coefficients are quantized to obtain quantized transformation coefficients, and entropy encoding is performed on the quantized transformation coefficients to generate a binary code stream, and then In order to complete the encoding of the point cloud attribute information to be encoded.
- the DCT transformation is performed on the points to be coded, the DCT transformation is performed on the points to be coded, and then the scattered distribution of attribute information in the spatial domain can be converted into a relatively concentrated distribution in the transformation domain, so that the signal energy is concentrated in a few coefficients, which is more convenient for quantization And encoding, so as to remove attribute redundancy, and achieve the purpose of improving attribute encoding efficiency and reconstruction performance.
- the transform coefficients include low-frequency coefficients and high-frequency coefficients
- quantizing the transform coefficients of the K points to be encoded includes:
- the method further includes:
- Inverse quantization is performed on the quantized-based transform coefficients, and inverse transforms are performed on the inverse transform coefficients obtained after inverse quantization, so as to obtain attribute reconstruction information of the K points to be encoded.
- inverse quantization and inverse DCT transformation may be further performed based on the quantized transformation coefficients to obtain the K points to be encoded Property reconstruction information.
- the encoded values obtained after the entropy encoding are dequantized, and the dequantized
- the inverse transformation coefficient obtained after performing inverse transformation, to obtain the attribute reconstruction information of the K points to be coded including:
- entropy encoding is performed on the quantized high-frequency coefficients and low-frequency coefficients respectively, and inverse quantization is performed on the coded values after entropy encoding to obtain inverse quantized inverse high-frequency coefficients and inverse low-frequency coefficients, and the inverse high-frequency coefficients and inverse
- the inverse low-frequency coefficients are subjected to inverse DCT transformation to obtain inverse attribute residual information of the K points to be encoded.
- the attribute residual information of K points to be encoded needs to be transformed by DCT, and then quantized, dequantized and inversely transformed. During this process, the attribute residual information may be lost , thus the obtained inverse attribute residual information may be inconsistent with the attribute residual information of the K to-be-encoded points before DCT transformation.
- the attribute reconstruction information of the K points to be encoded can also be obtained.
- the attribute reconstruction information may be the sum of attribute prediction information and inverse attribute residual information.
- quantizing the high-frequency coefficients and low-frequency coefficients corresponding to the K points to be encoded includes:
- the quantization step size may have a certain offset, and it is necessary to complete the scaling operation of the integer DCT based on the offset of the quantization step size, that is, the quantization of the quantized transform coefficients
- the step size is the quantization step size after the scaling operation is completed, so as to ensure the accuracy of the quantization result and avoid the quantization error caused by the quantization step offset.
- the obtaining the quantization step size of the high-frequency coefficient corresponding to the high-frequency coefficient, and obtaining the quantization step size of the low-frequency coefficient corresponding to the low-frequency coefficient include:
- the quantization step size of the high-frequency coefficient corresponding to the high-frequency coefficient is obtained, and the quantization step size of the low-frequency coefficient corresponding to the low-frequency coefficient is obtained.
- the quantization step size of the high-frequency transform coefficient is the original quantization step size, the preset quantization step size offset, and the high-frequency coefficient quantization
- the sum of step size offsets, the quantization step size of the low-frequency transform coefficient is the sum of the original quantization step size, the preset quantization step size offset and the low-frequency coefficient quantization step size offset;
- the quantization step size of the high-frequency coefficients is the original quantization step size, the preset quantization step size offset, and the low-frequency coefficient quantization step size offset
- the sum of shifts, the quantization step size of the low-frequency coefficients is equal to the quantization step size of the high-frequency coefficients.
- Q fin-AC Q ori +Offset coeff +Offset AC
- Q fin-DC Q ori +Offset coeff +Offset DC ; among them, Q fin-AC represents the quantization step size of high-frequency coefficients, Q ori represents the original quantization step size, Offset coeff represents the preset quantization step size offset, Offset AC represents the high-frequency coefficient quantization step size offset, Q fin -DC represents the quantization step size of the low-frequency coefficient, and Offset DC represents the quantization step size offset of the low-frequency coefficient.
- the quantization step size of is also the quantization step size of high-frequency coefficients
- Q ori represents the original quantization step size
- Offset coeff represents the preset quantization step size offset
- Offset DC represents the low-frequency coefficient quantization step size offset.
- the quantized value of the high-frequency coefficient is 0, and when the low-frequency coefficient is smaller than a second preset threshold, the low-frequency coefficient is quantized The subsequent value is 0.
- the first preset threshold and the second preset threshold may be the same or different.
- an encoding manner in which DCT transformation is not performed on the K points to be encoded may also be included.
- the method also includes:
- Entropy encoding is performed on the quantized attribute residual information of the K points to be encoded to generate a binary code stream.
- the attribute residual information of the K points to be encoded may be directly quantized, and then entropy encoding is performed based on the quantized attribute residual information to generate a binary code stream, thereby completing the encoding of the point cloud to be encoded.
- the method may further include:
- the method before the acquiring the first information, the method further includes:
- Acquire identification information where the identification information is used to indicate whether to execute the method
- the identification information may be identification information obtained based on user input, for example, it may be a parameter input by the user, or it may also be a pre-stored parameter at the encoding end, and the pre-stored parameter is obtained through user operation; the parameter is used for Indicates whether the encoding end implements the above method for encoding point cloud attribute information.
- the identification information can be characterized by 0 and 1, if the identification information is 0, it means that the above-mentioned point cloud attribute information encoding method does not need to be executed, and the encoding end does not execute the above-mentioned point cloud attribute information encoding method; if the If the identification information is 1, the encoding end executes the above-mentioned point cloud attribute information encoding method.
- the identification information may also be in other representation forms. For example, if the identification information is "true”, the encoding end executes the above method for encoding point cloud attribute information; if the identification information is "false", the encoding end does not Execute the above method for encoding point cloud attribute information.
- the identification information may also be in other representation forms, which are not listed too many in this embodiment of the present application.
- said method also includes:
- identification information it can be known based on the identification information whether the encoding end adopts the above-mentioned point cloud attribute information encoding method for encoding, and then the decoding end can also adopt the corresponding decoding method based on the identification information to ensure that the decoding end can Smooth decoding to ensure decoding efficiency.
- the attribute reconstruction information of K points to be encoded is obtained by dequantizing the encoded value after entropy encoding, or for the K points to be encoded that have undergone DCT transformation, the encoded value after entropy encoding Inverse quantization and DCT inverse transformation are performed sequentially to obtain the attribute reconstruction information of K points to be encoded. Based on the obtained attribute reconstruction information, it is also possible to judge whether the unencoded points in the unencoded point cloud need to be transformed by DCT, and then use the The scattered distribution of attribute information in the spatial domain is transformed into the centralized distribution in the transform domain, so as to achieve the purpose of removing spatial redundancy, improving the efficiency of attribute coding and reconstruction performance.
- FIG. 4 is a flow chart of another point cloud attribute encoding method provided by an embodiment of the present application.
- the process of this method is as follows: firstly, the points to be coded are sorted, either based on Morton codes or Hilbert codes; Whether the coded point is a repeated point, that is, to judge whether it is repeated with the coded point, if so, obtain the attribute prediction information of the repeated point, and quantify it based on the attribute prediction information; if the point to be coded is not a repeated point, perform Attribute information prediction, obtaining attribute prediction information, judging whether DCT transformation is required based on the attribute prediction information of the point to be encoded, and if so, performing K-order DCT transformation on the point to be encoded to obtain transformation coefficients, and quantize the transformation coefficients; if not Perform DCT transformation, then directly quantize the attribute residual information of the point to be encoded; reconstruct the attribute information based on the quantized related information, obtain the attribute reconstruction information, and judge
- FIG. 5 is a flow chart of a method for decoding point cloud attribute information provided by an embodiment of the present application.
- the method may be applied to terminals such as mobile phones, tablet computers, and computers. As shown in Figure 5, the method includes the following steps:
- Step 501 acquiring third information
- Step 502 based on the fourth information associated with the third information, determine whether to perform inverse DCT transformation on the K points to be decoded;
- Step 503. When it is determined to perform inverse DCT transformation on the K points to be decoded, perform inverse DCT transformation on the K points to be decoded to obtain attribute residual information of the K points to be decoded;
- Step 504 Obtain attribute reconstruction information of the K points to be decoded based on the attribute residual information and attribute prediction information of the K points to be decoded, so as to decode undecoded points in the point cloud to be decoded;
- the third information includes the K points to be decoded, and the fourth information includes attribute prediction information of the K points to be decoded; or, the third information includes the K points to be decoded
- the fourth information includes attribute reconstruction information of the N decoded points, K is a positive integer, and N is an integer greater than 1.
- the third information includes the K points to be decoded
- the fourth information includes attribute prediction information of the K points to be decoded
- the step 502 includes:
- the third information includes the first N decoded points of the K points to be decoded, and the fourth information includes attribute reconstruction information of the N decoded points; in this case, the step 502 includes:
- the third information includes K points to be decoded
- the step 501 may include:
- the point cloud to be decoded is sorted, and K points to be decoded in the sorted point cloud to be decoded are obtained.
- the sorting the point cloud to be decoded and obtaining the K points to be decoded in the sorted point cloud to be decoded includes:
- the method further includes:
- obtaining the S neighbor points closest to the target Manhattan to be decoded including:
- M, N1 and N2 are all positive integers.
- the search range of the Hilbert 1 order or the Morton 1 order is the first preset range
- the search range of the Hilbert 2 order or the Morton 2 order is the second preset Range
- the post-order search range of the Hilbert 2 order or Morton 2 order is the third preset range
- the binary code stream includes a first attribute parameter and a second attribute parameter, the first attribute parameter is used to characterize the first preset range, and the second attribute parameter is used to characterize the second preset range;
- the binary code stream further includes a third attribute parameter, which is used to represent The third preset range.
- the given preset search range is determined according to the correlation between the initial point number of the point cloud sequence and the volume of the bounding box of the input point cloud.
- the determining the initial attribute prediction information of the target point to be decoded based on the S neighbor points includes:
- the second weight is the difference between the target point to be decoded and the neighbor point The reciprocal of the Manhattan distance between .
- the third information includes the K points to be decoded, and before step 502, the method further includes:
- Obtaining the target point to be decoded among the K points to be decoded is L neighbor points with the closest Manhattan distance among the R neighbor points, and the target point to be decoded is any one of the K points to be decoded;
- T, R and L are all positive integers.
- the determining the initial attribute prediction information of the target point to be decoded based on the L neighbor points includes:
- the second weight is the difference between the target point to be decoded and the neighbor point The reciprocal of the Manhattan distance between .
- the sum of the first weights corresponding to the K nodes to be decoded is 1.
- the method before performing inverse DCT transformation on the K points to be decoded, the method further includes:
- the inverse DCT transformation of the K points to be decoded includes:
- the transform coefficients include high-frequency coefficients and low-frequency coefficients
- the dequantization of the transform coefficients to obtain dequantized transform coefficients includes:
- the obtaining the quantization step size of the high-frequency coefficient corresponding to the high-frequency coefficient, and obtaining the quantization step size of the low-frequency coefficient corresponding to the low-frequency coefficient include:
- the quantization step size of the high-frequency coefficient corresponding to the high-frequency coefficient is obtained, and the quantization step size of the low-frequency coefficient corresponding to the low-frequency coefficient is obtained.
- the quantization step of the high frequency coefficient corresponding to the high frequency coefficient is obtained, and the quantization step of the low frequency coefficient corresponding to the low frequency coefficient is obtained long, including:
- the quantization step size of the high-frequency transform coefficient is the original quantization step size, the preset quantization step size offset, and the high-frequency coefficient quantization step size
- the sum of offsets, the quantization step size of the low-frequency transform coefficient is the sum of the original quantization step size, the preset quantization step size offset and the low-frequency coefficient quantization step size offset;
- the quantization step size of the high-frequency coefficient is the original quantization step size, the preset quantization step size offset, and the low-frequency coefficient quantization step size offset
- the sum of shifts, the quantization step size of the low-frequency coefficients is equal to the quantization step size of the high-frequency coefficients.
- the quantized value of the high-frequency coefficient is 0, and when the low-frequency coefficient is smaller than a second preset threshold, the low-frequency The quantized value of the coefficient is 0.
- the method also includes:
- the method before the acquiring the third information, the method further includes:
- identification information from the binary code stream, where the identification information is used to indicate whether to execute the method
- decoding method provided in this embodiment corresponds to the encoding method in the embodiment described in FIG. 3 above, and the related concepts and specific implementation methods involved in this embodiment can refer to the encoding described in FIG. The description in the method will not be repeated in this embodiment.
- the decoding method in the process of decoding the point cloud to be decoded, it is necessary to determine whether to perform DCT transformation on the point to be decoded according to the attribute prediction information of the point to be decoded or the attribute reconstruction information of the decoded point.
- DCT transformation is performed on the point to be decoded, and then the scattered distribution of attribute information in the spatial domain can be converted into a relatively concentrated distribution in the transform domain, so that the signal energy is concentrated in a few coefficients, and more It is convenient for quantization and decoding, thereby removing attribute redundancy, and achieving the purpose of improving attribute decoding efficiency and reconstruction performance.
- the execution subject may be a point cloud attribute information encoding device, or the device used to execute the point cloud attribute information encoding method in the point cloud attribute information encoding device control module.
- the method for encoding point cloud attribute information performed by the point cloud attribute information encoding device is taken as an example to illustrate the point cloud attribute information encoding device provided in the embodiment of the present application.
- FIG. 6 is a structural diagram of a device for encoding point cloud attribute information provided by an embodiment of the present application.
- the point cloud attribute information encoding device 600 includes:
- the first determination module 602 is configured to determine whether to perform discrete cosine transform (DCT) transformation on the K points to be encoded based on the second information associated with the first information;
- DCT discrete cosine transform
- the first transformation module 603 is used to perform DCT transformation on the K points to be coded to obtain the transformation coefficients of the K points to be coded when it is determined to perform DCT transformation on the K points to be coded;
- An encoding module 604 configured to quantize the transformation coefficients of the K points to be encoded, perform entropy encoding based on the quantized transformation coefficients, and generate a binary code stream;
- the first information includes the K points to be encoded
- the second information includes attribute prediction information of the K points to be encoded
- the first information includes the K points to be encoded
- the second information includes attribute reconstruction information of the N encoded points
- K is a positive integer
- N is an integer greater than 1.
- the first information includes the K points to be encoded
- the second information includes attribute prediction information of the K points to be encoded
- the first determining module 602 is also used for:
- the first information includes the first N encoded points of the K points to be encoded, and the second information includes attribute reconstruction information of the N encoded points;
- the first determining module 602 is also used for:
- the first information includes the K points to be encoded, and the first acquiring module 601 is further configured to:
- Sorting the point cloud to be encoded and acquiring K points to be encoded in the sorted point cloud to be encoded.
- the first acquiring module 601 is also configured to:
- the device further includes an attribute information prediction module, configured to:
- the target point to be encoded is any one of the K points to be encoded
- the attribute information prediction module is also used for:
- M, N1 and N2 are all positive integers.
- the search range of the Hilbert 1 order or the Morton 1 order is the first preset range
- the search range of the Hilbert 2 order or the Morton 2 order is the second preset range
- the Hilbert 2 order or the subsequent search range of the Morton 2 order is the third preset range or the second preset range
- the binary code stream includes a first attribute parameter and a second attribute parameter, the first attribute parameter is used to characterize the first preset range, and the second attribute parameter is used to characterize the second preset scope;
- the binary code stream further includes a third attribute parameter, which is used to represent The third preset range.
- the given preset search range is determined according to the correlation between the initial point number of the point cloud sequence and the volume of the bounding box of the input point cloud.
- the attribute information prediction module is also used for:
- the first information includes the K points to be encoded
- the attribute information prediction module is further used for:
- Obtaining the target point to be encoded among the K points to be encoded is L neighbor points with the closest Manhattan distance among the R neighbor points, and the target point to be encoded is any one of the K points to be encoded;
- T, R and L are all positive integers.
- the attribute information prediction module is also used for:
- the sum of the first weights corresponding to the K nodes to be coded is 1.
- the device further includes a third acquisition module, configured to:
- the first conversion module 603 is also used for:
- the device further includes an attribute reconstruction module, configured to:
- Inverse quantization is performed on the quantized-based transform coefficients, and inverse transforms are performed on the inverse transform coefficients obtained after inverse quantization, so as to obtain attribute reconstruction information of the K points to be encoded.
- the transform coefficients include low-frequency coefficients and high-frequency coefficients
- the encoding module 604 is further configured to:
- the attribute reconstruction module is also used to:
- the coding module 604 is also used for:
- the coding module 604 is also used for:
- the quantization step size of the high-frequency coefficient corresponding to the high-frequency coefficient is obtained, and the quantization step size of the low-frequency coefficient corresponding to the low-frequency coefficient is obtained.
- the coding module 604 is also used for:
- the quantization step size of the high-frequency transform coefficient is the original quantization step size, the preset quantization step size offset, and the high-frequency coefficient quantization step size
- the sum of offsets, the quantization step size of the low-frequency transform coefficient is the sum of the original quantization step size, the preset quantization step size offset and the low-frequency coefficient quantization step size offset;
- the quantization step size of the high-frequency coefficients is the original quantization step size, the preset quantization step size offset, and the low-frequency coefficient quantization step size offset
- the sum of shifts, the quantization step size of the low-frequency coefficients is equal to the quantization step size of the high-frequency coefficients.
- the quantized value of the high-frequency coefficient is 0, and when the low-frequency coefficient is smaller than a second preset threshold, the low-frequency The quantized value of the coefficient is 0.
- the coding module 604 is also used for:
- Entropy encoding is performed on the quantized attribute residual information of the K points to be encoded to generate a binary code stream.
- the device further includes an attribute reconstruction module, configured to:
- the device also includes:
- a third determining module configured to acquire identification information, where the identification information is used to indicate whether the device performs encoding of point cloud attribute information
- a writing module configured to write the identification information into the binary code stream.
- the apparatus performing point cloud attribute information encoding means that the apparatus executes the above corresponding operations based on the above modules (such as the first acquisition module, the first determination module, etc.), which will not be repeated here.
- the point cloud attribute information encoding device needs to decide whether to perform DCT transformation on the point to be encoded according to the attribute prediction information of the point to be encoded or the attribute reconstruction information of the encoded point , when it is determined that DCT transformation needs to be performed on the code point to be coded, DCT transformation is performed on the code point to be coded, and then the scattered distribution of attribute information in the spatial domain can be converted into a relatively concentrated distribution in the transform domain, so that the signal energy is concentrated on a few coefficients In , it is more convenient to perform quantization and coding, so as to remove attribute redundancy and achieve the purpose of improving attribute coding efficiency and reconstruction performance.
- the device for encoding point cloud attribute information in the embodiment of the present application may be a device, a device with an operating system or an electronic device, or a component, an integrated circuit, or a chip in a terminal.
- the apparatus or electronic equipment may be a mobile terminal or a non-mobile terminal.
- a mobile terminal may include but not limited to the types of terminals listed above, and a non-mobile terminal may be a server, a network attached storage (Network Attached Storage, NAS), a personal computer (personal computer, PC), a television (television , TV), teller machines or self-service machines, etc., are not specifically limited in this embodiment of the present application.
- the device for encoding point cloud attribute information provided in the embodiment of the present application can realize each process realized by the method embodiment in FIG. 3 or FIG. 4 and achieve the same technical effect. To avoid repetition, details are not repeated here.
- FIG. 7 is a structural diagram of a device for decoding point cloud attribute information provided by an embodiment of the present application.
- the point cloud attribute information decoding device 700 includes:
- the second determining module 702 is configured to determine whether to perform inverse DCT transformation on the K points to be decoded based on fourth information associated with the third information;
- the second transformation module 703 is configured to perform an inverse DCT transformation on the K points to be decoded to obtain attribute residues of the K points to be decoded when it is determined to perform an inverse DCT transformation on the K points to be decoded. poor information;
- the decoding module 704 is configured to obtain attribute reconstruction information of the K points to be decoded based on the attribute residual information and attribute prediction information of the K points to be decoded, so as to decode undecoded points in the point cloud to be decoded;
- the third information includes the K points to be decoded, and the fourth information includes attribute prediction information of the K points to be decoded; or, the third information includes the K points to be decoded
- the fourth information includes attribute reconstruction information of the N decoded points, K is a positive integer, and N is an integer greater than 1.
- the third information includes the K points to be decoded, and the fourth information includes attribute prediction information of the K points to be decoded;
- the second determination module 702 is also used for:
- the absolute ratio of the maximum attribute prediction value to the minimum attribute prediction value is smaller than a second threshold, it is determined to perform inverse DCT transformation on the K points to be decoded.
- the third information includes the first N decoded points of the K points to be decoded, and the fourth information includes attribute reconstruction information of the N decoded points;
- the second determination module 702 is also used for:
- the third information includes K points to be decoded
- the second obtaining module 701 is further configured to:
- the second acquiring module 701 is also configured to:
- the device also includes an attribute prediction module, configured to:
- the attribute prediction module is also used for:
- M, N1 and N2 are all positive integers.
- the search range of the Hilbert 1 order or the Morton 1 order is the first preset range
- the search range of the Hilbert 2 order or the Morton 2 order is the second preset Range
- the post-order search range of the Hilbert 2 order or Morton 2 order is the third preset range
- the binary code stream includes a first attribute parameter and a second attribute parameter, the first attribute parameter is used to characterize the first preset range, and the second attribute parameter is used to characterize the second preset range;
- the binary code stream further includes a third attribute parameter, which is used to represent The third preset range.
- the given preset search range is determined according to the correlation between the initial point number of the point cloud sequence and the volume of the bounding box of the input point cloud.
- the attribute prediction module is also used for:
- the second weight is the difference between the target point to be decoded and the neighbor point The reciprocal of the Manhattan distance between .
- the third information includes the K points to be decoded, and the attribute prediction module is also used for:
- Obtaining the target point to be decoded among the K points to be decoded is L neighbor points with the closest Manhattan distance among the R neighbor points, and the target point to be decoded is any one of the K points to be decoded;
- T, R and L are all positive integers.
- the attribute prediction module is also used for:
- the second weight is the difference between the target point to be decoded and the neighbor point The reciprocal of the Manhattan distance between .
- the sum of the first weights corresponding to the K nodes to be decoded is 1.
- the device also includes an inverse quantization module for:
- the second transformation module 703 is also used for:
- the transform coefficients include high-frequency coefficients and low-frequency coefficients
- the inverse quantization module is also used for:
- the inverse quantization module is also used for:
- the quantization step size of the high-frequency coefficient corresponding to the high-frequency coefficient is obtained, and the quantization step size of the low-frequency coefficient corresponding to the low-frequency coefficient is obtained.
- the inverse quantization module is also used for:
- the quantization step size of the high-frequency transform coefficient is the original quantization step size, the preset quantization step size offset, and the high-frequency coefficient quantization step size
- the sum of offsets, the quantization step size of the low-frequency transform coefficient is the sum of the original quantization step size, the preset quantization step size offset and the low-frequency coefficient quantization step size offset;
- the quantization step size of the high-frequency coefficient is the original quantization step size, the preset quantization step size offset, and the low-frequency coefficient quantization step size offset
- the sum of shifts, the quantization step size of the low-frequency coefficients is equal to the quantization step size of the high-frequency coefficients.
- the quantized value of the high-frequency coefficient is 0, and when the low-frequency coefficient is smaller than a second preset threshold, the low-frequency The quantized value of the coefficient is 0.
- the device also includes a quantization module for:
- the device further includes a fourth determining module, configured to:
- identification information from the binary code stream, where the identification information is used to indicate whether the device performs point cloud attribute information decoding
- the device performs point cloud attribute information decoding means that the device performs corresponding operations based on the above modules (such as the second acquisition module, the second determination module, etc.), which will not be repeated here.
- the point cloud attribute information decoding device in the process of decoding the point cloud to be decoded, it is necessary to determine whether to perform DCT on the point to be decoded according to the attribute prediction information of the point to be decoded or the attribute reconstruction information of the decoded point. Transformation, when it is determined that DCT transformation is required for the point to be decoded, DCT transformation is performed on the point to be decoded, and then the scattered distribution of attribute information in the spatial domain can be converted into a relatively concentrated distribution in the transformed domain, so that the signal energy is concentrated in a few Among the coefficients, it is more convenient to quantize and decode, thereby removing attribute redundancy, and achieving the purpose of improving attribute decoding efficiency and reconstruction performance.
- the device for decoding point cloud attribute information in the embodiment of the present application may be a device, a device with an operating system or an electronic device, or a component, an integrated circuit, or a chip in a terminal.
- the apparatus or electronic equipment may be a mobile terminal or a non-mobile terminal.
- the mobile terminal may include but not limited to the types of terminals listed above, and the non-mobile terminal may be a server, a network attached storage (Network Attached Storage, NAS), a personal computer (personal computer, PC), a television (television , TV), teller machines or self-service machines, etc., are not specifically limited in this embodiment of the present application.
- the device for decoding point cloud attribute information provided by the embodiment of the present application can realize each process realized by the method embodiment in FIG. 5 and achieve the same technical effect. In order to avoid repetition, details are not repeated here.
- this embodiment of the present application further provides a communication device 800, including a processor 801, a memory 802, and programs or instructions stored in the memory 802 and operable on the processor 801,
- a communication device 800 including a processor 801, a memory 802, and programs or instructions stored in the memory 802 and operable on the processor 801
- the communication device 800 is a terminal
- the program or instruction is executed by the processor 801
- each process of the above-mentioned method embodiment described in FIG. 3 or FIG. 4 is implemented, or each process of the above-mentioned method embodiment described in FIG. 5 is implemented, And can achieve the same technical effect, in order to avoid repetition, no more details here.
- the embodiment of the present application also provides a terminal, including a processor and a communication interface, and the processor is configured to implement each process of the method embodiment described in FIG. 3 or FIG. 4 above, or implement each process of the method embodiment described in FIG. 5 above.
- This terminal embodiment corresponds to the above-mentioned terminal-side method embodiment, and each implementation process and implementation mode of the above-mentioned method embodiment can be applied to this terminal embodiment, and can achieve the same technical effect.
- FIG. 9 is a schematic diagram of a hardware structure of a terminal implementing an embodiment of the present application.
- the terminal 900 includes, but is not limited to: a radio frequency unit 901, a network module 902, an audio output unit 903, an input unit 904, a sensor 905, a display unit 906, a user input unit 907, an interface unit 908, a memory 909, and a processor 910, etc. at least some of the components.
- the terminal 900 can also include a power supply (such as a battery) for supplying power to various components, and the power supply can be logically connected to the processor 910 through the power management system, so as to manage charging, discharging, and power consumption through the power management system. Management and other functions.
- a power supply such as a battery
- the terminal structure shown in FIG. 9 does not constitute a limitation on the terminal, and the terminal may include more or fewer components than shown in the figure, or combine some components, or arrange different components, which will not be repeated here.
- the input unit 904 may include a graphics processor (Graphics Processing Unit, GPU) 9041 and a microphone 9042, and the graphics processor 9041 is used for the image capture device (such as the image data of the still picture or video obtained by the camera) for processing.
- the display unit 906 may include a display panel 9061, and the display panel 9061 may be configured in the form of a liquid crystal display, an organic light emitting diode, or the like.
- the user input unit 907 includes a touch panel 9071 and other input devices 9072.
- the touch panel 9071 is also called a touch screen.
- the touch panel 9071 may include two parts, a touch detection device and a touch controller.
- Other input devices 9072 may include, but are not limited to, physical keyboards, function keys (such as volume control buttons, switch buttons, etc.), trackballs, mice, and joysticks, which will not be repeated here.
- the radio frequency unit 901 receives the downlink data from the network side device, and processes it to the processor 910; in addition, sends the uplink data to the network side device.
- the radio frequency unit 901 includes, but is not limited to, an antenna, at least one amplifier, a transceiver, a coupler, a low noise amplifier, a duplexer, and the like.
- the memory 909 can be used to store software programs or instructions as well as various data.
- the memory 909 may mainly include a program or instruction storage area and a data storage area, wherein the program or instruction storage area may store an operating system, an application program or instructions required by at least one function (such as a sound playback function, an image playback function, etc.) and the like.
- the memory 909 may include a high-speed random access memory, and may also include a nonvolatile memory, wherein the nonvolatile memory may be a read-only memory (Read-Only Memory, ROM), a programmable read-only memory (Programmable ROM) , PROM), erasable programmable read-only memory (Erasable PROM, EPROM), electrically erasable programmable read-only memory (Electrically EPROM, EEPROM) or flash memory.
- ROM Read-Only Memory
- PROM programmable read-only memory
- PROM erasable programmable read-only memory
- Erasable PROM Erasable PROM
- EPROM electrically erasable programmable read-only memory
- EEPROM electrically erasable programmable read-only memory
- flash memory for example at least one magnetic disk storage device, flash memory device, or other non-volatile solid-state storage device.
- the processor 910 may include one or more processing units; optionally, the processor 910 may integrate an application processor and a modem processor, wherein the application processor mainly processes the operating system, user interface, application programs or instructions, etc., Modem processors mainly handle wireless communications, such as baseband processors. It can be understood that the foregoing modem processor may not be integrated into the processor 910 .
- the processor 910 is configured to:
- DCT discrete cosine transform
- Quantizing the transform coefficients of the K points to be encoded performing entropy encoding based on the quantized transform coefficients, and generating a binary code stream
- the first information includes the K points to be encoded
- the second information includes attribute prediction information of the K points to be encoded
- the first information includes the K points to be encoded
- the second information includes attribute reconstruction information of the N encoded points
- K is a positive integer
- N is an integer greater than 1.
- the first information includes the K points to be encoded
- the second information includes attribute prediction information of the K points to be encoded
- the processor 910 is further configured to:
- the first information includes the first N encoded points of the K points to be encoded, and the second information includes attribute reconstruction information of the N encoded points; the processor 910 is further configured to :
- the first information includes the K points to be encoded
- the processor 910 is further configured to:
- Sorting the point cloud to be encoded and acquiring K points to be encoded in the sorted point cloud to be encoded.
- processor 910 is also used for:
- processor 910 is also used for:
- the target point to be encoded is any one of the K points to be encoded
- processor 910 is also used for:
- M, N1 and N2 are all positive integers.
- the search range of the Hilbert 1 order or the Morton 1 order is the first preset range
- the search range of the Hilbert 2 order or the Morton 2 order is the second preset range
- the Hilbert 2 order or the subsequent search range of the Morton 2 order is the third preset range or the second preset range
- the binary code stream includes a first attribute parameter and a second attribute parameter, the first attribute parameter is used to characterize the first preset range, and the second attribute parameter is used to characterize the second preset scope;
- the binary code stream further includes a third attribute parameter, which is used to represent The third preset range.
- the given preset search range is determined according to the correlation between the initial point number of the point cloud sequence and the volume of the bounding box of the input point cloud.
- processor 910 is also used for:
- the first information includes the K points to be encoded
- the processor 910 is further configured to:
- Obtaining the target point to be encoded among the K points to be encoded is L neighbor points with the closest Manhattan distance among the R neighbor points, and the target point to be encoded is any one of the K points to be encoded;
- T, R and L are all positive integers.
- processor 910 is also used for:
- the sum of the first weights corresponding to the K nodes to be coded is 1.
- processor 910 is also used for:
- the DCT transformation of the K points to be encoded is performed to obtain the transformation coefficients of the K points to be encoded, including:
- processor 910 is also used for:
- Inverse quantization is performed on the quantized-based transform coefficients, and inverse transforms are performed on the inverse transform coefficients obtained after inverse quantization, so as to obtain attribute reconstruction information of the K points to be encoded.
- processor 910 is also used for:
- the said inverse quantization of the encoded value obtained after the entropy encoding, and the inverse transformation of the inverse transformation coefficient obtained after the inverse quantization, to obtain the attribute reconstruction information of the K points to be encoded include:
- processor 910 is also used for:
- processor 910 is also used for:
- the quantization step size of the high-frequency coefficient corresponding to the high-frequency coefficient is obtained, and the quantization step size of the low-frequency coefficient corresponding to the low-frequency coefficient is obtained.
- processor 910 is also used for:
- the quantization step size of the high-frequency transform coefficient is the original quantization step size, the preset quantization step size offset, and the high-frequency coefficient quantization step size
- the sum of offsets, the quantization step size of the low-frequency transform coefficient is the sum of the original quantization step size, the preset quantization step size offset and the low-frequency coefficient quantization step size offset;
- the quantization step size of the high-frequency coefficients is the original quantization step size, the preset quantization step size offset, and the low-frequency coefficient quantization step size offset
- the sum of shifts, the quantization step size of the low-frequency coefficients is equal to the quantization step size of the high-frequency coefficients.
- the quantized value of the high-frequency coefficient is 0, and when the low-frequency coefficient is smaller than a second preset threshold, the low-frequency The quantized value of the coefficient is 0.
- processor 910 is also used for:
- the attribute residual information of the K points to be encoded is quantized
- Entropy encoding is performed on the quantized attribute residual information of the K points to be encoded to generate a binary code stream.
- processor 910 is also used for:
- processor 910 is also used for:
- identification information where the identification information is used to indicate whether the processor 910 performs point cloud attribute information encoding
- the processor 910 executes the point cloud attribute information encoding means that the processor 910 executes the above steps to realize the point cloud attribute information encoding, or realizes the point cloud attribute information encoding method as shown in FIG. 3 or FIG. 4 . The method will not be repeated here.
- the terminal 900 needs to decide whether to perform DCT transformation on the point cloud to be encoded according to the attribute prediction information of the point to be encoded or the attribute reconstruction information of the encoded point.
- the code point is subjected to DCT transformation
- the code point to be coded is subjected to DCT transformation, and then the scattered distribution of attribute information in the space domain can be converted into a relatively concentrated distribution in the transformation domain, so that the signal energy is concentrated in a few coefficients, which is more convenient Quantization and coding, so as to remove attribute redundancy and achieve the purpose of improving attribute coding efficiency and reconstruction performance.
- the processor 910 is configured to:
- the attribute reconstruction information of the K points to be decoded is obtained to decode the undecoded points in the point cloud to be decoded;
- the third information includes the K points to be decoded, and the fourth information includes attribute prediction information of the K points to be decoded; or, the third information includes the K points to be decoded
- the fourth information includes attribute reconstruction information of the N decoded points, K is a positive integer, and N is an integer greater than 1.
- the third information includes the K points to be decoded
- the fourth information includes attribute prediction information of the K points to be decoded
- the processor 910 is further configured to:
- the absolute ratio of the maximum attribute prediction value to the minimum attribute prediction value is smaller than a second threshold, it is determined to perform inverse DCT transformation on the K points to be decoded.
- the third information includes the first N decoded points of the K points to be decoded, and the fourth information includes attribute reconstruction information of the N decoded points; the processor 910 is further configured to:
- the third information includes K points to be decoded
- the processor 910 is further configured to:
- processor 910 is also used for:
- processor 910 is also used for:
- processor 910 is also used for:
- M, N1 and N2 are all positive integers.
- the search range of the Hilbert 1 order or the Morton 1 order is the first preset range
- the search range of the Hilbert 2 order or the Morton 2 order is the second preset Range
- the post-order search range of the Hilbert 2 order or Morton 2 order is the third preset range
- the binary code stream includes a first attribute parameter and a second attribute parameter, the first attribute parameter is used to characterize the first preset range, and the second attribute parameter is used to characterize the second preset range;
- the binary code stream further includes a third attribute parameter, which is used to represent The third preset range.
- the given preset search range is determined according to the correlation between the initial point number of the point cloud sequence and the volume of the bounding box of the input point cloud.
- processor 910 is also used for:
- the second weight is the difference between the target point to be decoded and the neighbor point The reciprocal of the Manhattan distance between .
- processor 910 is also used for:
- Obtaining the target point to be decoded among the K points to be decoded is L neighbor points with the closest Manhattan distance among the R neighbor points, and the target point to be decoded is any one of the K points to be decoded;
- T, R and L are all positive integers.
- processor 910 is also used for:
- the second weight is the difference between the target point to be decoded and the neighbor point The reciprocal of the Manhattan distance between .
- the sum of the first weights corresponding to the K nodes to be decoded is 1.
- processor 910 is also used for:
- the inverse DCT transformation of the K points to be decoded includes:
- the transform coefficients include high-frequency coefficients and low-frequency coefficients
- the processor 910 is further configured to:
- processor 910 is also used for:
- the quantization step size of the high-frequency coefficient corresponding to the high-frequency coefficient is obtained, and the quantization step size of the low-frequency coefficient corresponding to the low-frequency coefficient is obtained.
- processor 910 is also used for:
- the quantization step size of the high-frequency transform coefficient is the original quantization step size, the preset quantization step size offset, and the high-frequency coefficient quantization step size
- the sum of offsets, the quantization step size of the low-frequency transform coefficient is the sum of the original quantization step size, the preset quantization step size offset and the low-frequency coefficient quantization step size offset;
- the quantization step size of the high-frequency coefficient is the original quantization step size, the preset quantization step size offset, and the low-frequency coefficient quantization step size offset
- the sum of shifts, the quantization step size of the low-frequency coefficients is equal to the quantization step size of the high-frequency coefficients.
- the quantized value of the high-frequency coefficient is 0, and when the low-frequency coefficient is smaller than a second preset threshold, the low-frequency The quantized value of the coefficient is 0.
- processor 910 is also used for:
- processor 910 is also used for:
- identification information from the binary code stream, where the identification information is used to indicate whether the processor 910 performs point cloud attribute information decoding
- the processor 910 executes the point cloud attribute information decoding means that the processor 910 executes the above steps to realize the point cloud attribute information decoding, or realizes the point cloud attribute information decoding method as shown in FIG. 5 , the specific implementation method is here No longer
- the terminal 900 needs to decide whether to perform DCT transformation on the point cloud to be decoded according to the attribute prediction information of the point to be decoded or the attribute reconstruction information of the decoded point.
- DCT transformation is performed on the decoding point
- the DCT transformation is performed on the point to be decoded, and then the scattered distribution of attribute information in the spatial domain can be converted into a relatively concentrated distribution in the transformation domain, so that the signal energy is concentrated in a few coefficients, which is more convenient.
- Quantization and decoding so as to remove attribute redundancy and achieve the purpose of improving attribute decoding efficiency and reconstruction performance.
- the embodiment of the present application also provides a readable storage medium, on which a program or instruction is stored, and when the program or instruction is executed by the processor, each process of the above-mentioned method embodiment described in FIG. 3 or FIG. 4 is implemented. , or realize each process of the method embodiment described in FIG. 5 above, and can achieve the same technical effect. To avoid repetition, details are not repeated here.
- the processor is the processor in the terminal described in the foregoing embodiments.
- the readable storage medium includes computer readable storage medium, such as computer read-only memory (Read-Only Memory, ROM), random access memory (Random Access Memory, RAM), magnetic disk or optical disk, etc.
- the embodiment of the present application further provides a chip, the chip includes a processor and a communication interface, the communication interface is coupled to the processor, and the processor is used to run programs or instructions to implement the above-mentioned Figure 3 or Figure 4.
- the chip includes a processor and a communication interface
- the communication interface is coupled to the processor
- the processor is used to run programs or instructions to implement the above-mentioned Figure 3 or Figure 4.
- the chip mentioned in the embodiment of the present application may also be called a system-on-chip, a system-on-chip, a system-on-a-chip, or a system-on-a-chip.
- An embodiment of the present application further provides a computer program product, the computer program product is stored in a non-transitory storage medium, and the computer program product is executed by at least one processor to implement the method described above in FIG. 3 or FIG. 4
- the computer program product is executed by at least one processor to implement the method described above in FIG. 3 or FIG. 4
- Each process of the embodiment, or each process of the method embodiment described in FIG. 5 above, can achieve the same technical effect. To avoid repetition, details are not repeated here.
- the embodiment of the present application also provides a communication device configured to execute the various processes of the method embodiment described in FIG. 3 or FIG. 4 above, or implement the various processes of the method embodiment described in FIG. 5 above, and can achieve the same To avoid repetition, the technical effects will not be repeated here.
- the term “comprising”, “comprising” or any other variation thereof is intended to cover a non-exclusive inclusion such that a process, method, article or apparatus comprising a set of elements includes not only those elements, It also includes other elements not expressly listed, or elements inherent in the process, method, article, or device. Without further limitations, an element defined by the phrase “comprising a " does not preclude the presence of additional identical elements in the process, method, article, or apparatus comprising that element.
- the scope of the methods and devices in the embodiments of the present application is not limited to performing functions in the order shown or discussed, and may also include performing functions in a substantially simultaneous manner or in reverse order according to the functions involved. Functions are performed, for example, the described methods may be performed in an order different from that described, and various steps may also be added, omitted, or combined. Additionally, features described with reference to certain examples may be combined in other examples.
- the methods of the above embodiments can be implemented by means of software plus a necessary general-purpose hardware platform, and of course also by hardware, but in many cases the former is better implementation.
- the technical solution of the present application can be embodied in the form of computer software products, which are stored in a storage medium (such as ROM/RAM, magnetic disk, etc.) , CD-ROM), including several instructions to make a terminal (which may be a mobile phone, a computer, a server, an air conditioner, or a network device, etc.) execute the methods described in the various embodiments of the present application.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Discrete Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Compression, Expansion, Code Conversion, And Decoders (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
Abstract
Description
Claims (52)
- 一种点云属性信息编码方法,包括:对K个待编码点进行离散余弦变换DCT变换,得到所述K个待编码点的变换系数;对所述K个待编码点的变换系数进行量化,基于量化后的变换系数进行熵编码,生成二进制码流。
- 根据权利要求1所述的方法,在对K个待编码点进行DCT变换之前,所述方法还包括:获取第一信息;基于所述第一信息关联的第二信息确定对K个待编码点进行DCT变换;其中,所述第一信息包括所述K个待编码点,所述第二信息包括所述K个待编码点的属性预测信息;或者,所述第一信息包括所述K个待编码点的前N个已编码点,所述第二信息包括所述N个已编码点的属性重建信息,K为正整数,N为大于1的整数。
- 根据权利要求2所述的方法,其中,所述第一信息包括所述K个待编码点,所述第二信息包括所述K个待编码点的属性预测信息;所述基于所述第一信息关联的第二信息确定对K个待编码点进行DCT变换,包括:获取所述K个待编码点对应的属性预测信息中的最大属性预测值与最小属性预测值,在所述最大属性预测值与所述最小属性预测值的绝对差值小于第一阈值的情况下,则确定对所述K个待编码点进行DCT变换;或者,在所述最大属性预测值与所述最小属性预测值的绝对比值小于第二阈值的情况下,则确定对所述K个待编码点进行DCT变换。
- 根据权利要求2所述的方法,其中,所述第一信息包括所述K个待编码点的前N个已编码点,所述第二信息包括所述N个已编码点的属性重建信息;所述基于所述第一信息关联的第二信息确定对K个待编码点进行DCT变换,包括:获取所述N个已编码点对应的属性重建信息中的最大属性重建值和最小属性重建值,在所述最大属性重建值与所述最小属性重建值的绝对差值小于第三阈值的情况下,则确定对所述K个待编码点进行DCT变换;或者,在所述最大属性重建值与所述最小属性重建值的绝对比值小于第四阈值的情况下,则确定对所述K个待编码点进行DCT变换。
- 根据权利要求1所述的方法,其中,所述方法还包括:对待编码点云进行排序,获取排序后的所述待编码点云中的K个待编码点。
- 根据权利要求5所述的方法,其中,所述对待编码点云进行排序,获取排序后的所述待编码点云中的K个待编码点,包括:计算所述待编码点云中每个点对应的希尔伯特码,将所述待编码点云按照希尔伯特码进行排序,在排序后的所述待编码点云中按序依次选取K个待编码点;或者,计算所述待编码点云中每个点对应的莫顿码,将所述待编码点云按照莫顿码进行排序,在排序后的所述待编码点云中按序依次选取K个待编码点。
- 根据权利要求2所述的方法,其中,所述基于所述第一信息关联的第二信息确定对K个待编码点进行DCT变换之前,所述方法还包括:按照双希尔伯特序或双莫顿序,获取与目标待编码点曼哈顿距离最近的S个邻居点,所述目标待编码点为所述K个待编码点中的任一个;基于所述S个邻居点确定所述目标待编码点的初始属性预测信息;根据所述目标待编码点对应的第一权重及所述初始属性预测信息,确定所述目标待编码点的属性预测信息。
- 根据权利要求7所述的方法,其中,所述按照双希尔伯特序或双莫顿序,获取与目标待编码点曼哈顿距离最近的S个邻居点,包括:在给定的预设搜索范围中,按照希尔伯特1序,获取目标待编码点的前序M个点,以及按照希尔伯特2序,获取所述目标待编码点的前序N1个点和后序N2个点,在基于M、N1及N2确定的目标范围内获取与目标待编码点曼哈顿距离最近的S个邻居点;或者,在所述给定的预设搜索范围中,按照莫顿1序,获取目标待编码点的前 序M个点,以及按照莫顿2序,获取所述目标待编码点的前序N1个点和后序N2个点,在基于M、N1及N2确定的目标范围内获取与目标待编码点曼哈顿距离最近的S个邻居点;其中,M、N1和N2均为正整数。
- 根据权利要求8所述的方法,其中,所述希尔伯特1序或莫顿1序的搜索范围为第一预设范围,所述希尔伯特2序或莫顿2序的前序搜索范围为第二预设范围,所述希尔伯特2序或莫顿2序的后序搜索范围为第三预设范围或所述第二预设范围;其中,所述二进制码流包括第一属性参数和第二属性参数,所述第一属性参数用于表征所述第一预设范围,所述第二属性参数用于表征所述第二预设范围;在所述希尔伯特2序或莫顿2序的后序搜索范围为第三预设范围的情况下,所述二进制码流还包括第三属性参数,所述第三属性参数用于表征所述第三预设范围。
- 根据权利要求8所述的方法,其中,所述给定的预设搜索范围根据点云序列的初始点数和输入点云包围盒体积的相关关系决定。
- 根据权利要求7所述的方法,其中,所述基于所述S个邻居点确定所述目标待编码点的初始属性预测信息,包括:基于所述S个邻居点中每一个邻居点及对应的第二权重,确定所述目标待编码点的初始属性预测信息,所述第二权重为所述目标待编码点与所述邻居点之间的曼哈顿距离的倒数。
- 根据权利要求2所述的方法,其中,所述第一信息包括所述K个待编码点,所述基于所述第一信息关联的第二信息确定对K个待编码点进行DCT变换之前,所述方法还包括:以所述K个待编码点中的第一个点作为参考点,获取所述参考点的T个邻居点;获取所述T个邻居点中与所述参考点曼哈顿距离最近的R个邻居点;获取所述K个待编码点中目标待编码点在所述R个邻居点中曼哈顿距离最近的L个邻居点,所述目标待编码点为所述K个待编码点中的任一个;基于L个邻居点确定所述目标待编码点的初始属性预测信息;根据所述目标待编码点对应的第一权重及所述初始属性预测信息,确定所述目标待编码点的属性预测信息;其中,T、R和L均为正整数。
- 根据权利要求12所述的方法,其中,所述基于L个邻居点确定所述目标待编码点的初始属性预测信息,包括:基于所述L个邻居点中每一个邻居点及对应的第二权重,确定所述目标待编码点的初始属性预测信息,所述第二权重为所述目标待编码点与所述邻居点之间的曼哈顿距离的倒数。
- 根据权利要求7-13任一项所述的方法,其中,所述K个待编码节点各自对应的所述第一权重的和为1。
- 根据权利要求1所述的方法,其中,在对K个待编码点进行DCT变换,得到所述K个待编码点的变换系数之前,所述方法还包括:基于所述K个待编码点的属性预测信息,获取所述K个待编码点的属性残差信息;所述对K个待编码点进行DCT变换,得到所述K个待编码点的变换系数,包括:对所述K个待编码点的属性残差信息进行DCT变换,得到所述K个待编码点对应的变换系数。
- 根据权利要求15所述的方法,其中,所述对所述K个待编码点的变换系数进行量化,基于量化后的变换系数进行熵编码之后,所述方法还包括:对所述基于量化后的变换系数进行反量化,对反量化后得到的反变换系数进行反变换,以获取所述K个待编码点的属性重建信息。
- 根据权利要求16所述的方法,其中,所述变换系数包括低频系数和高频系数,所述对所述K个待编码点的变换系数进行量化,基于量化后的变换系数进行熵编码,包括:对所述K个待编码点对应的高频系数和低频系数进行量化,并对量化后的高频系数和低频系数分别进行熵编码,以得到第一编码值和第二编码值;所述对所述熵编码后得到的编码值进行反量化,对反量化后得到的反变 换系数进行反变换,以获取所述K个待编码点的属性重建信息,包括:对所述第一编码值和所述第二编码值进行反量化,得到反量化后的反高频系数和反低频系数;基于所述反高频系数和所述反低频系数进行DCT反变换,得到所述K个待编码点对应的反属性残差信息;根据所述K个待编码点的所述属性预测信息及所述反属性残差信息,获取所述K个待编码点的属性重建信息。
- 根据权利要求17所述的方法,其中,所述对所述K个待编码点对应的高频系数和低频系数进行量化,包括:获取所述高频系数对应的高频系数量化步长,以及获取所述低频系数对应的低频系数量化步长;根据所述高频系数及所述高频系数量化步长对所述高频系数进行量化,根据所述低频系数及所述低频系数量化步长对所述低频系数进行量化。
- 根据权利要求18所述的方法,其中,所述获取所述高频系数对应的高频系数量化步长,以及获取所述低频系数对应的低频系数量化步长,包括:根据所述K个待编码点的属性信息对应的分量分布情况,获取所述高频系数对应的高频系数量化步长,以及获取所述低频系数对应的低频系数量化步长。
- 根据权利要求19所述的方法,其中,所述根据所述K个待编码点的属性信息对应的分量分布情况,获取所述高频系数对应的高频系数量化步长,以及获取所述低频系数对应的低频系数量化步长,包括:在所述K个待编码点的属性信息对应的分量分布平坦的情况下,所述高频变换系数的量化步长为原始量化步长、预设量化步长偏移和高频系数量化步长偏移之和,所述低频变换系数的量化步长为原始量化步长、预设量化步长偏移和低频系数量化步长偏移之和;在所述K个待编码点的属性信息对应的分量分布不平坦的情况下,所述高频系数的量化步长为原始量化步长、预设量化步长偏移和低频系数量化步长偏移之和,所述低频系数的量化步长与所述高频系数的量化步长相等。
- 根据权利要求18所述的方法,其中,在所述高频系数小于第一预设 阈值的情况下,所述高频系数量化后的值为0,在所述低频系数小于第二预设阈值的情况下,所述低频系数量化后的值为0。
- 根据权利要求15所述的方法,其中,所述方法还包括:在确定对所述K个待编码点不进行DCT变换的情况下,对所述K个待编码点的属性残差信息进行量化;对所述K个待编码点量化后的属性残差信息进行熵编码,生成二进制码流。
- 根据权利要求22所述的方法,其中,在对所述K个待编码点量化后的属性残差信息进行熵编码之后,所述方法还包括:对所述熵编码后得到的编码值进行反量化,得到所述K个待编码点反量化后的反属性残差信息;根据所述K个待编码点的所述属性预测信息及所述反属性残差信息,获取所述K个待编码点的属性重建信息。
- 根据权利要求2所述的方法,其中,在所述获取第一信息之前,所述方法还包括:获取标识信息,所述标识信息用于指示是否执行所述方法;根据所述标识信息确定是否执行所述方法;在所述生成二进制码流之后,所述方法还包括:将所述标识信息写入所述二进制码流。
- 一种点云属性信息解码方法,包括:对K个待解码点进行离散余弦变换DCT反变换,得到所述K个待解码点的属性残差信息;基于所述K个待解码点的属性残差信息和属性预测信息,获取所述K个待解码点的属性重建信息,以对待解码点云中的未解码点进行解码。
- 根据权利要求24所述的方法,在对所述K个待解码点进行DCT反变换之前,所述方法还包括:获取第三信息;基于所述第三信息关联的第四信息确定对K个待解码点进行DCT反变换,其中,所述第三信息包括所述K个待解码点,所述第四信息包括所述K 个待解码点的属性预测信息;或者,所述第三信息包括所述K个待解码点的前N个已解码点,所述第四信息包括所述N个已解码点的属性重建信息,K为正整数,N为大于1的整数。
- 根据权利要求26所述的方法,其中,所述第三信息包括所述K个待解码点,所述第四信息包括所述K个待解码点的属性预测信息;所述基于所述第三信息关联的第四信息确定对K个待解码点进行DCT反变换,包括:获取所述K个待解码点对应的属性预测信息中的最大属性预测值与最小属性预测值,在所述最大属性预测值与所述最小属性预测值的绝对差值小于第一阈值的情况下,则确定对所述K个待解码点进行DCT反变换;或者,在所述最大属性预测值与所述最小属性预测值的绝对比值小于第二阈值的情况下,则确定对所述K个待解码点进行DCT反变换。
- 根据权利要求26所述的方法,其中,所述第三信息包括所述K个待解码点的前N个已解码点,所述第四信息包括所N个已解码点的属性重建信息;所述基于所述第三信息关联的第四信息确定对K个待解码点进行DCT反变换,包括:获取所述N个已解码点对应的属性重建信息中的最大属性重建值和最小属性重建值,在所述最大属性重建值与所述最小属性重建值的绝对差值小于第三阈值的情况下,则确定对所述K个待解码点进行DCT反变换;或者,在所述最大属性重建值与所述最小属性重建值的绝对比值小于第四阈值的情况下,则确定对所述K个待解码点进行DCT反变换。
- 根据权利要求25所述的方法,其中,所述方法还包括:对待解码点云进行排序,获取排序后的所述待解码点云中的K个待解码点。
- 根据权利要求29所述的方法,其中,所述对待解码点云进行排序,获取排序后的所述待解码点云中的K个待解码点,包括:计算所述待解码点云中每个点对应的希尔伯特码,将所述待解码点云按照希尔伯特码进行排序,在排序后的所述待解码点云中按序依次选取K个待 解码点;或者,计算所述待解码点云中每个点对应的莫顿码,将所述待解码点云按照莫顿码进行排序,在排序后的所述待解码点云中按序依次选取K个待解码点。
- 根据权利要求26所述的方法,其中,所述基于所述第三信息关联的第四信息确定是否对K个待解码点进行DCT反变换之前,所述方法还包括:按照双希尔伯特序或双莫顿序,获取与目标待解码点曼哈顿距离最近的S个邻居点,所述目标待解码点为所述K个待解码点中的任一个;基于所述S个邻居点确定所述目标待解码点的初始属性预测信息;根据所述目标待解码点对应的第一权重及所述初始属性预测信息,确定所述目标待解码点的属性预测信息。
- 根据权利要求31所述的方法,其中,所述按照双希尔伯特序或双莫顿序,获取与目标待解码点曼哈顿距离最近的S个邻居点,包括:在给定的预设搜索范围中,按照希尔伯特1序,获取目标待解码点的前序M个点,以及按照希尔伯特2序,获取所述目标待解码点的前序N1个点和后序N2个点,在基于M、N1及N2确定的目标范围内获取与目标待解码点曼哈顿距离最近的S个邻居点;或者,在所述给定的预设搜索范围中,按照莫顿1序,获取目标待解码点的前序M个点,以及按照莫顿2序,获取所述目标待解码点的前序N1个点和后序N2个点,在基于M、N1及N2确定的目标范围内获取与目标待解码点曼哈顿距离最近的S个邻居点;其中,M、N1和N2均为正整数。
- 根据权利要求32所述的方法,其中,所述希尔伯特1序或莫顿1序的搜索范围为第一预设范围,所述希尔伯特2序或莫顿2序的前序搜索范围为第二预设范围,所述希尔伯特2序或莫顿2序的后序搜索范围为第三预设范围;其中,二进制码流包括第一属性参数和第二属性参数,所述第一属性参数用于表征所述第一预设范围,所述第二属性参数用于表征所述第二预设范围;在所述希尔伯特2序或莫顿2序的后序搜索范围为第三预设范围的情况 下,所述二进制码流还包括第三属性参数,所述第三属性参数用于表征所述第三预设范围。
- 根据权利要求32所述的方法,其中,所述给定的预设搜索范围根据点云序列的初始点数和输入点云包围盒体积的相关关系决定。
- 根据权利要求31所述的方法,其中,所述基于所述S个邻居点确定所述目标待解码点的初始属性预测信息,包括:基于所述S个邻居点中每一个邻居点及对应的第二权重,确定所述目标待解码点的初始属性预测信息,所述第二权重为所述目标待解码点与所述邻居点之间的曼哈顿距离的倒数。
- 根据权利要求26所述的方法,其中,所述第三信息包括所述K个待解码点,所述基于所述第三信息关联的第四信息确定是否对K个待解码点进行DCT反变换之前,所述方法还包括:以所述K个待解码点中的第一个点作为参考点,获取所述参考点的T个邻居点;获取所述T个邻居点中与所述参考点曼哈顿距离最近的R个邻居点;获取所述K个待解码点中目标待解码点在所述R个邻居点中曼哈顿距离最近的L个邻居点,所述目标待解码点为所述K个待解码点中的任一个;基于L个邻居点确定所述目标待解码点的初始属性预测信息;根据所述目标待解码点对应的第一权重及所述初始属性预测信息,确定所述目标待解码点的属性预测信息;其中,T、R和L均为正整数。
- 根据权利要求36所述的方法,其中,所述基于L个邻居点确定所述目标待解码点的初始属性预测信息,包括:基于所述L个邻居点中每一个邻居点及对应的第二权重,确定所述目标待解码点的初始属性预测信息,所述第二权重为所述目标待解码点与所述邻居点之间的曼哈顿距离的倒数。
- 根据权利要求31-37任一项所述的方法,其中,所述K个待解码节点各自对应的所述第一权重的和为1。
- 根据权利要求25所述的方法,其中,所述对K个待解码点进行DCT 反变换之前,所述方法还包括:获取所述K个待解码点的变换系数;对所述变换系数进行反量化,获得反量化后的变换系数;所述对K个待解码点进行DCT反变换,包括:基于所述反量化后的变换系数对所述K个待解码点进行DCT反变换。
- 根据权利要求39所述的方法,其中,所述变换系数包括高频系数和低频系数,所述对所述变换系数进行反量化,获得反量化后的变换系数,包括:获取所述高频系数对应的高频系数量化步长,以及获取所述低频系数对应的低频系数量化步长;根据所述高频系数及所述高频系数量化步长对所述高频系数进行反量化,根据所述低频系数及所述低频系数量化步长对所述低频系数进行反量化。
- 根据权利要求40所述的方法,其中,所述获取所述高频系数对应的高频系数量化步长,以及获取所述低频系数对应的低频系数量化步长,包括:根据所述K个待解码点的属性信息对应的分量分布情况,获取所述高频系数对应的高频系数量化步长,以及获取所述低频系数对应的低频系数量化步长。
- 根据权利要求41所述的方法,其中,所述根据所述K个待解码点的属性信息对应的分量分布情况,获取所述高频系数对应的高频系数量化步长,以及获取所述低频系数对应的低频系数量化步长,包括:在所述K个待解码点的属性信息对应的分量分布平坦的情况下,所述高频变换系数的量化步长为原始量化步长、预设量化步长偏移和高频系数量化步长偏移之和,所述低频变换系数的量化步长为原始量化步长、预设量化步长偏移和低频系数量化步长偏移之和;在所述K个待解码点的属性信息对应的分量分布不平坦的情况下,所述高频系数的量化步长为原始量化步长、预设量化步长偏移和低频系数量化步长偏移之和,所述低频系数的量化步长与所述高频系数的量化步长相等。
- 根据权利要求40所述的方法,其中,在所述高频系数小于第一预设阈值的情况下,所述高频系数量化后的值为0,在所述低频系数小于第二预 设阈值的情况下,所述低频系数量化后的值为0。
- 根据权利要求39所述的方法,其中,所述方法还包括:在确定对所述K个待解码点不进行DCT反变换的情况下,对所述K个待解码点的变换系数进行量化,以得到所述K个待解码点的属性残差信息。
- 根据权利要求26所述的方法,其中,在所述获取第三信息之前,所述方法还包括:从二进制码流中获取标识信息,所述标识信息用于指示是否执行所述方法;根据所述标识信息确定是否执行所述方法。
- 一种点云属性信息编码装置,包括:第一变换模块,用于对K个待编码点进行离散余弦变换DCT变换,得到所述K个待编码点的变换系数;编码模块,用于对所述K个待编码点的变换系数进行量化,基于量化后的变换系数进行熵编码,生成二进制码流。
- 一种点云属性信息解码装置,包括:第二变换模块,用于对K个待解码点进行离散余弦变换DCT反变换,得到所述K个待解码点的属性残差信息;解码模块,用于基于所述K个待解码点的属性残差信息和属性预测信息,获取所述K个待解码点的属性重建信息,以对待解码点云中的未解码点进行解码。
- 一种终端,包括处理器、存储器及存储在所述存储器上并可在所述处理器上运行的程序或指令,其中,所述程序或指令被所述处理器执行时实现如权利要求1-24中任一项所述的点云属性信息编码方法的步骤,或者实现如权利要求25-45中任一项所述的点云属性信息解码方法的步骤。
- 一种可读存储介质,所述可读存储介质上存储程序或指令,其中,所述程序或指令被处理器执行时实现如权利要求1-24中任一项所述的点云属性信息编码方法的步骤,或者实现如权利要求25-45中任一项所述的点云属性信息解码方法的步骤。
- 一种芯片,包括处理器和通信接口,其中,所述通信接口和所述处 理器耦合,所述处理器用于运行程序或指令,实现如权利要求1-24中任一项所述的点云属性信息编码方法的步骤,或者实现如权利要求25-45中任一项所述的点云属性信息解码方法的步骤。
- 一种计算机程序产品,其中,所述计算机程序产品被存储在非易失的存储介质中,所述计算机程序产品被至少一个处理器执行时实现如权利要求1-24中任一项所述的点云属性信息编码方法的步骤,或者实现如权利要求25-45中任一项所述的点云属性信息解码方法的步骤。
- 一种通信设备,被配置为执行如权利要求1-24中任一项所述的点云属性信息编码方法的步骤,或者如权利要求25-45中任一项所述的点云属性信息解码方法的步骤。
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP22819652.3A EP4354872A4 (en) | 2021-06-11 | 2022-06-10 | METHOD AND APPARATUS FOR ENCODING AND DECODING POINT CLOUD ATTRIBUTE INFORMATION, AND RELATED DEVICE |
KR1020237042781A KR20240006667A (ko) | 2021-06-11 | 2022-06-10 | 점군 속성 정보 부호화 방법, 복호화 방법, 장치 및 관련 기기 |
JP2023573432A JP2024520542A (ja) | 2021-06-11 | 2022-06-10 | 点群属性情報コーディング方法、デコーディング方法、装置及び関連機器 |
US18/530,090 US20240121439A1 (en) | 2021-06-11 | 2023-12-05 | Point cloud attribute information encoding method and apparatus, point cloud attribute information decoding method and apparatus, and related device |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110654116 | 2021-06-11 | ||
CN202110654116.7 | 2021-06-11 | ||
CN202110817311.7A CN115474046A (zh) | 2021-06-11 | 2021-07-20 | 点云属性信息编码方法、解码方法、装置及相关设备 |
CN202110817311.7 | 2021-07-20 |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US18/530,090 Continuation US20240121439A1 (en) | 2021-06-11 | 2023-12-05 | Point cloud attribute information encoding method and apparatus, point cloud attribute information decoding method and apparatus, and related device |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2022258055A1 true WO2022258055A1 (zh) | 2022-12-15 |
Family
ID=84365178
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2022/098193 WO2022258055A1 (zh) | 2021-06-11 | 2022-06-10 | 点云属性信息编码方法、解码方法、装置及相关设备 |
Country Status (6)
Country | Link |
---|---|
US (1) | US20240121439A1 (zh) |
EP (1) | EP4354872A4 (zh) |
JP (1) | JP2024520542A (zh) |
KR (1) | KR20240006667A (zh) |
CN (1) | CN115474046A (zh) |
WO (1) | WO2022258055A1 (zh) |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN118433395B (zh) * | 2024-07-02 | 2024-09-03 | 北京数原数字化城市研究中心 | 一种编码方法、装置及电子设备 |
Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20180053324A1 (en) * | 2016-08-19 | 2018-02-22 | Mitsubishi Electric Research Laboratories, Inc. | Method for Predictive Coding of Point Cloud Geometries |
CN108322742A (zh) * | 2018-02-11 | 2018-07-24 | 北京大学深圳研究生院 | 一种基于帧内预测的点云属性压缩方法 |
US20190007682A1 (en) * | 2017-07-03 | 2019-01-03 | Panasonic Intellectual Property Corporation Of America | Coding method, decoding method, encoder, and decoder |
US20190200046A1 (en) * | 2019-03-04 | 2019-06-27 | Intel Corporation | Volumetric video compression with motion history |
US20190208200A1 (en) * | 2016-09-30 | 2019-07-04 | Interdigital Vc Holdings, Inc. | Method and apparatus for omnidirectional video coding with adaptive intra most probable modes |
EP3699867A1 (en) * | 2019-02-19 | 2020-08-26 | Nokia Technologies Oy | An apparatus, a method and a computer program for volumetric video |
CN111953998A (zh) * | 2020-08-16 | 2020-11-17 | 西安电子科技大学 | 基于dct变换的点云属性编码及解码方法、装置及系统 |
CN112335242A (zh) * | 2018-06-25 | 2021-02-05 | 韩国电子通信研究院 | 用于使用量化参数对图像进行编码/解码的方法和设备以及存储比特流的记录介质 |
CN112449754A (zh) * | 2019-07-04 | 2021-03-05 | 深圳市大疆创新科技有限公司 | 一种数据编码、数据解码方法、设备及存储介质 |
WO2021062771A1 (zh) * | 2019-09-30 | 2021-04-08 | Oppo广东移动通信有限公司 | 颜色分量预测方法、编码器、解码器及计算机存储介质 |
Family Cites Families (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108632607B (zh) * | 2018-05-09 | 2019-06-21 | 北京大学深圳研究生院 | 一种基于多角度自适应帧内预测的点云属性压缩方法 |
US11010931B2 (en) * | 2018-10-02 | 2021-05-18 | Tencent America LLC | Method and apparatus for video coding |
US11122102B2 (en) * | 2019-07-03 | 2021-09-14 | Lg Electronics Inc. | Point cloud data transmission apparatus, point cloud data transmission method, point cloud data reception apparatus and point cloud data reception method |
US11895307B2 (en) * | 2019-10-04 | 2024-02-06 | Apple Inc. | Block-based predictive coding for point cloud compression |
CN112565757B (zh) * | 2020-12-03 | 2022-05-13 | 西安电子科技大学 | 基于通道差异化的点云属性编码及解码方法、装置及系统 |
-
2021
- 2021-07-20 CN CN202110817311.7A patent/CN115474046A/zh active Pending
-
2022
- 2022-06-10 JP JP2023573432A patent/JP2024520542A/ja active Pending
- 2022-06-10 EP EP22819652.3A patent/EP4354872A4/en active Pending
- 2022-06-10 WO PCT/CN2022/098193 patent/WO2022258055A1/zh active Application Filing
- 2022-06-10 KR KR1020237042781A patent/KR20240006667A/ko active Search and Examination
-
2023
- 2023-12-05 US US18/530,090 patent/US20240121439A1/en active Pending
Patent Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20180053324A1 (en) * | 2016-08-19 | 2018-02-22 | Mitsubishi Electric Research Laboratories, Inc. | Method for Predictive Coding of Point Cloud Geometries |
US20190208200A1 (en) * | 2016-09-30 | 2019-07-04 | Interdigital Vc Holdings, Inc. | Method and apparatus for omnidirectional video coding with adaptive intra most probable modes |
US20190007682A1 (en) * | 2017-07-03 | 2019-01-03 | Panasonic Intellectual Property Corporation Of America | Coding method, decoding method, encoder, and decoder |
CN108322742A (zh) * | 2018-02-11 | 2018-07-24 | 北京大学深圳研究生院 | 一种基于帧内预测的点云属性压缩方法 |
CN112335242A (zh) * | 2018-06-25 | 2021-02-05 | 韩国电子通信研究院 | 用于使用量化参数对图像进行编码/解码的方法和设备以及存储比特流的记录介质 |
EP3699867A1 (en) * | 2019-02-19 | 2020-08-26 | Nokia Technologies Oy | An apparatus, a method and a computer program for volumetric video |
US20190200046A1 (en) * | 2019-03-04 | 2019-06-27 | Intel Corporation | Volumetric video compression with motion history |
CN112449754A (zh) * | 2019-07-04 | 2021-03-05 | 深圳市大疆创新科技有限公司 | 一种数据编码、数据解码方法、设备及存储介质 |
WO2021062771A1 (zh) * | 2019-09-30 | 2021-04-08 | Oppo广东移动通信有限公司 | 颜色分量预测方法、编码器、解码器及计算机存储介质 |
CN111953998A (zh) * | 2020-08-16 | 2020-11-17 | 西安电子科技大学 | 基于dct变换的点云属性编码及解码方法、装置及系统 |
Also Published As
Publication number | Publication date |
---|---|
EP4354872A4 (en) | 2024-10-23 |
JP2024520542A (ja) | 2024-05-24 |
CN115474046A (zh) | 2022-12-13 |
KR20240006667A (ko) | 2024-01-15 |
EP4354872A1 (en) | 2024-04-17 |
US20240121439A1 (en) | 2024-04-11 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2022257970A1 (zh) | 点云的几何信息编码处理方法、解码处理方法及相关设备 | |
WO2022257971A1 (zh) | 点云编码处理方法、点云解码处理方法及相关设备 | |
WO2022257969A1 (zh) | 点云编码处理方法、解码处理方法及相关设备 | |
WO2022258009A1 (zh) | 熵编码、解码方法及装置 | |
US20240121439A1 (en) | Point cloud attribute information encoding method and apparatus, point cloud attribute information decoding method and apparatus, and related device | |
WO2022257968A1 (zh) | 点云编码方法、点云解码方法及终端 | |
CN116567246A (zh) | Avc编码方法和装置 | |
WO2023025024A1 (zh) | 点云属性编码方法、点云属性解码方法及终端 | |
WO2023098803A1 (zh) | 点云编码处理方法、点云解码处理方法及相关设备 | |
WO2022258010A1 (zh) | 点云编码处理方法、解码处理方法及装置 | |
WO2023098814A1 (zh) | 属性量化、反量化方法、装置及设备 | |
WO2023098820A1 (zh) | 点云编码、解码方法、装置及通信设备 | |
WO2022257978A1 (zh) | 点云编码方法、解码方法及装置 | |
WO2023155779A1 (zh) | 编码方法、解码方法、装置及通信设备 | |
WO2022257972A1 (zh) | 编码、解码控制方法及装置 | |
WO2023098802A1 (zh) | 点云属性编码方法、点云属性解码方法及终端 | |
WO2024217303A1 (zh) | 变换系数编码方法、变换系数解码方法及终端 | |
WO2023098807A1 (zh) | 点云编、解码处理方法、装置、编码设备及解码设备 | |
WO2024217301A1 (zh) | 点云编码处理方法、点云解码处理方法及相关设备 | |
WO2024149182A1 (zh) | 变换系数编码方法、变换系数解码方法及终端 | |
WO2024188138A1 (zh) | 点云编码处理方法、点云解码处理方法及相关设备 | |
CN115883832A (zh) | 基于人工智能的视频编码处理方法、装置、设备及介质 | |
CN118175334A (zh) | 属性变换编码方法、属性变换解码方法及终端 | |
CN118575200A (zh) | 用于网格压缩的自适应几何滤波 | |
CN115474041A (zh) | 点云属性的预测方法、装置及相关设备 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 22819652 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2023573432 Country of ref document: JP Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 20237042781 Country of ref document: KR Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 1020237042781 Country of ref document: KR |
|
WWE | Wipo information: entry into national phase |
Ref document number: 202327086833 Country of ref document: IN |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2022819652 Country of ref document: EP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
ENP | Entry into the national phase |
Ref document number: 2022819652 Country of ref document: EP Effective date: 20240111 |