WO2023174337A1 - Procédé de codage, procédé de décodage, appareil et dispositif - Google Patents

Procédé de codage, procédé de décodage, appareil et dispositif Download PDF

Info

Publication number
WO2023174337A1
WO2023174337A1 PCT/CN2023/081641 CN2023081641W WO2023174337A1 WO 2023174337 A1 WO2023174337 A1 WO 2023174337A1 CN 2023081641 W CN2023081641 W CN 2023081641W WO 2023174337 A1 WO2023174337 A1 WO 2023174337A1
Authority
WO
WIPO (PCT)
Prior art keywords
precision
geometric information
information
geometric
dimensional
Prior art date
Application number
PCT/CN2023/081641
Other languages
English (en)
Chinese (zh)
Inventor
吕卓逸
邹文杰
张伟
杨付正
Original Assignee
维沃移动通信有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 维沃移动通信有限公司 filed Critical 维沃移动通信有限公司
Publication of WO2023174337A1 publication Critical patent/WO2023174337A1/fr

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/20Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using video object coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/44Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder

Definitions

  • This application belongs to the field of coding and decoding technology, and specifically relates to a coding and decoding method, device and equipment.
  • Three-dimensional mesh can be considered the most popular representation method of three-dimensional models in the past many years, and it plays an important role in many applications. Its expression is simple, so it is widely integrated into the graphics processing unit (GPU) of computers, tablets and smartphones with hardware algorithms, specifically used to render three-dimensional meshes.
  • GPU graphics processing unit
  • 3D mesh geometry information can be compressed using point cloud compression algorithms.
  • point cloud compression algorithm is used to compress the geometric information of the three-dimensional mesh model, but the compression efficiency is not high.
  • Embodiments of the present application provide an encoding and decoding method, device and equipment, which can solve the problem of low compression efficiency in the compression method of three-dimensional grid geometric information in related technologies.
  • the first aspect provides an encoding method, including:
  • the encoding end quantifies the geometric information of the target three-dimensional grid and obtains the first-precision geometric information and the second-precision geometric information;
  • the encoding end obtains the first geometric map and placeholder map of the first precision geometric information
  • the encoding end obtains a second geometric diagram including the second precision geometric information and the first precision geometric information based on the second precision geometric information and the first geometric diagram;
  • the first precision geometric information is the geometric information after quantization of the target three-dimensional grid
  • the second precision geometric information is the geometric information lost during the quantization process of the target three-dimensional grid
  • an encoding device including:
  • the quantification module is used to quantify the geometric information of the target three-dimensional grid and obtain the first-precision geometric information and the second-precision geometric information;
  • a first acquisition module configured to acquire the first geometric diagram and placeholder diagram of the first precision geometric information
  • a second acquisition module configured to acquire a second geometric diagram including the second precision geometric information and the first precision geometric information according to the second precision geometric information and the first geometric diagram;
  • An encoding module used to encode the second geometric diagram and the placeholder diagram
  • the first precision geometric information is the geometric information after quantization of the target three-dimensional grid
  • the second precision geometric information is the geometric information lost during the quantization process of the target three-dimensional grid
  • the third aspect provides a decoding method, including:
  • the decoder decomposes the obtained code stream of the target three-dimensional grid, and obtains the placeholder map and the second geometric map containing the second-precision geometric information and the first-precision geometric information;
  • the decoding end obtains second precision geometric information and the first geometric map according to the second geometric map
  • the decoding end obtains first precision geometric information based on the first geometric diagram and the placeholder diagram;
  • the decoder performs inverse quantization according to the second precision geometric information and the first precision geometric information to obtain the geometric information of the target three-dimensional grid;
  • the first precision geometric information is the geometric information after quantization of the target three-dimensional grid
  • the second precision geometric information is the geometric information lost during the quantization process of the target three-dimensional grid
  • a decoding device including:
  • the third acquisition module is used to decompose the acquired code stream of the target three-dimensional grid, and acquire the placeholder map and the second geometric map containing the second-precision geometric information and the first-precision geometric information;
  • the fourth acquisition module is used to acquire the second precision geometric information and the first geometric figure according to the second geometric figure;
  • the fifth acquisition module is used to acquire the first precision geometric information according to the first geometric diagram and the placeholder diagram;
  • a sixth acquisition module configured to perform inverse quantization according to the second precision geometric information and the first precision geometric information, and obtain the geometric information of the target three-dimensional grid;
  • the first precision geometric information is the geometric information after quantization of the target three-dimensional grid
  • the second precision geometric information is the geometric information lost during the quantization process of the target three-dimensional grid
  • a coding device including a processor and a memory.
  • the memory stores programs or instructions that can be run on the processor. When the program or instructions are executed by the processor, the first The steps of the method described in this aspect.
  • an encoding device including a processor and a communication interface, wherein the processor is used to quantify the geometric information of the target three-dimensional grid and obtain first precision geometric information and second precision geometric information; obtain The first geometric diagram and placeholder diagram of the first precision geometric information; according to the second precision geometric information and the first geometric diagram, obtain a second geometry including the second precision geometric information and the first precision geometric information.
  • Figure encode the second geometric diagram and the placeholder diagram; wherein the first precision geometric information is the quantized geometric information of the target three-dimensional grid, and the second precision geometric information is the Geometric information lost during target 3D mesh quantization.
  • a decoding device including a processor and a memory.
  • the memory stores programs or instructions that can be run on the processor.
  • the program or instructions are executed by the processor, the third process is implemented. The steps of the method described in this aspect.
  • a decoding device including a processor and a communication interface, wherein the processor is used to decompose the obtained code stream of the target three-dimensional grid, obtain the placeholder map and the second precision geometric information and A second geometric diagram of the first precision geometric information; according to the second geometric diagram, obtain the second precision geometric information and the first geometric diagram; obtain the first precision geometric information according to the first geometric diagram and the placeholder map; Perform inverse quantization according to the second precision geometric information and the first precision geometric information to obtain the geometric information of the target three-dimensional grid;
  • the first precision geometric information is the geometric information after quantization of the target three-dimensional grid
  • the second precision geometric information is the geometric information lost during the quantization process of the target three-dimensional grid
  • a communication system including: an encoding device and a decoding device.
  • the encoding device can be used to perform the steps of the method described in the first aspect
  • the decoding device can be used to perform the steps of the method described in the third aspect. steps of the method.
  • a readable storage medium is provided. Programs or instructions are stored on the readable storage medium. When the programs or instructions are executed by a processor, the steps of the method described in the first aspect are implemented, or the steps of the method are implemented as described in the first aspect. The steps of the method described in the third aspect.
  • a chip in an eleventh aspect, includes a processor and a communication interface.
  • the communication interface is coupled to the processor.
  • the processor is used to run programs or instructions to implement the method described in the first aspect. method, or implement a method as described in the third aspect.
  • a computer program/program product is provided, the computer program/program product being stored in In the storage medium, the computer program/program product is executed by at least one processor to implement the steps of the method described in the first aspect.
  • the geometric map of the first-precision geometric information is used to encode the second-precision geometric information, thereby improving the coding efficiency of the second-precision geometric information, thereby improving the efficiency of compressing the geometric information using a quantization scheme.
  • Figure 1 is a schematic flow chart of the encoding method according to the embodiment of the present application.
  • Figure 2 is a schematic diagram of raw patch
  • Figure 3 is a schematic diagram of the fine division process based on grid
  • Figure 4 is a schematic diagram of the eight directions of patch arrangement
  • Figure 5 is a schematic diagram of a geometric diagram after arranging the arrangement values corresponding to the high-precision geometric information according to the positions of the low-precision geometric information;
  • Figure 6 is a schematic diagram of the geometric diagram after arranging high-precision arrangement values according to the distribution of closely spaced pixels that is translated to the left in the horizontal direction;
  • Figure 7 is a schematic diagram of the video-based three-dimensional grid geometric information encoding framework
  • Figure 8 is a schematic module diagram of an encoding device according to an embodiment of the present application.
  • Figure 9 is a schematic structural diagram of an encoding device according to an embodiment of the present application.
  • Figure 10 is a schematic flow chart of the decoding method according to the embodiment of the present application.
  • Figure 11 is a block diagram of geometric information reconstruction
  • Figure 12 is a schematic diagram of the video-based three-dimensional grid geometric information decoding framework
  • Figure 13 is a schematic module diagram of a decoding device according to an embodiment of the present application.
  • Figure 14 is a schematic structural diagram of a communication device according to an embodiment of the present application.
  • first, second, etc. in the description and claims of this application are used to distinguish similar objects. It is not used to describe a specific order or sequence. It is to be understood that the terms so used are interchangeable under appropriate circumstances so that the embodiments of the present application can be practiced in sequences other than those illustrated or described herein, and that "first" and “second” are distinguished objects It is usually one type, and the number of objects is not limited.
  • the first object can be one or multiple.
  • “and/or” in the description and claims indicates at least one of the connected objects, and the character “/" generally indicates that the related objects are in an "or” relationship.
  • LTE Long Term Evolution
  • LTE-Advanced, LTE-A Long Term Evolution
  • CDMA Code Division Multiple Access
  • TDMA Time Division Multiple Access
  • FDMA Frequency Division Multiple Access
  • OFDMA Orthogonal Frequency Division Multiple Access
  • SC-FDMA Single-carrier Frequency Division Multiple Access
  • system and “network” in the embodiments of this application are often used interchangeably, and the described technology can be used not only for the above-mentioned systems and radio technologies, but also for other systems and radio technologies.
  • NR New Radio
  • the following description describes a New Radio (NR) system for example purposes, and uses NR terminology in much of the following description, but these techniques can also be applied to applications other than NR system applications, such as 6th Generation , 6G) communication system.
  • NR New Radio
  • Three-dimensional models have become a new generation of digital media after audio, images, and videos.
  • Three-dimensional mesh and point cloud are two commonly used three-dimensional model representation methods.
  • 3D mesh models are more interactive and realistic, making them useful in various fields such as commerce, manufacturing, construction, education, medicine, entertainment, art, and military. has been increasingly widely used.
  • 3D mesh models obtained through 3D scanning equipment or 3D modeling software
  • the size and complexity of data are also growing rapidly. Therefore, how to efficiently compress 3D mesh data is the key to realizing convenient transmission, storage and processing of 3D mesh data.
  • a three-dimensional mesh often contains three main types of information: topological information, geometric information and attribute information.
  • Topological information is used to describe the connection relationship between elements such as vertices and patches in the mesh; geometric information is the three-dimensional coordinates of all vertices in the mesh; attribute information records other information attached to the mesh, such as normal vectors, Texture coordinates and colors, etc.
  • Some traditional general data compression methods can reduce a certain amount of 3D grid data, due to the particularity of 3D grid data, directly using these compression methods to compress 3D grid data often cannot achieve ideal results. Therefore, the compression of three-dimensional mesh data faces new challenges.
  • geometric data often takes up more storage space than topological data. Efficient compression of geometric data will be of extremely important significance in reducing the storage space of three-dimensional grid data. Therefore, the compression of three-dimensional mesh geometric information has become a research focus.
  • the 3D mesh geometric information compression algorithm can use the 3D geometric information compression algorithm of point cloud.
  • V-PCC Video-based Point Cloud Compression
  • G-PCC Geometry-based Point Cloud Compression, geometry-based point cloud compression
  • V-PCC The main idea of V-PCC is to project the geometric and attribute information of the point cloud into a two-dimensional video, and use existing video coding technology to compress the two-dimensional video to achieve the purpose of compressing the point cloud.
  • the geometric coding of V-PCC is achieved by projecting geometric information into placeholder video and geometric video, and using a video encoder to encode the two videos separately.
  • the process of V-PCC geometric information encoding mainly includes: first generating a three-dimensional patch (3D patch), which refers to a set of vertices with the same and connected projection plane in the point cloud.
  • the current method of generating 3D patches is to use nearby points to estimate the normal vector of each vertex, calculate the projection plane of each vertex based on the normal vector of each point and the normal vector of the preset plane, and connect the connected vertices with the same projection plane Make up a patch.
  • the 3D patch is projected onto a 2D plane to form a 2D patch, and the 2D patches are arranged on a two-dimensional image. This process is called patch packing.
  • a placeholder map is an image that represents the placeholder information of vertices in a two-dimensional image.
  • the position value of the vertex projection in the placeholder map is 1, and the other position values are 0.
  • a placeholder image is generated. What is stored in the geometric graph is the distance from each vertex to the projection plane. The depth information of each vertex can be directly calculated using the three-dimensional coordinates of the vertex, the projection plane of the vertex and the placeholder map, thereby generating a geometric map.
  • Image filling methods include "push-pull" background filling algorithm, filling method based on sparse linear model (Sparse Linear Model), harmonic background filling (Harmonic Background Filling) and other methods.
  • a video encoder in related technologies is used to compress the placeholder map and the geometric map to obtain a video code stream.
  • the placeholder video code stream, the geometric video code stream, and the sub-code stream containing patch information are synthesized into the final total code stream.
  • this embodiment of the present application provides an encoding method, including:
  • Step 101 The encoding end quantifies the geometric information of the target three-dimensional grid and obtains the first precision geometric information and the second precision geometric information;
  • the target three-dimensional grid mentioned in this application can be understood as the three-dimensional grid corresponding to any video frame.
  • the geometric information of the target three-dimensional grid can be understood as the coordinates of the vertices in the three-dimensional grid. These coordinates are usually Refers to three-dimensional coordinates.
  • the first precision geometric information can be understood as low-precision geometric information, that is, the low-precision geometric information refers to the quantized geometric information of the target three-dimensional grid, that is, the three-dimensional coordinates of each vertex included in the quantized target three-dimensional grid. coordinate information.
  • the second precision geometric information can be understood as high-precision geometric information, and the high-precision geometric information can be regarded as geometric information lost in the quantization process, that is, lost three-dimensional coordinate information.
  • the vertex spacing of the quantized three-dimensional mesh is reduced, thereby reducing the spacing of the two-dimensional vertices after projection, thereby improving compression efficiency.
  • Step 102 The encoding end obtains the first geometric diagram and placeholder diagram of the first precision geometric information
  • the first geometric diagram can be understood as a low-precision geometric diagram.
  • Step 103 The encoding end obtains a second geometric diagram including the second precision geometric information and the first precision geometric information based on the second precision geometric information and the first geometric diagram;
  • Step 104 The encoding end encodes the second geometric image and the placeholder image
  • step 101 is:
  • the encoding end quantizes each vertex in the target three-dimensional grid according to the quantization parameter of each component to obtain first precision geometric information
  • the encoding end obtains second precision geometric information based on the first precision geometric information and the quantization parameter of each component.
  • the quantization parameters of each component can be flexibly set according to usage requirements; the quantization parameters mainly include quantization parameters on the three components of X, Y and Z.
  • the f 1 function in Formula 1 to Formula 3 is a quantization function.
  • the input of the quantization function is the coordinates of a certain dimension and the quantization parameter of that dimension, and the output is the quantized coordinate value;
  • the f 2 function input in Formula 4 to Formula 6 is It is the original coordinate value, the quantized coordinate value and the quantized parameter of this dimension, and the output is a high-precision coordinate value.
  • the f 1 function can be calculated in a variety of ways.
  • a more common calculation method is as shown in Formula 7 to Formula 9, which is calculated by dividing the original coordinates of each dimension by the quantization parameter of that dimension.
  • / is the division operator, and the result of the division operation can be rounded in different ways, such as rounding, rounding down, rounding up, etc.
  • the implementation methods corresponding to Formula 7 to Formula 9 are as shown in Formula 10 to Formula 12, where, * is the multiplication operator.
  • the f 1 function and f 2 function can be implemented using bit operations, such as Formula 13 to Formula 18:
  • the quantization parameters QP x , QP y and QP z can be set flexibly.
  • the quantization parameters of different components are not necessarily equal. You can use the correlation of the quantization parameters of different components to establish the relationship between QP x , QP y and QP z , and set different quantization parameters for different components; secondly, different spaces
  • the quantization parameters of the regions are not necessarily equal.
  • the quantization parameters can be adaptively set according to the sparsity of the vertex distribution in the local region.
  • the high-precision geometric information contains detailed information of the outline of the three-dimensional mesh.
  • the high-precision geometric information (x h , y h , z h ) can be further processed.
  • the importance of high-precision geometric information of vertices in different areas is different. For areas where vertices are sparsely distributed, the distortion of high-precision geometric information will not have a major impact on the visual effect of the three-dimensional mesh.
  • the method further includes:
  • the information of the supplementary points is obtained;
  • the information of supplementary points refers to the information of points that require additional processing generated during the quantification process. That is to say, the supplementary points are points that require additional processing generated during the quantification process.
  • the coordinate positions overlap.
  • the supplementary point information includes at least one of the following:
  • the third-precision geometric information can be understood as low-precision geometric information of the supplementary points, that is, the quantized three-dimensional coordinate information of the supplementary points.
  • the fourth precision geometric information can be understood as the high-precision geometric information of the supplementary point, that is, the three-dimensional coordinate information of the supplementary point that is lost during the quantization process.
  • the specific method for obtaining the information of the supplementary points is: the encoding end determines the information of the supplementary points based on the geometric information of the target three-dimensional grid and the first precision geometric information.
  • the points with repeated low-precision geometric information are used as supplementary points and encoded separately.
  • the geometric information of supplementary points can also be divided into two parts: low-precision geometric information and high-precision geometric information. According to the application's requirements for compression distortion, you can choose to retain all supplementary points or only a part of them.
  • the high-precision geometric information of the supplementary points can also be further quantified, or the high-precision geometric information of only some points can be retained.
  • the geometric diagram of the supplementary point is obtained by:
  • the encoding end arranges the third precision geometric information of the supplementary points into a first original slice
  • the encoding end arranges the fourth precision geometric information of the supplementary points into a second original slice in the same order as the first original slice;
  • the encoding end compresses the first original slice and the second original slice to obtain a geometric map of the supplementary points.
  • the low-precision part and the high-precision part of the geometric information of the supplementary points are encoded separately.
  • the first step is to obtain the vertex arrangement order, scan the low-precision geometric map line by line from left to right, and The scanning order of each vertex is used as the order of vertices in the raw patch.
  • the second step is to generate raw patch.
  • a raw patch is a rectangular patch formed by arranging the three-dimensional coordinates of the vertices row by row as shown in Figure 2.
  • the low-precision geometric information of the vertices is arranged in order to obtain the low-precision geometric information raw patch; then, the high-precision geometric information is arranged into the supplementary point in the same order as the supplementary point low-precision raw patch.
  • Point high-precision raw patch Point high-precision raw patch; finally, to compress the supplementary point low-precision raw patch and high-precision raw patch, a variety of compression methods can be used. Among them, one method is to encode the values in the raw patch by run-length coding, entropy coding, etc. The other method is to add supplementary point low-precision raw patch and supplementary point high-precision raw patch to the low-precision geometric map. In the blank area, the geometric map of the supplementary points is obtained; finally, the geometric map is encoded by the video encoder to obtain the geometric map sub-stream of the supplementary points.
  • step 102 includes:
  • Step 1021 The encoding end divides the first precision geometric information into three-dimensional slices
  • the main step is to divide the low-precision geometric information into patches to obtain multiple three-dimensional patches;
  • the specific implementation method of this step is: the encoding end determines each element contained in the first-precision geometric information.
  • the projection plane of the vertices; the coding end performs slice division on the vertices contained in the first precision geometric information according to the projection plane; the coding end clusters the vertices contained in the first precision geometric information, and obtains Each piece after division.
  • the process of patch division mainly includes: first estimating the normal vector of each vertex, selecting the candidate projection plane with the smallest angle between the plane normal vector and the vertex normal vector as the projection plane of the vertex; then, according to the projection The plane initially divides the vertices, and vertices with the same and connected projection planes are composed into a patch; finally, a fine division algorithm is used to optimize the clustering results to obtain the final three-dimensional patch (3D patch).
  • the projection plane of each vertex is initially selected.
  • the normal vector of the candidate projection plane is Select the plane whose normal vector direction is closest to the vertex normal vector direction as the projection plane of the vertex.
  • the calculation process of plane selection is as shown in Equation 21:
  • the fine division process can use a grid-based algorithm to reduce the time complexity of the algorithm.
  • the grid-based fine division algorithm flow is shown in Figure 3, which specifically includes:
  • Step S301 Divide the (x, y, z) geometric coordinate space into voxels.
  • Step S302 Find filled voxels.
  • Filled voxels refer to voxels that contain at least one point in the grid.
  • Step S303 calculate the smoothing score of each filled voxel on each projection plane, recorded as voxScoreSmooth.
  • the voxel smoothing score of the voxel on a certain projection plane is the number of points gathered to the projection plane through the initial segmentation process.
  • Step S304 use KD-Tree partitioning to find neighboring filled voxels, recorded as nnFilledVoxels, that is, the nearest filled voxels of each filled voxel (within the search radius and/or limited to the maximum number of adjacent voxels).
  • Step S305 use the voxel smoothing score of the nearest neighbor filled voxel in each projection plane to calculate the smoothing score (scoreSmooth) of each filled voxel.
  • the calculation process is as shown in Equation 22:
  • Step S306 Calculate the normal score using the normal vector of the vertex and the normal vector of the candidate projection plane, recorded as scoreNormal.
  • the calculation process is as shown in Formula 23:
  • p is the index of the projection plane and i is the index of the vertex.
  • Step S307 use scoreSmooth and scoreNormal to calculate the final score of each voxel on each projection plane.
  • the calculation process is as shown in Equation 24:
  • i is the vertex index
  • p is the index of the projection plane
  • v is the voxel index where vertex i is located.
  • Step S308 Use the scores in step 307 to cluster the vertices to obtain finely divided patches.
  • Step 1022 The encoding end performs two-dimensional projection on the divided three-dimensional slice to obtain the two-dimensional slice;
  • this process is to project the 3D patch onto a two-dimensional plane to obtain a two-dimensional patch (2D patch).
  • Step 1023 The encoding end packages the two-dimensional slices to obtain two-dimensional image information
  • this step implements patch packing.
  • the purpose of patch packing is to arrange 2D patches on a two-dimensional image.
  • the basic principle of patch packing is to arrange patches on a two-dimensional image without overlapping or The pixel-free parts of the patch are partially overlapped and arranged on the two-dimensional image.
  • the patches are arranged more closely and have time domain consistency to improve coding performance.
  • the resolution of the 2D image is WxH
  • the minimum block size that defines the patch arrangement is T, which specifies the minimum distance between different patches placed on this 2D grid.
  • patches are inserted and placed on the 2D grid according to the non-overlapping principle.
  • Each patch occupies an area consisting of an integer number of T ⁇ T blocks.
  • the patches can choose a variety of different arrangement directions. For example, eight different arrangement directions can be adopted, as shown in Figure 4, including 0 degrees, 180 degrees, 90 degrees, 270 degrees, and mirror images of the first four directions.
  • a patch arrangement method with temporal consistency is adopted.
  • a GOF Group of frame, frame group
  • all patches of the first frame are arranged in order from largest to smallest.
  • the temporal consistency algorithm is used to adjust the order of patches.
  • the patch information can be obtained based on the information in the process of obtaining the two-dimensional image information, and then the patch information can be encoded to obtain the patch information sub-stream.
  • the patch information records the process of obtaining the two-dimensional image.
  • the information of each step operation that is, the patch information includes: patch division information, patch projection plane information, and patch packing position information.
  • Step 1024 The encoding end obtains the first geometric map and placeholder map based on the two-dimensional image information.
  • the process of obtaining the placeholder map is mainly: using the patch arrangement information obtained by patch packing, setting the position of the vertex in the two-dimensional image to 1, and setting the remaining positions to 0 to obtain the placeholder map.
  • the main process is: in the process of obtaining the 2D patch through projection, the distance from each vertex to the projection plane is saved. This distance is called depth.
  • the geometric map compression part is to compress each 2D patch in the 2D patch.
  • the depth value of the vertex is arranged to the position of the vertex in the placeholder map, and the first geometric map is obtained.
  • step 103 is:
  • the encoding end arranges the three-dimensional coordinate values of each vertex in the second precision geometric information, and obtains an arrangement value corresponding to each vertex;
  • the encoding end arranges the arrangement values corresponding to each vertex in the first area, and obtains a second geometric map containing second precision geometric information and first precision geometric information;
  • the first area is the area remaining after excluding the first precision geometric information projection distribution area from the two-dimensional projection distribution area of the target three-dimensional grid's geometric information.
  • the values of the three dimensions of xyz are first arranged vertex by vertex according to the preset arrangement rules (it should be noted that the preset arrangement rules can be pre-agreed, usually In this case, the encoding end encodes the preset arrangement rule into the code stream when encoding, and the decoding end obtains the preset arrangement rule when decoding the code stream, so as to use the preset arrangement rule for decoding) for arrangement.
  • the sequence of xyz, yzx or zxy is arranged as a value (arrangement value).
  • the arrangement value of each vertex is placed in the area saved after the projection distribution of the quantized low-precision geometric information is reduced (i.e., the first area), so that the high-precision arrangement value corresponding to the low-precision vertex is arranged in the low-precision geometry.
  • the high-precision geometric information area in the graph a geometric graph containing low-precision geometric information and high-precision geometric information is obtained.
  • the implementation of arranging the arrangement values corresponding to each vertex in the first area and obtaining the second geometric graph containing the second precision geometric information and the first precision geometric information can be implemented by one of the following:
  • the encoding end arranges the pixels of the arrangement value corresponding to each vertex in the first area according to the pixel distribution position of the projection point corresponding to the vertex in the first precision geometric information, and obtains the second precision geometric information. and a second geometric map of the first precision geometric information;
  • the projection point refers to the point corresponding to the vertex in the two-dimensional image after projecting the vertex of the three-dimensional grid into a two-dimensional slice.
  • a pixel in a two-dimensional image corresponds to the projection point of a vertex.
  • the pixel value corresponding to the projected point in the placeholder map is usually 1, and the depth value corresponding to the projected point in the geometric map is usually greater than 0.
  • the pixel distribution of the projection points corresponding to the vertices in the first precision geometric information can be the pixel distribution of the projection points corresponding to the low-precision vertices in the placeholder map and/or the geometry map.
  • the placeholder is used.
  • the pixel positions of the projection points corresponding to the low-precision vertices in the graph and/or geometric diagram correspond to the arrangement values of the vertices in the high-precision geometric information.
  • the encoding end arranges the pixels of the projection points corresponding to each vertex in the second precision geometric information in the first area according to the pixel distribution position of the projection point corresponding to the vertex in the first precision geometric information. , and perform translation processing on the pixels, arrange the arrangement values of the vertices at the pixel positions of the projection points corresponding to the vertices in the first area after translation processing, and obtain the second precision geometric information and the first precision geometric information.
  • this method is to closely arrange the arrangement values of the vertices in the second precision geometric information in the horizontal or vertical direction, that is, in the area saved after the projection distribution of the low-precision geometric information is reduced after quantization, the horizontal Translate the closely arranged distribution in the direction or vertical direction, arrange the high-precision arrangement values corresponding to the low-precision vertices in the high-precision geometric information area in the low-precision geometric map, and obtain a geometric map that also contains high-precision geometric information.
  • Figure 6 it is a schematic diagram of the geometric image after arranging the high-precision arrangement values according to the distribution of closely spaced pixels in the horizontal direction. The arrangement values corresponding to the high-precision geometric information are distributed at the bottom of the image in Figure 6. half part.
  • the encoding end scans the pixels of the projection points corresponding to the vertices in the second precision geometric information along the first direction row by row or column by column, and re-scans the position index of the pixels in each row or column within the row or column. number within;
  • the encoding end arranges the arrangement value corresponding to each vertex in the second precision geometric information in the first area at the pixel position specified by the position index of the pixel of the projection point corresponding to the vertex, and obtains the The second precision geometric information and the second geometric map of the first precision geometric information;
  • the first direction is a horizontal direction or a vertical direction.
  • the pixels of the projection points corresponding to the vertices in the second precision geometric information are not closely adjacent to each other, but have a certain interval.
  • the scan renumbering mentioned in the embodiment of the present application That is, pixels with gaps in the same row or column are renumbered in a pixel-adjacent manner. For example, in a certain row, there are 5 pixels, and their numbers from left to right are 1, 3, 5, 7, 8. After scanning from left to right, renumber the five pixels. After renumbering, the new numbers of the five pixels from left to right are 1, 2, 3, 4, and 5.
  • this implementation method is to first scan the pixels of the projection points corresponding to the vertices in the second precision geometric information in rows or columns, and renumber them.
  • the pixels are scanned according to the horizontal direction. Scan row by row in the direction, and re-number the position index within the row for the pixel position of the projection point corresponding to the vertex in each row.
  • scanning in the vertical direction scan column by column in the vertical direction, and scan each column column by column.
  • the pixel position of the projection point corresponding to the vertex in the column is re-numbered by the position index in the column; by translating the pixels of the projection point corresponding to the vertex in the second precision geometric information in the horizontal or vertical direction, and the arrangement value of the vertex Arrange to the pixel position of the projection point corresponding to the vertex to obtain a geometric map with closely arranged pixels.
  • a video encoder is used to encode high-precision geometric information.
  • the geometric graph is compressed with low-precision geometric information to obtain a geometric graph sub-stream.
  • the encoding efficiency of the video encoder can be improved, and the encoding efficiency of high-precision geometric information can be improved overall.
  • the three-dimensional grid is first quantized, from which low-precision geometric information, high-precision geometric information and supplementary point information may be obtained.
  • the projection method can be used to generate placeholder images and low-precision geometric images, which are encoded through a video encoder.
  • This application specifically proposes a method for representing and encoding high-precision geometric information.
  • the high-precision geometric information is first arranged vertex by vertex into one value according to the preset arrangement rules in the xyz three dimensions.
  • the closely arranged distribution is translated according to the two-dimensional projection distribution of the low-precision geometric information or in the horizontal direction (or vertical direction) to convert the high-precision vertices corresponding to the low-precision vertices.
  • Precision arrangement values are arranged in high-precision geometric information areas in low-precision geometry maps.
  • the geometric information of supplementary points is divided into low-precision parts and high-precision parts, which can be separately encoded into a code stream, or the low-precision parts and high-precision parts can be arranged in a low-precision geometric map using raw patches.
  • a video encoder is used to encode placeholder images, low-precision geometric maps containing high-precision geometric information.
  • the video decoder is used to decode the placeholder image and the low-precision geometric image containing high-precision geometric information.
  • the low-precision geometric information in the placeholder image and the low-precision geometric image containing high-precision geometric information can be used to reconstruct the placeholder image. Construct a low-precision three-dimensional grid, and use the high-precision geometric information (arrangement values) in the low-precision geometric map to reconstruct the low-precision three-dimensional grid into a high-precision three-dimensional grid.
  • the video-based three-dimensional mesh geometric information encoding framework of the embodiment of the present application is shown in Figure 7.
  • the overall encoding process is: first, you can choose whether to sample and simplify the three-dimensional mesh before quantization; then, quantify the three-dimensional mesh , which may produce three parts: low-precision geometric information, high-precision geometric information and supplementary point information; for low-precision geometric information, projection is used to divide patches, and patch arrangement generates patch sequence compression information (patch division information), Placement map and low-precision geometric map; for the existing high-precision geometric information, the high-precision geometric information can be arranged into one value in the xyz three dimensions vertex by vertex.
  • the closely arranged distribution is translated according to the two-dimensional projection distribution of the low-precision geometric information or in the horizontal direction (or vertical direction) to convert the high-precision vertices corresponding to the low-precision vertices.
  • the precision arrangement values are arranged in the high-precision geometric information area in the low-precision geometric map; for possible supplementary points, the geometric information of the supplementary points can be divided into low-precision parts and high-precision parts, which are arranged in raw patches and encoded separately. into one code stream, or add the raw patch to the geometric map; finally, encode the patch sequence to compress the information, placeholder map, and geometric map, and mix the multiple sub-streams to obtain the final output code stream.
  • this application provides an implementation method of encoding the geometric information of a three-dimensional grid.
  • the three components of the three-dimensional coordinate value of the quantified high-precision geometric information are The encoding is changed to a component encoding and arranged according to the position of the low-precision geometric map, which can improve the compression efficiency of the video encoder for high-precision geometric information; in addition, the tight arrangement of the high-precision geometric map can also improve the video
  • the encoder's compression efficiency for high-precision geometric information through the above scheme, the efficiency of compressing geometric information using quantization schemes can be further improved.
  • this embodiment of the present application provides an encoding device 800, which includes:
  • Quantization module 801 is used to quantify the geometric information of the target three-dimensional grid and obtain first-precision geometric information and second-precision geometric information;
  • the first acquisition module 802 is used to acquire the first geometric diagram and placeholder diagram of the first precision geometric information
  • the second acquisition module 803 is configured to acquire a second geometric diagram containing the second precision geometric information and the first precision geometric information according to the second precision geometric information and the first geometric diagram;
  • Encoding module 804 used to encode the second geometric image and the placeholder image
  • the first precision geometric information is the geometric information after quantization of the target three-dimensional grid
  • the second precision geometric information is the geometric information lost during the quantization process of the target three-dimensional grid
  • the second acquisition module 803 includes:
  • a first acquisition unit configured to arrange the coordinate values in three dimensions of each vertex in the second precision geometric information and obtain an arrangement value corresponding to each vertex
  • the second acquisition unit is used to arrange the arrangement values corresponding to each vertex in the first area, and acquire the second geometric map containing the second precision geometric information and the first precision geometric information;
  • the first area is the area remaining after excluding the first precision geometric information projection distribution area from the two-dimensional projection distribution area of the target three-dimensional grid's geometric information.
  • the second acquisition unit is used to implement one of the following:
  • the pixels of the projection points corresponding to each vertex in the second precision geometric information are arranged according to the pixel distribution position of the projection point corresponding to the vertex in the first precision geometric information, and the pixels are translated.
  • the pixels are translated, the arrangement values of the vertices are arranged at the pixel positions of the projection points corresponding to the vertices in the first area after translation processing, and the second precision geometric information and the second precision are obtained.
  • the implementation method of the second geometric graph of one-precision geometric information is:
  • the arrangement value corresponding to each vertex in the second precision geometric information is arranged in the first area at the pixel position specified by the position index of the pixel of the projection point corresponding to the vertex, and the second precision geometry is obtained.
  • information and a second geometric map of the first precision geometric information are arranged in the first area at the pixel position specified by the position index of the pixel of the projection point corresponding to the vertex, and the second precision geometry is obtained.
  • the first direction is a horizontal direction or a vertical direction.
  • the first acquisition module 802 includes:
  • a dividing unit used to divide the first precision geometric information into three-dimensional slices
  • the third acquisition unit is used to perform two-dimensional projection on the divided three-dimensional slices to obtain the two-dimensional slices;
  • the fourth acquisition unit is used to package the two-dimensional slices and acquire two-dimensional image information
  • the fifth acquisition unit is used to acquire the first geometric map and placeholder map according to the two-dimensional image information.
  • the fourth acquisition unit packages the two-dimensional slices and acquires the two-dimensional image information, it also includes:
  • the sixth acquisition unit is used to acquire slice information based on the information in the process of acquiring two-dimensional image information
  • the seventh acquisition unit is used to encode the slice information and obtain the slice information sub-stream.
  • the device also includes:
  • the third acquisition module is used to obtain supplementary point information based on quantification of the geometric information of the target three-dimensional grid
  • the information of the supplementary points is the information of points generated during the quantification process that require additional processing.
  • the third acquisition module is used for:
  • Information on supplementary points is determined based on the geometric information of the target three-dimensional grid and the first precision geometric information.
  • the supplementary point information includes at least one of the following:
  • Third-precision geometric information of the supplementary point where the third-precision geometric information is the quantified three-dimensional coordinate information of the supplementary point;
  • the fourth precision geometric information of the supplementary point is the three-dimensional coordinate information lost during the quantization process of the supplementary point.
  • the quantization module 801 includes:
  • the eighth acquisition unit is used to quantize each vertex in the target three-dimensional grid according to the quantization parameter of each component, and obtain the first precision geometric information
  • a ninth acquisition unit is configured to acquire second-precision geometric information based on the first-precision geometric information and the quantized parameters of each component.
  • This device embodiment corresponds to the above-mentioned encoding method embodiment.
  • Each implementation process and implementation manner of the above-mentioned method embodiment can be applied to this device embodiment, and can achieve the same technical effect.
  • Embodiments of the present application also provide an encoding device, including a processor and a communication interface, wherein the processor is used to quantify the geometric information of the target three-dimensional grid and obtain the first precision geometric information and the second precision geometric information; Obtain the first geometric diagram and placeholder diagram of the first precision geometric information; obtain the second precision geometric information including the second precision geometric information and the first precision geometric information according to the second precision geometric information and the first geometric diagram.
  • the first precision geometric information is the geometric information after quantization of the target three-dimensional grid
  • the second precision geometric information is the geometric information lost during the quantization process of the target three-dimensional grid
  • the encoding device 900 includes: a processor 901, a network interface 902, and a memory 903.
  • the network interface 902 is, for example, a common public radio interface (CPRI).
  • CPRI common public radio interface
  • the encoding device 900 in the embodiment of the present application also includes: stored in the memory 903 and available on the processor For the instructions or programs running on 901, the processor 901 calls the instructions or programs in the memory 903 to execute the method of executing each module shown in Figure 8, and achieves the same technical effect. To avoid repetition, it will not be described again here.
  • this embodiment of the present application also provides a decoding method, including:
  • Step 1001 The decoder decomposes the obtained code stream of the target three-dimensional grid, and obtains a placeholder map and a second geometric map containing second-precision geometric information and first-precision geometric information;
  • Step 1002 The decoder obtains the second precision geometric information and the first geometric map according to the second geometric map;
  • Step 1003 The decoder obtains first precision geometric information based on the first geometric map and the placeholder map;
  • Step 1004 The decoder performs inverse quantization based on the second precision geometric information and the first precision geometric information to obtain the geometric information of the target three-dimensional grid;
  • the first precision geometric information is the geometric information after quantization of the target three-dimensional grid
  • the second precision geometric information is the geometric information lost during the quantization process of the target three-dimensional grid
  • the specific implementation of decomposing the obtained code stream of the target three-dimensional grid and obtaining the placeholder map and the second geometric map containing the second precision geometric information and the first precision geometric information is:
  • the decoding end obtains a target sub-code stream according to the obtained code stream of the target three-dimensional grid.
  • the target sub-code stream includes: a slice information sub-stream, a placeholder map sub-stream and a geometric map sub-stream;
  • the decoding end obtains a placeholder map and a second geometric map including second precision geometric information and first precision geometric information according to the target sub-stream.
  • the specific implementation of obtaining the second precision geometric information and the first geometric diagram according to the second geometric diagram is:
  • the decoder obtains, in the second geometric graph, the arrangement value corresponding to each vertex in the second precision geometric information and the first geometric graph corresponding to the first precision geometric information;
  • the decoding end restores the three-dimensional coordinate values of the arrangement values corresponding to each vertex according to the arrangement order of the vertices to obtain second precision geometric information
  • the arrangement value corresponding to each vertex is arranged in the first area of the second geometric graph, and the first area is the two-dimensional projection distribution area of the geometric information of the target three-dimensional grid except for the first accuracy. The area remaining after the geometric information is projected onto the distribution area.
  • the implementation of obtaining the arrangement value corresponding to each vertex in the second precision geometric information in the second geometric graph includes one of the following:
  • the decoding end restores the arrangement value corresponding to each vertex in the second precision geometric information according to the pixel distribution position of the projection point corresponding to the vertex in the first precision geometric information in the first area in the second geometric diagram. ;
  • the decoder performs translation processing on the pixels of the projection points in the first area in the second geometric diagram, and obtains the arrangement value corresponding to each vertex in the second precision geometric information.
  • the translation process is performed on the pixels of the projection points in the first area in the second geometric diagram to obtain the arrangement value corresponding to each vertex in the second precision geometric information, including:
  • the decoding end scans the pixels of the projection points of the first geometric figure along the first direction row by row or column by column, and obtains the position index within the row or column where each pixel in each row or column is located;
  • the decoding end scans the pixels of the projection points in the first area in the second geometric figure row by row or column by column along the first direction, and scans each pixel in each row or column according to the first
  • the position index corresponding to the pixel of the projection point in a geometric figure is re-numbered within the row or column;
  • the renumbering here can be understood as: the pixels of the projection points corresponding to the vertices in the second precision geometric information are closely adjacent in the first area, and the pixels of the projection points corresponding to the actual vertices are There are certain intervals.
  • the scanning renumbering mentioned in the embodiment of this application means that the pixels with gaps in the same row or the same column are renumbered in a pixel-adjacent manner. For example, in a certain row, there are 5 pixels corresponding to each other. , the numbers from left to right are 1, 2, 3, 4, and 5 respectively. After scanning from left to right, the five pixels are numbered again according to the pixel distribution of the projection point of the first geometric figure. This The new numbers of the 5 pixels from left to right are: 1, 3, 5, 7, 8.
  • the decoding end will obtain the arrangement value corresponding to each vertex in the second precision geometric information from the first area in the second geometric graph according to the scanning order;
  • the first direction is a horizontal direction or a vertical direction.
  • the pixels of the projection points of the first geometric figure are not pixel translated during the encoding process, during decoding, it is necessary to restore the vertices in another pixel translated part according to the first geometric figure without pixel translation.
  • the position of the pixel of the corresponding projection point is used to ensure the accurate recovery of the vertices in the second precision geometric information.
  • the decoder obtains first precision geometric information based on the first geometric map and the placeholder map, including:
  • the decoding end obtains two-dimensional image information based on the first geometric diagram and the placeholder diagram;
  • the decoding end obtains a two-dimensional slice according to the two-dimensional image information
  • the decoding end performs three-dimensional back-projection on the two-dimensional slice according to the slice information corresponding to the slice information sub-stream to obtain the three-dimensional slice;
  • the decoder acquires first precision geometric information based on the three-dimensional slice.
  • the implementation method of performing inverse quantization based on the second precision geometric information and the first precision geometric information to obtain the geometric information of the target three-dimensional grid is:
  • the decoder determines the coordinates of each vertex in the first precision geometry information based on the first precision geometry information and the quantization parameter of each component;
  • the decoder determines the target three-dimensional mesh based on the coordinates of each vertex in the first precision geometry information and the second precision geometry information.
  • the method further includes:
  • the geometric map of the supplementary point determine the first original slice corresponding to the third precision geometric information of the supplementary point and the second original slice corresponding to the fourth precision geometric information of the supplementary point;
  • the decoding end determines the information of the supplementary point based on the first original slice and the second original slice;
  • the information of the supplementary points is the information of points generated during the quantification process that require additional processing.
  • the supplementary point information includes at least one of the following:
  • Third-precision geometric information of the supplementary point where the third-precision geometric information is the quantified three-dimensional coordinate information of the supplementary point;
  • the fourth precision geometric information of the supplementary point is the three-dimensional coordinate information lost during the quantization process of the supplementary point.
  • the geometric information of the supplementary points is divided into low-precision parts and high-precision parts and is decoded separately.
  • the geometric map of the supplementary points is decompressed.
  • Various decompression methods can be used. Among them, one method is to decode the geometric map through run-length decoding, entropy decoding, etc., and the other method is to extract the supplementary point low-precision raw patch and the supplementary point high-precision raw patch from the low-precision geometric map.
  • the low-precision geometric information of the supplementary points is obtained from the low-precision raw patch of the supplementary points in a specific order
  • the high-precision geometric information of the supplementary points is obtained from the high-precision raw patch of the supplementary points in a specific order;
  • the specific order is decoding
  • the end obtains it by parsing the code stream, that is, the encoding end uses which order to generate the supplementary point low-precision raw patch and the supplementary point high-precision raw patch, which will be informed to the decoder through the code stream.
  • the implementation method of determining the target three-dimensional mesh based on the coordinates of each vertex in the first precision geometric information and the second precision geometric information is:
  • the decoding end determines the target three-dimensional mesh based on the information of the supplementary points, the second precision geometric information, and the coordinates of each vertex in the first precision geometric information.
  • the video-based three-dimensional grid geometric information decoding process in the embodiment of the present application includes: decomposing the code stream into a patch information sub-stream, a placeholder image code stream and a geometric image code stream; and then decoding these three respectively.
  • Path code stream obtain patch information, placeholder map and geometric map; finally, use patch information, placeholder map and geometric map to reconstruct geometric information.
  • the most critical thing is the specific process of geometric information reconstruction:
  • Step S111 obtain 2D patch
  • obtaining a 2D patch refers to using the patch information to segment the placeholder information and depth information of the 2D patch from the placeholder map and geometric map.
  • the patch information contains the position and size of each 2D patch's bounding box in the placeholder map and low-precision geometric map.
  • the placeholder information and the 2D patch's placeholder information can be directly obtained using the patch information, placeholder map, and low-precision geometric map.
  • Low precision geometric information For high-precision geometric information, the vertex distribution of the low-precision geometric graph is used to correspond the high-precision geometric information arrangement values in the high-precision geometric information area to the vertices of the low-precision geometric graph, and then xyz is separated from them according to the preset arrangement rules used by the encoding end.
  • the low-precision raw patch and the high-precision raw patch of the supplementary points can be obtained by directly decoding the low-precision raw patch and the high-precision geometric information of the supplementary point.
  • Step S112 reconstruct the 3D patch
  • reconstructing a 3D patch refers to using the placeholder information and low-precision geometric information in the 2D patch to reconstruct the vertices in the 2D patch into a low-precision 3D patch.
  • the occupancy information of the 2D patch contains the position of the vertex relative to the coordinate origin in the local coordinate system of the patch projection plane, and the depth information contains the normal direction of the vertex in the projection plane. depth value on. Therefore, the 2D patch can be reconstructed into a low-precision 3D patch in the local coordinate system using the occupancy information and depth information.
  • Step S113 reconstruct the low-precision geometric model
  • reconstructing a low-precision geometric model refers to using the reconstructed low-precision 3D patch to reconstruct the entire low-precision three-dimensional geometric model.
  • the patch information contains the conversion relationship of the 3D patch from the local coordinate system to the global coordinate system of the three-dimensional geometric model. Using the coordinate conversion relationship to convert all 3D patches to the global coordinate system, a low-precision three-dimensional geometric model is obtained.
  • the geometric information in the low-precision raw patch is directly used to obtain the low-precision coordinate values of the supplementary points in the global coordinate system, thereby obtaining a complete low-precision three-dimensional geometric model.
  • Step S114 reconstruct the high-precision geometric model
  • reconstructing a high-precision geometric model refers to the process of using high-precision geometric information to reconstruct a high-precision geometric model based on a low-precision geometric model.
  • high-precision geometric information and low-precision geometric information are mapped, and the high-precision three-dimensional coordinates of the vertex can be reconstructed based on the high-precision geometric information and low-precision geometric information of the vertex.
  • the calculation process of high-precision three-dimensional coordinates (x r , y r , z r ) is as shown in Formula 25 to Formula 27:
  • the f 3 function is a reconstruction function.
  • the calculation process of the reconstruction function corresponds to the calculation process of the quantization function at the encoding end, and there are many ways to implement it. If the f 1 function adopts the implementation method of Formula 7 to Formula 12, then the reconstruction function is implemented as shown in Formula 28 to Formula 30:
  • the video-based three-dimensional grid geometric information decoding framework of the embodiment of this application is shown in Figure 12.
  • the overall decoding process is:
  • the code stream is decomposed into a patch information sub-stream, a placeholder map sub-stream, and a geometry map sub-stream, and decoded separately; the low-precision grid can be reconstructed using the low-precision parts of the placeholder map and geometry map. Geometric information, and then use the high-precision parts of the placeholder map and the geometric map to reconstruct the geometric information of the high-precision grid; finally, use the reconstructed geometric information and information such as connection relationships obtained by other encoding and decoding methods to reconstruct the grid.
  • the embodiment of the present application is a method embodiment of the opposite end corresponding to the embodiment of the above encoding method.
  • the decoding process is the inverse process of encoding. All the above implementation methods on the encoding side are applicable to the embodiment of the decoding end. The same technical effect can also be achieved, which will not be described again here.
  • this embodiment of the present application also provides a decoding device 1300, which includes:
  • the third acquisition module 1301 is used to decompose the acquired code stream of the target three-dimensional grid, and acquire the placeholder map and the second geometric map containing the second precision geometric information and the first precision geometric information;
  • the fourth acquisition module 1302 is used to acquire second precision geometric information and the first geometric graph according to the second geometric graph;
  • the fifth acquisition module 1303 is used to acquire first precision geometric information according to the first geometric diagram and the placeholder diagram;
  • the sixth acquisition module 1304 is used to perform inverse quantization according to the second precision geometric information and the first precision geometric information to obtain the geometric information of the target three-dimensional grid;
  • the first precision geometric information is the geometric information after quantization of the target three-dimensional grid
  • the second precision geometric information is the geometric information lost during the quantization process of the target three-dimensional grid
  • the third acquisition module 1301 includes:
  • the tenth acquisition unit is used to acquire a target sub-code stream according to the obtained code stream.
  • the target sub-code stream includes: a slice information sub-stream, a placeholder map sub-stream and a geometric map sub-stream;
  • An eleventh acquisition unit configured to acquire a placeholder image and a second geometric image containing second-precision geometric information and first-precision geometric information according to the target sub-code stream.
  • the fourth acquisition module 1302 includes:
  • a twelfth acquisition unit configured to obtain, in the second geometric graph, the arrangement value corresponding to each vertex in the second precision geometric information and the first geometric graph corresponding to the first precision geometric information;
  • the thirteenth acquisition unit is used to restore the three-dimensional coordinate values of the arrangement values corresponding to each vertex according to the arrangement order of the vertices to obtain second precision geometric information;
  • the arrangement value corresponding to each vertex is arranged in the first area of the second geometric graph, and the first area is the two-dimensional projection distribution area of the geometric information of the target three-dimensional grid except for the first accuracy. The area remaining after the geometric information is projected onto the distribution area.
  • the twelfth acquisition unit is used to implement one of the following:
  • Translation processing is performed on the pixels of the projection points in the first area in the second geometric diagram to obtain the arrangement value corresponding to each vertex in the second precision geometric information.
  • the implementation method of performing translation processing on the pixels of the projection points in the first area in the second geometric diagram to obtain the arrangement value corresponding to each vertex in the second precision geometric information is:
  • the pixels of the projection points in the first area in the second geometric figure are scanned row by row or column by column along the first direction, and each pixel in each row or column is scanned according to the same pattern as in the first geometric figure.
  • the position index corresponding to the pixel of the projection point is renumbered within the row or column;
  • the arrangement value corresponding to each vertex in the second precision geometric information will be obtained from the first area in the second geometric graph;
  • the first direction is a horizontal direction or a vertical direction.
  • the fifth acquisition module 1303 includes:
  • a fourteenth acquisition unit configured to acquire two-dimensional image information based on the first geometric diagram and the placeholder map
  • the fifteenth acquisition unit is used to acquire two-dimensional slices according to the two-dimensional image information
  • the sixteenth acquisition unit is used to perform three-dimensional back-projection of the two-dimensional slice according to the slice information corresponding to the slice information sub-stream, and obtain the three-dimensional slice;
  • the seventeenth acquisition unit is used to acquire first precision geometric information based on the three-dimensional slice.
  • the sixth acquisition module 1304 includes:
  • a first determination unit configured to determine the coordinates of each vertex in the first precision geometry information according to the first precision geometry information and the quantization parameter of each component
  • a second determination unit configured to determine the target three-dimensional mesh based on the coordinates of each vertex in the first precision geometric information and the second precision geometric information.
  • the device also includes:
  • the seventh acquisition module is used to obtain the geometric diagram of the supplementary point
  • a first determination module configured to determine the first original slice corresponding to the third precision geometric information of the supplementary point and the second original slice corresponding to the fourth precision geometric information of the supplementary point according to the geometric map of the supplementary point.
  • a second determination module configured to determine the information of the supplementary point according to the first original slice and the second original slice
  • the information of the supplementary points is the information of points generated during the quantification process that require additional processing.
  • the second determining unit is used to:
  • the target three-dimensional mesh is determined based on the information of the supplementary points, the second precision geometric information and the coordinates of each vertex in the first precision geometric information.
  • the supplementary point information includes at least one of the following:
  • Third-precision geometric information of the supplementary point where the third-precision geometric information is the quantified three-dimensional coordinate information of the supplementary point;
  • the fourth precision geometric information of the supplementary point is the three-dimensional coordinate information lost during the quantization process of the supplementary point.
  • this device embodiment is a device corresponding to the above-mentioned method. All implementation methods in the above-mentioned method embodiment are applicable to this device embodiment and can achieve the same technical effect, which will not be described again here.
  • the embodiment of the present application also provides a decoding device, including a processor, a memory, and storage on the memory. And can be a program or instruction that can be run on the processor.
  • a decoding device including a processor, a memory, and storage on the memory.
  • the program or instruction can be executed by the processor, the various processes of the above-mentioned decoding method embodiments can be achieved, and the same technical effect can be achieved. To avoid repetition, they will not be described again here. .
  • Embodiments of the present application also provide a readable storage medium.
  • Programs or instructions are stored on the computer-readable storage medium.
  • the program or instructions are executed by a processor, each process of the above-mentioned decoding method embodiment is implemented, and the same process can be achieved. To avoid repetition, the technical effects will not be repeated here.
  • the computer-readable storage medium is such as read-only memory (Read-Only Memory, ROM), random access memory (Random Access Memory, RAM), magnetic disk or optical disk, etc.
  • Embodiments of the present application also provide a decoding device, including a processor and a communication interface, wherein the processor is used to decompose the obtained code stream of the target three-dimensional grid, obtain the placeholder map and the second precision geometric information. and a second geometric diagram of the first precision geometric information; according to the second geometric diagram, obtain the second precision geometric information and the first geometric diagram; according to the first geometric diagram and the placeholder map, obtain the first precision geometric information ; Perform inverse quantization according to the second precision geometric information and the first precision geometric information to obtain the geometric information of the target three-dimensional grid;
  • the first precision geometric information is the geometric information after quantization of the target three-dimensional grid
  • the second precision geometric information is the geometric information lost during the quantization process of the target three-dimensional grid
  • This decoding device embodiment corresponds to the above-mentioned decoding method embodiment.
  • Each implementation process and implementation manner of the above-mentioned method embodiment can be applied to this decoding device embodiment, and can achieve the same technical effect.
  • the embodiment of the present application also provides a decoding device.
  • the structure of the decoding device is shown in Figure 9, which will not be described again here.
  • the decoding device in the embodiment of the present application also includes: instructions or programs stored in the memory and executable on the processor.
  • the processor calls the instructions or programs in the memory to execute the method executed by each module shown in Figure 13, and To achieve the same technical effect, to avoid repetition, we will not repeat them here.
  • Embodiments of the present application also provide a readable storage medium.
  • Programs or instructions are stored on the readable storage medium.
  • the program or instructions are executed by a processor, each process of the above decoding method embodiment is implemented, and the same process can be achieved. To avoid repetition, the technical effects will not be repeated here.
  • the processor is the processor in the decoding device described in the above embodiment.
  • the readable storage medium includes computer readable storage media, such as computer read-only memory ROM, random access memory RAM, magnetic disk or optical disk, etc.
  • this embodiment of the present application also provides a communication device 1400, which includes a processor 1401 and a memory 1402.
  • the memory 1402 stores programs or instructions that can be run on the processor 1401, such as , when the communication device 1400 is a coding device, when the program or instruction is executed by the processor 1401, each step of the above coding method embodiment is implemented, and the same technical effect can be achieved.
  • the communication device 1400 is a decoding device, when the program or instruction is executed by the processor 1401, each step of the above decoding method embodiment is implemented and the same technical effect can be achieved. To avoid duplication, the details are not repeated here.
  • An embodiment of the present application further provides a chip.
  • the chip includes a processor and a communication interface.
  • the communication interface is coupled to the processor.
  • the processor is used to run programs or instructions to implement the above encoding method or decoding method.
  • Each process in the example can achieve the same technical effect. To avoid repetition, we will not repeat it here.
  • chips mentioned in the embodiments of this application may also be called system-on-chip, system-on-a-chip, system-on-chip or system-on-chip, etc.
  • Embodiments of the present application further provide a computer program/program product.
  • the computer program/program product is stored in a storage medium.
  • the computer program/program product is executed by at least one processor to implement the above encoding method or decoding method.
  • Each process of the embodiment can achieve the same technical effect, so to avoid repetition, it will not be described again here.
  • Embodiments of the present application also provide a communication system, which at least includes: an encoding device and a decoding device.
  • the encoding device can be used to perform the steps of the encoding method as described above.
  • the decoding device can be used to perform the decoding method as described above. A step of. And can achieve the same technical effect. To avoid repetition, they will not be described again here.
  • the methods of the above embodiments can be implemented by means of software plus the necessary general hardware platform. Of course, it can also be implemented by hardware, but in many cases the former is better. implementation.
  • the technical solution of the present application can be embodied in the form of a computer software product that is essentially or contributes to related technologies.
  • the computer software product is stored in a storage medium (such as ROM/RAM, disk, CD), including several instructions to cause a terminal (which can be a mobile phone, computer, server, air conditioner, or network device, etc.) to execute the methods described in various embodiments of this application.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression, Expansion, Code Conversion, And Decoders (AREA)

Abstract

La présente demande concerne le domaine technique du codage et du décodage. Sont divulgués dans la présente demande un procédé de codage, un procédé de décodage, un appareil et un dispositif. Le procédé de codage comprend les étapes suivantes : une extrémité de codage quantifie des informations géométriques d'un maillage tridimensionnel cible, et acquiert des premières informations géométriques de précision et des secondes informations géométriques de précision ; l'extrémité de codage acquiert un premier graphe géométrique des premières informations géométriques de précision et une image de substitution ; selon les secondes informations géométriques de précision et le premier graphe géométrique, l'extrémité de codage acquiert un second graphe géométrique contenant les secondes informations géométriques de précision et les premières informations géométriques de précision ; et l'extrémité de codage code le second graphe géométrique et l'image de substitution, les premières informations géométriques de précision étant des informations géométriques obtenues après que le maillage tridimensionnel cible est quantifié, et les secondes informations géométriques de précision étant des informations géométriques perdues pendant le processus de quantification de maillage tridimensionnel cible.
PCT/CN2023/081641 2022-03-18 2023-03-15 Procédé de codage, procédé de décodage, appareil et dispositif WO2023174337A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202210272443.0 2022-03-18
CN202210272443.0A CN116800969A (zh) 2022-03-18 2022-03-18 编码、解码方法、装置及设备

Publications (1)

Publication Number Publication Date
WO2023174337A1 true WO2023174337A1 (fr) 2023-09-21

Family

ID=88022376

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2023/081641 WO2023174337A1 (fr) 2022-03-18 2023-03-15 Procédé de codage, procédé de décodage, appareil et dispositif

Country Status (2)

Country Link
CN (1) CN116800969A (fr)
WO (1) WO2023174337A1 (fr)

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2020145654A1 (fr) * 2019-01-09 2020-07-16 Samsung Electronics Co., Ltd. Sélection adaptative de précision de carte d'occupation
US20210090301A1 (en) * 2019-09-24 2021-03-25 Apple Inc. Three-Dimensional Mesh Compression Using a Video Encoder
CN112565764A (zh) * 2020-12-03 2021-03-26 西安电子科技大学 一种点云几何信息帧间编码及解码方法
CN112637608A (zh) * 2020-12-18 2021-04-09 南京邮电大学 一种动态点云几何信息压缩率失真优化方法及装置
CN113424225A (zh) * 2019-01-11 2021-09-21 三星电子株式会社 三维数据编码方法和设备以及三维数据解码方法和设备
US20210407149A1 (en) * 2019-03-25 2021-12-30 Panasonic Intellectual Property Corporation Of America Three-dimensional data encoding method, three-dimensional data decoding method, three-dimensional data encoding device, and three-dimensional data decoding device

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2020145654A1 (fr) * 2019-01-09 2020-07-16 Samsung Electronics Co., Ltd. Sélection adaptative de précision de carte d'occupation
CN113424225A (zh) * 2019-01-11 2021-09-21 三星电子株式会社 三维数据编码方法和设备以及三维数据解码方法和设备
US20210407149A1 (en) * 2019-03-25 2021-12-30 Panasonic Intellectual Property Corporation Of America Three-dimensional data encoding method, three-dimensional data decoding method, three-dimensional data encoding device, and three-dimensional data decoding device
US20210090301A1 (en) * 2019-09-24 2021-03-25 Apple Inc. Three-Dimensional Mesh Compression Using a Video Encoder
CN112565764A (zh) * 2020-12-03 2021-03-26 西安电子科技大学 一种点云几何信息帧间编码及解码方法
CN112637608A (zh) * 2020-12-18 2021-04-09 南京邮电大学 一种动态点云几何信息压缩率失真优化方法及装置

Also Published As

Publication number Publication date
CN116800969A (zh) 2023-09-22

Similar Documents

Publication Publication Date Title
JP2023184727A (ja) 画像処理装置および方法
CN103927746B (zh) 一种三维网格序列的配准及压缩方法
CN108028945A (zh) 通过使用单例系数更新执行变换的设备和方法
KR20230056750A (ko) 메시 압축을 위한 패치 지퍼링
KR20220063254A (ko) 세계 시그널링 정보에 대한 비디오 기반 포인트 클라우드 압축 모델
US20220180567A1 (en) Method and apparatus for point cloud coding
KR20230060534A (ko) 2차원 정규화 평면 투사에 기초한 포인트 클라우드 인코딩 및 디코딩 방법과 장치
KR20200007733A (ko) Pcc 데이터의 효과적인 압축을 위한 부호화/복호화 방법 및 장치
WO2023174337A1 (fr) Procédé de codage, procédé de décodage, appareil et dispositif
WO2023181875A1 (fr) Dispositif et procédé de traitement d'informations
WO2023179705A1 (fr) Procédés et appareils de codage et décodage, et dispositifs
WO2023174336A1 (fr) Procédé et appareil de codage, procédé et appareil de décodage et dispositif
KR20230127324A (ko) 메시 압축을 위한 패치 시간 정렬의 디코딩
WO2023155794A1 (fr) Procédé de codage, procédé de décodage, appareil et dispositif
WO2023174334A1 (fr) Procédé et appareil de codage, procédé et appareil de décodage, et dispositif
WO2023155778A1 (fr) Procédé et appareil de codage et dispositif
WO2023193709A1 (fr) Procédé et appareil de codage, procédé et appareil de décodage et dispositif
WO2023197990A1 (fr) Procédé de codage, procédé de décodage et terminal
WO2023179706A1 (fr) Procédé de codage, procédé de décodage, et terminal
WO2023179710A1 (fr) Procédé de codage et terminal
US20220392114A1 (en) Method and apparatus for calculating distance based weighted average for point cloud coding
CN116843771A (zh) 编码方法、解码方法及终端
CN117178299A (zh) 将3d网格快速且紧凑地打包到2d地图中的卷积方法
JP2024512915A (ja) 非マニホールドメッシュにおける境界ループの検出のための方法、装置、およびコンピュータプログラム
JP2024509988A (ja) チャートベースのメッシュ圧縮の方法、装置およびコンピュータプログラム

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 23769839

Country of ref document: EP

Kind code of ref document: A1