WO2023285998A1 - Compression de nuage de points utilisant des réseaux d'occupation - Google Patents

Compression de nuage de points utilisant des réseaux d'occupation Download PDF

Info

Publication number
WO2023285998A1
WO2023285998A1 PCT/IB2022/056480 IB2022056480W WO2023285998A1 WO 2023285998 A1 WO2023285998 A1 WO 2023285998A1 IB 2022056480 W IB2022056480 W IB 2022056480W WO 2023285998 A1 WO2023285998 A1 WO 2023285998A1
Authority
WO
WIPO (PCT)
Prior art keywords
occupancy
networks
probability
bitstream
function
Prior art date
Application number
PCT/IB2022/056480
Other languages
English (en)
Inventor
Danillo GRAZIOSI
Alexandre ZAGHETTO
Ali Tabatabai
Original Assignee
Sony Group Corporation
Sony Corporation Of America
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US17/828,326 external-priority patent/US20230013421A1/en
Application filed by Sony Group Corporation, Sony Corporation Of America filed Critical Sony Group Corporation
Publication of WO2023285998A1 publication Critical patent/WO2023285998A1/fr

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/90Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using coding techniques not provided for in groups H04N19/10-H04N19/85, e.g. fractals
    • H04N19/91Entropy coding, e.g. variable length coding [VLC] or arithmetic coding
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T9/00Image coding
    • G06T9/001Model-based coding, e.g. wire frame
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T9/00Image coding
    • G06T9/002Image coding using neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T9/00Image coding
    • G06T9/40Tree coding, e.g. quadtree, octree
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/90Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using coding techniques not provided for in groups H04N19/10-H04N19/85, e.g. fractals
    • H04N19/96Tree coding, e.g. quad-tree coding

Definitions

  • the present invention relates to three dimensional graphics. More specifically, the present invention relates to coding of three dimensional graphics.
  • Point clouds have been considered as a candidate format for transmission of 3D data, either captured by 3D scanners, LIDAR sensors, or used in popular applications such as VR/AR.
  • Point clouds are a set of points in 3D space.
  • each point usually have associated attributes, such as color (R, G, B) or even reflectance and temporal timestamps (e.g., in LIDAR images).
  • attributes such as color (R, G, B) or even reflectance and temporal timestamps (e.g., in LIDAR images).
  • devices capture point clouds in the order of thousands or even millions of points.
  • Occupancy networks implicitly represent the 3D surface as the continuous decision boundary of a deep neural network classifier.
  • the representation encodes a description of the 3D output at infinite resolution.
  • Occupancy networks enable efficient and flexible point cloud compression.
  • occupancy networks are able to handle points, meshes, or projected images of 3D objects, making them very flexible in terms of input signal representation.
  • the probability of occupancy of positions is estimated using occupancy networks instead of sparse convolutional neural networks.
  • a compression implementation using occupancy network enables scalability with infinite reconstruction resolution.
  • a method programmed in a non-transitory memory of a device comprises receiving a bitstream at one or more occupancy networks, determining a probability of a position in the bitstream being occupied with the one or more occupancy networks and generating a function based on the probability of positions being occupied.
  • the bitstream comprises voxels, points, meshes, or projected images of 3D objects.
  • the bitstream comprises one or more samples of a 3D space to be used to generate a 3D object with the one or more occupancy networks.
  • the probability is determined using machine learning to implement implicit neural functions.
  • the one or more occupancy networks implicitly represent 3D surfaces using a continuous decision boundary based on a deep neural network classifier, and decides based on a threshold whether data belongs inside or outside a 3D structure.
  • the probability is determined based neighboring position classification information.
  • the probability is used by an entropy encoder to define a code length of an occupancy code of points in 3D space.
  • the one or more occupancy networks learn the function to recover a specific shape based on a sparse input.
  • the function represents a set of classes, and an object is recovered based on an input. A size of the function is smaller than the bitstream.
  • an apparatus comprises a non-transitory memory for storing an application, the application for: receiving a bitstream at one or more occupancy networks, determining a probability of a position in the bitstream being occupied with the one or more occupancy networks and generating a function based on the probability of positions being occupied and a processor coupled to the memory, the processor configured for processing the application.
  • the bitstream comprises voxels, points, meshes, or projected images of 3D objects.
  • the bitstream comprises one or more samples of a 3D space to be used to generate a 3D object with the one or more occupancy networks.
  • the probability is determined using machine learning to implement implicit neural functions.
  • the one or more occupancy networks implicitly represent 3D surfaces using a continuous decision boundary based on a deep neural network classifier, and decides based on a threshold whether data belongs inside or outside a 3D structure.
  • the probability is determined based neighboring position classification information.
  • the probability is used by an entropy encoder to define a code length of an occupancy code of points in 3D space.
  • the one or more occupancy networks learn the function to recover a specific shape based on a sparse input.
  • the function represents a set of classes, and an object is recovered based on an input. A size of the function is smaller than the bitstream.
  • a system comprises an encoder configured for: receiving a bitstream at one or more occupancy networks, determining a probability of a position in the bitstream being occupied with the one or more occupancy networks and generating a function based on the probability of positions being occupied and a decoder configured for: recovering an object based on the function and an input.
  • the bitstream comprises voxels, points, meshes, or projected images of 3D objects.
  • the bitstream comprises one or more samples of a 3D space to be used to generate a 3D object with the one or more occupancy networks.
  • the probability is determined using machine learning to implement implicit neural functions.
  • the one or more occupancy networks implicitly represent 3D surfaces using a continuous decision boundary based on a deep neural network classifier, and decides based on a threshold whether data belongs inside or outside a 3D structure.
  • the probability is determined based neighboring position classification information.
  • the probability is used by to define a code length of an occupancy code of points in 3D space. A size of the function is smaller than the bitstream.
  • Figure 1 illustrates a diagram of occupancy networks according to some embodiments.
  • Figure 2 illustrates a diagram of point cloud compression using occupancy networks according to some embodiments.
  • Figure 3 illustrates a flowchart of a method of implementing point cloud compression using occupancy networks according to some embodiments.
  • Figure 4 illustrates a block diagram of an exemplary computing device configured to implement the method of implementing point cloud compression using occupancy networks according to some embodiments.
  • a point cloud compression scheme uses occupancy networks as an implicit representation of the points.
  • the implicit neural functions define an occupancy probability for points in 3D space. This probability is then used by an entropy encoder to define the code length of the occupancy code of points in 3D space.
  • the MPEG is currently concluding two standards for Point Cloud Compression (PCC).
  • Point clouds are used to represent three-dimensional scenes and objects, and are composed by volumetric elements (voxels) described by their position in 3D space and attributes such as color, reflectance, material, transparency, time stamp and others.
  • the planned outcome of the standardization activity is the Geometry-based Point Cloud Compression (G-PCC) and the
  • V-PCC Video-based Point Cloud Compression
  • a sparse convolutional network exploits the spatial dependency between neighbors to estimate the occupancy of voxels by means of probabilities used for entropy coding or binary classification, depending if one wants to perform lossless or lossy compression, respectively.
  • an occupancy network is described, which performs the same task by assigning to every location/position an occupancy probability between 0 and 1.
  • the embodiments described herein are more general since the method is able to be applied to points, meshes or projected images of 3D objects, and is not limited to a voxel-based representation. Scalability is able to be provided by voxebzing the volumetric space at an initial resolution and evaluating the occupancy network for all points in a grid.
  • Occupancy networks have several applications. Their usage is a scalable, and a more generic point cloud compression scheme is novel. Occupancy networks enable efficient and flexible point cloud compression. Although based on occupancy estimation, sparse convolutional neural networks are typically limited to voxel-based representation. In addition to the voxel-based representation, occupancy networks are able to deal with points, meshes, or projected images of 3D objects, making them more flexible in terms of input signal representation. The probability of occupancy of positions is estimated using occupancy networks instead of sparse convolutional neural networks.
  • the occupancy network implicitly represents 3D surfaces using a continuous decision boundary based on a deep neural network classifier, and decides based on a boundary (threshold) whether a point belongs inside or outside a 3D structure (e.g., mesh).
  • the occupancy network repetitively decides whether a point belongs inside or outside and by doing this, the occupancy network defines the surface of the volumetric representation.
  • the occupancy network is used to determine the probability of a position in space being occupied.
  • the occupancy network is able to be used to assist in compression as well.
  • Figure 1 illustrates a diagram of occupancy networks according to some embodiments.
  • Occupancy networks learn general characteristics of classes of objects.
  • occupancy maps learn a function that is able to recover a specific shape based on a sparse input.
  • an occupancy network 100 is able to represent chairs and tables.
  • the occupancy network 100 is then able to receive a sparse representation of an object 102 (e.g., chair) as an input to the function and produce a representation 104 of the object in its original form with a desired precision (e.g., a more detailed object).
  • a function receives a dataset of sparse points, and the function outputs an object similar to one of the classes the occupancy network function had learned.
  • the precision of the function is not mathematically limited.
  • point cloud compression In addition to recovering the object from the sparse point cloud, efficient and flexible point cloud compression is able to be performed.
  • the method is flexible because in addition to points, other forms of input are able to be used such as voxels, 2D images (projections) and meshes.
  • the input data is able to be compressed regardless of the input form using occupancy estimation.
  • FIG. 2 illustrates a diagram of point cloud compression using occupancy networks according to some embodiments.
  • a bitstream 200 is received at the occupancy networks 202.
  • the bitstream 200 is able to be voxels, points, projections, meshes or others.
  • the occupancy networks 202 are one or more neural networks able to obtain the implicit representation of a 3D object.
  • the bitstream 200 is able to comprise network coefficients and/or random samples of a 3D space instead of a point cloud, and the occupancy networks 202 are able to generate a 3D object based on the network coefficients and random samples to check occupied positions in 3D space.
  • Occupancy networks 202 progressively divide the space into smaller and smaller regions/divisions. For each division, the probability of positions being occupied is calculated.
  • the upper left region of the first block 210 has a 0.94 (or 94%) probability of having an occupied position.
  • the first block 210 is able to be divided further into the more refined second block 212 which has smaller divisions.
  • the blocks are able to be divided many more times, for example, to an n th block 214 which has the smallest divisions, in the example.
  • the probabilities of a position being occupied are shown for all four blocks, although probabilities below a threshold (e.g., 0.50) indicate the position is not likely occupied.
  • a threshold e.g. 0.50
  • the block is able to be divided theoretically infinitely (limited only by processing power and memory). By being able to divide the blocks many times, a system is able to be very scalable. For example, a system is able to output point clouds with different degrees of detail (e.g., coarse to fine detail).
  • the occupancy network assigns to every location an occupancy probability between 0 and 1.
  • An occupancy network is used but not necessarily the full capacity of a neural network.
  • the surface of an object is generated based on the observation of that object (input conditioning). Furthering the example, a full, continuous surface of an object may not be generated, where only a certain level of detail is included.
  • Scalability is provided in by voxebzing the volumetric space an initial resolution and evaluating the occupancy network for all points in the grid. Grid points p are marked as occupied if the evaluated value of the function at the point is bigger or equal to some threshold, which is given as a hyperparameter. In some embodiments, all voxel/points are marked as active, if at least two adjacent grid points have differing occupancy predictions.
  • the occupancy network is used to compress point clouds.
  • the implicit 3D surface representation not encoding the points themselves, rather a function is encoded.
  • the function is encoded.
  • the function is able to represent a set of classes, and then an object is able to be recovered based on an input.
  • the object itself is not encoded; rather, the function is encoded. This is also referred to as an implicit 3D surface representation.
  • different aspects of an object are able to have different amounts of refinement (e.g., coarse to fine).
  • Figure 3 illustrates a flowchart of a method of implementing point cloud compression using occupancy networks according to some embodiments.
  • a bitstream is received at occupancy networks.
  • the bitstream is able to include voxels, points, meshes, projected images of 3D objects or other data.
  • the probability of a position being occupied in the bitstream is determined.
  • the probability is determined in any manner such as based on machine learning (e.g., the implicit neural functions define an occupancy probability for points in 3D space) and/or classifications of the current object and previously classified objects.
  • the occupancy network implicitly represents 3D surfaces using a continuous decision boundary based on a deep neural network classifier, and decides based on a boundary (threshold) whether a point belongs inside or outside a 3D structure (e.g., mesh).
  • the occupancy network repetitively decides whether each point (or other data) belongs inside or outside and by doing this, the occupancy network defines the surface of the volumetric representation.
  • the probability is also able to be determined based on current information (e.g., a position that has two neighboring positions with a high probability of being occupied is also able to have a high probability of being occupied).
  • the probability is then used by an entropy encoder to define the code length of the occupancy code of points in 3D space.
  • a function is generated based on the probability of positions being occupied.
  • occupancy networks/maps learn a function that is able to recover a specific shape based on a sparse input.
  • the occupancy network is then able to receive a sparse representation of an object(e.g., chair) as an input to the function and produce a representation of the object in its original form with a desired precision (e.g., a more detailed object).
  • a function receives a dataset of sparse points, and the function outputs an object similar to one of the classes the occupancy network function had learned.
  • the function is able to represent a set of classes, and then an object is able to be recovered based on an input.
  • the object itself is not encoded; rather, the function is encoded. This is also referred to as an implicit 3D surface representation. Since the function does not include all of the data points, the representation is a compressed version of the input bitstream.
  • fewer or additional steps are implemented. In some embodiments, the order of the steps is modified.
  • Figure 4 illustrates a block diagram of an exemplary computing device configured to implement the method of implementing point cloud compression using occupancy networks according to some embodiments.
  • the computing device 400 is able to be used to acquire, store, compute, process, communicate and/or display information such as images and videos including 3D content.
  • the computing device 400 is able to implement any of the encoding/decoding aspects.
  • a hardware structure suitable for implementing the computing device 400 includes a network interface 402, a memory 404, a processor 406, I/O device(s) 408, a bus 410 and a storage device 412.
  • the choice of processor is not critical as long as a suitable processor with sufficient speed is chosen.
  • the memory 404 is able to be any conventional computer memory known in the art.
  • the storage device 412 is able to include a hard drive, CDROM, CDRW, DVD, DVDRW, High Definition disc/drive, ultra-HD drive, flash memory card or any other storage device.
  • the computing device 400 is able to include one or more network interfaces 402. An example of a network interface includes a network card connected to an Ethernet or other type of LAN.
  • the I/O device(s) 408 are able to include one or more of the following: keyboard, mouse, monitor, screen, printer, modem, touchscreen, button interface and other devices.
  • Compression application(s) 430 used to implement the compression implementation are likely to be stored in the storage device 412 and memory 404 and processed as applications are typically processed. More or fewer components shown in Figure 4 are able to be included in the computing device 400.
  • compression hardware 420 is included.
  • the computing device 400 in Figure 4 includes applications 430 and hardware 420 for the compression method, the compression method is able to be implemented on a computing device in hardware, firmware, software or any combination thereof.
  • the compression applications 430 are programmed in a memory and executed using a processor.
  • the compression hardware 420 is programmed hardware logic including gates specifically designed to implement the compression method.
  • the compression application(s) 430 include several applications and/or modules.
  • modules include one or more sub-modules as well. In some embodiments, fewer or additional modules are able to be included.
  • suitable computing devices include a personal computer, a laptop computer, a computer workstation, a server, a mainframe computer, a handheld computer, a personal digital assistant, a cellular/mobile telephone, a smart appliance, a gaming console, a digital camera, a digital camcorder, a camera phone, a smart phone, a portable music player, a tablet computer, a mobile device, a video player, a video disc writer/player (e.g., DVD writer/player, high definition disc writer/player, ultra high definition disc writer/player), a television, a home entertainment system, an augmented reality device, a virtual reality device, smart jewelry (e.g., smart watch), a vehicle (e.g., a self-driving vehicle) or any other suitable computing device.
  • a personal computer e.g., a laptop computer, a computer workstation, a server, a mainframe computer, a handheld computer, a personal digital assistant, a cellular/mobile telephone, a smart appliance, a gaming console
  • a device acquires or receives 3D content (e.g., point cloud content).
  • 3D content e.g., point cloud content.
  • the compression method is able to be implemented with user assistance or automatically without user involvement.
  • the compression method enables more efficient and more accurate 3D content encoding compared to previous implementations.
  • the compression method is highly scalable as well.
  • a method programmed in a non-transitory memory of a device comprising: receiving a bitstream at one or more occupancy networks; determining a probability of a position in the bitstream being occupied with the one or more occupancy networks; and generating a function based on the probability of positions being occupied.
  • the bitstream comprises voxels, points, meshes, or projected images of 3D objects.
  • bitstream comprises one or more samples of a 3D space to be used to generate a 3D object with the one or more occupancy networks.
  • An apparatus comprising: a non-transitory memory for storing an application, the application for: receiving a bitstream at one or more occupancy networks; determining a probability of a position in the bitstream being occupied with the one or more occupancy networks; and generating a function based on the probability of positions being occupied; and a processor coupled to the memory, the processor configured for processing the application.
  • the bitstream comprises voxels, points, meshes, or projected images of 3D objects.
  • bitstream comprises one or more samples of a 3D space to be used to generate a 3D object with the one or more occupancy networks.
  • the one or more occupancy networks implicitly represent 3D surfaces using a continuous decision boundary based on a deep neural network classifier, and decides based on a threshold whether data belongs inside or outside a 3D structure.
  • the probability is determined based neighboring position classification information.
  • the probability is used by an entropy encoder to define a code length of an occupancy code of points in 3D space.
  • a system comprising: an encoder configured for: receiving a bitstream at one or more occupancy networks; determining a probability of a position in the bitstream being occupied with the one or more occupancy networks; and generating a function based on the probability of positions being occupied; and a decoder configured for: recovering an object based on the function and an input.
  • bitstream comprises voxels, points, meshes, or projected images of 3D objects.
  • bitstream comprises one or more samples of a 3D space to be used to generate a 3D object with the one or more occupancy networks.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Signal Processing (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • Image Analysis (AREA)

Abstract

Selon l'invention, des réseaux d'occupation permettent une compression de nuage de points efficace et flexible. En plus de la représentation par voxels, les réseaux d'occupation peuvent gérer des points, des maillages, ou des images projetées d'objets 3D, ce qui les rend très souples pour la représentation de signal d'entrée. La probabilité d'occupation de positions est estimée en utilisant des réseaux d'occupation au lieu de réseaux neuronaux convolutifs clairsemés. Une mise en œuvre de compression utilisant un réseau d'occupation permet une extensibilité avec une résolution de reconstruction infinie.
PCT/IB2022/056480 2021-07-14 2022-07-14 Compression de nuage de points utilisant des réseaux d'occupation WO2023285998A1 (fr)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US202163221552P 2021-07-14 2021-07-14
US63/221,552 2021-07-14
US17/828,326 US20230013421A1 (en) 2021-07-14 2022-05-31 Point cloud compression using occupancy networks
US17/828,326 2022-05-31

Publications (1)

Publication Number Publication Date
WO2023285998A1 true WO2023285998A1 (fr) 2023-01-19

Family

ID=82748778

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/IB2022/056480 WO2023285998A1 (fr) 2021-07-14 2022-07-14 Compression de nuage de points utilisant des réseaux d'occupation

Country Status (1)

Country Link
WO (1) WO2023285998A1 (fr)

Non-Patent Citations (6)

* Cited by examiner, † Cited by third party
Title
EMRE CAN KAYA ET AL: "Neural Network Modeling of Probabilities for Coding the Octree Representation of Point Clouds", ARXIV.ORG, CORNELL UNIVERSITY LIBRARY, 201 OLIN LIBRARY CORNELL UNIVERSITY ITHACA, NY 14853, 11 June 2021 (2021-06-11), XP081988278 *
MENG JIA ET AL: "Learning Occupancy Function from Point Clouds for Surface Reconstruction", ARXIV.ORG, CORNELL UNIVERSITY LIBRARY, 201 OLIN LIBRARY CORNELL UNIVERSITY ITHACA, NY 14853, 22 October 2020 (2020-10-22), XP081793028 *
MESCHEDER LARS ET AL: "Occupancy Networks: Learning 3D Reconstruction in Function Space", 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), IEEE, 15 June 2019 (2019-06-15), pages 4455 - 4465, XP033686741, DOI: 10.1109/CVPR.2019.00459 *
WANG J ET AL: "[G-PCC EE13.54] A Geometry Compression Framework for AI-based PCC via Sparse Convolution", no. m57453, 6 July 2021 (2021-07-06), XP030297111, Retrieved from the Internet <URL:https://dms.mpeg.expert/doc_end_user/documents/135_OnLine/wg11/m57453-v1-m57453.zip m57453.docx> [retrieved on 20210706] *
WANG JIANQIANG ET AL: "Lossy Point Cloud Geometry Compression via End-to-End Learning", IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, IEEE, USA, vol. 31, no. 12, 13 January 2021 (2021-01-13), pages 4909 - 4923, XP011891210, ISSN: 1051-8215, [retrieved on 20211202], DOI: 10.1109/TCSVT.2021.3051377 *
WANG JIANQIANG ET AL: "Multiscale Point Cloud Geometry Compression", 2021 DATA COMPRESSION CONFERENCE (DCC), IEEE, 23 March 2021 (2021-03-23), pages 73 - 82, XP033912704, DOI: 10.1109/DCC50243.2021.00015 *

Similar Documents

Publication Publication Date Title
Huang et al. Octsqueeze: Octree-structured entropy model for lidar compression
EP3861755B1 (fr) Techniques et appareil de prédiction médiane pondérée pour codage d&#39;attributs de nuage de points
JP7261300B2 (ja) 適応ポイントクラウド属性コーディングのための方法、装置、及びコンピュータプログラム
JP2022542419A (ja) 点群表現を介したメッシュ圧縮
JP2023514853A (ja) 投影ベースのメッシュ圧縮
KR20140089426A (ko) 예측성 위치 디코딩
EP3343446A1 (fr) Procédé et appareil de codage et de décodage de listes de pixels
CN115336243A (zh) 基于哈尔的点云编码方法和装置
US20220337872A1 (en) Point cloud data transmission device, point cloud data transmission method, point cloud data reception device, and point cloud data reception method
Fan et al. Deep geometry post-processing for decompressed point clouds
WO2022131948A1 (fr) Dispositifs et procédés de codage séquentiel pour compression de nuage de points
US20230013421A1 (en) Point cloud compression using occupancy networks
Sarkis et al. Fast depth map compression and meshing with compressed tritree
WO2023285998A1 (fr) Compression de nuage de points utilisant des réseaux d&#39;occupation
WO2023272730A1 (fr) Procédé de codage et de décodage d&#39;un nuage de points
Ruivo et al. Double-deep learning-based point cloud geometry coding with adaptive super-resolution
US20230154052A1 (en) Point cloud data transmission device, point cloud data transmission method, point cloud data reception device and point cloud data reception method
CN112188199B (zh) 自适应点云属性编码的方法、装置、电子设备和存储介质
JP2023541271A (ja) 高密度メッシュ圧縮
US20230025378A1 (en) Task-driven machine learning-based representation and compression of point cloud geometry
CN116615753A (zh) 利用分层调色板表生成而进行的调色板模式视频编码
Mao et al. Learning to predict on octree for scalable point cloud geometry coding
US20230306643A1 (en) Mesh patch simplification
Li et al. Augmented normalizing flow for point cloud geometry coding
EP4325853A1 (fr) Dispositif d&#39;émission de données en nuage de points, procédé d&#39;émission de données en nuage de points, dispositif de réception de données en nuage de points, et procédé de réception de données en nuage de points

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22748448

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE