WO2023285998A1 - Compression de nuage de points utilisant des réseaux d'occupation - Google Patents
Compression de nuage de points utilisant des réseaux d'occupation Download PDFInfo
- Publication number
- WO2023285998A1 WO2023285998A1 PCT/IB2022/056480 IB2022056480W WO2023285998A1 WO 2023285998 A1 WO2023285998 A1 WO 2023285998A1 IB 2022056480 W IB2022056480 W IB 2022056480W WO 2023285998 A1 WO2023285998 A1 WO 2023285998A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- occupancy
- networks
- probability
- bitstream
- function
- Prior art date
Links
- 238000007906 compression Methods 0.000 title abstract description 42
- 230000006835 compression Effects 0.000 title abstract description 42
- 230000006870 function Effects 0.000 claims description 64
- 238000000034 method Methods 0.000 claims description 38
- 238000013528 artificial neural network Methods 0.000 claims description 15
- 238000010801 machine learning Methods 0.000 claims description 13
- 230000001537 neural effect Effects 0.000 claims description 11
- 238000013527 convolutional neural network Methods 0.000 abstract description 4
- 238000010586 diagram Methods 0.000 description 6
- 230000005540 biological transmission Effects 0.000 description 3
- 230000000694 effects Effects 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 230000003190 augmentative effect Effects 0.000 description 1
- 230000001413 cellular effect Effects 0.000 description 1
- 230000003750 conditioning effect Effects 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 238000013144 data compression Methods 0.000 description 1
- 238000011156 evaluation Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000002123 temporal effect Effects 0.000 description 1
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/90—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using coding techniques not provided for in groups H04N19/10-H04N19/85, e.g. fractals
- H04N19/91—Entropy coding, e.g. variable length coding [VLC] or arithmetic coding
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T9/00—Image coding
- G06T9/001—Model-based coding, e.g. wire frame
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T9/00—Image coding
- G06T9/002—Image coding using neural networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T9/00—Image coding
- G06T9/40—Tree coding, e.g. quadtree, octree
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/90—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using coding techniques not provided for in groups H04N19/10-H04N19/85, e.g. fractals
- H04N19/96—Tree coding, e.g. quad-tree coding
Definitions
- the present invention relates to three dimensional graphics. More specifically, the present invention relates to coding of three dimensional graphics.
- Point clouds have been considered as a candidate format for transmission of 3D data, either captured by 3D scanners, LIDAR sensors, or used in popular applications such as VR/AR.
- Point clouds are a set of points in 3D space.
- each point usually have associated attributes, such as color (R, G, B) or even reflectance and temporal timestamps (e.g., in LIDAR images).
- attributes such as color (R, G, B) or even reflectance and temporal timestamps (e.g., in LIDAR images).
- devices capture point clouds in the order of thousands or even millions of points.
- Occupancy networks implicitly represent the 3D surface as the continuous decision boundary of a deep neural network classifier.
- the representation encodes a description of the 3D output at infinite resolution.
- Occupancy networks enable efficient and flexible point cloud compression.
- occupancy networks are able to handle points, meshes, or projected images of 3D objects, making them very flexible in terms of input signal representation.
- the probability of occupancy of positions is estimated using occupancy networks instead of sparse convolutional neural networks.
- a compression implementation using occupancy network enables scalability with infinite reconstruction resolution.
- a method programmed in a non-transitory memory of a device comprises receiving a bitstream at one or more occupancy networks, determining a probability of a position in the bitstream being occupied with the one or more occupancy networks and generating a function based on the probability of positions being occupied.
- the bitstream comprises voxels, points, meshes, or projected images of 3D objects.
- the bitstream comprises one or more samples of a 3D space to be used to generate a 3D object with the one or more occupancy networks.
- the probability is determined using machine learning to implement implicit neural functions.
- the one or more occupancy networks implicitly represent 3D surfaces using a continuous decision boundary based on a deep neural network classifier, and decides based on a threshold whether data belongs inside or outside a 3D structure.
- the probability is determined based neighboring position classification information.
- the probability is used by an entropy encoder to define a code length of an occupancy code of points in 3D space.
- the one or more occupancy networks learn the function to recover a specific shape based on a sparse input.
- the function represents a set of classes, and an object is recovered based on an input. A size of the function is smaller than the bitstream.
- an apparatus comprises a non-transitory memory for storing an application, the application for: receiving a bitstream at one or more occupancy networks, determining a probability of a position in the bitstream being occupied with the one or more occupancy networks and generating a function based on the probability of positions being occupied and a processor coupled to the memory, the processor configured for processing the application.
- the bitstream comprises voxels, points, meshes, or projected images of 3D objects.
- the bitstream comprises one or more samples of a 3D space to be used to generate a 3D object with the one or more occupancy networks.
- the probability is determined using machine learning to implement implicit neural functions.
- the one or more occupancy networks implicitly represent 3D surfaces using a continuous decision boundary based on a deep neural network classifier, and decides based on a threshold whether data belongs inside or outside a 3D structure.
- the probability is determined based neighboring position classification information.
- the probability is used by an entropy encoder to define a code length of an occupancy code of points in 3D space.
- the one or more occupancy networks learn the function to recover a specific shape based on a sparse input.
- the function represents a set of classes, and an object is recovered based on an input. A size of the function is smaller than the bitstream.
- a system comprises an encoder configured for: receiving a bitstream at one or more occupancy networks, determining a probability of a position in the bitstream being occupied with the one or more occupancy networks and generating a function based on the probability of positions being occupied and a decoder configured for: recovering an object based on the function and an input.
- the bitstream comprises voxels, points, meshes, or projected images of 3D objects.
- the bitstream comprises one or more samples of a 3D space to be used to generate a 3D object with the one or more occupancy networks.
- the probability is determined using machine learning to implement implicit neural functions.
- the one or more occupancy networks implicitly represent 3D surfaces using a continuous decision boundary based on a deep neural network classifier, and decides based on a threshold whether data belongs inside or outside a 3D structure.
- the probability is determined based neighboring position classification information.
- the probability is used by to define a code length of an occupancy code of points in 3D space. A size of the function is smaller than the bitstream.
- Figure 1 illustrates a diagram of occupancy networks according to some embodiments.
- Figure 2 illustrates a diagram of point cloud compression using occupancy networks according to some embodiments.
- Figure 3 illustrates a flowchart of a method of implementing point cloud compression using occupancy networks according to some embodiments.
- Figure 4 illustrates a block diagram of an exemplary computing device configured to implement the method of implementing point cloud compression using occupancy networks according to some embodiments.
- a point cloud compression scheme uses occupancy networks as an implicit representation of the points.
- the implicit neural functions define an occupancy probability for points in 3D space. This probability is then used by an entropy encoder to define the code length of the occupancy code of points in 3D space.
- the MPEG is currently concluding two standards for Point Cloud Compression (PCC).
- Point clouds are used to represent three-dimensional scenes and objects, and are composed by volumetric elements (voxels) described by their position in 3D space and attributes such as color, reflectance, material, transparency, time stamp and others.
- the planned outcome of the standardization activity is the Geometry-based Point Cloud Compression (G-PCC) and the
- V-PCC Video-based Point Cloud Compression
- a sparse convolutional network exploits the spatial dependency between neighbors to estimate the occupancy of voxels by means of probabilities used for entropy coding or binary classification, depending if one wants to perform lossless or lossy compression, respectively.
- an occupancy network is described, which performs the same task by assigning to every location/position an occupancy probability between 0 and 1.
- the embodiments described herein are more general since the method is able to be applied to points, meshes or projected images of 3D objects, and is not limited to a voxel-based representation. Scalability is able to be provided by voxebzing the volumetric space at an initial resolution and evaluating the occupancy network for all points in a grid.
- Occupancy networks have several applications. Their usage is a scalable, and a more generic point cloud compression scheme is novel. Occupancy networks enable efficient and flexible point cloud compression. Although based on occupancy estimation, sparse convolutional neural networks are typically limited to voxel-based representation. In addition to the voxel-based representation, occupancy networks are able to deal with points, meshes, or projected images of 3D objects, making them more flexible in terms of input signal representation. The probability of occupancy of positions is estimated using occupancy networks instead of sparse convolutional neural networks.
- the occupancy network implicitly represents 3D surfaces using a continuous decision boundary based on a deep neural network classifier, and decides based on a boundary (threshold) whether a point belongs inside or outside a 3D structure (e.g., mesh).
- the occupancy network repetitively decides whether a point belongs inside or outside and by doing this, the occupancy network defines the surface of the volumetric representation.
- the occupancy network is used to determine the probability of a position in space being occupied.
- the occupancy network is able to be used to assist in compression as well.
- Figure 1 illustrates a diagram of occupancy networks according to some embodiments.
- Occupancy networks learn general characteristics of classes of objects.
- occupancy maps learn a function that is able to recover a specific shape based on a sparse input.
- an occupancy network 100 is able to represent chairs and tables.
- the occupancy network 100 is then able to receive a sparse representation of an object 102 (e.g., chair) as an input to the function and produce a representation 104 of the object in its original form with a desired precision (e.g., a more detailed object).
- a function receives a dataset of sparse points, and the function outputs an object similar to one of the classes the occupancy network function had learned.
- the precision of the function is not mathematically limited.
- point cloud compression In addition to recovering the object from the sparse point cloud, efficient and flexible point cloud compression is able to be performed.
- the method is flexible because in addition to points, other forms of input are able to be used such as voxels, 2D images (projections) and meshes.
- the input data is able to be compressed regardless of the input form using occupancy estimation.
- FIG. 2 illustrates a diagram of point cloud compression using occupancy networks according to some embodiments.
- a bitstream 200 is received at the occupancy networks 202.
- the bitstream 200 is able to be voxels, points, projections, meshes or others.
- the occupancy networks 202 are one or more neural networks able to obtain the implicit representation of a 3D object.
- the bitstream 200 is able to comprise network coefficients and/or random samples of a 3D space instead of a point cloud, and the occupancy networks 202 are able to generate a 3D object based on the network coefficients and random samples to check occupied positions in 3D space.
- Occupancy networks 202 progressively divide the space into smaller and smaller regions/divisions. For each division, the probability of positions being occupied is calculated.
- the upper left region of the first block 210 has a 0.94 (or 94%) probability of having an occupied position.
- the first block 210 is able to be divided further into the more refined second block 212 which has smaller divisions.
- the blocks are able to be divided many more times, for example, to an n th block 214 which has the smallest divisions, in the example.
- the probabilities of a position being occupied are shown for all four blocks, although probabilities below a threshold (e.g., 0.50) indicate the position is not likely occupied.
- a threshold e.g. 0.50
- the block is able to be divided theoretically infinitely (limited only by processing power and memory). By being able to divide the blocks many times, a system is able to be very scalable. For example, a system is able to output point clouds with different degrees of detail (e.g., coarse to fine detail).
- the occupancy network assigns to every location an occupancy probability between 0 and 1.
- An occupancy network is used but not necessarily the full capacity of a neural network.
- the surface of an object is generated based on the observation of that object (input conditioning). Furthering the example, a full, continuous surface of an object may not be generated, where only a certain level of detail is included.
- Scalability is provided in by voxebzing the volumetric space an initial resolution and evaluating the occupancy network for all points in the grid. Grid points p are marked as occupied if the evaluated value of the function at the point is bigger or equal to some threshold, which is given as a hyperparameter. In some embodiments, all voxel/points are marked as active, if at least two adjacent grid points have differing occupancy predictions.
- the occupancy network is used to compress point clouds.
- the implicit 3D surface representation not encoding the points themselves, rather a function is encoded.
- the function is encoded.
- the function is able to represent a set of classes, and then an object is able to be recovered based on an input.
- the object itself is not encoded; rather, the function is encoded. This is also referred to as an implicit 3D surface representation.
- different aspects of an object are able to have different amounts of refinement (e.g., coarse to fine).
- Figure 3 illustrates a flowchart of a method of implementing point cloud compression using occupancy networks according to some embodiments.
- a bitstream is received at occupancy networks.
- the bitstream is able to include voxels, points, meshes, projected images of 3D objects or other data.
- the probability of a position being occupied in the bitstream is determined.
- the probability is determined in any manner such as based on machine learning (e.g., the implicit neural functions define an occupancy probability for points in 3D space) and/or classifications of the current object and previously classified objects.
- the occupancy network implicitly represents 3D surfaces using a continuous decision boundary based on a deep neural network classifier, and decides based on a boundary (threshold) whether a point belongs inside or outside a 3D structure (e.g., mesh).
- the occupancy network repetitively decides whether each point (or other data) belongs inside or outside and by doing this, the occupancy network defines the surface of the volumetric representation.
- the probability is also able to be determined based on current information (e.g., a position that has two neighboring positions with a high probability of being occupied is also able to have a high probability of being occupied).
- the probability is then used by an entropy encoder to define the code length of the occupancy code of points in 3D space.
- a function is generated based on the probability of positions being occupied.
- occupancy networks/maps learn a function that is able to recover a specific shape based on a sparse input.
- the occupancy network is then able to receive a sparse representation of an object(e.g., chair) as an input to the function and produce a representation of the object in its original form with a desired precision (e.g., a more detailed object).
- a function receives a dataset of sparse points, and the function outputs an object similar to one of the classes the occupancy network function had learned.
- the function is able to represent a set of classes, and then an object is able to be recovered based on an input.
- the object itself is not encoded; rather, the function is encoded. This is also referred to as an implicit 3D surface representation. Since the function does not include all of the data points, the representation is a compressed version of the input bitstream.
- fewer or additional steps are implemented. In some embodiments, the order of the steps is modified.
- Figure 4 illustrates a block diagram of an exemplary computing device configured to implement the method of implementing point cloud compression using occupancy networks according to some embodiments.
- the computing device 400 is able to be used to acquire, store, compute, process, communicate and/or display information such as images and videos including 3D content.
- the computing device 400 is able to implement any of the encoding/decoding aspects.
- a hardware structure suitable for implementing the computing device 400 includes a network interface 402, a memory 404, a processor 406, I/O device(s) 408, a bus 410 and a storage device 412.
- the choice of processor is not critical as long as a suitable processor with sufficient speed is chosen.
- the memory 404 is able to be any conventional computer memory known in the art.
- the storage device 412 is able to include a hard drive, CDROM, CDRW, DVD, DVDRW, High Definition disc/drive, ultra-HD drive, flash memory card or any other storage device.
- the computing device 400 is able to include one or more network interfaces 402. An example of a network interface includes a network card connected to an Ethernet or other type of LAN.
- the I/O device(s) 408 are able to include one or more of the following: keyboard, mouse, monitor, screen, printer, modem, touchscreen, button interface and other devices.
- Compression application(s) 430 used to implement the compression implementation are likely to be stored in the storage device 412 and memory 404 and processed as applications are typically processed. More or fewer components shown in Figure 4 are able to be included in the computing device 400.
- compression hardware 420 is included.
- the computing device 400 in Figure 4 includes applications 430 and hardware 420 for the compression method, the compression method is able to be implemented on a computing device in hardware, firmware, software or any combination thereof.
- the compression applications 430 are programmed in a memory and executed using a processor.
- the compression hardware 420 is programmed hardware logic including gates specifically designed to implement the compression method.
- the compression application(s) 430 include several applications and/or modules.
- modules include one or more sub-modules as well. In some embodiments, fewer or additional modules are able to be included.
- suitable computing devices include a personal computer, a laptop computer, a computer workstation, a server, a mainframe computer, a handheld computer, a personal digital assistant, a cellular/mobile telephone, a smart appliance, a gaming console, a digital camera, a digital camcorder, a camera phone, a smart phone, a portable music player, a tablet computer, a mobile device, a video player, a video disc writer/player (e.g., DVD writer/player, high definition disc writer/player, ultra high definition disc writer/player), a television, a home entertainment system, an augmented reality device, a virtual reality device, smart jewelry (e.g., smart watch), a vehicle (e.g., a self-driving vehicle) or any other suitable computing device.
- a personal computer e.g., a laptop computer, a computer workstation, a server, a mainframe computer, a handheld computer, a personal digital assistant, a cellular/mobile telephone, a smart appliance, a gaming console
- a device acquires or receives 3D content (e.g., point cloud content).
- 3D content e.g., point cloud content.
- the compression method is able to be implemented with user assistance or automatically without user involvement.
- the compression method enables more efficient and more accurate 3D content encoding compared to previous implementations.
- the compression method is highly scalable as well.
- a method programmed in a non-transitory memory of a device comprising: receiving a bitstream at one or more occupancy networks; determining a probability of a position in the bitstream being occupied with the one or more occupancy networks; and generating a function based on the probability of positions being occupied.
- the bitstream comprises voxels, points, meshes, or projected images of 3D objects.
- bitstream comprises one or more samples of a 3D space to be used to generate a 3D object with the one or more occupancy networks.
- An apparatus comprising: a non-transitory memory for storing an application, the application for: receiving a bitstream at one or more occupancy networks; determining a probability of a position in the bitstream being occupied with the one or more occupancy networks; and generating a function based on the probability of positions being occupied; and a processor coupled to the memory, the processor configured for processing the application.
- the bitstream comprises voxels, points, meshes, or projected images of 3D objects.
- bitstream comprises one or more samples of a 3D space to be used to generate a 3D object with the one or more occupancy networks.
- the one or more occupancy networks implicitly represent 3D surfaces using a continuous decision boundary based on a deep neural network classifier, and decides based on a threshold whether data belongs inside or outside a 3D structure.
- the probability is determined based neighboring position classification information.
- the probability is used by an entropy encoder to define a code length of an occupancy code of points in 3D space.
- a system comprising: an encoder configured for: receiving a bitstream at one or more occupancy networks; determining a probability of a position in the bitstream being occupied with the one or more occupancy networks; and generating a function based on the probability of positions being occupied; and a decoder configured for: recovering an object based on the function and an input.
- bitstream comprises voxels, points, meshes, or projected images of 3D objects.
- bitstream comprises one or more samples of a 3D space to be used to generate a 3D object with the one or more occupancy networks.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Signal Processing (AREA)
- Artificial Intelligence (AREA)
- Evolutionary Computation (AREA)
- Image Analysis (AREA)
Abstract
Selon l'invention, des réseaux d'occupation permettent une compression de nuage de points efficace et flexible. En plus de la représentation par voxels, les réseaux d'occupation peuvent gérer des points, des maillages, ou des images projetées d'objets 3D, ce qui les rend très souples pour la représentation de signal d'entrée. La probabilité d'occupation de positions est estimée en utilisant des réseaux d'occupation au lieu de réseaux neuronaux convolutifs clairsemés. Une mise en œuvre de compression utilisant un réseau d'occupation permet une extensibilité avec une résolution de reconstruction infinie.
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US202163221552P | 2021-07-14 | 2021-07-14 | |
US63/221,552 | 2021-07-14 | ||
US17/828,326 US20230013421A1 (en) | 2021-07-14 | 2022-05-31 | Point cloud compression using occupancy networks |
US17/828,326 | 2022-05-31 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2023285998A1 true WO2023285998A1 (fr) | 2023-01-19 |
Family
ID=82748778
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/IB2022/056480 WO2023285998A1 (fr) | 2021-07-14 | 2022-07-14 | Compression de nuage de points utilisant des réseaux d'occupation |
Country Status (1)
Country | Link |
---|---|
WO (1) | WO2023285998A1 (fr) |
-
2022
- 2022-07-14 WO PCT/IB2022/056480 patent/WO2023285998A1/fr unknown
Non-Patent Citations (6)
Title |
---|
EMRE CAN KAYA ET AL: "Neural Network Modeling of Probabilities for Coding the Octree Representation of Point Clouds", ARXIV.ORG, CORNELL UNIVERSITY LIBRARY, 201 OLIN LIBRARY CORNELL UNIVERSITY ITHACA, NY 14853, 11 June 2021 (2021-06-11), XP081988278 * |
MENG JIA ET AL: "Learning Occupancy Function from Point Clouds for Surface Reconstruction", ARXIV.ORG, CORNELL UNIVERSITY LIBRARY, 201 OLIN LIBRARY CORNELL UNIVERSITY ITHACA, NY 14853, 22 October 2020 (2020-10-22), XP081793028 * |
MESCHEDER LARS ET AL: "Occupancy Networks: Learning 3D Reconstruction in Function Space", 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), IEEE, 15 June 2019 (2019-06-15), pages 4455 - 4465, XP033686741, DOI: 10.1109/CVPR.2019.00459 * |
WANG J ET AL: "[G-PCC EE13.54] A Geometry Compression Framework for AI-based PCC via Sparse Convolution", no. m57453, 6 July 2021 (2021-07-06), XP030297111, Retrieved from the Internet <URL:https://dms.mpeg.expert/doc_end_user/documents/135_OnLine/wg11/m57453-v1-m57453.zip m57453.docx> [retrieved on 20210706] * |
WANG JIANQIANG ET AL: "Lossy Point Cloud Geometry Compression via End-to-End Learning", IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, IEEE, USA, vol. 31, no. 12, 13 January 2021 (2021-01-13), pages 4909 - 4923, XP011891210, ISSN: 1051-8215, [retrieved on 20211202], DOI: 10.1109/TCSVT.2021.3051377 * |
WANG JIANQIANG ET AL: "Multiscale Point Cloud Geometry Compression", 2021 DATA COMPRESSION CONFERENCE (DCC), IEEE, 23 March 2021 (2021-03-23), pages 73 - 82, XP033912704, DOI: 10.1109/DCC50243.2021.00015 * |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Huang et al. | Octsqueeze: Octree-structured entropy model for lidar compression | |
EP3861755B1 (fr) | Techniques et appareil de prédiction médiane pondérée pour codage d'attributs de nuage de points | |
JP7261300B2 (ja) | 適応ポイントクラウド属性コーディングのための方法、装置、及びコンピュータプログラム | |
JP2022542419A (ja) | 点群表現を介したメッシュ圧縮 | |
JP2023514853A (ja) | 投影ベースのメッシュ圧縮 | |
KR20140089426A (ko) | 예측성 위치 디코딩 | |
EP3343446A1 (fr) | Procédé et appareil de codage et de décodage de listes de pixels | |
CN115336243A (zh) | 基于哈尔的点云编码方法和装置 | |
US20220337872A1 (en) | Point cloud data transmission device, point cloud data transmission method, point cloud data reception device, and point cloud data reception method | |
Fan et al. | Deep geometry post-processing for decompressed point clouds | |
WO2022131948A1 (fr) | Dispositifs et procédés de codage séquentiel pour compression de nuage de points | |
US20230013421A1 (en) | Point cloud compression using occupancy networks | |
Sarkis et al. | Fast depth map compression and meshing with compressed tritree | |
WO2023285998A1 (fr) | Compression de nuage de points utilisant des réseaux d'occupation | |
WO2023272730A1 (fr) | Procédé de codage et de décodage d'un nuage de points | |
Ruivo et al. | Double-deep learning-based point cloud geometry coding with adaptive super-resolution | |
US20230154052A1 (en) | Point cloud data transmission device, point cloud data transmission method, point cloud data reception device and point cloud data reception method | |
CN112188199B (zh) | 自适应点云属性编码的方法、装置、电子设备和存储介质 | |
JP2023541271A (ja) | 高密度メッシュ圧縮 | |
US20230025378A1 (en) | Task-driven machine learning-based representation and compression of point cloud geometry | |
CN116615753A (zh) | 利用分层调色板表生成而进行的调色板模式视频编码 | |
Mao et al. | Learning to predict on octree for scalable point cloud geometry coding | |
US20230306643A1 (en) | Mesh patch simplification | |
Li et al. | Augmented normalizing flow for point cloud geometry coding | |
EP4325853A1 (fr) | Dispositif d'émission de données en nuage de points, procédé d'émission de données en nuage de points, dispositif de réception de données en nuage de points, et procédé de réception de données en nuage de points |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 22748448 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |