WO2023110562A1 - Procédés et dispositifs de codage et décodage progressifs d'images à plans multiples - Google Patents

Procédés et dispositifs de codage et décodage progressifs d'images à plans multiples Download PDF

Info

Publication number
WO2023110562A1
WO2023110562A1 PCT/EP2022/084713 EP2022084713W WO2023110562A1 WO 2023110562 A1 WO2023110562 A1 WO 2023110562A1 EP 2022084713 W EP2022084713 W EP 2022084713W WO 2023110562 A1 WO2023110562 A1 WO 2023110562A1
Authority
WO
WIPO (PCT)
Prior art keywords
tile
atlas
tiles
depth
image
Prior art date
Application number
PCT/EP2022/084713
Other languages
English (en)
Inventor
Julien Fleureau
Bertrand Chupeau
Renaud Dore
Franck Thudor
Original Assignee
Interdigital Ce Patent Holdings, Sas
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Interdigital Ce Patent Holdings, Sas filed Critical Interdigital Ce Patent Holdings, Sas
Publication of WO2023110562A1 publication Critical patent/WO2023110562A1/fr

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/597Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/70Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards

Definitions

  • the present principles generally relate to the domain of three-dimensional (3D) scene and volumetric video content.
  • the present document is also understood in the context of the encoding, the formatting and the decoding of data representative of the texture and the geometry of a 3D scene for a rendering of volumetric content on end-user devices such as mobile devices or Head- Mounted Displays (HMD).
  • the present principles particularly relate to volumetric scenes represented by a multiplane image.
  • a multiplane image is a layered representation of a volumetric scene where each layer is actually a slice of the 3D space of the scene. Each slice is sampled according to an underlying central projection (e.g. perspective, spherical, ...) and a sampling law which defines the interlayer spacing.
  • a layer comprises color as well as transparency information of any 3D intersecting object of the scene. From this sliced representation, it is possible to recover / synthesize any viewpoint located in a limited region around the center of the underlying projection. It can be performed making use of algorithms (e.g. “reversed” Painter’s algorithm) which blend each layer with the proper weights (i.e. transparency) starting from the nearest to the furthest layer. Such techniques may run faster than other known view synthesis processes.
  • algorithms e.g. “reversed” Painter’s algorithm
  • an MPI-based view synthesis process allows for “smooth” representation of object borders and in a more general manner better robustness to contour and geometry artifacts than other view synthesis algorithms. Encoding a MPI may require a large amount of data as each layer is a full- size four-component image (i.e. three components for color and one for transparency).
  • a way to encode volumetric scenes is to pack samples of projected parts of the scene (called patches) into a large image (called atlas). It is possible to encode a MPI as a patch atlas. However, such a representation tends to require bigger atlases than other representations of the volumetric scene. Mechanisms to actively control the memory and processing resources at the rendering side are missing.
  • the present principles relate a method comprising:
  • Metadata comprise information associating each tile with a depth and each tile having a tile number determined as a monotonic function of the corresponding depth;
  • the present principles also relate to a device comprising a memory associated with a processor configured for implementing the method above.
  • the present principles also relate to a method comprising: obtaining a multiplane image comprising at least two layers, a layer having a depth and pixels of a layer having a transparency component; for each layer of the multiplane image, - splitting the layer in patch pictures based on the transparency component;
  • the present principles also relate to a device comprising a memory associated with a processor configured for implementing the method above.
  • FIG. 1 shows a representation of a multiplane image, according to a non-limiting embodiment of the present principles
  • FIG. 2 shows a representation of a volumetric scene sliced in layers according to a spherical projection, according to a non-limiting embodiment of the present principles
  • FIG. 3 shows an example architecture of a device which may be configured to implement a method according to a non-limiting embodiment of the present principles
  • FIG. 4 shows an example of an embodiment of the syntax of a stream when the data are transmitted over a packet-based transmission protocol, according to a non-limiting embodiment of the present principles
  • FIG. 5 illustrates the conversion of the MPI of Figure 1 to an atlas or to a set of atlases according to a non-limiting embodiment of the present principles
  • FIG. 6 illustrates the packing of groups of patches in a plurality of atlases according to a non-limiting embodiment of the present principles
  • FIG. 7 illustrates a non-limiting embodiment of the present principles wherein an interlayer prediction is performed at the decoding stage. 5. Detailed description of embodiments
  • each block represents a circuit element, module, or portion of code which comprises one or more executable instructions for implementing the specified logical function(s).
  • the function(s) noted in the blocks may occur out of the order noted. For example, two blocks shown in succession may, in fact, be executed substantially concurrently or the blocks may sometimes be executed in the reverse order, depending on the functionality involved.
  • Figure 1 shows a representation of a multiplane image.
  • the multiplane image is composed of three layers 11, 12 and 13 from the background to the foreground according to a perspective center of projection.
  • the number of layers and the part of the volumetric scene projected onto each layer may vary.
  • the two characters of layer 13 may be represented on two separate layers.
  • the distance between each layer may also vary to be representative of the difference of depth (i.e. distance to the center of projection) of objects represented onto two consecutive layers.
  • the type of projection may differ too.
  • a panoramic MPI would be represented as concentric cylinders.
  • a 360° MPI would be represented as concentric spheres or cubes, for example.
  • Pixels of the layers may carry different component information like color, specular and/or normal vectors.
  • each layer carries transparency information (e.g. a per-pixel scalar value comprised between 0 and 1) representing the level of transparency of each pixel of the layer frame.
  • transparency information e.g. a per-pixel scalar value comprised between 0 and 1 representing the level of transparency of each pixel of the layer frame.
  • Such a level is typically set to 1 when the pixel is associated with a non-contour part of an opaque surface, but it may be lower when it is associated with an object contour or even with a transparent object.
  • Figure 2 shows a representation of a volumetric scene sliced in layers according to a spherical projection.
  • the 3D volumetric scene is automatically sliced in numerous layers.
  • Such layers of a MPI comprise large areas entirely transparent. So, it is possible to cluster more or less opaque pixels (transparency greater than zero) of each layer and to pack the patch pictures of the clustered pixels into an atlas image.
  • MPEG-I / MPEG Immersive Video is a standardization work to encode volumetric scenes.
  • volumetric scenes are samples of projected parts of the scene (called patches or patch pictures) packed into a large image (called atlas or atlas image).
  • patches or patch pictures are samples of projected parts of the scene packed into a large image (called atlas or atlas image).
  • Each layer of the MPI is considered as an independent view with constant depth (the layer depth).
  • Each patch of the set of atlases is a subpart of a layer.
  • the footprint of a path may be cumulated / aggregated for several consecutive MPIs of a sequence of MPIs, for example on an intra-period and with constant depth.
  • the depth information of each patch (a constant) is stored in associated metadata. So, there is no need to encode a geometry component carrying the depth of each projected point.
  • the additional transparency information part of the MPI is however stored in a dedicated transparency component.
  • the atlas-based decoding process imposes to decode the whole set of atlases first and then, to provide the decoded atlases to the synthesis stage for a proper rendering.
  • Such a way of proceeding has two drawbacks in the case of a MPI encoding / decoding. First, it imposes that the whole active part of the MPI, which may be quite important, is in memory before the synthesis stage. Second, it does not take advantage of the fact that for a MPI synthesis, the patches could be processed from the closest one to the furthest.
  • FIG 3 shows an example architecture of a device 30 which may be configured to implement a method described according to the present principles.
  • each circuit of the encoder and/or the Tenderer may be a device according to the architecture of Figure 3, linked together, for instance, via their bus 31 and/or via I/O interface 36.
  • Device 30 comprises following elements that are linked together by a data and address bus 31, a microprocessor 32 (or CPU), which is, for example, a DSP (or Digital Signal Processor), a ROM (or Read Only Memory) 33, a RAM (or Random Access Memory) 34, a storage interface 35, an I/O interface 36 for reception of data to transmit, from an application, and a power supply (not illustrated), e.g. a battery.
  • a microprocessor 32 or CPU
  • DSP Digital Signal Processor
  • ROM Read Only Memory
  • RAM Random Access Memory
  • the power supply is external to the device.
  • the word « register » used in the specification may correspond to area of small capacity (some bits) or to very large area (e.g. a whole program or large amount of received or encoded or decoded data).
  • the ROM 33 comprises at least a program and parameters.
  • the ROM 33 may store algorithms and instructions to perform techniques in accordance with present principles.
  • the CPU 32 uploads the program in the RAM and executes the corresponding instructions.
  • the RAM 34 comprises, in a register, the program executed by the CPU 32 and uploaded after switch-on of the device 30, input data in a register, intermediate data in different states of the method in a register, and variables used for the execution of the method in a register.
  • the implementations described herein may be implemented in, for example, a method or a process, an apparatus, a computer program product, a data stream, or a signal. Even if only discussed in the context of a single form of implementation (for example, discussed only as a method or a device), the implementation of features discussed may also be implemented in other forms (for example a program).
  • An apparatus may be implemented in, for example, appropriate hardware, software, and firmware.
  • the methods may be implemented in, for example, an apparatus such as, for example, a processor, which refers to processing devices in general, including, for example, a computer, a microprocessor, an integrated circuit, or a programmable logic device. Processors also include communication devices, such as, for example, computers, cell phones, portable/personal digital assistants ("PDAs”), and other devices that facilitate communication of information between end-users.
  • PDAs portable/personal digital assistants
  • device 30 is configured to implement a method described according to the present principles, and belongs to a set including: a mobile device, a communication device, a game device, a tablet (or tablet computer), a laptop, a video camera, an encoding chip, and a server (e.g. a broadcast server, a video-on-demand server or a web server).
  • a server e.g. a broadcast server, a video-on-demand server or a web server.
  • Figure 4 shows an example of an embodiment of the syntax of a stream when the data are transmitted over a packet-based transmission protocol.
  • Figure 4 shows an example structure 4 of a volumetric video stream.
  • the structure consists in a container which organizes the stream in independent elements of syntax.
  • the structure may comprise a header part 41 which is a set of data common to every syntax element of the stream.
  • the header part comprises metadata about syntax elements, describing the nature and the role of each.
  • the header part may also comprise metadata like the coordinates of the underlying point of view used for generating the MPI.
  • the structure comprises a payload including video data 42 and metadata 43.
  • Syntax element 42 includes data representative of the transparency of layers of the MPI.
  • Element of syntax 43 is a part of the payload of the data stream and may include metadata about how frames of element of syntax 42 are encoded, for instance parameters used for projecting and packing points of a 3D scene onto frames.
  • metadata may be associated with each frame of the video or with a group of frames (also known as Group of Pictures (GoP) in video compression standards).
  • GoP Group of Pictures
  • FIG. 5 illustrates the conversion of the MPI of Figure 1 to an atlas or to a set of atlases according to the present principles.
  • the MPI contains layers 11, 12 and 13. Layers are split in patches on a transparency base. For instance, layer 11 is split in patches 511 and 512. Each ray of the sun may be represented in an independent patch as, in the example of Figure 5, they are not directly connected to the central circle of the sun. Patches of a layer are organized into groups. A group may contain patches belonging to one layer only. So, patches of a group may have the same depth. In Figure 5, patches 511 and 512 of layer 11 are organized in group 510. Patches of layer 12 are grouped in group 520.
  • Patches may be re-oriented in the group, for example to decreasing the size needed to pack them.
  • Patches of layer 13 are packed in group 530.
  • a group of patches is an image (that can be considered an atlas image) encoded in a V3C / HEVC / VVC tile.
  • the coding or the decoding of such tiles is independent from the coding or the decoding of other tiles.
  • a tile has a unique tile number to identify it. According to the present principles, each tile number is set as a monotonic function (i.e. only ascending or only descending) of the corresponding depth; that is, tile numbers increase according to the depth of the patches extracted from the layer the tile contains.
  • Metadata associating each tile with the depth of the corresponding layer is generated and encoded in the data stream in association with the tile.
  • the tile of group 530 has the lowest tile number (for instance 0) of the set of three tiles.
  • the tile of group 520 has a tile number greater than the tile of group 530 because layer 12 has a greater depth than layer 13 (for instance 1).
  • the tile of group 510 has a tile number greater than the tile of group 520 because layer 11 has a greater depth than layer 12 (for instance 2).
  • Figure 6 illustrates the packing of groups of patches in several atlases.
  • the tiles of groups of patches obtained as described in relation to Figure 5 are packed in bigger images called atlases or atlas images. Atlases have a unique atlas number used to identify them. The atlas number of a given atlas is set so that the depth associated with the first tile (i.e. the tile with the lowest tile number) that the atlas contains is greater than the depth of the last tile (the tile with the greatest number) of any other atlas with a lower atlas number. Within each atlas, the tile numbers are determined according to the principles described in relation to Figure 5.
  • each involved tile is decoded and rendered one after the other in monotonic (i.e. ascending or descending) order of atlas and tile numbers.
  • the current viewport to render is initially cleared and each decoded tile is sequentially blended over from the nearest one to the furthest due to the numbering of the set of atlases and tiles.
  • Pixels of a patch under rendering are projected onto pixels of the viewport image according to the depth of the tile comprising the patch and metadata indicating the position of the patch in the layer of the MPI the patch has been clustered from.
  • Proj ected pixels are blended from the closest to the furthest by taking the associated transparency values into account.
  • FIG. 7 illustrates an embodiment of the present principles wherein an inter-layer prediction is performed at the decoding stage.
  • each atlas corresponds to one layer.
  • Each atlas contains one tile with one patch that is an entire layer of the MPI.
  • layers of the MPI are not split in patches.
  • An inter-layer prediction process is performed at the decoding stage to reduce the overall bitrate while ensuring a progressive rendering. Indeed, significant similarities in the content of two consecutive layers of an MPI often occur. It is, for example, the case when parts of the volumetric scene fall just in between two layers. In that case some overlapping exist, allowing the prediction of a layer from the previous one.
  • a recursive scheme is implemented by the decoder where, in addition to being temporally predicted from the previous frame, a layer is also predicted from the previous layer in depth.
  • atlas 71 is temporally predicted by using corresponding atlas 72 in the previous 3D scene.
  • atlas 72 is the representation of the same layer of the previous MPI in the sequence of MPIs.
  • atlas 71 is also partially spatially predicted from atlas 73 which is the representation of the layer of the same MPI with a depth directly lower than the depth of atlas 71. This double prediction implies a bitrate reduction while ensuring a progressive rendering.
  • the decoding of the current atlas/tile requires keeping the previous atlas/tile in memory.
  • MV-HEVC and ML-VVC are extensions of, respectively, HEVC and VVC for multi-view / multi-layer extensions. They are designed to encode multi-view content (e.g., content acquired from a rig of cameras where views share a lot of similarity). They use spatial inter-view prediction (prediction from the other views at the same timestamp) in addition to temporal inter-view prediction (prediction from the same view at different timestamps). According to the present principles, such codec extensions may be used by replacing the multi-view input by the set of layers of an MPI.
  • the representation of a MPI (or a sequence of MPIs) is encoded for allowing a progressive rendering at the decoder side.
  • This feature and corresponding parameters are indicated in metadata associated with the payload content.
  • a possible way to structure metadata is provided in the following table: vme progressive rendering enabled flag is a binary value that indicates whether a progressive rendering is possible or not at the rendering and if so, an array indicating which subset of atlases of the entire atlas set is concerned by this feature is indicated. Indeed, some atlases may contain the description of a part of the 3D scene that has been organized to be progressively decoded and some other atlases not prepared for a progressive rendering.
  • the implementations described herein may be implemented in, for example, a method or a process, an apparatus, a computer program product, a data stream, or a signal. Even if only discussed in the context of a single form of implementation (for example, discussed only as a method or a device), the implementation of features discussed may also be implemented in other forms (for example a program).
  • An apparatus may be implemented in, for example, appropriate hardware, software, and firmware.
  • the methods may be implemented in, for example, an apparatus such as, for example, a processor, which refers to processing devices in general, including, for example, a computer, a microprocessor, an integrated circuit, or a programmable logic device. Processors also include communication devices, such as, for example, Smartphones, tablets, computers, mobile phones, portable/personal digital assistants ("PDAs”), and other devices that facilitate communication of information between end-users.
  • PDAs portable/personal digital assistants
  • Implementations of the various processes and features described herein may be embodied in a variety of different equipment or applications, particularly, for example, equipment or applications associated with data encoding, data decoding, view generation, texture processing, and other processing of images and related texture information and/or depth information.
  • equipment include an encoder, a decoder, a post-processor processing output from a decoder, a pre-processor providing input to an encoder, a video coder, a video decoder, a video codec, a web server, a set-top box, a laptop, a personal computer, a cell phone, a PDA, and other communication devices.
  • the equipment may be mobile and even installed in a mobile vehicle.
  • the methods may be implemented by instructions being performed by a processor, and such instructions (and/or data values produced by an implementation) may be stored on a processor-readable medium such as, for example, an integrated circuit, a software carrier or other storage device such as, for example, a hard disk, a compact diskette (“CD”), an optical disc (such as, for example, a DVD, often referred to as a digital versatile disc or a digital video disc), a random access memory (“RAM”), or a read-only memory (“ROM”).
  • the instructions may form an application program tangibly embodied on a processor-readable medium. Instructions may be, for example, in hardware, firmware, software, or a combination.
  • a processor may be characterized, therefore, as, for example, both a device configured to carry out a process and a device that includes a processor-readable medium (such as a storage device) having instructions for carrying out a process.
  • a processor-readable medium may store, in addition to or in lieu of instructions, data values produced by an implementation.
  • implementations may produce a variety of signals formatted to carry information that may be, for example, stored or transmitted. The information may include, for example, instructions for performing a method, or data produced by one of the described implementations.
  • a signal may be formatted to carry as data the rules for writing or reading the syntax of a described embodiment, or to carry as data the actual syntax-values written by a described embodiment.
  • Such a signal may be formatted, for example, as an electromagnetic wave (for example, using a radio frequency portion of spectrum) or as a baseband signal.
  • the formatting may include, for example, encoding a data stream and modulating a carrier with the encoded data stream.
  • the information that the signal carries may be, for example, analog or digital information.
  • the signal may be transmitted over a variety of different wired or wireless links, as is known.
  • the signal may be stored on a processor-readable medium.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

Des procédés et des dispositifs de codage, de décodage et de transmission d'une scène tridimensionnelle initialement représentée sous la forme d'une image à plans multiples (MPI) sont divulgués. Chaque couche de la MPI est divisée en parcelles sur la base d'un composant de transparence. Les parcelles d'une couche sont regroupées en une tuile. Plus la profondeur de la couche est importante, plus le numéro d'identification de la tuile est important. Lorsque plusieurs tuiles sont réunies dans une image d'atlas, la même fonction monotone (c'est-à-dire ascendante ou descendante) selon la profondeur s'applique aux numéros d'atlas. Côté décodage, la fenêtre d'affichage en cours de rendu est initialement effacée et chaque tuile décodée est incorporée séquentiellement, de la plus proche à la plus éloignée, en fonction de la numérotation de l'ensemble d'atlas et de tuiles. Les pixels d'une parcelle en cours de rendu sont projetés sur les pixels de l'image de fenêtre d'affichage selon la profondeur de la tuile comprenant la parcelle et les métadonnées indiquant la position de la parcelle dans la couche de la MPI à partir de laquelle la parcelle a été regroupée.
PCT/EP2022/084713 2021-12-17 2022-12-07 Procédés et dispositifs de codage et décodage progressifs d'images à plans multiples WO2023110562A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
EP21306825 2021-12-17
EP21306825.7 2021-12-17

Publications (1)

Publication Number Publication Date
WO2023110562A1 true WO2023110562A1 (fr) 2023-06-22

Family

ID=79317139

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/EP2022/084713 WO2023110562A1 (fr) 2021-12-17 2022-12-07 Procédés et dispositifs de codage et décodage progressifs d'images à plans multiples

Country Status (1)

Country Link
WO (1) WO2023110562A1 (fr)

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2021240069A1 (fr) * 2020-05-27 2021-12-02 Nokia Technologies Oy Couches de texture de décalage pour codage et signalisation de réflexion et réfraction pour vidéo immersive et procédés pour vidéo volumétrique multicouche associés

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2021240069A1 (fr) * 2020-05-27 2021-12-02 Nokia Technologies Oy Couches de texture de décalage pour codage et signalisation de réflexion et réfraction pour vidéo immersive et procédés pour vidéo volumétrique multicouche associés

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
"Text of ISO/IEC DIS 23090-5 Visual Volumetric Video-based Coding and Video-based Point Cloud Compression 2nd Edition", no. n20761, 23 July 2021 (2021-07-23), XP030296513, Retrieved from the Internet <URL:https://dms.mpeg.expert/doc_end_user/documents/135_OnLine/wg11/MDS20761_WG07_N00188.zip WG07N0188_ISO_IEC_23090-5_DIS_2ed.pdf> [retrieved on 20210723] *
BERTRAND CHUPEAU (INTERDIGITAL) ET AL: "[MIV] Editors draft of MIV specification", no. m58499, 14 December 2021 (2021-12-14), XP030299380, Retrieved from the Internet <URL:https://dms.mpeg.expert/doc_end_user/documents/137_OnLine/wg11/m58499-v1-m58499_ISO_IEC_23090-12_FDIS_Editors_Draft_v1.zip m58499_ISO_IEC_23090-12_FDIS_Editors_Draft_v1.docx> [retrieved on 20211214] *
BOYCE JILL M. ET AL: "MPEG Immersive Video Coding Standard", PROCEEDINGS OF THE IEEE, 1 January 2021 (2021-01-01), US, pages 1 - 16, XP055824808, ISSN: 0018-9219, DOI: 10.1109/JPROC.2021.3062590 *

Similar Documents

Publication Publication Date Title
US20220343549A1 (en) A method and apparatus for encoding, transmitting and decoding volumetric video
US12020461B2 (en) Method and apparatus for Haar-based point cloud coding
US11451836B2 (en) Techniques and apparatus for PCM patch creation using Morton codes
US20210383574A1 (en) Method and apparatus for enhanced patch boundary identification for point cloud compression
WO2023132919A1 (fr) Structure évolutive pour compression de nuage de points
US20240098235A1 (en) Method and apparatus for generating compact multiplane images
EP3554083A1 (fr) Procédé et appareil de codage et de décodage de métadonnées associées à une projection corrigée de nuages de points
US20230362409A1 (en) A method and apparatus for signaling depth of multi-plane images-based volumetric video
WO2023110562A1 (fr) Procédés et dispositifs de codage et décodage progressifs d&#39;images à plans multiples
US11250594B2 (en) Method and apparatus for geometry smoothing by local geometry projection
EP4260563A1 (fr) Signalisation de transfert radiatif pour vidéo immersive
US20230379495A1 (en) A method and apparatus for encoding mpi-based volumetric video
US11979606B2 (en) Conditional recolor for video based point cloud coding
US11956478B2 (en) Method and apparatus for point cloud chunking for improved patch packing and coding efficiency
US11606556B2 (en) Fast patch generation for video based point cloud coding
US20230011096A1 (en) Group of pictures based patch packing for video based point cloud coding
US20240195947A1 (en) Patch-based depth mapping method and apparatus for high-efficiency encoding/decoding of plenoptic video
US20220394294A1 (en) Non-binary occupancy map for video based point cloud coding
US20240029311A1 (en) Point cloud data transmission device, point cloud data transmission method, point cloud data reception device, and point cloud data reception method
US20230370636A1 (en) Image processing device and method
WO2024094540A1 (fr) Format de codage pour codage optimisé de vidéo volumétrique
WO2023113917A1 (fr) Structure hybride pour compression de nuage de points

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22830799

Country of ref document: EP

Kind code of ref document: A1