WO2022111343A1 - 非时序点云媒体的处理方法、装置、设备及存储介质 - Google Patents
非时序点云媒体的处理方法、装置、设备及存储介质 Download PDFInfo
- Publication number
- WO2022111343A1 WO2022111343A1 PCT/CN2021/131037 CN2021131037W WO2022111343A1 WO 2022111343 A1 WO2022111343 A1 WO 2022111343A1 CN 2021131037 W CN2021131037 W CN 2021131037W WO 2022111343 A1 WO2022111343 A1 WO 2022111343A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- gpcc
- point cloud
- area
- target
- entry
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 123
- 238000012545 processing Methods 0.000 title claims abstract description 55
- 230000003068 static effect Effects 0.000 claims abstract description 80
- 230000011664 signaling Effects 0.000 claims abstract description 35
- 238000004519 manufacturing process Methods 0.000 claims description 71
- 238000004590 computer program Methods 0.000 claims description 31
- 230000008569 process Effects 0.000 claims description 20
- 238000004891 communication Methods 0.000 claims description 18
- 238000005538 encapsulation Methods 0.000 claims description 16
- 230000006835 compression Effects 0.000 claims description 7
- 238000007906 compression Methods 0.000 claims description 7
- 238000010586 diagram Methods 0.000 description 22
- 230000006870 function Effects 0.000 description 7
- 230000001360 synchronised effect Effects 0.000 description 5
- 230000005540 biological transmission Effects 0.000 description 4
- 238000009877 rendering Methods 0.000 description 4
- 230000002123 temporal effect Effects 0.000 description 4
- 230000000007 visual effect Effects 0.000 description 4
- 230000006978 adaptation Effects 0.000 description 3
- 230000008878 coupling Effects 0.000 description 3
- 238000010168 coupling process Methods 0.000 description 3
- 238000005859 coupling reaction Methods 0.000 description 3
- 238000005516 engineering process Methods 0.000 description 3
- 210000000056 organ Anatomy 0.000 description 3
- 238000002591 computed tomography Methods 0.000 description 2
- 238000011960 computer-aided design Methods 0.000 description 2
- 230000002452 interceptive effect Effects 0.000 description 2
- 239000000463 material Substances 0.000 description 2
- 238000003672 processing method Methods 0.000 description 2
- 230000003044 adaptive effect Effects 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 235000019800 disodium phosphate Nutrition 0.000 description 1
- 239000011521 glass Substances 0.000 description 1
- 230000010354 integration Effects 0.000 description 1
- 238000002595 magnetic resonance imaging Methods 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 239000012092 media component Substances 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 239000013307 optical fiber Substances 0.000 description 1
- 230000008520 organization Effects 0.000 description 1
- 238000012858 packaging process Methods 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 230000000153 supplemental effect Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/20—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using video object coding
- H04N19/23—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using video object coding with coding of regions that are present throughout a whole video segment, e.g. sprites, background or mosaic
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/236—Assembling of a multiplex stream, e.g. transport stream, by combining a video stream with other content or additional data, e.g. inserting a URL [Uniform Resource Locator] into a video stream, multiplexing software data into a video stream; Remultiplexing of multiplex streams; Insertion of stuffing bits into the multiplex stream, e.g. to obtain a constant bit-rate; Assembling of a packetised elementary stream
- H04N21/23614—Multiplexing of additional data and video streams
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T17/00—Three dimensional [3D] modelling, e.g. data description of 3D objects
- G06T17/20—Finite element generation, e.g. wire-frame surface description, tesselation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
- G06T15/005—General purpose rendering architectures
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/119—Adaptive subdivision aspects, e.g. subdivision of a picture into rectangular or non-rectangular coding blocks
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/136—Incoming video signal characteristics or properties
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/164—Feedback from the receiver or from the transmission channel
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/167—Position within a video image, e.g. region of interest [ROI]
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/17—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/597—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/235—Processing of additional data, e.g. scrambling of additional data or processing content descriptors
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/81—Monomedia components thereof
- H04N21/816—Monomedia components thereof involving special video data, e.g 3D video
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/83—Generation or processing of protective or descriptive data associated with content; Content structuring
- H04N21/84—Generation or processing of descriptive data, e.g. content descriptors
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/85—Assembly of content; Generation of multimedia applications
- H04N21/854—Content authoring
- H04N21/85406—Content authoring involving a specific file format, e.g. MP4 format
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2200/00—Indexing scheme for image data processing or generation, in general
- G06T2200/04—Indexing scheme for image data processing or generation, in general involving 3D image data
Definitions
- the embodiments of the present application relate to the field of computer technologies, and in particular, to non-sequential point cloud media.
- the point cloud data of objects can be obtained in many ways, and the video production device can transmit the point cloud data to the video playback device in the form of point cloud media, that is, point cloud media files, so that the video playback device can play the point cloud media.
- point cloud media that is, point cloud media files
- point cloud data for the same object can be encapsulated into different point cloud media, for example: some point cloud media are the entire point cloud media of the object, while some point cloud media are only part of the object's point cloud. media.
- the present application provides a processing method, device, device and storage medium for time-series point cloud media, so that users can request non-sequential point cloud media of the same static object in multiple times, so as to improve processing efficiency and user experience.
- the present application provides a method for processing non-sequential point cloud media.
- the method is executed by a video production device, and the method includes: acquiring non-sequential point cloud data of a static object; The data is processed to obtain a GPCC bit stream; the GPCC bit stream is encapsulated to generate at least one entry of the GPCC area, and the entry of the GPCC area is used to represent the GPCC component of the three-dimensional 3D space area corresponding to the GPCC area; Encapsulate the entries in the GPCC area, and generate at least one non-sequential point cloud media of static objects, and the non-sequential point cloud media includes the identifier of the static object; send the MPD signaling of at least one non-sequential point cloud media to the video playback device ; receive the first request message sent by the video playback device according to the MPD signaling, the first request message is used to request the first non-sequential point cloud media in the at least one non-s
- the present application provides a method for processing non-sequential point cloud media.
- the method is performed by a video playback device.
- the method includes: receiving MPD signaling of at least one non-sequential point cloud media, the The cloud media includes the identifier of the static object; sends a first request message to the video production device according to the MPD signaling, where the first request message is used to request the first non-sequential point cloud media in the at least one non-sequential point cloud media point cloud media; receiving first non-sequential point cloud media from the video production device; playing the first non-sequential point cloud media; wherein at least one non-sequential point cloud media encapsulates at least one entry of point cloud compression GPCC area Generated, the entry of the at least one GPCC area is generated by encapsulating the GPCC bit stream, and the GPCC bit stream is obtained by processing the non-sequential point cloud data of the static object by GPCC encoding; for at least
- the present application provides a processing device for non-sequential point cloud media, including: a processing unit and a communication unit; the processing unit is used for: acquiring non-sequential point cloud data of static objects; The data is processed to obtain a GPCC bit stream; the GPCC bit stream is encapsulated to generate at least one entry of the GPCC area, and the entry of the GPCC area is used to represent the GPCC component of the three-dimensional 3D space area corresponding to the GPCC area; Encapsulate the entries in the GPCC area, and generate at least one non-sequential point cloud media of static objects, and the non-sequential point cloud media includes the identifier of the static object; send the MPD signaling of at least one non-sequential point cloud media to the video playback device
- the communication unit is configured to: receive a first request message sent by the video playback device according to the MPD signaling, where the first request message is used to request the first non-sequential point cloud
- the present application provides an apparatus for processing non-sequential point cloud media, comprising: a processing unit and a communication unit; the communication unit is configured to: receive MPD signaling of at least one non-sequential point cloud media, the non-sequential point cloud media The media includes the identifier of the static object; according to the MPD signaling, a first request message is sent to the video production device, where the first request message is used to request the first non-sequential point in the at least one non-sequential point cloud media cloud media; receiving the first non-sequential point cloud media from the video production device; the processing unit is configured to play the first non-sequential point cloud media; wherein, at least one non-sequential point cloud media is compressed GPCC area for at least one point cloud The entry is obtained by encapsulating and generating, and the entry of the at least one GPCC area is generated by encapsulating the GPCC bit stream, and the GPCC bit stream is obtained by processing the non
- a video production apparatus comprising: a processor and a memory for storing a computer program, the processor for invoking and running the computer program stored in the memory to perform the method of the above aspect.
- a video playback device comprising: a processor and a memory, where the memory is used for storing a computer program, and the processor is used for calling and running the computer program stored in the memory to execute the method of the above aspect.
- a computer-readable storage medium for storing a computer program, the computer program causing a computer to perform the method of the above aspect.
- the embodiments of the present application provide a computer program product including instructions, which, when executed on a computer, cause the computer to perform the method of the above aspect.
- the identification of the static object can be carried in the non-sequential point cloud media, so that the user can divide it multiple times and request the same Non-temporal point cloud media of static objects to improve user experience.
- the 3D space area corresponding to the entry of the GPCC area can be divided into multiple subspace areas. Combined with the independent encoding and decoding characteristics of GPCC tiles, users can decode and present non-sequential point cloud media more efficiently. delay lower.
- the video production device can flexibly combine the entries of multiple GPCC regions to form different non-sequential point cloud media, wherein the non-sequential point cloud media can constitute a complete GPCC frame or a partial GPCC frame.
- the flexibility of video production can be improved.
- FIG. 1 shows a schematic diagram of the architecture of a processing system for non-sequential point cloud media provided by an exemplary embodiment of the present application
- FIG. 2A shows a schematic structural diagram of a processing architecture of a non-sequential point cloud media provided by an exemplary embodiment of the present application
- FIG. 2B shows a schematic structural diagram of a sample provided by an exemplary embodiment of the present application
- FIG. 2C shows a schematic structural diagram of a container including multiple file tracks provided by an exemplary embodiment of the present application
- FIG. 2D shows a schematic structural diagram of a sample provided by another exemplary embodiment of the present application.
- FIG. 3 shows an interactive flowchart of a method for processing non-sequential point cloud media provided by an embodiment of the present application
- FIG. 4A shows a schematic diagram of encapsulation of a point cloud media provided by an embodiment of the present application
- FIG. 4B shows a schematic diagram of encapsulation of another point cloud media provided by an embodiment of the present application
- FIG. 5 shows a schematic diagram of an apparatus 500 for processing non-sequential point cloud media provided by an embodiment of the present application
- FIG. 6 shows a schematic diagram of an apparatus 600 for processing non-sequential point cloud media provided by an embodiment of the present application
- FIG. 7 shows a schematic block diagram of a video production device 700 provided by an embodiment of the present application.
- FIG. 8 shows a schematic block diagram of a video playback device 800 provided by an embodiment of the present application.
- Point cloud data is a specific recording form of point cloud.
- the point cloud data of each point in the point cloud can include geometric information (ie, three-dimensional position information) and attribute information.
- the geometry of each point in the point cloud The information refers to the Cartesian three-dimensional coordinate data of the point, and the attribute information of each point in the point cloud may include, but is not limited to, at least one of the following: color information, material information, and laser reflection intensity information.
- each point in the point cloud has the same amount of attribute information; for example, each point in the point cloud has both color information and laser reflection intensity; or each point in the point cloud has color information
- the way to obtain point cloud data may include but not limited to at least one of the following : (1) Generated by computer equipment.
- the computer device can generate point cloud data according to the virtual three-dimensional object and the virtual three-dimensional scene.
- the visual scene of the real world is acquired through a 3D photography device (ie, a set of cameras or a camera device with multiple lenses and sensors) to obtain point cloud data of the visual scene of the real world, and dynamic real-world three-dimensional objects can be obtained through 3D photography. or point cloud data of a 3D scene.
- a 3D photography device ie, a set of cameras or a camera device with multiple lenses and sensors
- point cloud data of a 3D scene (4) Obtain point cloud data of biological tissues and organs through medical equipment.
- medical equipment such as Magnetic Resonance Imaging (MRI), Computed Tomography (CT), and electromagnetic positioning information.
- point cloud media refers to point cloud media files formed by point cloud data.
- the point cloud media includes multiple media frames, and each media frame in the point cloud media is composed of point cloud data.
- Point cloud media can express the spatial structure and surface properties of 3D objects or 3D scenes flexibly and conveniently, so it is widely used in virtual reality (Virtual Reality, VR) games, computer aided design (Computer Aided Design, CAD), geographic information systems ( Geography Information System, GIS), Automatic Navigation System (ANS), digital cultural heritage, free viewpoint broadcasting, 3D immersive telepresence, 3D reconstruction of biological tissues and organs, etc.
- VR Virtual Reality
- CAD Computer Aided Design
- GIS Geographic Information System
- ANS Automatic Navigation System
- non-sequential point cloud media is aimed at the same static object, that is, for the same static object, its corresponding point cloud media is non-sequential.
- FIG. 1 shows a schematic diagram of the architecture of a non-sequential point cloud media processing system provided by an exemplary embodiment of the present application.
- the non-sequential point cloud media processing system 10 includes video playback.
- Device 101 and Video Production Device 102.
- the video production equipment refers to the computer equipment used by the provider of the non-sequential point cloud media (for example, the content producer of the non-sequential point cloud media).
- Smart mobile devices such as smart phones, etc.), servers, etc.
- video playback devices refer to computer devices used by users (such as users) of non-sequential point cloud media, which can be terminals (such as PCs), smart mobile devices devices (e.g. smartphones), VR devices (e.g. VR headsets, VR glasses, etc.).
- the video production device and the video playback device may be directly or indirectly connected through wired communication or wireless communication, which is not limited in this embodiment of the present application.
- FIG. 2A shows a schematic structural diagram of a processing architecture for non-sequential point cloud media provided by an exemplary embodiment of the present application.
- the following will combine the processing system for non-sequential point cloud media shown in FIG. 1 and the processing system shown in FIG. 2A .
- the processing architecture of the non-sequential point cloud media, the processing solution of the non-sequential point cloud media provided by the embodiment of the present application is introduced.
- the processing process of the non-sequential point cloud media includes the processing process on the video production device side and the processing process on the video playback device side. , the specific processing process is as follows:
- the acquisition method of point cloud data can be divided into two methods: acquiring point cloud data by collecting real-world visual scenes by a capture device, and generating by computer equipment.
- the capture device may be a hardware component provided in the video production device, for example, the capture device is a camera, a sensor, or the like of a terminal.
- the capture device may also be a hardware device connected to the content production device, such as a camera connected to a server. The capture device is used to provide the point cloud data acquisition service for the video production device.
- the capture device may include but is not limited to any of the following: a camera device, a sensor device, and a scanning device; wherein, the camera device may include a common camera, a stereo camera, Light field cameras, etc.; sensing devices may include laser devices, radar devices, etc.; scanning devices may include 3D laser scanning devices, etc.
- the number of capture devices can be multiple, and these capture devices are deployed in some specific positions in the real space to capture point cloud data from different angles in the space at the same time, and the captured point cloud data are synchronized in time and space.
- the computer device may generate point cloud data according to the virtual three-dimensional object and the virtual three-dimensional scene. Due to the different acquisition methods of point cloud data, the corresponding compression coding methods of point cloud data acquired by different methods may also be different.
- the video production device adopts a geometry-based point cloud compression (Geometry-Based Point Cloud Compression, GPCC) encoding method or a traditional video encoding-based point cloud compression (Video-Based Point Cloud Compression, VPCC) encoding method to obtain the
- GPCC geometry-based point cloud compression
- VPCC Video-Based Point Cloud Compression
- the video production device uses a file track to encapsulate the GPCC bit stream of the encoded point cloud data; the so-called file track refers to the GPCC bit stream of the encoded point cloud data.
- Encapsulation container GPCC bitstreams can be encapsulated in a single file track, GPCC bitstreams can also be encapsulated into multiple file tracks, GPCC bitstreams are encapsulated in a single file track and GPCC bitstreams are encapsulated in multiple file tracks. as follows:
- the GPCC bitstream is encapsulated in a single file track.
- the GPCC bitstream is required to be declared and represented according to the transport rules of the single file track.
- GPCC bitstreams encapsulated in a single file track do not require further processing and can be encapsulated by the International Organization for Standardization Base Media File Format (ISOBMFF).
- ISOBMFF International Organization for Standardization Base Media File Format
- each sample (Sample) packaged in a single file track contains one or more GPCC components, which are also called GPCC components, and the GPCC components may be GPCC geometric components or GPCC attribute components.
- the so-called sample refers to a set of encapsulation structures of one or more point clouds, that is, each sample consists of one or more Type-Length-Value ByteStream Format (TLV) encapsulation structures .
- Fig. 2B shows a schematic structural diagram of a sample provided by an exemplary embodiment of the present application. As shown in Fig. 2B, when a single file track is transmitted, the samples in the file track are composed of GPCC parameter set TLV, geometric bit stream Composed of TLV and attribute bitstream TLV, the sample is packed into a single file track.
- the GPCC bitstream is encapsulated in multiple file tracks.
- each sample in the file track contains at least one TLV encapsulation structure that carries a single GPCC component data, And the TLV encapsulation structure does not contain the encoded GPCC geometry bit stream and the encoded GPCC attribute bit stream at the same time.
- FIG. 2C shows a schematic structural diagram of a container including multiple file tracks provided by an exemplary embodiment of the present application. As shown in FIG.
- the package 1 transmitted in the file track 1 includes an encoded GPCC geometric bit stream , does not contain the coded GPCC attribute bitstream; the package 2 transmitted in the file track 2 contains the coded GPCC attribute bitstream and does not contain the coded GPCC geometry bitstream. Since the video playback device should first decode the encoded GPCC geometry bitstream when decoding, and the decoding of the encoded GPCC attribute bitstream depends on the decoded geometry information, different GPCC component bitstreams are encapsulated in separate file tracks , so that the video playback device can access the file track carrying the encoded GPCC geometry bitstream before the encoded GPCC attribute bitstream. Fig.
- FIG. 2D shows a schematic structural diagram of a sample provided by another exemplary embodiment of the present application.
- the encoded GPCC geometry bit stream and the encoded GPCC attribute bit stream Transmission is performed in different file tracks.
- the samples in the file track are composed of GPCC parameter set TLV and geometric bitstream TLV.
- the sample does not contain attribute bitstream TLV.
- the sample is encapsulated in any one of the multiple file tracks. in orbit.
- the acquired point cloud data is encoded and packaged by the video production device to form non-sequential point cloud media
- the non-sequential point cloud media may be the entire media file of the object, or may be a media segment of the object
- the video production device uses the media presentation description information (that is, the description signaling file) (Media presentation description, MPD) according to the file format requirements of the non-sequential point cloud media to record the metadata of the encapsulation file of the non-sequential point cloud media.
- Metadata is a general term for information related to the presentation of non-sequential point cloud media.
- the metadata may include description information for non-sequential point cloud media, description information for viewing windows, and signaling related to the presentation of non-sequential point cloud media. information and more.
- the video production device delivers the MPD to the video playback device, so that the video playback device requests to acquire point cloud media according to the relevant description information in the MPD.
- point cloud media and MPD are delivered by the video production device to the video playback device through a transmission mechanism (such as Dynamic Adaptive Streaming over HTTP (DASH), Smart Media Transport (SMT)) .
- DASH Dynamic Adaptive Streaming over HTTP
- SMT Smart Media Transport
- the video playback device may obtain non-sequential point cloud media through MPD signaling delivered by the video production device.
- the file decapsulation process on the video playback device side is opposite to the file encapsulation process on the video production device side.
- the video playback device decapsulates the encapsulated files of the non-sequential point cloud media according to the file format requirements of the non-sequential point cloud media, and obtains the encoding Bitstream (ie GPCC bitstream or VPCC bitstream).
- the decoding process on the video playback device side is opposite to the encoding process on the video production device side.
- the video playback device decodes the encoded bit stream and restores the point cloud data.
- the video playback device renders the point cloud data obtained by decoding the GPCC bitstream according to the metadata related to rendering and viewport in the MPD. After rendering, the visual scene corresponding to the point cloud data is presented.
- the point cloud data for the same object can be encapsulated into different point cloud media, for example: some point cloud media are the entire point cloud media of the object, and some point cloud media are part of the object's point cloud media.
- the user can request to play different point cloud media.
- the user requests he does not know whether the different point cloud media is the point cloud media of the same object, resulting in blind requests. This problem also exists for non-temporal point cloud media of static objects.
- the present application carries the identification of the static object in the non-sequential point cloud media, so that the user can request the non-sequential point cloud media of the same static object for multiple times and purposefully.
- FIG. 3 is an interactive flowchart of a method for processing non-sequential point cloud media provided by an embodiment of the present application.
- the main body of the method is a video production device and a video playback device. As shown in FIG. 3 , the method includes the following steps:
- the video production device obtains the non-sequential point cloud data of the static object.
- the video production device processes the non-sequential point cloud data through GPCC encoding to obtain a GPCC bit stream.
- the video production device encapsulates the GPCC bit stream to generate at least one entry of the GPCC area.
- the video production device encapsulates the entry of at least one GPCC area, and generates at least one non-sequential point cloud medium of the static object, and each non-sequential point cloud medium includes an identifier of the static object.
- the video production device sends at least one MPD signaling of non-sequential point cloud media to the video playback device.
- the video playback device sends a first request message.
- the first request message is sent by the video playback device according to MPD signaling, and the first request message is used to request the first non-sequential point cloud media in the at least one non-sequential point cloud media.
- the video production device sends the first non-sequential point cloud media to the video playback device according to the first request message.
- the video playback device plays the first non-sequential point cloud media.
- the entry of the GPCC area is used to represent the GPCC component of the 3D space area corresponding to the GPCC area.
- Each GPCC area corresponds to a 3D space area of the above-mentioned static object, and the 3D space area may be the whole or part of the 3D space area of the static object.
- a GPCC component is also called a GPCC component, and the GPCC component can be a GPCC geometry component or an attribute component.
- the identifier of the static object can be defined by the following code:
- ObjectInfoProperty indicates the property of the content corresponding to the entry, and both the GPCC geometric component and the property component can contain this property. If only the GPCC geometry component contains this attribute, the ObjectInfoProperty of all attribute components associated with the GPCC geometry component is the same.
- the object_ID indicates the identification of the static object, and the object_IDs of the entries in different GPCC areas of the same static object are the same.
- the identifier of the above-mentioned static object can be carried in the entry related to the GPCC geometric component in the point cloud medium, or carried in the entry related to the GPCC attribute component in the point cloud medium, or carried in the point cloud medium.
- Items related to GPCC geometric components and items related to GPCC attribute components are not limited in this application.
- FIG. 4A is a schematic diagram of encapsulation of a point cloud media provided by an embodiment of the present application.
- the point cloud media includes items related to GPCC geometric components and items related to GPCC attribute components. Among them, these items can be associated through the GPCC item group box in the point cloud media.
- items related to GPCC geometric components are associated with items related to GPCC attribute components.
- the entries related to the GPCC geometric components may include the following entry attributes: such as GPCC configuration (GPCC Configuration), 3D spatial region attributes (3D spatial region or ItemSpatialInfoProperty), and the identification of static objects.
- Items related to GPCC attribute components may include the following item attributes: such as GPCC Configuration (GPCC Configuration), identification of static objects, and the like.
- the GPCC configuration indicates the configuration information of the decoder required to decode the corresponding entry and information related to each GPCC component, but is not limited thereto.
- the items related to the GPCC attribute components may also include: 3D space area attributes, which are not limited in this application.
- FIG. 4B is a schematic diagram of encapsulation of another point cloud media provided by an embodiment of the present application.
- the point cloud media includes: an entry related to a GPCC geometric component , and the entry is associated with two GPCC attribute component related entries.
- FIG. 4A For the other attributes included in the entries related to the GPCC geometric components, and the attributes included in the entries related to the GPCC attribute components, reference may be made to FIG. 4A , which will not be repeated in this application.
- the identifier of the above-mentioned static object is not limited to be carried in the attribute corresponding to the entry of each GPCC area.
- the non-sequential point cloud media may be the whole or part of the point cloud media of the static object.
- the video playback device may send the first request message according to the above MPD signaling to request the first non-sequential point cloud media.
- the identification of the static object can be carried in the non-sequential point cloud media, so that the user can divide it multiple times and request the same Non-temporal point cloud media of static objects to improve user experience.
- each GPCC area corresponds to only one 3D space area, but in this application, the 3D space area can be further divided.
- the signaling has been updated accordingly, as follows:
- the entry of the target GPCC area includes: a 3D space area entry attribute, and the 3D space area entry attribute includes: a first identifier and a second identifier.
- the target GPCC area is one GPCC area in at least one GPCC area.
- the first identifier (Sub_region_contained) is used to identify whether the target 3D space region corresponding to the target GPCC region is divided into multiple subspace regions.
- the second identifier (tile_id_present) is used to identify whether the target GPCC area adopts the GPCC tile coding mode.
- tile_id_present 1
- the video production end must use the GPCC tile encoding method.
- the 3D space area entry attributes also include, but are not limited to: the respective information of the multiple subspace areas and the information of the target 3D space area. .
- the information of the subspace region includes at least one of the following, but is not limited to this: the identification of the subspace region, the location information of the subspace region, the target GPCC region adopts When GPCC tile is encoded, the tile (block) identifier in the subspace area.
- the location information of the subspace region includes, but is not limited to, the location information of an anchor point of the subspace region, and the lengths of the subspace region along the X axis, the Y axis, and the Z axis, respectively.
- the location information of the subspace region includes, but is not limited to, the location information of two anchor points of the subspace region.
- the information of the target 3D space area includes at least one of the following, but is not limited thereto: an identifier of the target 3D space area, location information of the target 3D space area, and the number of subspace areas included in the target 3D space area.
- the position information of the target 3D space area includes, but is not limited to: the position information of an anchor point of the target 3D space area, and the target 3D space area along the X-axis, Y-axis, Z-axis respectively. length.
- the location information of the target 3D space area includes, but is not limited to: location information of two anchor points of the target 3D space area.
- the 3D space region entry attribute further includes: a third identifier (initial_region_id).
- a third identifier initial_region_id.
- the value of the third identifier is the first value or empty, it indicates that the entry corresponding to the target GPCC area is the entry initially presented by the video playback device, and for the target 3D space area and the subspace area of the target 3D space area, when the video playback The device initially renders the target 3D space area.
- the value of the third identifier is the second value, it indicates that the entry corresponding to the target GPCC area is an entry initially presented by the video playback device, for the target 3D space area and the subspace area of the target 3D space area, at the initial stage of the video playback device What is presented is the subspace area corresponding to the second value in the target 3D space area.
- the above-mentioned first numerical value is 0, and the second numerical value is an identifier of a subspace area in the target 3D space area that needs to be initially presented.
- the 3D space region entry attribute further includes: information of the target 3D space region.
- the information of the target 3D space area includes at least one of the following items, but is not limited to this: the identification of the target 3D space area, the location information of the target 3D space area, and when the target GPCC area adopts GPCC tile encoding, the target 3D space area is in the target 3D space area.
- the tile identifier is not limited to this: the identification of the target 3D space area, the location information of the target 3D space area, and when the target GPCC area adopts GPCC tile encoding, the target 3D space area is in the target 3D space area.
- ItemSpatialInfoProperty represents the 3D spatial area property of the item of the GPCC area. If the entry is an entry corresponding to a geometric component, the attribute must be included; if the entry is an entry corresponding to an attribute component, the 3D space area attribute may not be included.
- sub_region_contained is 1, indicating that the 3D space region can be further divided into multiple subspace regions.
- tile_id_present When the value of this field is 1, the value of tile_id_present must be 1.
- the sub_region_contained value is 0, indicating that there is no further subspace region division in the 3D space.
- a value of 1 for tile_id_present indicates that the non-sequential point cloud data is encoded by GPCC tiles, and the tile id corresponding to the non-sequential point cloud is given in this attribute.
- inital_region_id indicates the ID of the space region that is initially presented in the overall space of the item when the current item is an item that is initially consumed or played. If the value of this field is 0 or the field does not exist, the area in which the entry is initially presented is the overall 3D space area. If the value of this field is the identifier of the subspace area, the area initially presented by the entry is the subspace area corresponding to the identifier.
- 3DSpatialRegionStruct represents a 3D space region
- the first 3DSpatialRegionStruct in the ItemSpatialInfoProperty indicates the 3D space region corresponding to the entry corresponding to the ItemSpatialInfoProperty
- the remaining 3DSpatialRegionStruct indicates each subspace region in the 3D space region corresponding to the entry.
- num_sub_regions indicates the number of subspace regions divided in the 3D space region corresponding to the entry.
- num_tiles indicates the number of tiles in the 3D space region corresponding to this entry, or the number of tiles corresponding to its subspace region.
- tile_id indicates the identifier of the GPCC tile.
- anchor_x, anchor_y, and anchor_z respectively represent the x, y, and z coordinates of the anchor point of the 3D space region or the subspace region of the region.
- region_dx, region_dy, and region_dz respectively represent the lengths of the 3D space region or the subspace regions of the region along the X axis, the Y axis, and the Z axis, respectively.
- the 3D space area can be divided into multiple subspace areas. Combined with the independent encoding and decoding characteristics of GPCC tiles, users can decode and present non-sequential point cloud media with higher efficiency and lower latency.
- the video production device may encapsulate the entry of at least one GPCC area to generate at least one non-sequential point cloud media of the static object.
- the entry of at least one GPCC area is one
- the entry of one GPCC area is encapsulated into one non-sequential point cloud media.
- the number of entries of at least one GPCC area is N
- the entries of the N GPCC areas are encapsulated into M non-sequential point cloud media.
- N is an integer greater than 1
- the value range of M is [1, N]
- M is an integer.
- the entries of the N GPCC areas can be encapsulated into one non-sequential point cloud medium.
- N non-sequential point cloud media In this case of encapsulation, each non-sequential point cloud media includes one entry.
- the second non-sequential point cloud media is any one of the at least one non-sequential point cloud media including entries of multiple GPCC regions.
- the second non-sequential point cloud media includes:
- GPCC item group box (GPCCItemGroupBox).
- the GPCC entry group box is used to associate entries of multiple GPCC areas, as shown in Figures 4A and 4B.
- the GPCC entry group box includes: identifiers of entries of multiple GPCC areas.
- the GPCC item group box includes: a fourth identification (initial_item_ID).
- the fourth identifier is the identifier of the entry initially presented by the video playback device in the entries of the multiple GPCC areas.
- the GPCC item group box includes: a fifth flag (partial_item_flag). If the fifth identifier takes the value of the third value, it indicates that the entries of the multiple GPCC areas constitute a complete GPCC frame of the static object. If the value of the fifth identifier is the fourth value, it indicates that the entries of the multiple GPCC areas constitute part of the GPCC frame of the static object.
- the third value may be 0, and the fourth value may be 1, but not limited thereto.
- the GPCC entry group box includes: location information of a GPCC area formed by a plurality of GPCC areas.
- the GPCC entry group box includes location information of the R1+R2 area.
- the items contained in the GPCCItemGroupBox are items that belong to the same static object, and items that have an associated relationship when presenting and consuming. All items contained in the GPCCItemGroupBox may constitute a complete GPCC frame, or may be part of a GPCC frame.
- initial_item_ID indicates the identification of the item initially consumed within an item group.
- the initial_item_ID is only valid when the current item group is the item group requested by the user for the first time.
- the same static object corresponds to two point cloud media, which are F1 and F2 respectively.
- partial_item_flag value When the partial_item_flag value is 0, it means that all items contained in the GPCCItemGroupBox and their associated items constitute a complete GPCC frame, and when the value is 1, it means that all the items contained in the GPCCItemGroupBox and their associated items constitute only a partial GPCC frame.
- the extension is as follows:
- the GPCC entry descriptor is used to describe the elements and attributes related to the GPCC entry, and the descriptor is a SupplementalProperty element.
- @schemeIdUri attribute is equal to "urn:mpeg:mpegI:gpcc:2020:gpsr".
- the descriptor can be located at the Adaptation Set level or the Representation level.
- Representation In DASH, a combination of one or more media components, such as a video file of a certain resolution, can be regarded as a Representation (description).
- Adaptation Sets In DASH, a collection of one or more video streams.
- One Adaptation Sets can contain multiple Representations.
- Table 1 GPCC entry description sub-elements and attributes
- the video production equipment can flexibly combine the entries of multiple GPCC regions to form different non-sequential point cloud media, wherein the non-sequential point cloud media can constitute a complete GPCC frame or a part of it. GPCC frame.
- the video production device can also improve the entries of the initial presentation.
- the non-sequential point cloud data has 4 versions of point cloud media on the video production equipment side: point cloud media F0 corresponding to all non-sequential point cloud data, corresponding to Some point cloud media F1 to F3 of non-sequential point cloud data, wherein F1 to F3 correspond to 3D space regions R1 to R3 respectively.
- point cloud media package contents of F0 ⁇ F3 are as follows:
- tile_id[] (3,4)
- inital_region_id 0;
- inital_region_id 0;
- tile_id[] (3,4)
- the video production device sends the MPD signaling of F0 to F3 to the user, and the Object_ID, space region, subspace region, and tile identification information therein are the same as those in the file encapsulation, and are not repeated here.
- user U1 Since user U1 has good network conditions and low data transmission delay, it can request F0; user U2 has poor network conditions and high data transmission delay, so it can request F1.
- the video production device transmits F0 to the video playback device corresponding to the user U1, and transmits F1 to the video playback device corresponding to the user U2.
- the initial viewing area is the SR1 area, and the corresponding tile ID is 1.
- U1 decodes and consumes it can decode tile '1' from the overall code stream for direct consumption and presentation, without the need to decode the overall file and present it, which improves the decoding efficiency and reduces the time required for rendering and presentation.
- the corresponding tile ID is 2, and the part corresponding to tile '2' in the overall code stream is directly decoded for presentation and consumption.
- the video playback device corresponding to user U2 After the video playback device corresponding to user U2 receives F1, it decodes F1 for consumption, and according to the area that the user may consume in the next step, combined with the information in the MPD file, that is, Object_ID and spatial area information, requests F2 or F3 for caching in advance.
- the video playback device can also make a purposeful request to the video production device again based on the user's consumption needs and possible consumption areas.
- Non-temporal point cloud media of the same static object can also make a purposeful request to the video production device again based on the user's consumption needs and possible consumption areas.
- the video playback device Since the video playback device obtains the identification of the static object in the point cloud media above, when it needs to obtain other point cloud media corresponding to the static object again, it can request the same static object multiple times in a targeted manner based on the identification of the static object.
- the object's point cloud media Since the video playback device obtains the identification of the static object in the point cloud media above, when it needs to obtain other point cloud media corresponding to the static object again, it can request the same static object multiple times in a targeted manner based on the identification of the static object.
- the object's point cloud media Since the video playback device obtains the identification of the static object in the point cloud media above, when it needs to obtain other point cloud media corresponding to the static object again, it can request the same static object multiple times in a targeted manner based on the identification of the static object.
- the object's point cloud media Since the video playback device obtains the identification of the static object in the point cloud media above, when it needs to obtain other point cloud media corresponding to the static object
- the non-sequential point cloud data exists in two versions of point cloud media on the video production device: F1 and F2, F1 contains item1 ⁇ item2, and F2 contains item3 to item4.
- the point cloud media package contents of F1 and F2 are as follows:
- inital_region_id 0;
- inital_region_id 0;
- GPCCItemGroupBox
- inital_region_id 0;
- inital_region_id 0;
- GPCCItemGroupBox
- the video production device sends the MPD signaling of F1 to F2 to the user, and the Object_ID, spatial area, and tile ID information are the same as those in the point cloud media encapsulation, and are not repeated here.
- User U1 requests F1 consumption; user U2 requests F2 consumption.
- the video production device transmits F1 to the video playback device corresponding to the user U1, and transmits F2 to the video playback device corresponding to the user U2.
- the video playback device corresponding to U1 After the video playback device corresponding to U1 receives F1, it initially watches item1.
- the initial viewing area of item1 is the entire viewing space of item1. Therefore, U1 consumes the entirety of item1. Since F1 contains item1 and item2, which correspond to tile1 and tile2 respectively, U1 can directly decode part of the code stream corresponding to tile1 for presentation when consuming item1. If U1 continues to consume, and the corresponding tile ID is 2 when viewing the item2 area, it will directly decode the part corresponding to tile'2' in the overall code stream for presentation and consumption. If U1 continues to consume and needs to watch the area corresponding to item3, it will request F2 according to the MPD file. After receiving F2, present consumption directly according to the area viewed by the user, and no longer judge the initial consumption item information and initial viewing area information in F2.
- the video playback device corresponding to U2 After the video playback device corresponding to U2 receives F2, it initially watches item3, and the initial viewing area of item3 is the overall viewing space of item3, so U2 consumes the entire item3. Since F2 contains item3 and item4, which correspond to tile3 and tile4 respectively, U2 can directly decode part of the code stream corresponding to tile3 for presentation when consuming item3.
- FIG. 5 is a schematic diagram of an apparatus 500 for processing non-sequential point cloud media according to an embodiment of the present application.
- the apparatus 500 includes a processing unit 510 and a communication unit 520 .
- the processing unit 510 is configured to: acquire non-sequential point cloud data of the static object.
- the non-sequential point cloud data is processed through the GPCC encoding method to obtain the GPCC bit stream.
- the GPCC bitstream is encapsulated to generate at least one entry of the GPCC area, where the entry of the GPCC area is used to represent the GPCC component of the three-dimensional 3D space area corresponding to the GPCC area.
- the communication unit 520 is configured to: receive a first request message sent by the video playback device according to the MPD signaling, where the first request message is used to request a first non-sequential point cloud media in the at least one non-sequential point cloud media . According to the first request message, the first non-sequential point cloud media is sent to the video playback device.
- the entry of the target GPCC area includes: a 3D space area entry attribute, and the 3D space area entry attribute includes: a first identifier and a second identifier.
- the target GPCC area is one GPCC area in at least one GPCC area.
- the first identifier is used to identify whether the target 3D space area corresponding to the target GPCC area is divided into multiple subspace areas.
- the second identifier is used to identify whether the target GPCC area adopts the GPCC tile encoding method.
- the 3D space region entry attribute further includes: respective information of the multiple subspace regions and information of the target 3D space region.
- the information of the subspace region includes at least one of the following: the identification of the subspace region, the location information of the subspace region, and when the target GPCC region adopts GPCC tile coding, The tile identifier in the subspace region.
- the information of the target 3D space area includes at least one of the following items: an identifier of the target 3D space area, location information of the target 3D space area, and the number of subspace areas included in the target 3D space area.
- the 3D space area entry attribute further includes: a third identifier.
- the value of the third identifier is the first value or empty, it indicates that the entry corresponding to the target GPCC area is the entry initially presented by the video playback device, and for the target 3D space area and the subspace area of the target 3D space area, when the video playback The device initially renders the target 3D space area.
- the value of the third identifier is the second value, it indicates that the entry corresponding to the target GPCC area is an entry initially presented by the video playback device, for the target 3D space area and the subspace area of the target 3D space area, at the initial stage of the video playback device What is presented is the subspace area corresponding to the second value in the target 3D space area.
- the 3D space region entry attribute further includes: information of the target 3D space region.
- the information of the target 3D space area includes at least one of the following: an identifier of the target 3D space area, location information of the target 3D space area, and a tile identifier in the target 3D space area when the target GPCC area adopts GPCC tile encoding.
- the processing unit 510 is specifically configured to: if the entry of at least one GPCC area is one, encapsulate the entry of one GPCC area into one non-sequential point cloud media. If the number of entries of at least one GPCC area is N, the entries of the N GPCC areas are encapsulated into M non-sequential point cloud media.
- N is an integer greater than 1, 1 ⁇ M ⁇ N, and M is an integer.
- the second non-sequential point cloud media includes: a GPCC entry group box.
- the second non-sequential point cloud media is any non-sequential point cloud media including entries of multiple GPCC areas in at least one non-sequential point cloud media, and the GPCC entry group box is used for associating entries of multiple GPCC areas.
- the GPCC entry group box includes: a fourth identifier.
- the fourth identifier is the identifier of the entry initially presented by the video playback device in the entries of the multiple GPCC areas.
- the GPCC entry group box includes: a fifth identifier. If the fifth identifier is a third value, it indicates that the entries of multiple GPCC areas constitute a complete GPCC frame of the static object. If the value of the fifth identifier is the fourth value, it indicates that the entries of the multiple GPCC areas constitute part of the GPCC frame of the static object.
- the GPCC entry group box includes: location information of a GPCC area formed by a plurality of GPCC areas.
- the communication unit 520 is further configured to: receive a second request message sent by the video playback device based on the identifier of the static object, where the second request message is used to request the at least one non-sequential point cloud media.
- the third non-temporal point cloud media According to the second request message, the third non-sequential point cloud media is sent to the video playback device.
- the apparatus embodiments and the method embodiments may correspond to each other, and similar descriptions may refer to the method embodiments. To avoid repetition, details are not repeated here.
- the apparatus 500 shown in FIG. 5 can execute the method embodiments corresponding to the video production equipment, and the aforementioned and other operations and/or functions of each module in the apparatus 500 are respectively for realizing the method embodiments corresponding to the video production equipment, in order to It is concise and will not be repeated here.
- the apparatus 500 of the embodiments of the present application is described above from the perspective of functional modules with reference to the accompanying drawings.
- the functional modules can be implemented in the form of hardware, can also be implemented by instructions in the form of software, and can also be implemented by a combination of hardware and software modules.
- the steps of the method embodiments in the embodiments of the present application may be completed by hardware integrated logic circuits in the processor and/or instructions in the form of software, and the steps of the methods disclosed in conjunction with the embodiments of the present application may be directly embodied as hardware
- the execution of the decoding processor is completed, or the execution is completed by a combination of hardware and software modules in the decoding processor.
- the software modules may be located in random access memory, flash memory, read-only memory, programmable read-only memory, electrically erasable programmable memory, registers, and other storage media mature in the art.
- the storage medium is located in the memory, and the processor reads the information in the memory, and completes the steps in the above method embodiments in combination with its hardware.
- FIG. 6 is a schematic diagram of an apparatus 600 for processing non-sequential point cloud media according to an embodiment of the present application.
- the apparatus 600 includes a processing unit 610 and a communication unit 620 .
- the communication unit 620 is configured to: receive MPD signaling of at least one non-sequential point cloud media, where the non-sequential point cloud media includes an identifier of the static object.
- a first non-temporal point cloud media is received from the video production device.
- the processing unit 610 is configured to play the first non-sequential point cloud media.
- At least one non-sequential point cloud media is generated by encapsulating at least one entry in the point cloud compressed GPCC area, and the entry in the at least one GPCC area is generated by encapsulating a GPCC bitstream, and the GPCC bitstream is It is obtained by processing the non-sequential point cloud data of static objects through GPCC encoding.
- the entry of the GPCC area is used to represent the GPCC component of the 3D space area corresponding to the GPCC area.
- the entry of the target GPCC area includes: a 3D space area entry attribute, and the 3D space area entry attribute includes: a first identifier and a second identifier.
- the target GPCC area is one GPCC area in at least one GPCC area.
- the first identifier is used to identify whether the target 3D space area corresponding to the target GPCC area is divided into multiple subspace areas.
- the second identifier is used to identify whether the target GPCC area adopts the GPCC tile encoding method.
- the 3D space region entry attribute further includes: respective information of the multiple subspace regions and information of the target 3D space region.
- the information of the subspace region includes at least one of the following: the identification of the subspace region, the location information of the subspace region, and when the target GPCC region adopts GPCC tile coding, The tile identifier in the subspace region.
- the information of the target 3D space area includes at least one of the following items: an identifier of the target 3D space area, location information of the target 3D space area, and the number of subspace areas included in the target 3D space area.
- the 3D space area entry attribute further includes: a third identifier.
- the value of the third identifier is the first value or empty, it indicates that the entry corresponding to the target GPCC area is the entry initially presented by the video playback device, and for the target 3D space area and the subspace area of the target 3D space area, when the video playback The device initially renders the target 3D space area.
- the value of the third identifier is the second value, it indicates that the entry corresponding to the target GPCC area is an entry initially presented by the video playback device, for the target 3D space area and the subspace area of the target 3D space area, at the initial stage of the video playback device What is presented is the subspace area corresponding to the second value in the target 3D space area.
- the 3D space region entry attribute further includes: information of the target 3D space region.
- the information of the target 3D space area includes at least one of the following: an identifier of the target 3D space area, location information of the target 3D space area, and a tile identifier in the target 3D space area when the target GPCC area adopts GPCC tile coding.
- the entry of at least one GPCC area is one, the entry of one GPCC area is encapsulated as one non-sequential point cloud media. If the number of entries of at least one GPCC area is N, then the entries of the N GPCC areas are packaged into M non-sequential point cloud media. Wherein, N is an integer greater than 1, 1 ⁇ M ⁇ N, and M is an integer.
- the second non-sequential point cloud media includes: a GPCC entry group box.
- the second non-sequential point cloud media is any one of the non-sequential point cloud media including entries of multiple GPCC regions in the at least one non-sequential point cloud media.
- the GPCC Entry Group box is used to associate entries from multiple GPCC areas.
- the GPCC entry group box includes: a fourth identifier.
- the fourth identifier is the identifier of the entry initially presented by the video playback device in the entries of the multiple GPCC areas.
- the GPCC entry group box includes: a fifth identification. If the fifth identifier takes the value of the third value, it indicates that the entries of the multiple GPCC areas constitute a complete GPCC frame of the static object. If the value of the fifth identifier is the fourth value, it indicates that the entries of the multiple GPCC areas constitute part of the GPCC frame of the static object.
- the GPCC entry group box includes: location information of a GPCC area formed by a plurality of GPCC areas.
- the communication unit 620 is further configured to send the second request message to the video production device according to the MPD signaling.
- a second non-temporal point cloud media is received.
- processing unit 610 is further configured to play the second non-sequential point cloud media.
- the apparatus embodiments and the method embodiments may correspond to each other, and similar descriptions may refer to the method embodiments. To avoid repetition, details are not repeated here.
- the apparatus 600 shown in FIG. 6 can execute the method embodiment corresponding to the video playback device, and the aforementioned and other operations and/or functions of each module in the apparatus 600 are for realizing the method embodiment corresponding to the video playback device, respectively, in order to It is concise and will not be repeated here.
- the apparatus 600 of the embodiment of the present application is described above from the perspective of functional modules with reference to the accompanying drawings.
- the functional modules can be implemented in the form of hardware, can also be implemented by instructions in the form of software, and can also be implemented by a combination of hardware and software modules.
- the steps of the method embodiments in the embodiments of the present application may be completed by hardware integrated logic circuits in the processor and/or instructions in the form of software, and the steps of the methods disclosed in conjunction with the embodiments of the present application may be directly embodied as hardware
- the execution of the decoding processor is completed, or the execution is completed by a combination of hardware and software modules in the decoding processor.
- the software modules may be located in random access memory, flash memory, read-only memory, programmable read-only memory, electrically erasable programmable memory, registers, and other storage media mature in the art.
- the storage medium is located in the memory, and the processor reads the information in the memory, and completes the steps in the above method embodiments in combination with its hardware.
- FIG. 7 is a schematic block diagram of a video production device 700 provided by an embodiment of the present application.
- the video production apparatus 700 may include:
- a memory 710 and a processor 720 the memory 710 is used to store computer programs and transmit the program codes to the processor 720.
- the processor 720 may call and run a computer program from the memory 710 to implement the method in the embodiments of the present application.
- the processor 720 may be configured to execute the above method embodiments according to the instructions in the computer program.
- the processor 720 may include, but is not limited to:
- DSP Digital Signal Processor
- ASIC Application Specific Integrated Circuit
- FPGA Field Programmable Gate Array
- the memory 710 includes but is not limited to:
- Non-volatile memory may be a read-only memory (Read-Only Memory, ROM), a programmable read-only memory (Programmable ROM, PROM), an erasable programmable read-only memory (Erasable PROM, EPROM), an electrically programmable read-only memory (Erasable PROM, EPROM). Erase programmable read-only memory (Electrically EPROM, EEPROM) or flash memory. Volatile memory may be Random Access Memory (RAM), which acts as an external cache.
- RAM Random Access Memory
- RAM Static RAM
- DRAM Dynamic RAM
- SDRAM Synchronous DRAM
- SDRAM double data rate synchronous dynamic random access memory
- Double Data Rate SDRAM DDR SDRAM
- enhanced SDRAM ESDRAM
- synchronous link dynamic random access memory SLDRAM
- Direct Rambus RAM Direct Rambus RAM
- the computer program can be divided into one or more modules, and the one or more modules are stored in the memory 710 and executed by the processor 720 to complete the steps provided by the present application.
- the one or more modules may be a series of computer program instruction segments capable of performing specific functions, and the instruction segments are used to describe the execution process of the computer program in the video production apparatus.
- the video production equipment may further include:
- a transceiver 730 which can be connected to the processor 720 or the memory 710 .
- the processor 720 can control the transceiver 730 to communicate with other devices, and specifically, can send information or data to other devices, or receive information or data sent by other devices.
- Transceiver 730 may include a transmitter and a receiver.
- the transceiver 730 may further include antennas, and the number of the antennas may be one or more.
- bus system includes a power bus, a control bus and a status signal bus in addition to a data bus.
- FIG. 8 is a schematic block diagram of a video playback device 800 provided by an embodiment of the present application.
- the video playback device 800 may include:
- a memory 810 and a processor 820 the memory 810 is used to store computer programs and transmit the program codes to the processor 820.
- the processor 820 can call and run a computer program from the memory 810 to implement the methods in the embodiments of the present application.
- the processor 820 may be configured to execute the above method embodiments according to the instructions in the computer program.
- the processor 820 may include, but is not limited to:
- the memory 810 includes but is not limited to:
- Volatile memory and/or non-volatile memory may be ROM, PROM, EPROM, EEPROM or flash memory.
- Volatile memory can be RAM, which acts as an external cache.
- many forms of RAM are available, such as SRAM, DRAM, SDRAM, DDR SDRAM, ESDRAM, SLDRAM, and DR RAM.
- the computer program may be divided into one or more modules, and the one or more modules are stored in the memory 810 and executed by the processor 820 to complete the steps provided by the present application.
- the one or more modules may be a series of computer program instruction segments capable of accomplishing specific functions, and the instruction segments are used to describe the execution process of the computer program in the video playback device.
- the video playback device may further include:
- a transceiver 830 which can be connected to the processor 820 or the memory 810 .
- the processor 820 may control the transceiver 830 to communicate with other devices, specifically, may send information or data to other devices, or receive information or data sent by other devices.
- Transceiver 830 may include a transmitter and a receiver.
- the transceiver 830 may further include antennas, and the number of the antennas may be one or more.
- each component in the video playback device is connected through a bus system, wherein the bus system includes a power bus, a control bus and a status signal bus in addition to a data bus.
- the present application also provides a computer storage medium on which a computer program is stored, and when the computer program is executed by a computer, enables the computer to execute the methods of the above method embodiments.
- the embodiments of the present application further provide a computer program product including instructions, when the instructions are executed by a computer, the instructions cause the computer to execute the methods of the above method embodiments.
- the computer program product includes one or more computer instructions.
- the computer may be a general purpose computer, a special purpose computer, a computer network, or other programmable device.
- the computer instructions may be stored on or transmitted from one computer readable storage medium to another computer readable storage medium, for example, the computer instructions may be transmitted over a wire from a website site, computer, server or data center (eg coaxial cable, optical fiber, digital subscriber line (DSL)) or wireless (eg infrared, wireless, microwave, etc.) means to another website site, computer, server or data center.
- the computer-readable storage medium can be any available medium that can be accessed by a computer or a data storage device such as a server, data center, etc. that includes one or more available media integrated.
- the available media may be magnetic media (eg, floppy disk, hard disk, magnetic tape), optical media (eg, digital video disc (DVD)), or semiconductor media (eg, solid state disk (SSD)), and the like.
- the disclosed system, apparatus and method may be implemented in other manners.
- the device embodiments described above are only illustrative.
- the division of the modules is only a logical function division. In actual implementation, there may be other division methods.
- multiple modules or components may be combined or Integration into another system, or some features can be ignored, or not implemented.
- the shown or discussed mutual coupling or direct coupling or communication connection may be through some interfaces, indirect coupling or communication connection of devices or modules, and may be in electrical, mechanical or other forms.
- Modules described as separate components may or may not be physically separated, and components shown as modules may or may not be physical modules, that is, may be located in one place, or may be distributed to multiple network elements. Some or all of the modules may be selected according to actual needs to achieve the purpose of the solution in this embodiment. For example, each functional module in each embodiment of the present application may be integrated into one processing module, or each module may exist physically alone, or two or more modules may be integrated into one module.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Physics & Mathematics (AREA)
- Computer Graphics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Geometry (AREA)
- Software Systems (AREA)
- Computer Security & Cryptography (AREA)
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
- Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
Abstract
Description
Claims (32)
- 一种非时序点云媒体的处理方法,所述方法由视频制作设备执行,所述方法包括:获取静态物体的非时序点云数据;通过基于几何模型的点云压缩GPCC编码方式对所述非时序点云数据进行处理,得到GPCC比特流;对所述GPCC比特流进行封装,生成至少一个GPCC区域的条目,所述GPCC区域的条目用于表示所述GPCC区域对应的三维3D空间区域的GPCC成分;对所述至少一个GPCC区域的条目进行封装,生成所述静态物体的至少一个非时序点云媒体,所述非时序点云媒体包括所述静态物体的标识;向视频播放设备发送所述至少一个非时序点云媒体的媒体演示描述MPD信令;接收所述视频播放设备根据所述MPD信令发送的第一请求消息,所述第一请求消息用于请求所述至少一个非时序点云媒体中的第一非时序点云媒体;根据所述第一请求消息,向所述视频播放设备发送所述第一非时序点云媒体。
- 根据权利要求1所述的方法,目标GPCC区域的条目包括:3D空间区域条目属性,所述3D空间区域条目属性包括:第一标识和第二标识;其中,所述目标GPCC区域为所述至少一个GPCC区域中的一个GPCC区域;所述第一标识用于标识所述目标GPCC区域对应的目标3D空间区域是否被划分为多个子空间区域;所述第二标识用于标识所述目标GPCC区域是否采用GPCC tile编码方式。
- 根据权利要求2所述的方法,若所述目标GPCC区域对应的目标3D空间区域被划分为多个子空间区域,则所述3D空间区域条目属性还包括:所述多个子空间区域各自的信息和所述目标3D空间区域的信息。
- 根据权利要求3所述的方法,针对所述多个子空间区域中的任一个子空间区域,所述子空间区域的信息包括以下至少一项:所述子空间区域的标识、所述子空间区域的位置信息、所述目标GPCC区域采用GPCC tile编码时,所述子空间区域中的tile标识;所述目标3D空间区域的信息包括以下至少一项:所述目标3D空间区域的标识、所述目标3D空间区域的位置信息、所述目标3D空间区域包括的子空间区域的数量。
- 根据权利要求2-4任一项所述的方法,若所述目标GPCC区域对应的目标3D空间区域被划分为多个子空间区域,则所述3D空间区域条目属性还包括:第三标识;当所述第三标识的取值为第一数值或者空时,表示所述目标GPCC区域对应的条目是所述视频播放设备初始呈现的条目时,针对所述目标3D空间区域和所述目标3D空间区域的子空间区域,在所述视频播放设备初始呈现的是所述目标3D空间区域;当所述第三标识的取值为第二数值时,表示所述目标GPCC区域对应的条目是所述视频播放设备初始呈现的条目时,针对所述目标3D空间区域和所述目标3D空间区域的子空间区域,在所述视频播放设备初始呈现的是所述目标3D空间区域中所述第二数值所对应的子空间区域。
- 根据权利要求2所述的方法,若所述目标GPCC区域对应的目标3D空间区域未被划 分为多个子空间区域,则所述3D空间区域条目属性还包括:所述目标3D空间区域的信息。
- 根据权利要求6所述的方法,所述目标3D空间区域的信息包括以下至少一项:所述目标3D空间区域的标识、所述目标3D空间区域的位置信息、所述目标GPCC区域采用GPCC tile编码时,所述目标3D空间区域中的tile标识。
- 根据权利要求1-4任一项所述的方法,所述对所述至少一个GPCC区域的条目进行封装,生成所述静态物体的至少一个非时序点云媒体,包括:若所述至少一个GPCC区域的条目是1个,则将1个GPCC区域的条目封装为1个非时序点云媒体;若所述至少一个GPCC区域的条目是N个,则将N个GPCC区域的条目封装为M个非时序点云媒体;其中,N为大于1的整数,1≤M≤N,M为整数。
- 根据权利要求1-4任一项所述的方法,第二非时序点云媒体包括:GPCC条目组盒子;其中,所述第二非时序点云媒体是所述至少一个非时序点云媒体中包括多个GPCC区域的条目的任一个非时序点云媒体,所述GPCC条目组盒子用于关联所述多个GPCC区域的条目。
- 根据权利要求9所述的方法,所述GPCC条目组盒子包括:第四标识;其中,所述第四标识是所述多个GPCC区域的条目中,在所述视频播放设备初始呈现的条目的标识。
- 根据权利要求9所述的方法,所述GPCC条目组盒子包括:第五标识;若所述第五标识取值为第三数值,则表示所述多个GPCC区域的条目构成所述静态物体的完整GPCC帧;若所述第五标识取值为第四数值,则表示所述多个GPCC区域的条目构成所述静态物体的部分GPCC帧。
- 根据权利要求9所述的方法,所述GPCC条目组盒子包括:所述多个GPCC区域构成的GPCC区域的位置信息。
- 根据权利要求1-4任一项所述的方法,在所述根据所述第一请求消息,向所述视频播放设备发送所述第一非时序点云媒体之后,所述方法还包括:接收所述视频播放设备基于所述静态物体的标识所发送的第二请求消息,所述第二请求消息用于请求所述至少一个非时序点云媒体中的第三非时序点云媒体;根据所述第二请求消息,向所述视频播放设备发送所述第三非时序点云媒体。
- 一种非时序点云媒体的处理方法,所述方法由视频播放设备执行,所述方法包括:接收至少一个非时序点云媒体的MPD信令,所述非时序点云媒体包括所述静态物体的标识;根据所述MPD信令向视频制作设备发送第一请求消息,所述第一请求消息用于请求所述至少一个非时序点云媒体中的第一非时序点云媒体;从所述视频制作设备接收所述第一非时序点云媒体;播放所述第一非时序点云媒体;其中,所述至少一个非时序点云媒体是对至少一个点云压缩GPCC区域的条目进行封装生成得到的,所述至少一个GPCC区域的条目是对GPCC比特流进行封装生成的,所述GPCC比特流是通过GPCC编码方式对静态物体的非时序点云数据进行处理得到的;针对所述至少一个GPCC区域的条目中的任一个GPCC区域的条目,所述GPCC区域的条目用于表示所述GPCC区域对应的3D空间区域的GPCC成分。
- 根据权利要求14所述的方法,目标GPCC区域的条目包括:3D空间区域条目属性,所述3D空间区域条目属性包括:第一标识和第二标识;其中,所述目标GPCC区域为所述至少一个GPCC区域中的一个GPCC区域;所述第一标识用于标识所述目标GPCC区域对应的目标3D空间区域是否被划分为多个子空间区域;所述第二标识用于标识所述目标GPCC区域是否采用GPCC tile编码方式。
- 根据权利要求15所述的方法,若所述目标GPCC区域对应的目标3D空间区域被划分为多个子空间区域,则所述3D空间区域条目属性还包括:所述多个子空间区域各自的信息和所述目标3D空间区域的信息。
- 根据权利要求16所述的方法,针对所述多个子空间区域中的任一个子空间区域,所述子空间区域的信息包括以下至少一项:所述子空间区域的标识、所述子空间区域的位置信息、所述目标GPCC区域采用GPCC tile编码时,所述子空间区域中的tile标识;所述目标3D空间区域的信息包括以下至少一项:所述目标3D空间区域的标识、所述目标3D空间区域的位置信息、所述目标3D空间区域包括的子空间区域的数量。
- 根据权利要求15-17任一项所述的方法,若所述目标GPCC区域对应的目标3D空间区域被划分为多个子空间区域,则所述3D空间区域条目属性还包括:第三标识;当所述第三标识的取值为第一数值或者空时,表示所述目标GPCC区域对应的条目是所述视频播放设备初始呈现的条目时,针对所述目标3D空间区域和所述目标3D空间区域的子空间区域,在所述视频播放设备初始呈现的是所述目标3D空间区域;当所述第三标识的取值为第二数值时,表示所述目标GPCC区域对应的条目是所述视频播放设备初始呈现的条目时,针对所述目标3D空间区域和所述目标3D空间区域的子空间区域,在所述视频播放设备初始呈现的是所述目标3D空间区域中所述第二数值所对应的子空间区域。
- 根据权利要求15所述的方法,若所述目标GPCC区域对应的目标3D空间区域未被划分为多个子空间区域,则所述3D空间区域条目属性还包括:所述目标3D空间区域的信息。
- 根据权利要求19所述的方法,所述目标3D空间区域的信息包括以下至少一项:所述目标3D空间区域的标识、所述目标3D空间区域的位置信息、所述目标GPCC区域采用GPCC tile编码时,所述目标3D空间区域中的tile标识。
- 根据权利要求14-17任一项所述的方法,若所述至少一个GPCC区域的条目是1个,则1个GPCC区域的条目被封装为1个非时序点云媒体;若所述至少一个GPCC区域的条目是N个,则N个GPCC区域的条目被封装为M个非时序点云媒体;其中,N为大于1的整数,1≤M≤N,M为整数。
- 根据权利要求14-17任一项所述的方法,第二非时序点云媒体包括:GPCC条目组盒子;其中,所述第二非时序点云媒体是所述至少一个非时序点云媒体中包括多个GPCC区域的条目的任一个非时序点云媒体;所述GPCC条目组盒子用于关联所述多个GPCC区域的条目。
- 根据权利要求22所述的方法,所述GPCC条目组盒子包括:第四标识;其中,所述第四标识是所述多个GPCC区域的条目中,在所述视频播放设备初始呈现的条目的标识。
- 根据权利要求22所述的方法,所述GPCC条目组盒子包括:第五标识;若所述第五标识取值为第三数值,则表示所述多个GPCC区域的条目构成所述静态物体的完整GPCC帧;若所述第五标识取值为第四数值,则表示所述多个GPCC区域的条目构成所述静态物体的部分GPCC帧。
- 根据权利要求22所述的方法,所述GPCC条目组盒子包括:所述多个GPCC区域构成的GPCC区域的位置信息。
- 根据权利要求14-17任一项所述的方法,在所述从所述视频制作设备接收所述第一非时序点云媒体之后,所述方法还包括:根据所述MPD信令,基于所述静态物体的标识向视频制作设备发送第二请求消息,所述第二请求消息用于请求所述至少一个非时序点云媒体中的第三非时序点云媒体;从所述视频制作设备接收所述第三非时序点云媒体;播放所述第三非时序点云媒体。
- 一种非时序点云媒体的处理装置,包括:处理单元和通信单元;所述处理单元用于:获取静态物体的非时序点云数据;通过GPCC编码方式对所述非时序点云数据进行处理,得到GPCC比特流;对所述GPCC比特流进行封装,生成至少一个GPCC区域的条目,所述GPCC区域的条目用于表示所述GPCC区域对应的三维3D空间区域的GPCC成分;对所述至少一个GPCC区域的条目进行封装,生成所述静态物体的至少一个非时序点云媒体,所述非时序点云媒体包括所述静态物体的标识;向视频播放设备发送所述至少一个非时序点云媒体的MPD信令;所述通信单元用于:接收所述视频播放设备根据所述MPD信令发送的第一请求消息,所述第一请求消息用于请求所述至少一个非时序点云媒体中的第一非时序点云媒体;根据所述第一请求消息,向所述视频播放设备发送所述第一非时序点云媒体。
- 一种非时序点云媒体的处理装置,包括:处理单元和通信单元;所述通信单元用于:接收至少一个非时序点云媒体的MPD信令,所述非时序点云媒体包括所述静态物体的 标识;根据所述MPD信令向视频制作设备发送第一请求消息,所述第一请求消息用于请求所述至少一个非时序点云媒体中的第一非时序点云媒体;从所述视频制作设备接收所述第一非时序点云媒体;所述处理单元用于播放所述第一非时序点云媒体;其中,所述至少一个非时序点云媒体是对至少一个点云压缩GPCC区域的条目进行封装生成得到的,所述至少一个GPCC区域的条目是对GPCC比特流进行封装生成的,所述GPCC比特流是通过GPCC编码方式对静态物体的非时序点云数据进行处理得到的;针对所述至少一个GPCC区域的条目中的任一个GPCC区域的条目,所述GPCC区域的条目用于表示所述GPCC区域对应的3D空间区域的GPCC成分。
- 一种视频制作设备,包括:处理器和存储器,所述存储器用于存储计算机程序,所述处理器用于调用并运行所述存储器中存储的计算机程序,以执行权利要求1至13中任一项所述的方法。
- 一种视频播放设备,包括:处理器和存储器,所述存储器用于存储计算机程序,所述处理器用于调用并运行所述存储器中存储的计算机程序,以执行权利要求14至26中任一项所述的方法。
- 一种计算机可读存储介质,用于存储计算机程序,所述计算机程序使得计算机执行如权利要求1至13中任一项所述的方法,或者执行如权利要求14至26中任一项所述的方法。
- 一种包括指令的计算机程序产品,当其在计算机上运行时,使得所述计算机执行权利要求1至13中任一项所述的方法,或者执行如权利要求14至26中任一项所述的方法。
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2023530295A JP7508710B2 (ja) | 2020-11-26 | 2021-11-17 | 非時系列ポイントクラウドメディアの処理方法、装置、機器、及びコンピュータプログラム |
KR1020237021494A KR20230110790A9 (ko) | 2020-11-26 | 2021-11-17 | 비-순차적 포인트 클라우드 매체를 프로세싱하기 위한 방법 및 장치, 디바이스 및 스토리지 매체 |
EP21896844.4A EP4254351A4 (en) | 2020-11-26 | 2021-11-17 | METHOD AND DEVICE FOR PROCESSING NON-SEQUENTIAL POINT CLOUD MEDIA, DEVICE AND STORAGE MEDIUM |
US17/969,627 US20230048474A1 (en) | 2020-11-26 | 2022-10-19 | Method and apparatus for processing non-sequential point cloud media, device, and storage medium |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011347626.1A CN114549778A (zh) | 2020-11-26 | 2020-11-26 | 非时序点云媒体的处理方法、装置、设备及存储介质 |
CN202011347626.1 | 2020-11-26 |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US17/969,627 Continuation US20230048474A1 (en) | 2020-11-26 | 2022-10-19 | Method and apparatus for processing non-sequential point cloud media, device, and storage medium |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2022111343A1 true WO2022111343A1 (zh) | 2022-06-02 |
Family
ID=81660424
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2021/131037 WO2022111343A1 (zh) | 2020-11-26 | 2021-11-17 | 非时序点云媒体的处理方法、装置、设备及存储介质 |
Country Status (5)
Country | Link |
---|---|
US (1) | US20230048474A1 (zh) |
EP (1) | EP4254351A4 (zh) |
KR (1) | KR20230110790A9 (zh) |
CN (1) | CN114549778A (zh) |
WO (1) | WO2022111343A1 (zh) |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20230276053A1 (en) * | 2020-06-22 | 2023-08-31 | Interdigital Patent Holdings, Inc. | Adaptive streaming of geometry-based point clouds |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110781894A (zh) * | 2019-09-29 | 2020-02-11 | 腾讯科技(深圳)有限公司 | 点云语义分割方法、装置及电子设备 |
WO2020060813A1 (en) * | 2018-09-18 | 2020-03-26 | Vid Scale, Inc. | Methods and apparatus for point cloud compression bitstream format |
CN111277904A (zh) * | 2020-02-17 | 2020-06-12 | 腾讯科技(深圳)有限公司 | 一种视频的播放控制方法、装置及计算设备 |
TW202029757A (zh) * | 2019-01-10 | 2020-08-01 | 新加坡商聯發科技(新加坡)私人有限公司 | 發信點雲多媒體資料的視埠以及興趣區域的方法及裝置 |
TW202041020A (zh) * | 2019-03-15 | 2020-11-01 | 新加坡商 聯發科技(新加坡)私人有限公司 | 用信號通知點雲多媒體資料軌道的空間關係的方法和裝置 |
-
2020
- 2020-11-26 CN CN202011347626.1A patent/CN114549778A/zh active Pending
-
2021
- 2021-11-17 EP EP21896844.4A patent/EP4254351A4/en active Pending
- 2021-11-17 WO PCT/CN2021/131037 patent/WO2022111343A1/zh active Application Filing
- 2021-11-17 KR KR1020237021494A patent/KR20230110790A9/ko active Search and Examination
-
2022
- 2022-10-19 US US17/969,627 patent/US20230048474A1/en active Pending
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2020060813A1 (en) * | 2018-09-18 | 2020-03-26 | Vid Scale, Inc. | Methods and apparatus for point cloud compression bitstream format |
TW202029757A (zh) * | 2019-01-10 | 2020-08-01 | 新加坡商聯發科技(新加坡)私人有限公司 | 發信點雲多媒體資料的視埠以及興趣區域的方法及裝置 |
TW202041020A (zh) * | 2019-03-15 | 2020-11-01 | 新加坡商 聯發科技(新加坡)私人有限公司 | 用信號通知點雲多媒體資料軌道的空間關係的方法和裝置 |
CN110781894A (zh) * | 2019-09-29 | 2020-02-11 | 腾讯科技(深圳)有限公司 | 点云语义分割方法、装置及电子设备 |
CN111277904A (zh) * | 2020-02-17 | 2020-06-12 | 腾讯科技(深圳)有限公司 | 一种视频的播放控制方法、装置及计算设备 |
Non-Patent Citations (1)
Title |
---|
See also references of EP4254351A4 |
Also Published As
Publication number | Publication date |
---|---|
US20230048474A1 (en) | 2023-02-16 |
EP4254351A1 (en) | 2023-10-04 |
EP4254351A4 (en) | 2024-05-08 |
JP2023550752A (ja) | 2023-12-05 |
CN114549778A (zh) | 2022-05-27 |
KR20230110790A9 (ko) | 2024-03-25 |
KR20230110790A (ko) | 2023-07-25 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN114079781B (zh) | 一种点云媒体的数据处理方法、装置、设备及存储介质 | |
US20230421810A1 (en) | Encapsulation and decapsulation methods and apparatuses for point cloud media file, and storage medium | |
CN113891117B (zh) | 沉浸媒体的数据处理方法、装置、设备及可读存储介质 | |
US20230169719A1 (en) | Method and Apparatus for Processing Immersive Media Data, Storage Medium and Electronic Apparatus | |
WO2022062860A1 (zh) | 一种点云媒体的数据处理方法、装置、设备及存储介质 | |
WO2024037247A1 (zh) | 一种点云媒体的数据处理方法及相关设备 | |
CN114095737A (zh) | 点云媒体文件封装方法、装置、设备及存储介质 | |
WO2022111343A1 (zh) | 非时序点云媒体的处理方法、装置、设备及存储介质 | |
WO2024041239A1 (zh) | 一种沉浸媒体的数据处理方法、装置、设备、存储介质及程序产品 | |
WO2024041238A1 (zh) | 一种点云媒体的数据处理方法及相关设备 | |
US20230086988A1 (en) | Method and apparatus for processing multi-view video, device and storage medium | |
WO2023226504A1 (zh) | 一种媒体数据处理方法、装置、设备以及可读存储介质 | |
US20230034937A1 (en) | Media file encapsulating method, media file decapsulating method, and related devices | |
JP7508710B2 (ja) | 非時系列ポイントクラウドメディアの処理方法、装置、機器、及びコンピュータプログラム | |
CN114581631A (zh) | 沉浸式媒体的数据处理方法、装置和计算机可读存储介质 | |
WO2023024839A1 (zh) | 媒体文件封装与解封装方法、装置、设备及存储介质 | |
CN115733576B (zh) | 点云媒体文件的封装与解封装方法、装置及存储介质 | |
WO2022111348A1 (zh) | 点云媒体的数据处理方法、装置、设备及存储介质 | |
US20230062933A1 (en) | Data processing method, apparatus, and device for non-sequential point cloud media | |
CN115426502A (zh) | 点云媒体的数据处理方法、装置、设备及存储介质 | |
CN117082262A (zh) | 点云文件封装与解封装方法、装置、设备及存储介质 | |
CN115061984A (zh) | 点云媒体的数据处理方法、装置、设备、存储介质 | |
CN115941995A (zh) | 媒体文件封装与解封装方法、装置、设备及存储介质 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 21896844 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2023530295 Country of ref document: JP |
|
ENP | Entry into the national phase |
Ref document number: 20237021494 Country of ref document: KR Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
ENP | Entry into the national phase |
Ref document number: 2021896844 Country of ref document: EP Effective date: 20230626 |