WO2020012968A1 - 画像処理装置および方法 - Google Patents
画像処理装置および方法 Download PDFInfo
- Publication number
- WO2020012968A1 WO2020012968A1 PCT/JP2019/025558 JP2019025558W WO2020012968A1 WO 2020012968 A1 WO2020012968 A1 WO 2020012968A1 JP 2019025558 W JP2019025558 W JP 2019025558W WO 2020012968 A1 WO2020012968 A1 WO 2020012968A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- point cloud
- information
- unit
- model
- decoding
- Prior art date
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/60—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/597—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
- G06T15/10—Geometric effects
- G06T15/20—Perspective computation
- G06T15/205—Image-based rendering
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T19/00—Manipulating 3D models or images for computer graphics
- G06T19/20—Editing of 3D images, e.g. changing shapes or colours, aligning objects or positioning parts
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T9/00—Image coding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/146—Data rate or code amount at the encoder output
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/46—Embedding additional information in the video signal during the compression process
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/85—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10028—Range image; Depth image; 3D point clouds
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2219/00—Indexing scheme for manipulating 3D models or images for computer graphics
- G06T2219/20—Indexing scheme for editing of 3D models
- G06T2219/2016—Rotation, translation, scaling
Definitions
- the present disclosure relates to an image processing apparatus and method, and more particularly, to an image processing apparatus and method capable of suppressing an increase in load of decoding processing of encoded data using a point cloud video-based approach.
- the present disclosure has been made in view of such a situation, and suppresses an increase in the number of instances required for decoding encoded data of a point cloud video-based approach, and suppresses an increase in decoding processing load. Is to be able to do.
- An image processing device includes a conversion unit configured to convert a parameter related to a plurality of point cloud models of a point cloud, and a two-dimensional projection of the plurality of point cloud models obtained by converting the parameters by the conversion unit.
- An image processing apparatus comprising: an encoding unit that encodes a planar image and generates a bit stream including encoded data of the two-dimensional planar image and conversion information that is information on the conversion of the parameter by the conversion unit. .
- An image processing method converts a parameter related to a plurality of point cloud models of a point cloud, encodes a two-dimensional plane image on which the plurality of point cloud models in which the parameters are converted are projected, and This is an image processing method for generating a bit stream including encoded data of a two-dimensional plane image and conversion information that is information relating to the conversion of the parameter.
- An image processing apparatus decodes a bit stream, converts a two-dimensional plane image on which a plurality of point cloud models are projected, and a conversion that is information on conversion of respective parameters of the plurality of point cloud models.
- a decoding unit that generates information, and reconstructs each of the plurality of point group models from the two-dimensional plane image generated by the decoding unit, based on the conversion information, And a reconstructing unit that inversely transforms the parameter of the image processing apparatus.
- An image processing method decodes a bit stream and converts a two-dimensional plane image onto which a plurality of point cloud models are projected, and information on conversion of parameters of the plurality of point cloud models. And an image for generating information, reconstructing each of the plurality of point cloud models from the generated two-dimensional plane image, and inversely converting respective parameters of the plurality of point cloud models based on the conversion information. Processing method.
- An image processing apparatus includes a projection unit that projects a plurality of partial point cloud models constituting a point cloud model onto projection planes that are set independently of each other.
- An encoding unit that encodes a two-dimensional planar image in which patches of the partial point cloud model projected on a plane are arranged and an occupancy map including projection plane information that is information on the projection plane, and generates a bit stream
- An image processing apparatus comprising:
- An image processing method includes projecting a plurality of partial point cloud models constituting a point cloud model onto projection planes set independently of each other, and This is an image processing method for encoding a two-dimensional plane image on which patches of a point cloud model are arranged, and an occupancy map including projection plane information that is information on the projection plane, and generating a bit stream.
- An image processing apparatus decodes a bit stream, projects a two-dimensional plane image on which a point cloud model is projected, and a projection plane of each of a plurality of partial point cloud models included in the point cloud model.
- a decoding unit that generates an occupancy map including projection plane information that is information about the two-dimensional plane image generated by the decoding unit, based on the projection plane information included in the occupancy map, And a reconstructing unit for reconstructing the point cloud model.
- An image processing method includes decoding a bit stream, a two-dimensional plane image on which a point cloud model is projected, and a projection plane of each of a plurality of partial point cloud models included in the point cloud model. And generating an occupancy map including projection plane information that is information relating to the two-dimensional planar image and the projection plane information included in the occupancy map. Image processing method.
- a parameter relating to a plurality of point cloud models of a point cloud is transformed, and a two-dimensional planar image on which a plurality of point cloud models having the transformed parameters are projected is encoded. Then, a bit stream including the encoded data of the two-dimensional plane image and the conversion information that is information on the conversion of the parameter is generated.
- a bit stream is decoded, a two-dimensional plane image on which a plurality of point cloud models are projected, and information regarding conversion of respective parameters of the plurality of point cloud models. Is generated, each of the plurality of point cloud models is reconstructed from the generated two-dimensional plane image, and based on the conversion information, each parameter of the plurality of point cloud models is inversely transformed. Is done.
- a plurality of partial point cloud models forming a point cloud model are respectively projected onto projection planes set independently of each other, and projected onto each projection plane.
- a two-dimensional plane image on which the patches of the partial point cloud model are arranged and an occupancy map including projection plane information that is information on the projection plane are encoded, and a bit stream is generated.
- a two-dimensional plane image on which a bit stream is decoded and a point cloud model is projected, and a plurality of partial point cloud models included in the point cloud model are respectively provided.
- An occupancy map including projection plane information, which is information relating to the projection plane, is generated. Based on the generated two-dimensional planar image and the projection plane information included in the occupancy map, a point cloud model is re-created. Be built.
- an image can be processed.
- it is possible to suppress an increase in the load of decoding processing of encoded data using the video-based approach of the point cloud.
- FIG. 3 is a diagram summarizing main features of the present technology.
- FIG. 14 is a diagram illustrating an overview of encoding and decoding of a point cloud to which the present technology is applied.
- FIG. 14 is a diagram illustrating an overview of encoding and decoding of a point cloud to which the present technology is applied.
- FIG. 39 is a block diagram illustrating a main configuration example of an encoding device. It is a flowchart explaining the example of the flow of an encoding process. 13 is a flowchart illustrating an example of the flow of a conversion process. It is a flowchart explaining the example of the flow of a bounding box setting process. It is a flowchart explaining the example of the flow of a packing process.
- FIG. 15 is a flowchart illustrating an example of the flow of a BB information generation process. It is a block diagram which shows the main structural examples of a decoding device. It is a flowchart explaining the example of the flow of a decoding process.
- FIG. 3 is a diagram summarizing main features of the present technology.
- FIG. 14 is a diagram illustrating an overview of encoding and decoding of a point cloud to which the present technology is applied.
- FIG. 39 is a block diagram illustrating a main configuration example of an encoding device.
- FIG. 3 is a block diagram illustrating a main configuration example of a patch decomposition unit. It is a flowchart explaining the example of the flow of an encoding process.
- Non-Patent Document 6 and QTBT (Quad ⁇ Tree ⁇ Binary Tree) ⁇ Block ⁇ Structure described in Non-Patent Document 7 are not directly described in the embodiment, It is within the disclosure range of the present technology and satisfies the support requirements of the claims. Further, for example, similarly, technical terms such as parsing, syntax, and semantics are within the disclosure range of the present technology even if there is no direct description in the embodiment. Satisfy the support requirements of the claims.
- ⁇ Point Cloud> Conventionally, there existed data such as a point cloud representing a three-dimensional structure based on position information and attribute information of a point cloud, and a mesh composed of vertices, edges, and surfaces and defining a three-dimensional shape using a polygonal representation. .
- a three-dimensional structure is represented as a set (point group) of many points. That is, the data of the point cloud is composed of position information and attribute information (for example, color) of each point of this point group. Therefore, the data structure is relatively simple, and an arbitrary three-dimensional structure can be expressed with sufficient accuracy by using sufficiently many points.
- a video-based approach has been proposed in which the position and color information of such a point cloud are projected onto a two-dimensional plane for each small area, and are encoded by an encoding method for a two-dimensional image. .
- an input point cloud (Point @ cloud) is divided into a plurality of segmentations (also referred to as regions) and projected onto a two-dimensional plane for each region.
- data for each position of the point cloud (that is, data for each point) is composed of position information (Geometry (also referred to as Depth)) and attribute information (Texture) as described above. Projected onto a dimensional plane.
- Each segmentation (also referred to as a patch) projected on the two-dimensional plane is arranged in a two-dimensional image.
- two-dimensional plane images such as AVC (Advanced Video Coding) and HEVC (High Efficiency Video Coding). Is encoded by the following encoding method.
- ⁇ Encoding of point cloud model> there may be a plurality of point cloud models composed of a dense point cloud. For example, when a plurality of people are scattered in a square, if a three-dimensional space including the entire square is converted into a point cloud, a dense point group (point cloud model) is formed in each person. That is, a plurality of point cloud models are formed.
- a plurality of point cloud models are stored in one bit stream. For example, as shown in the top row of the table in FIG. 1, a plurality of point cloud models are converted into one group, and encoded, thereby combining the plurality of point cloud models into one bit stream. To be stored.
- a point cloud model 11-1 to a point cloud model 11-4 exist in a point cloud, and bounding boxes 12-1 to 12-4 respectively correspond to the point cloud models 11-1 to 11-4. It is assumed that it is set.
- the point cloud model 11-1 to the point cloud model 11-4 will be referred to as a point cloud model 11 unless otherwise described.
- bounding boxes 12-1 to 12-4 are not distinguished from each other and described, they are referred to as bounding boxes 12. That is, these point cloud models 11 are separated from each other, and the bounding box 12 is set for each.
- the parameters of the point group model 11 are converted and collected to form a group having few sparse points.
- a group including the bounding boxes 12-1 to 12-4 is formed, and the encoding bounding box 13 is set for the entire group.
- the coordinates of the point cloud model 11 may be transformed as shown in the table of FIG.
- the coordinates of the point cloud model 11 may be shifted (Shift) (the position of the point cloud model 11 may be moved).
- the positions of the point cloud models 11-1 to 11-4 in the point cloud shown on the leftmost side of FIG. 2 may be spatially separated from the others.
- the coordinates of the point cloud model 11 may be rotated (the attitude of the point cloud model 11 may be rotated).
- the postures (orientations) of the point cloud models 11-1 to 11-4 in the point cloud illustrated on the leftmost side of FIG. 2 may be different from the other postures.
- the postures of the respective point cloud models can be made uniform, and sparse portions of the group can be reduced.
- the time (time stamp (TimeStump)) of the point cloud model 11 may be converted.
- the positions of the point cloud model 11-1 to the point cloud model 11-4 in the point cloud shown on the leftmost side of FIG. 2 may be temporally separated from the other (the point cloud model exists at a different time from the other). May be included).
- the times of the respective point cloud models can be aligned, and point cloud models existing at different times can be combined into one group.
- the size (Scale) of the point cloud model 11 may be converted. That is, the scale of each of the point cloud models 11-1 to 11-4 in the point cloud illustrated on the leftmost side of FIG. 2 may be different from the others.
- the size (scale) of each point group model can be made uniform, and the resolution in the spatial direction can be made uniform.
- the frame rate (Frame ⁇ Rate) of the point cloud model 11 may be converted. That is, the frame rates of the point cloud models 11-1 to 11-4 in the point cloud illustrated on the leftmost side of FIG. 2 may be different from the others. By such a conversion, for example, the frame rates (that is, resolutions in the time direction) of each point cloud model can be made uniform.
- a group of point group models in the encoding bounding box 13 shown second from the left is projected onto a two-dimensional plane, packed into a video frame, and placed third from the left.
- an occupancy map 16 are generated.
- encoding is performed using a two-dimensional image encoding method, and one 2D bit stream 17 (including an occupancy map 18 and a header (Header) 19) is encoded as shown in the rightmost part of FIG. Generate. That is, the encoded data of the point cloud models 11-1 to 11-4 is stored in the single 2D bit stream 17.
- ⁇ Conversion information signal> In the case of decryption, the reverse processing is performed. That is, an inverse transformation process of returning each point group model from the state of the second group from the left in FIG. 2 to the original state of the first group from the left is performed. Therefore, as shown in the row # 1 in the table of FIG. 1, parameter conversion relating to the point cloud model performed at the time of encoding (see FIG. 2) so that the inverse transform can be performed at the time of decoding. Conversion information indicating the contents of the conversion from the first state to the second state from the left is generated and transmitted to the decoding side (for example, included in a bit stream) (that is, the conversion information is signaled).
- the conversion information may be any information as long as it indicates the content of the conversion (the amount of change in the converted parameter). For example, as shown in the table of FIG. 1, information indicating the ratio of the converted parameters (for example, Shift, Rotate, Time @ Stump, Scale, Frame @ Rate, etc.) before and after the conversion may be used. Further, for example, as shown in the table of FIG. 1, information indicating the difference between the converted parameters (for example, Shift, Rotate, Time @ Stump, Scale, Frame @ Rate, etc.) before and after the conversion may be used.
- the ratio of the converted parameters for example, Shift, Rotate, Time @ Stump, Scale, Frame @ Rate, etc.
- information indicating the difference between the converted parameters for example, Shift, Rotate, Time @ Stump, Scale, Frame @ Rate, etc.
- the method of transmitting the conversion information is arbitrary. For example, you may make it transmit in connection with the bit stream containing the encoded data of a point cloud model. For example, it may be included in the bit stream. For example, as shown in FIG. 2, such conversion information may be included as BB information 21 in the header (Header) 19 of the 2D bit stream 17.
- the BB information 21 includes, as conversion information, information indicating a shift amount (Shift @ x / y / z), information indicating a rotation amount (Rotate @ x / y / z), and information indicating a change amount of a time stamp (Time).
- Stump information indicating the amount of change in scale (Scale @ x / y / z), information indicating the amount of change in frame rate (Frame @ Rate), and the like.
- the BB information 21 is generated and stored in the header 19, and at the time of decoding, the BB information 21 is read from the header 19 and used for inverse conversion.
- ⁇ Projection plane setting of point cloud model> As shown in the row # 2 in the table of FIG. 1, when projecting the group of the point group model shown second from the left in FIG. 2 onto a two-dimensional plane, the projection plane is changed to the point group model. It may be possible to set each time. By doing so, each point cloud model can be projected onto a more appropriate projection plane, so that a reduction in encoding efficiency due to inefficient projection of the point cloud model can be suppressed (code Efficiency can be improved).
- the projection plane of each of the point cloud models in six orthogonal directions may be made to be able to be rotated (Rotate). This makes it possible to relatively easily set the projection plane to a more appropriate plane (a plane on which efficient projection can be performed) with respect to the point cloud model.
- an arbitrary projection plane may be set (added). By doing so, the degree of freedom of the projection plane of each point cloud model can be improved, and more efficient projection can be expected (encoding efficiency can be expected to be further improved).
- this projection plane information may include any information as long as it is information necessary for specifying the projection plane.
- the projection plane information 22 may be stored in the occupancy map 16 (that is, the occupancy map 18).
- the projection plane information 22 includes information indicating the shift amount of the projection plane in the six orthogonal directions (Shift ⁇ x / y / z), information indicating the change amount of the scale (Scale x / y / z), and the rotation amount. Information (Rotate @ x / y / z) and the like are included. Also, for example, the projection plane information 22 may include information indicating an added arbitrary projection plane.
- the projection plane information 22 is generated and stored in the occupancy map 16, and at the time of decoding, the projection plane information 22 is read from the occupancy map 16 to reconstruct a point cloud. Used for
- Partial decoding support> ⁇ Patch placement control for each point cloud model> Also, as shown in the row # 3 in the table of FIG. 1, when the patches of each point group model are arranged in a two-dimensional image and packed as a video frame, the area where the patches are arranged is defined for each point group model. You may make it controllable. For example, patches belonging to the same point cloud model may be arranged in the same region. This area is optional. For example, it may be an independently decodable coding unit. That is, patches belonging to the same point group model may be arranged in the same, independently decodable coding unit.
- the independently decodable coding unit is arbitrary. For example, as shown in the table of FIG. 1, it may be a frame, a slice, or a tile.
- a point cloud model 31-1 and a point cloud model 31-2 exist in a point cloud.
- a bounding box 32-1 is set in the point cloud model 31-1
- a bounding box 32-2 is set in the point cloud model 31-2. That is, these are point cloud models independent of each other.
- patches 34-1 to 34-4 are patches of the point cloud model 31-1
- patches 35-1 to 35-5 are patches of the point cloud model 31-2.
- the patches 34-1 to 34-4 of the point cloud model 31-1 are arranged, for example, on the slice 36-2 of the two-dimensional image 33, and The two patches 35-1 to 35-5 are arranged on the slice 36-1 of the two-dimensional image 33.
- Independently decodable coding units for arranging patches in this way are controlled for each point group model.
- Patches belonging to the same point group model are allocated to the same, independently decodable coding units.
- partial decoding in which only some of the point cloud models are decoded and reconstructed.
- the slice 36-1 and the slice 36-2 are coding units that can be decoded independently of each other, only the patches 35-1 to 35-5 of the slice 36-1 are decoded.
- only the patches 34-1 to 34-4 of the slice 36-2 can be decoded. That is, it is possible to decode and reconstruct only the point cloud model 31-1 or to decode and reconstruct only the point cloud model 31-2.
- model information indicating the number of point cloud models packed in a video frame may be included in the model information.
- information indicating an area (an independently decodable coding unit) in which a patch of each point cloud model is arranged may be included in the model information.
- the method of transmitting the model information is arbitrary. For example, you may make it transmit in connection with the bit stream containing the encoded data of a point cloud model. For example, it may be included in the bit stream. For example, as shown in FIG. 2, such model information may be included as BB information 21 in the header (Header) 19 of the 2D bit stream 17.
- the BB information 21 includes, as model information, the number of models and information on the patch placement area of each point cloud model.
- the BB information 21 is generated and stored in the header 19, and at the time of decoding, the BB information 21 is read from the header 19 and used for partial decoding.
- FIG. 4 is a block diagram illustrating an example of a configuration of an encoding device that is an aspect of an image processing device to which the present technology is applied.
- An encoding device 100 shown in FIG. 4 is a device that projects 3D data such as a point cloud onto a two-dimensional plane and performs encoding using an encoding method for a two-dimensional image (an encoding device to which a video-based approach is applied). ).
- FIG. 4 shows main components such as a processing unit and a flow of data, and the components shown in FIG. 4 are not necessarily all. That is, in the encoding device 100, a processing unit not illustrated as a block in FIG. 4 may exist, or a process or data flow not illustrated as an arrow or the like in FIG. 4 may exist. This is the same in other drawings for explaining the processing unit and the like in the encoding device 100.
- the encoding device 100 includes a model conversion unit 111, a patch decomposition unit 112, a packing unit 113, an auxiliary patch information compression unit 114, a video encoding unit 115, a video encoding unit 116, and an OMap encoding unit. 117, a multiplexer 118, and a BB information generation unit 119.
- the model conversion unit 111 performs processing related to conversion of a parameter relating to the point cloud model. For example, the model conversion unit 111 acquires 3D data (for example, a point cloud (Point @ Cloud)) representing a three-dimensional structure, which is input to the encoding device 100. Further, the model conversion unit 111 converts parameters relating to the point cloud model included in the obtained point cloud. At this time, the model conversion unit 111 converts parameters related to each point cloud model by the method described above in ⁇ reduction of the number of bit streams> and ⁇ conversion of point cloud models>, and converts a plurality of point cloud models into one group. Put it together. The model conversion unit 111 supplies the point cloud model obtained by converting the parameters, that is, a point cloud including the point cloud models combined into one group, to the patch decomposition unit 112.
- 3D data for example, a point cloud (Point @ Cloud)
- the model conversion unit 111 converts parameters relating to the point cloud model included in the obtained point cloud.
- the model conversion unit 111 generates conversion information for the conversion by the method described above in ⁇ signal of conversion information>.
- the model conversion unit 111 supplies the generated conversion information to the BB information generation unit 119.
- the patch decomposing unit 112 performs processing related to decomposing 3D data. For example, the patch decomposition unit 112 acquires a point cloud (a point cloud including a plurality of point cloud models grouped together) supplied from the model conversion unit 111. Further, the patch decomposition unit 112 decomposes the obtained point cloud into a plurality of segmentations, projects the point cloud on a two-dimensional plane for each of the segmentations, and generates position information patches and attribute information patches. The patch disassembly unit 112 supplies information on each generated patch to the packing unit 113. Further, the patch disassembly unit 112 supplies auxiliary patch information, which is information on the disassembly, to the auxiliary patch information compression unit 114.
- a point cloud a point cloud including a plurality of point cloud models grouped together
- the packing unit 113 performs processing related to data packing. For example, the packing unit 113 obtains information on a patch of position information (Geometry) indicating a position of a point and information on a patch of attribute information (Texture) such as color information added to the position information from the patch decomposition unit 112. I do.
- a patch of position information (Geometry) indicating a position of a point
- a patch of attribute information (Texture) such as color information added to the position information from the patch decomposition unit 112. I do.
- the packing unit 113 arranges the acquired patches in a two-dimensional image and packs them as a video frame.
- the packing unit 113 arranges patches of position information in a two-dimensional image and packs them as a video frame (also referred to as a geometry video frame) of position information.
- the packing unit 113 arranges a patch of attribute information in a two-dimensional image and packs the patch as a video frame (also referred to as a color video frame) of attribute information.
- the packing unit 113 controls the area in which the patches are arranged (independently decodable coding units) for each point group model by the method described above in ⁇ Patch arrangement control for each point group model>. . Then, the packing unit 113 generates model information by the method described above in ⁇ Signal of model information>. The packing unit 113 supplies the generated model information to the BB information generation unit 119.
- the packing unit 113 generates an occupancy map corresponding to these video frames. Further, the packing unit 113 performs a Dilation process on the color video frame.
- the packing unit 113 supplies the generated geometry video frame to the video encoding unit 115.
- the packing unit 113 supplies the color video frame generated in this way to the video encoding unit 116. Further, the packing unit 113 supplies the generated occupancy map to the OMap encoding unit 117. Further, the packing unit 113 supplies control information regarding such packing to the multiplexer 118.
- the auxiliary patch information compression unit 114 performs processing related to compression of the auxiliary patch information. For example, the auxiliary patch information compression unit 114 acquires the data supplied from the patch decomposition unit 112. The auxiliary patch information compression unit 114 encodes (compresses) the auxiliary patch information included in the acquired data. The auxiliary patch information compression unit 114 supplies the encoded data of the obtained auxiliary patch information to the multiplexer 118.
- the video encoding unit 115 performs a process related to encoding of a video frame of position information (Geometry). For example, the video encoding unit 115 acquires a geometry video frame supplied from the packing unit 113. In addition, the video encoding unit 115 encodes the acquired geometry video frame using an arbitrary two-dimensional image encoding method such as AVC or HEVC. The video encoding unit 115 supplies the encoded data (encoded data of the geometry video frame) obtained by the encoding to the multiplexer 118.
- the video encoding unit 116 performs a process related to encoding of a video frame of attribute information (Texture). For example, the video encoding unit 116 acquires a color video frame supplied from the packing unit 113. In addition, the video encoding unit 116 encodes the obtained color video frame by using an arbitrary two-dimensional image encoding method such as AVC or HEVC. The video encoding unit 116 supplies encoded data (encoded data of a color video frame) obtained by the encoding to the multiplexer 118.
- AVC arbitrary two-dimensional image encoding method
- the OMap encoding unit 117 performs a process related to encoding an occupancy map. For example, the OMap encoding unit 117 acquires the occupancy map supplied from the packing unit 113. Further, the OMap encoding unit 117 encodes the obtained occupancy map by an arbitrary encoding method such as arithmetic encoding. The OMap encoding unit 117 supplies the encoded data (encoded data of the occupancy map) obtained by the encoding to the multiplexer 118.
- the multiplexer 118 performs processing related to multiplexing. For example, the multiplexer 118 acquires encoded data of the auxiliary patch information supplied from the auxiliary patch information compression unit 114. Further, the multiplexer 118 acquires control information on packing supplied from the packing unit 113. Further, the multiplexer 118 acquires encoded data of the geometry video frame supplied from the video encoding unit 115. Further, the multiplexer 118 acquires the encoded data of the color video frame supplied from the video encoding unit 116. Further, the multiplexer 118 acquires the encoded data of the occupancy map supplied from the OMap encoding unit 117. Further, the multiplexer 118 acquires the BB information supplied from the BB information generation unit 119.
- the multiplexer 118 multiplexes the acquired information to generate a bitstream.
- the multiplexer 118 outputs the generated bit stream to the outside of the encoding device 100.
- BB information generation section 119 performs processing related to generation of BB information. For example, the BB information generation unit 119 acquires the conversion information supplied from the model conversion unit 111. Further, the BB information generation unit 119 acquires the model information supplied from the packing unit 113. The BB information generation unit 119 generates BB information including the acquired conversion information and model information. The BB information generation unit 119 supplies the generated BB information to the multiplexer 118. That is, the BB information is transmitted to the decoding side.
- the encoding device 100 can convert and group a plurality of point cloud models of the point cloud, encode the encoded point cloud models, and store them in one bitstream. That is, even when a plurality of point cloud models exist in the point cloud, it is possible to suppress an increase in the number of bit streams while suppressing a decrease in encoding efficiency. Therefore, encoding apparatus 100 can suppress an increase in the number of instances required for decoding. That is, an increase in the load of the decoding process can be suppressed. Thereby, an increase in the cost of the decoder can be suppressed. In addition, it is possible to suppress an increase in the processing time of the decoding process.
- step S101 the model conversion unit 111 of the encoding device 100 executes a conversion process to convert a plurality of point cloud models included in the point cloud into one group. Further, the model conversion unit 111 generates conversion information on the conversion.
- the patch decomposing unit 112 projects the plurality of point group models (groups) compiled in step S101 on a two-dimensional plane and decomposes them into patches.
- the patch disassembly unit 112 may set the projection plane of each point cloud model by the method described above in, for example, ⁇ Setting of projection plane of point cloud model>. By doing so, the patch decomposition unit 112 can project each point cloud model onto a more appropriate projection plane, and thus suppresses a reduction in coding efficiency due to inefficient projection of the point cloud model. (Encoding efficiency can be improved).
- the patch disassembly unit 112 may generate projection plane information related to the setting of the projection plane by the method described above in ⁇ Signal of projection plane information>, for example, and transmit the generated projection plane information to the decoding side. By doing so, decoding can be correctly performed (correctly reconstruct a point cloud) on the decoding side.
- the patch disassembly unit 112 generates auxiliary patch information on the disassembly.
- step S103 the auxiliary patch information compression unit 114 compresses (encodes) the auxiliary patch information generated in step S102.
- step S104 the packing unit 113 executes a packing process, arranges the patches of the position information and attribute information generated in step S102 in a two-dimensional image, and packs them as a video frame.
- the packing unit 113 generates model information and an occupancy map. Further, the packing unit 113 performs a Dilation process on the color video frame. Further, the packing unit 113 generates control information on such packing.
- step S105 the BB information generation unit 119 executes BB information generation processing, and generates BB information including the conversion information generated in step S101, the model information generated in step S104, and the like.
- step S106 the video encoding unit 115 encodes the geometry video frame generated in step S104 by using a two-dimensional image encoding method.
- the video encoding unit 115 encodes the geometry video frame according to the setting of the independently decodable encoding unit (the encoding unit area setting) in which the patches of the respective point cloud models are arranged in step S104. That is, for example, when a slice or a tile is set, a geometry video frame is encoded independently for each slice or tile.
- step S107 the video encoding unit 116 encodes the color video frame generated in step S104 by a two-dimensional image encoding method.
- the video encoding unit 116 encodes the color video frame in step S104 according to the setting of the independently decodable coding unit (the coding unit area setting) in which the patches of the respective point cloud models are arranged. That is, for example, when a slice or tile is set, a color video frame is encoded independently for each slice or tile.
- step S108 the OMap encoding unit 117 encodes the occupancy map generated in step S104 by a predetermined encoding method.
- step S109 the multiplexer 118 outputs the various information generated as described above (for example, the encoded data of the auxiliary patch information generated in step S103, the control information on the packing generated in step S104, and the information generated in step S105.
- BB information encoded data of the geometry video frame generated in step S106, encoded data of the color video frame generated in step S107, and encoded data of the occupancy map generated in step S108).
- step S110 the multiplexer 118 outputs the bit stream generated in step S109 to the outside of the encoding device 100.
- step S110 ends, the encoding processing ends.
- step S131 the model conversion unit 111 executes a bounding box setting process to set a bounding box of each point cloud model.
- step S132 the model conversion unit 111 converts the parameters of each bounding box (each point cloud model) set in step S131.
- the model conversion unit 111 converts parameters such as shift, rotation, time stamp, scale, and frame rate as described above in ⁇ Conversion of point cloud model> and the like.
- the model conversion unit 111 can suppress an increase in the number of bit streams while suppressing a decrease in encoding efficiency even when a plurality of point cloud models exist in the point cloud. Therefore, the model conversion unit 111 can suppress an increase in the number of instances required for decoding. That is, it is possible to suppress an increase in the load of the decoding process. Thereby, an increase in the cost of the decoder can be suppressed. In addition, it is possible to suppress an increase in the processing time of the decoding process.
- step S133 the model conversion unit 111 generates the conversion information of the parameters of each bounding box as described above in ⁇ signal of conversion information> and the like. For example, the model conversion unit 111 generates conversion information indicating a ratio and a difference before and after the conversion. By doing so, the model conversion unit 111 can perform the inverse conversion correctly on the decoding side.
- step S133 ends, the conversion processing ends.
- step S141 the model conversion unit 111 derives a normal vector of a part of each point cloud model included in the point cloud.
- step S142 the model conversion unit 111 sets the bounding box so that the one having a large distribution in the normal direction derived in step S141 matches the orthogonal projection vector.
- the packing unit 113 sets a coding unit area according to the point cloud model in step S151, and allocates different coding unit areas to each point cloud model.
- step S152 the packing unit 113 searches and arranges, for each patch of each point group model, an optimal position in the coding unit area allocated to the point group model in step S151.
- the packing unit 113 converts the patches belonging to the same point group model into the same, independently decodable coding units by the method described above in ⁇ Patch arrangement control for each point group model> and the like. Arrange to generate geometry video frames or color video frames. In this manner, on the decoding side, a so-called “partial decoding” in which only a part of the point cloud models is decoded and reconstructed, can be realized.
- step S153 the packing unit 113 generates model information on the arrangement of the point cloud model in step S152 by the method described above in ⁇ Signal of Model Information>. In this manner, on the decoding side, a so-called “partial decoding” in which only a part of the point cloud models is decoded and reconstructed, can be realized.
- step S154 the packing unit 113 generates an occupancy map.
- step S155 the packing unit 113 performs a Dilation process on the color video frame.
- step S155 When the processing in step S155 ends, the packing processing ends, and the processing returns to FIG.
- step S161 the BB information generation unit 119 acquires the conversion information generated in step S133 in FIG.
- step S162 the BB information generation unit 119 acquires the model information generated in step S153 in FIG.
- the BB information generation unit 119 generates BB information including the conversion information and the model information. This BB information is multiplexed together with encoded data and the like by the multiplexer 118 in step S109 (FIG. 5).
- step S163 ends, the BB information generation processing ends, and the processing returns to FIG.
- FIG. 10 is a block diagram illustrating an example of a configuration of a decoding device that is an aspect of an image processing device to which the present technology is applied.
- the decoding device 200 illustrated in FIG. 10 decodes encoded data obtained by projecting and encoding 3D data such as a point cloud on a two-dimensional plane by a decoding method for a two-dimensional image, and projecting the encoded data on a three-dimensional space. (A decoding device to which a video-based approach is applied). For example, the decoding device 200 decodes the bitstream generated by encoding the point cloud by the encoding device 100 (FIG. 4), and reconstructs the point cloud.
- FIG. 10 shows main components such as the processing unit and the flow of data, and the components shown in FIG. 10 are not necessarily all. That is, in the decoding device 200, a processing unit not illustrated as a block in FIG. 10 may exist, or a process or data flow not illustrated as an arrow or the like in FIG. 10 may exist. This is the same in other drawings for explaining the processing unit and the like in the decoding device 200.
- the decoding device 200 includes a demultiplexer 211, an auxiliary patch information decoding unit 212, a point cloud model selection unit 213, a video decoding unit 214, a video decoding unit 215, an OMap decoding unit 216, an unpacking unit 217, And a 3D reconstruction unit 218.
- the demultiplexer 211 performs processing relating to demultiplexing of data. For example, the demultiplexer 211 acquires a bit stream input to the decoding device 200. This bit stream is supplied from the encoding device 100, for example. The demultiplexer 211 demultiplexes this bit stream, extracts encoded data of the auxiliary patch information, and supplies it to the auxiliary patch information decoding unit 212. In addition, the demultiplexer 211 extracts encoded data of the geometry video frame from the bit stream by demultiplexing, and supplies the extracted encoded data to the video decoding unit 214. Further, the demultiplexer 211 extracts coded data of the color video frame from the bit stream by demultiplexing, and supplies the coded data to the video decoding unit 215.
- the demultiplexer 211 extracts encoded data of the occupancy map from the bit stream by demultiplexing, and supplies the extracted data to the OMap decoding unit 216. Further, the demultiplexer 211 extracts control information related to packing from the bit stream by demultiplexing, and supplies the extracted control information to the unpacking unit 217. The demultiplexer 211 extracts BB information from the bit stream by demultiplexing, and supplies the extracted BB information to the point cloud model selection unit 213 and the 3D reconstruction unit 218.
- the auxiliary patch information decoding unit 212 performs a process related to decoding of encoded data of the auxiliary patch information. For example, the auxiliary patch information decoding unit 212 acquires the encoded data of the auxiliary patch information supplied from the demultiplexer 211. Further, the auxiliary patch information decoding unit 212 decodes (decompresses) the coded data of the auxiliary patch information included in the obtained data. The auxiliary patch information decoding unit 212 supplies the auxiliary patch information obtained by decoding to the 3D reconstruction unit 218.
- the point cloud model selection unit 213 performs processing related to selection of a point cloud model to be partially decoded. For example, the point cloud model selection unit 213 acquires BB information from the demultiplexer 211. Further, the point cloud model selection unit 213 receives designation of a point cloud model input by a user or the like based on the model information included in the BB information. For example, the point cloud model selection unit 213 presents the point cloud model included in the model information to the user or the like as an option, and allows the user to select a point cloud model to be decoded.
- the point cloud model selection unit 213 outputs information specifying an area (an independently decodable coding unit) corresponding to the selected point cloud model (where the patch of the point cloud model is arranged) to the video decoding unit 214. , A video decoding unit 215, and an OMap decoding unit 216.
- the video decoding unit 214 performs a process related to decoding of encoded data of the geometry video frame. For example, the video decoding unit 214 acquires encoded data of a geometry video frame supplied from the demultiplexer 211. The video decoding unit 214 decodes the encoded data of the geometry video frame by an arbitrary two-dimensional image decoding method such as AVC or HEVC.
- the video decoding unit 214 can partially decode an area (an independently decodable coding unit) specified by the point cloud model selection unit 213. For example, when a region to be decoded is specified by the point cloud model selection unit 213, the video decoding unit 214 converts the specified region of the encoded data of the geometry video frame into an arbitrary two-dimensional data such as AVC or HEVC. Decoding is performed by an image decoding method. For example, the video decoding unit 214 decodes frames / slices / tiles specified by the point cloud model selection unit 213 in the encoded data of the geometry video frame.
- the video decoding unit 214 can partially decode a geometry video frame.
- the video decoding unit 214 supplies the geometry video frame (or a part of the area) obtained by the decoding to the unpacking unit 217.
- the video decoding unit 215 performs a process related to decoding of encoded data of a color video frame. For example, the video decoding unit 215 obtains encoded data of a color video frame supplied from the demultiplexer 211. The video decoding unit 215 decodes the encoded data of the color video frame by an arbitrary two-dimensional image decoding method such as AVC or HEVC.
- the video decoding unit 215 can partially decode an area (an independently decodable coding unit) specified by the point cloud model selection unit 213. For example, when an area to be decoded is specified by the point cloud model selection unit 213, the video decoding unit 215 converts the specified area of the encoded data of the color video frame into an arbitrary two-dimensional image such as AVC or HEVC. Decoding is performed by an image decoding method. For example, the video decoding unit 215 decodes a frame / slice / tile or the like specified by the point cloud model selection unit 213 in the encoded data of the color video frame.
- the video decoding unit 215 can partially decode a color video frame.
- the video decoding unit 215 supplies the color video frame obtained by the decoding (or a partial area thereof) to the unpacking unit 217.
- the OMap decoding unit 216 performs a process related to decoding the encoded data of the occupancy map. For example, the OMap decoding unit 216 obtains encoded data of an occupancy map supplied from the demultiplexer 211. The OMap decoding unit 216 decodes the encoded data of the occupancy map by an arbitrary decoding method corresponding to the encoding method.
- the OMap decoding unit 216 can partially decode an area (an independently decodable coding unit) specified by the point cloud model selection unit 213. For example, when a region to be decoded is specified by the point cloud model selection unit 213, the OMap decoding unit 216 converts the specified region of the coded occupancy map data into an arbitrary decoding code corresponding to the coding method. Decrypt by the method. For example, the OMap decoding unit 216 decodes the coded data of the occupancy map, such as a frame / slice / tile specified by the point cloud model selection unit 213.
- the OMap decoding unit 216 can partially decode the occupancy map.
- the OMap decoding unit 216 supplies the decoded occupancy map (or a partial area thereof) to the unpacking unit 217.
- the unpacking unit 217 performs a process related to unpacking. For example, the unpacking unit 217 obtains a geometry video frame from the video decoding unit 214, obtains a color video frame from the video decoding unit 215, and obtains an occupancy map from the OMap decoding unit 216. Further, the unpacking unit 217 unpacks the geometry video frame and the color video frame based on the control information on the packing. The unpacking unit 217 converts the position information (Geometry) data (geometry patches and the like) and the attribute information (Texture) data (texture patches and the like) obtained by unpacking and the occupancy map and the like into a 3D reconstruction unit. 218.
- the 3D reconstruction unit 218 performs a process related to the reconstruction of the point cloud.
- the 3D reconstruction unit 218 includes BB information supplied from the demultiplexer 211, auxiliary patch information supplied from the auxiliary patch information decoding unit 212, and position information (Geometry) data supplied from the unpacking unit 217.
- BB information supplied from the demultiplexer 211
- auxiliary patch information supplied from the auxiliary patch information decoding unit 212
- position information (Geometry) data supplied from the unpacking unit 217.
- a point cloud is reconstructed based on (geometry patches and the like), attribute information (Texture) data (texture patches and the like), occupancy maps and the like.
- the 3D reconstruction unit 218 specifies a projection plane corresponding to each point group model in the group based on the projection plane information, and reconstructs a point cloud from a patch or the like using the projection plane. Accordingly, since the decoding device 200 can project each point cloud model onto a more appropriate projection plane, it is possible to suppress a reduction in coding efficiency due to inefficient projection of the point cloud model (code Efficiency can be improved).
- the 3D reconstruction unit 218 can perform an inverse transformation on the reconstructed point cloud model using the transformation information included in the BB information. Therefore, the 3D reconstruction unit 218 can perform an inverse transform so as to correctly correspond to the transform process performed on the encoding side.
- the 3D reconstruction unit 218 outputs the reconstructed point cloud to the outside of the decoding device 200.
- the point cloud is supplied to, for example, a display unit to be imaged, and the image is displayed, recorded on a recording medium, or supplied to another device via communication.
- the decoding device 200 can correctly decode a plurality of point cloud models combined into one bit stream. Therefore, even when a plurality of point cloud models exist in the point cloud, the decoding device 200 can suppress an increase in the number of bit streams while suppressing a decrease in encoding efficiency. Therefore, the decoding device 200 can suppress an increase in the number of instances required for itself. That is, an increase in the load of the decoding process can be suppressed. Thereby, an increase in the cost of the decoder can be suppressed. In addition, it is possible to suppress an increase in the processing time of the decoding process.
- the demultiplexer 211 of the decoding device 200 demultiplexes the bit stream in step S201.
- step S202 the auxiliary patch information decoding unit 212 decodes the auxiliary patch information extracted from the bit stream in step S201.
- step S203 the point cloud model selection unit 213 receives designation of a point cloud model to be decoded.
- step S204 based on the BB information, the point cloud model selection unit 213 determines an independently decodable encoding unit corresponding to the received specification of the point cloud model (that is, a patch of the specified point cloud model is arranged). (Independently decodable coding unit).
- steps S203 and S204 may be omitted.
- step S205 the video decoding unit 214 decodes the encoded data of the geometry video frame (the video frame of the position information) extracted from the bit stream in step S201.
- the video decoding unit 214 When partial decoding is performed, that is, when the designation of the point cloud model to be decoded is received in step S203 and the coding unit to be decoded is selected in step S204, the video decoding unit 214 performs the processing in step S204 of the geometry video frame. Decoding the coding unit (for example, frame / slice / tile) selected in.
- step S206 the video decoding unit 215 decodes the encoded data of the color video frame (the video frame of the attribute information) extracted from the bit stream in step S201.
- the video decoding unit 215 decodes the coding unit (for example, frame / slice / tile) of the color video frame selected in step S204.
- step S207 the OMap decoding unit 216 decodes the encoded data of the occupancy map extracted from the bit stream in step S201.
- the OMap decoding unit 216 decodes the coding unit (for example, frame / slice / tile) of the occupancy map selected in step S204.
- step S209 the 3D reconstruction unit 218 determines a point cloud (each point cloud) based on the auxiliary patch information obtained in step S202 and the geometry patch, texture patch, occupancy map, and the like obtained in step S208. Model).
- step S210 the 3D reconstruction unit 218 performs an inverse transform process, which is an inverse process of the transform process performed on the encoding side, on each reconstructed point group model based on the transform information included in the BB information. I do.
- step S210 ends, the decoding processing ends.
- the decoding device 200 can suppress an increase in the number of bit streams and an increase in the load of the decoding process.
- the projection plane set for the entire point group model is a part where the points included in the point group model are dense (also referred to as a partial point group model).
- the projection direction is not always optimal. That is, the partial point cloud model may be projected in an inefficient direction, and the coding efficiency may be reduced.
- the projection plane of the point cloud model may be locally controlled.
- a point cloud model including a sparse point portion is divided into a plurality of partial point cloud models, and a projection plane is set for each of the partial point cloud models. You may do so.
- ⁇ ⁇ The method of dividing the projection model is arbitrary. For example, as shown in the table of FIG. 12, a dense point group included in the point cloud model may be divided into partial point cloud models.
- the point cloud model corresponding to the bounding box 312 includes a sparse part and a dense point group 311-1 and a partial point cloud model 311. -2.
- the partial point cloud model 311-1 and the partial point cloud model 311-2 are divided, and the projection planes are set independently.
- projection planes 313-1 in six orthogonal directions are set for the partial point cloud model 311-1, and orthogonal The projection plane 313-2 in the direction is set. Then, a patch is generated by projecting the patch on the projection plane, and the patch is arranged on a two-dimensional image.
- the color video frame (Texture) 314 and the geometry video frame (Depth) are arranged as shown in FIG. 315, and an occupancy map (Occupancy @ Map) 315 is generated. Then, they are encoded to generate a 2D bit stream 317 and an occupancy map 318. At the time of decoding, these inverse processes are performed.
- a projection plane can be set for each partial point cloud model, so that each partial point cloud model can be projected onto a more appropriate projection plane. Therefore, it is possible to suppress a decrease in coding efficiency due to inefficient projection of the point cloud model (encoding efficiency can be improved).
- the projection plane of each of the partial point cloud models in the six orthogonal directions may be rotated (Rotated).
- Rotated the projection surface to a more appropriate surface for the partial point cloud model (a surface on which efficient projection can be performed).
- an arbitrary projection plane may be set (added). By doing so, the degree of freedom of the projection plane of each partial point cloud model can be improved, and more efficient projection can be expected (enhancement of coding efficiency can be expected).
- projection plane information 321 is generated for those projection planes.
- this projection plane information may include any information as long as it is information necessary for specifying the projection plane.
- the projection plane information 321 includes information indicating the amount of rotation of the projection plane (Rotate @ x / y / z).
- information indicating the shift amount of the projection plane in the six orthogonal directions Shift x / y / z
- information indicating the change amount of the scale Scale x / y / z
- the projection plane information 321 may include information indicating an added arbitrary projection plane.
- this projection plane information may be stored in an occupancy map.
- the projection plane information 321 is stored in the occupancy map 316 (that is, the occupancy map 318).
- the projection plane information 321 is generated and stored in the occupancy map 16, and at the time of decoding, the projection plane information 321 is read from the occupancy map 316 to reconstruct a point cloud.
- FIG. 14 is a block diagram illustrating an example of a configuration of an encoding device that is an aspect of an image processing device to which the present technology is applied.
- An encoding device 400 illustrated in FIG. 14 is a device similar to the encoding device 100 (FIG. 4), and projects 3D data such as a point cloud onto a two-dimensional plane by using an encoding method for a two-dimensional image.
- This is an apparatus that performs encoding (an encoding apparatus to which a video-based approach is applied).
- FIG. 14 shows main components such as the processing unit and the flow of data, and the components shown in FIG. 14 are not necessarily all. That is, in the encoding device 400, a processing unit not illustrated as a block in FIG. 14 may exist, or a process or data flow not illustrated as an arrow or the like in FIG. 14 may exist. This is the same in other drawings for explaining the processing unit and the like in the encoding device 400.
- the encoding device 400 includes a patch decomposition unit 411, a packing unit 412, an auxiliary patch information compression unit 413, a video encoding unit 414, a video encoding unit 415, an OMap encoding unit 416, and a multiplexer 417. Having.
- the patch decomposing unit 411 performs a process related to decomposing 3D data. For example, the patch decomposition unit 411 obtains a point cloud input to the encoding device 400. Further, the patch decomposition unit 411 decomposes the obtained point cloud into a plurality of segmentations, projects the point cloud on a two-dimensional plane for each of the segmentations, and generates position information patches and attribute information patches. The patch disassembly unit 411 supplies information on each generated patch to the packing unit 412. Further, the patch decomposing unit 411 supplies auxiliary patch information, which is information on the decomposition, to the auxiliary patch information compression unit 413.
- the packing unit 412 performs a process related to data packing. For example, the packing unit 412 obtains from the patch disassembly unit 411 information on a patch of position information (Geometry) indicating a position of a point and information on a patch of attribute information (Texture) such as color information added to the position information. I do.
- a patch of position information indicating a position of a point
- a patch of attribute information indicating a position of a point
- Texture patch of attribute information
- the packing unit 412 arranges the acquired patches in a two-dimensional image and packs them as video frames. For example, the packing unit 412 arranges patches of position information in a two-dimensional image and packs them as a geometry video frame. Further, for example, the packing unit 412 arranges the patch of the attribute information in a two-dimensional image and packs it as a color video frame. Further, the packing unit 412 generates an occupancy map corresponding to these video frames. Further, the packing unit 412 performs a Dilation process on the color video frame.
- the packing unit 412 supplies the generated geometry video frame to the video encoding unit 414.
- the packing unit 412 supplies the color video frame generated in this manner to the video encoding unit 415. Further, the packing unit 412 supplies the occupancy map thus generated to the OMap encoding unit 416. Further, the packing unit 412 supplies control information regarding such packing to the multiplexer 417.
- the auxiliary patch information compression unit 413 performs a process related to compression of the auxiliary patch information. For example, the auxiliary patch information compression unit 413 acquires the data supplied from the patch decomposition unit 411. The auxiliary patch information compression unit 413 encodes (compresses) the auxiliary patch information included in the acquired data. The auxiliary patch information compression unit 413 supplies the encoded data of the obtained auxiliary patch information to the multiplexer 417.
- the video encoding unit 414 performs a process related to encoding of a video frame of position information (Geometry). For example, the video encoding unit 414 acquires a geometry video frame supplied from the packing unit 412. Further, the video encoding unit 414 encodes the acquired geometry video frame by using an arbitrary two-dimensional image encoding method such as AVC or HEVC. The video encoding unit 414 supplies the encoded data (encoded data of the geometry video frame) obtained by the encoding to the multiplexer 417.
- AVC arbitrary two-dimensional image encoding method
- the video encoding unit 415 performs a process related to encoding of a video frame of attribute information (Texture). For example, the video encoding unit 415 acquires a color video frame supplied from the packing unit 412. Further, the video encoding unit 415 encodes the obtained color video frame by using an arbitrary two-dimensional image encoding method such as AVC or HEVC. The video encoding unit 415 supplies encoded data (encoded data of a color video frame) obtained by the encoding to the multiplexer 417.
- AVC arbitrary two-dimensional image encoding method
- the OMap encoding unit 416 performs a process related to encoding an occupancy map. For example, the OMap encoding unit 416 acquires the occupancy map supplied from the packing unit 412. Further, the OMap encoding unit 416 encodes the obtained occupancy map by an arbitrary encoding method such as arithmetic encoding. The OMap encoding unit 416 supplies the encoded data (encoded data of the occupancy map) obtained by the encoding to the multiplexer 417.
- the multiplexer 417 performs processing related to multiplexing. For example, the multiplexer 417 acquires the encoded data of the auxiliary patch information supplied from the auxiliary patch information compression unit 413. Further, the multiplexer 417 acquires control information on packing supplied from the packing unit 412. Further, the multiplexer 417 acquires the encoded data of the geometry video frame supplied from the video encoding unit 414. Further, the multiplexer 417 acquires the encoded data of the color video frame supplied from the video encoding unit 415. Further, the multiplexer 417 acquires the encoded data of the occupancy map supplied from the OMap encoding unit 416.
- the multiplexer 417 multiplexes the acquired information to generate a bitstream.
- the multiplexer 417 outputs the generated bit stream to the outside of the encoding device 400.
- FIG. 15 is a block diagram illustrating a main configuration example of the patch decomposition unit 411 in FIG.
- the patch decomposition unit 411 includes a model division unit 431, a partial point group model projection plane setting unit 432, a projection plane information generation unit 433, a partial point group model projection unit 434, and a projection plane information addition unit. 435.
- the model division unit 431 performs processing related to division of the point cloud model. For example, the model dividing unit 431 acquires a (point cloud model) of the point cloud input to the encoding device 400. In addition, the model division unit 431 performs, for example, ⁇ 5. By using the method described above in ⁇ Local control of projection plane> of Local Projection Plane Control, etc., a portion where points included in the point cloud model are dense is set as a partial point cloud model, and the partial point cloud model is used. The point cloud model is divided for each. The model dividing unit 431 supplies the information of each set partial point cloud model to the partial point cloud model projection plane setting unit 432.
- the partial point cloud model projection plane setting unit 432 performs a process related to setting a projection plane for each partial point cloud model. For example, the partial point cloud model projection plane setting unit 432 acquires information on the partial point cloud model supplied from the model dividing unit 431. In addition, the partial point cloud model projection plane setting unit 432 outputs, for example, ⁇ 5. A projection plane is set for each of the partial point cloud models by the method described above in ⁇ Local control of projection plane> of ⁇ Local control of projection plane>. The partial point cloud model projection plane setting unit 432 supplies the information on the projection plane for each partial point cloud model thus set to the projection plane information generation unit 433 and the partial point cloud model projection unit 434.
- the projection plane information generation unit 433 performs processing related to generation of projection plane information. For example, the projection plane information generation unit 433 acquires information about the projection plane set for each partial point cloud model, supplied from the partial point cloud model projection plane setting unit 432. In addition, the projection plane information generation unit 433 performs, for example, ⁇ 5. Projection plane information for each projection plane is generated by the method described above in ⁇ Local projection plane control> ⁇ Signal of projection plane information>. The projection plane information generation unit 433 supplies the generated projection plane information to the projection plane information addition unit 435.
- the partial point cloud model projection unit 434 performs processing related to projection for each partial point cloud model. For example, the partial point cloud model projection unit 434 acquires information on the projection plane set for each partial point cloud model, supplied from the partial point cloud model projection plane setting unit 432. Further, the partial point cloud model projection unit 434 uses the acquired information on the projection plane, for example, in ⁇ 5. By the method described above in ⁇ Local control of projection plane> in ⁇ Local control of projection plane>, etc., each partial point group model is projected on the projection plane for each small area to generate a patch. The partial point cloud model projection unit 434 supplies the generated patch to the projection plane information addition unit 435.
- the projection plane information adding unit 435 performs processing related to addition of projection plane information. For example, the projection plane information addition unit 435 acquires the projection plane information supplied from the projection plane information generation unit 433. Further, the projection plane information addition unit 435 acquires a patch or the like supplied from the partial point cloud model projection unit 434.
- the projection plane information adding unit 435 is configured to output, for example, ⁇ 5.
- the projection plane related to the projection plane used to generate the patch is obtained from the patch acquired from the partial point cloud model projection unit 434. Add information.
- the projection plane information addition unit 435 supplies the patch to which the projection plane information is added to the packing unit 412. Note that the packing unit 412 stores the projection plane information added to the patch in an occupancy map.
- the encoding device 400 can set and project a projection plane independently for each of the plurality of partial point cloud models of the point cloud. By doing so, each partial point cloud model can be projected onto a more appropriate projection plane. Therefore, it is possible to suppress a decrease in coding efficiency due to inefficient projection of the point cloud model (encoding efficiency can be improved).
- the patch disassembly unit 411 of the encoding device 400 executes the patch disassembly process in step S401 to disassemble the point cloud model into patches. Further, the patch disassembly unit 112 generates auxiliary patch information on the disassembly.
- step S402 the auxiliary patch information compression unit 413 compresses (encodes) the auxiliary patch information generated in step S401.
- the packing unit 412 performs a packing process, arranges each of the position information and attribute information patches generated in step S401 in a two-dimensional image, and packs them as a video frame.
- the packing unit 412 generates model information and an occupancy map. Further, the packing unit 412 performs a Dilation process on the color video frame. Further, the packing unit 412 generates control information regarding such packing.
- step S404 the video encoding unit 414 encodes the geometry video frame generated in step S403 by using a two-dimensional image encoding method.
- step S405 the video encoding unit 415 encodes the color video frame generated in step S403 by using a two-dimensional image encoding method.
- step S406 the OMap encoding unit 416 encodes the occupancy map generated in step S403 by a predetermined encoding method.
- step S407 the multiplexer 417 outputs the various information generated as described above (for example, the coded data of the auxiliary patch information generated in step S402, the control information on the packing generated in step S403, and the information generated in step S404).
- step S408 the multiplexer 417 outputs the bit stream generated in step S407 to the outside of the encoding device 400.
- step S408 ends, the encoding processing ends.
- the model division unit 431 determines in step S421, for example, ⁇ 5.
- the point group model to be processed including the sparse points is divided into a plurality of partial point group models by the method described above in ⁇ Local control of projection plane> of ⁇ Local control of local projection plane>. .
- step S 422 the partial point cloud model projection plane setting unit 432 determines, for example, ⁇ 5.
- the projection plane of each partial point group model set in step S421 is set by the method described above in ⁇ Local control of projection plane> of ⁇ Local control of projection plane>.
- step S423 the projection plane information generation unit 433 determines, for example, ⁇ 5.
- the projection plane information on the projection plane of each partial point group model set in step S422 is generated by the method described above in ⁇ Local projection plane control> ⁇ Signal of projection plane information>.
- step S424 the partial point cloud model projection unit 434 determines, for example, ⁇ 5.
- each partial point group model is projected on the projection plane set in step S422 to generate a patch.
- step S425 the projection plane information adding unit 435 determines, for example, ⁇ 5.
- the projection plane information generated in step S423 is added (added) to the patch generated in step S424 by the method described above in ⁇ Local projection plane control> ⁇ signal of projection plane information>.
- the projection plane information given to this patch is stored in the occupancy map by the packing unit 412.
- step S425 When the process in step S425 is completed, the patch disassembly process is completed, and the process returns to FIG.
- step S441 the model division unit 431 specifies a denser point group based on the histogram of the point group model to be processed.
- step S442 the model dividing unit 431 estimates the projection direction of each of the dense point cloud identified in step S441 based on the normal vector of the point.
- step S443 the model dividing unit 431 sets a dense point group whose projection directions estimated in step S442 are different from each other as a partial point group model.
- step S443 ends, the division processing ends, and the processing returns to FIG.
- the packing unit 412 places each patch of each partial point cloud model on a two-dimensional image in step S461.
- step S462 the packing unit 412 generates an occupancy map including the projection plane information generated in step S423 in FIG.
- step S463 the packing unit 412 performs a Dilation process on the color video frame.
- step S463 When the processing in step S463 is completed, the packing processing is completed, and the processing returns to FIG.
- the projection plane can be set independently for each of the plurality of partial point cloud models of the point cloud, and each partial point cloud model is projected onto a more appropriate projection plane. be able to. Therefore, it is possible to suppress a decrease in coding efficiency due to inefficient projection of the point cloud model (encoding efficiency can be improved).
- FIG. 20 is a block diagram illustrating an example of a configuration of a decoding device that is an aspect of an image processing device to which the present technology is applied.
- a decoding device 500 shown in FIG. 20 is a device similar to the decoding device 200 (FIG. 10), and outputs encoded data obtained by projecting and encoding 3D data such as a point cloud onto a two-dimensional plane.
- This is a device (a decoding device to which a video-based approach is applied) that decodes by a decoding method for an image and projects it on a three-dimensional space.
- the decoding device 500 decodes the bitstream generated by encoding the point cloud by the encoding device 400 (FIG. 14), and reconstructs the point cloud.
- FIG. 20 shows main components such as the processing unit and the flow of data, and the components shown in FIG. 20 are not necessarily all. That is, in the decoding device 500, a processing unit not shown as a block in FIG. 20 may exist, or a process or data flow not shown as an arrow or the like in FIG. 20 may exist. This is the same in other drawings for explaining the processing unit and the like in the decoding device 500.
- the decoding device 500 includes a demultiplexer 511, an auxiliary patch information decoding unit 512, a video decoding unit 513, a video decoding unit 514, an OMap decoding unit 515, an unpacking unit 516, and a 3D reconstruction unit 517.
- a demultiplexer 511 an auxiliary patch information decoding unit 512, a video decoding unit 513, a video decoding unit 514, an OMap decoding unit 515, an unpacking unit 516, and a 3D reconstruction unit 517.
- the demultiplexer 511 performs processing relating to demultiplexing of data. For example, the demultiplexer 511 acquires a bit stream input to the decoding device 500. This bit stream is supplied from the encoding device 400, for example. The demultiplexer 511 demultiplexes this bit stream, extracts encoded data of the auxiliary patch information, and supplies it to the auxiliary patch information decoding unit 512. The demultiplexer 511 extracts encoded data of the geometry video frame from the bit stream by demultiplexing, and supplies the extracted encoded data to the video decoding unit 513. Further, the demultiplexer 511 extracts coded data of the color video frame from the bit stream by demultiplexing, and supplies the coded data to the video decoding unit 514.
- the demultiplexer 511 extracts coded occupancy map data from the bit stream by demultiplexing, and supplies the coded data to the OMap decoding unit 515. Further, the demultiplexer 511 extracts control information related to packing from the bit stream by demultiplexing, and supplies the extracted control information to the unpacking unit 516.
- the auxiliary patch information decoding unit 512 performs a process related to decoding of encoded data of the auxiliary patch information. For example, the auxiliary patch information decoding unit 512 acquires encoded data of the auxiliary patch information supplied from the demultiplexer 511. The auxiliary patch information decoding unit 512 decodes (decompresses) the encoded data of the auxiliary patch information included in the acquired data. The auxiliary patch information decoding unit 512 supplies the auxiliary patch information obtained by decoding to the 3D reconstruction unit 517.
- the video decoding unit 513 performs a process related to decoding encoded data of a geometry video frame. For example, the video decoding unit 513 acquires encoded data of a geometry video frame supplied from the demultiplexer 511. The video decoding unit 513 decodes the encoded data of the geometry video frame by an arbitrary two-dimensional image decoding method such as AVC or HEVC. The video decoding unit 513 supplies the geometry video frame (or a part of the area) obtained by the decoding to the unpacking unit 516.
- the video decoding unit 514 performs a process related to decoding of encoded data of a color video frame. For example, the video decoding unit 514 acquires encoded data of a color video frame supplied from the demultiplexer 511. The video decoding unit 514 decodes the encoded data of the color video frame by an arbitrary two-dimensional image decoding method such as AVC or HEVC. The video decoding unit 514 supplies the color video frame obtained by the decoding (or a partial area thereof) to the unpacking unit 516.
- the OMap decoding unit 515 performs a process related to decoding encoded data of an occupancy map. For example, the OMap decoding unit 515 acquires the encoded data of the occupancy map supplied from the demultiplexer 511. The OMap decoding unit 515 decodes the encoded data of the occupancy map by an arbitrary decoding method corresponding to the encoding method.
- the OMap decoding unit 515 supplies the occupancy map (or a part of the occupancy map) obtained by the decoding to the unpacking unit 516.
- the unpacking unit 516 performs a process related to unpacking. For example, the unpacking unit 516 acquires a geometry video frame from the video decoding unit 513, acquires a color video frame from the video decoding unit 514, and acquires an occupancy map from the OMap decoding unit 515. The unpacking unit 516 unpacks the geometry video frame and the color video frame based on the control information on the packing. The unpacking unit 516 converts the position information (Geometry) data (geometry patches and the like) and the attribute information (Texture) data (texture patches and the like) obtained by unpacking and the occupancy map and the like into a 3D reconstruction unit. 517.
- the 3D reconstruction unit 517 performs a process related to the reconstruction of the point cloud.
- the 3D reconstruction unit 517 includes the auxiliary patch information supplied from the auxiliary patch information decoding unit 512, the position information (Geometry) data (geometry patch and the like) supplied from the unpacking unit 516, and the attribute information (Texture). ) Is reconstructed based on the data (texture patches and the like) and the occupancy map.
- the 3D reconstruction unit 517 identifies a projection plane corresponding to each partial point cloud model based on the projection plane information, and reconstructs a point cloud from a patch or the like using the projection plane. Therefore, the decoding apparatus 500 can reconstruct each partial point cloud model from the patches projected on a more appropriate projection plane, and thus the encoding efficiency is reduced due to the inefficient projection of the partial point cloud model. Can be suppressed (encoding efficiency can be improved).
- the 3D reconstruction unit 517 outputs the reconstructed point cloud to the outside of the decoding device 500.
- the point cloud is supplied to, for example, a display unit to be imaged, and the image is displayed, recorded on a recording medium, or supplied to another device via communication.
- the decoding device 500 can suppress a decrease in coding efficiency even when a plurality of partial point cloud models exist in the point cloud.
- the demultiplexer 511 of the decoding device 500 demultiplexes the bit stream in step S501.
- step S502 the auxiliary patch information decoding unit 512 decodes the auxiliary patch information extracted from the bit stream in step S501.
- step S503 the video decoding unit 513 decodes the encoded data of the geometry video frame (the video frame of the position information) extracted from the bit stream in step S501.
- step S504 the video decoding unit 514 decodes the encoded data of the color video frame (the video frame of the attribute information) extracted from the bit stream in step S501.
- step S505 the OMap decoding unit 515 decodes the encoded data of the occupancy map extracted from the bit stream in step S501.
- This occupancy map includes the above-described projection plane information.
- the unpacking unit 516 performs unpacking. For example, the unpacking unit 516 unpacks the geometry video frame obtained by decoding the encoded data in step S503, and generates a geometry patch. The unpacking unit 516 unpacks the color video frame obtained by decoding the encoded data in step S504, and generates a texture patch. Further, the unpacking unit 516 unpacks the occupancy map obtained by decoding the encoded data in step S505, and extracts an occupancy map corresponding to a geometry patch or a texture patch.
- step S507 the 3D reconstruction unit 517 determines the auxiliary patch information obtained in step S502, the geometry patch, texture patch, and occupancy map obtained in step S506, and the projection plane information included in the occupancy map. Based on the above, the point cloud (each point cloud model) is reconstructed.
- step S507 ends, the decoding processing ends.
- the decoding device 500 can suppress a decrease in encoding efficiency.
- control information related to the present technology described in each of the above embodiments may be transmitted from the encoding side to the decoding side.
- control information for example, enabled_flag
- control for designating a range for example, an upper limit or a lower limit of a block size, or both, a slice, a picture, a sequence, a component, a view, a layer, and the like
- Information may be transmitted.
- ⁇ Computer> The above-described series of processes can be executed by hardware or can be executed by software.
- a program constituting the software is installed in a computer.
- the computer includes a computer incorporated in dedicated hardware, a general-purpose personal computer that can execute various functions by installing various programs, and the like.
- FIG. 22 is a block diagram illustrating a configuration example of hardware of a computer that executes the series of processes described above by a program.
- a CPU Central Processing Unit
- ROM Read Only Memory
- RAM Random Access Memory
- the input / output interface 910 is also connected to the bus 904.
- An input unit 911, an output unit 912, a storage unit 913, a communication unit 914, and a drive 915 are connected to the input / output interface 910.
- the input unit 911 includes, for example, a keyboard, a mouse, a microphone, a touch panel, an input terminal, and the like.
- the output unit 912 includes, for example, a display, a speaker, an output terminal, and the like.
- the storage unit 913 includes, for example, a hard disk, a RAM disk, a nonvolatile memory, and the like.
- the communication unit 914 includes, for example, a network interface.
- the drive 915 drives a removable medium 921 such as a magnetic disk, an optical disk, a magneto-optical disk, or a semiconductor memory.
- the CPU 901 loads, for example, a program stored in the storage unit 913 into the RAM 903 via the input / output interface 910 and the bus 904, and executes the program. Is performed.
- the RAM 903 also appropriately stores data necessary for the CPU 901 to execute various processes.
- the program executed by the computer can be applied, for example, by recording it on a removable medium 921 as a package medium or the like.
- the program can be installed in the storage unit 913 via the input / output interface 910 by attaching the removable medium 921 to the drive 915.
- This program can also be provided via a wired or wireless transmission medium such as a local area network, the Internet, or digital satellite broadcasting. In that case, the program can be received by the communication unit 914 and installed in the storage unit 913.
- a wired or wireless transmission medium such as a local area network, the Internet, or digital satellite broadcasting.
- the program can be received by the communication unit 914 and installed in the storage unit 913.
- this program can be installed in the ROM 902 or the storage unit 913 in advance.
- the present technology can be applied to any configuration.
- the present technology is applicable to a transmitter or a receiver (for example, a television receiver or a mobile phone) in satellite broadcasting, cable broadcasting such as cable TV, distribution on the Internet, and distribution to a terminal by cellular communication, or
- the present invention can be applied to various electronic devices such as an apparatus (for example, a hard disk recorder or a camera) that records an image on a medium such as an optical disk, a magnetic disk, and a flash memory, and reproduces an image from the storage medium.
- the present technology is applicable to a processor (eg, a video processor) as a system LSI (Large Scale Integration), a module using a plurality of processors (eg, a video module), a unit using a plurality of modules (eg, a video unit).
- a processor eg, a video processor
- the present invention can be implemented as a part of the configuration of a device such as a set (for example, a video set) in which other functions are added to the unit.
- the present technology can be applied to a network system including a plurality of devices.
- the present technology may be implemented as cloud computing in which a plurality of devices share and process in a shared manner via a network.
- the present technology is implemented in a cloud service that provides a service relating to an image (moving image) to an arbitrary terminal such as a computer, an AV (Audio Visual) device, a portable information processing terminal, and an IoT (Internet of Things) device. You may make it.
- a system refers to a set of a plurality of components (devices, modules (parts), and the like), and it does not matter whether all components are in the same housing. Therefore, a plurality of devices housed in separate housings and connected via a network and one device housing a plurality of modules in one housing are all systems. .
- the system, device, processing unit, and the like to which the present technology is applied can be used in any field such as, for example, transportation, medical care, crime prevention, agriculture, livestock industry, mining, beauty, factories, home appliances, weather, and nature monitoring. . Further, its use is arbitrary.
- “flag” is information for identifying a plurality of states, and is not limited to information used for identifying two states of true (1) or false (0), as well as three or more. Information that can identify the state is also included. Therefore, the value that the “flag” can take may be, for example, a binary value of 1/0, or may be a ternary value or more. That is, the number of bits constituting the "flag” is arbitrary, and may be 1 bit or a plurality of bits. Further, the identification information (including the flag) may include not only a form in which the identification information is included in the bit stream but also a form in which the difference information of the identification information with respect to certain reference information is included in the bit stream. In, "flag” and “identification information” include not only the information but also difference information with respect to reference information.
- association means, for example, that one data can be used (linked) when one data is processed. That is, the data associated with each other may be collected as one data or may be individual data. For example, the information associated with the encoded data (image) may be transmitted on a different transmission path from the encoded data (image). Further, for example, information associated with encoded data (image) may be recorded on a recording medium different from the encoded data (image) (or another recording area of the same recording medium). Good. Note that this “association” may be a part of the data instead of the entire data. For example, an image and information corresponding to the image may be associated with each other in an arbitrary unit such as a plurality of frames, one frame, or a part of the frame.
- Embodiments of the present technology are not limited to the above-described embodiments, and various modifications can be made without departing from the gist of the present technology.
- the configuration described as one device (or processing unit) may be divided and configured as a plurality of devices (or processing units).
- the configuration described above as a plurality of devices (or processing units) may be configured as one device (or processing unit).
- a configuration other than those described above may be added to the configuration of each device (or each processing unit).
- a part of the configuration of a certain device (or processing unit) may be included in the configuration of another device (or other processing unit).
- the above-described program may be executed by an arbitrary device.
- the device only has to have necessary functions (functional blocks and the like) and be able to obtain necessary information.
- each step of one flowchart may be executed by one device, or a plurality of devices may share and execute the steps.
- the plurality of processes may be executed by one device, or may be shared and executed by a plurality of devices.
- a plurality of processes included in one step may be executed as a plurality of steps.
- the processing described as a plurality of steps may be collectively executed as one step.
- a program executed by a computer may be configured so that processing of steps for describing a program is executed in chronological order according to the order described in this specification, or may be executed in parallel or when calls are executed. It may be executed individually at a necessary timing such as when it is touched. That is, as long as no inconsistency arises, the processing of each step may be performed in an order different from the order described above. Further, the processing of the steps for describing this program may be executed in parallel with the processing of another program, or may be executed in combination with the processing of another program.
- a plurality of technologies related to the present technology can be independently and independently implemented unless there is a contradiction.
- some or all of the present technology described in any of the embodiments may be combined with some or all of the present technology described in other embodiments.
- some or all of the above-described arbitrary technology may be implemented in combination with another technology that is not described above.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Computer Graphics (AREA)
- Architecture (AREA)
- Computer Hardware Design (AREA)
- General Engineering & Computer Science (AREA)
- Software Systems (AREA)
- Computing Systems (AREA)
- Geometry (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
- Image Processing (AREA)
Abstract
Description
1.点群モデルの変換
2.部分デコード対応
3.第1の実施の形態(符号化装置)
4.第2の実施の形態(復号装置)
5.局所的な投影面制御
6.第3の実施の形態(符号化装置)
7.第4の実施の形態(復号装置)
8.付記
<技術内容・技術用語をサポートする文献等>
本技術で開示される範囲は、実施の形態に記載されている内容だけではなく、出願当時において公知となっている以下の非特許文献に記載されている内容も含まれる。
非特許文献2:(上述)
非特許文献3:(上述)
非特許文献4:(上述)
非特許文献5:TELECOMMUNICATION STANDARDIZATION SECTOR OF ITU(International Telecommunication Union), "Advanced video coding for generic audiovisual services", H.264, 04/2017
非特許文献6:TELECOMMUNICATION STANDARDIZATION SECTOR OF ITU(International Telecommunication Union), "High efficiency video coding", H.265, 12/2016
非特許文献7:Jianle Chen, Elena Alshina, Gary J. Sullivan, Jens-Rainer, Jill Boyce, "Algorithm Description of Joint Exploration Test Model 4", JVET-G1001_v1, Joint Video Exploration Team (JVET) of ITU-T SG 16 WP 3 and ISO/IEC JTC 1/SC 29/WG 11 7th Meeting: Torino, IT, 13-21 July 2017
従来、点群の位置情報や属性情報等により3次元構造を表すポイントクラウドや、頂点、エッジ、面で構成され、多角形表現を使用して3次元形状を定義するメッシュ等のデータが存在した。
このようなポイントクラウドの位置と色情報それぞれを、小領域毎に2次元平面に投影し、2次元画像用の符号化方法で符号化するビデオベースドアプローチ(Video-based approach)が提案されている。
このようなポイントクラウドにおいて、密なポイント群からなる点群モデルが複数存在する場合がある。例えば、広場に複数の人物が点在する場合にその広場全体を含む3次元空間をポイントクラウド化すると、各人物の部分に密なポイント群(点群モデル)が形成される。つまり、複数の点群モデルが形成される。
そこで、複数の点群モデルを1本のビットストリームに格納するようにする。例えば、図1の表の一番上の行に示されるように、複数の点群モデルを変換して1つのグループにまとめ、符号化することにより、複数の点群モデルを1本のビットストリームに格納するようにする。
より具体的には、図1の表の#1の行に示されるように、各点群モデルに関するパラメータを変換して、ポイントが疎な部分を低減させた1つのグループにまとめる(各点群モデルの変換)。つまり、そのグループをまとめて符号化し、1本のビットストリームとする。
復号の場合は、以上の逆処理が行われる。つまり、各点群モデルを、図2の左から2番目のグループの状態から、左から1番目の元の状態に戻す逆変換処理が行われる。そこで、復号の際にこの逆変換を行うことができるように、図1の表の#1の行に示されるように、符号化の際に行われる点群モデルに関するパラメータの変換(図2の左から1番目状態から2番目の状態への変換)の内容を示す変換情報を生成し、(例えばビットストリームに含めて)復号側に伝送する(つまり、変換情報のシグナリングを行う)。
なお、図1の表の#2の行に示されるように、図2の左から2番目に示される点群モデルのグループを2次元平面に投影する場合に、その投影面を、点群モデル毎に設定することができるようにしてもよい。このようにすることにより、各点群モデルをより適切な投影面に投影することができるので、点群モデルの投影が非効率であることによる符号化効率の低減を抑制することができる(符号化効率を向上させることができる)。
なお、このように点群モデル毎に投影面を設定する場合(グループ共通の投影面としない場合)、復号の際(ポイントクラウドを再構築する際)に、その投影面の情報も必要になる。そこで、図1の表の#2の行に示されるように、以上のように設定する投影面に関する情報である投影面情報を復号側に伝送する(投影面情報のシグナリング)。
<点群モデル毎のパッチ配置制御>
また、図1の表の#3の行に示されるように、各点群モデルのパッチを2次元画像に配置し、ビデオフレームとしてパッキングする場合に、パッチを配置する領域を点群モデル毎に制御することができるようにしてもよい。例えば、互いに同一の点群モデルに属するパッチを、互いに同一の領域に配置するようにしてもよい。この領域は任意である。例えば、独立に復号可能な符号化単位であってもよい。すなわち、互いに同一の点群モデルに属するパッチを、互いに同一の、独立に復号可能な符号化単位に配置するようにしてもよい。
なお、このような部分デコードを行うためには、復号側においてどの領域(どの独立に復号可能な符号化単位)にどの点群モデルのパッチが配置されているかを把握する必要がある。そこで、図1の表の#3の行に示されるように、点群モデルに関する情報であるモデル情報を生成し、符号化側から復号側に伝送する(シグナルする)。
<符号化装置>
次に、以上のような各手法を実現する構成について説明する。図4は、本技術を適用した画像処理装置の一態様である符号化装置の構成の一例を示すブロック図である。図4に示される符号化装置100は、ポイントクラウドのような3Dデータを2次元平面に投影して2次元画像用の符号化方法により符号化を行う装置(ビデオベースドアプローチを適用した符号化装置)である。
次に、符号化装置100により実行される符号化処理の流れの例を、図5のフローチャートを参照して説明する。
次に、図5のステップS101において実行される変換処理の流れの例を、図6のフローチャートを参照して説明する。
次に、図6のステップS131において実行されるバウンディングボックス設定処理の流れの例を、図7のフローチャートを参照して説明する。
次に、図5のステップS104において実行されるパッキング処理の流れの例を、図8のフローチャートを参照して説明する。
次に、図5のステップS105において実行されるBB情報生成処理の流れの例を、図9のフローチャートを参照して説明する。
<復号装置>
図10は、本技術を適用した画像処理装置の一態様である復号装置の構成の一例を示すブロック図である。図10に示される復号装置200は、ポイントクラウドのような3Dデータが2次元平面に投影されて符号化された符号化データを、2次元画像用の復号方法により復号し、3次元空間に投影する装置(ビデオベースドアプローチを適用した復号装置)である。例えば、復号装置200は、符号化装置100(図4)がポイントクラウドを符号化して生成したビットストリームを復号し、ポイントクラウドを再構築する。
次に、この復号装置200により実行される復号処理の流れの例を、図11のフローチャートを参照して説明する。
点群モデルにポイントが疎な部分がある場合、その点群モデル全体に対して設定される投影面が、点群モデルに含まれるポイントが密な部分(部分点群モデルとも称する)にとって、その最適な投影方向であるとは限らない。つまり、部分点群モデルを非効率な方向に投影し、符号化効率が低減するおそれがあった。
そこで、図12の表の上から1段目に示されるように、点群モデルの投影面を局所的に制御するようにしてもよい。例えば、図12の表の上から2段目に示されるように、ポイントが疎な部分を含む点群モデルを複数の部分点群モデルに分割し、部分点群モデル毎に投影面を設定するようにしてもよい。
なお、このように部分点群モデル毎に投影面を設定する場合(点群モデル全体の投影面としない場合)、復号の際(ポイントクラウドを再構築する際)に、その投影面の情報も必要になる。そこで、図12の表の上から2行目に示されるように、以上のように設定する投影面に関する情報である投影面情報を生成し、復号側に伝送する(投影面情報のシグナリング)。
<符号化装置>
次に、以上のような手法を実現する構成について説明する。図14は、本技術を適用した画像処理装置の一態様である符号化装置の構成の一例を示すブロック図である。図14に示される符号化装置400は、符号化装置100(図4)と同様の装置であり、ポイントクラウドのような3Dデータを2次元平面に投影して2次元画像用の符号化方法により符号化を行う装置(ビデオベースドアプローチを適用した符号化装置)である。
図15は、図14のパッチ分解部411の主な構成例を示すブロック図である。図15に示されるように、パッチ分解部411は、モデル分割部431、部分点群モデル投影面設定部432、投影面情報生成部433、部分点群モデル投影部434、および投影面情報付加部435を有する。
次に、符号化装置400により実行される符号化処理の流れの例を、図16のフローチャートを参照して説明する。
次に、図17のフローチャートを参照して、図16のステップS401において実行されるパッチ分解処理の流れの例を説明する。
次に、図17のステップS421において実行される分割処理の流れの例を、図18のフローチャートを参照して説明する。
次に、図16のステップS403において実行されるパッキング処理の流れの例を、図19のフローチャートを参照して説明する。
<復号装置>
図20は、本技術を適用した画像処理装置の一態様である復号装置の構成の一例を示すブロック図である。図20に示される復号装置500は、復号装置200(図10)と同様の装置であり、ポイントクラウドのような3Dデータが2次元平面に投影されて符号化された符号化データを、2次元画像用の復号方法により復号し、3次元空間に投影する装置(ビデオベースドアプローチを適用した復号装置)である。例えば、復号装置500は、符号化装置400(図14)がポイントクラウドを符号化して生成したビットストリームを復号し、ポイントクラウドを再構築する。
次に、この復号装置500により実行される復号処理の流れの例を、図21のフローチャートを参照して説明する。
<制御情報>
以上の各実施の形態において説明した本技術に関する制御情報を符号化側から復号側に伝送するようにしてもよい。例えば、上述した本技術を適用することを許可(または禁止)するか否かを制御する制御情報(例えばenabled_flag)を伝送するようにしてもよい。また、例えば、上述した本技術を適用することを許可(または禁止)する範囲(例えばブロックサイズの上限若しくは下限、またはその両方、スライス、ピクチャ、シーケンス、コンポーネント、ビュー、レイヤ等)を指定する制御情報を伝送するようにしてもよい。
上述した一連の処理は、ハードウエアにより実行させることもできるし、ソフトウエアにより実行させることもできる。一連の処理をソフトウエアにより実行する場合には、そのソフトウエアを構成するプログラムが、コンピュータにインストールされる。ここでコンピュータには、専用のハードウエアに組み込まれているコンピュータや、各種のプログラムをインストールすることで、各種の機能を実行することが可能な、例えば汎用のパーソナルコンピュータ等が含まれる。
以上においては、ポイントクラウドデータの符号化・復号に本技術を適用する場合について説明したが、本技術は、これらの例に限らず、任意の規格の3Dデータの符号化・復号に対して適用することができる。つまり、上述した本技術と矛盾しない限り、符号化・復号方式等の各種処理、並びに、3Dデータやメタデータ等の各種データの仕様は任意である。また、本技術と矛盾しない限り、上述した一部の処理や仕様を省略してもよい。
本技術を適用したシステム、装置、処理部等は、例えば、交通、医療、防犯、農業、畜産業、鉱業、美容、工場、家電、気象、自然監視等、任意の分野に利用することができる。また、その用途も任意である。
なお、本明細書において「フラグ」とは、複数の状態を識別するための情報であり、真(1)または偽(0)の2状態を識別する際に用いる情報だけでなく、3以上の状態を識別することが可能な情報も含まれる。したがって、この「フラグ」が取り得る値は、例えば1/0の2値であってもよいし、3値以上であってもよい。すなわち、この「フラグ」を構成するbit数は任意であり、1bitでも複数bitでもよい。また、識別情報(フラグも含む)は、その識別情報をビットストリームに含める形だけでなく、ある基準となる情報に対する識別情報の差分情報をビットストリームに含める形も想定されるため、本明細書においては、「フラグ」や「識別情報」は、その情報だけではなく、基準となる情報に対する差分情報も包含する。
Claims (20)
- ポイントクラウドの複数の点群モデルに関するパラメータを変換する変換部と、
前記変換部により前記パラメータが変換された前記複数の点群モデルが投影された2次元平面画像を符号化し、前記2次元平面画像の符号化データと、前記変換部による前記パラメータの変換に関する情報である変換情報とを含むビットストリームを生成する符号化部と
を備える画像処理装置。 - 前記変換部は、前記パラメータとして、前記点群モデルの座標を変換する
請求項1に記載の画像処理装置。 - 前記変換部は、前記座標の変換として、前記座標のシフト若しくは回転、またはその両方を行う
請求項2に記載の画像処理装置。 - 前記変換部は、前記パラメータとして、前記点群モデルの時刻を変換する
請求項1に記載の画像処理装置。 - 前記変換部は、前記パラメータとして、前記点群モデルのスケールを変換する
請求項1に記載の画像処理装置。 - 前記変換部は、前記パラメータとして、前記点群モデルのフレームレートを変換する
請求項1に記載の画像処理装置。 - 前記変換情報は、前記変換部により変換された前記パラメータの変化量を示す情報を含む
請求項1に記載の画像処理装置。 - 前記点群モデルのパッチを、前記点群モデル毎に領域を変えて配置し、ビデオフレームとしてパッキングするパッキング部をさらに備え、
前記符号化部は、前記パッキング部により前記パッチがパッキングされた前記ビデオフレームを符号化し、前記点群モデルに関する情報であるモデル情報をさらに含む前記ビットストリームを生成する
請求項1に記載の画像処理装置。 - 前記領域は、独立に復号可能な符号化単位である
請求項8に記載の画像処理装置。 - 前記モデル情報は、前記ビットストリームに含まれる点群モデルの数と、各点群モデルのパッチが配置された領域を示す情報とを含む
請求項8に記載の画像処理装置。 - ポイントクラウドの複数の点群モデルに関するパラメータを変換し、
前記パラメータが変換された前記複数の点群モデルが投影された2次元平面画像を符号化し、前記2次元平面画像の符号化データと、前記パラメータの変換に関する情報である変換情報とを含むビットストリームを生成する
画像処理方法。 - ビットストリームを復号し、複数の点群モデルが投影された2次元平面画像と、前記複数の点群モデルのそれぞれのパラメータの変換に関する情報である変換情報とを生成する復号部と、
前記復号部により生成された前記2次元平面画像から前記複数の点群モデルのそれぞれを再構築し、前記変換情報に基づいて、前記複数の点群モデルのそれぞれのパラメータを逆変換する再構築部と
を備える画像処理装置。 - ビットストリームを復号し、複数の点群モデルが投影された2次元平面画像と、前記複数の点群モデルのそれぞれのパラメータの変換に関する情報である変換情報とを生成し、
生成された前記2次元平面画像から前記複数の点群モデルのそれぞれを再構築し、前記変換情報に基づいて、前記複数の点群モデルのそれぞれのパラメータを逆変換する
画像処理方法。 - 点群モデルを構成する複数の部分点群モデルを、それぞれ、互いに独立に設定された投影面に投影する投影部と、
前記投影部により各投影面に投影された前記部分点群モデルのパッチが配置された2次元平面画像と、前記投影面に関する情報である投影面情報を含むオキュパンシーマップとを符号化し、ビットストリームを生成する符号化部と
備える画像処理装置。 - 前記投影面情報は、前記投影面の回転量を示す情報を含む
請求項14に記載の画像処理装置。 - 前記点群モデルを前記複数の部分点群モデルに分割する分割部と、
前記複数の部分点群モデルのそれぞれについて投影面を設定する設定部と
をさらに備える請求項14に記載の画像処理装置。 - 前記分割部は、
前記点群モデルのヒストグラムにより密な点群を特定し、
法線ベクトルに基づいて前記密な点群のそれぞれの投影方向を推定し、
前記投影方向が互いに異なる密な点群を、部分点群モデルとして設定する
請求項16に記載の画像処理装置。 - 点群モデルを構成する複数の部分点群モデルを、それぞれ、互いに独立に設定された投影面に投影し、
各投影面に投影された前記部分点群モデルのパッチが配置された2次元平面画像と、前記投影面に関する情報である投影面情報を含むオキュパンシーマップとを符号化し、ビットストリームを生成する
画像処理方法。 - ビットストリームを復号し、点群モデルが投影された2次元平面画像と、前記点群モデルに含まれる複数の部分点群モデルのそれぞれの投影面に関する情報である投影面情報を含むオキュパンシーマップとを生成する復号部と、
前記復号部により生成された前記2次元平面画像と、前記オキュパンシーマップに含まれる前記投影面情報とに基づいて、前記点群モデルを再構築する再構築部と
を備える画像処理装置。 - ビットストリームを復号し、点群モデルが投影された2次元平面画像と、前記点群モデルに含まれる複数の部分点群モデルのそれぞれの投影面に関する情報である投影面情報を含むオキュパンシーマップとを生成し、
生成された前記2次元平面画像と、前記オキュパンシーマップに含まれる前記投影面情報とに基づいて、前記点群モデルを再構築する
画像処理方法。
Priority Applications (10)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201980045067.8A CN112470480A (zh) | 2018-07-11 | 2019-06-27 | 图像处理装置和方法 |
EP19834544.9A EP3823286A4 (en) | 2018-07-11 | 2019-06-27 | IMAGE PROCESSING DEVICE AND METHOD |
KR1020207036753A KR20210030276A (ko) | 2018-07-11 | 2019-06-27 | 화상 처리 장치 및 방법 |
MX2020013700A MX2020013700A (es) | 2018-07-11 | 2019-06-27 | Metodo y aparato de procesamiento de imagen. |
US15/734,567 US11699248B2 (en) | 2018-07-11 | 2019-06-27 | Image processing apparatus and method |
CA3106234A CA3106234A1 (en) | 2018-07-11 | 2019-06-27 | Image processing apparatus and method |
BR112021000044-5A BR112021000044A2 (pt) | 2018-07-11 | 2019-06-27 | Aparelho e método de processamento de imagem |
JP2020530096A JP7327399B2 (ja) | 2018-07-11 | 2019-06-27 | 画像処理装置および方法 |
AU2019302075A AU2019302075A1 (en) | 2018-07-11 | 2019-06-27 | Image processing device and method |
PH12021550063A PH12021550063A1 (en) | 2018-07-11 | 2021-01-10 | Image processing apparatus and method |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2018-131300 | 2018-07-11 | ||
JP2018131300 | 2018-07-11 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2020012968A1 true WO2020012968A1 (ja) | 2020-01-16 |
Family
ID=69141449
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2019/025558 WO2020012968A1 (ja) | 2018-07-11 | 2019-06-27 | 画像処理装置および方法 |
Country Status (12)
Country | Link |
---|---|
US (1) | US11699248B2 (ja) |
EP (1) | EP3823286A4 (ja) |
JP (1) | JP7327399B2 (ja) |
KR (1) | KR20210030276A (ja) |
CN (1) | CN112470480A (ja) |
AU (1) | AU2019302075A1 (ja) |
BR (1) | BR112021000044A2 (ja) |
CA (1) | CA3106234A1 (ja) |
MX (1) | MX2020013700A (ja) |
PH (1) | PH12021550063A1 (ja) |
TW (1) | TW202017374A (ja) |
WO (1) | WO2020012968A1 (ja) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2021200919A1 (ja) * | 2020-03-30 | 2021-10-07 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカ | 三次元データ符号化方法、三次元データ復号方法、三次元データ符号化装置、及び三次元データ復号装置 |
WO2022149589A1 (ja) * | 2021-01-07 | 2022-07-14 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカ | 三次元データ復号方法、三次元データ符号化方法、三次元データ復号装置、及び三次元データ符号化装置 |
Families Citing this family (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111566703B (zh) * | 2018-01-17 | 2023-10-20 | 索尼公司 | 图像处理装置和方法 |
JP2022003716A (ja) * | 2018-10-01 | 2022-01-11 | ソニーグループ株式会社 | 画像処理装置および方法 |
CN113261297A (zh) * | 2019-01-07 | 2021-08-13 | 索尼集团公司 | 图像处理装置和方法 |
CN114930860A (zh) * | 2020-01-07 | 2022-08-19 | Lg电子株式会社 | 点云数据发送装置、点云数据发送方法、点云数据接收装置和点云数据接收方法 |
US11706450B2 (en) * | 2020-09-18 | 2023-07-18 | Samsung Electronics Co., Ltd. | Partial decoding and reconstruction of a video-based point cloud compression bitstream |
WO2023047021A2 (en) * | 2021-09-27 | 2023-03-30 | Nokia Technologies Oy | A method, an apparatus and a computer program product for video encoding and video decoding |
CN115049750B (zh) * | 2022-05-31 | 2023-06-16 | 九识智行(北京)科技有限公司 | 基于八叉树的体素地图生成方法、装置、存储介质及设备 |
KR20240043994A (ko) | 2022-09-28 | 2024-04-04 | 증안리약초마을 협동조합 | 누룽지 제조 방법 |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2017126314A1 (en) * | 2016-01-22 | 2017-07-27 | Mitsubishi Electric Corporation | Method for compressing point cloud |
WO2018034253A1 (en) * | 2016-08-19 | 2018-02-22 | Mitsubishi Electric Corporation | Method, encoder system and non-transitory computer readable recording medium storing thereon program for encoding point cloud of representing scene |
WO2019055963A1 (en) * | 2017-09-18 | 2019-03-21 | Apple Inc. | COMPRESSION OF CLOUD OF POINTS |
WO2019142666A1 (ja) * | 2018-01-16 | 2019-07-25 | ソニー株式会社 | 画像処理装置および方法 |
Family Cites Families (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2017142353A1 (ko) * | 2016-02-17 | 2017-08-24 | 엘지전자 주식회사 | 360 비디오를 전송하는 방법, 360 비디오를 수신하는 방법, 360 비디오 전송 장치, 360 비디오 수신 장치 |
US11019257B2 (en) * | 2016-05-19 | 2021-05-25 | Avago Technologies International Sales Pte. Limited | 360 degree video capture and playback |
WO2018043905A1 (ko) | 2016-08-29 | 2018-03-08 | 엘지전자 주식회사 | 360 비디오를 전송하는 방법, 360 비디오를 수신하는 방법, 360 비디오 전송 장치, 360 비디오 수신 장치 |
CN109644279B (zh) | 2016-09-02 | 2023-09-22 | Vid拓展公司 | 用于用信号通知360度视频信息的方法和系统 |
EP3349182A1 (en) | 2017-01-13 | 2018-07-18 | Thomson Licensing | Method, apparatus and stream for immersive video format |
US11514613B2 (en) * | 2017-03-16 | 2022-11-29 | Samsung Electronics Co., Ltd. | Point cloud and mesh compression using image/video codecs |
US10499054B2 (en) * | 2017-10-12 | 2019-12-03 | Mitsubishi Electric Research Laboratories, Inc. | System and method for inter-frame predictive compression for point clouds |
-
2019
- 2019-06-27 US US15/734,567 patent/US11699248B2/en active Active
- 2019-06-27 MX MX2020013700A patent/MX2020013700A/es unknown
- 2019-06-27 WO PCT/JP2019/025558 patent/WO2020012968A1/ja unknown
- 2019-06-27 JP JP2020530096A patent/JP7327399B2/ja active Active
- 2019-06-27 AU AU2019302075A patent/AU2019302075A1/en not_active Abandoned
- 2019-06-27 CN CN201980045067.8A patent/CN112470480A/zh active Pending
- 2019-06-27 EP EP19834544.9A patent/EP3823286A4/en active Pending
- 2019-06-27 KR KR1020207036753A patent/KR20210030276A/ko unknown
- 2019-06-27 CA CA3106234A patent/CA3106234A1/en not_active Abandoned
- 2019-06-27 BR BR112021000044-5A patent/BR112021000044A2/pt unknown
- 2019-07-02 TW TW108123292A patent/TW202017374A/zh unknown
-
2021
- 2021-01-10 PH PH12021550063A patent/PH12021550063A1/en unknown
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2017126314A1 (en) * | 2016-01-22 | 2017-07-27 | Mitsubishi Electric Corporation | Method for compressing point cloud |
WO2018034253A1 (en) * | 2016-08-19 | 2018-02-22 | Mitsubishi Electric Corporation | Method, encoder system and non-transitory computer readable recording medium storing thereon program for encoding point cloud of representing scene |
WO2019055963A1 (en) * | 2017-09-18 | 2019-03-21 | Apple Inc. | COMPRESSION OF CLOUD OF POINTS |
WO2019142666A1 (ja) * | 2018-01-16 | 2019-07-25 | ソニー株式会社 | 画像処理装置および方法 |
Non-Patent Citations (8)
Title |
---|
GOLLA, TIM ET AL.: "Real-time Point Cloud Compression", 2015 IEEE /RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS, 2015, pages 5087 - 5092, XP032832361, DOI: 10.1109/IROS.2015.7354093 * |
JIANLE CHENELENA ALSHINAGARY J. SULLIVANJENS-RAINERJILL BOYCE: "Algorithm Description of Joint Exploration Test Model 4", JVET- G1001 V1, JOINT VIDEO EXPLORATION TEAM (JVET) OF ITU-T SG 16 WP 3 AND ISO/IEC JTC 1/SC 29/WG 11 7TH MEETING: TORINO, IT, 13 July 2017 (2017-07-13) |
K. MAMMOU: "PCC Test Model Category 2 vO", N17248 MPEG OUTPUT DOCUMENT, October 2017 (2017-10-01) |
K. MAMMOU: "Video-based and Hierarchical Approaches Point Cloud Compression", MPEG M41649, October 2017 (2017-10-01) |
R. MEKURIAK. BLOMP. CESAR., DESIGN, IMPLEMENTATION AND EVALUATION OF A POINT CLOUD CODEC FOR TELE-IMMERSIVE VIDEO |
TELECOMMUNICATION STANDARDIZATION SECTOR OF ITU (INTERNATIONAL TELECOMMUNICATION UNION, ADVANCED VIDEO CODING FOR GENERIC AUDIOVISUAL SERVICES, April 2017 (2017-04-01) |
TELECOMMUNICATION STANDARDIZATION SECTOR OF ITU (INTERNATIONAL TELECOMMUNICATION UNION, HIGH EFFICIENCY VIDEO CODING, December 2016 (2016-12-01) |
TIM GOLLAREINHARD KLEIN: "Real-time Point Cloud Compression", 2015, IEEE |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2021200919A1 (ja) * | 2020-03-30 | 2021-10-07 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカ | 三次元データ符号化方法、三次元データ復号方法、三次元データ符号化装置、及び三次元データ復号装置 |
WO2022149589A1 (ja) * | 2021-01-07 | 2022-07-14 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカ | 三次元データ復号方法、三次元データ符号化方法、三次元データ復号装置、及び三次元データ符号化装置 |
Also Published As
Publication number | Publication date |
---|---|
AU2019302075A1 (en) | 2021-01-28 |
KR20210030276A (ko) | 2021-03-17 |
EP3823286A4 (en) | 2021-12-01 |
CN112470480A (zh) | 2021-03-09 |
US20210233278A1 (en) | 2021-07-29 |
US11699248B2 (en) | 2023-07-11 |
EP3823286A1 (en) | 2021-05-19 |
MX2020013700A (es) | 2021-02-26 |
JPWO2020012968A1 (ja) | 2021-08-02 |
JP7327399B2 (ja) | 2023-08-16 |
BR112021000044A2 (pt) | 2021-03-30 |
PH12021550063A1 (en) | 2021-12-06 |
CA3106234A1 (en) | 2020-01-16 |
TW202017374A (zh) | 2020-05-01 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP7327399B2 (ja) | 画像処理装置および方法 | |
WO2019198523A1 (ja) | 画像処理装置および方法 | |
US11721048B2 (en) | Image processing apparatus and method | |
WO2020012967A1 (ja) | 画像処理装置および方法 | |
US20200359053A1 (en) | Image processing apparatus and image processing method | |
US11310518B2 (en) | Image processing apparatus and method | |
JP7331852B2 (ja) | 画像処理装置および方法 | |
US11399189B2 (en) | Image processing apparatus and method | |
WO2020188932A1 (ja) | 情報処理装置および情報生成方法 | |
JP7396302B2 (ja) | 画像処理装置および方法 | |
WO2020071101A1 (ja) | 画像処理装置および方法 | |
JP2022047546A (ja) | 情報処理装置および方法 | |
US20230370637A1 (en) | Image processing device and method | |
WO2022050088A1 (ja) | 画像処理装置および方法 | |
WO2024057903A1 (ja) | 情報処理装置および方法 | |
WO2022230941A1 (ja) | 情報処理装置および方法 | |
JP2022102267A (ja) | 画像処理装置および方法 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 19834544 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2020530096 Country of ref document: JP Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 3106234 Country of ref document: CA |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
REG | Reference to national code |
Ref country code: BR Ref legal event code: B01A Ref document number: 112021000044 Country of ref document: BR |
|
ENP | Entry into the national phase |
Ref document number: 2019302075 Country of ref document: AU Date of ref document: 20190627 Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 112021000044 Country of ref document: BR Kind code of ref document: A2 Effective date: 20210104 |