WO2019142666A1 - 画像処理装置および方法 - Google Patents
画像処理装置および方法 Download PDFInfo
- Publication number
- WO2019142666A1 WO2019142666A1 PCT/JP2019/000051 JP2019000051W WO2019142666A1 WO 2019142666 A1 WO2019142666 A1 WO 2019142666A1 JP 2019000051 W JP2019000051 W JP 2019000051W WO 2019142666 A1 WO2019142666 A1 WO 2019142666A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- data
- dimensional
- unit
- image processing
- processing apparatus
- Prior art date
Links
- 238000012545 processing Methods 0.000 title claims abstract description 163
- 238000000034 method Methods 0.000 title abstract description 159
- 238000003672 processing method Methods 0.000 claims abstract description 13
- 238000000605 extraction Methods 0.000 claims description 14
- 230000010365 information processing Effects 0.000 abstract description 7
- 238000005516 engineering process Methods 0.000 description 47
- 230000011218 segmentation Effects 0.000 description 42
- 238000012856 packing Methods 0.000 description 34
- 238000004458 analytical method Methods 0.000 description 21
- 230000006835 compression Effects 0.000 description 13
- 238000007906 compression Methods 0.000 description 13
- 238000013459 approach Methods 0.000 description 12
- 238000010586 diagram Methods 0.000 description 9
- 239000000284 extract Substances 0.000 description 7
- 238000000354 decomposition reaction Methods 0.000 description 5
- 238000004891 communication Methods 0.000 description 4
- 238000007796 conventional method Methods 0.000 description 4
- 230000006866 deterioration Effects 0.000 description 4
- 230000006870 function Effects 0.000 description 3
- 230000005540 biological transmission Effects 0.000 description 2
- 238000003384 imaging method Methods 0.000 description 2
- 239000000203 mixture Substances 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 238000012360 testing method Methods 0.000 description 2
- VBRBNWWNRIMAII-WYMLVPIESA-N 3-[(e)-5-(4-ethylphenoxy)-3-methylpent-3-enyl]-2,2-dimethyloxirane Chemical compound C1=CC(CC)=CC=C1OC\C=C(/C)CCC1C(C)(C)O1 VBRBNWWNRIMAII-WYMLVPIESA-N 0.000 description 1
- 101000638069 Homo sapiens Transmembrane channel-like protein 2 Proteins 0.000 description 1
- 102100032054 Transmembrane channel-like protein 2 Human genes 0.000 description 1
- 230000003796 beauty Effects 0.000 description 1
- 230000010267 cellular communication Effects 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 238000011143 downstream manufacturing Methods 0.000 description 1
- 230000010354 integration Effects 0.000 description 1
- 244000144972 livestock Species 0.000 description 1
- 238000005065 mining Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 230000002265 prevention Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T11/00—2D [Two Dimensional] image generation
- G06T11/003—Reconstruction from projections, e.g. tomography
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T9/00—Image coding
- G06T9/001—Model-based coding, e.g. wire frame
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/50—Depth or shape recovery
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T9/00—Image coding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/597—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10028—Range image; Depth image; 3D point clouds
Definitions
- the present disclosure relates to an image processing apparatus and method, and more particularly to an image processing apparatus and method capable of suppressing reduction in quality due to two-dimensional projection of 3D data.
- the projection onto a two-dimensional plane may be difficult in some cases, and there is a possibility that the quality may be reduced by the encoding accompanied by the projection onto the two-dimensional plane.
- This indication is made in view of such a situation, and enables it to control quality reduction by two-dimensional projection of 3D data.
- An image processing apparatus is an image processing apparatus including a two-dimensional projection unit that projects data at every position included in 3D data representing a three-dimensional structure on a two-dimensional plane of a plurality of layers.
- An image processing method is an image processing method for projecting data at every position included in 3D data representing a three-dimensional structure on a two-dimensional plane of a plurality of layers.
- An image processing apparatus includes a three-dimensional projection unit that projects, to a three-dimensional space, data at every position of 3D data projected onto a two-dimensional plane with the number of layers indicated by the layer number information. It is an image processing device.
- An image processing method is an image processing method of projecting data of every position of 3D data projected onto a two-dimensional plane of the number of layers indicated by the layer number information into a three-dimensional space.
- data for every position included in 3D data representing a three-dimensional structure is projected onto a two-dimensional plane of a plurality of layers.
- data for every position of 3D data projected on a two-dimensional plane of the number of layers indicated by the layer number information is projected on a three-dimensional space.
- information can be processed.
- it is possible to suppress the reduction in quality due to two-dimensional projection of 3D data.
- Non-Patent Document 1 (described above)
- Non-Patent Document 2 (described above)
- Non-Patent Document 3 (described above)
- Non-patent document 4 TELECOM MUNICATION STANDARDIZATION SECTOR OF ITU (International Telecommunication Union), "Advanced video coding for generic audiovisual services", H.264, 04/2017
- Non-Patent Document 5 TELECOM MUNICATION STANDARDIZATION SECTOR OF ITU (International Telecommunication Union), "High efficiency video coding", H. 265, 12/2016
- Non-Patent Document 6 Jianle Chen, Maria Alshina, Gary J.
- Non-Patent Document 5 the Quad Tree Plus Binary Tree (QTBT) Block Structure described in Non-Patent Document 6
- this embodiment is not limited thereto. It shall be within the scope of disclosure of technology and shall meet the support requirements of the scope of claims. Also, technical terms such as, for example, Parsing, Syntax, Semantics, etc. are also within the disclosure scope of the present technology, even if they are not directly described in the embodiments. To meet the support requirements of the range.
- ⁇ Point cloud> Conventionally, there has been data such as a point cloud representing a three-dimensional structure based on position information and attribute information of a point group, and data such as meshes that define a three-dimensional shape using polygons, composed of vertices, edges, and faces. .
- a three-dimensional structure as shown in A of FIG. 1 is expressed as a set of many points (point group) as shown in B of FIG. That is, the data of the point cloud is composed of position information and attribute information (for example, color etc.) of each point of this point group. Therefore, while the data structure is relatively simple, an arbitrary three-dimensional structure can be represented with sufficient accuracy by using a sufficiently large number of points.
- a video-based approach has been proposed in which the position and color information of such a point cloud are projected on a two-dimensional plane for each small area and encoded by an encoding method for a two-dimensional image. .
- an input Point cloud is divided into a plurality of segmentations (also referred to as regions or patches), and projected into a two-dimensional plane for each region.
- the data for each position of the point cloud (that is, the data for each point) is composed of position information (also referred to as geometry (also referred to as depth)) and attribute information (texture) as described above. Projected onto a dimensional plane.
- 3D data (point cloud) projected on a two-dimensional plane is encoded by a coding method for a two-dimensional plane image, such as AVC (Advanced Video Coding) or HEVC (High Efficiency Video Coding), for example.
- AVC Advanced Video Coding
- HEVC High Efficiency Video Coding
- FIG. 3 is a list of the present technology described in each embodiment.
- the first row from the top of this table (except for the item name row) describes the video-based approach in the conventional (TMC2). That is, in the conventional video based approach, two-dimensional projection of 3D data is performed on a two-dimensional (two-layer) plane. This specification was common to the whole screen (the same projection was performed for every segment). Therefore, projection control information used to control such projection is signaled from the encoding side to the decoding side as information in units of frames.
- the point cloud to be encoded, there are also points other than the object surface due to noise and the characteristics of the imaging system. Therefore, the projection onto a two-dimensional plane of two layers as in the above-described conventional method may be difficult. Therefore, there are points that can not be projected on a two-dimensional plane, and there is a possibility that the quality of data may be reduced by encoding with the projection on the two-dimensional plane.
- the second row from the top of the table in FIG. 3 (except for the row of item names) describes the present technology described in the first embodiment (Example 1).
- the number of layers in the two-dimensional plane on which 3D data is projected can be made variable, and the number of layers can be projected so that all data (data at each position) of all points overlapping in the depth direction can be projected. It is a method to set.
- information indicating the number of layers may be signaled from the encoding side to the decoding side for each area.
- the third row from the top of the table in FIG. 3 (excluding the row of item names) describes the present technology described in the second embodiment (Example 2).
- This method is a method of adding the definition of "absent point" when projecting 3D data onto a two-dimensional plane.
- the definition of the pixel value of the point to be deleted on the decoding side may be signaled from the encoding side to the decoding side.
- the fourth row from the top of the table in FIG. 3 (excluding the row of item names) describes the present technology described in the third embodiment (Example 3).
- This method is a method for enabling setting of depth parameters for controlling the depth range of 3D data projected onto a two-dimensional plane for each area.
- the depth parameter may be signaled from the encoding side to the decoding side.
- FIG. 4 is a block diagram showing an example of a configuration of an encoding apparatus which is an aspect of an image processing apparatus to which the present technology is applied.
- the encoding apparatus 100 shown in FIG. 4 is an apparatus that projects 3D data such as a point cloud onto a two-dimensional plane and performs encoding using a two-dimensional image encoding method.
- the encoding apparatus 100 implements the techniques described in Non-Patent Document 1 to Non-Patent Document 6, and encodes 3D data in a method according to the standard described in any of those documents. .
- FIG. 4 shows main processing units, data flows, etc., and the one shown in FIG. 4 is not limited to all. That is, in the encoding device 100, there may be a processing unit not shown as a block in FIG. 4, or there may be processing or data flow not shown as an arrow or the like in FIG. The same applies to the other drawings for explaining the processing unit and the like in the encoding device 100.
- the encoding apparatus 100 includes a patch decomposing unit 111, a packing unit 112, an auxiliary patch information compression unit 113, a video encoding unit 114, a video encoding unit 115, an OMap encoding unit 116, and a multiplexer 117.
- a patch decomposing unit 111 the encoding apparatus 100 includes a patch decomposing unit 111, a packing unit 112, an auxiliary patch information compression unit 113, a video encoding unit 114, a video encoding unit 115, an OMap encoding unit 116, and a multiplexer 117.
- the patch disassembly unit 111 performs a process related to the disassembly of 3D data. For example, the patch decomposition unit 111 acquires 3D data (for example, point cloud) representing a three-dimensional structure, which is input to the encoding device 100 (arrow 121). Also, the patch disassembly unit 111 resolves the acquired 3D data into a plurality of patches, and projects the 3D data onto a two-dimensional plane for each of the patches.
- 3D data for example, point cloud
- the patch disassembly unit 111 supplies 3D data projected on a two-dimensional plane for each patch to the packing unit 112 (arrow 122). Also, the patch disassembling unit 111 supplies auxiliary patch information, which is information related to the disassembly, to the auxiliary patch information compression unit 113 (arrow 123).
- the packing unit 112 performs processing related to data packing. For example, the packing unit 112 acquires data of a two-dimensional plane on which 3D data is projected for each patch supplied from the patch disassembly unit 111 (arrow 122). Also, the packing unit 112 packs the acquired layers of the two-dimensional plane as different video frames. For example, the packing unit 112 may include position information (Gepmetry) indicating the position of the point, attribute information (Texture) such as color information added to the position information, and an occupancy map (Occupancy Map) indicating the presence or absence of the point. ), Each packed as a video frame.
- position information indicating the position of the point
- attribute information such as color information added to the position information
- Occupancy Map occupancy map
- the packing unit 112 supplies the generated video frame to the downstream processing unit (arrow 124). For example, the packing unit 112 supplies the video frame of the generated location information (Geometry) to the video encoding unit 114. Also, for example, the packing unit 112 supplies the generated video frame of the attribute information (Texture) to the video encoding unit 115. Furthermore, for example, the packing unit 112 supplies the generated Ocupancy map video frame to the OMap encoding unit 116.
- the packing unit 112 supplies the generated video frame to the downstream processing unit (arrow 124). For example, the packing unit 112 supplies the video frame of the generated location information (Geometry) to the video encoding unit 114. Also, for example, the packing unit 112 supplies the generated video frame of the attribute information (Texture) to the video encoding unit 115. Furthermore, for example, the packing unit 112 supplies the generated Ocupancy map video frame to the OMap encoding unit 116.
- the packing unit 112 also supplies control information on the packing to the multiplexer 117 (arrow 125).
- the auxiliary patch information compression unit 113 performs processing relating to compression of auxiliary patch information. For example, the auxiliary patch information compression unit 113 acquires data supplied from the patch decomposition unit 111 (arrow 123). The auxiliary patch information compression unit 113 encodes (compresses) auxiliary patch information included in the acquired data. The auxiliary patch information compression unit 113 supplies the encoded data of the obtained auxiliary patch information to the multiplexer 117 (arrow 126).
- the video encoding unit 114 performs processing relating to encoding of a video frame of location information (Geometry). For example, the video encoding unit 114 obtains a video frame of location information (Geometry) supplied from the packing unit 112 (arrow 124). Also, the video encoding unit 114 encodes the obtained video frame of the position information (Geometry) by an encoding method for an arbitrary two-dimensional image such as AVC or HEVC. The video encoding unit 114 supplies the encoded data (encoded data of a video frame of position information (Geometry)) obtained by the encoding to the multiplexer 117 (arrow 127).
- the video encoding unit 115 performs processing regarding encoding of a video frame of attribute information (Texture). For example, the video encoding unit 115 obtains a video frame of attribute information (Texture) supplied from the packing unit 112 (arrow 124). Also, the video encoding unit 115 encodes the obtained video frame of the attribute information (Texture) according to an arbitrary two-dimensional image encoding method such as AVC or HEVC. The video encoding unit 115 supplies the encoded data (encoded data of a video frame of attribute information (Texture)) obtained by the encoding to the multiplexer 117 (arrow 128).
- a video frame of attribute information (Texture) supplied from the packing unit 112 (arrow 124).
- the video encoding unit 115 encodes the obtained video frame of the attribute information (Texture) according to an arbitrary two-dimensional image encoding method such as AVC or HEVC.
- the video encoding unit 115 supplies the encoded data (encoded data of a video frame
- the OMap encoding unit 116 performs processing relating to encoding of a video frame of an occupancy map. For example, the OMap encoding unit 116 obtains a video frame of the occupancy map supplied from the packing unit 112 (arrow 124). In addition, the OMap encoding unit 116 encodes the video frame of the acquired occupancy map according to an arbitrary two-dimensional image encoding method such as AVC or HEVC. The OMap encoding unit 116 supplies the encoded data (encoded data of the video frame of the occupancy map) obtained by the encoding to the multiplexer 117 (arrow 129).
- the multiplexer 117 performs processing related to multiplexing. For example, the multiplexer 117 obtains encoded data of the auxiliary patch information supplied from the auxiliary patch information compression unit 113 (arrow 126). Also, for example, the multiplexer 117 obtains control information on packing supplied from the packing unit 112 (arrow 125). Also, for example, the multiplexer 117 obtains encoded data of a video frame of position information (Geometry) supplied from the video encoding unit 114 (arrow 127). Also, for example, the multiplexer 117 obtains encoded data of a video frame of attribute information (Texture) supplied from the video encoding unit 115 (arrow 128). Also, for example, the multiplexer 117 acquires encoded data of the video frame of the occupancy map supplied from the OMap encoding unit 116 (arrow 129).
- the multiplexer 117 obtains encoded data of the video frame of the occupancy map supplied from the OMap encoding unit 116 (arrow 129).
- the multiplexer 117 multiplexes the acquired pieces of information to generate a bit stream.
- the multiplexer 117 outputs the generated bit stream to the outside of the coding apparatus 100 (arrow 130).
- FIG. 5 is a block diagram showing a main configuration example of the patch disassembly unit 111.
- the patch decomposing unit 111 in this case includes a normal direction estimation unit 151, a segmentation initial setting unit 152, a segmentation updating unit 153, a two-dimensional projection unit 154, and a pixel distribution analysis unit 155.
- the normal direction estimation unit 151 performs processing regarding estimation of the normal direction of the surface of the 3D data. For example, the normal direction estimation unit 151 acquires input 3D data. Further, the normal direction estimation unit 151 estimates the normal direction of the surface of the object represented by the acquired 3D data. For example, the normal direction estimation unit 151 constructs a kd-tree, searches for a neighborhood, and calculates an optimal approximate tangent plane to estimate the normal direction. The normal direction estimation unit 151 supplies the estimation result of the normal direction to the segmentation initial setting unit 152 along with other data.
- the segmentation initialization unit 152 performs processing related to the segmentation initialization. For example, the segmentation initial setting unit 152 acquires data supplied from the normal direction estimation unit 151. Also, for example, the segmentation initial setting unit 152 sets the plane corresponding to the normal direction of the 3D data based on the component of each of the six axes in the normal direction estimated by the normal direction estimation unit 151. Classify. The segmentation initial setting unit 152 supplies the classification result together with other data to the segmentation update unit 153.
- the segmentation update unit 153 performs processing related to segmentation update. For example, the segmentation update unit 153 acquires data supplied from the segmentation initial setting unit 152. Then, the segmentation update unit 153 collects too small regions in the initial setting segmentation set by the segmentation initial setting unit 152 so as to be a sufficiently large region. The segmentation update unit 153 supplies the updated information on the segmentation to the two-dimensional projection unit 154 along with other information.
- the two-dimensional projection unit 154 performs processing relating to two-dimensional projection of 3D data. For example, the two-dimensional projection unit 154 acquires data supplied from the segmentation update unit 153. Further, the two-dimensional projection unit 154 generates an occupancy map of 3D data included in the acquired data using the pixel distribution analysis unit 155, or generates a two-dimensional plane of the 3D data and the occupancy data for each area. Project to The two-dimensional projection unit 154 supplies the 3D data projected on the two-dimensional plane to the packing unit 112 along with other data.
- the pixel distribution analysis unit 155 performs processing regarding analysis of pixel distribution of 3D data to be processed by the two-dimensional projection unit 154.
- FIG. 6 is a block diagram illustrating an example of a configuration of a decoding device which is an aspect of an image processing device to which the present technology is applied.
- the decoding apparatus 200 shown in FIG. 6 decodes encoded data obtained by projecting and encoding 3D data such as a point cloud on a two-dimensional plane by a decoding method for a two-dimensional image, and projecting the data in a three-dimensional space Device.
- the decoding apparatus 200 implements the techniques described in Non-Patent Document 1 to Non-Patent Document 6, and decodes encoded data of 3D data in a method compliant with the standard described in any of those documents. I do.
- FIG. 6 shows main processing units, data flows, etc., and the one shown in FIG. 6 is not limited to all. That is, in the decoding apparatus 200, there may be a processing unit not shown as a block in FIG. 6, or there may be a process or data flow not shown as an arrow or the like in FIG. The same applies to the other drawings for explaining the processing unit and the like in the decoding apparatus 200.
- the decoding apparatus 200 includes a demultiplexer 211, an auxiliary patch information decoding unit 212, a video decoding unit 213, a video decoding unit 214, an OMap decoding unit 215, an unpacking unit 216, and a 3D reconstruction unit 217.
- a demultiplexer 211 the demultiplexer 211
- an auxiliary patch information decoding unit 212 the decoding apparatus 200
- a video decoding unit 213, a video decoding unit 214 includes a video decoding unit 214, an OMap decoding unit 215, an unpacking unit 216, and a 3D reconstruction unit 217.
- the demultiplexer 211 performs processing related to data demultiplexing. For example, the demultiplexer 211 obtains a bit stream input to the decoding device 200. This bit stream is supplied from, for example, the coding apparatus 100. The demultiplexer 211 demultiplexes this bit stream, extracts the encoded data of the auxiliary patch information, and supplies it to the auxiliary patch information decoding unit 212. Further, the demultiplexer 211 extracts encoded data of a video frame of position information (Geometory) from the bit stream by demultiplexing, and supplies it to the video decoding unit 213.
- a video frame of position information GPU
- the demultiplexer 211 extracts encoded data of the video frame of the attribute information (Texture) from the bit stream by demultiplexing, and supplies it to the video decoding unit 214. Also, the demultiplexer 211 extracts the encoded data of the video frame of the occupancy map from the bit stream by demultiplexing, and supplies this to the OMap decoding unit 215.
- the auxiliary patch information decoding unit 212 performs processing regarding decoding of encoded data of auxiliary patch information. For example, the auxiliary patch information decoding unit 212 acquires encoded data of the auxiliary patch information supplied from the demultiplexer 211. In addition, the auxiliary patch information decoding unit 212 decodes encoded data of auxiliary patch information included in the acquired data. The auxiliary patch information decoding unit 212 supplies the auxiliary patch information obtained by the decoding to the 3D reconstruction unit 217.
- the video decoding unit 213 performs processing relating to decoding of encoded data of a video frame of position information (Geometory). For example, the video decoding unit 213 acquires encoded data of a video frame of position information (Geometory) supplied from the demultiplexer 211. Also, for example, the video decoding unit 213 decodes the acquired encoded data to obtain a video frame of position information (Geometory). The video decoding unit 213 supplies the video frame of the position information (Geometory) to the unpacking unit 216.
- the video decoding unit 214 performs processing related to decoding of encoded data of a video frame of attribute information (Texture). For example, the video decoding unit 214 acquires encoded data of a video frame of attribute information (Texture) supplied from the demultiplexer 211. Also, for example, the video decoding unit 214 decodes the acquired encoded data to obtain a video frame of attribute information (Texture). The video decoding unit 214 supplies the video frame of the attribute information (Texture) to the unpacking unit 216.
- a video frame of attribute information (Texture) supplied from the demultiplexer 211.
- the video decoding unit 214 decodes the acquired encoded data to obtain a video frame of attribute information (Texture).
- the video decoding unit 214 supplies the video frame of the attribute information (Texture) to the unpacking unit 216.
- the OMap decoding unit 215 performs processing regarding decoding of encoded data of a video frame of an occupancy map. For example, the OMap decoding unit 215 acquires encoded data of a video frame of the occupancy map supplied from the demultiplexer 211. Also, for example, the OMap decoding unit 215 decodes the acquired encoded data to obtain a video frame of an occupancy map. The OMap decoding unit 215 supplies the video frame of the occupancy map to the unpacking unit 216.
- the unpacking unit 216 performs processing related to unpacking. For example, the unpacking unit 216 acquires a video frame of position information (Geometory) from the video decoding unit 213, acquires a video frame of attribute information (Texture) from the video decoding unit 214, and an occupancy map from the OMap decoding unit 215. Get the video frame of. The unpacking unit 216 unpacks these video frames. The unpacking unit 216 supplies the data of the location information (Geometory), the data of the attribute information (Texture), and the data of the occupancy map obtained by the unpacking to the 3D reconstruction unit 217.
- the 3D reconstruction unit 217 performs processing regarding reconstruction of 3D data.
- the 3D reconstruction unit 217 may use auxiliary patch information supplied from the auxiliary patch information decoding unit 212, data of location information (Geometory) supplied from the unpacking unit 216, data of attribute information (Texture), Reconstruct 3D data based on data of Pancy map etc.
- the 3D reconstruction unit 217 outputs the 3D data obtained by such processing to the outside of the decoding device 200.
- the 3D data is, for example, supplied to the display unit and the image is displayed, recorded on a recording medium, or supplied to another apparatus via communication.
- FIG. 7 is a block diagram showing an example of the main configuration of the 3D reconstruction unit 217 of FIG.
- the 3D reconstruction unit 217 includes a three-dimensional projection unit 251, a pixel distribution analysis unit 252, a reverse segmentation update unit 253, a reverse segmentation initial setting unit 254, and a reverse normal direction estimation unit 255. .
- the three-dimensional projection unit 251 projects, onto a three-dimensional space, 3D data projected on a two-dimensional plane for each area.
- the pixel distribution analysis unit 252 performs processing such as analysis of the pixel distribution when projecting the three-dimensional projection unit 251 to the three-dimensional space.
- the reverse segmentation update unit 253 performs reverse processing of the segmentation update unit 153.
- the reverse segmentation initial setting unit 254 performs reverse processing of the segmentation initial setting unit 152.
- the reverse normal direction estimation unit 255 performs reverse processing of the normal direction estimation unit 151.
- First embodiment> ⁇ Variable number of layers in 2D plane to project 3D data>
- 3D data is projected on a two-dimensional plane of two layers (layer 0 and layer 1) as in the example shown in FIG.
- Layer 0 (Layer 0) is projected with data of points on the surface viewed from the projection plane of 3D data.
- Layer 1 Layer 1
- the difference value of the distance from the layer 0 is the pixel value.
- the number of layers in a two-dimensional plane on which 3D data is projected may be variable. For example, data at every position included in 3D data representing a three-dimensional structure is projected onto a two-dimensional plane of a plurality of layers.
- the image processing apparatus is provided with a two-dimensional projection unit that projects data for every position included in 3D data representing a three-dimensional structure on a two-dimensional plane of a plurality of layers.
- the two-dimensional projection unit 154 may project data of each 3D data at each position where the position overlaps in the depth direction as viewed from the projection plane onto different layers of the two-dimensional plane of the plurality of layers. .
- the two-dimensional projection unit 154 may generate the same number of layers as the maximum number of data at each position where the position overlaps in the depth direction when viewed from the projection plane of the 3D data on this two-dimensional plane. .
- the maximum number of data overlapping in the depth direction is 4 in the 3D data of the region (Local Bounding Box). Therefore, this 3D data is projected onto a two-dimensional plane of four layers (layer 0 to layer 3).
- information indicating the number of layers in the two-dimensional plane on which the three-dimensional data is projected by the two-dimensional projection unit 154 may be signaled to the bit stream. That is, the multiplexer 117 functioning as a bit stream generation unit has information indicating the number of layers of the two-dimensional plane on which the 3D data is projected by the two-dimensional projection unit 154, and the two-dimensional plane is encoded by the video encoding unit 114 or the like. A bit stream including the encoded data obtained as a result is generated.
- the 3D data projected onto all the layers of the two-dimensional plane can be easily projected to the three-dimensional space by referring to the information indicating the number of layers of this two-dimensional plane. be able to.
- the patch deconstructing unit 111 of the encoding device 100 deconstructs 3D data into patches and projects the data of each patch on a two-dimensional plane in step S101.
- the auxiliary patch information compression unit 113 compresses the auxiliary patch information obtained by the process of step S101.
- step S103 the packing unit 112 packs the 3D data projected on the two-dimensional plane for each patch by the patch disassembly unit 111 as a video frame.
- step S104 the video encoding unit 114 encodes a geometry video frame, which is a video frame of the position information obtained by the process of step S103, using the encoding method for a two-dimensional image.
- step S105 the video encoding unit 114 encodes a color video frame, which is a video frame of the attribute information obtained by the process of step S103, using the encoding method for a two-dimensional image.
- step S106 the video encoding unit 114 encodes the occupancy map obtained by the process of step S103 according to the encoding method for a two-dimensional image.
- step S107 the multiplexer 117 multiplexes the various information generated as described above, and generates a bit stream including the information.
- step S108 the multiplexer 117 outputs the bit stream generated by the process of step S107 to the outside of the encoding apparatus 100.
- step S108 ends, the encoding process ends.
- the normal direction estimation unit 151 estimates the normal direction in step S121.
- the segmentation initialization unit 152 performs segmentation initialization.
- the segmentation update unit 153 updates the segmentation of the initial state set in step S122 as necessary.
- the two-dimensional projection unit 154 projects the 3D data onto a two-dimensional plane.
- step S124 ends, the patch disassembly process ends, and the process returns to FIG.
- step S143 the pixel distribution analysis unit 155 determines whether or not there are unprojected pixels (data for each position of 3D data not projected on a two-dimensional plane). If it is determined that there is an unprojected pixel, the process proceeds to step S144.
- step S144 the two-dimensional projection unit 154 projects the area onto the layer i to be processed.
- step S145 the two-dimensional projection unit 154 increments the variable i (i ++).
- step S143 If it is determined in step S143 that there are no unprojected pixels (all the pixels in the area have been projected), the process proceeds to step S146.
- step S146 the two-dimensional projection unit 154 supplies and encodes information indicating the number of layers i. Also, in step S147, the two-dimensional projection unit 154 encodes the geometry image for the corresponding i frame. That is, the two-dimensional projection unit 154 supplies the two-dimensional plane on which the 3D data is projected to the packing unit 112, and packs the layers as different frames.
- step S147 ends, the two-dimensional projection process ends, and the process returns to FIG.
- the image processing apparatus is provided with a three-dimensional projection unit that projects, on a three-dimensional space, data of every position of 3D data projected onto a two-dimensional plane with the number of layers indicated by the layer number information.
- the demultiplexer 211 of the decoding apparatus 200 demultiplexes the bit stream in step S201.
- step S202 the auxiliary patch information decoding unit 212 decodes the auxiliary patch information extracted from the bit stream in the process of step S201.
- step S203 the video decoding unit 213 decodes the encoded data of the geometry video frame (video frame of position information) extracted from the bit stream in the process of step S201.
- step S204 the video decoding unit 214 decodes the encoded data of the color video frame (video frame of attribute information) extracted from the bit stream in the process of step S201.
- step S205 the OMap decoding unit 215 decodes the encoded data of the occupancy map extracted from the bit stream in the process of step S201.
- step S206 the unpacking unit 216 unpacks the geometry video frame, the color video frame, and the occupancy map decoded in steps S203 to S205, respectively.
- step S207 the 3D reconstruction unit 217 reconstructs 3D data, such as a point cloud, based on the auxiliary patch information obtained in step S202 and the various information obtained in step S206.
- step S207 ends, the decoding process ends.
- the three-dimensional projection unit 251 projects a two-dimensional image on the three-dimensional space in step S221.
- step S222 the reverse segmentation update unit 253 reversely updates the segmentation, and separates the summarized segmentation.
- step S223 the reverse segmentation initial setting unit 254 performs reverse processing of the initial setting of segmentation, and collects classified points.
- step S224 the antinormal direction estimation unit 255 performs inverse processing of normal direction estimation to reconstruct a point cloud.
- step S224 ends, the point cloud reconstruction process ends, and the process returns to FIG.
- the three-dimensional projection unit 251 decodes information indicating the number of layers i in step S241.
- step S242 the three-dimensional projection unit 251 decodes the geometry image for the corresponding i frame.
- step S243 the pixel distribution analysis unit 252 initializes the variable k to 0.
- step S244 the pixel distribution analysis unit 252 determines whether or not the variable k ⁇ i. If it is determined that k ⁇ i, the process proceeds to step S245.
- step S245 the three-dimensional projection unit 251 projects the 3D data of the layer k (Layer k) onto the three-dimensional space.
- step S246 the pixel distribution analysis unit 252 increments the variable k (k ++).
- step S246 ends, the process returns to step S244.
- step S244 If it is determined in step S244 that the variable k is not less than i, the three-dimensional projection process ends, and the process returns to FIG.
- Second embodiment> ⁇ Signal of a value indicating "non-pixel">
- pixels where the 3D data is not projected on the two-dimensional plane that is, pixels where pixel values are not set (also referred to as absent pixels).
- pixel values are complemented with respect to such “missing pixels”.
- the pixel value is copied from the pixel on the left thereof (pixel complementation is performed).
- pixel complementation is performed because the encoding method for two-dimensional images (for example, AVC, HEVC, etc.) has no concept of no data (blank).
- the data for each position included in the 3D data representing the three-dimensional structure is projected onto a two-dimensional plane, and it is indicated that the data for each position does not exist at the position where the data for each position of the two-dimensional plane does not exist.
- Set a predetermined value For example, in an image processing apparatus, data for each position included in 3D data representing a three-dimensional structure is projected on a two-dimensional plane, and data for each position is not available at each position of the two-dimensional plane.
- a two-dimensional projection unit is provided which sets a predetermined value indicating absence.
- a specified value X is used as a pixel value indicating “non-pixel” on a two-dimensional plane on which 3D data is projected.
- the value of X may be, for example, a predetermined fixed value.
- X may be set to 255 (upper limit of 8 bits).
- an arbitrary value of X> D may be used as the value of X.
- D indicates the maximum value of the depth of the bounding box.
- a pixel value of a two-dimensional plane can not be set to a value greater than or equal to the maximum value of the depth of the bounding box. Therefore, the unused value may be used as X.
- the encoding device 100 can express “non-pixel” without requiring data complementation, so that deterioration of 3D data can be suppressed. That is, it is possible to suppress the reduction in quality due to two-dimensional projection of 3D data.
- the encoding device 100 can reduce the number of “non-encoding pixels” in the two-dimensional plane, so the region is set larger compared to the case where the complementing process is not performed. be able to. Therefore, the projection in a minute area (Point) unit becomes unnecessary, and it is possible to suppress an increase in processing amount and a reduction in encoding efficiency.
- the two-dimensional projection unit 154 When the two-dimensional projection processing is started, the two-dimensional projection unit 154 performs region extraction by segmentation in step S301. In step S302, the pixel distribution analysis unit 155 determines whether the area includes a sparse point cloud. If it is determined that the area includes a sparse point cloud, the process proceeds to step S303.
- step S303 the two-dimensional projection unit 154 signals information defining “non-pixel” in the area. That is, the above-described X is set, and the “non-existent point” projected on the two-dimensional plane is expressed using the X.
- step S305 the process proceeds to step S305.
- step S302 If it is determined in step S302 that the area does not include the sparse point cloud, the process proceeds to step S304.
- step S304 the two-dimensional projection unit 154 projects the region on each layer of the two-dimensional plane.
- step S305 the process proceeds to step S305.
- step S305 the two-dimensional projection unit 154 encodes the geometry image of the area.
- step S305 ends, the two-dimensional projection process ends, and the process returns to FIG.
- the encoding device 100 can express “non-pixel” without requiring data complementation. Therefore, deterioration of 3D data can be suppressed. That is, it is possible to suppress the reduction in quality due to two-dimensional projection of 3D data.
- the encoding device 100 can reduce the number of “non-encoding pixels” in the two-dimensional plane, so the region is set larger compared to the case where the complementing process is not performed. be able to. Therefore, the projection in a minute area (Point) unit becomes unnecessary, and it is possible to suppress an increase in processing amount and a reduction in encoding efficiency.
- data other than data of a predetermined value indicating that data at each position does not exist in three-dimensional space Project to For example, in the image processing apparatus, among data of each position included in 3D data representing a three-dimensional structure projected on a two-dimensional plane, data other than data of a predetermined value indicating that data at each position does not exist Is provided with a three-dimensional projection unit that projects into three-dimensional space.
- the number of “non-codeable pixels” in the two-dimensional plane can be reduced, so the area can be set larger compared to the case where the complementing process is not performed. Therefore, the projection in a minute area (Point) unit becomes unnecessary, and it is possible to suppress an increase in processing amount and a reduction in encoding efficiency.
- the three-dimensional projection unit 251 decodes the geometry image of the area to be processed in step S321.
- step S322 the pixel distribution analysis unit 252 determines whether or not information (a pixel value indicating “non-pixel”) defining “non-pixel” is present in the geometry image. If it is determined that there is, the process proceeds to step S323.
- step S323 the three-dimensional projection unit 251 deletes the pixel of the pixel value indicating the detected “non-pixel” (setting so as not to be projected in the three-dimensional space).
- step S323 ends, the process proceeds to step S324.
- step S322 If it is determined in step S322 that there is no pixel value indicating "non-existent pixel", the process of step S323 is omitted, and the process proceeds to step S324.
- detection of a pixel value indicating “a non-pixel” is performed on the geometry image of the area to be processed, and if it is detected, the pixel is deleted.
- step S324 the three-dimensional projection unit 251 projects the geometry image of the processing target area onto the three-dimensional space.
- step S324 ends, the three-dimensional projection process ends, and the process returns to FIG.
- the decoding device 200 can express “no pixels” without requiring data complementation, so that deterioration of 3D data can be suppressed. . That is, it is possible to suppress the reduction in quality due to two-dimensional projection of 3D data.
- the decoding device 200 can reduce the number of “non-coding pixels” in the two-dimensional plane, so setting the area larger compared to the case where only the complementing process is not performed. Can. Therefore, the projection in a minute area (Point) unit becomes unnecessary, and it is possible to suppress an increase in processing amount and a reduction in encoding efficiency.
- a depth parameter th that controls the range in the depth direction of the 3D data projected onto the two-dimensional plane is used. Since the points in the range specified by the depth parameter th are to be projected onto the two-dimensional plane, the value of the depth parameter th is related to the length in the depth direction of the region (Local bounding box). For example, when the value of the depth parameter th is larger than the length in the depth direction of the area, points in other areas may be projection targets. That is, the length in the depth direction of the area needs to be longer than the depth parameter th.
- this depth parameter th is controlled on a frame basis.
- all depth parameters th in a frame have a common value.
- the value of the depth parameter th is not always optimal, and there is a possibility that the coding efficiency may be reduced. For example, there may be a case where the coding efficiency is better by dividing the area in the depth direction, but the value of the depth parameter th is large and the division can not be performed.
- the image processing apparatus when data for each position of 3D data representing a three-dimensional structure is projected on a two-dimensional plane for each predetermined area in three-dimensional space, it can be projected on one layer set for each area A two-dimensional projection unit that projects, on a two-dimensional plane, data at each position within a range in the depth direction indicated by a depth parameter that limits the range in the depth direction of data at each position of 3D data representing a three-dimensional structure. To do.
- the depth parameter is expanded so as to be able to be transmitted for each patch, and the position (TH) of a pixel to be projected onto a layer is transmitted for each area.
- the depth parameter th can be set for the area 1 and the depth parameter th ′ can be set for the area 2. Therefore, the efficiency (Geometry coding efficiency) of layer 1 (Layer 1) is improved.
- the pixel distribution analysis unit 155 initializes the area number i to 0 in step S401.
- step S402 the pixel distribution analysis unit 155 determines whether there is an unprocessed pixel. If it is determined that there is an unprocessed pixel, the process proceeds to step S403.
- step S403 the two-dimensional projection unit 154 extracts the area i by segmentation.
- step S404 the two-dimensional projection unit 154 performs adjustment of the depth parameter and encoding (RD determination). That is, the setting of the optimum depth parameter th and the range of the area is performed by the RD determination.
- step S405 the two-dimensional projection unit 154 projects the geometry image of the processing target area i on the two-dimensional surface based on the setting performed in step S404.
- step S406 the two-dimensional projection unit 154 encodes the geometry image projected on the two-dimensional plane.
- step S407 the pixel distribution analysis unit 155 increments the variable i (i ++).
- the process of step S407 ends, the process returns to step S402, and the subsequent processes are repeated.
- step S402 When it is determined in step S402 that there is no unprocessed pixel, the process proceeds to step S408.
- step S408 the two-dimensional projection unit 154 encodes an occlusion map.
- the encoding device 100 can obtain the depth parameter th more suitable for each area, and can suppress the reduction in encoding efficiency.
- the data for each position included in the 3D data representing the three-dimensional structure projected on the two-dimensional plane is projected on the three-dimensional space for each predetermined region of the three-dimensional space
- the data for each position is To project within the range in the depth direction indicated by the depth parameter that limits the range in the depth direction of the data of each position of the 3D data that can be projected onto one layer set for each region in the three-dimensional space Do.
- the image processing apparatus when the data for each position included in the 3D data representing the three-dimensional structure projected on the two-dimensional plane is projected onto the three-dimensional space for each predetermined region of the three-dimensional space Within the range in the depth direction indicated by the depth parameter that limits the range in the depth direction of the data for each position of 3D data that can be projected to one layer set for each region in the three-dimensional space to provide a three-dimensional projection unit that projects onto
- the three-dimensional projection unit 251 decodes the occlusion map in step S421.
- step S422 the pixel distribution analysis unit 252 initializes the region number k to 0.
- step S423 the pixel distribution analysis unit 252 determines whether the variable k ⁇ i. If it is determined that k ⁇ i, the process proceeds to step S424.
- step S424 the three-dimensional projection unit 251 decodes the depth parameter th.
- step S425 the three-dimensional projection unit 251 decodes the geometry image.
- step S426 the three-dimensional projection unit 251 projects the image of the area onto a three-dimensional space.
- step S427 the pixel distribution analysis unit 252 increments the variable k (k ++).
- step S427 ends, the process returns to step S423, and the subsequent processes are repeated.
- step S423 If it is determined in step S423 that k does not satisfy i ⁇ i, the three-dimensional projection process ends, and the process returns to FIG.
- the decoding apparatus 200 can perform projection of 3D data into the three-dimensional space by using the depth parameter th set for each area. The reduction of the efficiency can be suppressed.
- control information related to the present technology described in each of the above embodiments may be transmitted from the encoding side to the decoding side.
- control information for example, enabled_flag
- control for specifying a range for example, upper limit or lower limit of block size, or both, slice, picture, sequence, component, view, layer, etc.
- a range for example, upper limit or lower limit of block size, or both, slice, picture, sequence, component, view, layer, etc.
- the series of processes described above can be performed by hardware or software.
- a program that configures the software is installed on a computer.
- the computer includes, for example, a general-purpose personal computer that can execute various functions by installing a computer incorporated in dedicated hardware and various programs.
- FIG. 24 is a block diagram showing an example of a hardware configuration of a computer that executes the series of processes described above according to a program.
- a central processing unit (CPU) 901, a read only memory (ROM) 902, and a random access memory (RAM) 903 are mutually connected via a bus 904.
- An input / output interface 910 Also connected to the bus 904 is an input / output interface 910.
- An input unit 911, an output unit 912, a storage unit 913, a communication unit 914, and a drive 915 are connected to the input / output interface 910.
- the input unit 911 includes, for example, a keyboard, a mouse, a microphone, a touch panel, an input terminal, and the like.
- the output unit 912 includes, for example, a display, a speaker, and an output terminal.
- the storage unit 913 is formed of, for example, a hard disk, a RAM disk, a non-volatile memory, or the like.
- the communication unit 914 includes, for example, a network interface.
- the drive 915 drives removable media 921 such as a magnetic disk, an optical disk, a magneto-optical disk, or a semiconductor memory.
- the CPU 901 loads the program stored in the storage unit 913 into the RAM 903 via the input / output interface 910 and the bus 904 and executes the program. Processing is performed.
- the RAM 903 also stores data necessary for the CPU 901 to execute various processes.
- the program executed by the computer can be recorded and applied to, for example, a removable medium 921 as a package medium or the like.
- the program can be installed in the storage unit 913 via the input / output interface 910 by attaching the removable media 921 to the drive 915.
- the program can also be provided via a wired or wireless transmission medium such as a local area network, the Internet, or digital satellite broadcasting. In that case, the program can be received by the communication unit 914 and installed in the storage unit 913.
- a wired or wireless transmission medium such as a local area network, the Internet, or digital satellite broadcasting.
- the program can be received by the communication unit 914 and installed in the storage unit 913.
- this program can be installed in advance in the ROM 902 or the storage unit 913.
- the encoding apparatus 100 and the decoding apparatus 200 have been described above as application examples of the present technology, the present technology can be applied to any configuration.
- the present technology includes a transmitter or a receiver (for example, a television receiver or a mobile phone) in satellite broadcasting, cable broadcasting such as cable TV, distribution over the Internet, and distribution to terminals by cellular communication, or
- a transmitter or a receiver for example, a television receiver or a mobile phone
- satellite broadcasting for example, cable TV
- cable broadcasting such as cable TV
- distribution over the Internet distribution to terminals by cellular communication
- the present invention can be applied to various electronic devices such as devices for recording an image on a medium such as an optical disk, a magnetic disk, and a flash memory and reproducing an image from these storage media (for example, a hard disk recorder and a camera).
- the present technology includes a processor (for example, a video processor) as a system LSI (Large Scale Integration) or the like, a module (for example, a video module) using a plurality of processors or the like, a unit (for example, a video unit) using a plurality of modules Or, it may be implemented as part of the configuration of the device, such as a set (for example, a video set) in which other functions are added to the unit.
- a processor for example, a video processor
- LSI Large Scale Integration
- module for example, a video module
- a unit for example, a video unit
- it may be implemented as part of the configuration of the device, such as a set (for example, a video set) in which other functions are added to the unit.
- the present technology can also be applied to a network system configured by a plurality of devices.
- the present technology may be implemented as cloud computing in which multiple devices share and process jointly via a network.
- the present technology is implemented in a cloud service that provides services related to images (moving images) to arbitrary terminals such as computers, AV (Audio Visual) devices, portable information processing terminals, and IoT (Internet of Things) devices. You may do it.
- the system means a set of a plurality of components (apparatus, modules (parts), etc.), and it does not matter whether all the components are in the same case. Therefore, a plurality of devices housed in separate housings and connected via a network and one device housing a plurality of modules in one housing are all systems. .
- the system, apparatus, processing unit, etc. to which the present technology is applied can be used in any field such as traffic, medical care, crime prevention, agriculture, animal husbandry, mining, beauty, factory, home appliance, weather, nature monitoring, etc. . Moreover, the use is also arbitrary.
- the present technology can be applied to systems and devices provided for provision of ornamental content and the like.
- the present technology can be applied to systems and devices provided for traffic, such as traffic condition supervision and automatic operation control.
- the present technology can be applied to systems and devices provided for security.
- the present technology can be applied to a system or device provided for automatic control of a machine or the like.
- the present technology can be applied to systems and devices provided for use in agriculture and livestock.
- the present technology can also be applied to systems and devices that monitor natural conditions such as, for example, volcanoes, forests, and oceans, and wildlife.
- the present technology can be applied to systems and devices provided for sports.
- “flag” is information for identifying a plurality of states, and is not limited to information used to identify two states of true (1) or false (0), and may be three or more. Information that can identify the state is also included. Therefore, the possible value of this "flag” may be, for example, a binary value of 1/0 or a ternary value or more. That is, the number of bits constituting this "flag” is arbitrary, and may be one bit or plural bits. Further, not only the identification information (including the flag) is included in the bitstream, but also the difference information of the identification information with respect to a certain reference information is assumed to be included in the bitstream. In the above, “flag” and “identification information” include not only the information but also difference information with respect to the reference information.
- association means, for example, that one data can be processed (linked) in processing the other data. That is, the data associated with each other may be collected as one data or may be individual data.
- the information associated with the coded data (image) may be transmitted on a transmission path different from that of the coded data (image).
- information associated with encoded data (image) may be recorded on a recording medium (or another recording area of the same recording medium) different from the encoded data (image). Good.
- this “association” may not be the entire data but a part of the data. For example, an image and information corresponding to the image may be associated with each other in an arbitrary unit such as a plurality of frames, one frame, or a part in a frame.
- the configuration described as one device (or processing unit) may be divided and configured as a plurality of devices (or processing units).
- the configuration described as a plurality of devices (or processing units) in the above may be collectively configured as one device (or processing unit).
- configurations other than those described above may be added to the configuration of each device (or each processing unit).
- part of the configuration of one device (or processing unit) may be included in the configuration of another device (or other processing unit) if the configuration or operation of the entire system is substantially the same. .
- the above-described program may be executed on any device.
- the device may have necessary functions (functional blocks and the like) so that necessary information can be obtained.
- one apparatus may execute each step of one flowchart, or a plurality of apparatuses may share and execute. Furthermore, when a plurality of processes are included in one step, one apparatus may execute the plurality of processes, or a plurality of apparatuses may share and execute the processes. In other words, a plurality of processes included in one step can be executed as a process of a plurality of steps. Conversely, the processes described as a plurality of steps can be collectively performed as one step.
- processing of steps for writing a program may be performed in chronological order according to the order described in this specification, in parallel, or in a call. It may be individually executed at necessary timing such as when it is received. That is, as long as no contradiction arises, the processing of each step may be performed in an order different from the order described above. Furthermore, the process of the step of writing this program may be executed in parallel with the process of another program, or may be executed in combination with the process of another program.
- a plurality of techniques relating to the present technology can be independently implemented as long as no contradiction arises.
- any number of the present techniques may be used in combination.
- part or all of the present technology described in any of the embodiments can be implemented in combination with part or all of the present technology described in the other embodiments.
- some or all of the above-described optional present techniques may be implemented in combination with other techniques not described above.
- An image processing apparatus comprising: a two-dimensional projection unit configured to project data at every position included in 3D data representing a three-dimensional structure on a two-dimensional plane of a plurality of layers.
- the two-dimensional projection unit projects the data for each of the positions of the 3D data, the positions of which overlap in the depth direction when viewed from the projection plane, on mutually different layers of the two-dimensional plane of the plurality of layers.
- the image processing apparatus according to (1).
- the two-dimensional projection unit generates, on the two-dimensional plane, the same number of layers as the maximum number of data for each position where the position overlaps in the depth direction when viewed from the projection plane of the 3D data. Image processing apparatus as described.
- the image processing apparatus according to any one of (1) to (3), further including: an encoding unit that encodes the 3D data projected onto the two-dimensional plane by the two-dimensional projection unit.
- an encoding unit that encodes the 3D data projected onto the two-dimensional plane by the two-dimensional projection unit.
- the encoding unit encodes position information, attribute information, and an occupancy map of the 3D data projected onto each layer of the two-dimensional plane.
- the image processing apparatus according to (4) or (5), further including: a bitstream generation unit that generates a bitstream including the image.
- the video camera further includes a packing unit packing the two-dimensional plane on which the 3D data is projected by the two-dimensional projection unit as a video frame,
- the image processing apparatus according to any one of (4) to (6), wherein the encoding unit is configured to encode the video frame in which the two-dimensional plane is packed by the packing unit.
- the image processing apparatus according to any one of (1) to (7), wherein the two-dimensional projection unit projects the 3D data onto the two-dimensional plane for each predetermined area.
- An image processing apparatus comprising: a three-dimensional projection unit configured to project, onto a three-dimensional space, data of all positions of 3D data projected onto a two-dimensional plane of the number of layers indicated by the layer number information.
- the information processing apparatus further comprises an extraction unit for extracting the layer number information contained in the bit stream, The three-dimensional projection unit projects, to the three-dimensional space, data at every position of the 3D data projected onto the two-dimensional plane in the number of layers indicated by the layer number information extracted by the extraction unit.
- the image processing apparatus which is configured to: (13)
- the information processing apparatus further includes a decoding unit that decodes encoded data of the 3D data projected onto the two-dimensional plane included in the bit stream,
- the three-dimensional projection unit projects, to the three-dimensional space, the data at every position of the 3D data projected on the two-dimensional plane, obtained by decoding the encoded data by the decoding unit.
- the image processing apparatus which is configured to: (14)
- the decoding unit decodes each piece of encoded data of position information, attribute information, and occupancy map of the 3D data projected on each layer of the two-dimensional plane.
- the image according to (13) Processing unit is configured to:
- the video decoding apparatus further comprises an unpacking unit for unpacking a video frame obtained by decoding the encoded data by the decoding unit and obtained by packing the 3D data projected onto the two-dimensional plane.
- the three-dimensional projection unit projects, to the three-dimensional space, the data at every position of the 3D data projected on the two-dimensional plane, obtained by unpacking a video frame by the unpacking unit.
- the image processing apparatus according to (13) or (14), which is configured to: (16) The three-dimensional projection unit projects, onto the three-dimensional space, the 3D data projected onto the two-dimensional plane for each predetermined area.
- the data for each position included in the 3D data representing the three-dimensional structure is projected on a two-dimensional plane, and the data for each position does not exist at a position where the data for each position on the two-dimensional plane does not exist
- An image processing apparatus comprising: a two-dimensional projection unit that sets a predetermined value indicating.
- the image processing apparatus according to (21), wherein the predetermined value is a predetermined fixed value.
- the image processing apparatus according to (21), wherein the predetermined value is a value larger than a maximum value of depth of the 3D data.
- the image processing apparatus according to (24), wherein the encoding unit encodes position information, attribute information, and an occupancy map of the 3D data projected on the two-dimensional plane.
- a bitstream generation unit configured to generate a bitstream including information indicating the predetermined value and encoded data obtained by encoding the two-dimensional plane by the encoding unit.
- the video camera further includes a packing unit packing the two-dimensional plane on which the 3D data is projected by the two-dimensional projection unit as a video frame, The image processing apparatus according to any one of (24) to (26), wherein the encoding unit is configured to encode the video frame in which the two-dimensional plane is packed by the packing unit.
- the information processing apparatus further comprises an extraction unit for extracting information indicating the predetermined value included in the bit stream,
- the three-dimensional projection unit projects, to a three-dimensional space, data other than the data of the predetermined value indicated in the information extracted by the extraction unit among the data at each position included in the 3D data.
- the image processing apparatus according to any one of (31) to (33).
- the image processing apparatus further comprises a decoding unit that decodes encoded data of the 3D data projected onto the two-dimensional plane included in the bit stream, The three-dimensional projection unit is extracted by the extraction unit among the data for each position included in the 3D data projected on the two-dimensional plane, obtained by decoding the encoded data by the decoding unit.
- the image processing apparatus configured to project data other than the data of the predetermined value indicated in the received information into a three-dimensional space.
- the decoding unit decodes the encoded data of the position information, the attribute information, and the occupancy map of the 3D data projected onto each layer of the two-dimensional plane.
- the video decoding apparatus further comprises an unpacking unit for unpacking a video frame obtained by decoding the encoded data by the decoding unit and obtained by packing the 3D data projected onto the two-dimensional plane.
- the three-dimensional projection unit is extracted by the extraction unit among the data for each position included in the 3D data projected on the two-dimensional plane, obtained by unpacking a video frame by the unpacking unit.
- the image processing apparatus configured to project data other than the data of the predetermined value indicated in the received information into a three-dimensional space.
- the image processing apparatus according to any one of (31) to (38), wherein the 3D data is a point cloud. (40) Of the data at each position included in the 3D data representing the three-dimensional structure projected on the two-dimensional plane, the data other than the data of the predetermined value indicating that the data at each position does not exist is An image processing method that projects to space.
- Image processing comprising a two-dimensional projection unit for projecting the data at each position within the range in the depth direction indicated by the depth parameter that limits the range in the depth direction of the data at each position of 3D data representing apparatus.
- the image processing apparatus according to (41), further including: an encoding unit that encodes the 3D data projected onto the two-dimensional plane by the two-dimensional projection unit.
- the image processing device according to (42), wherein the encoding unit encodes position information, attribute information, and an occupancy map of the 3D data projected onto the two-dimensional plane.
- a bit stream generating unit is further provided which generates a bit stream including the depth parameter set for each of the regions and encoded data obtained by encoding the two-dimensional plane by the encoding unit.
- the image processing apparatus according to (42) or (43).
- the image processing apparatus further comprises a packing unit packing the two-dimensional plane on which the 3D data is projected by the two-dimensional projection unit as a video frame,
- the image processing apparatus according to any one of (42) to (44), wherein the encoding unit is configured to encode the video frame in which the two-dimensional plane is packed by the packing unit.
- a three-dimensional structure that can be projected on one layer set for each area when projecting data for each position of 3D data representing a three-dimensional structure on a two-dimensional plane for each predetermined area in three-dimensional space An image processing method for projecting, on the two-dimensional plane, data at each position within the range in the depth direction indicated by a depth parameter that limits the range in the depth direction of data at each position of 3D data representing.
- An image processing apparatus comprising: a three-dimensional projection unit.
- the method further comprises an extraction unit for extracting the depth parameter included in the bit stream,
- the three-dimensional projection unit is configured to project the data for each position included in the 3D data within a range in the depth direction indicated by the depth parameter extracted by the extraction unit (51). Image processing apparatus as described.
- the information processing apparatus further comprising a decoding unit that decodes encoded data of the 3D data projected onto the two-dimensional plane included in the bit stream,
- the three-dimensional projection unit extracts the data for each position of the 3D data projected on the two-dimensional plane, obtained by decoding the encoded data by the decoding unit, extracted by the extraction unit.
- the image processing apparatus according to (52), configured to project within a range in the depth direction indicated by a depth parameter.
- the decoding unit decodes the encoded data of the position information, the attribute information, and the occupancy map of the 3D data projected onto each layer of the two-dimensional plane.
- the video decoding apparatus further comprises an unpacking unit for unpacking a video frame obtained by decoding the encoded data by the decoding unit and obtained by packing the 3D data projected onto the two-dimensional plane.
- the three-dimensional projection unit extracts data of the 3D data projected on the two-dimensional plane at each position, obtained by the video frame being unpacked by the unpacking unit, by the extraction unit.
- the image processing apparatus according to (53) or (54), configured to project within a range in the depth direction indicated by a depth parameter.
- the image processing apparatus according to any one of (51) to (55), wherein the 3D data is a point cloud.
- 100 encoding device 111 patch decomposing unit, 112 packing unit, 113 auxiliary patch information compression unit, 114 video encoding unit, 115 video encoding unit, 116 OMap encoding unit, 117 multiplexer, 151 normal direction estimation unit, 152 Segmentation initial setting unit, 153 segmentation update unit, 154 two-dimensional projection unit, 155 pixel distribution analysis unit, 200 decoding device, 211 demultiplexer, 212 auxiliary patch information decoding unit, 213 video decoding unit, 214 video decoding unit, 215 OMap decoding Unit, 216 Unpacking Unit, 217 3D Reconstruction Unit, 251 3D Projection Unit, 252 Pixel Distribution Analysis Unit, 253 Reverse Segmentation Update Unit, 254 Reverse Segmentation Unit Initial setting unit, 255 reverse normal direction estimation unit
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Multimedia (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Signal Processing (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
- Image Processing (AREA)
- Image Generation (AREA)
- Computer Graphics (AREA)
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
- Image Analysis (AREA)
Abstract
Description
1.ビデオベースドアプローチ
2.第1の実施の形態(可変レイヤ数)
3.第2の実施の形態(無いポイントの定義)
4.第3の実施の形態(可変奥行パラメータ)
5.付記
<技術内容・技術用語をサポートする文献等>
本技術で開示される範囲は、実施例に記載されている内容だけではなく、出願当時において公知となっている以下の非特許文献に記載されている内容も含まれる。
非特許文献2:(上述)
非特許文献3:(上述)
非特許文献4:TELECOMMUNICATION STANDARDIZATION SECTOR OF ITU(International Telecommunication Union), "Advanced video coding for generic audiovisual services", H.264, 04/2017
非特許文献5:TELECOMMUNICATION STANDARDIZATION SECTOR OF ITU(International Telecommunication Union), "High efficiency video coding", H.265, 12/2016
非特許文献6:Jianle Chen, Elena Alshina, Gary J. Sullivan, Jens-Rainer, Jill Boyce, "Algorithm Description of Joint Exploration Test Model 4", JVET-G1001_v1, Joint Video Exploration Team (JVET) of ITU-T SG 16 WP 3 and ISO/IEC JTC 1/SC 29/WG 11 7th Meeting: Torino, IT, 13-21 July 2017
従来、点群の位置情報や属性情報等により3次元構造を表すポイントクラウドや、頂点、エッジ、面で構成され、多角形表現を使用して3次元形状を定義するメッシュ等のデータが存在した。
このようなポイントクラウドの位置と色情報それぞれを、小領域毎に2次元平面に投影し、2次元画像用の符号化方法で符号化するビデオベースドアプローチ(Video-based approach)が提案されている。
以上に説明したようなビデオベースドアプローチに関する本技術について説明する。図3は、各実施の形態において説明する本技術の一覧である。
次に、以上のような各手法を実現する構成について説明する。図4は、本技術を適用した画像処理装置の一態様である符号化装置の構成の一例を示すブロック図である。図4に示される符号化装置100は、ポイントクラウドのような3Dデータを2次元平面に投影して2次元画像用の符号化方法により符号化を行う装置である。
図5は、パッチ分解部111の主な構成例を示すブロック図である。図5に示されるように、この場合のパッチ分解部111は、法線方向推定部151、セグメンテーション初期設定部152、セグメンテーション更新部153、2次元投影部154、および画素分布解析部155を有する。
図6は、本技術を適用した画像処理装置の一態様である復号装置の構成の一例を示すブロック図である。図6に示される復号装置200は、ポイントクラウドのような3Dデータが2次元平面に投影されて符号化された符号化データを、2次元画像用の復号方法により復号し、3次元空間に投影する装置である。
図7は、図6の3D再構築部217の主な構成例を示すブロック図である。図7に示されるように、3D再構築部217は、3次元投影部251、画素分布解析部252、逆セグメンテーション更新部253、逆セグメンテーション初期設定部254、および逆法線方向推定部255を有する。
<3Dデータを投影する2次元平面のレイヤ数可変>
従来の方法では、3Dデータは、図8に示される例のように、2つのレイヤ(レイヤ0およびレイヤ1)の2次元平面に投影されていた。
符号化装置100により実行される符号化処理の流れの例を、図10のフローチャートを参照して説明する。
次に、図11のフローチャートを参照して、図10のステップS101において実行されるパッチ分解処理の流れの例を説明する。
次に、図12のフローチャートを参照して、図11のステップS124において実行される2次元投影処理の流れの例を説明する。
復号側においては、符号化側より提供される2次元平面のレイヤ数を示す情報を用いることにより、上述のようにレイヤ数が可変の2次元平面に投影された3Dデータの再構築を実現することができる。
復号装置200により実行される復号処理の流れの例を、図13のフローチャートを参照して説明する。
次に、図14のフローチャートを参照して、図13のステップS207において実行されるポイントクラウド再構築処理の流れの例を説明する。
次に、図15のフローチャートを参照して、図14のステップS221において実行される3次元投影処理の流れの例を説明する。
<「無い画素」を示す値のシグナル>
疎(Sparse)な3Dデータを2次元平面に投影すると、その2次元平面上において3Dデータが投影されない画素、すなわち、画素値が設定されない画素(無い画素とも称する)が生じる場合がある。
この場合も、符号化処理およびパッチ分解処理は、第1の実施の形態において説明した場合と同様に行われる。したがってそれらの説明は省略する。
復号側においては、2次元平面に投影された3Dデータを3次元空間に投影する際に、以上のようにしてシグナルされた2次元平面上の「無い画素」を示す画素値(上述のX)を検出し、削除する(投影しないようにする)。
この場合も、復号処理およびポイントクラウド再構築処理は、第1の実施の形態において説明した場合と同様に行われる。したがってそれらの説明は省略する。
<奥行パラメータの制御>
3Dデータを2次元平面に投影する際に、その2次元平面に投影する3Dデータの奥行方向の範囲を制御する奥行パラメータthが利用される。この奥行パラメータthにより指定される範囲内のポイントが2次元平面への投影対象となるため、この奥行パラメータthの値は、領域(Local bounding box)の奥行方向の長さに関係する。例えば、領域の奥行方向の長さよりも奥行パラメータthの値が大きいと、他の領域のポイントも投影対象になり得る。つまり、領域の奥行方向の長さは、奥行パラメータthよりも長くする必要がある。
この場合も、符号化処理およびパッチ分解処理は、第1の実施の形態において説明した場合と同様に行われる。したがってそれらの説明は省略する。
復号側においては、2次元平面に投影された3Dデータを3次元空間に投影する際に、以上のように制御された奥行パラメータthに示される範囲に3Dデータを投影するようにする。
この場合も、復号処理およびポイントクラウド再構築処理は、第1の実施の形態において説明した場合と同様に行われる。したがってそれらの説明は省略する。
<制御情報>
以上の各実施の形態において説明した本技術に関する制御情報を符号化側から復号側に伝送するようにしてもよい。例えば、上述した本技術を適用することを許可(または禁止)するか否かを制御する制御情報(例えばenabled_flag)を伝送するようにしてもよい。また、例えば、上述した本技術を適用することを許可(または禁止)する範囲(例えばブロックサイズの上限若しくは下限、またはその両方、スライス、ピクチャ、シーケンス、コンポーネント、ビュー、レイヤ等)を指定する制御情報を伝送するようにしてもよい。
上述した一連の処理は、ハードウエアにより実行させることもできるし、ソフトウエアにより実行させることもできる。一連の処理をソフトウエアにより実行する場合には、そのソフトウエアを構成するプログラムが、コンピュータにインストールされる。ここでコンピュータには、専用のハードウエアに組み込まれているコンピュータや、各種のプログラムをインストールすることで、各種の機能を実行することが可能な、例えば汎用のパーソナルコンピュータ等が含まれる。
以上においては、ポイントクラウドデータのVoxel化に本技術を適用する場合について説明したが、本技術は、これらの例に限らず、任意の規格の3DデータのVoxel化に対して適用することができる。つまり、上述した本技術と矛盾しない限り、符号化・復号方式等の各種処理、並びに、3Dデータやメタデータ等の各種データの仕様は任意である。また、本技術と矛盾しない限り、上述した一部の処理や仕様を省略してもよい。
本技術を適用したシステム、装置、処理部等は、例えば、交通、医療、防犯、農業、畜産業、鉱業、美容、工場、家電、気象、自然監視等、任意の分野に利用することができる。また、その用途も任意である。
なお、本明細書において「フラグ」とは、複数の状態を識別するための情報であり、真(1)または偽(0)の2状態を識別する際に用いる情報だけでなく、3以上の状態を識別することが可能な情報も含まれる。したがって、この「フラグ」が取り得る値は、例えば1/0の2値であってもよいし、3値以上であってもよい。すなわち、この「フラグ」を構成するbit数は任意であり、1bitでも複数bitでもよい。また、識別情報(フラグも含む)は、その識別情報をビットストリームに含める形だけでなく、ある基準となる情報に対する識別情報の差分情報をビットストリームに含める形も想定されるため、本明細書においては、「フラグ」や「識別情報」は、その情報だけではなく、基準となる情報に対する差分情報も包含する。
(1) 3次元構造を表す3Dデータに含まれる全ての位置毎のデータを、複数レイヤの2次元平面に投影する2次元投影部
を備える画像処理装置。
(2) 前記2次元投影部は、前記3Dデータの、投影面から見て位置が奥行方向に重なる前記位置毎のデータを、前記複数のレイヤの前記2次元平面の、互いに異なるレイヤに投影する
(1)に記載の画像処理装置。
(3) 前記2次元投影部は、前記2次元平面について、前記3Dデータの投影面から見て位置が奥行方向に重なる前記位置毎のデータの最大数と同数のレイヤを生成する
(2)に記載の画像処理装置。
(4) 前記2次元投影部により前記2次元平面に投影された前記3Dデータを、符号化する符号化部をさらに備える
(1)乃至(3)のいずれかに記載の画像処理装置。
(5) 前記符号化部は、前記2次元平面の各レイヤに投影された前記3Dデータの、位置情報、属性情報、およびオキュパンシーマップをそれぞれ符号化する
(4)に記載の画像処理装置。
(6) 前記2次元投影部により前記3Dデータが投影された前記2次元平面のレイヤ数を示す情報と、前記符号化部により前記2次元平面が符号化されて得られた符号化データとを含むビットストリームを生成するビットストリーム生成部をさらに備える
(4)または(5)に記載の画像処理装置。
(7) 前記2次元投影部により前記3Dデータが投影された前記2次元平面を、ビデオフレームとしてパッキングするパッキング部をさらに備え、
前記符号化部は、前記パッキング部により前記2次元平面がパッキングされた前記ビデオフレームを符号化するように構成される
(4)乃至(6)のいずれかに記載の画像処理装置。
(8) 前記2次元投影部は、前記3Dデータを、所定の領域毎に前記2次元平面に投影する
(1)乃至(7)のいずれかに記載の画像処理装置。
(9) 前記3Dデータは、ポイントクラウドである
(1)乃至(8)のいずれかに記載の画像処理装置。
(10) 3次元構造を表す3Dデータに含まれる全ての位置毎のデータを、複数レイヤの2次元平面に投影する
画像処理方法。
を備える画像処理装置。
(12) ビットストリームに含まれる前記レイヤ数情報を抽出する抽出部をさらに備え、
前記3次元投影部は、前記抽出部により抽出された前記レイヤ数情報が示すレイヤ数の前記2次元平面に投影された前記3Dデータの全ての前記位置毎のデータを、前記3次元空間に投影するように構成される
(11)に記載の画像処理装置。
(13) 前記ビットストリームに含まれる前記2次元平面に投影された前記3Dデータの符号化データを復号する復号部をさらに備え、
前記3次元投影部は、前記復号部により前記符号化データが復号されて得られた、前記2次元平面に投影された前記3Dデータの全ての前記位置毎のデータを、前記3次元空間に投影するように構成される
(12)に記載の画像処理装置。
(14) 前記復号部は、前記2次元平面の各レイヤに投影された前記3Dデータの、位置情報、属性情報、およびオキュパンシーマップのそれぞれの符号化データを復号する
(13)に記載の画像処理装置。
(15) 前記復号部により前記符号化データが復号されて得られた、前記2次元平面に投影された前記3Dデータがパッキングされたビデオフレームをアンパッキングするアンパッキング部をさらに備え、
前記3次元投影部は、前記アンパッキング部によりビデオフレームがアンパッキングされて得られた、前記2次元平面に投影された前記3Dデータの全ての前記位置毎のデータを、前記3次元空間に投影するように構成される
(13)または(14)に記載の画像処理装置。
(16) 前記3次元投影部は、所定の領域毎に前記2次元平面に投影された前記3Dデータを、前記3次元空間に投影する
(11)乃至(15)のいずれかに記載の画像処理装置。
(17) 前記3Dデータは、ポイントクラウドである
(11)乃至(16)のいずれかに記載の画像処理装置。
(18) レイヤ数情報が示すレイヤ数の2次元平面に投影された3Dデータの全ての位置毎のデータを、3次元空間に投影する
画像処理方法。
を備える画像処理装置。
(22) 前記所定の値は、予め定められた固定値である
(21)に記載の画像処理装置。
(23) 前記所定の値は、前記3Dデータの奥行の最大値より大きな値である
(21)に記載の画像処理装置。
(24) 前記2次元投影部により前記2次元平面に投影された前記3Dデータを、符号化する符号化部をさらに備える
(21)乃至(23)のいずれかに記載の画像処理装置。
(25) 前記符号化部は、前記2次元平面に投影された前記3Dデータの、位置情報、属性情報、およびオキュパンシーマップをそれぞれ符号化する
(24)に記載の画像処理装置。
(26) 前記所定の値を示す情報と、前記符号化部により前記2次元平面が符号化されて得られた符号化データとを含むビットストリームを生成するビットストリーム生成部をさらに備える
(24)または(25)に記載の画像処理装置。
(27) 前記2次元投影部により前記3Dデータが投影された前記2次元平面を、ビデオフレームとしてパッキングするパッキング部をさらに備え、
前記符号化部は、前記パッキング部により前記2次元平面がパッキングされた前記ビデオフレームを符号化するように構成される
(24)乃至(26)のいずれかに記載の画像処理装置。
(28) 前記2次元投影部は、前記3Dデータを、所定の領域毎に前記2次元平面に投影する
(21)乃至(27)のいずれかに記載の画像処理装置。
(29) 前記3Dデータは、ポイントクラウドである
(21)乃至(28)のいずれかに記載の画像処理装置。
(30) 3次元構造を表す3Dデータに含まれる位置毎のデータを2次元平面に投影し、前記2次元平面の前記位置毎のデータが存在しない位置に、前記位置毎のデータが存在しないことを示す所定の値をセットする
画像処理方法。
を備える画像処理装置。
(32) 前記所定の値は、予め定められた固定値である
(31)に記載の画像処理装置。
(33) 前記所定の値は、前記3Dデータの奥行の最大値より大きな値である
(31)に記載の画像処理装置。
(34) ビットストリームに含まれる前記所定の値を示す情報を抽出する抽出部をさらに備え、
前記3次元投影部は、前記3Dデータに含まれる前記位置毎のデータの内、前記抽出部により抽出された前記情報に示される前記所定の値のデータ以外のデータを、3次元空間に投影するように構成される
(31)乃至(33)のいずれかに記載の画像処理装置。
(35) 前記ビットストリームに含まれる前記2次元平面に投影された前記3Dデータの符号化データを復号する復号部をさらに備え、
前記3次元投影部は、前記復号部により前記符号化データが復号されて得られた、前記2次元平面に投影された前記3Dデータに含まれる前記位置毎のデータの内、前記抽出部により抽出された前記情報に示される前記所定の値のデータ以外のデータを、3次元空間に投影するように構成される
(34)に記載の画像処理装置。
(36) 前記復号部は、前記2次元平面の各レイヤに投影された前記3Dデータの、位置情報、属性情報、およびオキュパンシーマップのそれぞれの符号化データを復号する
(35)に記載の画像処理装置。
(37) 前記復号部により前記符号化データが復号されて得られた、前記2次元平面に投影された前記3Dデータがパッキングされたビデオフレームをアンパッキングするアンパッキング部をさらに備え、
前記3次元投影部は、前記アンパッキング部によりビデオフレームがアンパッキングされて得られた、前記2次元平面に投影された前記3Dデータに含まれる前記位置毎のデータの内、前記抽出部により抽出された前記情報に示される前記所定の値のデータ以外のデータを、3次元空間に投影するように構成される
(35)または(36)に記載の画像処理装置。
(38) 前記3次元投影部は、所定の領域毎に前記2次元平面に投影された前記3Dデータを、前記3次元空間に投影する
(31)乃至(37)のいずれかに記載の画像処理装置。
(39) 前記3Dデータは、ポイントクラウドである
(31)乃至(38)のいずれかに記載の画像処理装置。
(40) 2次元平面に投影された3次元構造を表す3Dデータに含まれる位置毎のデータの内、前記位置毎のデータが存在しないことを示す所定の値のデータ以外のデータを、3次元空間に投影する
画像処理方法。
を備える画像処理装置。
(42) 前記2次元投影部により前記2次元平面に投影された前記3Dデータを、符号化する符号化部をさらに備える
(41)に記載の画像処理装置。
(43) 前記符号化部は、前記2次元平面に投影された前記3Dデータの、位置情報、属性情報、およびオキュパンシーマップをそれぞれ符号化する
(42)に記載の画像処理装置。
(44) 前記領域毎に設定された前記奥行パラメータと、前記符号化部により前記2次元平面が符号化されて得られた符号化データとを含むビットストリームを生成するビットストリーム生成部をさらに備える
(42)または(43)に記載の画像処理装置。
(45) 前記2次元投影部により前記3Dデータが投影された前記2次元平面を、ビデオフレームとしてパッキングするパッキング部をさらに備え、
前記符号化部は、前記パッキング部により前記2次元平面がパッキングされた前記ビデオフレームを符号化するように構成される
(42)乃至(44)のいずれかに記載の画像処理装置。
(46) 前記3Dデータは、ポイントクラウドである
(41)乃至(45)のいずれかに記載の画像処理装置。
(47) 3次元構造を表す3Dデータの位置毎のデータを3次元空間の所定の領域毎に2次元平面に投影する際に、前記領域毎に設定された1レイヤに投影可能な3次元構造を表す3Dデータの位置毎のデータの奥行方向の範囲を制限する奥行パラメータが示す前記奥行方向の範囲内の前記位置毎のデータを、前記2次元平面に投影する
画像処理方法。
を備える画像処理装置。
(52) ビットストリームに含まれる前記奥行パラメータを抽出する抽出部をさらに備え、
前記3次元投影部は、前記3Dデータに含まれる前記位置毎のデータを、前記抽出部により抽出された前記奥行パラメータが示す前記奥行方向の範囲内に投影するように構成される
(51)に記載の画像処理装置。
(53) 前記ビットストリームに含まれる前記2次元平面に投影された前記3Dデータの符号化データを復号する復号部をさらに備え、
前記3次元投影部は、前記復号部により前記符号化データが復号されて得られた、前記2次元平面に投影された前記3Dデータの前記位置毎のデータを、前記抽出部により抽出された前記奥行パラメータが示す前記奥行方向の範囲内に投影するように構成される
(52)に記載の画像処理装置。
(54) 前記復号部は、前記2次元平面の各レイヤに投影された前記3Dデータの、位置情報、属性情報、およびオキュパンシーマップのそれぞれの符号化データを復号する
(53)に記載の画像処理装置。
(55) 前記復号部により前記符号化データが復号されて得られた、前記2次元平面に投影された前記3Dデータがパッキングされたビデオフレームをアンパッキングするアンパッキング部をさらに備え、
前記3次元投影部は、前記アンパッキング部によりビデオフレームがアンパッキングされて得られた、前記2次元平面に投影された前記3Dデータの前記位置毎のデータを、前記抽出部により抽出された前記奥行パラメータが示す前記奥行方向の範囲内に投影するように構成される
(53)または(54)に記載の画像処理装置。
(56) 前記3Dデータは、ポイントクラウドである
(51)乃至(55)のいずれかに記載の画像処理装置。
(57) 3次元空間の所定の領域毎に、2次元平面に投影された3次元構造を表す3Dデータに含まれる位置毎のデータを前記3次元空間に投影する際に、前記位置毎のデータを、前記3次元空間の、前記領域毎に設定された1レイヤに投影可能な前記3Dデータの前記位置毎のデータの奥行方向の範囲を制限する奥行パラメータが示す前記奥行方向の範囲内に投影する
画像処理方法。
Claims (20)
- 3次元構造を表す3Dデータに含まれる全ての位置毎のデータを、複数レイヤの2次元平面に投影する2次元投影部
を備える画像処理装置。 - 前記2次元投影部は、前記3Dデータの、投影面から見て位置が奥行方向に重なる前記位置毎のデータを、前記複数のレイヤの前記2次元平面の、互いに異なるレイヤに投影する
請求項1に記載の画像処理装置。 - 前記2次元投影部は、前記2次元平面について、前記3Dデータの投影面から見て位置が奥行方向に重なる前記位置毎のデータの最大数と同数のレイヤを生成する
請求項2に記載の画像処理装置。 - 前記2次元投影部は、前記2次元平面の前記位置毎のデータが存在しない位置に、前記位置毎のデータが存在しないことを示す所定の値をセットする
請求項1に記載の画像処理装置。 - 前記所定の値は、予め定められた固定値である
請求項4に記載の画像処理装置。 - 前記所定の値は、前記3Dデータの奥行の最大値より大きな値である
請求項4に記載の画像処理装置。 - 前記2次元投影部は、前記3Dデータの前記位置毎のデータを、3次元空間の所定の領域毎に2次元平面に投影する際に、前記領域毎に設定された1レイヤに投影可能な前記3Dデータの前記位置毎のデータの奥行方向の範囲を制限する奥行パラメータが示す前記奥行方向の範囲内の前記位置毎のデータを、前記2次元平面に投影する
請求項1に記載の画像処理装置。 - 前記2次元投影部により前記2次元平面に投影された前記3Dデータを、符号化する符号化部をさらに備える
請求項1に記載の画像処理装置。 - 前記2次元投影部により前記3Dデータが投影された前記2次元平面のレイヤ数を示す情報と、前記符号化部により前記2次元平面が符号化されて得られた符号化データとを含むビットストリームを生成するビットストリーム生成部をさらに備える
請求項8に記載の画像処理装置。 - 前記3Dデータは、ポイントクラウドである
請求項1に記載の画像処理装置。 - 3次元構造を表す3Dデータに含まれる全ての位置毎のデータを、複数レイヤの2次元平面に投影する
画像処理方法。 - レイヤ数情報が示すレイヤ数の2次元平面に投影された3Dデータの全ての位置毎のデータを、3次元空間に投影する3次元投影部
を備える画像処理装置。 - 前記3次元投影部は、前記3Dデータに含まれる前記位置毎のデータの内、前記位置毎のデータが存在しないことを示す所定の値のデータ以外のデータを、前記3次元空間に投影する
請求項12に記載の画像処理装置。 - 前記所定の値は、予め定められた固定値である
請求項13に記載の画像処理装置。 - 前記所定の値は、前記3Dデータの奥行の最大値より大きな値である
請求項13に記載の画像処理装置。 - 前記3次元投影部は、前記3次元空間の所定の領域毎に、前記3Dデータに含まれる前記位置毎のデータを前記3次元空間に投影する際に、前記位置毎のデータを、前記3次元空間の、前記領域毎に設定された1レイヤに投影可能な3次元構造を表す3Dデータの位置毎のデータの奥行方向の範囲を制限する奥行パラメータが示す前記奥行方向の範囲内に投影する
請求項12に記載の画像処理装置。 - ビットストリームに含まれる前記レイヤ数情報を抽出する抽出部をさらに備え、
前記3次元投影部は、前記抽出部により抽出された前記レイヤ数情報が示すレイヤ数の前記2次元平面に投影された前記3Dデータの全ての前記位置毎のデータを、前記3次元空間に投影するように構成される
請求項12に記載の画像処理装置。 - 前記ビットストリームに含まれる前記2次元平面に投影された前記3Dデータの符号化データを復号する復号部をさらに備え、
前記3次元投影部は、前記復号部により前記符号化データが復号されて得られた、前記2次元平面に投影された前記3Dデータの全ての前記位置毎のデータを、前記3次元空間に投影するように構成される
請求項17に記載の画像処理装置。 - 前記3Dデータは、ポイントクラウドである
請求項12に記載の画像処理装置。 - レイヤ数情報が示すレイヤ数の2次元平面に投影された3Dデータの全ての位置毎のデータを、3次元空間に投影する
画像処理方法。
Priority Applications (14)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
AU2019210233A AU2019210233B2 (en) | 2018-01-16 | 2019-01-07 | Image processing device and method |
CN201980007870.2A CN111566702B (zh) | 2018-01-16 | 2019-01-07 | 图像处理装置和方法 |
US16/960,910 US11721048B2 (en) | 2018-01-16 | 2019-01-07 | Image processing apparatus and method |
SG11202005374UA SG11202005374UA (en) | 2018-01-16 | 2019-01-07 | Image processing apparatus and method |
EP19741709.0A EP3742401A4 (en) | 2018-01-16 | 2019-01-07 | IMAGE PROCESSING DEVICE AND METHOD |
BR112020013599-2A BR112020013599A2 (pt) | 2018-01-16 | 2019-01-07 | aparelho e método de processamento de imagem. |
KR1020237037047A KR20230155019A (ko) | 2018-01-16 | 2019-01-07 | 화상 처리 장치 및 방법 |
JP2019566411A JP7327166B2 (ja) | 2018-01-16 | 2019-01-07 | 画像処理装置および方法 |
CA3087980A CA3087980A1 (en) | 2018-01-16 | 2019-01-07 | Image processing apparatus and method |
MX2020007416A MX2020007416A (es) | 2018-01-16 | 2019-01-07 | Aparato y metodo de procesamiento de imagenes. |
RU2020122782A RU2020122782A (ru) | 2018-01-16 | 2019-01-07 | Устройство и способ обработки изображения |
KR1020207019305A KR102596507B1 (ko) | 2018-01-16 | 2019-01-07 | 화상 처리 장치 및 방법 |
PH12020500587A PH12020500587A1 (en) | 2018-01-16 | 2020-07-09 | Image processing apparatus and method |
ZA2020/05038A ZA202005038B (en) | 2018-01-16 | 2020-08-14 | Image processing device and method |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2018004983 | 2018-01-16 | ||
JP2018-004983 | 2018-01-16 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2019142666A1 true WO2019142666A1 (ja) | 2019-07-25 |
Family
ID=67301752
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2019/000051 WO2019142666A1 (ja) | 2018-01-16 | 2019-01-07 | 画像処理装置および方法 |
Country Status (15)
Country | Link |
---|---|
US (1) | US11721048B2 (ja) |
EP (1) | EP3742401A4 (ja) |
JP (1) | JP7327166B2 (ja) |
KR (2) | KR20230155019A (ja) |
CN (1) | CN111566702B (ja) |
AU (1) | AU2019210233B2 (ja) |
BR (1) | BR112020013599A2 (ja) |
CA (1) | CA3087980A1 (ja) |
MX (1) | MX2020007416A (ja) |
PH (1) | PH12020500587A1 (ja) |
RU (1) | RU2020122782A (ja) |
SG (1) | SG11202005374UA (ja) |
TW (1) | TWI815842B (ja) |
WO (1) | WO2019142666A1 (ja) |
ZA (1) | ZA202005038B (ja) |
Cited By (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2020012968A1 (ja) * | 2018-07-11 | 2020-01-16 | ソニー株式会社 | 画像処理装置および方法 |
WO2020017359A1 (ja) * | 2018-07-20 | 2020-01-23 | ソニー株式会社 | 画像処理装置および方法 |
JP2021511712A (ja) * | 2018-01-19 | 2021-05-06 | インターデジタル ヴイシー ホールディングス, インコーポレイテッド | 点群処理 |
WO2021095565A1 (ja) * | 2019-11-13 | 2021-05-20 | ソニーグループ株式会社 | 画像処理装置および方法 |
JP2021518681A (ja) * | 2018-04-11 | 2021-08-02 | インターデジタル ヴイシー ホールディングス, インコーポレイテッド | 三次元物体を表すポイントクラウドを符号化/復号する方法及び装置 |
JP2021518694A (ja) * | 2018-04-11 | 2021-08-02 | インターデジタル ヴイシー ホールディングス, インコーポレイテッド | 三次元物体を表すポイントクラウドを符号化/復号する方法及び装置 |
JP2021518689A (ja) * | 2018-04-11 | 2021-08-02 | インターデジタル ヴイシー ホールディングス, インコーポレイテッド | 点群の幾何学的形状をコーディングする方法およびデバイス |
JP2021519533A (ja) * | 2018-04-11 | 2021-08-10 | インターデジタル ヴイシー ホールディングス, インコーポレイテッド | 投影平面の少なくとも1つの画像領域に一度直交投影された3d点のセットの深度値を符号化するための方法 |
WO2021193088A1 (ja) * | 2020-03-25 | 2021-09-30 | ソニーグループ株式会社 | 画像処理装置および方法 |
JP2021531542A (ja) * | 2018-07-11 | 2021-11-18 | インターデジタル ヴイシー ホールディングス, インコーポレイテッド | 点群の点のテクスチャを符号化/復号する方法 |
JP2022525599A (ja) * | 2019-03-12 | 2022-05-18 | ホアウェイ・テクノロジーズ・カンパニー・リミテッド | 点群コーディングのためのパッチデータユニットのコーディングおよび復号 |
JP2022550150A (ja) * | 2019-09-30 | 2022-11-30 | 中興通訊股▲ふん▼有限公司 | 点群データのためのファイルフォーマット |
US12002243B2 (en) | 2021-09-03 | 2024-06-04 | Huawei Technologies Co., Ltd. | Patch data unit coding and decoding for point-cloud coding |
Families Citing this family (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11463681B2 (en) * | 2018-02-23 | 2022-10-04 | Nokia Technologies Oy | Encoding and decoding of volumetric video |
EP3780613A4 (en) * | 2018-04-11 | 2021-05-19 | Sony Corporation | IMAGE PROCESSING DEVICE AND METHOD |
JP7100523B2 (ja) * | 2018-07-27 | 2022-07-13 | 京セラ株式会社 | 表示装置、表示システムおよび移動体 |
US11956478B2 (en) * | 2019-01-09 | 2024-04-09 | Tencent America LLC | Method and apparatus for point cloud chunking for improved patch packing and coding efficiency |
US20220159261A1 (en) * | 2019-03-21 | 2022-05-19 | Lg Electronics Inc. | Point cloud data transmission device, point cloud data transmission method, point cloud data reception device, and point cloud data reception method |
KR102533985B1 (ko) * | 2021-05-21 | 2023-05-17 | 한국항공대학교산학협력단 | 포인트 클라우드 복원 시스템 및 방법 |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2012506573A (ja) * | 2008-10-21 | 2012-03-15 | コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ | シーンの階層化深さモデルを提供するための方法及び装置並びにシーンの階層化深さモデルを有する信号 |
JP2016511457A (ja) * | 2013-01-18 | 2016-04-14 | マイクロソフト テクノロジー ライセンシング,エルエルシー | 深度マップ上への再投影を使用するサーフェスコーデック |
Family Cites Families (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR100446635B1 (ko) * | 2001-11-27 | 2004-09-04 | 삼성전자주식회사 | 깊이 이미지 기반 3차원 객체 표현 장치 및 방법 |
JP3957620B2 (ja) * | 2001-11-27 | 2007-08-15 | 三星電子株式会社 | 深さイメージ基盤3次元客体を表現するための装置及び方法 |
JP5480914B2 (ja) * | 2009-12-11 | 2014-04-23 | 株式会社トプコン | 点群データ処理装置、点群データ処理方法、および点群データ処理プログラム |
JP5522367B2 (ja) | 2009-12-13 | 2014-06-18 | 国立大学法人 千葉大学 | 表面再現方法及び表面再現プログラム |
JP5161991B2 (ja) * | 2011-03-25 | 2013-03-13 | 株式会社東芝 | 画像処理装置 |
WO2012141235A1 (ja) * | 2011-04-13 | 2012-10-18 | 株式会社トプコン | 三次元点群位置データ処理装置、三次元点群位置データ処理システム、三次元点群位置データ処理方法およびプログラム |
CN103959342A (zh) * | 2011-11-25 | 2014-07-30 | 汤姆逊许可公司 | 基于具有复制点的空间树的位置编码 |
US9075933B2 (en) * | 2012-10-11 | 2015-07-07 | Adobe Systems Incorporated | 3D transformation of objects using 2D controls projected in 3D space and contextual face selections of a three dimensional bounding box |
WO2015006224A1 (en) * | 2013-07-08 | 2015-01-15 | Vangogh Imaging, Inc. | Real-time 3d computer vision processing engine for object recognition, reconstruction, and analysis |
WO2015055607A2 (en) * | 2013-10-14 | 2015-04-23 | Koninklijke Philips N.V. | Remapping a depth map for 3d viewing |
US9465361B2 (en) | 2014-03-31 | 2016-10-11 | Disney Enterprises, Inc. | Image based multiview multilayer holographic rendering algorithm |
CA2948903C (en) * | 2014-05-13 | 2020-09-22 | Pcp Vr Inc. | Method, system and apparatus for generation and playback of virtual reality multimedia |
US10142608B2 (en) * | 2014-07-08 | 2018-11-27 | Samsung Electronics Co., Ltd. | Electronic apparatus and method for processing three-dimensional information using image |
US20170214943A1 (en) * | 2016-01-22 | 2017-07-27 | Mitsubishi Electric Research Laboratories, Inc. | Point Cloud Compression using Prediction and Shape-Adaptive Transforms |
CN106934861B (zh) * | 2017-02-09 | 2020-06-26 | 深圳先进技术研究院 | 物体三维重建方法及装置 |
CN107481313A (zh) * | 2017-08-18 | 2017-12-15 | 深圳市唯特视科技有限公司 | 一种基于学习有效点云生成的密集三维物体重建方法 |
-
2018
- 2018-12-27 TW TW107147408A patent/TWI815842B/zh active
-
2019
- 2019-01-07 BR BR112020013599-2A patent/BR112020013599A2/pt unknown
- 2019-01-07 US US16/960,910 patent/US11721048B2/en active Active
- 2019-01-07 MX MX2020007416A patent/MX2020007416A/es unknown
- 2019-01-07 WO PCT/JP2019/000051 patent/WO2019142666A1/ja unknown
- 2019-01-07 RU RU2020122782A patent/RU2020122782A/ru unknown
- 2019-01-07 SG SG11202005374UA patent/SG11202005374UA/en unknown
- 2019-01-07 JP JP2019566411A patent/JP7327166B2/ja active Active
- 2019-01-07 CA CA3087980A patent/CA3087980A1/en active Pending
- 2019-01-07 AU AU2019210233A patent/AU2019210233B2/en active Active
- 2019-01-07 CN CN201980007870.2A patent/CN111566702B/zh active Active
- 2019-01-07 EP EP19741709.0A patent/EP3742401A4/en active Pending
- 2019-01-07 KR KR1020237037047A patent/KR20230155019A/ko active Search and Examination
- 2019-01-07 KR KR1020207019305A patent/KR102596507B1/ko active IP Right Grant
-
2020
- 2020-07-09 PH PH12020500587A patent/PH12020500587A1/en unknown
- 2020-08-14 ZA ZA2020/05038A patent/ZA202005038B/en unknown
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2012506573A (ja) * | 2008-10-21 | 2012-03-15 | コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ | シーンの階層化深さモデルを提供するための方法及び装置並びにシーンの階層化深さモデルを有する信号 |
JP2016511457A (ja) * | 2013-01-18 | 2016-04-14 | マイクロソフト テクノロジー ライセンシング,エルエルシー | 深度マップ上への再投影を使用するサーフェスコーデック |
Non-Patent Citations (5)
Title |
---|
"PCC Test Model Category 2 vO", N17248 MPEG OUTPUT DOCUMENT, October 2017 (2017-10-01) |
GOLLA, TIM ET AL.: "Real-time Point Cloud Compression", IEEE /RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS, 2015, pages 5087 - 5092, XP032832361, doi:10.1109/IROS.2015.7354093 * |
JIANLE CHENELENA ALSHINAGARY J. SULLIVANJENS-RAINERJILL BOYCE: "Algorithm Description of Joint Exploration Test Model 4", JVET-G1001_V1, JOINT VIDEO EXPLORATION TEAM (JVET) OF ITU-T SG 16 WP 3 AND ISO/IEC JTC 1/SC 29/WG 11 7TH MEETING, 13 July 2017 (2017-07-13) |
K. MAMMOU: "Video-based and Hierarchical Approaches Point Cloud Compression", MPEG M41649, October 2017 (2017-10-01) |
TIM GOLLAREINHARD KLEIN: "Real-time Point Cloud Compression", 2015, IEEE |
Cited By (28)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11949889B2 (en) | 2018-01-19 | 2024-04-02 | Interdigital Vc Holdings, Inc. | Processing a point cloud |
JP7476104B2 (ja) | 2018-01-19 | 2024-04-30 | インターデジタル ヴイシー ホールディングス, インコーポレイテッド | 点群処理 |
JP2021511712A (ja) * | 2018-01-19 | 2021-05-06 | インターデジタル ヴイシー ホールディングス, インコーポレイテッド | 点群処理 |
US11856222B2 (en) | 2018-04-11 | 2023-12-26 | Interdigital Vc Holdings, Inc. | Method and apparatus for encoding/decoding a point cloud representing a 3D object |
JP7389751B2 (ja) | 2018-04-11 | 2023-11-30 | インターデジタル ヴイシー ホールディングス, インコーポレイテッド | 三次元物体を表すポイントクラウドを符号化/復号する方法及び装置 |
JP2021518694A (ja) * | 2018-04-11 | 2021-08-02 | インターデジタル ヴイシー ホールディングス, インコーポレイテッド | 三次元物体を表すポイントクラウドを符号化/復号する方法及び装置 |
JP7488189B2 (ja) | 2018-04-11 | 2024-05-21 | インターデジタル ヴイシー ホールディングス, インコーポレイテッド | 点群の幾何学的形状をコーディングする方法およびデバイス |
JP2021518689A (ja) * | 2018-04-11 | 2021-08-02 | インターデジタル ヴイシー ホールディングス, インコーポレイテッド | 点群の幾何学的形状をコーディングする方法およびデバイス |
JP2021519533A (ja) * | 2018-04-11 | 2021-08-10 | インターデジタル ヴイシー ホールディングス, インコーポレイテッド | 投影平面の少なくとも1つの画像領域に一度直交投影された3d点のセットの深度値を符号化するための方法 |
JP7446234B2 (ja) | 2018-04-11 | 2024-03-08 | インターデジタル ヴイシー ホールディングス, インコーポレイテッド | 投影平面の少なくとも1つの画像領域に一度直交投影された3d点のセットの深度値を符号化するための方法 |
JP2021518681A (ja) * | 2018-04-11 | 2021-08-02 | インターデジタル ヴイシー ホールディングス, インコーポレイテッド | 三次元物体を表すポイントクラウドを符号化/復号する方法及び装置 |
JP7431742B2 (ja) | 2018-04-11 | 2024-02-15 | インターデジタル ヴイシー ホールディングス, インコーポレイテッド | 三次元物体を表すポイントクラウドを符号化/復号する方法及び装置 |
US11838547B2 (en) | 2018-04-11 | 2023-12-05 | Interdigital Vc Holdings, Inc. | Method and device for encoding the geometry of a point cloud |
JP2021531542A (ja) * | 2018-07-11 | 2021-11-18 | インターデジタル ヴイシー ホールディングス, インコーポレイテッド | 点群の点のテクスチャを符号化/復号する方法 |
US11699248B2 (en) | 2018-07-11 | 2023-07-11 | Sony Corporation | Image processing apparatus and method |
JP7327399B2 (ja) | 2018-07-11 | 2023-08-16 | ソニーグループ株式会社 | 画像処理装置および方法 |
JP7351858B2 (ja) | 2018-07-11 | 2023-09-27 | インターデジタル ヴイシー ホールディングス, インコーポレイテッド | 点群の点のテクスチャを符号化/復号する方法 |
JPWO2020012968A1 (ja) * | 2018-07-11 | 2021-08-02 | ソニーグループ株式会社 | 画像処理装置および方法 |
WO2020012968A1 (ja) * | 2018-07-11 | 2020-01-16 | ソニー株式会社 | 画像処理装置および方法 |
WO2020017359A1 (ja) * | 2018-07-20 | 2020-01-23 | ソニー株式会社 | 画像処理装置および方法 |
US11356690B2 (en) | 2018-07-20 | 2022-06-07 | Sony Corporation | Image processing apparatus and method |
JP7267447B2 (ja) | 2019-03-12 | 2023-05-01 | ホアウェイ・テクノロジーズ・カンパニー・リミテッド | 点群コーディングのためのパッチデータユニットのコーディングおよび復号 |
JP2022525599A (ja) * | 2019-03-12 | 2022-05-18 | ホアウェイ・テクノロジーズ・カンパニー・リミテッド | 点群コーディングのためのパッチデータユニットのコーディングおよび復号 |
JP7402976B2 (ja) | 2019-09-30 | 2023-12-21 | 中興通訊股▲ふん▼有限公司 | 点群データのためのファイルフォーマット |
JP2022550150A (ja) * | 2019-09-30 | 2022-11-30 | 中興通訊股▲ふん▼有限公司 | 点群データのためのファイルフォーマット |
WO2021095565A1 (ja) * | 2019-11-13 | 2021-05-20 | ソニーグループ株式会社 | 画像処理装置および方法 |
WO2021193088A1 (ja) * | 2020-03-25 | 2021-09-30 | ソニーグループ株式会社 | 画像処理装置および方法 |
US12002243B2 (en) | 2021-09-03 | 2024-06-04 | Huawei Technologies Co., Ltd. | Patch data unit coding and decoding for point-cloud coding |
Also Published As
Publication number | Publication date |
---|---|
PH12020500587A1 (en) | 2021-06-14 |
TWI815842B (zh) | 2023-09-21 |
CN111566702B (zh) | 2024-03-26 |
RU2020122782A (ru) | 2022-01-10 |
MX2020007416A (es) | 2020-09-14 |
EP3742401A1 (en) | 2020-11-25 |
CA3087980A1 (en) | 2019-07-25 |
US11721048B2 (en) | 2023-08-08 |
TW201933284A (zh) | 2019-08-16 |
KR20230155019A (ko) | 2023-11-09 |
EP3742401A4 (en) | 2020-12-09 |
KR20200108833A (ko) | 2020-09-21 |
AU2019210233A1 (en) | 2020-07-23 |
CN111566702A (zh) | 2020-08-21 |
BR112020013599A2 (pt) | 2020-12-01 |
ZA202005038B (en) | 2021-05-26 |
JP7327166B2 (ja) | 2023-08-16 |
US20210174559A1 (en) | 2021-06-10 |
JPWO2019142666A1 (ja) | 2021-01-28 |
SG11202005374UA (en) | 2020-07-29 |
AU2019210233B2 (en) | 2023-10-19 |
KR102596507B1 (ko) | 2023-11-01 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2019142666A1 (ja) | 画像処理装置および方法 | |
JP7384159B2 (ja) | 画像処理装置および方法 | |
WO2019198523A1 (ja) | 画像処理装置および方法 | |
WO2019142667A1 (ja) | 画像処理装置および方法 | |
US11699248B2 (en) | Image processing apparatus and method | |
JP7331852B2 (ja) | 画像処理装置および方法 | |
KR20200140256A (ko) | 화상 처리 장치 및 방법 | |
US11399189B2 (en) | Image processing apparatus and method | |
WO2020188932A1 (ja) | 情報処理装置および情報生成方法 | |
JP7396302B2 (ja) | 画像処理装置および方法 | |
WO2022145214A1 (ja) | 情報処理装置および方法 | |
WO2020137603A1 (ja) | 画像処理装置および方法 | |
WO2022201787A1 (ja) | 画像処理装置および方法 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 19741709 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2019566411 Country of ref document: JP Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 3087980 Country of ref document: CA |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
ENP | Entry into the national phase |
Ref document number: 2019210233 Country of ref document: AU Date of ref document: 20190107 Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 2019741709 Country of ref document: EP Effective date: 20200817 |
|
REG | Reference to national code |
Ref country code: BR Ref legal event code: B01A Ref document number: 112020013599 Country of ref document: BR |
|
ENP | Entry into the national phase |
Ref document number: 112020013599 Country of ref document: BR Kind code of ref document: A2 Effective date: 20200702 |