TW201933284A - Image processing device and method - Google Patents

Image processing device and method Download PDF

Info

Publication number
TW201933284A
TW201933284A TW107147408A TW107147408A TW201933284A TW 201933284 A TW201933284 A TW 201933284A TW 107147408 A TW107147408 A TW 107147408A TW 107147408 A TW107147408 A TW 107147408A TW 201933284 A TW201933284 A TW 201933284A
Authority
TW
Taiwan
Prior art keywords
data
dimensional
unit
recorded
image processing
Prior art date
Application number
TW107147408A
Other languages
Chinese (zh)
Other versions
TWI815842B (en
Inventor
中神央二
矢野幸司
隈智
加藤毅
Original Assignee
日商索尼股份有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 日商索尼股份有限公司 filed Critical 日商索尼股份有限公司
Publication of TW201933284A publication Critical patent/TW201933284A/en
Application granted granted Critical
Publication of TWI815842B publication Critical patent/TWI815842B/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T11/002D [Two Dimensional] image generation
    • G06T11/003Reconstruction from projections, e.g. tomography
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T9/00Image coding
    • G06T9/001Model-based coding, e.g. wire frame
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/50Depth or shape recovery
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T9/00Image coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/597Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10028Range image; Depth image; 3D point clouds

Abstract

The present disclosure relates to an image processing device and method which make it possible to suppress a reduction in quality resulting from a two-dimensional projection of 3D data. Data for every position included in 3D data representing a three-dimensional structure are projected onto two-dimensional planes in a plurality of layers. Further, data for every position in 3D data projected onto two-dimensional planes in a number of layers indicated by layer number information are projected into a three-dimensional space. The present disclosure is applicable, for example, to information processing devices, image processing devices, electronic equipment, information processing methods and programs.

Description

影像處理裝置及方法Image processing device and method

本揭露係有關於影像處理裝置及方法,特別是有關於可抑制3D資料的2維投影所致之品質的降低的影像處理裝置及方法。The present disclosure relates to an image processing apparatus and method, and more particularly to an image processing apparatus and method that can suppress a decrease in quality due to two-dimensional projection of 3D data.

先前,作為例如像是點雲(Point cloud)這類表示3維結構之3D資料的編碼方法,係有分別將點雲的位置與色彩資訊,每一小領域地投影至2維平面,以2維影像用之編碼方法加以編碼的方式(以下亦稱為視訊基礎方式(Video-based approach)),已被提出(例如參照非專利文獻1乃至非專利文獻3)。
[先前技術文獻]
[非專利文獻]
Previously, as a coding method for representing a 3D structure such as a point cloud, such as a point cloud, the position and color information of the point cloud are respectively projected to a 2-dimensional plane, and 2 A method of encoding a coding method for a video image (hereinafter also referred to as a video-based approach) has been proposed (for example, refer to Non-Patent Document 1 to Non-Patent Document 3).
[Previous Technical Literature]
[Non-patent literature]

[非專利文獻1] Tim Golla and Reinhard Klein, "Real-time Point Cloud Compression ", IEEE, 2015
[非專利文獻2] K. Mammou, "Video-based and Hierarchical Approaches Point Cloud Compression", MPEG m41649, Oct. 2017
[非專利文獻3] "PCC Test Model Category 2 v0", N17248 MPEG output document, October 2017
[Non-Patent Document 1] Tim Golla and Reinhard Klein, "Real-time Point Cloud Compression", IEEE, 2015
[Non-Patent Document 2] K. Mammou, "Video-based and Hierarchical Approaches Point Cloud Compression", MPEG m41649, Oct. 2017
[Non-Patent Document 3] "PCC Test Model Category 2 v0", N17248 MPEG output document, October 2017

[發明所欲解決之課題][Problems to be solved by the invention]

然而,編碼對象的點雲,係會因為雜訊或攝像系統之特性,而導致在物件表面以外也有Point存在。因此,有時候難以投影至2維平面,藉由伴隨對該2維平面之投影的編碼,可能會導致品質降低。However, the point cloud of the encoding object may have a Point outside the surface of the object due to the characteristics of the noise or camera system. Therefore, sometimes it is difficult to project to a 2-dimensional plane, and the encoding accompanying the projection of the 2-dimensional plane may result in a decrease in quality.

本揭露係有鑑於此種狀況而研發,目的在於能夠抑制3D資料的2維投影所致之品質的降低。

[用以解決課題之手段]
The present disclosure has been developed in view of such a situation, and aims to suppress a decrease in quality due to two-dimensional projection of 3D data.

[Means to solve the problem]

本技術之一側面之影像處理裝置,係為一種影像處理裝置,係具備:2維投影部,係將表示3維結構之3D資料中所含之所有的每一位置之資料,投影至複數圖層之2維平面。An image processing device according to one aspect of the present invention is an image processing device comprising: a two-dimensional projection unit that projects data of each position included in a 3D structure of a three-dimensional structure onto a plurality of layers 2D plane.

本技術之一側面之影像處理方法,係為一種影像處理裝置,係將表示3維結構之3D資料中所含之所有的每一位置之資料,投影至複數圖層之2維平面。An image processing method according to one aspect of the present technology is an image processing apparatus that projects data of each position included in 3D data of a three-dimensional structure onto a two-dimensional plane of a plurality of layers.

本技術之另一側面的影像處理裝置,係為一種影像處理裝置,係具備:3維投影部,係將已被投影至圖層數資訊所表示之圖層數之2維平面上的3D資料的所有的每一位置之資料,投影至3維空間。The image processing device according to another aspect of the present technology is an image processing device comprising: a three-dimensional projection unit that is a 3D image that has been projected onto a two-dimensional plane of the number of layers indicated by the layer number information. All the information of each position is projected into the 3-dimensional space.

本技術之另一側面的影像處理方法,係為一種影像處理方法,係將已被投影至圖層數資訊所表示之圖層數之2維平面上的3D資料的所有的每一位置之資料,投影至3維空間。The image processing method on the other side of the technology is an image processing method, which is data of all the positions of the 3D data on the two-dimensional plane that has been projected onto the number of layers indicated by the layer number information. , projected to 3D space.

於本技術之一側面的影像處理裝置及方法中,表示3維結構之3D資料中所含之所有的每一位置之資料,會被投影至複數圖層之2維平面。In the image processing apparatus and method of one aspect of the present technology, data representing each of the positions included in the 3D data of the three-dimensional structure is projected onto the two-dimensional plane of the complex layer.

於本技術之另一側面的影像處理裝置及方法中,已被投影至圖層數資訊所表示之圖層數之2維平面上的3D資料的所有的每一位置之資料,會被投影至3維空間。

[發明效果]
In the image processing apparatus and method of the other aspect of the present technology, data of each position of the 3D data that has been projected onto the two-dimensional plane of the number of layers indicated by the layer number information is projected to 3 dimensional space.

[Effect of the invention]

若依據本揭露,則可處理資訊。尤其是,可抑制3D資料的2維投影所致之品質的降低。According to the disclosure, information can be processed. In particular, it is possible to suppress a decrease in quality due to 2-dimensional projection of 3D data.

以下,說明用以實施本揭露的形態(以下稱作實施形態)。此外,說明係用以下順序來進行。
1.視訊基礎方式
2.第1實施形態(可變圖層數)
3.第2實施形態(空點之定義)
4.第3實施形態(可變縱深參數)
5.附記
Hereinafter, a mode for carrying out the disclosure (hereinafter referred to as an embodiment) will be described. Further, the description is made in the following order.
Video based method
2. First embodiment (variable layer number)
3. Second embodiment (definition of empty points)
4. Third embodiment (variable depth parameter)
5. Attachment

<1.視訊基礎方式>
<支持技術內容、技術用語的文獻等>
本技術所揭露之範圍,係不只有實施例中所記載的內容,還包含了於申請當時已為公知的以下之非專利文獻中所記載的內容。
<1. Video Basics>
<Documents supporting technical content and technical terms, etc.>
The scope of the present invention is not limited to the contents described in the examples, but also includes the contents described in the following non-patent documents which are known at the time of application.

非專利文獻1:(上述)
非專利文獻2:(上述)
非專利文獻3:(上述)
非專利文獻4:TELECOMMUNICATION STANDARDIZATION SECTOR OF ITU(International Telecommunication Union), "Advanced video coding for generic audiovisual services", H.264, 04/2017
非專利文獻5:TELECOMMUNICATION STANDARDIZATION SECTOR OF ITU(International Telecommunication Union), "High efficiency video coding", H.265, 12/2016
非專利文獻6:Jianle Chen, Elena Alshina, Gary J. Sullivan, Jens-Rainer, Jill Boyce, "Algorithm Description of Joint Exploration Test Model 4", JVET-G1001_v1, Joint Video Exploration Team (JVET) of ITU-T SG 16 WP 3 and ISO/IEC JTC 1/SC 29/WG 11 7th Meeting: Torino, IT, 13-21 July 2017
Non-Patent Document 1: (above)
Non-Patent Document 2: (above)
Non-Patent Document 3: (above)
Non-Patent Document 4: TELECOMMUNICATION STANDARDIZATION SECTOR OF ITU (International Telecommunication Union), "Advanced video coding for generic audiovisual services", H.264, 04/2017
Non-Patent Document 5: TELECOMMUNICATION STANDARDIZATION SECTOR OF ITU (International Telecommunication Union), "High efficiency video coding", H.265, 12/2016
Non-Patent Document 6: Jianle Chen, Elena Alshina, Gary J. Sullivan, Jens-Rainer, Jill Boyce, "Algorithm Description of Joint Exploration Test Model 4", JVET-G1001_v1, Joint Video Exploration Team (JVET) of ITU-T SG 16 WP 3 and ISO/IEC JTC 1/SC 29/WG 11 7th Meeting: Torino, IT, 13-21 July 2017

亦即,上述的非專利文獻中所記載之內容亦為判斷支持要件之際的根據。例如,非專利文獻5中所記載之Quad-Tree Block Structure、非專利文獻6中所記載之QTBT(Quad Tree Plus Binary Tree) Block Structure即使於實施例中沒有直接記載的情況下,仍屬於本技術的揭露範圍內,並視為滿足申請專利範圍的支持要件。又,例如,關於剖析(Parsing)、語法(Syntax)、語意(Semantics)等之技術用語也是同樣地,即使於實施例中沒有直接記載的情況下,仍屬於本技術的揭露範圍內,並視為滿足申請專利範圍的支持要件。That is, the content described in the above non-patent document is also a basis for judging the support requirement. For example, the Quad-Tree Block Structure described in Non-Patent Document 5 and the QTBT (Quad Tree Plus Binary Tree) Block Structure described in Non-Patent Document 6 are still in the present technology even if they are not directly described in the examples. Within the scope of disclosure, and deemed to meet the support requirements of the scope of patent application. Further, for example, the technical terms such as Parsing, Syntax, and Semantics are also the same, and even if they are not directly described in the examples, they are still within the scope of the present technology. Supporting requirements for the scope of patent application.

<點雲>
先前,藉由點群之位置資訊或屬性資訊等來表現3維結構的點雲,或以頂點、邊緣、面所構成,使用多角形表現來定義3維形狀的網格等之資料,係為存在。
<point cloud>
Previously, a point cloud of a three-dimensional structure was represented by position information or attribute information of a point group, or a vertex, an edge, and a surface were used, and a polygon representation was used to define a three-dimensional shape of the mesh, etc. presence.

例如點雲的情況下,是將如圖1的A所示的立體結構物,以如圖1的B所示的多數的點之集合(點群),來加以表現。亦即,點雲之資料,係由該點群的各點之位置資訊或屬性資訊(例如色彩等)而被構成。因此資料結構會比較單純,同時,藉由使用足夠多的點,就可以足夠的精度來表現任意的立體結構。For example, in the case of a point cloud, the three-dimensional structure shown in A of FIG. 1 is represented by a set of points (point groups) of a plurality of points as shown in FIG. 1B. That is, the data of the point cloud is composed of position information or attribute information (for example, color, etc.) of each point of the point group. Therefore, the data structure will be relatively simple, and at the same time, by using enough points, the arbitrary three-dimensional structure can be expressed with sufficient precision.

<視訊基礎方式之概要>
將此種點雲的各位置與色彩資訊,每一小領域地投影至2維平面,以2維影像用之編碼方法來進行編碼的視訊基礎方式(Video-based approach),已被提出。
<Overview of video basic methods>
A video-based approach in which each position of the point cloud and color information, each small field is projected onto a two-dimensional plane, and encoded by a two-dimensional image coding method has been proposed.

在該視訊基礎方式中,例如如圖2所示,已被輸入之點雲(Point cloud)是被分割成複數個分段(亦稱領域或斑塊),每一領域地被投影至2維平面。此外,點雲的每一位置之資料(亦即各點之資料),係如上述是由位置資訊(Geometry(亦稱Depth))和屬性資訊(Texture)所構成,分別被每一領域地投影至2維平面。In the video basic mode, for example, as shown in FIG. 2, the point cloud that has been input is divided into a plurality of segments (also called fields or plaques), and each field is projected to 2 dimensions. flat. In addition, the data of each location of the point cloud (that is, the data of each point) is composed of position information (Geometry (also known as Depth)) and attribute information (Texture) as described above, and is projected by each field. To the 2-dimensional plane.

然後,已被投影至2維平面上的3D資料(點雲),係藉由例如AVC(Advanced Video Coding)或HEVC (High Efficiency Video Coding)等這類2維平面影像用之編碼方式,而被編碼。Then, the 3D data (point cloud) that has been projected onto the 2-dimensional plane is encoded by a two-dimensional plane image such as AVC (Advanced Video Coding) or HEVC (High Efficiency Video Coding). coding.

<有關於視訊基礎方式的本技術>
說明有關於如以上所說明之視訊基礎方式的本技術。圖3係為於各實施形態所說明的本技術之一覽。
<There is a technology related to the basic method of video>
The present technology is described with respect to the basic method of video as explained above. Fig. 3 is a list of the present technology described in each embodiment.

該表從上起算第1段(項目名的段除外)中係記載了有關於先前(TMC2)中的視訊基礎方式。亦即,於先前的視訊基礎方式中,3D資料的2維投影,係對2圖層(2層)之2維平面進行。該規格係於畫面全體中為共通(無論針對哪個區段都是進行相同的投影)。因此,此種被使用於投影之控制的投影控制資訊,係作為畫格單位之資訊,從編碼側而被傳訊至解碼側。From the top of the table, the first paragraph (excluding the section of the project name) describes the basic method of video in the previous (TMC2). That is, in the previous video basic mode, the 2-dimensional projection of the 3D data is performed on the 2-dimensional plane of the 2 layers (2 layers). This specification is common to all screens (the same projection is performed regardless of which section is used). Therefore, the projection control information used for the control of the projection is transmitted as information of the frame unit from the encoding side to the decoding side.

然而,編碼對象的點雲,係會因為雜訊或攝像系統之特性,而導致在物件表面以外也有Point存在。因此,如上述的先前的方法那樣,對2圖層之2維平面的投影,可能會有困難。因此,會發生無法投影至2維平面的點,會因為往2維平面之投影所伴隨的編碼,而有可能導致資料品質降低。However, the point cloud of the encoding object may have a Point outside the surface of the object due to the characteristics of the noise or camera system. Therefore, as in the previous method described above, it may be difficult to project a two-dimensional plane of two layers. Therefore, a point that cannot be projected onto a two-dimensional plane occurs, which may result in degradation of data quality due to coding accompanying projection onto a two-dimensional plane.

又,例如,在壓縮對象是具有Sparse(稀疏)之性質的情況下,就會需要微小領域(Point)單位之投影,而為了將許多的小領域予以編碼,處理量、編碼效率都可能會變差。Further, for example, in the case where the compressed object has the property of Sparse, a projection of a minute unit is required, and in order to encode many small fields, the processing amount and coding efficiency may change. difference.

圖3的表從上起算第2段(項目名的段除外)中係記載了,有關於第1實施形態(實施例1)中所說明的本技術。該方法,係將投影3D資料的2維平面之圖層(Layer)數設為可變,以使得可將縱深方向上重疊的所有的點之資料(每一位置之資料)進行投影的方式,來設定圖層數的方法。The table of Fig. 3 is described in the second paragraph (excluding the segment of the item name) from the top, and the present technology described in the first embodiment (the first embodiment) is described. In this method, the number of layers of the two-dimensional plane of the projected 3D data is set to be variable so that the data of all the points (the data of each position) overlapping in the depth direction can be projected. The method of setting the number of layers.

此情況下,亦可為,例如,表示圖層數的資訊,是每一領域地,從編碼側被傳訊至解碼側。In this case, for example, information indicating the number of layers may be transmitted from the encoding side to the decoding side in each field.

藉由如此設計,就可較正確地重現在物件表面具有厚度的點雲。亦即,可抑制3D資料的2維投影所致之品質的降低。By designing in this way, it is possible to more accurately reproduce the point cloud having a thickness on the surface of the object. That is, it is possible to suppress a decrease in quality due to 2-dimensional projection of 3D data.

圖3的表從上起算第3段(項目名的段除外)中係記載了,有關於第2實施形態(實施例2)中所說明的本技術。該方法係為,將3D資料投影至2維平面之際,追加「空點」之定義的方法。The table of Fig. 3 is described in the third paragraph (excluding the segment of the item name) from the top, and the present technology described in the second embodiment (the second embodiment) is described. This method is a method of adding a definition of "empty point" when projecting 3D data to a two-dimensional plane.

此情況下,例如,於解碼側做刪除的點的像素值之定義,亦可從編碼側被傳訊至解碼側。In this case, for example, the definition of the pixel value of the point to be deleted on the decoding side may be transmitted from the encoding side to the decoding side.

藉由如此設計,就可較正確地重現Sparse的點雲。亦即,不需要微小領域(Point)單位之投影,可抑制處理量的增大或編碼效率的降低。By designing this way, Sparse's point cloud can be reproduced more correctly. That is, it is not necessary to project a minute unit, and it is possible to suppress an increase in the amount of processing or a decrease in coding efficiency.

圖3的表從上起算第4段(項目名的段除外)中係記載了,有關於第3實施形態(實施例3)中所說明的本技術。該方法係為,可將控制投影至2維平面的3D資料之縱深之範圍的縱深參數,對每一領域加以設定的方法。The table of Fig. 3 is described in the fourth paragraph (excluding the segment of the item name) from the top, and the present technology described in the third embodiment (Example 3) is described. This method is a method in which the depth parameter of the range of the depth of the 3D data projected onto the two-dimensional plane can be set for each field.

此情況下,例如,其縱深參數亦可從編碼側被傳訊至解碼側。In this case, for example, its depth parameter can also be transmitted from the encoding side to the decoding side.

藉由如此設計,可控制每一領域之畫質,可抑制位置資訊(Geometry)的效率之降低。With this design, the quality of each field can be controlled, and the efficiency of positional information (Geometry) can be suppressed.

<編碼裝置>
接著說明,實現如以上之各手法的構成。圖4係適用了本技術的影像處理裝置之一態樣的編碼裝置之構成之一例的區塊圖。圖4所示的編碼裝置100係為,將像是點雲的3D資料投影至2維平面而以2維影像用之編碼方法進行編碼的裝置。
<encoding device>
Next, the configuration of each of the above methods will be realized. Fig. 4 is a block diagram showing an example of a configuration of an encoding device to which one aspect of the image processing device of the present technology is applied. The coding apparatus 100 shown in FIG. 4 is a device that projects 3D data such as a point cloud onto a two-dimensional plane and encodes it by a coding method for two-dimensional video.

例如,編碼裝置100,係實作非專利文獻1乃至非專利文獻6中所記載之技術,以符合這些文獻之任一者所記載之規格的方法,進行3D資料之編碼。For example, the encoding device 100 implements the techniques described in Non-Patent Document 1 or Non-Patent Document 6, and encodes 3D data in a method conforming to the specifications described in any of these documents.

此外,於圖4中係主要表示處理部或資料的流動等,圖4所示者並非全部。亦即,於編碼裝置100中,亦可存在有於圖4中未以區塊而被表示的處理部,亦可存在有於圖4中未以箭頭等而表示的處理或資料之流動。這在編碼裝置100內的處理部等的說明的其他圖中也是同樣如此。In addition, in FIG. 4, the flow of a processing part or a material is mainly shown, and it is not shown in FIG. That is, the encoding device 100 may have a processing unit that is not represented by a block in FIG. 4, and may have a flow of processing or data that is not indicated by an arrow or the like in FIG. This is also the same in other figures of the description of the processing unit and the like in the encoding device 100.

如圖4所示編碼裝置100係具有:斑塊分解部111、打包部112、輔助斑塊資訊壓縮部113、視訊編碼部114、視訊編碼部115、OMap編碼部116、及多工器117。As shown in FIG. 4, the encoding apparatus 100 includes a plaque decomposition unit 111, a packing unit 112, an auxiliary patch information compressing unit 113, a video encoding unit 114, a video encoding unit 115, an OMap encoding unit 116, and a multiplexer 117.

斑塊分解部111,係進行3D資料之分解的相關處理。例如,斑塊分解部111,係將被輸入至編碼裝置100的,表示3維結構之3D資料(例如點雲),加以取得(箭頭121)。又,斑塊分解部111,係將已取得之該3D資料分解成複數斑塊,該每一斑塊地將3D資料投影至2維平面。The plaque decomposition unit 111 performs correlation processing for decomposition of 3D data. For example, the plaque decomposition unit 111 acquires 3D data (for example, a point cloud) indicating a three-dimensional structure, which is input to the encoding device 100, and acquires it (arrow 121). Further, the plaque decomposition unit 111 decomposes the acquired 3D data into a plurality of plaques, and the plaque projects the 3D data onto the two-dimensional plane.

斑塊分解部111,係將每一斑塊地被投影至2維平面上的3D資料,供給至打包部112(箭頭122)。又,斑塊分解部111,係將該分解的相關之資訊亦即輔助斑塊資訊,供給至輔助斑塊資訊壓縮部113(箭頭123)。The plaque disassembling unit 111 supplies the 3D material projected onto the two-dimensional plane for each plaque to the packing unit 112 (arrow 122). Further, the plaque decomposition unit 111 supplies the information related to the decomposition, that is, the auxiliary plaque information, to the auxiliary plaque information compressing unit 113 (arrow 123).

打包部112,係進行資料之打包的相關處理。例如,打包部112,係將從斑塊分解部111所被供給之每一斑塊地被投影了3D資料的2維平面之資料,加以取得(箭頭122)。又,打包部112,係將已取得之該2維平面之各圖層打包成為彼此互異之視訊畫格。例如,打包部112,係將表示點的位置的位置資訊(Gepmetry)、被附加至該位置資訊的色彩資訊等之屬性資訊(Texture)、及表示點的存在有無的佔用地圖(Occupancy Map),分別打包成為視訊畫格。The packing unit 112 performs related processing of packing the data. For example, the packing unit 112 acquires data of a two-dimensional plane on which 3D data is projected from each plaque supplied from the plaque decomposition unit 111 (arrow 122). Further, the packing unit 112 packs the acquired layers of the two-dimensional plane into mutually different video frames. For example, the packing unit 112 is a position information (Gepmetry) indicating the position of the point, attribute information (Texture) such as color information attached to the position information, and an Occupancy Map indicating the presence or absence of the point. Packed separately into a video frame.

打包部112,係將已生成之視訊畫格,供給至後段的處理部(箭頭124)。例如,打包部112係將已生成之位置資訊(Geometry)之視訊畫格,供給至視訊編碼部114。又,例如,打包部112係將已生成之屬性資訊(Texture)之視訊畫格,供給至視訊編碼部115。再者,例如,打包部112係將已生成之佔用地圖之視訊畫格,供給至OMap編碼部116。The packing unit 112 supplies the generated video frame to the processing unit (arrow 124) in the subsequent stage. For example, the packing unit 112 supplies the video frame of the generated position information (Geometry) to the video encoding unit 114. Further, for example, the packing unit 112 supplies the video frame of the generated attribute information (Texture) to the video encoding unit 115. Further, for example, the packing unit 112 supplies the generated video frame of the occupied map to the OMap encoding unit 116.

又,打包部112係將該打包的相關之控制資訊,供給至多工器117(箭頭125)。Further, the packing unit 112 supplies the packaged control information to the multiplexer 117 (arrow 125).

輔助斑塊資訊壓縮部113,係進行輔助斑塊資訊之壓縮的相關處理。例如,輔助斑塊資訊壓縮部113,係將從斑塊分解部111所被供給之資料,加以取得(箭頭123)。輔助斑塊資訊壓縮部113,係將已取得之資料中所含之輔助斑塊資訊,予以編碼(壓縮)。輔助斑塊資訊壓縮部113,係將所得到之輔助斑塊資訊之編碼資料,供給至多工器117(箭頭126)。The auxiliary patch information compressing unit 113 performs correlation processing for assisting compression of the patch information. For example, the auxiliary patch information compressing unit 113 acquires the data supplied from the plaque decomposition unit 111 (arrow 123). The auxiliary patch information compressing unit 113 encodes (compresses) the auxiliary patch information included in the acquired data. The auxiliary patch information compressing unit 113 supplies the encoded data of the obtained auxiliary patch information to the multiplexer 117 (arrow 126).

視訊編碼部114,係進行位置資訊(Geometry)之視訊畫格之編碼的相關處理。例如,視訊編碼部114,係將從打包部112所被供給之位置資訊(Geometry)之視訊畫格,加以取得(箭頭124)。又,視訊編碼部114,係將該已取得之位置資訊(Geometry)之視訊畫格,例如以AVC或HEVC等之任意的2維影像用之編碼方法,進行編碼。視訊編碼部114,係將該編碼所得之編碼資料(位置資訊(Geometry)之視訊畫格之編碼資料),供給至多工器117(箭頭127)。The video encoding unit 114 performs a process of encoding the video frame of the position information (Geometry). For example, the video encoding unit 114 acquires the video frame of the position information (Geometry) supplied from the packing unit 112 (arrow 124). Further, the video encoding unit 114 encodes the video frame of the acquired position information (Geometry), for example, by an encoding method for any two-dimensional video such as AVC or HEVC. The video encoding unit 114 supplies the encoded encoded data (coded data of the video frame of the position information (Geometry)) to the multiplexer 117 (arrow 127).

視訊編碼部115,係進行屬性資訊(Texture)之視訊畫格之編碼的相關處理。例如,視訊編碼部115,係將從打包部112所被供給之屬性資訊(Texture)之視訊畫格,加以取得(箭頭124)。又,視訊編碼部115,係將該已取得之屬性資訊(Texture)之視訊畫格,例如以AVC或HEVC等之任意的2維影像用之編碼方法,進行編碼。視訊編碼部115,係將該編碼所得之編碼資料(屬性資訊(Texture)之視訊畫格之編碼資料),供給至多工器117(箭頭128)。The video encoding unit 115 performs processing related to encoding of the video frame of the attribute information (Texture). For example, the video encoding unit 115 acquires the video frame of the attribute information (Texture) supplied from the packing unit 112 (arrow 124). Further, the video encoding unit 115 encodes the acquired video frame of the attribute information (Texture), for example, by an encoding method for any two-dimensional video such as AVC or HEVC. The video encoding unit 115 supplies the encoded encoded data (encoded data of the video frame of the attribute information) to the multiplexer 117 (arrow 128).

OMap編碼部116,係進行佔用地圖之視訊畫格之編碼的相關處理。例如,OMap編碼部116,係將從打包部112所被供給之佔用地圖之視訊畫格,加以取得(箭頭124)。又,OMap編碼部116,係將該已取得之佔用地圖之視訊畫格,例如以AVC或HEVC等之任意的2維影像用之編碼方法,進行編碼。OMap編碼部116,係將該編碼所得之編碼資料(佔用地圖之視訊畫格之編碼資料),供給至多工器117(箭頭129)。The OMap encoding unit 116 performs a process of encoding the video frame occupying the map. For example, the OMap encoding unit 116 acquires the video frame of the occupied map supplied from the packing unit 112 (arrow 124). Further, the OMap encoding unit 116 encodes the acquired video frame of the occupied map, for example, by an encoding method for any two-dimensional video such as AVC or HEVC. The OMap encoding unit 116 supplies the coded encoded data (encoded data of the video frame occupying the map) to the multiplexer 117 (arrow 129).

多工器117,係進行多工化的相關處理。例如,多工器117,係將從輔助斑塊資訊壓縮部113所被供給之輔助斑塊資訊之編碼資料,加以取得(箭頭126)。又,例如,多工器117,係將從打包部112所被供給之打包的相關之控制資訊,加以取得(箭頭125)。又,例如,多工器117,係將從視訊編碼部114所被供給之位置資訊(Geometry)之視訊畫格之編碼資料,加以取得(箭頭127)。又,例如,多工器117,係將從視訊編碼部115所被供給之屬性資訊(Texture)之視訊畫格之編碼資料,加以取得(箭頭128)。又,例如,多工器117,係將從OMap編碼部116所被供給之佔用地圖之視訊畫格之編碼資料,加以取得(箭頭129)。The multiplexer 117 is a related process of multiplexing. For example, the multiplexer 117 acquires the encoded data of the auxiliary patch information supplied from the auxiliary patch information compressing unit 113 (arrow 126). Further, for example, the multiplexer 117 acquires control information related to the package supplied from the packing unit 112 (arrow 125). Further, for example, the multiplexer 117 acquires the coded data of the video frame of the position information (Geometry) supplied from the video encoding unit 114 (arrow 127). Further, for example, the multiplexer 117 acquires the coded data of the video frame of the attribute information (Texture) supplied from the video encoding unit 115 (arrow 128). Further, for example, the multiplexer 117 acquires the coded data of the video frame of the occupied map supplied from the OMap encoding unit 116 (arrow 129).

多工器117,係將已取得之這些資訊進行多工化,而生成位元串流。多工器117,係將該已生成之位元串流,輸出至編碼裝置100之外部(箭頭130)。The multiplexer 117 multiplexes the obtained information to generate a bit stream. The multiplexer 117 streams the generated bit stream to the outside of the encoding device 100 (arrow 130).

<斑塊分解部>
圖5係斑塊分解部111的主要構成例的區塊圖。如圖5所示,此時的斑塊分解部111係具有:法線方向推定部151、分段初期設定部152、分段更新部153、2維投影部154、及像素分布解析部155。
<plaque decomposition department>
FIG. 5 is a block diagram of a main configuration example of the plaque decomposition unit 111. As shown in FIG. 5, the plaque decomposition unit 111 at this time includes a normal direction estimation unit 151, a segment initial setting unit 152, a segment update unit 153, a two-dimensional projection unit 154, and a pixel distribution analysis unit 155.

法線方向推定部151,係進行3D資料之表面的法線方向之推定的相關處理。例如,法線方向推定部151,係將所被輸入的3D資料,加以取得。又,法線方向推定部151,係推定該已取得之3D資料所呈現的物件之表面的法線方向。例如,法線方向推定部151,係建構kd-tree,並探索附近,算出最適近似接平面等而推定法線方向。法線方向推定部151,係將該法線方向之推定結果連同其他資料,供給至分段初期設定部152。The normal direction estimating unit 151 is a correlation process for estimating the normal direction of the surface of the 3D material. For example, the normal direction estimating unit 151 acquires the input 3D material. Further, the normal direction estimating unit 151 estimates the normal direction of the surface of the object presented by the acquired 3D material. For example, the normal direction estimating unit 151 constructs a kd-tree, searches for a nearby area, calculates an optimum approximation plane, and estimates a normal direction. The normal direction estimating unit 151 supplies the estimated result of the normal direction together with other data to the segment initial setting unit 152.

分段初期設定部152,係進行分段之初期設定的相關處理。例如,分段初期設定部152,係將從法線方向推定部151所被供給之資料,加以取得。又,例如,分段初期設定部152,係基於已被法線方向推定部151所推定出來的法線方向的6軸的各方向之成分,而將3D資料的其法線方向所對應的面,加以分類。分段初期設定部152,係將該分類結果,連同其他資料一起供給至分段更新部153。The segment initial setting unit 152 performs correlation processing for initial setting of the segment. For example, the segment initial setting unit 152 acquires the data supplied from the normal direction estimating unit 151. In addition, for example, the segment initial setting unit 152 sets the surface corresponding to the normal direction of the 3D material based on the components of the six directions in the normal direction estimated by the normal direction estimating unit 151. , to classify. The segment initial setting unit 152 supplies the classification result to the segment update unit 153 together with other materials.

分段更新部153,係進行分段之更新的相關處理。例如,分段更新部153,係將從分段初期設定部152所被供給之資料,加以取得。然後,分段更新部153,係已被分段初期設定部152所設定之初期設定之分段中的太小的領域予以集結,使其變成足夠大的領域。分段更新部153,係將已更新之分段的相關之資訊,連同其他資訊,一起供給至2維投影部154。The segment update unit 153 performs correlation processing of segment update. For example, the segment update unit 153 acquires the data supplied from the segment initial setting unit 152. Then, the segment update unit 153 is a field that is too small among the segments of the initial settings set by the segment initial setting unit 152, so that it becomes a sufficiently large field. The segment update unit 153 supplies the related information of the updated segment to the two-dimensional projection unit 154 together with other information.

2維投影部154,係進行3D資料之2維投影的相關處理。例如,2維投影部154,係將從分段更新部153所被供給之資料,加以取得。又,2維投影部154,係利用像素分布解析部155,而生成已取得之資料中所含之3D資料之佔用地圖,或是將該3D資料或佔用資料,每一領域地投影至2維平面等等。2維投影部154,係將已投影至2維平面的3D資料,連同其他資料,一起供給至打包部112。The two-dimensional projection unit 154 performs correlation processing of two-dimensional projection of 3D data. For example, the two-dimensional projection unit 154 acquires the data supplied from the segment update unit 153. Further, the two-dimensional projection unit 154 uses the pixel distribution analysis unit 155 to generate an occupied map of the 3D data included in the acquired data, or project the 3D data or the occupied data to each of the fields to the 2D. Plane and so on. The two-dimensional projection unit 154 supplies the 3D data that has been projected onto the two-dimensional plane to the packing unit 112 together with other materials.

像素分布解析部155,係進行2維投影部154之處理對象的3D資料的像素分布之解析的相關處理。The pixel distribution analysis unit 155 performs correlation processing for analyzing the pixel distribution of the 3D material to be processed by the two-dimensional projection unit 154.

<解碼裝置>
圖6係適用了本技術的影像處理裝置之一態樣的解碼裝置之構成之一例的區塊圖。圖6所示的解碼裝置200係為,將像是點雲之類的3D資料被投影至2維平面而被編碼成的編碼資料,以2維影像用之解碼方法進行解碼,並投影至3維空間的裝置。
<decoding device>
Fig. 6 is a block diagram showing an example of a configuration of a decoding device to which one aspect of the image processing device of the present technology is applied. The decoding device 200 shown in FIG. 6 is a coded data obtained by projecting 3D data such as a point cloud onto a two-dimensional plane, and decoding it by a decoding method for two-dimensional images, and projecting it to 3 A device for dimensional space.

例如,解碼裝置200,係實作非專利文獻1乃至非專利文獻6中所記載之技術,以符合這些文獻之任一者所記載之規格的方法,進行3D資料的編碼資料之解碼。For example, the decoding device 200 implements the techniques described in Non-Patent Document 1 or Non-Patent Document 6, and decodes the encoded data of the 3D data in a method conforming to the specifications described in any of these documents.

此外,於圖6中係主要表示處理部或資料的流動等,圖6所示者並非全部。亦即,於解碼裝置200中,亦可存在有於圖6中未以區塊而被表示的處理部,亦可存在有於圖6中未以箭頭等而表示的處理或資料之流動。這在解碼裝置200內的處理部等的說明的其他圖中也是同樣如此。In addition, in FIG. 6, the flow of a processing part or a material is mainly shown, and it is not shown in FIG. That is, in the decoding device 200, there may be a processing unit that is not represented by a block in FIG. 6, and there may be a flow of processing or data that is not indicated by an arrow or the like in FIG. This is also the same in other figures of the description of the processing unit and the like in the decoding device 200.

如圖6所示,解碼裝置200係具有:解多工器211、輔助斑塊資訊解碼部212、視訊解碼部213、視訊解碼部214、OMap解碼部215、解包部216、及3D重新建構部217。As shown in FIG. 6, the decoding apparatus 200 includes a demultiplexer 211, an auxiliary patch information decoding unit 212, a video decoding unit 213, a video decoding unit 214, an OMap decoding unit 215, an unpacking unit 216, and a 3D reconstruction. Part 217.

解多工器211,係進行資料之逆多工化的相關處理。例如,解多工器211,係將被輸入至解碼裝置200的位元串流,加以取得。該位元串流係例如,是由編碼裝置100所供給。解多工器211,係將該位元串流進行逆多工化,將輔助斑塊資訊之編碼資料予以抽出,將其供給至輔助斑塊資訊解碼部212。又,解多工器211,係藉由逆多工化,而從位元串流,抽出位置資訊(Geometory)之視訊畫格之編碼資料,將其供給至視訊解碼部213。再者,解多工器211,係藉由逆多工化,而從位元串流,抽出屬性資訊(Texture)之視訊畫格之編碼資料,將其供給至視訊解碼部214。又,解多工器211,係藉由逆多工化,而從位元串流,抽出佔用地圖之視訊畫格之編碼資料,將其供給至OMap解碼部215。The multiplexer 211 performs correlation processing of inverse multiplexing of data. For example, the demultiplexer 211 is obtained by acquiring a bit stream input to the decoding device 200. This bit stream is supplied, for example, by the encoding device 100. The demultiplexer 211 performs inverse multiplexing on the bit stream, extracts the coded data of the auxiliary patch information, and supplies it to the auxiliary patch information decoding unit 212. Further, the demultiplexer 211 extracts the encoded data of the video frame of the position information (Geometory) from the bit stream by inverse multiplexing, and supplies it to the video decoding unit 213. Further, the demultiplexer 211 extracts the encoded data of the video frame of the attribute information from the bit stream by inverse multiplexing, and supplies it to the video decoding unit 214. Further, the demultiplexer 211 extracts the encoded data of the video frame occupying the map from the bit stream by inverse multiplexing, and supplies it to the OMap decoding unit 215.

輔助斑塊資訊解碼部212,係進行輔助斑塊資訊之編碼資料之解碼的相關處理。例如,輔助斑塊資訊解碼部212,係將從解多工器211所被供給之輔助斑塊資訊之編碼資料,加以取得。又,輔助斑塊資訊解碼部212,係將該已取得之資料中所含之輔助斑塊資訊之編碼資料,予以解碼。輔助斑塊資訊解碼部212,係將該解碼所得之輔助斑塊資訊,供給至3D重新建構部217。The auxiliary patch information decoding unit 212 performs correlation processing for decoding the encoded data of the auxiliary patch information. For example, the auxiliary patch information decoding unit 212 acquires the encoded data of the auxiliary patch information supplied from the demultiplexer 211. Further, the auxiliary patch information decoding unit 212 decodes the encoded data of the auxiliary patch information included in the acquired data. The auxiliary patch information decoding unit 212 supplies the decoded auxiliary patch information to the 3D reconstruction unit 217.

視訊解碼部213,係進行位置資訊(Geometory)之視訊畫格之編碼資料之解碼的相關處理。例如,視訊解碼部213,係將從解多工器211所被供給之位置資訊(Geometory)之視訊畫格之編碼資料,加以取得。又,例如,視訊解碼部213,係將該已取得之編碼資料予以解碼,獲得位置資訊(Geometory)之視訊畫格。視訊解碼部213,係將該位置資訊(Geometory)之視訊畫格,供給至解包部216。The video decoding unit 213 performs a process of decoding the encoded data of the video frame of the position information (Geometory). For example, the video decoding unit 213 acquires the encoded data of the video frame of the location information (Geometory) supplied from the demultiplexer 211. Further, for example, the video decoding unit 213 decodes the acquired encoded data to obtain a video frame of the location information (Geometory). The video decoding unit 213 supplies the video frame of the position information (Geometory) to the unpacking unit 216.

視訊解碼部214,係進行屬性資訊(Texture)之視訊畫格之編碼資料之解碼的相關處理。例如,視訊解碼部214,係將從解多工器211所被供給之屬性資訊(Texture)之視訊畫格之編碼資料,加以取得。又,例如,視訊解碼部214,係將該已取得之編碼資料予以解碼,獲得屬性資訊(Texture)之視訊畫格。視訊解碼部214,係將該屬性資訊(Texture)之視訊畫格,供給至解包部216。The video decoding unit 214 performs a process of decoding the encoded data of the video frame of the attribute information (Texture). For example, the video decoding unit 214 acquires the encoded data of the video frame of the attribute information (Texture) supplied from the demultiplexer 211. Further, for example, the video decoding unit 214 decodes the acquired encoded data to obtain a video frame of the attribute information (Texture). The video decoding unit 214 supplies the video frame of the attribute information to the unpacking unit 216.

OMap解碼部215,係進行佔用地圖之視訊畫格之編碼資料之解碼的相關處理。例如,OMap解碼部215,係將從解多工器211所被供給之佔用地圖之視訊畫格之編碼資料,加以取得。又,例如,OMap解碼部215,係將該已取得之編碼資料予以解碼,獲得佔用地圖之視訊畫格。OMap解碼部215,係將該佔用地圖之視訊畫格,供給至解包部216。The OMap decoding unit 215 performs a process of decoding the encoded data of the video frame occupying the map. For example, the OMap decoding unit 215 acquires the encoded data of the video frame of the occupied map supplied from the demultiplexer 211. Further, for example, the OMap decoding unit 215 decodes the acquired encoded data to obtain a video frame occupying the map. The OMap decoding unit 215 supplies the video frame of the occupied map to the unpacking unit 216.

解包部216,係進行解包的相關處理。例如,解包部216,係從視訊解碼部213取得位置資訊(Geometory)之視訊畫格,從視訊解碼部214取得屬性資訊(Texture)之視訊畫格,從OMap解碼部215取得佔用地圖之視訊畫格。解包部216,係將這些視訊畫格予以解包。解包部216,係將解包所得到的位置資訊(Geometory)之資料、屬性資訊(Texture)之資料、及佔用地圖之資料,供給至3D重新建構部217。The unpacking unit 216 performs related processing of unpacking. For example, the unpacking unit 216 acquires the video frame of the location information (Geometory) from the video decoding unit 213, acquires the video frame of the attribute information (Texture) from the video decoding unit 214, and acquires the video of the occupied map from the OMap decoding unit 215. Frame. The unpacking unit 216 unpacks these video frames. The unpacking unit 216 supplies the data of the location information (Geometory) obtained by the unpacking, the information of the attribute information (Texture), and the data of the occupied map to the 3D reconstruction unit 217.

3D重新建構部217,係進行3D資料之重新建構的相關處理。例如,3D重新建構部217,係基於從輔助斑塊資訊解碼部212所被供給之輔助斑塊資訊、或從解包部216所被供給之位置資訊(Geometory)之資料、屬性資訊(Texture)之資料、及佔用地圖之資料等,來重新建構3D資料。3D重新建構部217,係將如此處理所得之3D資料,輸出至解碼裝置200之外部。The 3D reconstruction unit 217 performs processing related to reconstruction of 3D data. For example, the 3D reconstruction unit 217 is based on the auxiliary patch information supplied from the auxiliary patch information decoding unit 212 or the location information (Geometory) supplied from the unpacking unit 216, and attribute information (Texture). The data and the information on the map are used to reconstruct the 3D data. The 3D reconstruction unit 217 outputs the 3D data thus processed to the outside of the decoding device 200.

該3D資料係例如,被供給至顯示部而使其影像被顯示、或被記錄至記錄媒體、或透過通訊而被供給至其他裝置。The 3D data is supplied to the display unit, for example, and the video is displayed, recorded on a recording medium, or supplied to another device via communication.

<3D重新建構部>
圖7係圖6的3D重新建構部217的主要構成例的區塊圖。如圖7所示,3D重新建構部217係具有:3維投影部251、像素分布解析部252、逆分段更新部253、逆分段初期設定部254、及逆法線方向推定部255。
<3D Reconstruction Department>
Fig. 7 is a block diagram showing a main configuration example of the 3D reconstruction unit 217 of Fig. 6. As shown in FIG. 7, the 3D reconstruction unit 217 includes a three-dimensional projection unit 251, a pixel distribution analysis unit 252, an inverse segment update unit 253, an inverse segment initial setting unit 254, and an inverse normal direction estimation unit 255.

3維投影部251,係進行每一領域地被投影至2維平面上的3D資料往3維空間之投影。像素分布解析部252,係在3維投影部251所致之往3維空間之投影之際,進行像素分布之解析等之處理。The three-dimensional projection unit 251 performs projection of 3D data projected onto a two-dimensional plane into a three-dimensional space in each field. The pixel distribution analysis unit 252 performs processing such as analysis of pixel distribution when the projection of the three-dimensional space by the three-dimensional projection unit 251 is performed.

逆分段更新部253,係進行分段更新部153的逆處理。逆分段初期設定部254,係進行分段初期設定部152的逆處理。逆法線方向推定部255,係進行法線方向推定部151的逆處理。The inverse segment update unit 253 performs inverse processing of the segment update unit 153. The reverse segment initial setting unit 254 performs the inverse processing of the segment initial setting unit 152. The inverse normal direction estimating unit 255 performs the inverse processing of the normal direction estimating unit 151.

<2.第1實施形態>
<投影3D資料的2維平面之圖層數可變>
在先前的方法中,3D資料,係如圖8所示的例子,是被投影至2個圖層(圖層0及圖層1)之2維平面。
<2. First embodiment>
<The number of layers of the 2-dimensional plane of the projected 3D data is variable>
In the previous method, the 3D data, as shown in the example of FIG. 8, is a 2-dimensional plane projected onto two layers (layer 0 and layer 1).

在圖層0(Layer0),係有3D資料的從投影面來看而為表面的點的資料,會被投影。在圖層1(Layer1),係有從圖層0起算在所定之閾值(Default Th = 4Voxel)以內而距離最遠的點的資料,會被投影。遠離超過該所定之閾值Th者,係被丟棄(Drop)。此外,於圖層1中,從圖層0起算之距離之差分值係為像素值。In layer 0 (Layer 0), the data of the point of the 3D data viewed from the projection surface as the surface is projected. In Layer 1 (Layer 1), data from the point 0 at the threshold (Default Th = 4Voxel) and the farthest distance from the layer 0 is projected. If it is far away from the threshold value Th, it is discarded. Further, in the layer 1, the difference value of the distance from the layer 0 is a pixel value.

因為是如此只藉由2張固定之圖層來表現物件表面的演算法,所以會喪失其他點的資訊(無法重現)。因此,會發生無法投影至2維平面的點,會因為往2維平面之投影所伴隨的編碼,而有可能導致資料品質降低。Because the algorithm of the surface of the object is represented by only two fixed layers, the information of other points is lost (cannot be reproduced). Therefore, a point that cannot be projected onto a two-dimensional plane occurs, which may result in degradation of data quality due to coding accompanying projection onto a two-dimensional plane.

於是,亦可將投影3D資料的2維平面之圖層數,設成可變。例如,使表示3維結構之3D資料中所含之所有的每一位置之資料,投影至複數圖層之2維平面。例如,於影像處理裝置中,使其具備:2維投影部,係將表示3維結構之3D資料中所含之所有的每一位置之資料,投影至複數圖層之2維平面。Therefore, the number of layers of the two-dimensional plane of the projected 3D data can be set to be variable. For example, data representing each of the positions included in the 3D material of the 3-dimensional structure is projected onto a 2-dimensional plane of the complex layer. For example, the image processing apparatus includes a two-dimensional projection unit that projects data of each of the positions included in the 3D data of the three-dimensional structure onto a two-dimensional plane of the complex layer.

例如圖9的情況下,設定圖層0(Layer0)乃至圖層3(Layer3)之2維平面(比2圖層還多圖層之2維平面),對這些圖層投影3D資料。藉由如此設計,就可較正確地重現在物件表面具有厚度的點雲。亦即,可抑制3D資料的2維投影所致之品質的降低。For example, in the case of FIG. 9, a two-dimensional plane of layer 0 (Layer 0) or even layer 3 (Layer 3) (a two-dimensional plane having more layers than two layers) is set, and 3D data is projected on these layers. By designing in this way, it is possible to more accurately reproduce the point cloud having a thickness on the surface of the object. That is, it is possible to suppress a decrease in quality due to 2-dimensional projection of 3D data.

例如,2維投影部154係亦可將3D資料的從投影面來看位置是在縱深方向上重疊的每一位置之資料,投影至複數圖層之2維平面的彼此互異之圖層。For example, the two-dimensional projection unit 154 may project data of each position in which the position of the 3D material is superimposed in the depth direction from the projection surface, and project to a mutually different layer of the two-dimensional plane of the plurality of layers.

圖9之例子的情況下,縱深方向上重疊的資料,係被投影至圖層0乃至圖層3之中的彼此互異之圖層。藉由如此設計,可將3D資料的,從投影面來看位置是在縱深方向上重疊的每一位置之資料,全部投影至2維平面。亦即,可抑制資訊的損失。因此,可抑制3D資料的2維投影所致之品質的降低。In the case of the example of Fig. 9, the data superimposed in the depth direction is projected onto the layers different from each other in the layer 0 or the layer 3. With such a design, the position of the 3D material at each position in the depth direction as viewed from the projection surface can be projected onto the 2-dimensional plane. That is, the loss of information can be suppressed. Therefore, it is possible to suppress a decrease in quality due to 2-dimensional projection of 3D data.

又,例如,2維投影部154,係亦可針對該2維平面,生成與3D資料的從投影面來看位置是在縱深方向上重疊的每一位置之資料之最大筆數相同數量之圖層。Further, for example, the two-dimensional projection unit 154 can generate, for the two-dimensional plane, the same number of layers as the maximum number of pieces of data of each position where the position of the 3D material is superimposed in the depth direction as viewed from the projection surface. .

圖9之例子的情況下,該當領域(Local Bounding Box)之3D資料中,縱深方向上重疊的資料之最大筆數係為4。因此,該3D資料,係被投影至4圖層(圖層0乃至圖層3)之2維平面。In the case of the example of Fig. 9, in the 3D data of the Local Bounding Box, the maximum number of data superimposed in the depth direction is 4. Therefore, the 3D data is projected onto a 2-dimensional plane of 4 layers (layer 0 or layer 3).

藉由如此設計,就可將3D資料之所有的每一位置之資料,投影至2維平面。因此,可抑制資訊的損失,所亦可抑制3D資料的2維投影所致之品質的降低。With this design, the data of each position of the 3D data can be projected onto the 2-dimensional plane. Therefore, the loss of information can be suppressed, and the quality degradation caused by the two-dimensional projection of the 3D data can be suppressed.

此外,此情況下,亦可將表示被2維投影部154投影了3D資料的2維平面之圖層數的資訊,在位元串流中進行傳訊。亦即,作為位元串流生成部而發揮機能的多工器117,係生成含有表示被2維投影部154投影了3D資料的2維平面之圖層數的資訊、和藉由視訊編碼部114等將2維平面予以編碼所得之編碼資料的位元串流。Further, in this case, information indicating the number of layers of the two-dimensional plane on which the 3D data is projected by the two-dimensional projection unit 154 may be transmitted in the bit stream. In other words, the multiplexer 117 functioning as the bit stream generating unit generates information including the number of layers indicating the two-dimensional plane on which the 3D data is projected by the two-dimensional projection unit 154, and the video encoding unit. A bit stream of encoded data obtained by encoding a 2-dimensional plane, such as 114.

藉由如此設計,於解碼側可藉由參照該表示2維平面之圖層數的資訊,就可容易地將已被投影至2維平面之所有圖層的3D資料,投影至3維空間。With such a design, the 3D data of all the layers projected onto the two-dimensional plane can be easily projected onto the three-dimensional space by referring to the information indicating the number of layers of the two-dimensional plane on the decoding side.

<編碼處理之流程>
被編碼裝置100所執行的編碼處理之流程之例子,參照圖10的流程圖來加以說明。
<Process of Encoding Process>
An example of the flow of the encoding process executed by the encoding device 100 will be described with reference to the flowchart of FIG.

一旦編碼處理被開始,則編碼裝置100的斑塊分解部111,係於步驟S101中,將3D資料分解成斑塊,將各斑塊之資料投影至2維平面。於步驟S102中,輔助斑塊資訊壓縮部113,係將步驟S101之處理所得到之輔助斑塊資訊,予以壓縮。Once the encoding process is started, the plaque decomposition unit 111 of the encoding device 100 decomposes the 3D data into plaques and projects the data of each plaque onto the two-dimensional plane in step S101. In step S102, the auxiliary patch information compressing unit 113 compresses the auxiliary patch information obtained by the processing of step S101.

於步驟S103中,打包部112,係將藉由斑塊分解部111而每一斑塊地被投影至2維平面上的3D資料,打包成為視訊畫格。於步驟S104中,視訊編碼部114,係將步驟S103之處理所得到之位置資訊之視訊畫格亦即幾何視訊畫格,以2維影像用之編碼方法進行編碼。In step S103, the packing unit 112 packs the 3D material projected onto the two-dimensional plane by the plaque decomposition unit 111, and packs it into a video frame. In step S104, the video encoding unit 114 encodes the video frame of the position information obtained by the processing of step S103, that is, the geometric video frame, by the encoding method for the two-dimensional video.

於步驟S105中,視訊編碼部114,係將步驟S103之處理所得到之屬性資訊之視訊畫格亦即彩色視訊畫格,以2維影像用之編碼方法進行編碼。於步驟S106中,視訊編碼部114,係將步驟S103之處理所得到之佔用地圖,以2維影像用之編碼方法進行編碼。In step S105, the video encoding unit 114 encodes the video frame of the attribute information obtained by the processing of step S103, that is, the color video frame, by the encoding method for the two-dimensional image. In step S106, the video encoding unit 114 encodes the occupied map obtained by the processing of step S103 by the encoding method for the two-dimensional video.

於步驟S107中,多工器117,係將如以上而被生成之各種資訊進行多工化,生成含有這些資訊的位元串流。In step S107, the multiplexer 117 multiplexes various pieces of information generated as described above to generate a bit stream containing the information.

於步驟S108中,多工器117,係將步驟S107之處理所生成之位元串流,輸出至編碼裝置100的外部。In step S108, the multiplexer 117 outputs the bit stream generated by the processing of step S107 to the outside of the encoding device 100.

一旦步驟S108的處理結束,則編碼處理就結束。Once the processing of step S108 ends, the encoding process ends.

<斑塊分解處理之流程>
其次,參照圖11的流程圖,說明圖10之步驟S101中所執行的斑塊分解處理之流程例。
<Flow of plaque decomposition processing>
Next, an example of the flow of the plaque decomposition processing executed in step S101 of Fig. 10 will be described with reference to the flowchart of Fig. 11 .

一旦斑塊分解處理被開始,則法線方向推定部151,係於步驟S121中,推定法線方向。於步驟S122中,分段初期設定部152,係進行分段之初期設定。於步驟S123中,分段更新部153,係將步驟S122中所被設定之初期狀態之分段,因應需要而加以更新。於步驟S124中,2維投影部154,係將3D資料投影至2維平面。When the plaque decomposition processing is started, the normal direction estimating unit 151 estimates the normal direction in step S121. In step S122, the segment initial setting unit 152 performs initial setting of the segmentation. In step S123, the segment update unit 153 updates the segment of the initial state set in step S122 as necessary. In step S124, the two-dimensional projection unit 154 projects the 3D material onto the two-dimensional plane.

一旦步驟S124的處理結束,則斑塊分解處理就結束,處理係回到圖10。Once the processing of step S124 ends, the plaque decomposition processing ends, and the processing returns to FIG.

<2維投影處理之流程>
其次,參照圖12的流程圖,說明圖11之步驟S124中所執行的2維投影處理之流程例。
<Process of 2-dimensional projection processing>
Next, an example of the flow of the two-dimensional projection processing executed in step S124 of Fig. 11 will be described with reference to the flowchart of Fig. 12 .

一旦2維投影處理被開始,則2維投影部154,係於步驟S141中,進行分段所致之領域抽出。於步驟S142中,2維投影部154,係進行初期化以使圖層(Layer)號碼i=0。When the two-dimensional projection processing is started, the two-dimensional projection unit 154 is in step S141, and performs field extraction by segmentation. In step S142, the two-dimensional projection unit 154 is initialized so that the layer number i=0.

於步驟S143中,像素分布解析部155係判定,是否有未投影像素(未被投影至2維平面的3D資料的每一位置之資料)存在。若判定為有未投影像素存在,則處理係前進至步驟S144。In step S143, the pixel distribution analysis unit 155 determines whether or not there is an unprojected pixel (data that is not projected to each position of the 3D material of the two-dimensional plane). If it is determined that there is an unprojected pixel, the processing proceeds to step S144.

於步驟S144中,2維投影部154係將領域投影至處理對象之圖層i(Layer i)。於步驟S145中,2維投影部154,係將變數i予以增值(i++)。一旦步驟S145之處理結束,則處理係回到步驟S143,重複其以後之處理。In step S144, the two-dimensional projection unit 154 projects the domain to the layer i (Layer i) of the processing target. In step S145, the two-dimensional projection unit 154 adds the variable i (i++). Once the process of step S145 is ended, the process returns to step S143, and the subsequent processes are repeated.

於步驟S143中,若判定為未投影像素不存在(領域內之所有的像素都已被投影)的情況,則處理係前進至步驟S146。In step S143, if it is determined that the unprojected pixel does not exist (all the pixels in the field have been projected), the processing proceeds to step S146.

於步驟S146中,2維投影部154係將表示圖層枚數i的資訊予以供給,令其被編碼。又,於步驟S147中,2維投影部154係令對應之i畫格份的幾何影像被編碼。亦即,2維投影部154,係將已投影有3D資料的2維平面供給至打包部112,令該各圖層被打包成為彼此互異之畫格。In step S146, the two-dimensional projection unit 154 supplies information indicating the number of layers i, and encodes it. Further, in step S147, the two-dimensional projection unit 154 causes the geometric image of the corresponding i-frame portion to be encoded. That is, the two-dimensional projection unit 154 supplies a two-dimensional plane on which 3D data has been projected to the packing unit 112, and the layers are packaged into mutually different frames.

一旦步驟S147的處理結束,則2維投影處理就結束,處理係回到圖11。Once the process of step S147 is completed, the two-dimensional projection process ends, and the process returns to FIG.

藉由如以上般地執行各處理,藉由如此設計,就可較正確地重現在物件表面具有厚度的點雲。亦即,可抑制3D資料的2維投影所致之品質的降低。By performing the respective processes as above, by designing such, it is possible to more accurately reproduce the point cloud having a thickness on the surface of the object. That is, it is possible to suppress a decrease in quality due to 2-dimensional projection of 3D data.

<已被投影至圖層數可變的2維平面上的3D資料之重新建構>
於解碼側,藉由使用由編碼側所提供的表示2維平面之圖層數的資訊,就可如上述般地實現已被投影至圖層數為可變之2維平面上的3D資料之重新建構。
<Reconstruction of 3D data that has been projected onto a 2-dimensional plane with variable number of layers>
On the decoding side, by using the information indicating the number of layers of the two-dimensional plane provided by the encoding side, the 3D data that has been projected onto the two-dimensional plane whose number of layers is variable can be realized as described above. Reconstructed.

亦即,將已被投影至圖層數資訊所表示之圖層數之2維平面上的3D資料的所有的每一位置之資料,投影至3維空間。例如,於影像處理裝置中,使其具備3維投影部,係將已被投影至圖層數資訊所表示之圖層數之2維平面上的3D資料的所有的每一位置之資料,投影至3維空間。That is, the data of each position of the 3D material on the two-dimensional plane that has been projected onto the number of layers indicated by the layer number information is projected to the three-dimensional space. For example, in the image processing device, a three-dimensional projection unit is provided, and data of each position of the 3D material on the two-dimensional plane that has been projected onto the number of layers indicated by the layer number information is projected. To 3D space.

藉由如此設計,可實現已被投影至圖層數為可變之2維平面上的3D資料之重新建構。亦即,可較正確地重現在物件表面具有厚度的點雲。亦即,可抑制3D資料的2維投影所致之品質的降低。With such a design, reconstruction of 3D data that has been projected onto a two-dimensional plane whose number of layers is variable can be realized. That is, the point cloud having a thickness on the surface of the object can be reproduced more correctly. That is, it is possible to suppress a decrease in quality due to 2-dimensional projection of 3D data.

<解碼處理之流程>
被解碼裝置200所執行的解碼處理之流程之例子,參照圖13的流程圖來加以說明。
<Process of decoding processing>
An example of the flow of the decoding process executed by the decoding device 200 will be described with reference to the flowchart of FIG.

一旦解碼處理被開始,則解碼裝置200的解多工器211係於步驟S201中,將位元串流進行逆多工化。Once the decoding process is started, the demultiplexer 211 of the decoding device 200 is in step S201, and the bit stream is inversely multiplexed.

於步驟S202中,輔助斑塊資訊解碼部212,係將藉由步驟S201之處理而從位元串流所被抽出的輔助斑塊資訊,予以解碼。於步驟S203中,視訊解碼部213,係將藉由步驟S201之處理而從位元串流所被抽出的幾何視訊畫格(位置資訊之視訊畫格)之編碼資料,予以解碼。In step S202, the auxiliary patch information decoding unit 212 decodes the auxiliary patch information extracted from the bit stream by the processing of step S201. In step S203, the video decoding unit 213 decodes the encoded data of the geometric video frame (the video frame of the position information) extracted from the bit stream by the processing of step S201.

於步驟S204中,視訊解碼部214,係將藉由步驟S201之處理而從位元串流所被抽出的彩色視訊畫格(屬性資訊之視訊畫格)之編碼資料,予以解碼。於步驟S205中,OMap解碼部215,係將藉由步驟S201之處理而從位元串流所被抽出的佔用地圖之編碼資料,予以解碼。In step S204, the video decoding unit 214 decodes the coded data of the color video frame (the video frame of the attribute information) extracted from the bit stream by the process of step S201. In step S205, the OMap decoding unit 215 decodes the encoded data of the occupied map extracted from the bit stream by the processing of step S201.

於步驟S206中,解包部216,係將步驟S203乃至步驟S205中所被解碼的幾何視訊畫格、彩色視訊畫格、及佔用地圖,分別予以解包。In step S206, the unpacking unit 216 unpacks the geometric video frame, the color video frame, and the occupied map decoded in step S203 to step S205, respectively.

於步驟S207中,3D重新建構部217,係基於步驟S202中所得到之輔助斑塊資訊、和步驟S206中所得到之各種資訊,而將例如點雲等之3D資料予以重新建構。In step S207, the 3D reconstruction unit 217 reconstructs the 3D material such as the point cloud based on the auxiliary patch information obtained in step S202 and various pieces of information obtained in step S206.

一旦步驟S207之處理結束則解碼處理就結束。Once the processing of step S207 ends, the decoding process ends.

<點雲重新建構處理之流程>
其次,參照圖14的流程圖,說明圖13之步驟S207中所執行的點雲重新建構處理之流程例。
<Process of point cloud re-construction processing>
Next, an example of the flow of the point cloud reconstruction process executed in step S207 of Fig. 13 will be described with reference to the flowchart of Fig. 14.

一旦點雲重新建構處理被開始,則3維投影部251係於步驟S221中,將2維影像投影至3維空間。When the point cloud reconstruction process is started, the three-dimensional projection unit 251 is in step S221 to project the two-dimensional image into the three-dimensional space.

於步驟S222中,逆分段更新部253,係將分段做逆向地更新,將已被集結的分段予以劃分。In step S222, the inverse segment update unit 253 updates the segments in the reverse direction and divides the segments that have been assembled.

於步驟S223中,逆分段初期設定部254,係進行分段之初期設定的逆處理,將已被分類的點予以集結。In step S223, the inverse segmentation initial setting unit 254 performs inverse processing of the initial setting of the segmentation, and aggregates the classified points.

於步驟S224中,逆法線方向推定部255,係進行法線方向推定的逆處理,將點雲予以重新建構。In step S224, the inverse normal direction estimating unit 255 performs inverse processing of the normal direction estimation to reconstruct the point cloud.

一旦步驟S224的處理結束,則點雲重新建構處理就結束,處理係回到圖13。Once the process of step S224 ends, the point cloud reconstruction process ends and the process returns to FIG.

<3維投影處理之流程>
其次,參照圖15的流程圖,說明圖14之步驟S221中所執行的3維投影處理之流程例。
<3D projection processing flow>
Next, an example of the flow of the three-dimensional projection processing executed in step S221 of Fig. 14 will be described with reference to the flowchart of Fig. 15 .

一旦3維投影處理被開始,則3維投影部251,係於步驟S241中,令表示圖層枚數i的資訊被解碼。When the three-dimensional projection processing is started, the three-dimensional projection unit 251, in step S241, causes the information indicating the number of layers i to be decoded.

於步驟S242中,3維投影部251,係令對應之i畫格份的幾何影像被解碼。In step S242, the three-dimensional projection unit 251 causes the geometric image of the corresponding i-frame to be decoded.

於步驟S243中,像素分布解析部252,係進行初期化使變數k=0。In step S243, the pixel distribution analysis unit 252 performs initialization to make the variable k=0.

於步驟S244中,像素分布解析部252,係判定是否變數k<i。若判定為k<i,則處理係前進至步驟S245。In step S244, the pixel distribution analysis unit 252 determines whether or not the variable k<i. If it is determined that k < i, the processing proceeds to step S245.

於步驟S245中,3維投影部251係將圖層k (Layer k)的3D資料,投影至3維空間。In step S245, the three-dimensional projection unit 251 projects the 3D material of the layer k (Layer k) into a three-dimensional space.

於步驟S246中,像素分布解析部252,係將變數k予以增值(k++)。In step S246, the pixel distribution analysis unit 252 adds the variable k (k++).

一旦步驟S246之處理結束,則處理係回到步驟S244。Once the process of step S246 ends, the process returns to step S244.

又,於步驟S244中若判定為並非變數k<i的情況下,則結束3維投影處理,處理係回到圖14。Moreover, if it is determined in step S244 that the variable k<i is not present, the three-dimensional projection processing is terminated, and the processing returns to FIG.

藉由如以上般地執行各處理,可實現已被投影至圖層數為可變之2維平面上的3D資料之重新建構。亦即,可較正確地重現在物件表面具有厚度的點雲。亦即,可抑制3D資料的2維投影所致之品質的降低。By performing the respective processes as above, reconstruction of the 3D data that has been projected onto the two-dimensional plane whose number of layers is variable can be realized. That is, the point cloud having a thickness on the surface of the object can be reproduced more correctly. That is, it is possible to suppress a decrease in quality due to 2-dimensional projection of 3D data.

<3.第2實施形態>
<表示「空像素」的值之傳訊>
一旦將稀疏(Sparse)的3D資料投影至2維平面,則會發生於該2維平面上未被投影3D資料的像素,亦即未被設定像素值的像素(亦稱為空像素)的情況。
<3. Second embodiment>
<Message indicating the value of "empty pixel">
Once the sparse (Sparse) 3D data is projected onto the 2D plane, it will occur in the 2D plane where the 3D data is not projected, that is, the pixel (also referred to as the empty pixel) that is not set to the pixel value. .

在先前的方法中,3D資料為稀疏(Sparse)的情況下,如圖16所示,對此種「空像素」會補足像素值。例如圖16的情況下,對圖層0(Layer0)之「空像素」,是從其左鄰之像素複製像素值(進行像素補足)。會進行此種補足處理的原因是,2維影像用之編碼方法(例如AVC或HEVC等)並沒有無資料存在(空白)的概念。In the previous method, in the case where the 3D data is sparse (Sparse), as shown in FIG. 16, such "empty pixels" complement the pixel values. For example, in the case of FIG. 16, for the "empty pixel" of the layer 0 (Layer0), the pixel value is copied from the pixel adjacent to the left side (the pixel complement is performed). The reason for this complement processing is that the coding method for 2D images (such as AVC or HEVC, etc.) does not have the concept of no data (blank).

然而,一旦進行如此的補足處理,就會對3D資料附加原本不存在的點。因此,恐怕會導致3D資料劣化。亦即,恐怕會因為3D資料的2維投影而導致品質降低。However, once such a complementary process is performed, a point that does not exist originally is added to the 3D material. Therefore, it is feared that the 3D data will deteriorate. That is, I am afraid that the quality will be degraded due to the 2-dimensional projection of 3D data.

此外,若不進行此種補足處理的情況下,則必須要以使得在3D資料所被投影的2維平面上不會產生「空像素」的方式,來設定領域。因此,就會需要微小領域(Point)單位之投影,而恐怕會導致處理量增大,或編碼效率降低等等。Further, if such a complement processing is not performed, it is necessary to set the field so that "empty pixels" are not generated on the two-dimensional plane on which the 3D data is projected. Therefore, a projection of a small point unit is required, which may lead to an increase in throughput or a decrease in coding efficiency.

於是,將表示3維結構之3D資料中所含之每一位置之資料投影至2維平面,並在該2維平面之每一位置之資料為不存在的位置上,設置表示該每一位置之資料為不存在的所定之值。例如,於影像處理裝置中,使其具備2維投影部,係將表示3維結構之3D資料中所含之每一位置之資料投影至2維平面,並在該2維平面之每一位置之資料為不存在的位置上,設置表示該每一位置之資料為不存在的所定之值。Then, the data representing each position included in the 3D data of the 3-dimensional structure is projected onto the 2-dimensional plane, and at each position where the data of each position of the 2-dimensional plane is non-existent, the setting indicates each position. The data is the value that does not exist. For example, in the image processing apparatus, the two-dimensional projection unit is provided to project data of each position included in the 3D data of the three-dimensional structure to a two-dimensional plane, and at each position of the two-dimensional plane. The data is a location where the data does not exist, and the data indicating that each location is a predetermined value that does not exist.

例如如圖17所示,作為3D資料所被投影之2維平面上的表示「空像素」的像素值,是使用規定之值X。該X之值係亦可設成例如,已被預先決定之所定之固定值。例如,亦可設成X=255(8bit之上限)。For example, as shown in FIG. 17, the pixel value indicating "empty pixel" on the two-dimensional plane on which the 3D data is projected is a predetermined value X. The value of X can also be set, for example, to a predetermined fixed value that has been predetermined. For example, it can also be set to X=255 (the upper limit of 8 bits).

此情況下,必須要在位元串流中,將X之值予以傳訊(必須要通知給解碼側)。In this case, the value of X must be signaled in the bit stream (must be notified to the decoding side).

又,例如,X之值,係亦可利用X>D的任意之值。此處的D係表示,該當定界框之縱深之最大值。在2維平面的像素值中,不可能設置超出該定界框之縱深之最大值以上之值。因此,該亦可將未使用的值當作X來使用。Further, for example, the value of X may be any value of X>D. The D system here indicates the maximum value of the depth of the bounding box. In the pixel value of the 2-dimensional plane, it is impossible to set a value exceeding the maximum value of the depth of the bounding frame. Therefore, it is also possible to use unused values as X.

此情況下,於解碼側,係可根據定界框之尺寸的相關之資訊,來判別已被設置至2維平面的像素值是否為未使用之值。因此,不需要將該X之值在位元串流中進行傳訊(亦即不需要將該X之值通知給解碼側)。但是,當然還是可以把X之值通知給解碼側。In this case, on the decoding side, it is possible to determine whether the pixel value set to the 2-dimensional plane is an unused value based on the related information of the size of the bounding frame. Therefore, there is no need to signal the value of X in the bit stream (ie, there is no need to inform the decoding side of the value of X). However, of course, the value of X can be notified to the decoding side.

藉由如此設計,編碼裝置100,就不需要進行資料之補足,即可表現「空像素」,因此可抑制3D資料的劣化。亦即,可抑制3D資料的2維投影所致之品質的降低。With such a design, the encoding apparatus 100 can express "empty pixels" without complementing the data, thereby suppressing deterioration of 3D data. That is, it is possible to suppress a decrease in quality due to 2-dimensional projection of 3D data.

又,藉由如此設計,編碼裝置100係可降低2維平面中的「無法編碼之像素」的數量,因此相較於單純不進行補足處理的情況,可將領域設定得較大。因此,不需要微小領域(Point)單位之投影,可抑制處理量的增大或編碼效率的降低。Further, with such a design, the encoding apparatus 100 can reduce the number of "pixels that cannot be encoded" in the two-dimensional plane. Therefore, the field can be set larger than when the complementary processing is not performed. Therefore, projection of a minute unit is not required, and an increase in the amount of processing or a decrease in coding efficiency can be suppressed.

<2維投影處理之流程>
此情況下也是,編碼處理及斑塊分解處理,係和第1實施形態中所說明的情況同樣地被進行。因此省略這些的說明。
<Process of 2-dimensional projection processing>
Also in this case, the encoding process and the plaque decomposition process are performed in the same manner as described in the first embodiment. Therefore, the description of these is omitted.

於圖11的步驟S124中所被執行的,此情況的2維投影處理之流程之例子,參照圖18的流程圖而加以說明。An example of the flow of the two-dimensional projection processing in this case executed in step S124 of Fig. 11 will be described with reference to the flowchart of Fig. 18.

一旦2維投影處理被開始,則2維投影部154,係於步驟S301中,進行分段所致之領域抽出。於步驟S302中,像素分布解析部155係判定是否為含有稀疏點雲的領域。若判定為是含有稀疏點雲的領域的情況,則處理係前進至步驟S303。When the two-dimensional projection processing is started, the two-dimensional projection unit 154 performs the field extraction by the segmentation in step S301. In step S302, the pixel distribution analysis unit 155 determines whether or not it is a field including a sparse point cloud. If it is determined that it is the field including the sparse point cloud, the processing proceeds to step S303.

於步驟S303中,2維投影部154,係將對該當領域定義「空像素」的資訊,予以傳訊。亦即,設定上述的X,利用該X來表現,已被投影至2維平面上的「不存在的點」。一旦步驟S303之處理結束,則處理係前進至步驟S305。In step S303, the two-dimensional projection unit 154 transmits information that defines "empty pixels" for the field. That is, the X described above is set and expressed by the X, and has been projected onto the "non-existing point" on the two-dimensional plane. Once the process of step S303 ends, the process proceeds to step S305.

又,於步驟S302中,若判定為是不含稀疏點雲的領域,則處理係前進至步驟S304。Moreover, if it is determined in step S302 that the field of the sparse point cloud is not included, the processing proceeds to step S304.

於步驟S304中,2維投影部154,係將該當領域投影至2維平面之各圖層。一旦步驟S304之處理結束,則處理係前進至步驟S305。In step S304, the two-dimensional projection unit 154 projects the field to each layer of the two-dimensional plane. Once the process of step S304 ends, the process proceeds to step S305.

於步驟S305中,2維投影部154係令該當領域的幾何影像被編碼。In step S305, the two-dimensional projection unit 154 causes the geometric image of the field to be encoded.

一旦步驟S305的處理結束,則2維投影處理就結束,處理係回到圖11。Once the process of step S305 ends, the 2-dimensional projection process ends and the process returns to FIG.

如以上所述,藉由進行2維投影處理,編碼裝置100係可不必進行資料之補足,就能表現「空像素」。因此,可抑制3D資料的劣化。亦即,可抑制3D資料的2維投影所致之品質的降低。As described above, by performing the two-dimensional projection processing, the encoding apparatus 100 can express "empty pixels" without complementing the data. Therefore, deterioration of 3D data can be suppressed. That is, it is possible to suppress a decrease in quality due to 2-dimensional projection of 3D data.

又,藉由如此設計,編碼裝置100係可降低2維平面中的「無法編碼之像素」的數量,因此相較於單純不進行補足處理的情況,可將領域設定得較大。因此,不需要微小領域(Point)單位之投影,可抑制處理量的增大或編碼效率的降低。Further, with such a design, the encoding apparatus 100 can reduce the number of "pixels that cannot be encoded" in the two-dimensional plane. Therefore, the field can be set larger than when the complementary processing is not performed. Therefore, projection of a minute unit is not required, and an increase in the amount of processing or a decrease in coding efficiency can be suppressed.

<已被傳訊的表示「空像素」之值的利用>
於解碼側,將已被投影至2維平面上的3D資料,投影至3維空間之際,會將如以上所述般地所被傳訊的表示2維平面上之「空像素」的像素值(上述的X)予以偵測、並刪除(使其不進行投影)。
<Usage of the value of "empty pixel" that has been transmitted >
On the decoding side, when the 3D data that has been projected onto the 2-dimensional plane is projected into the 3-dimensional space, the pixel value of the "empty pixel" on the two-dimensional plane that is transmitted as described above will be transmitted. (X above) is detected and deleted (so that it is not projected).

亦即,在已被投影至2維平面上的表示3維結構之3D資料中所含之每一位置之資料之內,將表示該每一位置之資料為不存在的所定之值的資料以外之資料,投影至3維空間。例如,於影像處理裝置中,使其具備3維投影部,係在已被投影至2維平面上的表示3維結構之3D資料中所含之每一位置之資料之內,將表示該每一位置之資料為不存在的所定之值的資料以外之資料,投影至3維空間。That is, in the data of each position included in the 3D data representing the 3-dimensional structure that has been projected onto the 2-dimensional plane, it is indicated that the data of each position is the data of the predetermined value that does not exist. The data is projected into a 3-dimensional space. For example, in the image processing apparatus, the three-dimensional projection unit is provided, and is displayed in the data of each position included in the 3D data representing the three-dimensional structure that has been projected onto the two-dimensional plane. The data of a location is data other than the data of the predetermined value that does not exist, and is projected into a 3-dimensional space.

藉由如此設計,就不需要進行資料之補足,即可表現「空像素」,因此可抑制3D資料的劣化。亦即,可抑制3D資料的2維投影所致之品質的降低。With this design, it is possible to express "empty pixels" without having to supplement the data, thereby suppressing deterioration of 3D data. That is, it is possible to suppress a decrease in quality due to 2-dimensional projection of 3D data.

又,藉由如此設計,可降低2維平面中的「無法編碼之像素」的數量,因此相較於單純不進行補足處理的情況,可將領域設定得較大。因此,不需要微小領域(Point)單位之投影,可抑制處理量的增大或編碼效率的降低。Further, by designing this, the number of "pixels that cannot be encoded" in the two-dimensional plane can be reduced, so that the field can be set larger than when the complement processing is not performed alone. Therefore, projection of a minute unit is not required, and an increase in the amount of processing or a decrease in coding efficiency can be suppressed.

<3維投影處理之流程>
此情況下也是,解碼處理及點雲重新建構處理,係和第1實施形態中所說明的情況同樣地被進行。因此省略這些的說明。
<3D projection processing flow>
Also in this case, the decoding process and the point cloud reconstruction process are performed in the same manner as described in the first embodiment. Therefore, the description of these is omitted.

於圖14的步驟S221中所被執行的,此情況的3維投影處理之流程之例子,參照圖19的流程圖而加以說明。An example of the flow of the three-dimensional projection processing in this case executed in step S221 of Fig. 14 will be described with reference to the flowchart of Fig. 19.

一旦3維投影處理被開始,則3維投影部251,係於步驟S321中,令處理對象之領域的幾何影像被解碼。When the three-dimensional projection processing is started, the three-dimensional projection unit 251, in step S321, decodes the geometric image of the domain of the processing target.

於步驟S322中,像素分布解析部252係判定,在該幾何影像中,是否存在有定義「空像素」的資訊(表示「空像素」的像素值)。若判定為存在,則處理係前進至步驟S323。In step S322, the pixel distribution analysis unit 252 determines whether or not there is information (a pixel value indicating "empty pixel") in which the "empty pixel" is defined in the geometric image. If it is determined to be present, the process proceeds to step S323.

於步驟S323中,3維投影部251,係將該已被偵測到的表示「空像素」的像素值之像素,予以刪除(設定使其不投影至3維空間)。一旦步驟S323之處理結束,則處理係前進至步驟S324。In step S323, the three-dimensional projection unit 251 deletes the pixel of the detected pixel value indicating "empty pixel" (set it so as not to project into a three-dimensional space). Once the process of step S323 ends, the process proceeds to step S324.

又,於步驟S322中,若判定為表示「空像素」的像素值是不存在的情況下,則省略步驟S323之處理,處理係前進至步驟S324。In addition, if it is determined in step S322 that the pixel value indicating "empty pixel" does not exist, the processing of step S323 is omitted, and the processing proceeds to step S324.

亦即,對處理對象之領域的幾何影像,進行表示「空像素」的像素值之偵測,在有偵測到的情況下,則該像素會被刪除。That is, the detection of the pixel value indicating "empty pixel" is performed on the geometric image of the domain of the processing object, and if detected, the pixel is deleted.

於步驟S324中,3維投影部251,係將處理對象之領域的幾何影像,投影至3維空間。In step S324, the three-dimensional projection unit 251 projects the geometric image of the domain to be processed into a three-dimensional space.

一旦步驟S324的處理結束,則3維投影處理就結束,處理係回到圖14。Once the process of step S324 ends, the 3-dimensional projection process ends and the process returns to FIG.

藉由如以上般地執行3維投影處理,解碼裝置200就不需要進行資料之補足,即可表現「空像素」,因此可抑制3D資料的劣化。亦即,可抑制3D資料的2維投影所致之品質的降低。By performing the three-dimensional projection processing as described above, the decoding apparatus 200 can express "empty pixels" without complementing the data, and thus it is possible to suppress degradation of the 3D data. That is, it is possible to suppress a decrease in quality due to 2-dimensional projection of 3D data.

又,藉由如此設計,解碼裝置200係可降低2維平面中的「無法編碼之像素」的數量,因此相較於單純不進行補足處理的情況,可將領域設定得較大。因此,不需要微小領域(Point)單位之投影,可抑制處理量的增大或編碼效率的降低。Further, with such a design, the decoding apparatus 200 can reduce the number of "pixels that cannot be encoded" in the two-dimensional plane. Therefore, the field can be set larger than when the complement processing is not performed alone. Therefore, projection of a minute unit is not required, and an increase in the amount of processing or a decrease in coding efficiency can be suppressed.

<4.第3實施形態>
<縱深參數之控制>
將3D資料投影至2維平面之際,控制該投影至2維平面的3D資料之縱深方向之範圍的縱深參數th,會被利用。由於被該縱深參數th所指定之範圍內的點會成為往2維平面之投影對象,因此該縱深參數th之值,係與領域(Local bounding box)的縱深方向之長度有所關連。例如,若相較於領域的縱深方向之長度而縱深參數th之值為較大,則可能其他領域的點也會變成投影對象。亦即,領域的縱深方向之長度,必須要比縱深參數th還長。
<4. Third embodiment>
<Control of depth parameters>
When the 3D data is projected onto the two-dimensional plane, the depth parameter th that controls the depth direction of the 3D data projected onto the two-dimensional plane is used. Since the point within the range specified by the depth parameter th becomes a projection object to the two-dimensional plane, the value of the depth parameter th is related to the length of the depth direction of the local bounding box. For example, if the value of the depth parameter th is larger than the length of the depth direction of the field, then points in other fields may become projection objects. That is, the length of the field in the depth direction must be longer than the depth parameter th.

在先前的方法中,係如圖20所示,該縱深參數th,係以畫格單位而被控制。因此,畫格內之所有的縱深參數th,係具有共通的值。In the previous method, as shown in Fig. 20, the depth parameter th is controlled in frame units. Therefore, all the depth parameters th in the frame have a common value.

然而,例如像是臉與腳、中心部與週邊部,於畫格內點雲之表面密度並非一定的情況,也是可能發生的。因此,縱深參數th之值並非總是最佳,而恐怕會導致編碼效率降低。例如,明明朝縱深方向來分割領域是編碼效率較佳,可是會有縱深參數th之值會變大,而導致無法分割的案例。However, for example, the face and the feet, the center and the peripheral portion, the surface density of the point cloud in the frame is not constant, and it is also possible. Therefore, the value of the depth parameter th is not always optimal, and it is feared that the coding efficiency is lowered. For example, it is better to divide the field in the depth direction of the Ming Dynasty, but the value of the depth parameter th will become larger, resulting in an inseparable case.

於是,將表示3維結構之3D資料的每一位置之資料,按照3維空間之所定之每一領域地投影至2維平面之際,將對該每一領域所被設定之用來限制可投影至1圖層的表示3維結構之3D資料的每一位置之資料之縱深方向之範圍的縱深參數所表示的縱深方向之範圍內之每一位置之資料,投影至該2維平面。例如,於影像處理裝置中,使其具備2維投影部,係將表示3維結構之3D資料的每一位置之資料,按照3維空間之所定之每一領域地投影至2維平面之際,將對該每一領域所被設定之用來限制可投影至1圖層的表示3維結構之3D資料的每一位置之資料之縱深方向之範圍的縱深參數所表示的縱深方向之範圍內之每一位置之資料,投影至該2維平面。Therefore, the data of each position of the 3D data representing the 3-dimensional structure is projected to the 2-dimensional plane according to each field defined by the 3-dimensional space, and the setting for each field is limited. The data projected to each of the positions in the depth direction indicated by the depth parameter in the depth direction of the data of each position of the 3D data of the one-dimensional structure projected onto the one-layer layer is projected onto the two-dimensional plane. For example, in the image processing device, the two-dimensional projection unit is provided, and the data of each position of the 3D data representing the three-dimensional structure is projected onto the two-dimensional plane in accordance with each field defined by the three-dimensional space. And each of the fields is set to limit the depth direction indicated by the depth parameter of the range of the depth direction of the data of each position of the 3D data representing the three-dimensional structure that can be projected to one layer. The data for each location is projected onto the 2D plane.

例如,如圖21所示般地,擴充成可將縱深參數每一斑塊地予以傳輸,而每一領域地傳輸要投影至圖層(Layer)的像素之位置(TH)。藉由如此設計,例如,可對領域1設定縱深參數th,並對領域2設定縱深參數th’。因此,可改善圖層1(Layer1)的效率(Geometry編碼效率)。For example, as shown in FIG. 21, it is expanded to transmit the depth parameter for each patch, and each field transmits the position (TH) of the pixel to be projected to the layer. By designing in this way, for example, the depth parameter th can be set for the field 1 and the depth parameter th' can be set for the field 2. Therefore, the efficiency of Layer 1 (Geyer coding efficiency) can be improved.

<2維投影處理之流程>
此情況下也是,編碼處理及斑塊分解處理,係和第1實施形態中所說明的情況同樣地被進行。因此省略這些的說明。
<Process of 2-dimensional projection processing>
Also in this case, the encoding process and the plaque decomposition process are performed in the same manner as described in the first embodiment. Therefore, the description of these is omitted.

於圖11的步驟S124中所被執行的,此情況的2維投影處理之流程之例子,參照圖22的流程圖而加以說明。An example of the flow of the two-dimensional projection processing in this case executed in step S124 of Fig. 11 will be described with reference to the flowchart of Fig. 22 .

一旦2維投影處理被開始,則像素分布解析部155係於步驟S401中,進行初期化以使領域號碼i=0。When the two-dimensional projection processing is started, the pixel distribution analysis unit 155 performs initialization in step S401 so that the field number i=0.

於步驟S402中,像素分布解析部155係判定是否有未處理像素存在。若判定為有未處理像素存在,則處理係前進至步驟S403。In step S402, the pixel distribution analysis unit 155 determines whether or not there is an unprocessed pixel. If it is determined that there is an unprocessed pixel, the processing proceeds to step S403.

於步驟S403中,2維投影部154係進行分段所致之領域i之抽出。In step S403, the two-dimensional projection unit 154 performs extraction of the field i by segmentation.

於步驟S404中,2維投影部154係進行縱深參數的調整與編碼(RD判定)。亦即,最佳的縱深參數th與領域之範圍的設定,是藉由RD判定而被進行。In step S404, the two-dimensional projection unit 154 performs adjustment and encoding of the depth parameter (RD determination). That is, the setting of the optimum depth parameter th and the range of the field is performed by the RD determination.

於步驟S405中,2維投影部154,係基於步驟S404中所進行的設定,而將處理對象領域i之幾何影像,投影至2維表面。In step S405, the two-dimensional projection unit 154 projects the geometric image of the processing target domain i onto the two-dimensional surface based on the settings made in step S404.

於步驟S406中,2維投影部154,係令已被投影至該2維平面上的幾何影像被編碼。In step S406, the two-dimensional projection unit 154 causes the geometric image that has been projected onto the two-dimensional plane to be encoded.

於步驟S407中,像素分布解析部155,係將變數i予以增值(i++)。一旦步驟S407之處理結束,則處理係回到步驟S402,重複其以後之處理。In step S407, the pixel distribution analysis unit 155 adds the variable i (i++). Once the process of step S407 ends, the process returns to step S402, and the subsequent processes are repeated.

又,於步驟S402中,若判定為未處理像素是不存在的情況下,則處理係前進至步驟S408。Moreover, if it is determined in step S402 that the unprocessed pixel does not exist, the processing proceeds to step S408.

於步驟S408中,2維投影部154,係令遮蔽貼圖(Occulusion Map)被編碼。一旦步驟S408的處理結束,則2維投影處理就結束,處理係回到圖11。In step S408, the two-dimensional projection unit 154 causes the Occulusion Map to be encoded. Once the processing of step S408 ends, the two-dimensional projection processing ends, and the processing returns to FIG.

藉由如以上所述般地進行2維投影處理,編碼裝置100係可獲得較適合於各領域的縱深參數th,可抑制編碼效率的降低。By performing the two-dimensional projection processing as described above, the encoding apparatus 100 can obtain the depth parameter th suitable for each field, and can suppress the reduction in encoding efficiency.

<被控制的縱深參數之利用>
於解碼側,將已被投影至2維平面上的3D資料,投影至3維空間之際,會在如以上般地所被控制的縱深參數th所示的範圍內,投影3D資料。
<Use of controlled depth parameters>
On the decoding side, when the 3D data that has been projected onto the two-dimensional plane is projected into the three-dimensional space, the 3D data is projected within the range indicated by the depth parameter th controlled as described above.

亦即,按照3維空間之所定之每一領域,將已被投影至2維平面上的表示3維結構之3D資料中所含之每一位置之資料投影至該3維空間之際,是將該每一位置之資料,投影至該3維空間的,對該每一領域所被設定之用來限制可投影至1圖層的3D資料之每一位置之資料之縱深方向之範圍的縱深參數所表示的該縱深方向之範圍內。例如,於影像處理裝置中,使其具備3維投影部,係按照3維空間之所定之每一領域,將已被投影至2維平面上的表示3維結構之3D資料中所含之每一位置之資料投影至該3維空間之際,是將該每一位置之資料,投影至該3維空間的,對該每一領域所被設定之用來限制可投影至1圖層的3D資料之每一位置之資料之縱深方向之範圍的縱深參數所表示的該縱深方向之範圍內。That is, according to each field defined by the three-dimensional space, when the data of each position included in the 3D data representing the three-dimensional structure projected onto the two-dimensional plane is projected onto the three-dimensional space, Projecting the data of each position into the 3D space, and setting the depth parameter of the range of the depth direction of the data of each position of the 3D data that can be projected to the 1 layer for each field Within the range of the depth direction indicated. For example, in the image processing device, the three-dimensional projection unit is provided, and each of the 3D data representing the three-dimensional structure that has been projected onto the two-dimensional plane is included in each of the fields defined by the three-dimensional space. When the data of a position is projected into the three-dimensional space, the data of each position is projected into the three-dimensional space, and the set of each field is used to limit the 3D data that can be projected to one layer. The depth direction of the range of the depth direction of the data of each position is within the range of the depth direction.

藉由如此設計,就可利用對每一領域所被設定之縱深參數th來進行3D資料往3維空間之投影,因此可抑制編碼效率的降低。With this design, the depth parameter th set for each field can be used to project the 3D data into the three-dimensional space, so that the reduction in coding efficiency can be suppressed.

<3維投影處理之流程>
此情況下也是,解碼處理及點雲重新建構處理,係和第1實施形態中所說明的情況同樣地被進行。因此省略這些的說明。
<3D projection processing flow>
Also in this case, the decoding process and the point cloud reconstruction process are performed in the same manner as described in the first embodiment. Therefore, the description of these is omitted.

於圖14的步驟S221中所被執行的,此情況的3維投影處理之流程之例子,參照圖23的流程圖而加以說明。An example of the flow of the three-dimensional projection processing in this case, which is executed in step S221 of Fig. 14, will be described with reference to the flowchart of Fig. 23.

一旦3維投影處理被開始,則3維投影部251,係於步驟S421中,令Occlusion地圖被解碼。Once the three-dimensional projection processing is started, the three-dimensional projection unit 251 is caused to decode the Occlusion map in step S421.

於步驟S422中,像素分布解析部252,係進行初期化使領域號碼k=0。In step S422, the pixel distribution analysis unit 252 initializes the field number k=0.

於步驟S423中,像素分布解析部252,係判定是否變數k<i。若判定為k<i,則處理係前進至步驟S424。In step S423, the pixel distribution analysis unit 252 determines whether or not the variable k<i. If it is determined that k < i, the processing proceeds to step S424.

於步驟S424中,3維投影部251係令縱深參數th被解碼。於步驟S425中,3維投影部251係令幾何影像被解碼。於步驟S426中,3維投影部251,係將該當領域的影像,投影至3維空間。In step S424, the three-dimensional projection unit 251 causes the depth parameter th to be decoded. In step S425, the three-dimensional projection unit 251 causes the geometric image to be decoded. In step S426, the three-dimensional projection unit 251 projects the image of the field into a three-dimensional space.

於步驟S427中,像素分布解析部252,係將變數k予以增值(k++)。一旦步驟S427之處理結束,則處理係回到步驟S423,重複其以後之處理。In step S427, the pixel distribution analysis unit 252 adds the variable k (k++). Once the process of step S427 is ended, the process returns to step S423, and the subsequent processes are repeated.

又,於步驟S423中,若判定為並非k<i的情況下,則結束3維投影處理,處理係回到圖14。Further, if it is determined in step S423 that k < i is not satisfied, the three-dimensional projection processing is terminated, and the processing returns to FIG.

藉由如以上般地執行3維投影處理,解碼裝置200係可利用對每一領域所被設定之縱深參數th來進行3D資料往3維空間之投影,因此可抑制編碼效率的降低。By performing the three-dimensional projection processing as described above, the decoding apparatus 200 can project the 3D data into the three-dimensional space by using the depth parameter th set for each field, thereby suppressing the reduction in coding efficiency.

<5.附記>
<控制資訊>
以上的各實施形態中所說明的關於本技術的控制資訊,亦可從編碼側傳輸至解碼側。例如,亦可將用來控制是否許可(或禁止)適用上述本技術的控制資訊(例如enabled_flag),予以傳輸。又,例如,亦可將用來指定許可(或禁止)適用上述本技術之範圍(例如區塊大小之上限或是下限、或其雙方、切片、圖像、序列、分量、視點、圖層等)的控制資訊,予以傳輸。
<5. Attachment>
<Control Information>
The control information about the present technology described in the above embodiments may be transmitted from the encoding side to the decoding side. For example, control information (e.g., enabled_flag) for applying the above-described technique may be controlled to be permitted (or disabled) for transmission. Also, for example, the scope of the above-described technology (eg, the upper or lower limit of the block size, or both sides, slices, images, sequences, components, viewpoints, layers, etc.) may be used to specify permission (or prohibition). Control information is transmitted.

<電腦>
上述之一連串之處理,係可藉由硬體來執行,亦可藉由軟體來執行。在以軟體來執行一連串之處理時,構成該軟體的程式,係可安裝至電腦。此處,電腦係包含:被組裝在專用硬體中的電腦、或藉由安裝各種程式而可執行各種機能的例如通用之個人電腦等。
<computer>
The series of processing described above can be performed by hardware or by software. When a series of processes are executed in software, the program constituting the software can be installed to a computer. Here, the computer system includes a computer that is incorporated in a dedicated hardware, or a general-purpose personal computer that can perform various functions by installing various programs.

圖24係以程式來執行上述一連串處理的電腦的硬體之構成例的區塊圖。Fig. 24 is a block diagram showing a configuration example of a hardware of a computer that executes the above-described series of processes in a program.

於圖24所示的電腦900中,CPU(Central Processing Unit)901、ROM(Read Only Memory)902、RAM (Random Access Memory)903,係透過匯流排904而被彼此連接。In the computer 900 shown in FIG. 24, a CPU (Central Processing Unit) 901, a ROM (Read Only Memory) 902, and a RAM (Random Access Memory) 903 are connected to each other via the bus bar 904.

匯流排904,係還連接著輸出入介面910。輸出入介面910上係連接有:輸入部911、輸出部912、記憶部913、通訊部914、及驅動機915。The bus 904 is also connected to the input/output interface 910. The input/output interface 910 is connected to an input unit 911, an output unit 912, a storage unit 913, a communication unit 914, and a drive unit 915.

輸入部911,係例如由鍵盤、滑鼠、麥克風、觸控面板、輸入端子等所成。輸出部912係由例如顯示器、揚聲器、輸出端子等所成。記憶部913,係由例如硬碟、RAM碟、非揮發性記憶體等所成。通訊部914係由例如網路介面所成。驅動機915係驅動:磁碟、光碟、光磁碟、或半導體記憶體等之可移除式媒體921。The input unit 911 is formed, for example, by a keyboard, a mouse, a microphone, a touch panel, an input terminal, or the like. The output unit 912 is formed of, for example, a display, a speaker, an output terminal, or the like. The memory unit 913 is formed of, for example, a hard disk, a RAM disk, a non-volatile memory, or the like. The communication unit 914 is formed by, for example, a network interface. The drive unit 915 is driven by a removable medium 921 such as a magnetic disk, a compact disk, a magneto-optical disk, or a semiconductor memory.

在如以上構成的電腦中,藉由CPU901而例如將記憶部913中所記憶之程式,透過輸出入介面910及匯流排904,而載入至RAM903裡並加以執行,就可進行上述一連串處理。RAM903中,還適宜地記憶著CPU901在執行各種處理時所必需的資料等。In the computer having the above configuration, the program stored in the memory unit 913 is loaded into the RAM 903 through the input/output interface 910 and the bus bar 904, for example, by the CPU 901, and the above-described series of processes can be performed. In the RAM 903, information necessary for the CPU 901 to execute various processes and the like are also appropriately stored.

電腦(CPU901)所執行的程式,係可記錄在例如封裝媒體等之可移除式媒體921中而適用。此情況下,程式係藉由將可移除式媒體921裝著至驅動機915,就可透過輸出入介面910,安裝至記憶部913。The program executed by the computer (CPU 901) can be recorded in a removable medium 921 such as a package medium. In this case, the program is attached to the memory unit 913 via the input/output interface 910 by attaching the removable medium 921 to the drive unit 915.

又,該程式係可透過區域網路、網際網路、數位衛星播送這類有線或無線的傳輸媒體而提供。此情況,程式,係可以通訊部914而接收之,並安裝至記憶部913。In addition, the program can be provided via a wired or wireless transmission medium such as a regional network, an Internet, or a digital satellite. In this case, the program can be received by the communication unit 914 and installed in the memory unit 913.

除此以外,該程式係可事前安裝在ROM902或記憶部913中。In addition to this, the program can be installed in advance in the ROM 902 or the memory unit 913.

<本技術的適用對象>
以上雖然針對在點雲資料之Voxel化中適用本技術的情況加以說明,但本技術係不限於這些例子,亦可對任意規格的3D資料之Voxel化做適用。亦即,只要不與上述的本技術產生矛盾,編碼、解碼方式等之各種處理、以及3D資料或後設資料等之各種資料的規格,係為任意。又,只要不與本技術產生矛盾,亦可省略上述的一部分之處理或規格。
<Applicable object of the technology>
Although the above description has been given of the case where the present technology is applied to Voxelization of point cloud data, the present technology is not limited to these examples, and may be applied to Voxelization of 3D data of any specification. In other words, as long as there is no contradiction with the above-described technique, various processing such as encoding and decoding methods, and specifications of various materials such as 3D data or post-data are arbitrary. Further, as long as there is no contradiction with the present technology, a part of the processing or specifications described above may be omitted.

又,以上作為本技術之適用例雖然說明了編碼裝置100及解碼裝置200,但本技術係可適用於任意之構成。Further, although the encoding device 100 and the decoding device 200 have been described above as an application example of the present technology, the present technology can be applied to any configuration.

例如,本技術係可適用於衛星播送、有線TV等之有線播送、網際網路上的配訊、及藉由蜂巢基地台通訊而對終端之配訊等時候的送訊機或收訊機(例如電視受像機或行動電話機)、或是在光碟、磁碟及快閃記憶體等之媒體中記錄影像、或從這些記憶媒體中再生出影像的裝置(例如硬碟錄影機或攝影機)等,各式各樣的電子機器。For example, the present technology can be applied to a satellite broadcast, a wired TV broadcast of a cable TV, a communication on the Internet, and a transmitter or a receiver at the time of communication with a terminal by communication with a cellular base station (for example, a television receiver or a mobile phone, or a device that records images on or reproduces images from media such as a compact disc, a flash memory, or the like (such as a hard disk recorder or a video camera). A wide range of electronic machines.

又例如,本技術係亦可作為系統LSI(Large Scale Integration)等的處理器(例如視訊處理器)、使用複數處理器等的模組(例如視訊模組)、使用複數模組等的單元(例如視訊單元)、對單元再附加其他機能而成的套組(例如視訊套組)等,作為裝置的部分構成而實施。Further, for example, the present technology can also be used as a processor such as a system LSI (Large Scale Integration) (for example, a video processor), a module using a complex processor (for example, a video module), or a unit using a complex module (for example). For example, a video unit, a kit (for example, a video set) to which other functions are added to the unit, and the like are implemented as part of the device.

又例如,本技術係亦可對由複數裝置所構成的網路系統做適用。例如,亦可將本技術,透過網路而分擔給複數台裝置,以共同進行處理的雲端運算的方式,來加以實施。例如,亦可在對電腦、AV(Audio Visual)機器、攜帶型資訊處理終端、IoT(Internet of Things)裝置等之任意之終端,提供影像(動態影像)的相關之服務的雲端服務中,實施本技術。For another example, the present technology can also be applied to a network system composed of a plurality of devices. For example, the present technology can be implemented by sharing a plurality of devices over a network and performing cloud computing for processing together. For example, it can be implemented in a cloud service that provides services related to video (motion video) to any terminal such as a computer, an AV (Audio Visual) device, a portable information processing terminal, and an IoT (Internet of Things) device. This technology.

此外,於本說明書中,所謂的系統,係意味著複數構成要素(裝置、模組(零件)等)的集合,所有構成要素是否位於同一框體內則在所不問。因此,被收納在個別的框體中,透過網路而連接的複數台裝置、及在1個框體中收納有複數模組的1台裝置,均為系統。In addition, in this specification, a system means a collection of a plurality of components (devices, modules (parts), etc.), and all components are not in the same frame. Therefore, a plurality of devices that are housed in individual frames, connected through a network, and one device in which a plurality of modules are housed in one frame are all systems.

<可適用本技術的領域、用途>
適用了本技術的系統、裝置、處理部等係可利用於例如:交通、醫療、防盜、農業、畜產業、礦業、美容、工場、家電、氣象、自然監視等任意之領域。又,其用途也為任意。
<Fields and uses in which the technology can be applied>
The system, the device, the processing unit, and the like to which the present technology is applied can be utilized in, for example, transportation, medical care, theft prevention, agriculture, animal industry, mining, beauty, workshops, home appliances, weather, natural surveillance, and the like. Moreover, its use is also arbitrary.

例如,本技術係可是用於,供作觀賞用內容等之提供之用的系統或裝置。又例如,本技術係亦可適用於交通狀況之監理或自動駕駛控制等,供作交通之用的系統或裝置。甚至例如,本技術係亦可適用於供作安全之用的系統或裝置。又例如,本技術係亦可適用於供作機械等之自動控制之用的系統或裝置。甚至例如,本技術係亦可適用於供作農業或畜產業之用的系統或裝置。又,本技術係亦可適用於例如監視火山、森林、海洋等之自然狀態或野生生物等的系統或裝置。甚至例如,本技術係亦可適用於供作運動之用的系統或裝置。For example, the present technology may be a system or device for providing for viewing content or the like. For example, the present technology can also be applied to a system or device for traffic use, such as supervision of traffic conditions or automatic driving control. Even, for example, the technology can be applied to systems or devices for security purposes. For another example, the present technology can also be applied to a system or device for automatic control of a machine or the like. Even, for example, the present technology can be applied to systems or devices for use in the agricultural or livestock industry. Further, the present technology can also be applied to, for example, a system or device that monitors natural states such as volcanoes, forests, oceans, and the like, or wildlife. Even, for example, the present technology can be applied to systems or devices for use in sports.

<其他>
此外,於本說明書中所謂「旗標」,係為用來識別複數狀態所需之資訊,不只包含在用來識別真(1)或偽(0)之2種狀態之際所使用的資訊,亦包含可以識別3種以上之狀態的資訊。因此,該「旗標」所能採取的值,係可為例如1/0之2值,亦可為3值以上。亦即,構成該「旗標」的bit數係為任意,可為1bit亦可為複數bit。又,識別資訊(亦包含旗標),係不只將該識別資訊含入至位元串流的形式,也想定了將識別資訊相對於某個作為基準之資訊的差分資訊含入至位元串流的形式,因此於本說明書中,「旗標」或「識別資訊」,係不只包含該資訊,也還包含了相對於作為基準之資訊的差分資訊。
<Other>
In addition, the "flag" in this specification is the information required to identify the complex state, and not only the information used to identify the two states of true (1) or pseudo (0), It also contains information that can identify more than three states. Therefore, the value that the "flag" can take is, for example, a value of 1/0, or a value of 3 or more. In other words, the number of bits constituting the "flag" is arbitrary, and may be 1 bit or a complex bit. Moreover, the identification information (including the flag) is not only included in the form of the bit stream, but also the differential information of the identification information relative to a certain reference information is included in the bit string. In the form of a stream, in this specification, "flag" or "identification information" includes not only the information but also differential information relative to the information as a reference.

又,編碼資料(位元串流)的相關之各種資訊(後設資料等),係只要與編碼資料建立關連,則無論是以哪種形態而被傳輸或記錄皆可。此處,「建立關連」此一用語係意味著例如:使得在一方之資料進行處理之際可能利用到他方之資料的情況(可建立連結)。亦即,被彼此建立關連的資料,係亦可整體視為1個資料,也可分別視為個別之資料。例如,與編碼資料(影像)建立關連的資訊,係亦可在有別於該編碼資料(影像)的其他傳輸路上被傳輸。又,例如,與編碼資料(影像)建立關連的資訊,係亦可在有別於該編碼資料(影像)的其他記錄媒體(或是同一記錄媒體的其他區域)中被記錄。此外,該「建立關連」,係亦可不是資料全體,而是資料的一部分。例如,影像與對應於該影像的資訊,係亦可使用複數畫格、1畫格、或畫格內之一部分等之任意之單位,而被彼此建立關連。Further, various kinds of information (post-design data, etc.) related to the encoded data (bit stream) may be transmitted or recorded in any form as long as it is related to the encoded data. Here, the term "establishing a connection" means, for example, a situation in which the information of the other party may be utilized when the data of one party is processed (a link may be established). That is to say, information that is related to each other can be regarded as one piece of information as a whole or as individual information. For example, information associated with encoded data (images) may also be transmitted on other transmission paths different from the encoded material (images). Further, for example, the information associated with the encoded material (image) may be recorded in another recording medium (or other area of the same recording medium) different from the encoded material (image). In addition, the "establishment of connection" may not be the whole of the information, but part of the information. For example, the image and the information corresponding to the image may be associated with each other using any of a plurality of frames, a frame, or a part of the frame.

此外,於本說明書中,「合成」、「多工化」、「附加」、「一體化」、「含入」、「儲存」、「放入」、「插進」、「插入」等之用語,係意味著例如將編碼資料與後設資料總結成1個資料,把複數個物綁成1個的意思,是意味著上述的「建立關連」的1種方法。In addition, in this specification, "synthesis", "multiplexing", "additional", "integration", "inclusion", "storage", "input", "insertion", "insertion", etc. The term means that, for example, the combination of the coded data and the post-set data into one piece of information and the binding of the plurality of items into one means a method of "establishing a connection" as described above.

又,本技術的實施形態係不限定於上述實施形態,在不脫離本技術主旨的範圍內可做各種變更。Further, the embodiments of the present invention are not limited to the above-described embodiments, and various modifications can be made without departing from the spirit and scope of the invention.

例如,亦可將以1個裝置(或處理部)做說明的構成加以分割,成為複數裝置(或處理部)而構成之。反之,亦可將以上說明中以複數裝置(或處理部)做說明的構成總結成1個裝置(或處理部)而構成之。又,對各裝置(或各處理部)之構成,當然亦可附加上述以外之構成。再者,若系統全體的構成或動作是實質相同,則亦可使某個裝置(或處理部)之構成的一部分被包含在其他裝置(或其他處理部)之構成中。For example, a configuration in which one device (or processing unit) is described may be divided into a plurality of devices (or processing units). On the other hand, the configuration described by the plurality of devices (or processing units) in the above description may be summarized into one device (or processing unit). Further, it is needless to say that the configuration of each device (or each processing unit) may be other than the above. Furthermore, if the entire configuration or operation of the system is substantially the same, a part of the configuration of a certain device (or processing unit) may be included in the configuration of another device (or other processing unit).

又,例如,上述的程式,係亦可於任意的裝置中被執行。此情況下,只要讓該裝置,具有必要的機能(機能區塊等),能夠獲得必要的資訊即可。Further, for example, the above-described program can be executed in any device. In this case, as long as the device has the necessary functions (function blocks, etc.), necessary information can be obtained.

又,例如,1個流程圖的各步驟,亦可由1個裝置來執行,也可由複數裝置來分擔而執行。甚至,1個步驟中包含有複數個處理的情況下,該複數個處理亦可由1個裝置來執行,也可由複數裝置來分擔而執行。換言之,亦可將1個步驟中所含之複數個處理,以複數個步驟之處理的方式而執行之。反之,亦可將以複數個步驟的方式做說明的處理,整合成1個步驟而執行之。Further, for example, each step of one flowchart may be executed by one device or may be performed by a plurality of devices. Even when a plurality of processes are included in one step, the plurality of processes may be executed by one device or may be performed by a plurality of devices. In other words, the plurality of processes included in one step can also be performed in a plurality of steps. On the other hand, the processing described in the plural steps can be integrated into one step and executed.

又,例如,電腦所執行的程式,描述程式的步驟之處理,係可為依照本說明書所說明之順序而在時間序列上被執行,也可平行地,或可在進行呼叫時等必要之時序上,而被個別地執行。亦即,只要不產生矛盾,各步驟之處理係亦可以和上述之順序不同的順序而被執行。甚至,描述該程式的步驟之處理,亦可與其他程式之處理平行地執行,也可和其他程式之處理組合而執行。Further, for example, the program executed by the computer and the processing of the steps describing the program may be executed in time series in accordance with the order described in the specification, or may be in parallel, or may be necessary to perform a call. On, but executed individually. That is, as long as no contradiction occurs, the processing of each step can be performed in a different order from the above. Even the processing of the steps describing the program can be performed in parallel with the processing of other programs, or can be performed in combination with the processing of other programs.

又,例如,本技術所相關之複數個技術,係只要不產生矛盾的況下,都可分別獨立以單體而加以實施。當然,亦可將任意的複數個本技術加以併用而實施。例如,可以將任一實施形態中所說明的本技術的部分或全部,與其他實施形態中所說明的本技術的部分或全部,加以組合而實施。又,亦可將上述的任意之本技術的部分或全部,與未上述的其他技術加以併用而實施。Further, for example, a plurality of techniques related to the present technology can be implemented independently as a single unit as long as no contradiction occurs. Of course, any of a plurality of the techniques can be used in combination. For example, part or all of the present technology described in any of the embodiments may be combined with some or all of the present technology described in the other embodiments. Further, some or all of the above-described techniques may be used in combination with other techniques not described above.

此外,本技術係亦可採取如以下之構成。
(1) 一種影像處理裝置,係具備:
2維投影部,係將表示3維結構之3D資料中所含之所有的每一位置之資料,投影至複數圖層之2維平面。
(2) 如(1)所記載之影像處理裝置,其中,
前記2維投影部,係將前記3D資料的從投影面來看位置是在縱深方向上重疊的前記每一位置之資料,投影至前記複數圖層之前記2維平面的彼此互異之圖層。
(3) 如(2)所記載之影像處理裝置,其中,
前記2維投影部,係針對前記2維平面,生成與前記3D資料的從投影面來看位置是在縱深方向上重疊的前記每一位置之資料之最大筆數相同數量之圖層。
(4) 如(1)乃至(3)之任一項所記載之影像處理裝置,其中,
還具備:編碼部,係將已被前記2維投影部投影至前記2維平面的前記3D資料,予以編碼。
(5) 如(4)所記載之影像處理裝置,其中,
前記編碼部,係將已被投影至前記2維平面之各圖層的前記3D資料的位置資訊、屬性資訊、及佔用地圖,分別予以編碼。
(6) 如(4)或(5)所記載之影像處理裝置,其中,
還具備:位元串流生成部,係生成位元串流,其中含有:表示藉由前記2維投影部而被投影了前記3D資料的前記2維平面之圖層數的資訊、和藉由前記編碼部將前記2維平面予以編碼所得之編碼資料。
(7) 如(4)乃至(6)之任一項所記載之影像處理裝置,其中,
還具備:打包部,係將藉由前記2維投影部而被投影了前記3D資料的前記2維平面,打包成為視訊畫格;
前記編碼部係被構成為,將藉由前記打包部把前記2維平面打包而成的前記視訊畫格,予以編碼。
(8) 如(1)乃至(7)之任一項所記載之影像處理裝置,其中,
前記2維投影部,係將前記3D資料,按照所定之每一領域地投影至前記2維平面。
(9) 如(1)乃至(8)之任一項所記載之影像處理裝置,其中,
前記3D資料,係為點雲。
(10) 一種影像處理方法,係
將表示3維結構之3D資料中所含之所有的每一位置之資料,投影至複數圖層之2維平面。
Further, the present technology can also adopt the following constitution.
(1) An image processing apparatus having:
The two-dimensional projection unit projects data of each position included in the 3D data of the three-dimensional structure onto a two-dimensional plane of the complex layer.
(2) The image processing device according to (1), wherein
In the second-dimensional projection unit, the position of the 3D data from the projection surface is the position of each position in the depth direction, and is projected on the two-dimensional plane before the multi-layer layer.
(3) The image processing device according to (2), wherein
The pre-recorded two-dimensional projection unit generates the same number of layers as the maximum number of pieces of data of each position of the pre-recorded 3D data from the projection surface as viewed from the projection surface in the depth direction.
(4) The image processing device according to any one of (1) to (3), wherein
Further, the encoding unit is configured to project the 3D data which has been projected by the pre-recorded two-dimensional projection unit to the pre-recorded two-dimensional plane.
(5) The image processing device according to (4), wherein
The preamble encoding unit encodes the position information, the attribute information, and the occupied map of the pre-recorded 3D data that has been projected onto each layer of the pre-recorded two-dimensional plane.
(6) The image processing device according to (4) or (5), wherein
Further, the present invention further includes a bit stream generation unit that generates a bit stream including information indicating a number of layers of a pre-recorded two-dimensional plane on which the 3D data is projected by the two-dimensional projection unit, and The precoding unit encodes the encoded data obtained by encoding the 2D plane.
(7) The image processing device according to any one of (4) to (6), wherein
Further, the packaging unit is configured to project a two-dimensional plane of the pre-recorded 3D data by a two-dimensional projection unit, and package it into a video frame;
The preamble encoding unit is configured to encode a pre-recorded video frame in which a pre-recorded two-dimensional plane is packaged by a pre-packaging unit.
(8) The image processing device according to any one of (1) to (7), wherein
The pre-recorded 2D projection unit projects the 3D data of the predecessor into the pre-recorded 2D plane according to each field.
(9) The image processing device according to any one of (1) to (8) wherein
The pre-recorded 3D data is a point cloud.
(10) An image processing method for projecting data of each position included in a 3D material of a three-dimensional structure onto a two-dimensional plane of a complex layer.

(11) 一種影像處理裝置,係具備:
3維投影部,係將已被投影至圖層數資訊所表示之圖層數之2維平面上的3D資料的所有的每一位置之資料,投影至3維空間。
(12) 如(11)所記載之影像處理裝置,其中,
還具備:抽出部,係將位元串流中所含之前記圖層數資訊,予以抽出;
前記3維投影部係被構成為,將已被前記抽出部所抽出之前記圖層數資訊所表示之圖層數之前記2維平面上已被投影的前記3D資料的所有的前記每一位置之資料,投影至前記3維空間。
(13) 如(12)所記載之影像處理裝置,其中,
還具備:解碼部,係將前記位元串流中所含之已被投影至前記2維平面上的前記3D資料之編碼資料,予以解碼;
前記3維投影部係被構成為,將藉由前記解碼部將前記編碼資料予以解碼所得的,已被投影至前記2維平面上的前記3D資料的所有的前記每一位置之資料,投影至前記3維空間。
(14) 如(13)所記載之影像處理裝置,其中,
前記解碼部,係將已被投影至前記2維平面之各圖層的前記3D資料的位置資訊、屬性資訊、及佔用地圖之每一者的編碼資料,予以解碼。
(15) 如(13)或(14)所記載之影像處理裝置,其中,
還具備:解包部,係將藉由前記解碼部把前記編碼資料進行解碼所得到的,已被投影至前記2維平面的前記3D資料所被打包而成的視訊畫格,予以解包;
前記3維投影部係被構成為,將藉由前記解包部把視訊畫格進行解包所得到的,已被投影至前記2維平面上的前記3D資料的所有的前記每一位置之資料,投影至前記3維空間。
(16) 如(11)乃至(15)之任一項所記載之影像處理裝置,其中,
前記3維投影部,係將按照所定之每一領域而被投影至前記2維平面的前記3D資料,投影至前記3維空間。
(17) 如(11)乃至(16)之任一項所記載之影像處理裝置,其中,
前記3D資料,係為點雲。
(18) 一種影像處理方法,係
將已被投影至圖層數資訊所表示之圖層數之2維平面上的3D資料的所有的每一位置之資料,投影至3維空間。
(11) An image processing apparatus comprising:
The three-dimensional projection unit projects data of each position of the 3D data on the two-dimensional plane that has been projected onto the number of layers indicated by the layer number information into a three-dimensional space.
(12) The image processing device according to (11), wherein
It also has a drawing unit for extracting the information of the number of layers in the bit stream which is included in the bit stream;
The pre-recorded three-dimensional projection unit is configured to record all the pre-recorded positions of the pre-recorded 3D data that have been projected on the two-dimensional plane before the number of layers indicated by the number of layers before the extracted portion is extracted. The data is projected to the front 3D space.
(13) The image processing device according to (12), wherein
Further, the decoding unit further includes: decoding the encoded data of the pre-recorded 3D data included in the pre-recorded bit stream that has been projected onto the pre-recorded two-dimensional plane;
The pre-recorded three-dimensional projection unit is configured to decode the pre-recorded data by the pre-decoding unit, and project the data onto all the pre-recorded positions of the pre-recorded 3D data on the two-dimensional plane. Pre-record 3D space.
(14) The image processing device according to (13), wherein
The predecessor decoding unit decodes the positional information, the attribute information, and the coded data of each of the occupied maps of the 3D data of each layer that has been projected onto each of the pre-recorded two-dimensional planes.
(15) The image processing device according to (13) or (14), wherein
Further, the unpacking unit is obtained by decoding the preamble encoded data by the pre-recording unit, and the video frame that has been projected onto the pre-recorded 3D data of the pre-recorded two-dimensional plane is unpacked;
The pre-recorded three-dimensional projection unit is configured to unpack the video frame by the pre-decoding unit, and project the data into all the pre-recorded positions of the pre-recorded 3D data on the two-dimensional plane. Projected to the front 3D space.
(16) The image processing device according to any one of (11) to (15), wherein
The 3D projection unit is projected to the 3D space of the pre-recorded 2D plane in accordance with each of the defined fields.
(17) The image processing device according to any one of (11) to (16), wherein
The pre-recorded 3D data is a point cloud.
(18) An image processing method for projecting data of each position of 3D data on a two-dimensional plane that has been projected onto the number of layers indicated by the layer number information into a three-dimensional space.

(21) 一種影像處理裝置,係具備:
2維投影部,係將表示3維結構之3D資料中所含之每一位置之資料投影至2維平面,並在前記2維平面之前記每一位置之資料為不存在的位置上,設置表示前記每一位置之資料為不存在的所定之值。
(22) 如(21)所記載之影像處理裝置,其中,
前記所定之值,係為已被預先決定之固定值。
(23) 如(21)所記載之影像處理裝置,其中,
前記所定之值,係比前記3D資料之縱深之最大值還大的值。
(24) 如(21)乃至(23)之任一項所記載之影像處理裝置,其中,
還具備:編碼部,係將已被前記2維投影部投影至前記2維平面的前記3D資料,予以編碼。
(25) 如(24)所記載之影像處理裝置,其中,
前記編碼部,係將已被投影至前記2維平面的前記3D資料的位置資訊、屬性資訊、及佔用地圖,分別予以編碼。
(26) 如(24)或(25)所記載之影像處理裝置,其中,
還具備:位元串流生成部,係生成位元串流,其中含有:表示前記所定之值的資訊、和藉由前記編碼部將前記2維平面予以編碼所得之編碼資料。
(27) 如(24)乃至(26)之任一項所記載之影像處理裝置,其中,
還具備:打包部,係將藉由前記2維投影部而被投影了前記3D資料的前記2維平面,打包成為視訊畫格;
前記編碼部係被構成為,將藉由前記打包部把前記2維平面打包而成的前記視訊畫格,予以編碼。
(28) 如(21)乃至(27)之任一項所記載之影像處理裝置,其中,
前記2維投影部,係將前記3D資料,按照所定之每一領域地投影至前記2維平面。
(29) 如(21)乃至(28)之任一項所記載之影像處理裝置,其中,
前記3D資料,係為點雲。
(30) 一種影像處理方法,係
將表示3維結構之3D資料中所含之每一位置之資料投影至2維平面,並在前記2維平面之前記每一位置之資料為不存在的位置上,設置表示前記每一位置之資料為不存在的所定之值。
(21) An image processing apparatus comprising:
The two-dimensional projection unit projects data of each position included in the 3D data of the three-dimensional structure onto a two-dimensional plane, and records the position of each position before the two-dimensional plane as a non-existent position. Indicates that the information in each of the preceding positions is the value that does not exist.
(22) The image processing device according to (21), wherein
The value stated in the preceding paragraph is a fixed value that has been predetermined.
(23) The image processing device according to (21), wherein
The value set in the preceding paragraph is a value larger than the maximum value of the depth of the 3D data.
(24) The image processing device according to any one of (21) to (23), wherein
Further, the encoding unit is configured to project the 3D data which has been projected by the pre-recorded two-dimensional projection unit to the pre-recorded two-dimensional plane.
(25) The image processing device according to (24), wherein
The preamble encoding unit encodes the position information, the attribute information, and the occupied map of the 3D data that has been projected onto the pre-recorded 2D plane.
(26) The image processing device according to (24) or (25), wherein
Further, the present invention further includes a bit stream generating unit that generates a bit stream including: information indicating a value set in the foregoing, and coded data obtained by encoding the pre-recorded two-dimensional plane by the pre-encoding unit.
(27) The image processing device according to any one of (24) to (26), wherein
Further, the packaging unit is configured to project a two-dimensional plane of the pre-recorded 3D data by a two-dimensional projection unit, and package it into a video frame;
The preamble encoding unit is configured to encode a pre-recorded video frame in which a pre-recorded two-dimensional plane is packaged by a pre-packaging unit.
(28) The image processing device according to any one of (21) to (27), wherein
The pre-recorded 2D projection unit projects the 3D data of the predecessor into the pre-recorded 2D plane according to each field.
(29) The image processing device according to any one of (21) to (28), wherein
The pre-recorded 3D data is a point cloud.
(30) An image processing method for projecting data of each position included in a 3D data representing a 3-dimensional structure onto a 2-dimensional plane, and recording the information of each position before the 2-dimensional plane as a non-existent position On the top, the data indicating that each position of the preceding record is a non-existent value.

(31) 一種影像處理裝置,係具備:
3維投影部,係在已被投影至2維平面上的表示3維結構之3D資料中所含之每一位置之資料之內,將表示前記每一位置之資料為不存在的所定之值的資料以外之資料,投影至3維空間。
(32) 如(31)所記載之影像處理裝置,其中,
前記所定之值,係為已被預先決定之固定值。
(33) 如(31)所記載之影像處理裝置,其中,
前記所定之值,係比前記3D資料之縱深之最大值還大的值。
(34) 如(31)乃至(33)之任一項所記載之影像處理裝置,其中,
還具備:抽出部,係將位元串流中所含之表示前記所定之值的資訊,予以抽出;
前記3維投影部,係被構成為,在前記3D資料中所含之前記每一位置之資料之內,將已被前記抽出部所抽出之前記資訊所表示的前記所定之值的資料以外之資料,投影至3維空間。
(35) 如(34)所記載之影像處理裝置,其中,
還具備:解碼部,係將前記位元串流中所含之已被投影至前記2維平面上的前記3D資料之編碼資料,予以解碼;
前記3維投影部,係被構成為,在藉由前記解碼部把前記編碼資料進行解碼所得到的,已被投影至前記2維平面的前記3D資料中所含之前記每一位置之資料之內,將已被前記抽出部所抽出之前記資訊所表示的前記所定之值的資料以外之資料,投影至3維空間。
(36) 如(35)所記載之影像處理裝置,其中,
前記解碼部,係將已被投影至前記2維平面之各圖層的前記3D資料的位置資訊、屬性資訊、及佔用地圖之每一者的編碼資料,予以解碼。
(37) 如(35)或(36)所記載之影像處理裝置,其中,
還具備:解包部,係將藉由前記解碼部把前記編碼資料進行解碼所得到的,已被投影至前記2維平面的前記3D資料所被打包而成的視訊畫格,予以解包;
前記3維投影部,係被構成為,在藉由前記解包部把視訊畫格進行解包所得到的,已被投影至前記2維平面的前記3D資料中所含之前記每一位置之資料之內,將已被前記抽出部所抽出之前記資訊所表示的前記所定之值的資料以外之資料,投影至3維空間。
(38) 如(31)乃至(37)之任一項所記載之影像處理裝置,其中,
前記3維投影部,係將按照所定之每一領域而被投影至前記2維平面的前記3D資料,投影至前記3維空間。
(39) 如(31)乃至(38)之任一項所記載之影像處理裝置,其中,
前記3D資料,係為點雲。
(40) 一種影像處理方法,係
在已被投影至2維平面上的表示3維結構之3D資料中所含之每一位置之資料之內,將表示前記每一位置之資料為不存在的所定之值的資料以外之資料,投影至3維空間。
(31) An image processing apparatus comprising:
The 3-dimensional projection unit, within the data of each position included in the 3D data representing the 3-dimensional structure that has been projected onto the 2-dimensional plane, will indicate that the data of each position in the preceding record is a non-existent value. The data other than the data is projected into the 3-dimensional space.
(32) The image processing device according to (31), wherein
The value stated in the preceding paragraph is a fixed value that has been predetermined.
(33) The image processing device according to (31), wherein
The value set in the preceding paragraph is a value larger than the maximum value of the depth of the 3D data.
(34) The image processing device according to any one of (31) to (33), wherein
Further, the extracting unit extracts information indicating a value specified in the preceding stream contained in the bit stream;
The pre-recorded three-dimensional projection unit is configured to include, in the data of each position previously included in the pre-recorded 3D data, data that has been extracted by the pre-recording unit and the value of the pre-recorded information indicated by the information. Data, projected to 3D space.
(35) The image processing device according to (34), wherein
Further, the decoding unit further includes: decoding the encoded data of the pre-recorded 3D data included in the pre-recorded bit stream that has been projected onto the pre-recorded two-dimensional plane;
The pre-recorded three-dimensional projection unit is configured to decode the pre-recorded encoded data by the pre-decoding unit, and has been projected onto the pre-recorded 3D data of the pre-recorded two-dimensional plane. In the case, the data other than the data of the value indicated in the previous record indicated by the pre-recording unit is projected to the three-dimensional space.
(36) The image processing device according to (35), wherein
The predecessor decoding unit decodes the positional information, the attribute information, and the coded data of each of the occupied maps of the 3D data of each layer that has been projected onto each of the pre-recorded two-dimensional planes.
(37) The image processing device of (35) or (36), wherein
Further, the unpacking unit is obtained by decoding the preamble encoded data by the pre-recording unit, and the video frame that has been projected onto the pre-recorded 3D data of the pre-recorded two-dimensional plane is unpacked;
The pre-recorded three-dimensional projection unit is configured to unpack the video frame by the pre-decoding unit, and has been projected to the pre-recorded 3D plane of the pre-recorded 2D plane. Within the data, the data other than the data of the previous value indicated by the previous information extracted by the pre-recording department is projected into the 3-dimensional space.
(18) The image processing device according to any one of (31) to (37), wherein
The 3D projection unit is projected to the 3D space of the pre-recorded 2D plane in accordance with each of the defined fields.
(39) The image processing device according to any one of (31) to (38), wherein
The pre-recorded 3D data is a point cloud.
(40) An image processing method in which data representing each position included in a 3D structure of a 3D structure that has been projected onto a 2-dimensional plane is non-existent Projects other than the data of the specified value are projected into the 3-dimensional space.

(41) 一種影像處理裝置,係具備:
2維投影部,係將表示3維結構之3D資料的每一位置之資料,按照3維空間之所定之每一領域地投影至2維平面之際,將對前記每一領域所被設定之用來限制可投影至1圖層的表示3維結構之3D資料的每一位置之資料之縱深方向之範圍的縱深參數所表示的前記縱深方向之範圍內的前記每一位置之資料,投影至前記2維平面。
(42) 如(41)所記載之影像處理裝置,其中,
還具備:編碼部,係將已被前記2維投影部投影至前記2維平面的前記3D資料,予以編碼。
(43) 如(42)所記載之影像處理裝置,其中,
前記編碼部,係將已被投影至前記2維平面的前記3D資料的位置資訊、屬性資訊、及佔用地圖,分別予以編碼。
(44) 如(42)或(43)所記載之影像處理裝置,其中,
還具備:位元串流生成部,係生成位元串流,其中含有:對前記每一領域所被設定之前記縱深參數、和藉由前記編碼部將前記2維平面予以編碼所得之編碼資料。
(45) 如(42)乃至(44)之任一項所記載之影像處理裝置,其中,
還具備:打包部,係將藉由前記2維投影部而被投影了前記3D資料的前記2維平面,打包成為視訊畫格;
前記編碼部係被構成為,將藉由前記打包部把前記2維平面打包而成的前記視訊畫格,予以編碼。
(46) 如(41)乃至(45)之任一項所記載之影像處理裝置,其中,
前記3D資料,係為點雲。
(47) 一種影像處理方法,係
將表示3維結構之3D資料的每一位置之資料,按照3維空間之所定之每一領域地投影至2維平面之際,將對前記每一領域所被設定之用來限制可投影至1圖層的表示3維結構之3D資料的每一位置之資料之縱深方向之範圍的縱深參數所表示的前記縱深方向之範圍內的前記每一位置之資料,投影至前記2維平面。
(41) An image processing apparatus comprising:
The two-dimensional projection unit sets the data of each position of the 3D data representing the three-dimensional structure to each of the fields defined in the three-dimensional space to the two-dimensional plane. It is used to limit the data of each position in the range of the depth in the depth direction indicated by the depth parameter of the depth direction of the data of each position of the 3D data which can be projected to one layer, and project to the pre-record 2D plane.
(42) The image processing device according to (41), wherein
Further, the encoding unit is configured to project the 3D data which has been projected by the pre-recorded two-dimensional projection unit to the pre-recorded two-dimensional plane.
(43) The image processing device according to (42), wherein
The preamble encoding unit encodes the position information, the attribute information, and the occupied map of the 3D data that has been projected onto the pre-recorded 2D plane.
(44) The image processing device of (42) or (43), wherein
A bit stream generation unit is further configured to generate a bit stream, wherein: the depth parameter selected before each field of the preceding record and the coded data obtained by encoding the previous two-dimensional plane by the preceding coding unit .
(45) The image processing device according to any one of (42) to (44), wherein
Further, the packaging unit is configured to project a two-dimensional plane of the pre-recorded 3D data by a two-dimensional projection unit, and package it into a video frame;
The preamble encoding unit is configured to encode a pre-recorded video frame in which a pre-recorded two-dimensional plane is packaged by a pre-packaging unit.
(4) The image processing device according to any one of (41) to (45), wherein
The pre-recorded 3D data is a point cloud.
(47) An image processing method in which each piece of data representing a 3D structure of a 3D structure is projected onto a 2D plane in accordance with a field defined by a 3D space, and will be recorded in each field of the predecessor. It is set to limit the data of each position in the range of the depth direction indicated by the depth parameter in the depth direction of the data indicating the depth direction of the data of each position of the 3D structure which can be projected to one layer, Project to the front 2D plane.

(51) 一種影像處理裝置,係具備:
3維投影部,係按照3維空間之所定之每一領域,將已被投影至2維平面上的表示3維結構之3D資料中所含之每一位置之資料投影至前記3維空間之際,是將前記每一位置之資料,投影至前記3維空間的,對前記每一領域所被設定之用來限制可投影至1圖層的前記3D資料的前記每一位置之資料的縱深方向之範圍的縱深參數所表示的前記縱深方向之範圍內。
(52) 如(51)所記載之影像處理裝置,其中,
還具備:抽出部,係將位元串流中所含之前記縱深參數,予以抽出;
前記3維投影部,係被構成為,將前記3D資料中所含之前記每一位置之資料,投影至已被前記抽出部所抽出之前記縱深參數所表示的前記縱深方向之範圍內。
(53) 如(52)所記載之影像處理裝置,其中,
還具備:解碼部,係將前記位元串流中所含之已被投影至前記2維平面上的前記3D資料之編碼資料,予以解碼;
前記3維投影部係被構成為,將藉由前記解碼部將前記編碼資料予以解碼所得的,已被投影至前記2維平面上的前記3D資料的前記每一位置之資料,投影至已被前記抽出部所抽出之前記縱深參數所表示的前記縱深方向之範圍內。
(54) 如(53)所記載之影像處理裝置,其中,
前記解碼部,係將已被投影至前記2維平面之各圖層的前記3D資料的位置資訊、屬性資訊、及佔用地圖之每一者的編碼資料,予以解碼。
(55) 如(53)或(54)所記載之影像處理裝置,其中,
還具備:解包部,係將藉由前記解碼部把前記編碼資料進行解碼所得到的,已被投影至前記2維平面的前記3D資料所被打包而成的視訊畫格,予以解包;
前記3維投影部係被構成為,將藉由前記解包部把視訊畫格進行解包所得的,已被投影至前記2維平面上的前記3D資料的前記每一位置之資料,投影至已被前記抽出部所抽出之前記縱深參數所表示的前記縱深方向之範圍內。
(56) 如(51)乃至(55)之任一項所記載之影像處理裝置,其中,
前記3D資料,係為點雲。
(57) 一種影像處理方法,係
按照3維空間之所定之每一領域,將已被投影至2維平面上的表示3維結構之3D資料中所含之每一位置之資料投影至前記3維空間之際,是將前記每一位置之資料,投影至前記3維空間的,對前記每一領域所被設定之用來限制可投影至1圖層的前記3D資料的前記每一位置之資料的縱深方向之範圍的縱深參數所表示的前記縱深方向之範圍內。
(51) An image processing apparatus comprising:
The three-dimensional projection unit projects the data of each position included in the 3D data representing the three-dimensional structure onto the two-dimensional space according to each field defined by the three-dimensional space to the pre-recorded three-dimensional space. In the third dimension, the data of each position of the pre-record is projected to the depth direction of the data of each position of the pre-recorded 3D data that can be projected to the first layer in each field. The depth parameter of the range is within the range of the depth direction of the preceding paragraph.
(52) The image processing device according to (51), wherein
The utility model further has: an extracting unit, which extracts a previous depth parameter included in the bit stream;
The pre-recorded three-dimensional projection unit is configured to project the data of each position previously included in the 3D data into the range of the depth direction indicated by the depth parameter before being extracted by the pre-recording unit.
(53) The image processing device according to (52), wherein
Further, the decoding unit further includes: decoding the encoded data of the pre-recorded 3D data included in the pre-recorded bit stream that has been projected onto the pre-recorded two-dimensional plane;
The pre-recorded three-dimensional projection unit is configured to decode the data of the pre-recorded data by the pre-recording unit, and project the data into the pre-recorded 3D data of the pre-recorded two-dimensional plane. The depth of the preceding depth direction indicated by the depth parameter is recorded before the extraction of the pre-extraction unit.
(54) The image processing device according to (53), wherein
The predecessor decoding unit decodes the positional information, the attribute information, and the coded data of each of the occupied maps of the 3D data of each layer that has been projected onto each of the pre-recorded two-dimensional planes.
(55) The image processing device according to (53) or (54), wherein
Further, the unpacking unit is obtained by decoding the preamble encoded data by the pre-recording unit, and the video frame that has been projected onto the pre-recorded 3D data of the pre-recorded two-dimensional plane is unpacked;
The pre-recorded three-dimensional projection unit is configured to unpack the video frame by the pre-decoding unit, and project the data into the pre-recorded 3D data of the pre-recorded 2D plane. It has been extracted by the pre-recording unit before the depth direction indicated by the depth parameter.
(56) The image processing device according to any one of (51) to (55), wherein
The pre-recorded 3D data is a point cloud.
(57) An image processing method for projecting data of each position included in a 3D data representing a 3-dimensional structure onto a 2-dimensional plane in accordance with each field defined by a 3-dimensional space to a pre-record 3 In the dimension space, the data of each position of the predecessor is projected into the pre-recorded 3-dimensional space, and the information set in each field of the pre-record is used to limit the pre-recorded 3D data that can be projected to the 1 layer. The depth parameter of the range of the depth direction is within the range of the depth direction indicated in the preceding paragraph.

100‧‧‧編碼裝置100‧‧‧ coding device

111‧‧‧斑塊分解部 111‧‧‧Potment Decomposition Department

112‧‧‧打包部 112‧‧‧Packing Department

113‧‧‧輔助斑塊資訊壓縮部 113‧‧‧Auxiliary Plaque Information Compression Department

114‧‧‧視訊編碼部 114‧‧•Video Coding Department

115‧‧‧視訊編碼部 115‧‧•Video Coding Department

116‧‧‧OMap編碼部 116‧‧‧OMap Encoding Department

117‧‧‧多工器 117‧‧‧Multiplexer

151‧‧‧法線方向推定部 151‧‧‧ Normal direction estimation

152‧‧‧分段初期設定部 152‧‧ Section initial setting

153‧‧‧分段更新部 153‧‧ Section Update Department

154‧‧‧2維投影部 154‧‧2D Dimensional Projector

155‧‧‧像素分布解析部 155‧‧‧Pixel Distribution Analysis Department

200‧‧‧解碼裝置 200‧‧‧ decoding device

211‧‧‧解多工器 211‧‧ ‧ multiplexer

212‧‧‧輔助斑塊資訊解碼部 212‧‧‧Auxiliary plaque information decoding department

213‧‧‧視訊解碼部 213‧‧‧Video Decoding Department

214‧‧‧視訊解碼部 214‧‧‧Video Decoding Department

215‧‧‧OMap解碼部 215‧‧‧OMap Decoding Department

216‧‧‧解包部 216‧‧·Unpacking Department

217‧‧‧3D重新建構部 217‧‧3D reconstruction department

251‧‧‧3維投影部 251‧‧3D projection department

252‧‧‧像素分布解析部 252‧‧‧Pixel Distribution Analysis Department

253‧‧‧逆分段更新部 253‧‧‧ Reverse Segment Update Department

254‧‧‧逆分段初期設定部 254‧‧‧Inverse Segmentation Initial Setting Department

255‧‧‧逆法線方向推定部 255‧‧‧Inverse normal direction estimation

900‧‧‧電腦 900‧‧‧ computer

901‧‧‧CPU 901‧‧‧CPU

902‧‧‧ROM 902‧‧‧ROM

903‧‧‧RAM 903‧‧‧RAM

904‧‧‧匯流排 904‧‧ ‧ busbar

910‧‧‧輸出入介面 910‧‧‧Import interface

911‧‧‧輸入部 911‧‧‧ Input Department

912‧‧‧輸出部 912‧‧‧Output Department

913‧‧‧記憶部 913‧‧‧Memory Department

914‧‧‧通訊部 914‧‧‧Communication Department

915‧‧‧驅動機 915‧‧‧ drive machine

921‧‧‧可移除式媒體 921‧‧‧Removable media

[圖1]點雲之例子的說明圖。[Fig. 1] An explanatory diagram of an example of a point cloud.

[圖2]視訊基礎方式之概要之例子的說明圖。 [Fig. 2] An explanatory diagram of an example of an outline of a video basic mode.

[圖3]於各實施形態所說明的本技術之總整理的圖。 Fig. 3 is a view showing the overall arrangement of the present technology described in each embodiment.

[圖4]編碼裝置的主要構成例的區塊圖。 Fig. 4 is a block diagram showing a main configuration example of an encoding device.

[圖5]斑塊分解部的主要構成例的說明圖。 Fig. 5 is an explanatory diagram of a main configuration example of a plaque decomposition unit.

[圖6]解碼裝置的主要構成例的區塊圖。 Fig. 6 is a block diagram showing a main configuration example of a decoding device.

[圖7]3D重新建構部的主要構成例的區塊圖。 [Fig. 7] A block diagram of a main configuration example of the 3D reconstruction unit.

[圖8]先前的2維投影之樣子之例子的圖示。 [Fig. 8] An illustration of an example of a previous two-dimensional projection.

[圖9]適用了本技術的2維投影之樣子之例子的圖示。 [Fig. 9] An illustration of an example of a two-dimensional projection to which the present technique is applied.

[圖10]編碼處理之流程例的說明用流程圖。 [Fig. 10] A flowchart for explaining an example of the flow of the encoding process.

[圖11]斑塊分解處理之流程之例子的說明用流程圖。 [Fig. 11] A flow chart for explaining an example of the flow of the plaque decomposition processing.

[圖12]2維投影處理之流程之例子的說明用流程圖。 [Fig. 12] A flow chart for explaining an example of the flow of the two-dimensional projection processing.

[圖13]解碼處理之流程例的說明用流程圖。 [Fig. 13] A flowchart for explaining an example of the flow of the decoding process.

[圖14]點雲重新建構處理之流程之例子的說明用流程圖。 [Fig. 14] A flow chart for explaining an example of the flow of the point cloud reconstruction process.

[圖15]3維投影處理之流程之例子的說明用流程圖。 [Fig. 15] A flow chart for explaining an example of the flow of the three-dimensional projection processing.

[圖16]先前的2維投影之樣子之例子的圖示。 [Fig. 16] An illustration of an example of a previous two-dimensional projection.

[圖17]適用了本技術的2維投影之樣子之例子的圖示。 [Fig. 17] An illustration of an example of a two-dimensional projection to which the present technique is applied.

[圖18]2維投影處理之流程之例子的說明用流程圖。 [Fig. 18] A flow chart for explaining an example of the flow of the two-dimensional projection processing.

[圖19]3維投影處理之流程之例子的說明用流程圖。 [Fig. 19] A flow chart for explaining an example of the flow of the three-dimensional projection processing.

[圖20]先前的2維投影之樣子之例子的圖示。 [Fig. 20] A diagram showing an example of the appearance of the previous two-dimensional projection.

[圖21]適用了本技術的2維投影之樣子之例子的圖示。 [Fig. 21] An illustration of an example of a two-dimensional projection to which the present technique is applied.

[圖22]2維投影處理之流程之例子的說明用流程圖。 [Fig. 22] A flow chart for explaining an example of the flow of the two-dimensional projection processing.

[圖23]3維投影處理之流程之例子的說明用流程圖。 [Fig. 23] A flow chart for explaining an example of the flow of the three-dimensional projection processing.

[圖24]電腦的主要構成例的區塊圖。 [Fig. 24] A block diagram of a main configuration example of a computer.

Claims (20)

一種影像處理裝置,係具備: 2維投影部,係將表示3維結構之3D資料中所含之所有的每一位置之資料,投影至複數圖層之2維平面。An image processing device having: The two-dimensional projection unit projects data of each position included in the 3D data of the three-dimensional structure onto a two-dimensional plane of the complex layer. 如請求項1所記載之影像處理裝置,其中, 前記2維投影部,係將前記3D資料的從投影面來看位置是在縱深方向上重疊的前記每一位置之資料,投影至前記複數圖層之前記2維平面的彼此互異之圖層。The image processing device of claim 1, wherein In the second-dimensional projection unit, the position of the 3D data from the projection surface is the position of each position in the depth direction, and is projected on the two-dimensional plane before the multi-layer layer. 如請求項2所記載之影像處理裝置,其中, 前記2維投影部,係針對前記2維平面,生成與前記3D資料的從投影面來看位置是在縱深方向上重疊的前記每一位置之資料之最大筆數相同數量之圖層。The image processing device of claim 2, wherein The pre-recorded two-dimensional projection unit generates the same number of layers as the maximum number of pieces of data of each position of the pre-recorded 3D data from the projection surface as viewed from the projection surface in the depth direction. 如請求項1所記載之影像處理裝置,其中, 前記2維投影部,係在前記2維平面之前記每一位置之資料為不存在的位置上,設置表示前記每一位置之資料為不存在的所定之值。The image processing device of claim 1, wherein The pre-recorded two-dimensional projection unit is provided at a position where the data of each position is not present before the two-dimensional plane, and the data indicating that each position of the previous position is a non-existent value. 如請求項4所記載之影像處理裝置,其中, 前記所定之值,係為已被預先決定之固定值。The image processing device of claim 4, wherein The value stated in the preceding paragraph is a fixed value that has been predetermined. 如請求項4所記載之影像處理裝置,其中, 前記所定之值,係比前記3D資料之縱深之最大值還大的值。The image processing device of claim 4, wherein The value set in the preceding paragraph is a value larger than the maximum value of the depth of the 3D data. 如請求項1所記載之影像處理裝置,其中, 前記2維投影部,係在將前記3D資料的前記每一位置之資料,按照3維空間之所定之每一領域地投影至2維平面之際,將對前記每一領域所被設定之用來限制可投影至1圖層的前記3D資料的前記每一位置之資料的縱深方向之範圍的縱深參數所表示的前記縱深方向之範圍內的前記每一位置之資料,投影至前記2維平面。The image processing device of claim 1, wherein The 2D projection section of the pre-recorded 3D data is used to map each field of the pre-recording to the 2D plane according to the position of each position in the 3D space. The data of each position in the range of the preceding depth direction indicated by the depth parameter indicated by the depth parameter in the depth direction of the data of each position of the pre-recorded 3D data which can be projected to the first layer is projected to the pre-recorded two-dimensional plane. 如請求項1所記載之影像處理裝置,其中, 還具備:編碼部,係將已被前記2維投影部投影至前記2維平面的前記3D資料,予以編碼。The image processing device of claim 1, wherein Further, the encoding unit is configured to project the 3D data which has been projected by the pre-recorded two-dimensional projection unit to the pre-recorded two-dimensional plane. 如請求項8所記載之影像處理裝置,其中, 還具備:位元串流生成部,係生成位元串流,其中含有:表示藉由前記2維投影部而被投影了前記3D資料的前記2維平面之圖層數的資訊、和藉由前記編碼部將前記2維平面予以編碼所得之編碼資料。The image processing device of claim 8, wherein Further, the present invention further includes a bit stream generation unit that generates a bit stream including information indicating a number of layers of a pre-recorded two-dimensional plane on which the 3D data is projected by the two-dimensional projection unit, and The precoding unit encodes the encoded data obtained by encoding the 2D plane. 如請求項1所記載之影像處理裝置,其中, 前記3D資料,係為點雲。The image processing device of claim 1, wherein The pre-recorded 3D data is a point cloud. 一種影像處理方法,係 將表示3維結構之3D資料中所含之所有的每一位置之資料,投影至複數圖層之2維平面。An image processing method The data representing each of the positions included in the 3D data of the 3-dimensional structure is projected onto the 2-dimensional plane of the complex layer. 一種影像處理裝置,係具備: 3維投影部,係將已被投影至圖層數資訊所表示之圖層數之2維平面上的3D資料的所有的每一位置之資料,投影至3維空間。An image processing device having: The three-dimensional projection unit projects data of each position of the 3D data on the two-dimensional plane that has been projected onto the number of layers indicated by the layer number information into a three-dimensional space. 如請求項12所記載之影像處理裝置,其中, 前記3維投影部,係在前記3D資料中所含之前記每一位置之資料之內,將表示前記每一位置之資料為不存在的所定之值的資料以外之資料,投影至前記3維空間。The image processing device of claim 12, wherein The 3D projection part of the pre-recorded 3D data is included in the data of each position before the 3D data, and the data indicating the position of each position in the previous record is the data other than the value of the non-existent value, and is projected to the pre-record 3D. space. 如請求項13所記載之影像處理裝置,其中, 前記所定之值,係為已被預先決定之固定值。The image processing device of claim 13, wherein The value stated in the preceding paragraph is a fixed value that has been predetermined. 如請求項13所記載之影像處理裝置,其中, 前記所定之值,係比前記3D資料之縱深之最大值還大的值。The image processing device of claim 13, wherein The value set in the preceding paragraph is a value larger than the maximum value of the depth of the 3D data. 如請求項12所記載之影像處理裝置,其中, 前記3維投影部,係在按照前記3維空間之所定之每一領域地,將前記3D資料中所含之前記每一位置之資料投影至前記3維空間之際,將前記每一位置之資料,投影至前記3維空間的,對前記每一領域所被設定之用來限制可投影至1圖層的表示3維結構之3D資料的每一位置之資料之縱深方向之範圍的縱深參數所表示的前記縱深方向之範圍內。The image processing device of claim 12, wherein The 3D projection part of the pre-recorded 3D space is used to project each position of the previous 3D data into the 3D space in the pre-recorded 3D space. The data is projected into the pre-recorded 3-dimensional space, and the depth parameter of the range of the depth direction of the data of each position of the 3D data representing the 3-dimensional structure that can be projected to the first layer is limited. The front of the representation is in the depth direction. 如請求項12所記載之影像處理裝置,其中, 還具備:抽出部,係將位元串流中所含之前記圖層數資訊,予以抽出; 前記3維投影部係被構成為,將已被前記抽出部所抽出之前記圖層數資訊所表示之圖層數之前記2維平面上已被投影的前記3D資料的所有的前記每一位置之資料,投影至前記3維空間。The image processing device of claim 12, wherein It also has a drawing unit for extracting the information of the number of layers in the bit stream which is included in the bit stream; The pre-recorded three-dimensional projection unit is configured to record all the pre-recorded positions of the pre-recorded 3D data that have been projected on the two-dimensional plane before the number of layers indicated by the number of layers before the extracted portion is extracted. The data is projected to the front 3D space. 如請求項17所記載之影像處理裝置,其中, 還具備:解碼部,係將前記位元串流中所含之已被投影至前記2維平面上的前記3D資料之編碼資料,予以解碼; 前記3維投影部係被構成為,將藉由前記解碼部將前記編碼資料予以解碼所得的,已被投影至前記2維平面上的前記3D資料的所有的前記每一位置之資料,投影至前記3維空間。The image processing device of claim 17, wherein Further, the decoding unit further includes: decoding the encoded data of the pre-recorded 3D data included in the pre-recorded bit stream that has been projected onto the pre-recorded two-dimensional plane; The pre-recorded three-dimensional projection unit is configured to decode the pre-recorded data by the pre-decoding unit, and project the data onto all the pre-recorded positions of the pre-recorded 3D data on the two-dimensional plane. Pre-record 3D space. 如請求項12所記載之影像處理裝置,其中, 前記3D資料,係為點雲。The image processing device of claim 12, wherein The pre-recorded 3D data is a point cloud. 一種影像處理方法,係 將已被投影至圖層數資訊所表示之圖層數之2維平面上的3D資料的所有的每一位置之資料,投影至3維空間。An image processing method The data of each position of the 3D material on the two-dimensional plane that has been projected onto the number of layers indicated by the layer number information is projected into the three-dimensional space.
TW107147408A 2018-01-16 2018-12-27 Image processing device and method TWI815842B (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2018004983 2018-01-16
JP2018-004983 2018-12-20

Publications (2)

Publication Number Publication Date
TW201933284A true TW201933284A (en) 2019-08-16
TWI815842B TWI815842B (en) 2023-09-21

Family

ID=67301752

Family Applications (1)

Application Number Title Priority Date Filing Date
TW107147408A TWI815842B (en) 2018-01-16 2018-12-27 Image processing device and method

Country Status (15)

Country Link
US (1) US11721048B2 (en)
EP (1) EP3742401A4 (en)
JP (1) JP7327166B2 (en)
KR (2) KR102596507B1 (en)
CN (1) CN111566702B (en)
AU (1) AU2019210233B2 (en)
BR (1) BR112020013599A2 (en)
CA (1) CA3087980A1 (en)
MX (1) MX2020007416A (en)
PH (1) PH12020500587A1 (en)
RU (1) RU2020122782A (en)
SG (1) SG11202005374UA (en)
TW (1) TWI815842B (en)
WO (1) WO2019142666A1 (en)
ZA (1) ZA202005038B (en)

Families Citing this family (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11949889B2 (en) 2018-01-19 2024-04-02 Interdigital Vc Holdings, Inc. Processing a point cloud
EP3756170A4 (en) * 2018-02-23 2021-11-17 Nokia Technologies Oy Encoding and decoding of volumetric video
CN111971967A (en) * 2018-04-11 2020-11-20 交互数字Vc控股公司 Method and apparatus for encoding/decoding a point cloud representing a 3D object
JP7446234B2 (en) * 2018-04-11 2024-03-08 インターデジタル ヴイシー ホールディングス, インコーポレイテッド Method for encoding depth values of a set of 3D points once orthogonally projected onto at least one image region of a projection plane
EP3777180B1 (en) 2018-04-11 2023-08-09 InterDigital VC Holdings, Inc. A method and apparatus for encoding/decoding a point cloud representing a 3d object
CN111937388A (en) * 2018-04-11 2020-11-13 索尼公司 Image processing apparatus and method
EP3554082A1 (en) 2018-04-11 2019-10-16 InterDigital VC Holdings, Inc. A method and device for coding the geometry of a point cloud
US11533505B2 (en) * 2018-07-11 2022-12-20 Interdigital Vc Holdings, Inc. Method for encoding/decoding texture of points of a point cloud
AU2019302075A1 (en) * 2018-07-11 2021-01-28 Sony Corporation Image processing device and method
JP2021182650A (en) * 2018-07-20 2021-11-25 ソニーグループ株式会社 Image processing device and method
JP7100523B2 (en) * 2018-07-27 2022-07-13 京セラ株式会社 Display devices, display systems and moving objects
US11956478B2 (en) * 2019-01-09 2024-04-09 Tencent America LLC Method and apparatus for point cloud chunking for improved patch packing and coding efficiency
SG11202109967SA (en) * 2019-03-12 2021-10-28 Huawei Tech Co Ltd Patch data unit coding and decoding for point-cloud data
KR20230153532A (en) * 2019-03-21 2023-11-06 엘지전자 주식회사 Method of encoding point cloud data, apparatus of encoding point cloud data, method of decoding point cloud data, and apparatus of decoding point cloud data
EP4042381A4 (en) * 2019-09-30 2022-11-02 ZTE Corporation File format for point cloud data
US20220303578A1 (en) * 2019-11-13 2022-09-22 Sony Group Corporation Image processing apparatus and method
JPWO2021193088A1 (en) * 2020-03-25 2021-09-30
KR102533985B1 (en) * 2021-05-21 2023-05-17 한국항공대학교산학협력단 System and method for restoration of point cloud

Family Cites Families (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3957620B2 (en) * 2001-11-27 2007-08-15 三星電子株式会社 Apparatus and method for representing a depth image-based 3D object
KR100450823B1 (en) * 2001-11-27 2004-10-01 삼성전자주식회사 Node structure for representing 3-dimensional objects using depth image
EP2180449A1 (en) * 2008-10-21 2010-04-28 Koninklijke Philips Electronics N.V. Method and device for providing a layered depth model of a scene
JP5480914B2 (en) * 2009-12-11 2014-04-23 株式会社トプコン Point cloud data processing device, point cloud data processing method, and point cloud data processing program
JP5522367B2 (en) 2009-12-13 2014-06-18 国立大学法人 千葉大学 Surface reproduction method and surface reproduction program
JP5161991B2 (en) * 2011-03-25 2013-03-13 株式会社東芝 Image processing device
WO2012141235A1 (en) * 2011-04-13 2012-10-18 株式会社トプコン Three-dimensional point group position data processing device, three-dimensional point group position data processing system, three-dimensional point group position data processing method and program
US10592490B2 (en) * 2011-11-25 2020-03-17 Interdigital Vc Holdings, Inc. Position coding based on spatial tree with duplicate points
US9075933B2 (en) * 2012-10-11 2015-07-07 Adobe Systems Incorporated 3D transformation of objects using 2D controls projected in 3D space and contextual face selections of a three dimensional bounding box
US20140204088A1 (en) * 2013-01-18 2014-07-24 Microsoft Corporation Surface codec using reprojection onto depth maps
US9715761B2 (en) * 2013-07-08 2017-07-25 Vangogh Imaging, Inc. Real-time 3D computer vision processing engine for object recognition, reconstruction, and analysis
CN105612742A (en) * 2013-10-14 2016-05-25 皇家飞利浦有限公司 Remapping a depth map for 3D viewing
US9465361B2 (en) 2014-03-31 2016-10-11 Disney Enterprises, Inc. Image based multiview multilayer holographic rendering algorithm
EP3143774A4 (en) * 2014-05-13 2018-04-25 PCP VR Inc. Method, system and apparatus for generation and playback of virtual reality multimedia
WO2016006728A1 (en) * 2014-07-08 2016-01-14 삼성전자 주식회사 Electronic apparatus and method for processing three-dimensional information using image
US20170214943A1 (en) * 2016-01-22 2017-07-27 Mitsubishi Electric Research Laboratories, Inc. Point Cloud Compression using Prediction and Shape-Adaptive Transforms
CN106934861B (en) * 2017-02-09 2020-06-26 深圳先进技术研究院 Object three-dimensional reconstruction method and device
CN107481313A (en) * 2017-08-18 2017-12-15 深圳市唯特视科技有限公司 A kind of dense three-dimensional object reconstruction method based on study available point cloud generation

Also Published As

Publication number Publication date
RU2020122782A (en) 2022-01-10
KR20230155019A (en) 2023-11-09
KR102596507B1 (en) 2023-11-01
EP3742401A1 (en) 2020-11-25
PH12020500587A1 (en) 2021-06-14
CN111566702A (en) 2020-08-21
MX2020007416A (en) 2020-09-14
CA3087980A1 (en) 2019-07-25
US20210174559A1 (en) 2021-06-10
AU2019210233A1 (en) 2020-07-23
US11721048B2 (en) 2023-08-08
TWI815842B (en) 2023-09-21
SG11202005374UA (en) 2020-07-29
JPWO2019142666A1 (en) 2021-01-28
JP7327166B2 (en) 2023-08-16
WO2019142666A1 (en) 2019-07-25
BR112020013599A2 (en) 2020-12-01
AU2019210233B2 (en) 2023-10-19
ZA202005038B (en) 2021-05-26
EP3742401A4 (en) 2020-12-09
CN111566702B (en) 2024-03-26
KR20200108833A (en) 2020-09-21

Similar Documents

Publication Publication Date Title
TW201933284A (en) Image processing device and method
JP7384159B2 (en) Image processing device and method
US11611774B2 (en) Image processing apparatus and image processing method for 3D data compression
WO2019198523A1 (en) Image processing device and method
US11699248B2 (en) Image processing apparatus and method
JP7331852B2 (en) Image processing device and method
TW201946449A (en) Image processing device and method
US11399189B2 (en) Image processing apparatus and method
CN114946179B (en) Point cloud data transmitting device, point cloud data transmitting method, point cloud data receiving device and point cloud data receiving method
JP7396302B2 (en) Image processing device and method
US11948337B2 (en) Image processing apparatus and method