TW201926258A - Three-dimensional point cloud tracking apparatus and method by recurrent neural network - Google Patents

Three-dimensional point cloud tracking apparatus and method by recurrent neural network Download PDF

Info

Publication number
TW201926258A
TW201926258A TW106143026A TW106143026A TW201926258A TW 201926258 A TW201926258 A TW 201926258A TW 106143026 A TW106143026 A TW 106143026A TW 106143026 A TW106143026 A TW 106143026A TW 201926258 A TW201926258 A TW 201926258A
Authority
TW
Taiwan
Prior art keywords
point cloud
dimensional point
memory
environment
computing
Prior art date
Application number
TW106143026A
Other languages
Chinese (zh)
Other versions
TWI657407B (en
Inventor
王志維
蔡岳廷
Original Assignee
財團法人資訊工業策進會
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 財團法人資訊工業策進會 filed Critical 財團法人資訊工業策進會
Priority to TW106143026A priority Critical patent/TWI657407B/en
Priority to CN201810043041.7A priority patent/CN109901189B/en
Application granted granted Critical
Publication of TWI657407B publication Critical patent/TWI657407B/en
Publication of TW201926258A publication Critical patent/TW201926258A/en

Links

Classifications

    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02ATECHNOLOGIES FOR ADAPTATION TO CLIMATE CHANGE
    • Y02A90/00Technologies having an indirect contribution to adaptation to climate change
    • Y02A90/10Information and communication technologies [ICT] supporting adaptation to climate change, e.g. for weather forecasting or climate simulation

Landscapes

  • Image Analysis (AREA)
  • Processing Or Creating Images (AREA)

Abstract

The present disclosure illustrates a three-dimensional point cloud tracking apparatus and method by recurrent neural network. The three-dimensional point cloud tracking apparatus and method would track the three-dimensional point cloud of the entire environment and model the entire environment by using recurrent neural network. This invention can be used to reconstruct the three-dimensional point cloud of the entire environment in the current moment, but also can be used to predict the three-dimensional point cloud of the entire environment in the follow-up moment.

Description

利用遞歸神經網路的三維點雲追蹤裝置及方法Three-dimensional point cloud tracking device and method using recurrent neural network

本發明是有關於一種點雲(point cloud)追蹤裝置及方法,且特別是一種利用遞歸神經網路(Recurrent Neural Network,RNN)的三維(three-dimensional,3D)點雲追蹤裝置及方法。The present invention relates to a point cloud tracking device and method, and more particularly to a three-dimensional (3D) point cloud tracking device and method using a Recurrent Neural Network (RNN).

「點雲」是指透過三維雷射掃描器所取得到的資料形式。現今中,三維雷射掃描器又可稱作為「光達(LiDAR)」,它是主要利用感測反射的雷射光束而來快速獲得到大量密佈於掃描物體表面上的多個點,並且因為這些點皆可包含有三維座標,所以光達便能建立起有關此掃描物體的三維點雲,以用來描述此掃描物體的表面形狀。"Point cloud" refers to the form of data obtained through a three-dimensional laser scanner. Nowadays, the three-dimensional laser scanner can also be called "LiDAR", which is mainly used to sense the reflected laser beam to quickly obtain a large number of points densely on the surface of the scanned object, and because These points can all contain three-dimensional coordinates, so the light can establish a three-dimensional point cloud about the scanned object to describe the surface shape of the scanned object.

因此,近年來光達則通常被用於自駕車系統或道路感測系統中,以作為避障或追蹤車輛之用途。然而,當此掃描物體因被遮蔽或在光達視野死角時,現有技術即無法能夠建立起有關此掃描物體的三維點雲,進而也就失去了以上用途。有鑑於此,本領域亟需一種能夠用來重建並預測三維點雲的方式。Therefore, in recent years, Ganda has been used in self-driving systems or road sensing systems as a means of avoiding obstacles or tracking vehicles. However, when the scanned object is obscured or is in a dead angle of light, the prior art cannot establish a three-dimensional point cloud about the scanned object, and thus loses the above use. In view of this, there is a need in the art for a way to reconstruct and predict a three-dimensional point cloud.

本發明之目的在於提供一種利用遞歸神經網路的三維點雲追蹤裝置及方法,並且為了因應具有多移動目標的複雜環境,所以本發明是以整個環境的三維點雲來作為追蹤對象。也就是說,本發明是用來重建並預測整個環境的三維點雲。It is an object of the present invention to provide a three-dimensional point cloud tracking apparatus and method using a recurrent neural network, and in order to cope with a complex environment having multiple moving targets, the present invention uses a three-dimensional point cloud of the entire environment as a tracking object. That is, the present invention is a three-dimensional point cloud used to reconstruct and predict the entire environment.

為達上述目的,本發明實施例提供一種利用遞歸神經網路的三維點雲追蹤裝置。所述三維點雲追蹤裝置包括輸入/輸出介面、儲存器以及處理器。輸入/輸出介面用來接收環境在不同時刻下的不同觀測三維點雲,且這些觀測三維點雲係由至少一光達所掃描取得。儲存器用來儲存有關環境的至少一記憶三維點雲。處理器則分別與輸入/輸出介面及儲存器電性連接,用來接收這些觀測三維點雲及記憶三維點雲,並且當收到環境在第一時刻下的觀測三維點雲時,處理器利用至少一遞歸神經網路模型,來對觀測三維點雲及記憶三維點雲進行環境重建運算,以得到環境在此第一時刻下的重建三維點雲,並且再利用遞歸神經網路模型,來對記憶三維點雲及空白三維點雲進行環境預測運算,以得到環境在第二時刻下的預測三維點雲。其中,所述第二時刻係晚於第一時刻。To achieve the above objective, an embodiment of the present invention provides a three-dimensional point cloud tracking device using a recurrent neural network. The three-dimensional point cloud tracking device includes an input/output interface, a storage, and a processor. The input/output interface is used to receive different observed 3D point clouds of the environment at different times, and the observed 3D point clouds are obtained by scanning at least one light. The storage is used to store at least one memory three-dimensional point cloud about the environment. The processor is electrically connected to the input/output interface and the storage respectively for receiving the observed three-dimensional point cloud and the memory three-dimensional point cloud, and the processor utilizes when the environment receives the three-dimensional point cloud at the first moment. At least one recursive neural network model is used to perform an environmental reconstruction operation on the observed three-dimensional point cloud and the memory three-dimensional point cloud to obtain a reconstructed three-dimensional point cloud of the environment at the first moment, and then recursive neural network model is used to The three-dimensional point cloud and the blank three-dimensional point cloud are used to perform environmental prediction operations to obtain a predicted three-dimensional point cloud of the environment at the second moment. The second moment is later than the first moment.

除此之外,本發明實施例另提供一種利用遞歸神經網路的三維點雲追蹤方法,執行於前述三維點雲追蹤裝置中。所述三維點雲追蹤方法包括如下步驟。令輸入/輸出介面接收環境在不同時刻下的不同觀測三維點雲,其中這些觀測三維點雲係由至少一光達所掃描取得。令儲存器儲存有關環境的至少一記憶三維點雲。令處理器接收這些觀測三維點雲及記憶三維點雲,並且當收到環境在第一時刻下的觀測三維點雲時,令處理器利用至少一遞歸神經網路模型,來對觀測三維點雲及記憶三維點雲進行環境重建運算,以得到環境在此第一時刻下的重建三維點雲,並且令處理器再利用遞歸神經網路模型,來對記憶三維點雲及空白三維點雲進行環境預測運算,以得到環境在第二時刻下的預測三維點雲。其中,所述第二時刻係晚於第一時刻。In addition, the embodiment of the present invention further provides a three-dimensional point cloud tracking method using a recurrent neural network, which is implemented in the foregoing three-dimensional point cloud tracking device. The three-dimensional point cloud tracking method includes the following steps. The input/output interface receives different observed three-dimensional point clouds of the environment at different times, wherein the observed three-dimensional point clouds are obtained by scanning at least one light. The storage stores at least one memory three-dimensional point cloud of the environment. Having the processor receive the observed 3D point cloud and the memory 3D point cloud, and when receiving the 3D point cloud in the environment at the first moment, let the processor utilize at least one recursive neural network model to observe the 3D point cloud And the memory 3D point cloud performs the environment reconstruction operation to obtain the reconstructed 3D point cloud of the environment at the first moment, and the processor reuses the recurrent neural network model to perform the environment on the memory 3D point cloud and the blank 3D point cloud. The prediction operation is performed to obtain a predicted three-dimensional point cloud of the environment at the second moment. The second moment is later than the first moment.

為使能更進一步瞭解本發明之特徵及技術內容,請參閱以下有關本發明之詳細說明與附圖,但是此等說明與所附圖式僅係用來說明本發明,而非對本發明的權利範圍作任何的限制。The detailed description of the present invention and the accompanying drawings are to be understood by the claims The scope is subject to any restrictions.

在下文中,將藉由圖式說明本發明之各種實施例來詳細描述本發明。然而,本發明概念可能以許多不同形式來體現,且不應解釋為限於本文中所闡述之例示性實施例。此外,在圖式中相同參考數字可用以表示類似的元件。In the following, the invention will be described in detail by way of illustration of various embodiments of the invention. However, the inventive concept may be embodied in many different forms and should not be construed as being limited to the illustrative embodiments set forth herein. In addition, the same reference numerals may be used in the drawings to represent similar elements.

請參閱圖1,圖1是本發明實施例所提供的利用遞歸神經網路的三維點雲追蹤裝置的功能方塊示意圖。三維點雲追蹤裝置1包括輸入/輸出介面11、處理器13以及儲存器15。值得一提的是,上述各元件可以是透過純硬件電路來實現,或者是透過硬件電路搭配固件或軟件來實現,但本發明皆不以此為限制。除此之外,上述各元件可以是整合或是分開設置,且本發明亦不以此為限制。總而言之,本發明並不限制三維點雲追蹤裝置1的具體實現方式。Please refer to FIG. 1. FIG. 1 is a functional block diagram of a three-dimensional point cloud tracking device using a recurrent neural network according to an embodiment of the present invention. The three-dimensional point cloud tracking device 1 includes an input/output interface 11, a processor 13, and a storage 15. It should be noted that the above components may be implemented by a pure hardware circuit, or by hardware or software, or the software, but the invention is not limited thereto. In addition, the above components may be integrated or separately provided, and the present invention is not limited thereto. In summary, the present invention does not limit the specific implementation of the three-dimensional point cloud tracking device 1.

在本實施例中,輸入/輸出介面11是用來接收環境(未繪示)在不同時刻下的不同觀測三維點雲S,且這些觀測三維點雲S係由至少一光達(未繪示)所掃描取得。由於光達的掃描原理已為本技術領域中具有通常知識者所習知,因此有關上述觀測三維點雲S的細部內容於此就不再多加贅述。另外,儲存器15是用來儲存有關環境的至少一記憶三維點雲M。有關於記憶三維點雲M的具體內容,將會於下文中藉由其他實施例而作詳盡說明,故於此就先不再多加贅述。處理器13則分別與輸入/輸出介面11及儲存器15電性連接,並用來接收這些觀測三維點雲S及記憶三維點雲M。In this embodiment, the input/output interface 11 is used to receive different observed three-dimensional point clouds S at different times in the environment (not shown), and the observed three-dimensional point cloud S is composed of at least one light (not shown) ) was scanned. Since the scanning principle of the optical signal is well known to those of ordinary skill in the art, the details of the above-described observation of the three-dimensional point cloud S will not be further described herein. In addition, the storage 15 is for storing at least one memory three-dimensional point cloud M related to the environment. The specific content of the memory three-dimensional point cloud M will be described in detail below through other embodiments, and thus will not be further described herein. The processor 13 is electrically connected to the input/output interface 11 and the storage device 15, and is configured to receive the observed three-dimensional point cloud S and the memory three-dimensional point cloud M.

請一併參閱圖2,圖2係將用以來解釋圖1中的處理器13的具體運作方式。如圖2所示,當收到環境在某第一時刻(例如,時刻t)下的觀測三維點雲S(t)時,處理器13則利用至少一遞歸神經網路模型17,來對觀測三維點雲S(t)及記憶三維點雲M進行環境重建運算,以得到環境在此第一時刻下的重建三維點雲R(t),並且再利用遞歸神經網路模型17,來對記憶三維點雲M及一個空白三維點雲(未繪示)進行環境預測運算,以得到環境在第二時刻(例如,時刻t+1)下的預測三維點雲P(t+1)。對此,應當理解的是,所述第二時刻係晚於第一時刻。Please refer to FIG. 2 together. FIG. 2 is a diagram for explaining the specific operation mode of the processor 13 in FIG. As shown in FIG. 2, when the observed three-dimensional point cloud S(t) is received at a certain time (eg, time t), the processor 13 uses at least one recurrent neural network model 17 to observe the observation. The three-dimensional point cloud S(t) and the memory three-dimensional point cloud M perform an environment reconstruction operation to obtain a reconstructed three-dimensional point cloud R(t) of the environment at the first moment, and reuse the recurrent neural network model 17 to memorize the memory. The three-dimensional point cloud M and a blank three-dimensional point cloud (not shown) perform an environment prediction operation to obtain a predicted three-dimensional point cloud P(t+1) of the environment at a second time (for example, time t+1). In this regard, it should be understood that the second time is later than the first time.

然而,為了方便以下說明,本實施例的第一及第二時刻則是僅先採用時刻t及時刻t+1的例子來進行說明,但其並非用以限制本發明。同樣地,為了方便以下說明,圖2中的遞歸神經網路模型17則也是僅先採用數量為1的例子來進行說明,但其亦非用以限制本發明。也就是說,圖2中所分別進行環境重建運算或環境預測運算的遞歸神經網路模型17,可以指的是同一個遞歸神經網路模型17,或是不同的遞歸神經網路模型17,但本發明皆不以此為限制。However, for convenience of the following description, the first and second timings of the present embodiment are described by using only the time t and the time t+1 first, but are not intended to limit the present invention. Similarly, for convenience of the following description, the recursive neural network model 17 of FIG. 2 is also described by using only an example of a number of 1, but it is not intended to limit the present invention. That is to say, the recursive neural network model 17 for performing the environmental reconstruction operation or the environmental prediction operation in Fig. 2 may refer to the same recurrent neural network model 17, or a different recurrent neural network model 17, but The present invention is not limited thereto.

另外,有關圖2的遞歸神經網路模型17中進行環境重建運算,或進行環境預測運算的具體運作方式,將會於下文中藉由其他實施例而作詳盡說明,故於此就先不再多加贅述。需要說明的是,雖然圖2只用了再一次的遞歸神經網路模型17而來獲得到環境在時刻t+1下的預測三維點雲P(t+1),但本發明卻不以此為限制。也就是說,只要是能夠在處理時間及計算能力充足的條件下,本發明實施例都能夠是依照預測的下一時刻遠近(亦即,第二與第一時刻間的時距)而來自行決定是否使用遞迴數次的遞歸神經網路模型17,以來進而取得到環境在其他第二時刻(例如,時刻t+2)下的預測三維點雲。In addition, the specific operation mode for performing the environment reconstruction operation or the environment prediction operation in the recurrent neural network model 17 of FIG. 2 will be described in detail below by other embodiments, so that it is no longer More details. It should be noted that although FIG. 2 only uses the recursive neural network model 17 again to obtain the predicted three-dimensional point cloud P(t+1) of the environment at time t+1, the present invention does not For the limit. That is, as long as the processing time and the computing power are sufficient, the embodiment of the present invention can follow the predicted next time (ie, the time interval between the second and the first time). It is decided whether or not to use the recursive neural network model 17 that is recursed several times, and then the predicted three-dimensional point cloud of the environment at other second moments (for example, time t+2) is obtained.

總而言之,根據以上內容的教示,本技術領域中具有通常知識者應可以理解到,因為本發明特別是以整個環境的三維點雲來作為追蹤對象,所以當此環境在目前時刻(例如,第一時刻t)下卻有著因某移動物體被遮蔽而未能獲得到此環境的部分區域三維點雲時,本發明是將能夠利用過去所儲存的此環境的記憶三維點雲而來估測出此移動物體的三維點雲資訊,以進而對目前未能獲知的上述部分區域三維點雲進行補充。也就是說,正確地重建起此環境在目前時刻下的三維點雲。In summary, according to the teachings of the above, it should be understood by those of ordinary skill in the art, because the present invention is particularly a three-dimensional point cloud of the entire environment as a tracking object, so when the environment is at the current time (for example, the first At time t), when there is a partial 3D point cloud in which a moving object is obscured and the environment is not obtained, the present invention estimates that this can be utilized by using a memory three-dimensional point cloud stored in the past. The three-dimensional point cloud information of the moving object is further supplemented by the above-mentioned partial area three-dimensional point cloud which is not currently known. In other words, correctly reconstruct the 3D point cloud of this environment at the current moment.

另一方面,對於上述被遮蔽的移動物體來說,因為現有技術通常僅能夠以等速度運動的方式而來預測此移動物體的未來變化,所以當此移動物體是以非等速度方式運動,或是此移動物體被遮蔽的時間過長時,現有技術都將容易失去了對此移動物體的追蹤。但因為本發明是利用了遞歸神經網路模型來對整個環境的三維點雲進行編碼,所以即使是在此環境內的某移動物體的運動方式較為複雜,或者是在此移動物體被遮蔽的時間過長時,本發明是都將能夠仍預測得到此環境在後續時刻下的三維點雲。也就是說,準確地追蹤到此環境內的移動物體的未來變化。On the other hand, for the above-mentioned shaded moving object, since the prior art generally can only predict the future change of the moving object in a manner of moving at a constant speed, when the moving object moves in a non-equal speed manner, or When the moving object is obscured for too long, the prior art will easily lose track of the moving object. However, since the present invention utilizes a recursive neural network model to encode a three-dimensional point cloud of the entire environment, even a moving object in the environment moves in a complicated manner or is in a time when the moving object is obscured. When the length is too long, the present invention will be able to still predict the three-dimensional point cloud of this environment at a subsequent moment. That is, accurately tracking future changes in moving objects within this environment.

接著,請一併參閱圖3A,圖3A係將用以來解釋圖2的遞歸神經網路模型17中進行環境重建運算的具體運作方式。值得一提的是,為了方便以下說明,圖3A中的記憶三維點雲M則是採用數量為2的例子來進行說明。也就是說,記憶三維點雲M可包括第一記憶三維點雲M1及第二記憶三維點雲M2,但其亦非用以限制本發明。Next, please refer to FIG. 3A together. FIG. 3A is a detailed operation mode for explaining the environment reconstruction operation in the recurrent neural network model 17 of FIG. 2 . It is worth mentioning that, in order to facilitate the following description, the memory three-dimensional point cloud M in FIG. 3A is described by using an example of a number of two. That is to say, the memory three-dimensional point cloud M may include the first memory three-dimensional point cloud M1 and the second memory three-dimensional point cloud M2, but it is not intended to limit the present invention.

如圖3A所示,遞歸神經網路模型17是會先對觀測三維點雲S(t)進行第一稀疏捲積(sparse convolution)運算,以得到經此第一稀疏捲積SP1後的運算三維點雲Q1(t)。然後,遞歸神經網路模型17是將運算三維點雲Q1(t)與第一記憶三維點雲M1進行第二稀疏捲積運算,以得到經此第二稀疏捲積SP2後的運算三維點雲Q2(t),並且以運算三維點雲Q2(t)更新(update)第一記憶三維點雲M1。最後,遞歸神經網路模型17是再將運算三維點雲Q2(t)與第二記憶三維點雲M2進行第三稀疏捲積運算,以得到環境在此時刻t(亦即,第一時刻)下的重建三維點雲R(t),並且以重建三維點雲R(t)更新第二記憶三維點雲M2。As shown in FIG. 3A, the recurrent neural network model 17 first performs a first sparse convolution operation on the observed three-dimensional point cloud S(t) to obtain a three-dimensional operation after the first sparse convolution SP1. Point cloud Q1(t). Then, the recursive neural network model 17 performs a second sparse convolution operation on the computed three-dimensional point cloud Q1(t) and the first memory three-dimensional point cloud M1 to obtain an operation three-dimensional point cloud after the second sparse convolution SP2. Q2(t), and the first memory three-dimensional point cloud M1 is updated by computing the three-dimensional point cloud Q2(t). Finally, the recurrent neural network model 17 performs a third sparse convolution operation on the computed three-dimensional point cloud Q2(t) and the second memory three-dimensional point cloud M2 to obtain an environment at this time t (ie, the first moment). The 3D point cloud R(t) is reconstructed, and the second memory 3D point cloud M2 is updated by reconstructing the 3D point cloud R(t).

從上述內容可知,因為圖3A是可以用到了稀疏捲積的運算特性,所以本實施例的三維點雲追蹤裝置1是將能夠在合理的時間及計算能力上處理複雜的三維點雲資訊。然而,由於稀疏捲積的運算原理已為本技術領域中具有通常知識者所習知,因此有關稀疏捲積SP1~SP3的細部內容於此就不再多加贅述。需要說明的是,圖3A所使用的三層稀疏捲積方式(亦即,稀疏捲積SP1~SP3)在此僅只是舉例,其並非用以限制本發明。換句話說,本技術領域中具有通常知識者應可依據實際需求或應用來進行不同層級數目的稀疏捲積設計。As can be seen from the above, since FIG. 3A is an operational characteristic in which sparse convolution can be used, the three-dimensional point cloud tracking device 1 of the present embodiment is capable of processing complex three-dimensional point cloud information with reasonable time and computational power. However, since the arithmetic principle of the sparse convolution is well known to those of ordinary skill in the art, the details of the sparse convolutions SP1 to SP3 will not be further described herein. It should be noted that the three-layer sparse convolution mode (that is, the sparse convolutions SP1 to SP3) used in FIG. 3A is merely an example, and is not intended to limit the present invention. In other words, those of ordinary skill in the art should be able to perform different levels of sparse convolution design depending on actual needs or applications.

另外,在其中一種應用中,遞歸神經網路模型17是可以僅將部分的稀疏捲積運算結果用於下一層的稀疏捲積運算或輸出,並且將另一部分的稀疏捲積運算結果用於更新記憶三維點雲。舉例來說,在圖3A的稀疏捲積SP2後,遞歸神經網路模型17是可以僅將部分的運算三維點雲Q2(t)用於下一層的稀疏捲積運算(亦即,稀疏捲積SP3),並且將另一部分的運算三維點雲Q2(t)用於更新第一記憶三維點雲M1。In addition, in one of the applications, the recursive neural network model 17 is capable of using only part of the sparse convolution operation result for the next layer of the sparse convolution operation or output, and the other part of the sparse convolution operation result for updating. Memory 3D point cloud. For example, after the sparse convolution SP2 of FIG. 3A, the recursive neural network model 17 can use only part of the computational three-dimensional point cloud Q2(t) for the next layer of sparse convolution operations (ie, sparse convolution). SP3), and another part of the operational three-dimensional point cloud Q2(t) is used to update the first memory three-dimensional point cloud M1.

然而,因為經第二稀疏捲積SP2後的運算三維點雲Q2(t)的捲積核數目可以分為數個特徵(或稱,頻道),所以上述這兩部分的運算三維點雲Q2(t)可以是指包含有不同頻道的數據。也就是說,上述這兩部分的運算三維點雲Q2(t)的數據可以是完全沒有重疊。總而言之,本發明並不限制進行稀疏捲積運算或更新記憶三維點雲時的具體實現方式。However, since the number of convolution kernels of the operation three-dimensional point cloud Q2(t) after the second sparse convolution SP2 can be divided into several features (or channels), the operation of the above two parts is three-dimensional point cloud Q2 (t ) can refer to data that contains different channels. That is to say, the data of the above three parts of the operation three-dimensional point cloud Q2(t) may be completely unoverlapping. In summary, the present invention does not limit the specific implementation when performing a sparse convolution operation or updating a memory three-dimensional point cloud.

另一方面,若考量到第一或第二記憶三維點雲M1、M2並不是只能夠完全用運算三維點雲Q2(t)或重建三維點雲R(t)來作取代,因此,在其中一種應用中,處理器13更可用以定義至少一權重自定函數f 、至少一第一稀疏捲積核K1以及至少一第二稀疏捲積核K2。值得一提的是,上述權重自定函數f 、第一稀疏捲積核K1及第二稀疏捲積核K2可以是經由三維點雲追蹤裝置1在進行完一訓練模式後所定義,但本發明並不以此為限制。On the other hand, if it is considered that the first or second memory three-dimensional point cloud M1, M2 is not only completely replaceable by computing the three-dimensional point cloud Q2(t) or reconstructing the three-dimensional point cloud R(t), In one application, the processor 13 is further operable to define at least one weighted custom function f , at least one first sparse convolution kernel K1, and at least one second sparse convolution kernel K2. It is worth mentioning that the weight custom function f , the first sparse convolution kernel K1 and the second sparse convolution kernel K2 may be defined after the completion of a training mode by the three-dimensional point cloud tracking device 1, but the present invention This is not a limitation.

舉例來說,所述訓練模式可以是利用遞歸神經網路模型17來對已知的三維點雲(未繪示)進行環境重建運算,以獲得到其所重建後的三維點雲,並且可藉由比對已知的三維點雲及其所重建後的三維點雲間的誤差關係,來進而制定出遞歸神經網路模型17中的各項特徵參數(例如,權重自定函數f 、第一稀疏捲積核K1及第二稀疏捲積核K2等,又或者是,稀疏捲積SP1~SP3內的各捲積核參數)。For example, the training mode may be a recursive neural network model 17 to perform an environmental reconstruction operation on a known three-dimensional point cloud (not shown) to obtain a reconstructed three-dimensional point cloud, and may borrow The characteristic parameters in the recursive neural network model 17 are further determined by comparing the error relationship between the known three-dimensional point cloud and the reconstructed three-dimensional point cloud (for example, the weight custom function f , the first sparse volume) The product K1 and the second sparse convolution kernel K2, etc., or the convolution kernel parameters in the sparse convolutions SP1 to SP3).

另外,在所述訓練模式中,遞歸神經網路模型17還可以通過使用線性損失函數(linear hinge loss)並加入L1懲罰(L1 penalty),以提升稀疏捲積的稀疏程度。由於訓練模式的原理已皆為本技術領域中具有通常知識者所習知,因此上述內容將僅是做為示意,以下即不再多做說明。總而言之,本發明並不限制三維點雲追蹤裝置1進行訓練模式時的具體實現方式,故本技術領域中具有通常知識者應可依據實際需求或應用來進行相關設計。In addition, in the training mode, the recurrent neural network model 17 can also increase the sparsity of the sparse convolution by using a linear hinge loss and adding a L1 penalty. Since the principles of the training mode are well known to those of ordinary skill in the art, the above description will be merely illustrative, and the following description will not be repeated. In summary, the present invention does not limit the specific implementation manner when the three-dimensional point cloud tracking device 1 performs the training mode. Therefore, those skilled in the art should be able to perform related designs according to actual needs or applications.

更進一步來說,在以運算三維點雲Q2(t)更新第一記憶三維點雲M1的一較佳實施中,遞歸神經網路模型17是會利用權重自定函數f ,來由第一記憶三維點雲M1、運算三維點雲Q2(t)、第一稀疏捲積核K1及第二稀疏捲積核K2中,決定一個權重向量p,並且將第一記憶三維點雲M1更新為第一記憶三維點雲M1、運算三維點雲Q2(t)及權重向量p代入一權重方程式後的結果。Further, in a preferred implementation of updating the first memory three-dimensional point cloud M1 by computing the three-dimensional point cloud Q2(t), the recursive neural network model 17 uses the weighted custom function f to be used by the first memory. In the three-dimensional point cloud M1, the three-dimensional point cloud Q2(t), the first sparse convolution kernel K1, and the second sparse convolution kernel K2, a weight vector p is determined, and the first memory three-dimensional point cloud M1 is updated to the first The result of memorizing a three-dimensional point cloud M1, computing a three-dimensional point cloud Q2(t), and a weight vector p into a weighting equation.

類似地,在以重建三維點雲R(t)更新第二記憶三維點雲M2的一較佳實施中,遞歸神經網路模型17則是會利用權重自定函數f ,來由第二記憶三維點雲M2、重建三維點雲R(t)、第一稀疏捲積核K1及第二稀疏捲積核K2中,決定權重向量p,並且將第二記憶三維點雲M2更新為第二記憶三維點雲M2、重建三維點雲R(t)及權重向量p代入前述權重方程式後的結果。Similarly, in a preferred implementation of updating the second memory three-dimensional point cloud M2 by reconstructing the three-dimensional point cloud R(t), the recursive neural network model 17 uses the weighted custom function f to obtain the second memory three-dimensional The point cloud M2, the reconstructed three-dimensional point cloud R(t), the first sparse convolution kernel K1 and the second sparse convolution kernel K2, determine the weight vector p, and update the second memory three-dimensional point cloud M2 to the second memory three-dimensional The result of the point cloud M2, the reconstruction of the three-dimensional point cloud R(t) and the weight vector p into the aforementioned weight equation.

因此,應當理解的是,在上述更新第一及第二記憶三維點雲M1、M2的這兩較佳實施例中,每一較佳實施例所能用到的權重自定函數f 、第一及第二稀疏捲積核K1、K2即可是互不相同的。總而言之,本發明亦不限制權重自定函數f 、第一及第二稀疏捲積核K1、K2的具體實現方式。接著,請一併參閱圖3B,圖3B係將用以來解釋圖3A的環境重建運算中更新第一或第二記憶三維點雲的一較佳實施例下的具體運作方式。在圖3B中,權重方程式即為p*C1+(1-p)*C2,且權重向量p則表示為:p=f (C1*K1+C2*K2),而C1及C2則分別為第一記憶三維點雲M1與運算三維點雲Q2(t),或第二記憶三維點雲M2與重建三維點雲R(t)。Therefore, it should be understood that in the above two preferred embodiments for updating the first and second memory three-dimensional point clouds M1, M2, the weighting function f , the first one that can be used in each preferred embodiment And the second sparse convolution kernels K1, K2 may be different from each other. In summary, the present invention also does not limit the specific implementation of the weight custom function f , the first and second sparse convolution kernels K1, K2. Next, please refer to FIG. 3B together. FIG. 3B is a specific operation mode of a preferred embodiment for updating the first or second memory three-dimensional point cloud in the environment reconstruction operation of FIG. 3A. In FIG. 3B, the weight equation is p*C1+(1-p)*C2, and the weight vector p is expressed as: p= f (C1*K1+C2*K2), and C1 and C2 are respectively the first. Memorize the 3D point cloud M1 with the computational 3D point cloud Q2(t), or the second memory 3D point cloud M2 and reconstruct the 3D point cloud R(t).

根據以上內容的教示,應當亦理解的是,權重向量p的分量係介於0至1之間。也就是說,假設在權重向量p全為0的情況下,遞歸神經網路模型17就會是只利用運算三維點雲Q2(t)或重建三維點雲R(t) (亦即,C2)來取代目前的第一或第二記憶三維點雲M1、M2,以此類推,假設在權重向量p全為1的情況下,遞歸神經網路模型17也就會是只利用原先的第一或第二記憶三維點雲M1、M2(亦即,C1)來維持作為目前的第一或第二記憶三維點雲M1、M2,而不會是利用到運算三維點雲Q2(t)或重建三維點雲R(t)來更新目前的第一或第二記憶三維點雲M1、M2。總而言之,圖3B所使用更新第一或第二記憶三維點雲M1、M2的具體實現方式在此亦僅只是舉例,其並非用以限制本發明。Based on the teachings above, it should also be understood that the component of the weight vector p is between 0 and 1. That is to say, assuming that the weight vector p is all zero, the recursive neural network model 17 will use only the computational three-dimensional point cloud Q2(t) or reconstruct the three-dimensional point cloud R(t) (ie, C2). To replace the current first or second memory three-dimensional point cloud M1, M2, and so on, assuming that the weight vector p is all one, the recursive neural network model 17 will only use the original first or The second memory three-dimensional point cloud M1, M2 (ie, C1) is maintained as the current first or second memory three-dimensional point cloud M1, M2, and is not utilized to calculate the three-dimensional point cloud Q2 (t) or reconstruct three-dimensional The point cloud R(t) updates the current first or second memory three-dimensional point cloud M1, M2. In summary, the specific implementation of updating the first or second memory three-dimensional point cloud M1, M2 used in FIG. 3B is merely an example, and is not intended to limit the present invention.

更進一步來說,從上述內容可知,圖1中的儲存器15所儲存的記憶三維點雲M(亦即,包括第一及第二記憶三維點雲M1、M2)並非只是輸入/輸出介面11所收到的不同時刻下的這些觀測三維點雲S,而是這些觀測三維點雲S所經由數次捲積及更新處理(例如,圖3A)後的數據結果。也就是說,圖2中的記憶三維點雲M即為過去收到的觀測三維點雲S(t-1)(未繪示)所經由數次捲積及更新處理後的數據結果。因此,在其中一種應用中,記憶三維點雲M更可以是在當三維點雲追蹤裝置1所開始探測此環境後才產生,而非是一開始就預先儲存在儲存器15中。另外,假設觀測三維點雲S(t)即為最一開始的觀測數據時,儲存器15所儲存的記憶三維點雲M則可以是由空白三維點雲經數次捲積及更新處理後所產生。總而言之,本發明亦不限制記憶三維點雲M的具體實現方式。Furthermore, it can be seen from the above that the memory three-dimensional point cloud M (that is, including the first and second memory three-dimensional point clouds M1, M2) stored in the storage 15 in FIG. 1 is not just the input/output interface 11 These observations of the three-dimensional point cloud S at different times are received, but the data results of the observations of the three-dimensional point cloud S after several convolution and update processes (for example, FIG. 3A). That is to say, the memory three-dimensional point cloud M in FIG. 2 is the data result after several times of convolution and update processing of the observed three-dimensional point cloud S(t-1) (not shown) received in the past. Therefore, in one of the applications, the memory three-dimensional point cloud M can be generated only after the three-dimensional point cloud tracking device 1 starts detecting the environment, instead of being stored in the storage 15 in advance. In addition, assuming that the three-dimensional point cloud S(t) is the first observation data, the memory three-dimensional point cloud M stored in the memory 15 can be processed by the blank three-dimensional point cloud after several times of convolution and update processing. produce. In summary, the present invention also does not limit the specific implementation of the memory three-dimensional point cloud M.

再者,請一併參閱圖3C,圖3C係將用以來解釋圖2的遞歸神經網路模型17中進行環境預測運算的具體運作方式。如圖3C所示,遞歸神經網路模型17是會先將一個空白三維點雲(未繪示)與第一記憶三維點雲M1進行第四稀疏捲積運算,以得到經此第四稀疏捲積SP4後的運算三維點雲Q3(t)。然後,遞歸神經網路模型17是會再將運算三維點雲Q3(t)與第二記憶三維點雲M2進行第五稀疏捲積運算,以得到環境在時刻t+1(亦即,第二時刻)下的預測三維點雲P(t+1)。由於圖3C中的部分技術原理與圖3A相同,故於此也就不再多加贅述。總而言之,圖3C所使用到的環境預測運算的具體實現方式在此也僅只是舉例,其並非用以限制本發明。Furthermore, please refer to FIG. 3C together. FIG. 3C is a detailed operation mode for explaining the environment prediction operation in the recurrent neural network model 17 of FIG. 2 . As shown in FIG. 3C, the recurrent neural network model 17 first performs a fourth sparse convolution operation on a blank three-dimensional point cloud (not shown) and the first memory three-dimensional point cloud M1 to obtain a fourth thinned volume. Computation of the 3D point cloud Q3(t) after SP4. Then, the recurrent neural network model 17 performs a fifth sparse convolution operation on the computed three-dimensional point cloud Q3(t) and the second memory three-dimensional point cloud M2 to obtain an environment at time t+1 (ie, second The predicted three-dimensional point cloud P(t+1) under time). Since some of the technical principles in FIG. 3C are the same as those in FIG. 3A, no further details are provided herein. In summary, the specific implementation of the environment prediction operation used in FIG. 3C is merely an example here, and is not intended to limit the present invention.

最後,為了更進一步說明關於三維點雲追蹤裝置1的運作流程,本發明進一步提供其三維點雲追蹤方法的一種實施方式。請參閱圖4,圖4是本發明實施例所提供的利用遞歸神經網路的三維點雲追蹤方法的流程示意圖。其中,圖4的三維點雲追蹤方法是可以執行於圖1的三維點雲追蹤裝置1中,但本發明並不限制圖4的三維點雲追蹤方法僅能夠執行於圖1的三維點雲追蹤裝置1中。另外,詳細步驟流程如前述實施例所述,於此僅作概述而不再多加冗述。Finally, in order to further explain the operational flow of the three-dimensional point cloud tracking device 1, the present invention further provides an embodiment of its three-dimensional point cloud tracking method. Please refer to FIG. 4. FIG. 4 is a schematic flowchart diagram of a three-dimensional point cloud tracking method using a recurrent neural network according to an embodiment of the present invention. The three-dimensional point cloud tracking method of FIG. 4 can be executed in the three-dimensional point cloud tracking device 1 of FIG. 1 , but the present invention does not limit the three-dimensional point cloud tracking method of FIG. 4 to only perform the three-dimensional point cloud tracking of FIG. 1 . In device 1. In addition, the detailed step flow is as described in the foregoing embodiments, and is merely summarized herein and will not be redundant.

如圖4所示,首先,在步驟S410中,令輸入/輸出介面接收環境在不同時刻下的不同觀測三維點雲,其中這些觀測三維點雲係由至少一光達所掃描取得。其次,在步驟S420中,令儲存器儲存有關環境的至少一記憶三維點雲。接著,在步驟S430中,令處理器接收這些觀測三維點雲及記憶三維點雲,並且當收到環境在第一時刻下的觀測三維點雲時,則進行步驟S440至步驟S450。As shown in FIG. 4, first, in step S410, the input/output interface is configured to receive different observed three-dimensional point clouds of the environment at different times, wherein the observed three-dimensional point clouds are obtained by scanning at least one light. Next, in step S420, the storage is caused to store at least one memory three-dimensional point cloud of the environment. Next, in step S430, the processor is caused to receive the observed three-dimensional point cloud and the memory three-dimensional point cloud, and when the observed three-dimensional point cloud is received at the first time, the process proceeds to step S440 to step S450.

在步驟S440中,令處理器利用至少一遞歸神經網路模型,來對觀測三維點雲及記憶三維點雲進行環境重建運算,以得到環境在此第一時刻下的重建三維點雲,並且在步驟S450中,令處理器再利用遞歸神經網路模型,來對記憶三維點雲及空白三維點雲進行環境預測運算,以得到環境在第二時刻下的預測三維點雲。其中,所述第二時刻係晚於第一時刻。In step S440, the processor is configured to perform an environment reconstruction operation on the observed three-dimensional point cloud and the memory three-dimensional point cloud by using at least one recurrent neural network model to obtain a reconstructed three-dimensional point cloud of the environment at the first moment, and In step S450, the processor uses the recurrent neural network model to perform an environment prediction operation on the memory three-dimensional point cloud and the blank three-dimensional point cloud to obtain a predicted three-dimensional point cloud of the environment at the second moment. The second moment is later than the first moment.

根據以上內容的教示,本技術領域中具有通常知識者應可以理解到,步驟S410、步驟S420及步驟S430應該為並行執行而未衝突之步驟。另外,以下為了更進一步說明關於步驟S440的實現細節,本發明進一步提供其步驟S440的一種實施方式。請參閱圖5A,圖5A是圖4的三維點雲追蹤方法中利用遞歸神經網路模型進行環境重建運算的流程示意圖。其中,圖5A中部分與圖4相同之流程步驟以相同之圖號標示,故於此便不再多加詳述其細節。In light of the above teachings, those of ordinary skill in the art will appreciate that steps S410, S420, and S430 should be performed in parallel without conflicting steps. In addition, in order to further explain the implementation details regarding step S440, the present invention further provides an embodiment of step S440 thereof. Please refer to FIG. 5A. FIG. 5A is a schematic flowchart of an environment reconstruction operation using a recurrent neural network model in the three-dimensional point cloud tracking method of FIG. 4. The steps in FIG. 5A that are the same as those in FIG. 4 are denoted by the same reference numerals, and thus the details thereof will not be described in detail.

在圖5A的實施例中,步驟S440更可以包括有步驟S441至步驟S445。首先,在步驟S441中,遞歸神經網路模型是會先對觀測三維點雲進行第一稀疏捲積運算,以得到經第一稀疏捲積後的第一運算三維點雲。接著,在步驟S443中,遞歸神經網路模型是將第一運算三維點雲與儲存器所儲存的第一記憶三維點雲進行第二稀疏捲積運算,以得到經第二稀疏捲積後的第二運算三維點雲,並且以第二運算三維點雲更新第一記憶三維點雲。In the embodiment of FIG. 5A, step S440 may further include steps S441 to S445. First, in step S441, the recursive neural network model first performs a first sparse convolution operation on the observed three-dimensional point cloud to obtain a first computed three-dimensional point cloud after the first sparse convolution. Next, in step S443, the recursive neural network model performs a second sparse convolution operation on the first computed three-dimensional point cloud and the first memory three-dimensional point cloud stored in the storage to obtain the second sparsely convolved convolution The second computing three-dimensional point cloud, and updating the first memory three-dimensional point cloud with the second computing three-dimensional point cloud.

然後,在步驟S445中,遞歸神經網路模型是再將第二運算三維點雲與儲存器所儲存的第二記憶三維點雲進行第三稀疏捲積運算,以得到環境在第一時刻下的重建三維點雲,並且以重建三維點雲更新第二記憶三維點雲。值得注意的是,圖5A所採用的實施方式在此僅是用以舉例,其並非用以限制本發明。另外,圖5A中更新第一或第二記憶三維點雲的一較佳實施方式,可請參閱到圖3B所示,故於此便不再多加贅述。Then, in step S445, the recursive neural network model performs a third sparse convolution operation on the second computed three-dimensional point cloud and the second memory three-dimensional point cloud stored in the storage to obtain the environment at the first moment. The 3D point cloud is reconstructed, and the second memory 3D point cloud is updated by reconstructing the 3D point cloud. It is to be noted that the embodiment adopted in FIG. 5A is merely for exemplification, and is not intended to limit the present invention. In addition, a preferred embodiment of updating the first or second memory three-dimensional point cloud in FIG. 5A can be referred to FIG. 3B, and thus no further details are provided herein.

另外,以下為了更進一步說明關於步驟S450的實現細節,本發明進一步提供其步驟S450的一種實施方式。請參閱圖5B,圖5B是圖4的三維點雲追蹤方法中利用遞歸神經網路模型進行環境預測運算的流程示意圖。其中,圖5B中部分與圖4相同之流程步驟以相同之圖號標示,故於此便不再多加詳述其細節。In addition, in order to further explain the implementation details regarding step S450, the present invention further provides an embodiment of step S450 thereof. Please refer to FIG. 5B. FIG. 5B is a schematic flowchart of the environment prediction operation using the recurrent neural network model in the three-dimensional point cloud tracking method of FIG. 4. The process steps in FIG. 5B that are the same as those in FIG. 4 are denoted by the same reference numerals, and thus the details thereof will not be described in detail.

在圖5B的實施例中,步驟S450更可以包括有步驟S451至步驟S453。首先,在步驟S451中,遞歸神經網路模型是會先將一個空白三維點雲與儲存器所儲存的第一記憶三維點雲進行第四稀疏捲積運算,以得到經第四稀疏捲積後的第三運算三維點雲。然後,在步驟S453中,遞歸神經網路模型是會再將第三運算三維點雲與儲存器所儲存的第二記憶三維點雲進行第五稀疏捲積運算,以得到環境在第二時刻下的預測三維點雲。由於詳細步驟流程如前述實施例所述,故於此也就不再多加詳述其細節。In the embodiment of FIG. 5B, step S450 may further include steps S451 to S453. First, in step S451, the recursive neural network model first performs a fourth sparse convolution operation on a blank three-dimensional point cloud and a first memory three-dimensional point cloud stored in the storage to obtain a fourth sparse convolution. The third operation of the 3D point cloud. Then, in step S453, the recursive neural network model performs a fifth sparse convolution operation on the second computed three-dimensional point cloud and the second memory three-dimensional point cloud stored in the storage to obtain the environment at the second moment. The prediction of 3D point clouds. Since the detailed step flow is as described in the foregoing embodiment, the details thereof will not be described in detail herein.

綜上所述,本發明實施例所提供的利用遞歸神經網路的三維點雲追蹤裝置及方法,不僅可以是用來重建整個環境的三維點雲,還可以是用來預測整個環境在後續時刻下的三維點雲。特別地是,因為本發明是以整個環境的三維點雲來作為追蹤對象,所以在重建的過程中,本發明是可以利用過去的點雲資訊,來對此環境目前所因某移動物體被遮蔽而未能探測的部分區域點雲進行補充,以藉此正確地重建起此環境在目前時刻下的三維點雲,並且在預測的過程中,本發明是可以通過利用遞歸神經網路模型來對整個環境進行建模(亦即,編碼)的方式,以進而預測出此環境在後續時刻下的三維點雲,並藉此準確地追蹤到此環境內的某移動物體的未來變化。除此之外,本發明還可以是運用到了稀疏捲積的運算特性,因此在合理的時間及計算能力上,本發明是能夠有效地處理複雜的三維點雲資訊,以進而實踐上述重建及預測的最佳效果。In summary, the 3D point cloud tracking device and method using the recurrent neural network provided by the embodiment of the present invention can be used not only to reconstruct a 3D point cloud of the entire environment, but also to predict the entire environment at a subsequent time. The 3D point cloud below. In particular, since the present invention uses a three-dimensional point cloud of the entire environment as a tracking object, in the process of reconstruction, the present invention can utilize past point cloud information to mask the current moving object of the environment. The partial region point cloud that cannot be detected is supplemented to thereby correctly reconstruct the three-dimensional point cloud of the environment at the current moment, and in the process of prediction, the present invention can be utilized by using a recurrent neural network model. The entire environment is modeled (ie, coded) in order to predict the 3D point cloud of the environment at subsequent times and thereby accurately track future changes in a moving object within the environment. In addition, the present invention can also apply the operational characteristics of sparse convolution, so the present invention can effectively process complex three-dimensional point cloud information in a reasonable time and computing power, thereby implementing the above reconstruction and prediction. The best results.

以上所述僅為本發明之實施例,其並非用以侷限本發明之專利範圍。The above description is only an embodiment of the present invention, and is not intended to limit the scope of the invention.

1‧‧‧三維點雲追蹤裝置1‧‧‧3D point cloud tracking device

11‧‧‧輸入/輸出介面11‧‧‧Input/Output Interface

13‧‧‧處理器13‧‧‧ Processor

15‧‧‧儲存器15‧‧‧Storage

S、S(t)‧‧‧觀測三維點雲S, S (t) ‧ ‧ observation three-dimensional point cloud

M‧‧‧記憶三維點雲M‧‧‧ memory three-dimensional point cloud

17‧‧‧遞歸神經網路模型17‧‧‧Recurrent neural network model

R(t)‧‧‧重建三維點雲R(t)‧‧‧Reconstruct 3D point cloud

P(t+1)‧‧‧預測三維點雲P(t+1)‧‧‧ forecast 3D point cloud

M1‧‧‧第一記憶三維點雲M1‧‧‧ first memory 3D point cloud

M2‧‧‧第二記憶三維點雲M2‧‧‧Second memory 3D point cloud

SP1~SP5‧‧‧稀疏捲積SP1~SP5‧‧‧Sparse Convolution

Q1(t)~Q3(t)‧‧‧運算三維點雲Q1(t)~Q3(t)‧‧‧ computing three-dimensional point cloud

C1、C2‧‧‧三維點雲C1, C2‧‧‧3D point cloud

K1‧‧‧第一稀疏捲積核K1‧‧‧first sparse convolution kernel

K2‧‧‧第二稀疏捲積核K2‧‧‧Second sparse convolution kernel

f‧‧‧權重自定函數 f ‧‧‧weight custom function

p‧‧‧權重向量P‧‧‧weight vector

S410~S450、S441~S445、S451~S453‧‧‧流程步驟S410~S450, S441~S445, S451~S453‧‧‧ Process steps

圖1是本發明實施例所提供的利用遞歸神經網路的三維點雲追蹤裝置的功能方塊示意圖。 圖2是圖1的三維點雲追蹤裝置中的處理器的運作示意圖。 圖3A是圖2的遞歸神經網路模型中進行環境重建運算的運作示意圖。 圖3B是圖3A的環境重建運算中更新第一或第二記憶三維點雲的一較佳實施例下的運作示意圖。 圖3C是圖2的遞歸神經網路模型中進行環境預測運算的運作示意圖。 圖4是本發明實施例所提供的利用遞歸神經網路的三維點雲追蹤方法的流程示意圖。 圖5A是圖4的三維點雲追蹤方法中利用遞歸神經網路模型進行環境重建運算的流程示意圖。 圖5B是圖4的三維點雲追蹤方法中利用遞歸神經網路模型進行環境預測運算的流程示意圖。FIG. 1 is a functional block diagram of a three-dimensional point cloud tracking device using a recurrent neural network according to an embodiment of the present invention. 2 is a schematic diagram of the operation of a processor in the three-dimensional point cloud tracking device of FIG. 1. FIG. 3A is a schematic diagram of the operation of performing an environment reconstruction operation in the recurrent neural network model of FIG. 2. FIG. FIG. 3B is a schematic diagram of the operation of a preferred embodiment of updating the first or second memory three-dimensional point cloud in the environment reconstruction operation of FIG. 3A. FIG. 3C is a schematic diagram of the operation of the environment prediction operation in the recurrent neural network model of FIG. 2. FIG. 4 is a schematic flow chart of a three-dimensional point cloud tracking method using a recurrent neural network according to an embodiment of the present invention. FIG. 5A is a schematic flow chart of an environment reconstruction operation using a recurrent neural network model in the three-dimensional point cloud tracking method of FIG. 4. FIG. FIG. 5B is a schematic flow chart of the environment prediction operation using the recurrent neural network model in the three-dimensional point cloud tracking method of FIG. 4. FIG.

Claims (14)

一種利用遞歸神經網路的三維點雲追蹤裝置,包括: 一輸入/輸出介面,用來接收一環境在不同時刻下的不同觀測三維點雲,其中該些觀測三維點雲係由至少一光達所掃描取得; 一儲存器,用來儲存有關該環境的至少一記憶三維點雲;以及 一處理器,分別與該輸入/輸出介面及該儲存器電性連接,用來接收該些觀測三維點雲及該至少一記憶三維點雲,並且當收到該環境在一第一時刻下的該觀測三維點雲時,該處理器利用至少一遞歸神經網路模型,來對該觀測三維點雲及該至少一記憶三維點雲進行環境重建運算,以得到該環境在該第一時刻下的一重建三維點雲,並且再利用該遞歸神經網路模型,來對該至少一記憶三維點雲及一空白三維點雲進行環境預測運算,以得到該環境在一第二時刻下的一預測三維點雲,其中該第二時刻係晚於該第一時刻。A three-dimensional point cloud tracking device using a recurrent neural network, comprising: an input/output interface for receiving different observed three-dimensional point clouds of an environment at different times, wherein the three-dimensional point cloud systems are at least one light Obtaining a memory for storing at least one memory three-dimensional point cloud of the environment; and a processor electrically connected to the input/output interface and the storage for receiving the observed three-dimensional points a cloud and the at least one memory three-dimensional point cloud, and when receiving the observed three-dimensional point cloud of the environment at a first time, the processor utilizes at least one recurrent neural network model to observe the three-dimensional point cloud and Performing an environment reconstruction operation on the at least one memory three-dimensional point cloud to obtain a reconstructed three-dimensional point cloud of the environment at the first moment, and reusing the recursive neural network model to the at least one memory three-dimensional point cloud and one The blank three-dimensional point cloud performs an environment prediction operation to obtain a predicted three-dimensional point cloud of the environment at a second moment, wherein the second moment is later than the first moment. 如請求項第1項所述的三維點雲追蹤裝置,其中該至少一記憶三維點雲包括一第一記憶三維點雲及一第二記憶三維點雲,並且在利用該遞歸神經網路模型,來對該觀測三維點雲及該至少一記憶三維點雲進行該環境重建運算,以得到該環境在該第一時刻下的該重建三維點雲的步驟中,包括: 對該觀測三維點雲進行一第一稀疏捲積運算,以得到一第一運算三維點雲; 將該第一運算三維點雲與該第一記憶三維點雲進行一第二稀疏捲積運算,以得到一第二運算三維點雲,並且以該第二運算三維點雲更新該第一記憶三維點雲;以及 將該第二運算三維點雲與該第二記憶三維點雲進行一第三稀疏捲積運算,以得到該環境在該第一時刻下的該重建三維點雲,並且以該重建三維點雲更新該第二記憶三維點雲。The three-dimensional point cloud tracking device of claim 1, wherein the at least one memory three-dimensional point cloud comprises a first memory three-dimensional point cloud and a second memory three-dimensional point cloud, and using the recurrent neural network model, Performing the environment reconstruction operation on the observed three-dimensional point cloud and the at least one memory three-dimensional point cloud to obtain the reconstructed three-dimensional point cloud in the environment at the first moment, comprising: performing the observed three-dimensional point cloud a first sparse convolution operation to obtain a first computing three-dimensional point cloud; performing a second sparse convolution operation on the first computing three-dimensional point cloud and the first memory three-dimensional point cloud to obtain a second computing three-dimensional Point cloud, and updating the first memory three-dimensional point cloud with the second computing three-dimensional point cloud; and performing a third sparse convolution operation on the second computing three-dimensional point cloud and the second memory three-dimensional point cloud to obtain the The environment reconstructs the three-dimensional point cloud at the first moment, and updates the second memory three-dimensional point cloud with the reconstructed three-dimensional point cloud. 如請求項第2項所述的三維點雲追蹤裝置,其中在利用該遞歸神經網路模型,來對該至少一記憶三維點雲及該空白三維點雲進行環境預測運算,以得到該環境在該第二時刻下的該預測三維點雲的步驟中,包括: 將該空白三維點雲與該第一記憶三維點雲進行一第四稀疏捲積運算,以得到一第三運算三維點雲;以及 將該第三運算三維點雲與該第二記憶三維點雲進行一第五稀疏捲積運算,以得到該環境在該第二時刻下的該預測三維點雲。The three-dimensional point cloud tracking device of claim 2, wherein the recursive neural network model is used to perform an environment prediction operation on the at least one memory three-dimensional point cloud and the blank three-dimensional point cloud to obtain the environment. The step of predicting the three-dimensional point cloud in the second moment includes: performing a fourth sparse convolution operation on the blank three-dimensional point cloud and the first memory three-dimensional point cloud to obtain a third computing three-dimensional point cloud; And performing a fifth sparse convolution operation on the third computing three-dimensional point cloud and the second memory three-dimensional point cloud to obtain the predicted three-dimensional point cloud of the environment at the second moment. 如請求項第3項所述的三維點雲追蹤裝置,其中該處理器更用以定義至少一權重自定函數、至少一第一稀疏捲積核及至少一第二稀疏捲積核,並且在以該第二運算三維點雲更新該第一記憶三維點雲的步驟中,更包括: 利用該權重自定函數,來由該第一記憶三維點雲、該第二運算三維點雲、該第一稀疏捲積核及該第二稀疏捲積核中,決定一權重向量,並且將該第一記憶三維點雲更新為該第一記憶三維點雲、該第二運算三維點雲及該權重向量代入一權重方程式後的結果。The three-dimensional point cloud tracking device of claim 3, wherein the processor is further configured to define at least one weight custom function, at least one first sparse convolution kernel, and at least one second sparse convolution kernel, and And the step of updating the first memory three-dimensional point cloud by using the second computing three-dimensional point cloud, further comprising: using the weight custom function to generate the first memory three-dimensional point cloud, the second computing three-dimensional point cloud, the first a sparse convolution kernel and the second sparse convolution kernel, determining a weight vector, and updating the first memory three-dimensional point cloud to the first memory three-dimensional point cloud, the second computing three-dimensional point cloud, and the weight vector The result of substituting a weight equation. 如請求項第4項所述的三維點雲追蹤裝置,其中在以該重建三維點雲更新該第二記憶三維點雲的步驟中,更包括: 利用該權重自定函數,來由該第二記憶三維點雲、該重建三維點雲、該第一稀疏捲積核及該第二稀疏捲積核中,決定該權重向量,並且將該第二記憶三維點雲更新為該第二記憶三維點雲、該重建三維點雲及該權重向量代入該權重方程式後的結果。The third-dimensional point cloud tracking device of claim 4, wherein, in the step of updating the second memory three-dimensional point cloud with the reconstructed three-dimensional point cloud, the method further comprises: using the weight custom function to be used by the second Determining the weight vector in the memory three-dimensional point cloud, the reconstructed three-dimensional point cloud, the first sparse convolution kernel, and the second sparse convolution kernel, and updating the second memory three-dimensional point cloud to the second memory three-dimensional point The cloud, the reconstructed three-dimensional point cloud, and the result of the weight vector being substituted into the weighting equation. 如請求項第5項所述的三維點雲追蹤裝置,其中該權重自定函數、該第一稀疏捲積核及該第二稀疏捲積核係經由該三維點雲追蹤裝置在進行完一訓練模式後所定義,且該權重向量的分量係介於0至1之間。The three-dimensional point cloud tracking device of claim 5, wherein the weight self-determination function, the first sparse convolution kernel, and the second sparse convolution kernel are performed through the three-dimensional point cloud tracking device. The mode is defined after the mode, and the component of the weight vector is between 0 and 1. 如請求項第6項所述的三維點雲追蹤裝置,其中該權重方程式為p*C1+(1-p)*C2,其中p為該權重向量,且C1及C2分別為該第一記憶三維點雲及該第二運算三維點雲,或該第二記憶三維點雲及該重建三維點雲。The three-dimensional point cloud tracking device of claim 6, wherein the weighting equation is p*C1+(1-p)*C2, where p is the weight vector, and C1 and C2 are respectively the first memory three-dimensional point The cloud and the second computing three-dimensional point cloud, or the second memory three-dimensional point cloud and the reconstructed three-dimensional point cloud. 一種利用遞歸神經網路的三維點雲追蹤方法,執行於一三維點雲追蹤裝置中,該三維點雲追蹤裝置包括一輸入/輸出介面、一儲存器以及一處理器,該三維點雲追蹤方法包括: 令該輸入/輸出介面接收一環境在不同時刻下的不同觀測三維點雲,其中該些觀測三維點雲係由至少一光達所掃描取得; 令該儲存器儲存有關該環境的至少一記憶三維點雲;以及 令該處理器接收該些觀測三維點雲及該至少一記憶三維點雲,並且當收到該環境在一第一時刻下的該觀測三維點雲時,令該處理器利用至少一遞歸神經網路模型,來對該觀測三維點雲及該至少一記憶三維點雲進行環境重建運算,以得到該環境在該第一時刻下的一重建三維點雲,並且令該處理器再利用該遞歸神經網路模型,來對該至少一記憶三維點雲及一空白三維點雲進行環境預測運算,以得到該環境在第二時刻下的一預測三維點雲,其中該第二時刻係晚於該第一時刻。A three-dimensional point cloud tracking method using a recurrent neural network is implemented in a three-dimensional point cloud tracking device, the three-dimensional point cloud tracking device comprising an input/output interface, a storage and a processor, and the three-dimensional point cloud tracking method The method includes: causing the input/output interface to receive a different observed three-dimensional point cloud of an environment at different times, wherein the observed three-dimensional point cloud is obtained by scanning at least one light; and causing the storage to store at least one of the environment Memorizing the three-dimensional point cloud; and causing the processor to receive the observed three-dimensional point cloud and the at least one memory three-dimensional point cloud, and when the environment receives the observed three-dimensional point cloud at a first time, the processor is Performing an environment reconstruction operation on the observed three-dimensional point cloud and the at least one memory three-dimensional point cloud by using at least one recurrent neural network model to obtain a reconstructed three-dimensional point cloud of the environment at the first moment, and causing the processing Using the recursive neural network model, the environment prediction operation is performed on the at least one memory three-dimensional point cloud and a blank three-dimensional point cloud to obtain the environment. He predicted a 3D point cloud in two moments, wherein the first time point in the time line of the second night. 如請求項第8項所述的三維點雲追蹤方法,其中該至少一記憶三維點雲包括一第一記憶三維點雲及一第二記憶三維點雲,並且在利用該遞歸神經網路模型,來對該觀測三維點雲及該至少一記憶三維點雲進行該環境重建運算,以得到該環境在該第一時刻下的該重建三維點雲的步驟中,包括: 對該觀測三維點雲進行一第一稀疏捲積運算,以得到一第一運算三維點雲; 將該第一運算三維點雲與該第一記憶三維點雲進行一第二稀疏捲積運算,以得到一第二運算三維點雲,並且以該第二運算三維點雲更新該第一記憶三維點雲;以及 將該第二運算三維點雲與該第二記憶三維點雲進行一第三稀疏捲積運算,以得到該環境在該第一時刻下的該重建三維點雲,並且以該重建三維點雲更新該第二記憶三維點雲。The method of claim 3, wherein the at least one memory three-dimensional point cloud comprises a first memory three-dimensional point cloud and a second memory three-dimensional point cloud, and using the recurrent neural network model, Performing the environment reconstruction operation on the observed three-dimensional point cloud and the at least one memory three-dimensional point cloud to obtain the reconstructed three-dimensional point cloud in the environment at the first moment, comprising: performing the observed three-dimensional point cloud a first sparse convolution operation to obtain a first computing three-dimensional point cloud; performing a second sparse convolution operation on the first computing three-dimensional point cloud and the first memory three-dimensional point cloud to obtain a second computing three-dimensional Point cloud, and updating the first memory three-dimensional point cloud with the second computing three-dimensional point cloud; and performing a third sparse convolution operation on the second computing three-dimensional point cloud and the second memory three-dimensional point cloud to obtain the The environment reconstructs the three-dimensional point cloud at the first moment, and updates the second memory three-dimensional point cloud with the reconstructed three-dimensional point cloud. 如請求項第9項所述的三維點雲追蹤方法,其中在利用該遞歸神經網路模型,來對該至少一記憶三維點雲及該空白三維點雲進行環境預測運算,以得到該環境在該第二時刻下的該預測三維點雲的步驟中,包括: 將該空白三維點雲與該第一記憶三維點雲進行一第四稀疏捲積運算,以得到一第三運算三維點雲;以及 將該第三運算三維點雲與該第二記憶三維點雲進行一第五稀疏捲積運算,以得到該環境在該第二時刻下的該預測三維點雲。The method of claim 3, wherein the recursive neural network model is used to perform an environment prediction operation on the at least one memory three-dimensional point cloud and the blank three-dimensional point cloud to obtain the environment. The step of predicting the three-dimensional point cloud in the second moment includes: performing a fourth sparse convolution operation on the blank three-dimensional point cloud and the first memory three-dimensional point cloud to obtain a third computing three-dimensional point cloud; And performing a fifth sparse convolution operation on the third computing three-dimensional point cloud and the second memory three-dimensional point cloud to obtain the predicted three-dimensional point cloud of the environment at the second moment. 如請求項第10項所述的三維點雲追蹤方法,其中該處理器更用以定義至少一權重自定函數、至少一第一稀疏捲積核及至少一第二稀疏捲積核,並且在以該第二運算三維點雲更新該第一記憶三維點雲的步驟中,更包括: 利用該權重自定函數,來由該第一記憶三維點雲、該第二運算三維點雲、該第一稀疏捲積核及該第二稀疏捲積核中,決定一權重向量,並且將該第一記憶三維點雲更新為該第一記憶三維點雲、該第二運算三維點雲及該權重向量代入一權重方程式後的結果。The method of claim 3, wherein the processor is further configured to define at least one weight custom function, at least one first sparse convolution kernel, and at least one second sparse convolution kernel, and And the step of updating the first memory three-dimensional point cloud by using the second computing three-dimensional point cloud, further comprising: using the weight custom function to generate the first memory three-dimensional point cloud, the second computing three-dimensional point cloud, the first a sparse convolution kernel and the second sparse convolution kernel, determining a weight vector, and updating the first memory three-dimensional point cloud to the first memory three-dimensional point cloud, the second computing three-dimensional point cloud, and the weight vector The result of substituting a weight equation. 如請求項第11項所述的三維點雲追蹤方法,其中在以該重建三維點雲更新該第二記憶三維點雲的步驟中,更包括: 利用該權重自定函數,來由該第二記憶三維點雲、該重建三維點雲、該第一稀疏捲積核及該第二稀疏捲積核中,決定該權重向量,並且將該第二記憶三維點雲更新為該第二記憶三維點雲、該重建三維點雲及該權重向量代入該權重方程式後的結果。The method of claim 3, wherein in the step of updating the second three-dimensional point cloud with the reconstructed three-dimensional point cloud, the method further comprises: using the weighting function to obtain the second Determining the weight vector in the memory three-dimensional point cloud, the reconstructed three-dimensional point cloud, the first sparse convolution kernel, and the second sparse convolution kernel, and updating the second memory three-dimensional point cloud to the second memory three-dimensional point The cloud, the reconstructed three-dimensional point cloud, and the result of the weight vector being substituted into the weighting equation. 如請求項第12項所述的三維點雲追蹤方法,其中該權重自定函數、該第一稀疏捲積核及該第二稀疏捲積核係經由該三維點雲追蹤裝置在進行完一訓練模式後所定義,且該權重向量的分量係介於0至1之間。The three-dimensional point cloud tracking method of claim 12, wherein the weight self-determination function, the first sparse convolution kernel, and the second sparse convolution kernel are performed through the three-dimensional point cloud tracking device. The mode is defined after the mode, and the component of the weight vector is between 0 and 1. 如請求項第13項所述的三維點雲追蹤方法,其中該權重方程式為p*C1+(1-p)*C2,其中p為該權重向量,且C1及C2分別為該第一記憶三維點雲及該第二運算三維點雲,或該第二記憶三維點雲及該重建三維點雲。The method of claim 3, wherein the weighting equation is p*C1+(1-p)*C2, where p is the weight vector, and C1 and C2 are the first memory three-dimensional point, respectively. The cloud and the second computing three-dimensional point cloud, or the second memory three-dimensional point cloud and the reconstructed three-dimensional point cloud.
TW106143026A 2017-12-07 2017-12-07 Three-dimensional point cloud tracking apparatus and method by recurrent neural network TWI657407B (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
TW106143026A TWI657407B (en) 2017-12-07 2017-12-07 Three-dimensional point cloud tracking apparatus and method by recurrent neural network
CN201810043041.7A CN109901189B (en) 2017-12-07 2018-01-17 Three-dimensional point cloud tracking device and method using recurrent neural network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
TW106143026A TWI657407B (en) 2017-12-07 2017-12-07 Three-dimensional point cloud tracking apparatus and method by recurrent neural network

Publications (2)

Publication Number Publication Date
TWI657407B TWI657407B (en) 2019-04-21
TW201926258A true TW201926258A (en) 2019-07-01

Family

ID=66943202

Family Applications (1)

Application Number Title Priority Date Filing Date
TW106143026A TWI657407B (en) 2017-12-07 2017-12-07 Three-dimensional point cloud tracking apparatus and method by recurrent neural network

Country Status (2)

Country Link
CN (1) CN109901189B (en)
TW (1) TWI657407B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11100702B2 (en) 2019-12-27 2021-08-24 Industrial Technology Research Institute 3D image labeling method based on labeling information of 2D image and 3D image labeling device
CN112623027A (en) * 2020-12-26 2021-04-09 浙江天行健智能科技有限公司 Road feel simulation method based on BP neural network

Family Cites Families (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020164067A1 (en) * 2001-05-02 2002-11-07 Synapix Nearest neighbor edge selection from feature tracking
CN101833786B (en) * 2010-04-06 2011-12-28 清华大学 Method and system for capturing and rebuilding three-dimensional model
US8718380B2 (en) * 2011-02-14 2014-05-06 Mitsubishi Electric Research Laboratories, Inc. Representing object shapes using radial basis function support vector machine classification
CN104123730B (en) * 2014-07-31 2016-09-14 武汉大学 Remote sensing image based on roadway characteristic and laser point cloud method for registering and system
US9285805B1 (en) * 2015-07-02 2016-03-15 Geodigital International Inc. Attributed roadway trajectories for self-driving vehicles
US9710714B2 (en) * 2015-08-03 2017-07-18 Nokia Technologies Oy Fusion of RGB images and LiDAR data for lane classification
CN105678689B (en) * 2015-12-31 2020-01-31 百度在线网络技术(北京)有限公司 High-precision map data registration relation determining method and device
CN107025642B (en) * 2016-01-27 2018-06-22 百度在线网络技术(北京)有限公司 Vehicle's contour detection method and device based on point cloud data
CN106097334B (en) * 2016-06-08 2023-09-12 北京航空航天大学 Point cloud registration method based on deep learning
US9836853B1 (en) * 2016-09-06 2017-12-05 Gopro, Inc. Three-dimensional convolutional neural networks for video highlight detection
CN206691107U (en) * 2017-03-08 2017-12-01 深圳市速腾聚创科技有限公司 Pilotless automobile system and automobile
CN106951847B (en) * 2017-03-13 2020-09-29 百度在线网络技术(北京)有限公司 Obstacle detection method, apparatus, device and storage medium
CN107226087B (en) * 2017-05-26 2019-03-26 西安电子科技大学 A kind of structured road automatic Pilot transport vehicle and control method
CN107423679A (en) * 2017-05-31 2017-12-01 深圳市鸿逸达科技有限公司 A kind of pedestrian is intended to detection method and system

Also Published As

Publication number Publication date
CN109901189B (en) 2023-07-07
CN109901189A (en) 2019-06-18
TWI657407B (en) 2019-04-21

Similar Documents

Publication Publication Date Title
US10705216B2 (en) Three-dimensional point cloud tracking apparatus and method using recurrent neural network
US10984545B2 (en) Estimating depth for a video stream captured with a monocular rgb camera
Ming et al. Deep learning for monocular depth estimation: A review
Dong et al. Towards real-time monocular depth estimation for robotics: A survey
EP3427008B1 (en) Laser scanner with real-time, online ego-motion estimation
CN108230361B (en) Method and system for enhancing target tracking by fusing unmanned aerial vehicle detector and tracker
US11520037B2 (en) Perception system
US9193075B1 (en) Apparatus and methods for object detection via optical flow cancellation
CN108897836B (en) Method and device for robot to map based on semantics
US11074438B2 (en) Disentangling human dynamics for pedestrian locomotion forecasting with noisy supervision
CN110646787A (en) Self-motion estimation method and device and model training method and device
JP2021532373A (en) Positioning system, positioning method and storage medium
Henson et al. Attitude-trajectory estimation for forward-looking multibeam sonar based on acoustic image registration
US11822621B2 (en) Systems and methods for training a machine-learning-based monocular depth estimator
CN108367436B (en) Active camera movement determination for object position and range in three-dimensional space
US20200265294A1 (en) Object Animation Using Generative Neural Networks
CN111739066B (en) Visual positioning method, system and storage medium based on Gaussian process
JP7332238B2 (en) Methods and Apparatus for Physics-Guided Deep Multimodal Embedding for Task-Specific Data Utilization
US20220277581A1 (en) Hand pose estimation method, device and storage medium
TWI657407B (en) Three-dimensional point cloud tracking apparatus and method by recurrent neural network
US20230237342A1 (en) Adaptive lookahead for planning and learning
WO2020213099A1 (en) Object detection/tracking device, method, and program recording medium
Abdulaaty et al. Real-time depth completion using radar and camera
JP7310927B2 (en) Object tracking device, object tracking method and recording medium
Mandal et al. Unsupervised Learning of Depth, Camera Pose and Optical Flow from Monocular Video