TWI485651B - Method for depth estimation and device usnig the same - Google Patents

Method for depth estimation and device usnig the same Download PDF

Info

Publication number
TWI485651B
TWI485651B TW100108826A TW100108826A TWI485651B TW I485651 B TWI485651 B TW I485651B TW 100108826 A TW100108826 A TW 100108826A TW 100108826 A TW100108826 A TW 100108826A TW I485651 B TWI485651 B TW I485651B
Authority
TW
Taiwan
Prior art keywords
data
depth
timing
frame data
time
Prior art date
Application number
TW100108826A
Other languages
Chinese (zh)
Other versions
TW201237805A (en
Inventor
Houng Jyh Wang
Chih Wei Kao
Tzu Hung Chen
Original Assignee
Teco Elec & Machinery Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Teco Elec & Machinery Co Ltd filed Critical Teco Elec & Machinery Co Ltd
Priority to TW100108826A priority Critical patent/TWI485651B/en
Publication of TW201237805A publication Critical patent/TW201237805A/en
Application granted granted Critical
Publication of TWI485651B publication Critical patent/TWI485651B/en

Links

Landscapes

  • Image Analysis (AREA)
  • Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)

Description

深度資料估計方法及其裝置Depth data estimation method and device thereof

本發明是有關於一種深度估計方法及其裝置,且特別是應用雙視域影像比對(Stereo Matching)技術來進行估計之深度估計方法及其裝置。The present invention relates to a depth estimation method and apparatus thereof, and more particularly to a depth estimation method and apparatus for estimating using a dual-view image matching (Stereo Matching) technique.

在科技發展日新月異的現今時代中,立體影像多媒體系統逐漸被業界所重視。一般來說,在立體影像/視訊的應用中,雙視域影像比對(Stereo Matching)影像處理技術,是目前業界急需開發的立體影像核心技術。在現有技術中,雙視域影像比對技術係先根據雙視域影像計算出影像深度分佈圖。In today's fast-changing technology era, stereoscopic multimedia systems are gradually being valued by the industry. In general, in the application of stereoscopic video/video, the dual-view image matching (Stereo Matching) image processing technology is a stereoscopic image core technology that is urgently needed in the industry. In the prior art, the dual-view image comparison technique first calculates an image depth map according to the dual-view image.

一般來說,雙視域影像比對技術具有計算複雜度較高之問題。據此,如何設計出具有計算複雜度較低之雙視域影像比對深度估計方法為業界不斷致力的方向之一。In general, the dual-view image comparison technique has a problem of high computational complexity. Based on this, how to design a dual-view image comparison depth estimation method with low computational complexity is one of the directions that the industry is constantly striving for.

本發明係有關於一種深度估計方法及其裝置,相較於傳統深度估計方法,本發明相關之深度估計方法及其裝置具有運算複雜度較低之優點。The present invention relates to a depth estimation method and apparatus thereof, and the depth estimation method and apparatus thereof according to the present invention have the advantage of low computational complexity compared to the conventional depth estimation method.

根據本發明(之第一方面),提出一種深度估計方法,用以針對輸入雙眼影像(Binocular Video)資料進行深度估計,深度估計方法包括下列步驟。首先接收輸入雙眼影像資料中對應至第一及第二時間序的第一及第二時序圖框資料組,其中各第一及第二時序圖框資料組包括一第一視角及第二視角圖框資料。然後找出第二時序圖框資料組相對於第一時序圖框資料組之第一移動向量資料。接著根據第一時序圖框資料組之第一視角及第二視角圖框資料進行雙視域影像比對(Stereo Matching),以找出對應至第一時間序的第一時序深度資料。然後根據移動向量資料及第一時序深度資料找出第二時序估計深度資料。之後根據第二時序估計深度資料、第二時序圖框資料組之第一視角及第二視角圖框資料,找出第二時序深度資料。According to a first aspect of the present invention, a depth estimation method is proposed for performing depth estimation on input binocular video data, and the depth estimation method comprises the following steps. First receiving the first and second timing frame data sets corresponding to the first and second time sequences in the input binocular image data, wherein each of the first and second timing frame data sets includes a first viewing angle and a second viewing angle Frame information. Then, the first moving vector data of the second timing frame data group relative to the first timing frame data group is found. Then, the two-view image matching (Stereo Matching) is performed according to the first view and the second view frame data of the first time series frame data group to find the first time-series depth data corresponding to the first time sequence. Then, the second time series estimated depth data is found according to the motion vector data and the first timing depth data. Then, according to the second timing estimation depth data, the first viewing angle of the second timing frame data group, and the second viewing angle frame data, the second timing depth data is found.

根據本發明(之第二方面),提出一種深度資料估計裝置,用以針對輸入雙眼影像資料進行深度估計,其中包括輸入單元、移動向量產生單元、雙視域影像比對單元及深度估計單元。輸入單元接收輸入雙眼影像資料中對應至第一時間序的第一時序圖框資料組及對應至第二時間序的第二時序圖框資料組,其中各第一及第二時序圖框資料組包括第一視角圖框資料及第二視角圖框資料。移動向量產生單元找出第二時序圖框資料組相對於第一時序圖框資料組之第一移動向量資料。雙視域影像比對單元根據第一時序圖框資料組之第一視角及第二視角圖框資料進行雙視域影像比對,以找出對應至第一時間序的第一時序深度資料。深度估計單元根據移動向量資料及第一時序深度資料找出第二時序估計深度資料。雙視域影像比對單元根據第二時序估計深度資料、第二時序圖框資料組之第一視角及第二視角圖框資料,找出第二時序深度資料。According to a second aspect of the present invention, a depth data estimating apparatus is provided for performing depth estimation on input binocular image data, including an input unit, a motion vector generating unit, a dual-view image matching unit, and a depth estimating unit. . The input unit receives a first timing frame data group corresponding to the first time sequence and a second timing frame data group corresponding to the second time sequence in the input binocular image data, wherein each of the first and second timing frames The data set includes first view frame data and second view frame data. The motion vector generation unit finds the first motion vector data of the second timing frame data group with respect to the first timing frame data group. The dual-view image comparison unit performs dual-view image comparison according to the first view and the second view frame data of the first time frame data group to find the first time depth corresponding to the first time sequence data. The depth estimating unit finds the second time series estimated depth data according to the motion vector data and the first timing depth data. The dual-view image comparison unit finds the second time-series depth data according to the second time-series estimation depth data, the first view angle of the second time-series frame data group, and the second view frame data.

為了對本發明之上述及其他方面有更佳的瞭解,下文特舉較佳實施例,並配合所附圖式,作詳細說明如下:In order to better understand the above and other aspects of the present invention, the preferred embodiments are described below, and in conjunction with the drawings, the detailed description is as follows:

本實施例之深度估計方法係參考輸入之雙眼影像(Binocular Video)資料之移動向量資料來降低雙視域影像比對(Stereo Matching)之運算量。The depth estimation method of this embodiment reduces the computational complexity of the dual-view image matching (Stereo Matching) by referring to the motion vector data of the input binocular video data.

第一實施例First embodiment

本實施例之深度估計方法係參考輸入之雙眼影像資料之移動向量資料來估產生估計深度資料,並參考此估計深度資料來簡化其產生對應之深度資料的操作。The depth estimation method in this embodiment estimates the estimated depth data by referring to the motion vector data of the input binocular image data, and refers to the estimated depth data to simplify the operation of generating the corresponding depth data.

請參照第1圖,其繪示依照本發明第一實施例之深度估計裝置的方塊圖。本實施例之深度估計裝置1用以針對輸入雙眼影像(Binocular Video)資料Vi進行深度估計。輸入雙眼影像資料Vi例如包括複數個雙視域圖框資料組,該複數個雙視域圖框資料組係可與各個時間序相對應,而於各雙視域圖框資料組各包括兩筆對應至不同視角的圖框資料。在其他例子中,輸入雙眼影像資料Vi更可包括三個或三個以上之多視角圖框資料。Referring to FIG. 1, a block diagram of a depth estimating apparatus according to a first embodiment of the present invention is shown. The depth estimating device 1 of the present embodiment is configured to perform depth estimation on the input binocular video data Vi. The input binocular image data Vi includes, for example, a plurality of dual-view frame data sets, the plurality of double-view frame data sets may correspond to respective time sequences, and each of the double-view frame data sets includes two The pen corresponds to the frame data of different perspectives. In other examples, the input binocular image data Vi may further include three or more multi-view frame data.

舉例來說,輸入雙眼影像資料Vi中對應至第一時間序t1之時序圖框資料組Vi_t1包括第一視角圖框資料F1_t1及第二視角圖框資料F2_t1;輸入雙眼影像資料Vi中對應至第二時間序t2之時序圖框資料組Vi_t2包括第一視角圖框資料F1_t2及第二視角圖框資料F2_t2。其中,第一視角圖框資料F1_t1及F1_t2例如分別為第一及第二時間序t1及t2時之左眼視角圖框資料;第二視角圖框資料F2_t1及F2_t2例如分別為第一及第二時間序t1及t2時之右眼視角圖框資料;第一及第二時間序t1及t2例如對應至互相鄰近的操作時點。For example, the timing frame data group Vi_t1 corresponding to the first time sequence t1 in the input binocular image data Vi includes the first view frame data F1_t1 and the second view frame data F2_t1; and the input binocular image data Vi corresponds to The timing frame data set Vi_t2 to the second time sequence t2 includes a first view frame data F1_t2 and a second view frame data F2_t2. The first view frame data F1_t1 and F1_t2 are, for example, left eye view frame data in the first and second time sequences t1 and t2, respectively; the second view frame data F2_t1 and F2_t2 are, for example, first and second, respectively. The right eye view frame data in time sequence t1 and t2; the first and second time sequences t1 and t2 correspond, for example, to operating time points adjacent to each other.

深度估計裝置1包括輸入單元102、移動向量產生單元104及雙視域影像比對單元108。The depth estimation device 1 includes an input unit 102, a motion vector generation unit 104, and a dual view image comparison unit 108.

輸入單元102接收時序圖框資料組Vi_t1及Vi_t2,其中包括輸入雙眼影像資料Vi中對應至第一及第二時間序t1及t2之第一視角圖框資料F1_t1及F1_t2與第二視角圖框資料F2_t1及F2_t2。移動向量產生單元104找出時序圖框資料組Vi_t2相對於時序圖框資料組Vi_t1之第一移動向量資料M_12。The input unit 102 receives the timing frame data sets Vi_t1 and Vi_t2, and includes the first view frame data F1_t1 and F1_t2 and the second view frame corresponding to the first and second time sequences t1 and t2 in the input binocular image data Vi. Information F2_t1 and F2_t2. The motion vector generation unit 104 finds the first motion vector data M_12 of the timing frame data group Vi_t2 with respect to the timing frame data group Vi_t1.

在一個操作實例中,輸入單元102及移動向量產生單元104可以影像解壓縮器來實現;影像解壓縮器針對輸入之壓縮雙眼影像資料進行解壓縮操作,以還原未壓縮之原始影像資料及其對應之移動向量資料,該輸入之壓縮雙眼影像資料其可為左右眼視角圖框資料彼此獨立之兩筆影像資料或一個同時儲存有左右眼視角圖框資料之一筆雙眼影像資料。在一個操作實例中,移動向量產生單元104更例如包括平均濾波器(Mean Filter),可用以對移動向量資料進行相關之影像處理操作。In an operation example, the input unit 102 and the motion vector generation unit 104 can be implemented by an image decompressor; the image decompressor performs a decompression operation on the input compressed binocular image data to restore the uncompressed original image data and Corresponding moving vector data, the input compressed binocular image data can be two image data independently of each other, and one binocular image data of one side of the left and right eye view frame data. In one example of operation, the motion vector generation unit 104 further includes, for example, an averaging filter (Mean Filter) that can be used to perform related image processing operations on the motion vector data.

雙視域影像比對單元106根據第一視角及第二視角圖框資料F1_t1及F2_t1進行雙視域影像比對,以找出對應至該第一時間序t1的第一時序深度資料D_t1。The dual-view image comparison unit 106 performs dual-view image comparison according to the first view and the second view frame data F1_t1 and F2_t1 to find the first time-series data D_t1 corresponding to the first time sequence t1.

深度估計單元108根據移動向量資料M_12及第一時序深度資料D_t1,找出對應至第二時間序t2的第二時序估計深度資料DE_t2。由於第一及第二時間序t1及t2為相互鄰近之時序時點,一般來說,輸入雙眼影像資料Vi在第一及第二時間序t1及t2中具有線性變化之深度分佈圖形,深度估計單元108係根據第一時序深度資料D_t1及移動向量資料M_12來估計得到第二時序估計深度資料DE_t2。The depth estimating unit 108 finds the second time-series estimated depth data DE_t2 corresponding to the second time sequence t2 according to the motion vector data M_12 and the first timing depth data D_t1. Since the first and second time sequences t1 and t2 are timing points adjacent to each other, in general, the input binocular image data Vi has a linearly varying depth distribution pattern in the first and second time sequences t1 and t2, and the depth estimation is performed. The unit 108 estimates the second time series estimated depth data DE_t2 according to the first time series depth data D_t1 and the motion vector data M_12.

深度估計單元108所估計得到之第二時序估計深度資料DE_t2更被提供至雙視域影像比對單元106;雙視域影像比對單元106在已有對應至第二時間序t2之第二時序估計深度資料DE_t2之基礎下,執行針對第一視角及第二視角圖框資料F1_t2及F2_t2之雙視域影像比對操作,以找出對應至第二時間序t2之第二時序深度資料D_t2。The second timing estimation depth data DE_t2 estimated by the depth estimation unit 108 is further provided to the dual-view image comparison unit 106; the dual-view image comparison unit 106 has a second timing corresponding to the second time sequence t2. Based on the estimated depth data DE_t2, the dual view image comparison operation for the first view and the second view frame data F1_t2 and F2_t2 is performed to find the second time depth depth data D_t2 corresponding to the second time sequence t2.

舉例來說,深度估計單元108係根據第二時序估計深度資料DE_t2來決定第一視角及第二視角圖框資料F1_t2及F2_t2之雙視域影像比對操作中,欲進行比對之搜尋視窗(Search Window)。在一個操作實例中,針對第一視角圖框資料F1_t2中對應至座標(i1,j1)之畫素資料P(i1,j1)來說,第二時序估計深度資料DE_t2指示其對應至深度值X,其中i1及j1為自然數。據此,雙視域影像比對單元106可對應將搜尋視窗的中心位置對應至第二視角圖框資料F2_t2中之座標位置(i1+X,j1)。For example, the depth estimation unit 108 determines, according to the second time series estimation depth data DE_t2, the search window for comparing the dual view image comparison operations of the first view and the second view frame data F1_t2 and F2_t2 ( Search Window). In an operation example, for the pixel data P(i1, j1) corresponding to the coordinates (i1, j1) in the first view frame data F1_t2, the second time series estimation depth data DE_t2 indicates that it corresponds to the depth value X. , where i1 and j1 are natural numbers. Accordingly, the dual-view image matching unit 106 can correspond to the center position of the search window to the coordinate position (i1+X, j1) in the second view frame data F2_t2.

據此,根據第二時序估計深度資料DE_t2,雙視域影像比對單元106可有效地得到相關於第一視角及第二視角圖框資料F1_t2及F2_t2間的可能深度值,藉此可有效地縮小欲進行雙視域影像比對操作之搜尋視窗的大小,並對應地降低比對操作所需之運算量。According to this, according to the second timing estimation depth data DE_t2, the dual-view image comparison unit 106 can effectively obtain the possible depth values between the first view and the second view frame data F1_t2 and F2_t2, thereby effectively The size of the search window for the dual-view image comparison operation is reduced, and the amount of calculation required for the comparison operation is correspondingly reduced.

請參照第2圖,其繪示依照本發明第一實施例之深度估計方法的流程圖。本實施例之深度估計方法包括下列之步驟。首先如步驟(a),輸入單元102接收輸入雙眼影像資料Vi中對應至第一時間序t1的時序圖框資料組Vi_t1及對應至第二時間序t2的時序圖框資料組Vi_t2。接著如步驟(b),移動向量產生單元104找出時序圖框資料組Vi_t2相對於時序圖框資料組Vi_t1之移動向量資料M_12。Referring to FIG. 2, a flow chart of a depth estimation method according to a first embodiment of the present invention is shown. The depth estimation method of this embodiment includes the following steps. First, as step (a), the input unit 102 receives the timing frame data set Vi_t1 corresponding to the first time sequence t1 and the time series frame data set Vi_t2 corresponding to the second time sequence t2 in the input binocular image data Vi. Next, as in step (b), the motion vector generation unit 104 finds the motion vector data M_12 of the timing frame data set Vi_t2 with respect to the timing frame data set Vi_t1.

然後如步驟(c),雙視域影像比對單元106根據第一視角及第二視角圖框資料F1_t1及F2_t1進行雙視域影像比對,以找出對應至第一時間序t1的第一時序深度資料D_t1。接著如步驟(d),深度估計單元108根據移動向量資料M_12及第一時序深度資料D_t1找出第二時序估計深度資料DE_t2。Then, in step (c), the dual-view image comparison unit 106 performs dual-view image comparison according to the first view and the second view frame data F1_t1 and F2_t1 to find the first corresponding to the first time sequence t1. Timing depth data D_t1. Then, as step (d), the depth estimating unit 108 finds the second time series estimated depth data DE_t2 according to the motion vector data M_12 and the first timing depth data D_t1.

之後如步驟(e),雙視域影像比對單元106根據第二時序估計深度資料DE_t2、第一視角及第二視角圖框資料F1_t2及F2_t2,找出並輸出第二時序深度資料D_t2。Then, in step (e), the dual view image comparison unit 106 finds and outputs the second time series depth data D_t2 according to the second time series estimation depth data DE_t2, the first view and the second view frame data F1_t2 and F2_t2.

在本實施例中,雖僅以本實施例之深度估計裝置1參考對應至第一時間序t1之第一時序深度資料D_t1來估計出對應至第二時間序t2之第二時序估計深度資料DE_t2,並據以簡化其找出對應至第二時序深度資料D_t2之操作的情形為例做說明,然,本實施例之深度估計裝置1並不侷限於此。In this embodiment, the depth estimation device 1 of the present embodiment estimates the second time-series estimated depth data corresponding to the second time sequence t2 with reference to the first time-series depth data D_t1 corresponding to the first time sequence t1. DE_t2 is used as an example to simplify the operation of finding the operation corresponding to the second timing depth data D_t2. However, the depth estimating apparatus 1 of the present embodiment is not limited thereto.

舉例來說,深度估計裝置1’中之輸入單元102’更用以接收時序圖框資料組Vi_t3;移動向量產生單元104’更用以產生時序圖框資料組Vi_t3相對於時序圖框資料組Vi_t2之移動向量資料M_23;深度估計單元108’更用以參考移動向量資料M_12及M23及對應至第一、第二時間序t1及t2之第一、第二時序深度資料D_t1及D_t2估計出對應至第三時間序t3之第三時序估計深度資料DE_t3;雙視域影像比對單元106’參考第三時序估計深度資料DE_t3以簡化其找出對應至第三時序深度資料D_t3之操作,如第3圖所示。於此實施例中,第三時間序t3為介於第一及第二時間序t1及t2間的操作時序,然並不以此為限,舉例而言,第三時間序t3亦可為於第二時間序t2之後的操作時序。For example, the input unit 102' in the depth estimating device 1' is further configured to receive the timing frame data set Vi_t3; the motion vector generating unit 104' is further configured to generate the timing frame data set Vi_t3 relative to the timing frame data set Vi_t2 The motion vector data M_23; the depth estimating unit 108' is further configured to refer to the motion vector data M_12 and M23 and the first and second timing depth data D_t1 and D_t2 corresponding to the first and second time sequences t1 and t2 to estimate corresponding to The third time series estimation depth data DE_t3 of the third time sequence t3; the dual view image comparison unit 106' refers to the third time series estimation depth data DE_t3 to simplify the operation of finding the corresponding to the third time series depth data D_t3, such as the third The figure shows. In this embodiment, the third time sequence t3 is an operation sequence between the first and second time sequences t1 and t2, but not limited thereto. For example, the third time sequence t3 may also be The timing of the operation after the second time sequence t2.

本實施例之深度估計方法及其裝置係參考對應至目標時間序之時序圖框資料組與對應至另一時間序之圖框資料組間的移動向量資料及對應至此另一時間序之深度資料,來產生對應至此目標時間序的估計深度資料。本實施例之深度估計方法及其裝置更參考此估計深度資料,來簡化其產生對應至此目標時間序之深度資料的操作。據此相較於傳統深度估計方法,本發明相關之深度估計方法及其裝置具有運算複雜度較低之優點。The depth estimation method and apparatus of the embodiment refer to the motion vector data corresponding to the time series frame data group corresponding to the target time sequence and the frame data group corresponding to another time sequence, and the depth data corresponding to the other time sequence. To generate estimated depth data corresponding to the time sequence of this target. The depth estimation method and apparatus of the present embodiment further refer to the estimated depth data to simplify the operation of generating depth data corresponding to the target time sequence. Accordingly, the depth estimation method and apparatus related to the present invention have the advantage of low computational complexity compared to the conventional depth estimation method.

第二實施例Second embodiment

本實施例之深度估計方法係參考輸入之雙眼影像資料之物件分割資訊來修正其產生對應之深度資料的操作。The depth estimation method in this embodiment corrects the operation of generating the corresponding depth data by referring to the object segmentation information of the input binocular image data.

請參照第4圖,其繪示依照本發明第二實施例之深度估計裝置的方塊圖。本實施例之深度估計裝置2與第一實施例之深度估計裝置不同之處在於深度估計裝置2中更包括物件資訊估計單元210、物件資訊修正單元212及控制單元214。Referring to FIG. 4, a block diagram of a depth estimating apparatus according to a second embodiment of the present invention is shown. The depth estimating device 2 of the present embodiment is different from the depth estimating device of the first embodiment in that the depth estimating device 2 further includes an object information estimating unit 210, an object information correcting unit 212, and a control unit 214.

物件資訊估計單元210根據移動向量資料M_12找出對應至第二時序圖框資料組Vi_t2之第二時序估計物件分配資料O_t2。舉例來說,物件資訊估計單元210可根據移動向量資料M_12中各筆畫素移動向量資料與其周圍鄰近畫素移動向量資料的一致性,來對輸入雙眼影像資料Vi中之物件進行辨識及追蹤。The object information estimating unit 210 finds the second time-series estimated object allocation data O_t2 corresponding to the second timing frame data set Vi_t2 based on the motion vector data M_12. For example, the object information estimating unit 210 can identify and track the objects in the input binocular image data Vi according to the consistency of each pen pixel moving vector data in the moving vector data M_12 and the surrounding pixel moving vector data.

物件資訊修正單元212根據第一視角及第二視角圖框資料F1_t2及F2_t2修正第二時序估計物件分配資料O_t2,以得到第二時序物件分配資料E_t2。舉例來說,物件資訊修正單元212係參考對應至第二時間序t2之兩筆對應至不同視角之圖框資料的畫素資料,來驗證第二時序估計物件分配資料O_t2的精確性,並產生對應至第二時間序t2的第二時序物件分配資料E_t2。The object information correcting unit 212 corrects the second time-series estimated object allocation data O_t2 according to the first viewing angle and the second viewing angle frame data F1_t2 and F2_t2 to obtain the second time-series object allocation data E_t2. For example, the object information correction unit 212 verifies the accuracy of the second time-scheduled object allocation data O_t2 by referring to the pixel data corresponding to the frame data of the different viewing angles corresponding to the second time sequence t2. Corresponding to the second time series object allocation data E_t2 of the second time sequence t2.

控制單元214根據第二時序物件分配資料E_t2修正第二時序深度資料D_t2,以得到修正後之輸出第二時序深度資料Dx_t2。舉例來說,控制單元214例如參考第二時序物件分配資料E_t2中相關於物件之邊緣資訊(Boundary),來將對應至相同物件之深度資料修正為接近之深度值。The control unit 214 corrects the second time-series depth data D_t2 according to the second time-series object allocation data E_t2 to obtain the corrected output second-time depth data Dx_t2. For example, the control unit 214 corrects the depth data corresponding to the same object to a close depth value, for example, by referring to the edge information related to the object in the second time-series object allocation data E_t2.

由於參考第一及第三時序物件分配資料E_t1及E_t3修正第一及第三時序深度資料D_t1及D_t3進行修正之操作與根據第二時序物件分配資料E_t2修正第二時序深度資料D_t2之操作實質上相同,深度估計裝置2亦可經由實質上相同之操作,參考第一及第三時序物件分配資料E_t1及E_t3來針對第一或第三時序深度資料D_t1及D_t3進行修正操作。換言之,根據第一及第三時序物件分配資料E_t1及E_t3修正第一及第三時序深度資料D_t1及D_t3之操作可依據前述根據第二時序物件分配資料E_t2修正第二時序深度資料D_t2之操作類推得到,而於此並不再對其進行贅述。The operation of correcting the first and third timing depth data D_t1 and D_t3 by referring to the first and third time-series object allocation data E_t1 and E_t3 and the operation of correcting the second timing depth data D_t2 according to the second time-series object allocation data E_t2 are substantially Similarly, the depth estimating device 2 can perform the correcting operation for the first or third time-series depth data D_t1 and D_t3 with reference to the first and third time-series object allocation data E_t1 and E_t3 via substantially the same operation. In other words, the operations of modifying the first and third timing depth data D_t1 and D_t3 according to the first and third time-series object allocation data E_t1 and E_t3 may be based on the foregoing operation of correcting the second time-series depth data D_t2 according to the second time-series object allocation data E_t2. Obtained, and will not be repeated here.

請參照第5圖,其繪示依照本發明第二實施例之深度估計方法的流程圖。本實施例之深度估計方法與第一實施例之深度估計方法不同之處在於其於步驟(b)之後更包括步驟(i)-(k)。進一步的說,於步驟(b)之後首先執行步驟(i),物件資訊估計單元210根據移動向量資料M_12找出對應至第二時序圖框資料組Vi_t2之第二時序估計物件分配資料O_t2。Referring to FIG. 5, a flow chart of a depth estimation method according to a second embodiment of the present invention is shown. The depth estimation method of the present embodiment is different from the depth estimation method of the first embodiment in that it further includes steps (i)-(k) after the step (b). Further, after step (b), step (i) is first performed, and the object information estimating unit 210 finds the second time-series estimated object allocation data O_t2 corresponding to the second timing frame data set Vi_t2 according to the motion vector data M_12.

接著如步驟(j),物件資料修正單元212根據第一視角及第二視角圖框資料F1_t2及F2_t2修正第二時序估計物件分配資料O_t2,以得到第二時序物件分配資料E_t2。之後如步驟(k),控制單元214根據第二時序物件分配資料E_t2修正第二時序深度資料D_t2,以得到輸出第二時序深度資料Dx_t2。Then, in step (j), the object data correcting unit 212 corrects the second time-series estimated object allocation data O_t2 according to the first viewing angle and the second viewing angle frame data F1_t2 and F2_t2 to obtain the second time-series object allocation data E_t2. Then, as step (k), the control unit 214 corrects the second timing depth data D_t2 according to the second time-series object allocation data E_t2 to obtain the output second timing depth data Dx_t2.

本實施例之深度估計方法及其裝置係參考對應至目標時間序之時序圖框資料組與對應至另一時間序之圖框資料組間的移動向量資料及對應至此另一時間序之深度資料,來產生對應至此目標時間序的估計深度資料。本實施例之深度估計方法及其裝置更參考此估計深度資料,來簡化其產生對應至此目標時間序之深度資料的操作。據此相較於傳統深度估計方法,本發明相關之深度估計方法及其裝置具有運算複雜度較低之優點。The depth estimation method and apparatus of the embodiment refer to the motion vector data corresponding to the time series frame data group corresponding to the target time sequence and the frame data group corresponding to another time sequence, and the depth data corresponding to the other time sequence. To generate estimated depth data corresponding to the time sequence of this target. The depth estimation method and apparatus of the present embodiment further refer to the estimated depth data to simplify the operation of generating depth data corresponding to the target time sequence. Accordingly, the depth estimation method and apparatus related to the present invention have the advantage of low computational complexity compared to the conventional depth estimation method.

綜上所述,雖然本發明已以較佳實施例揭露如上,然其並非用以限定本發明。另,本發明所稱之深度資料、估計深度資料係指具有深度性質之資料,並不侷限於深度值,如視差值等其他具有深度性質之資料,亦在本發明之範疇中。本發明所屬技術領域中具有通常知識者,在不脫離本發明之精神和範圍內,當可作各種之更動與潤飾。因此,本發明之保護範圍當視後附之申請專利範圍所界定者為準。In conclusion, the present invention has been disclosed in the above preferred embodiments, and is not intended to limit the present invention. In addition, the depth data and the estimated depth data referred to in the present invention refer to data having a deep nature, and are not limited to depth values, such as disparity values and other materials having deep nature, and are also within the scope of the present invention. A person skilled in the art can make various changes and modifications without departing from the spirit and scope of the invention. Therefore, the scope of the invention is defined by the scope of the appended claims.

1、1’、2...深度估計裝置1, 1', 2. . . Depth estimating device

102、102’、202...輸入單元102, 102', 202. . . Input unit

104、104’、204...移動向量產生單元104, 104', 204. . . Moving vector generation unit

106、106’、206...雙視域影像比對單元106, 106', 206. . . Dual-view image comparison unit

108、108’、208...深度估計單元108, 108', 208. . . Depth estimation unit

210...物件資訊估計單元210. . . Object information estimation unit

212...物件資訊修正單元212. . . Object information correction unit

214...控制單元214. . . control unit

第1圖繪示依照本發明第一實施例之深度估計裝置的方塊圖。Fig. 1 is a block diagram showing a depth estimating apparatus according to a first embodiment of the present invention.

第2圖繪示依照本發明第一實施例之深度估計方法的流程圖。2 is a flow chart showing a depth estimation method according to a first embodiment of the present invention.

第3圖繪示依照本發明第一實施例之深度估計裝置的另一方塊圖。Fig. 3 is a block diagram showing another embodiment of the depth estimating apparatus according to the first embodiment of the present invention.

第4圖繪示依照本發明第二實施例之深度估計裝置的方塊圖。Fig. 4 is a block diagram showing a depth estimating apparatus according to a second embodiment of the present invention.

第5圖繪示依照本發明第二實施例之深度估計方法的流程圖。FIG. 5 is a flow chart showing a depth estimation method according to a second embodiment of the present invention.

(a)-(e)...操作步驟(a)-(e). . . Steps

Claims (12)

一種深度估計方法,用以針對一輸入雙眼影像(Binocular Video)資料進行深度估計,該深度估計方法包括:(a) 接收該輸入雙眼影像資料中對應至一第一時間序的一第一時序圖框資料組及對應至一第二時間序的一第二時序圖框資料組,其中該第一時序圖框資料組與第二時序圖框資料組各自包括一第一視角圖框資料及一第二視角圖框資料;(b) 找出該第二時序圖框資料組相對於該第一時序圖框資料組之一第一移動向量資料;(c) 根據該第一時序圖框資料組之該第一視角圖框資料及該第二視角圖框資料進行雙視域影像比對(Stereo Matching),以找出對應至該第一時間序的一第一時序深度資料;(d) 根據該第一移動向量資料及該第一時序深度資料找出一第二時序估計深度資料;以及(e) 根據該第二時序估計深度資料、該第二時序圖框資料組之該第一視角圖框資料及該第二視角圖框資料,找出該第二時序深度資料。A depth estimation method for performing depth estimation on an input binocular video (Binocular Video) data, the depth estimation method comprising: (a) receiving a first one of the input binocular image data corresponding to a first time sequence a timing frame data group and a second timing frame data group corresponding to a second time sequence, wherein the first timing frame data group and the second timing frame data group each include a first perspective frame Data and a second view frame data; (b) finding a first movement vector data of the second time series frame data group relative to the first time frame data group; (c) according to the first time The first view frame data of the sequence frame data group and the second view frame data are subjected to two-view image matching (Stereo Matching) to find a first time series depth corresponding to the first time sequence. Data; (d) finding a second time-series estimation depth data according to the first motion vector data and the first timing depth data; and (e) estimating the depth data according to the second timing, the second timing frame data The first view frame data of the group and The second view frame data is used to find the second time series depth data. 如申請專利範圍第1項所述之深度資料估計方法,其中步驟(a)更包括:接收該輸入雙眼影像資料中對應至一第三時間序的一第三時序圖框資料組,該第三時序圖框資料組包括一第一視角圖框資料及一第二視角圖框資料。 The method for estimating a depth data according to claim 1, wherein the step (a) further comprises: receiving a third timing frame data group corresponding to a third time sequence in the input binocular image data, the The three timing frame data group includes a first perspective frame data and a second perspective frame data. 如申請專利範圍第2項所述之深度資料估計方法,更包括:(f)找出該第三時序圖框資料組相對於該第二時序圖框資料組之一第二移動向量資料;(g)根據該第一移動向量資料、該第一時序深度資料、該第二移動向量資料及該第二時序深度資料找出一第三時序估計深度資料;及(h)根據該第三時序估計深度資料、該第三時序圖框資料組之該第一視角圖框資料及該第二視角圖框資料,找出該第三時序深度資料。 The method for estimating the depth data according to item 2 of the patent application scope further includes: (f) finding a second movement vector data of the third time series frame data group relative to the second time frame data group; g) finding a third timing estimation depth data according to the first motion vector data, the first timing depth data, the second motion vector data, and the second timing depth data; and (h) according to the third timing The depth data, the first view frame data of the third time series frame data group, and the second view frame data are estimated to find the third time series depth data. 如申請專利範圍第3項所述之深度資料估計方法,更包括:(i)根據該第二移動向量資料找出對應至該第三時序圖框資料組之一估計物件分配資料;(j)根據該第三時序圖框資料組之該第一視角圖框資料及該第二視角圖框資料修正該估計物件分配資料,以得到一第三時序物件分配資料;及(k)根據該第三時序物件分配資料修正該第三時序深度資料,以得到一輸出第三時序深度資料。 The method for estimating a depth data according to item 3 of the patent application scope further includes: (i) finding an estimated object allocation data corresponding to one of the third time series frame data groups according to the second motion vector data; (j) And correcting the estimated object allocation data according to the first view frame data of the third time series frame data group and the second view frame data to obtain a third time series object allocation data; and (k) according to the third The time series object allocation data corrects the third time series depth data to obtain an output third time series depth data. 如申請專利範圍第1項所述之深度資料估計方法,更包括: (i)根據該第一移動向量資料找出對應至該第二時序圖框資料組之一估計物件分配資料;(j)根據該第二時序圖框資料組之該第一視角圖框資料及該第二視角圖框資料修正該估計物件分配資料,以得到一第二時序物件分配資料;及(k)根據該第二時序物件分配資料修正該第二時序深度資料,以得到一輸出第二時序深度資料。 For example, the method for estimating depth data as described in item 1 of the patent application scope includes: (i) finding an estimated object allocation data corresponding to one of the second timing frame data sets according to the first motion vector data; (j) the first viewing frame data according to the second timing frame data group and The second view frame data is corrected to obtain the second time-series object allocation data; and (k) correcting the second time-series depth data according to the second time-series object allocation data to obtain an output second Timing depth data. 如申請專利範圍第1項所述之深度資料估計方法,更包括:(i)根據該第一移動向量資料找出對應至該第一時序圖框資料組之一估計物件分配資料;(j)根據該第一時序圖框資料組之該第一視角及該第二視角圖框資料修正該估計物件分配資料,以得到一第一時序物件分配資料;及(k)根據該第一時序物件分配資料修正該第一時序深度資料,以得到一輸出第一時序深度資料。 The method for estimating a depth data according to the first aspect of the patent application, further comprising: (i) finding, according to the first motion vector data, an estimated object allocation data corresponding to one of the first timing frame data sets; And correcting the estimated object allocation data according to the first viewing angle and the second viewing angle frame data of the first timing frame data group to obtain a first time series object allocation data; and (k) according to the first The time-series object allocation data corrects the first timing depth data to obtain an output first timing depth data. 一種深度資料估計裝置,用以針對一輸入雙眼影像(Binocular Video)資料進行深度估計,該深度估計裝置包括:一輸入單元,用以接收該輸入雙眼影像資料中對應至一第一時間序的一第一時序圖框資料組及對應至一第二時間序的一第二時序圖框資料組,其中各該第一及該第二時序圖框資料組包括一第一視角圖框資料及一第二視角 圖框資料;一移動向量產生單元,耦接於該輸入單元,用以找出該第二時序圖框資料組相對於該第一時序圖框資料組之一第一移動向量資料;一雙視域影像比對(Stereo Matching)單元,耦接於該輸入單元,用以根據該第一時序圖框資料組之該第一視角及該第二視角圖框資料進行雙視域影像比對,以找出對應至該第一時間序的一第一時序深度資料;以及一深度估計單元,耦接於該移動向量產生單元與該雙視域影像比對單元,用以根據該第一移動向量資料及該第一時序深度資料找出一第二時序估計深度資料;其中,該雙視域影像比對單元根據該第二時序估計深度資料、該第二時序圖框資料組之該第一視角及該第二視角圖框資料,找出該第二時序深度資料。 A depth data estimating device for performing depth estimation on an input binocular video (Binocular Video) data, the depth estimating device comprising: an input unit configured to receive the input binocular image data corresponding to a first time sequence a first timing frame data group and a second timing frame data group corresponding to a second time sequence, wherein each of the first and second timing frame data sets includes a first perspective frame data And a second perspective a frame vector data, coupled to the input unit, for finding a first motion vector data of the second timing frame data group relative to the first time frame data group; The viewfinder unit is coupled to the input unit for performing dual-view image comparison according to the first view and the second view frame data of the first time frame data group a first time depth data corresponding to the first time sequence; and a depth estimation unit coupled to the motion vector generation unit and the dual view image comparison unit for Locating a second time-series estimation depth data by using the motion vector data and the first timing depth data; wherein the dual-view image comparison unit estimates the depth data according to the second timing, and the second timing frame data group The first view angle and the second view frame data are used to find the second time series depth data. 如申請專利範圍第7項所述之深度資料估計裝置,其中該輸入單元更接收該輸入雙眼影像資料中對應至一第三時間序的一第三時序圖框資料組,該第三時序圖框資料組包括一第一視角圖框資料及一第二視角圖框資料。 The depth data estimating device of claim 7, wherein the input unit further receives a third timing frame data group corresponding to a third time sequence in the input binocular image data, the third timing chart The frame data group includes a first view frame data and a second view frame data. 如申請專利範圍第8項所述之深度資料估計裝置,其中:該移動向量產生單元更用以找出該第三時序圖框資料組相對於該第二時序圖框資料組之一第二移動向量資料; 該雙視域影像比對單元更根據該第一移動向量資料、該第一時序深度資料、該第二移動向量資料及該第二時序深度資料找出一第三時序估計深度資料;及該深度估計單元更根據該第三時序估計深度資料、該第三時序圖框資料組之該第一視角及該第二視角圖框資料,找出該第三時序深度資料。 The depth data estimating device of claim 8, wherein: the motion vector generating unit is further configured to find a second movement of the third timing frame data group with respect to one of the second timing frame data groups. Vector data The dual-view image comparison unit further finds a third time-series estimated depth data according to the first motion vector data, the first timing depth data, the second motion vector data, and the second timing depth data; The depth estimation unit further finds the third time series depth data according to the third time series estimation depth data, the first view angle of the third time series frame data group, and the second view frame data. 如申請專利範圍第9項所述之深度資料估計裝置,更包括:一物件資訊估計單元,耦接於該移動向量產生單元,用以根據該第二移動向量資料找出對應至該第三時序圖框資料組之一估計物件分配資料;一物件資訊修正單元,耦接於該物件資訊估計單元,根據該第三時序圖框資料組之該第一視角及該第二視角圖框資料修正該估計物件分配資料,以得到一第三時序物件分配資料;及一控制單元,耦接於該物件資訊修正單元,用以根據該第三時序物件分配資料修正該第三時序深度資料,以得到一輸出第三時序深度資料。 The depth data estimating device of claim 9, further comprising: an object information estimating unit coupled to the motion vector generating unit, configured to find the corresponding to the third timing according to the second motion vector data One of the frame data sets estimates the object allocation data; an object information correction unit is coupled to the object information estimating unit, and corrects the first viewing angle and the second viewing frame data according to the third time series frame data group Estimating the object allocation data to obtain a third time-series object allocation data; and a control unit coupled to the object information correction unit for correcting the third time-series depth data according to the third time-series object allocation data to obtain a The third timing depth data is output. 如申請專利範圍第7項所述之深度資料估計裝置,更包括:一物件資訊估計單元,耦接於該移動向量產生單元,用以根據該第一移動向量資料找出對應至該第二時序圖框資料組之一估計物件分配資料;及 一物件資訊修正單元,耦接於該物件資訊估計單元,根據該第二時序圖框資料組之該第一視角及該第二視角圖框資料修正該估計物件分配資料,以得到一第二時序物件分配資料;及一控制單元,耦接於該物件資訊修正單元,用以根據該第二時序物件分配資料修正該第二時序深度資料,以得到一輸出第二時序深度資料。 The depth data estimating device of claim 7, further comprising: an object information estimating unit coupled to the motion vector generating unit, configured to find the corresponding to the second timing according to the first motion vector data One of the frame data sets estimates the item allocation information; and An object information correction unit is coupled to the object information estimating unit, and corrects the estimated object allocation data according to the first viewing angle and the second viewing angle frame data of the second timing frame data group to obtain a second timing The object allocation data; and a control unit coupled to the object information correction unit for correcting the second time series depth data according to the second time series object allocation data to obtain an output second time series depth data. 如申請專利範圍第7項所述之深度資料估計裝置,更包括:一物件資訊估計單元,耦接於該移動向量產生單元,用以根據該第一移動向量資料找出對應至該第一時序圖框資料組之一估計物件分配資料;及一物件資訊修正單元,耦接於該物件資訊估計單元,根據該第一時序圖框資料組之該第一視角及該第二視角圖框資料修正該估計物件分配資料,以得到一第一時序物件分配資料;及一控制單元,耦接於該物件資訊修正單元,用以根據該第一時序物件分配資料修正該第一時序深度資料,以得到一輸出第一時序深度資料。 The depth data estimating device of claim 7, further comprising: an object information estimating unit coupled to the motion vector generating unit, configured to find the corresponding to the first time according to the first motion vector data One of the sequence frame data sets estimates the object allocation data; and an object information correction unit is coupled to the object information estimating unit, according to the first view and the second view frame of the first time series frame data set Correcting the estimated object allocation data to obtain a first time-series object allocation data; and a control unit coupled to the object information correction unit for correcting the first timing according to the first time-series object allocation data Depth data to get an output first timing depth data.
TW100108826A 2011-03-15 2011-03-15 Method for depth estimation and device usnig the same TWI485651B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
TW100108826A TWI485651B (en) 2011-03-15 2011-03-15 Method for depth estimation and device usnig the same

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
TW100108826A TWI485651B (en) 2011-03-15 2011-03-15 Method for depth estimation and device usnig the same

Publications (2)

Publication Number Publication Date
TW201237805A TW201237805A (en) 2012-09-16
TWI485651B true TWI485651B (en) 2015-05-21

Family

ID=47223239

Family Applications (1)

Application Number Title Priority Date Filing Date
TW100108826A TWI485651B (en) 2011-03-15 2011-03-15 Method for depth estimation and device usnig the same

Country Status (1)

Country Link
TW (1) TWI485651B (en)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7003136B1 (en) * 2002-04-26 2006-02-21 Hewlett-Packard Development Company, L.P. Plan-view projections of depth image data for object tracking
CN101051386A (en) * 2007-05-23 2007-10-10 北京航空航天大学 Precision matching method for multiple depth image
TW201023649A (en) * 2008-12-02 2010-06-16 Himax Tech Ltd Method and apparatus of tile-based belief propagation
CN101540926B (en) * 2009-04-15 2010-10-27 南京大学 Stereo video coding-decoding method based on H.264
CN101883283A (en) * 2010-06-18 2010-11-10 北京航空航天大学 Control method for code rate of three-dimensional video based on SAQD domain

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7003136B1 (en) * 2002-04-26 2006-02-21 Hewlett-Packard Development Company, L.P. Plan-view projections of depth image data for object tracking
CN101051386A (en) * 2007-05-23 2007-10-10 北京航空航天大学 Precision matching method for multiple depth image
TW201023649A (en) * 2008-12-02 2010-06-16 Himax Tech Ltd Method and apparatus of tile-based belief propagation
CN101540926B (en) * 2009-04-15 2010-10-27 南京大学 Stereo video coding-decoding method based on H.264
CN101883283A (en) * 2010-06-18 2010-11-10 北京航空航天大学 Control method for code rate of three-dimensional video based on SAQD domain

Also Published As

Publication number Publication date
TW201237805A (en) 2012-09-16

Similar Documents

Publication Publication Date Title
JP5153940B2 (en) System and method for image depth extraction using motion compensation
Pham et al. Domain transformation-based efficient cost aggregation for local stereo matching
US9208541B2 (en) Apparatus and method for correcting disparity map
CN109963048B (en) Noise reduction method, noise reduction device and noise reduction circuit system
KR102380862B1 (en) Method and apparatus for image processing
JP2020515931A (en) Method and apparatus for combining scene segmentation and 3D reconstruction
US20190095694A1 (en) Apparatus and method for performing 3d estimation based on locally determined 3d information hypotheses
WO2009097714A1 (en) Depth searching method and depth estimating method for multi-viewing angle video image
EP3110149B1 (en) A system and a method for depth-image-based rendering
Oliveira et al. Selective hole-filling for depth-image based rendering
Gong Enforcing temporal consistency in real-time stereo estimation
JP2014515197A (en) Multi-view rendering apparatus and method using background pixel expansion and background-first patch matching
TW201436552A (en) Method and apparatus for increasing frame rate of an image stream using at least one higher frame rate image stream
JP5178538B2 (en) Method for determining depth map from image, apparatus for determining depth map
KR100943635B1 (en) Method and apparatus for generating disparity map using digital camera image
US9208549B2 (en) Method and apparatus for color transfer between images
JP2009530701A5 (en)
CN105791795A (en) Three-dimensional image processing method and device and three-dimensional video display device
Jantet et al. Joint projection filling method for occlusion handling in depth-image-based rendering
Kneip et al. SDICP: Semi-Dense Tracking based on Iterative Closest Points.
KR20150097251A (en) Camera alignment method using correspondences between multi-images
JP2004356747A (en) Method and apparatus for matching image
EP2525324A2 (en) Method and apparatus for generating a depth map and 3d video
KR20160085708A (en) Method and apparatus for generating superpixels for multi-view images
JP2009186287A (en) Plane parameter estimating device, plane parameter estimating method, and plane parameter estimating program

Legal Events

Date Code Title Description
MM4A Annulment or lapse of patent due to non-payment of fees