TW201426466A - Optical touch system - Google Patents

Optical touch system Download PDF

Info

Publication number
TW201426466A
TW201426466A TW101150858A TW101150858A TW201426466A TW 201426466 A TW201426466 A TW 201426466A TW 101150858 A TW101150858 A TW 101150858A TW 101150858 A TW101150858 A TW 101150858A TW 201426466 A TW201426466 A TW 201426466A
Authority
TW
Taiwan
Prior art keywords
dimensional plane
plane information
operator
dimensional
information
Prior art date
Application number
TW101150858A
Other languages
Chinese (zh)
Inventor
Wei-Hong Wang
sheng-zhi Wang
Zeng-Yu Zeng
bo-xiang Chen
Original Assignee
Disk King Technology Co Ltd
Univ Nat Chiao Tung
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Disk King Technology Co Ltd, Univ Nat Chiao Tung filed Critical Disk King Technology Co Ltd
Priority to TW101150858A priority Critical patent/TW201426466A/en
Publication of TW201426466A publication Critical patent/TW201426466A/en

Links

Landscapes

  • Length Measuring Devices By Optical Means (AREA)

Abstract

An optical touch system is disclosed to identify the location of an operating object in a three dimensional space. The system includes a first camera, a second camera and a processing unit. The first camera and the second camera obtains a first two dimensional data and a second two dimensional data from the three dimensional space respectively, and the processing unit identifies the location of the operating object by using the data. In detail, the processing unit executes a processing mechanism described as following to calculate the location. First, a three dimensional coordination is established by using the two dimensional data. Second, each of the gradient variation in the two dimensional data is applied to identify the location of the operating object. Third, each of the shape information in the two dimensional data is applied to identify the location of the operating object. Forth, the parallax pitch of the operating objects in the two dimensional data is applied to identify the actual location of the operating object in the three dimensional space.

Description

光學式觸控系統 Optical touch system

本發明是有關於一種觸控技術,特別是關於一種光學觸控技術。 The present invention relates to a touch technology, and more particularly to an optical touch technology.

請參考第1圖,第1圖是習知之桌上型光學觸控裝置的操作原理示意圖。第1圖中,桌上型光學觸控裝置100主要包括一投影機110、一紅外線接收鏡頭120與一紅外線發射器130。其原理如下:首先,投影機110在桌面上定義出一個操作平面101,操作平面101上可以顯示出投影機110所要展現的多媒體內容。然後,紅外線發射器130在操作平面101上鋪設出一層紅外線光幕;接下來,當操作物102,例如手指或筆,在操作平面101上動作時,紅外線會擊中操作物102然後反射到紅外線接收鏡頭120。因此,紅外線接收鏡頭120可以取得一個平面上有一個特定位置是紅外線亮點的資訊,此特定位置即為操作物102的所在位置。然而,這種以紅外光偵測操作物102位置的觸控技術,容易受到外在環境光中的紅外線干擾。 Please refer to FIG. 1 , which is a schematic diagram of the operation principle of a conventional desktop optical touch device. In the first embodiment, the desktop optical touch device 100 mainly includes a projector 110, an infrared receiving lens 120 and an infrared emitter 130. The principle is as follows: First, the projector 110 defines an operation plane 101 on the desktop, and the operation plane 101 can display the multimedia content to be displayed by the projector 110. Then, the infrared emitter 130 lays a layer of infrared light curtain on the operation plane 101; next, when the operator 102, such as a finger or a pen, moves on the operation plane 101, the infrared rays hit the operator 102 and then reflect to the infrared rays. The lens 120 is received. Therefore, the infrared receiving lens 120 can obtain information that a specific position is an infrared bright spot on a plane, and the specific position is the position of the operating object 102. However, such a touch technology that detects the position of the object 102 by infrared light is susceptible to infrared interference in the external ambient light.

故,本發明之發明人有鑑於上述缺失,乃蒐集相關資料,經由多方評估及考量,並以從事於此行業累積之多年經驗,經由不斷試作及修改,始設計出此一光學式觸控系統。 Therefore, the inventors of the present invention have collected the relevant materials in view of the above-mentioned deficiencies, and have designed and evaluated the optical touch system through continuous trial and modification through multi-party evaluation and consideration, and with years of experience accumulated in the industry. .

依據本發明一實施方式,提出一種光學式觸控系統,係用以辨識一操作物於一三維操作空間中的位置,其包括一第一攝影鏡頭、一第二攝影鏡頭與一運算單元。第一攝影鏡頭係用以自三維操作空間中取得一第一二維平面資訊;第二攝影鏡頭係用以自三維操作空間中取得一第二二維平面資訊。運算單元利用第一二維平面資訊與第二二維平面資訊,辨識操作物的位置;其中,運算單元係執行一處理機制來辨識操作物的位置,處理機制包括下列步驟:步驟一,利用第一二維平面資訊與第二二維平面資訊建立一立體座標系;步驟二,以第一二維平面資訊與第二二維平面資訊中的梯度變化辨識操作物;步驟三,以第一二維平面資訊與第二二維平面資訊中的線形資訊辨識操作物;步驟四,以操作物於第一二維平面資訊與第二二維平面資訊中的視差間距,來定義操作物的位置。 According to an embodiment of the present invention, an optical touch system is provided for identifying a position of an operator in a three-dimensional operation space, and includes a first photographic lens, a second photographic lens, and an arithmetic unit. The first photographic lens is used to obtain a first two-dimensional plane information from the three-dimensional operation space; the second photographic lens is used to obtain a second two-dimensional plane information from the three-dimensional operation space. The operation unit uses the first two-dimensional plane information and the second two-dimensional plane information to identify the position of the operation object; wherein the operation unit performs a processing mechanism to identify the position of the operation object, and the processing mechanism includes the following steps: Step 1 A two-dimensional plane information and a second two-dimensional plane information establish a three-dimensional coordinate system; in step two, the operator is identified by the gradient change in the first two-dimensional plane information and the second two-dimensional plane information; step three, the first two The dimension information and the linear information in the second two-dimensional plane information identify the operator; in step four, the position of the operator is defined by the parallax distance between the first two-dimensional plane information and the second two-dimensional plane information.

值得注意的是,在本發明其他實施方式中,運算單元於執行處理機制時,在以第一二維平面資訊與第二二維平面資訊中的梯度變化辨識操作物中,係事先建立操作物之一方向梯度圖(HOG)模型,且將方向梯度圖(HOG)模型當作過濾器,分別在第一二維平面資訊與第二二維平面資訊中辨識操作物。 It should be noted that, in other embodiments of the present invention, when the processing unit performs the processing mechanism, the operation object is established in advance in the operation of identifying the gradient in the first two-dimensional plane information and the second two-dimensional plane information. One direction gradient map (HOG) model, and a direction gradient map (HOG) model is used as a filter to identify an operator in the first two-dimensional plane information and the second two-dimensional plane information, respectively.

舉例來說,上述方向梯度圖(HOG)模型可由一本過濾器(root filter)與多個部分過濾器(part filters)來組成。在可選擇的設計上,多個部分過濾器(part filters)可各自有相對應之變形成本,以定錨 (anchor)在本過濾器(root filter)之周邊。 For example, the above-described directional gradient map (HOG) model can be composed of a root filter and a plurality of partial filters. In an alternative design, multiple partial filters can each have a corresponding deformation cost to anchor (anchor) is around the root filter.

另一方面,運算單元在執行處理機制時,以第一二維平面資訊與第二二維平面資訊中的線形資訊,辨識操作物時,可事先建立操作物之一線形模型來描述操作物的輪廓,進而分別在第一二維平面資訊與第二二維平面資訊中辨識該操作物。例如,運算單元在處理第一二維平面資訊與第二二維平面資訊中的線形資訊時,可將多個破碎的邊緣鍊以至少一段橋,修補連接成至少一輪廓段,進而與基於操作物的輪廓所預設的線形模型進行相似權重的投射比對。 On the other hand, when the processing unit performs the processing mechanism, when the operator objects are identified by the first two-dimensional plane information and the linear information in the second two-dimensional plane information, a linear model of the operator may be established in advance to describe the operator. And contouring, and then identifying the operator in the first two-dimensional plane information and the second two-dimensional plane information, respectively. For example, when the processing unit processes the linear information in the first two-dimensional plane information and the second two-dimensional plane information, the plurality of broken edge chains may be repaired and connected into at least one contour segment by at least one section of the bridge, and then based on the operation. The linear model preset by the contour of the object performs a projection comparison of similar weights.

從另一個角度觀之,在本發明其他實施方式中,當前述立體座標系、梯度變化辨識及線形資訊辨識功能完備後,當一物件介入第一二維平面資訊與第二二維平面資訊,且偵測此物件之視差間距介於一上限值與一下限值間,即可執行梯度變化辨識及線形資訊辨識功能,以判定物件是否為操作物。 From another perspective, in another embodiment of the present invention, when the stereo coordinate system, the gradient change identification, and the linear information identification function are completed, when an object is involved in the first two-dimensional plane information and the second two-dimensional plane information, And detecting that the parallax distance of the object is between an upper limit value and a lower limit value, the gradient change identification and the linear information recognition function can be performed to determine whether the object is an operation object.

申言之,若物件之視差間距低於下限值,可直接判定物件非操作物;若物件之視差間距高於下限值,可直接判定物件未觸發一觸控事件,而不論其為操作物與否。 It is claimed that if the parallax distance of the object is lower than the lower limit value, the object non-operating object can be directly determined; if the parallax distance of the object is higher than the lower limit value, it can be directly determined that the object does not trigger a touch event, regardless of the operation. Things or not.

為達成上述目的及功效,本發明所採用之技術手段及構造,茲繪圖就本發明各實施方式詳加說明其特徵與功能如下,俾利完全了解。 In order to achieve the above object and effect, the technical means and structure adopted by the present invention are described in detail in the embodiments of the present invention as follows.

請參閱第2圖,第2圖是本實施方式之光學式觸控系統200的結構示意圖。第2圖中,光學式觸控系統200包括第一攝影鏡頭210、第二攝影鏡頭220與一處理單元230。第一攝影鏡頭210從操作平面101與操作物102所在的三維立體空間中,取得一第一二維平面資訊211,第二攝影鏡頭220則從相同的三維立體空間中取得一第二二維平面資訊221。然後,處理單元230再根據第一二維平面資訊211與第二二維平面資訊221,偵測出操作物102的具體位置。更進一步的說明之,處理單元230必須先建立一個三維系統來對應操作平面101與操作物102所在的立體操作空間,然後再分析第一二維平面資訊211與第二二維平面資訊221的資訊來確認操作物102的位置,進而達到觸控的效果。 Please refer to FIG. 2 , which is a schematic structural diagram of the optical touch control system 200 of the present embodiment. In FIG. 2 , the optical touch system 200 includes a first photographic lens 210 , a second photographic lens 220 , and a processing unit 230 . The first photographic lens 210 obtains a first two-dimensional plane information 211 from the three-dimensional space in which the operation plane 101 and the operator 102 are located, and the second photographic lens 220 obtains a second two-dimensional plane from the same three-dimensional space. Information 221. Then, the processing unit 230 detects the specific location of the operator 102 according to the first two-dimensional plane information 211 and the second two-dimensional plane information 221 . To further illustrate, the processing unit 230 must first establish a three-dimensional system to correspond to the stereoscopic operation space in which the operation plane 101 and the operator 102 are located, and then analyze the information of the first two-dimensional plane information 211 and the second two-dimensional plane information 221 The position of the operator 102 is confirmed to achieve the touch effect.

具體而言,處理單元230必須執行一處理機制240,處理機制240至少包括下列步驟:首先執行步驟一,建立立體座標系程序241:利用第一二維平面資訊211與第二二維平面資訊221間的差異,來建立一個三維的立體座標系。舉例來說,對三維直角坐標系xyz而言,第一二維平面資訊211與第二二維平面資訊221是從不同的z軸高度位置,取得的相同x-y兩軸平面資訊。因此,利用第一二維平面資訊211與第二二維平面資訊221的像差,即可建立上述的立體座標系。更詳細的立體座標系建立技術,將於後面舉例說明之。 Specifically, the processing unit 230 must execute a processing mechanism 240. The processing mechanism 240 includes at least the following steps: First, step 1 is performed to establish a stereo coordinate system program 241: using the first two-dimensional plane information 211 and the second two-dimensional plane information 221 The difference between them is to create a three-dimensional coordinate system. For example, for the three-dimensional Cartesian coordinate system xyz, the first two-dimensional plane information 211 and the second two-dimensional plane information 221 are the same x-y two-axis plane information obtained from different z-axis height positions. Therefore, the above-mentioned stereo coordinate system can be established by using the aberration of the first two-dimensional plane information 211 and the second two-dimensional plane information 221 . A more detailed three-dimensional coordinate system creation technique will be exemplified later.

然後執行步驟二,以梯度辨識操作物程序242:在三維的立體座標系已被建立妥善後,分別從有操作物102介入的第一二維平 面資訊211與第二二維平面資訊221中,辨識出操作物102在三維的立體座標系中的位置。此處,操作平面101上的背景畫面可以是主動式的畫面,例如液晶顯示屏,也可以是被動式的投影畫面,例如由投影機110所界定具有多媒體內容的操作平面101。以第一攝影鏡頭210取得的第一二維平面資訊211為例,當操作物102出現在操作平面101上時,處理單元230在本程序中,需於第一二維平面資訊211中,利用物件邊框的顏色梯度變化,界定出個別物件的位置與形狀;藉此,符合預設操作物102形狀之物件,即可被鎖定進而取得位置資訊。當然,第二二維平面資訊221亦比照辦理;更詳細的以梯度辨識操作物技術,將於後面舉例說明之。 Then step 2 is performed to identify the operator program 242 with a gradient: after the three-dimensional coordinate system has been properly established, respectively, from the first two-dimensional plane with the operator 102 intervening In the face information 211 and the second two-dimensional plane information 221, the position of the operator 102 in the three-dimensional coordinate system is recognized. Here, the background picture on the operation plane 101 may be an active picture, such as a liquid crystal display, or a passive projection picture, such as an operation plane 101 having multimedia content defined by the projector 110. Taking the first two-dimensional plane information 211 obtained by the first photographic lens 210 as an example, when the operator 102 appears on the operation plane 101, the processing unit 230 needs to be utilized in the first two-dimensional plane information 211 in the present program. The color gradient of the object frame changes to define the position and shape of the individual objects; thereby, the object conforming to the shape of the preset operator 102 can be locked to obtain position information. Of course, the second two-dimensional plane information 221 is also processed in comparison; a more detailed gradient identification operator technique will be exemplified later.

接下來執行步驟三,以線形辨識操作物程序243:在上述以梯度辨識操作物程序242中,純以梯度變化來界定出符合預設操作物102形狀之物件時,有相當程度的誤判可能性,亦即將操作平面101上的背景畫面中形似操作物102的其他物件,誤認為操作物102。因此,處理單元230在本程序中,進一步利用在前述程序242中各個疑似操作物102形狀之物件的線形,作進一步的判定,利用預設操作物102線形資訊,排除基於梯度資訊而類似操作物102的物件。當然,第二二維平面資訊221亦比照辦理;更詳細的以線形辨識操作物技術,將於後面舉例說明之。 Next, step 3 is executed to identify the operator program 243 in line: in the gradient identification operator program 242, when the object conforming to the shape of the preset operator 102 is defined by the gradient change, there is a considerable degree of misjudgment possibility. That is, the other objects in the background image on the operation plane 101 that are similar to the operator 102 are mistaken for the operator 102. Therefore, in the present program, the processing unit 230 further utilizes the linear shape of the object shape of each suspected operation object 102 in the foregoing program 242, and further determines the use of the preset operation object 102 linear information to exclude similar operations based on the gradient information. 102 objects. Of course, the second two-dimensional plane information 221 is also processed in comparison; a more detailed line identification operation technique will be exemplified later.

最後執行步驟四,以視差間距確認操作物位置程序244:請一併參考第3圖,第3圖是本實施方式中,以視差間距確認操作物 位置的原理示意圖。從第一二維平面資訊211與第二二維平面資訊221中分別取出操作物102位置後,處理單元230在本程序中,即可利用兩者間的視差間距界定出操作物102的真實位置。具體而言,如第3圖所繪示,以手指作為操作物102時,左邊為指尖碰觸投影面,右邊為指尖尚未碰觸投影面。無論指尖是否碰觸投影面,對上方的第一攝影鏡頭210的俯視視野來說,指尖中心幾乎會落在投影面相同的位置,但是對下方的第二攝影鏡頭220而言,指尖碰觸投影面與尚未碰觸投影面會造成視野觀測的指尖中心分別落在投影面上不同的位置;兩者間產生的距離即為視差間距,若操作物102觸及平面,視差間距201是小於一個臨界值的;反之,若操作物102遠離平面,亦即非觸控狀態,視差間距202必大於上述臨界值。因此,本程序即可由此判斷操作物102是否處於觸控狀態,以決定是否利用操作物102的位置執行各種觸控功能。 Finally, step 4 is executed to confirm the operator position program 244 with the parallax spacing: Please refer to FIG. 3 together. FIG. 3 is the embodiment of the present invention, and the operator is confirmed by the parallax distance. Schematic diagram of the location. After the position of the operator 102 is respectively taken out from the first two-dimensional plane information 211 and the second two-dimensional plane information 221, the processing unit 230 can define the true position of the operator 102 by using the parallax distance between the two in the program. . Specifically, as shown in FIG. 3 , when the finger is used as the operator 102 , the fingertip touches the projection surface on the left side and the fingertip does not touch the projection surface on the right side. Regardless of whether the fingertip touches the projection surface, the center of the fingertip is almost at the same position on the projection surface for the top view of the upper first photographic lens 210, but for the second photographic lens 220 below, the fingertip Touching the projection surface and the projection surface that has not touched the projection surface will cause the center of the fingertip to be in different positions on the projection surface; the distance generated between the two is the parallax distance. If the operator 102 touches the plane, the parallax spacing 201 is If the operator 102 is away from the plane, that is, the non-touch state, the parallax spacing 202 must be greater than the above threshold. Therefore, the program can thereby determine whether the operator 102 is in a touch state to determine whether to perform various touch functions using the position of the operator 102.

值得注意的是,從第3圖所繪示之左右兩邊的狀態,可以發現指尖碰觸投影面時,上方的第一攝影鏡頭210所觀測的指尖中心和下方的第二攝影鏡頭220所觀測的指尖中心落於投影面的位置還是有些許的差異,事實上這是為何定義觸控點為指尖中心而不採用指尖最前端的原因。指尖中心是有一定的厚度,所以當指尖碰觸投影面時,上方的第一攝影鏡頭210與下方的第二攝影鏡頭220所決定的指尖中心經過映射運算(Homography)之後還是有預期的落差存在,這是因為指尖中心並不是完全的落在平面上。 由於本實施方式只採用可見光攝影機,所以投影畫面的內容對指尖偵測有直接的影響,當投影的影像出現類似手指的東西或是影像中包含真正的手指,會造成系統誤判,所以利用有厚度的指尖中心當作觸控點,可以有效避免畫面中出現的物體造成的誤判,亦即畫面中的手指影像反而會因為對應得太完美而不被當成操作物102。 It should be noted that, from the state of the left and right sides shown in FIG. 3, it can be found that the center of the fingertip observed by the upper first photographing lens 210 and the second photographing lens 220 below the fingertip when the fingertip touches the projection surface There is still a slight difference in the position of the center of the fingertip on the projection surface. In fact, this is why the touch point is defined as the center of the fingertip without the tip of the fingertip. The center of the fingertip has a certain thickness, so when the fingertip touches the projection surface, the center of the fingertip determined by the upper first photographing lens 210 and the second photographing lens 220 below is still expected after the mapping operation (Homography). The drop exists because the center of the fingertip does not completely fall on the plane. Since the present embodiment only uses a visible light camera, the content of the projected image has a direct influence on the fingertip detection. When the projected image appears like a finger or the image contains a real finger, the system may be misjudged, so The center of the fingertip is used as a touch point, which can effectively avoid misjudgment caused by objects appearing in the picture, that is, the finger image in the picture may not be regarded as the operator 102 because the correspondence is too perfect.

易言之,處理機制240還可包括步驟五如下:利用第一攝影鏡頭210與第二攝影鏡頭220所取得之影像間的落差,可以建議一個立體系統;雖然,在後續的使用上,兩鏡頭對同一物件取值仍有落差,但可以糾偏校正之。經校正後,若一物件之視差間距201為零,反而表示其非操作物102,而是背景畫面上的影像物件。考量糾偏校正無法百分之百,本實施方式可定義操作物102之視差間距201必須大於一下限值。 In other words, the processing mechanism 240 may further include the following steps: using a difference between the images obtained by the first photographic lens 210 and the second photographic lens 220, a stereoscopic system may be suggested; although, in subsequent use, the two lenses are used. There is still a drop in the value of the same object, but it can be corrected. After correction, if the parallax spacing 201 of an object is zero, it means the non-operating object 102, but the image object on the background image. Considering that the correction correction cannot be 100%, the present embodiment can define that the parallax distance 201 of the operator 102 must be greater than the lower limit.

另一方面,隨著操作平面101與操作物102所在的立體操作空間縱深變化,會因為第一攝影鏡頭210與第二攝影鏡頭220對指尖中心斜投影角度的不同,而無法使用固定的臨界值來判斷觸碰事件。易言之,雖然指尖離投影面的高度是相同的,卻因為第一攝影鏡頭210與第二攝影鏡頭220的位置不同而出現不同的視差,因此本實施方式另提出了動態臨界值方案來解決這樣的問題;亦即,判定指尖是否碰觸到投影面是由下列方程式(1)所決定: 其中,TH low (x,y)和TH high (x,y)為動態臨界值的上下界,會隨著指尖 中心出現在投影面不同的位置而改變。理想上,TH low (x,y)在任何位置應該都是0,因為映射運算(Homography)可以把平面上出現的點,幾乎沒有視差的在兩個視野之間對應;但實際上,卻必須考量尋找指尖中心時可能有些許誤差。若定義指尖碰觸投影面時指尖中心在兩個視野之間的視差為D touch (x,y),本實施方式採用TH low (x,y)=0.5D touch (x,y)及TH high (x,y)=1.5D touch (x,y)。事實上,TH high (x,y)的值影響了觸控的精準度,也就是指尖離投影面多近會被判定為觸碰事件;若指尖中心的位置愈精準則可以將TH high (x,y)改為愈接近D touch (x,y)的值,以提供更高精準度的觸碰判定。因此,當上方視野所找到的指尖中心經過映射運算(Homography)映射到下方視野時,和下方視野找到的指尖中心相比,若差距D ts (x,y)介於臨界值之間,則可以判定指尖碰觸投影面。 On the other hand, as the operation plane 101 and the stereoscopic operation space where the operator 102 is located vary in depth, the fixed angle may not be used because the first photographic lens 210 and the second photographic lens 220 have different oblique projection angles to the center of the fingertip. The value is used to determine the touch event. In other words, although the height of the fingertip from the projection surface is the same, different parallax occurs because the positions of the first photographic lens 210 and the second photographic lens 220 are different. Therefore, the present embodiment further proposes a dynamic threshold value scheme. Solve the problem; that is, determine whether the fingertip touches the projection surface is determined by the following equation (1): Among them, TH low ( x , y ) and TH high ( x , y ) are the upper and lower bounds of the dynamic critical value, which will change as the center of the fingertip appears at different positions on the projection surface. Ideally, TH low ( x , y ) should be 0 at any position, because the mapping operation (Homography) can correspond to the points appearing on the plane, and there is almost no parallax between the two fields of view; but in reality, it must There may be some errors in the search for the center of the fingertip. If the parallax of the center of the fingertip between the two fields of view is D touch ( x , y ) when the fingertip touches the projection surface, the embodiment adopts TH low ( x , y )=0.5 D touch ( x , y ) and TH high ( x , y )=1.5 D touch ( x , y ). In fact, the value of TH high ( x , y ) affects the accuracy of the touch, that is, how close the fingertip is to the projection surface, it will be judged as a touch event; if the position of the center of the fingertip is more precise, TH high can be ( x , y ) is changed to the value of D touch ( x , y ) to provide a higher-accuracy touch determination. Therefore, when the center of the fingertip found by the upper field of view is mapped to the lower field of view by the mapping operation (Homography), if the difference D ts ( x , y ) is between the critical values, compared with the center of the fingertip found by the lower field of view, Then you can determine that the fingertip touches the projection surface.

在實作上,建立立體座標系程序241已經建立了第一攝影鏡頭210與第二攝影鏡頭220之間在一個共同平面的對應關係;所以只需要在第一攝影鏡頭210的視野當中尋找操作物102可能的位置並定義可能的操作點,例如指尖,透過映射運算(Homography)即可輕易的決定當指尖碰觸投影面時,第二攝影鏡頭220中的指尖會落在這些可能的操作點的相對位置上,進而只需要在第二攝影鏡頭220視野中有興趣的位置尋找是否出現指尖中心並判斷該指尖是否碰觸投影面即可。因為實務上只在意觸碰的事件,所以當指尖尚未碰觸投影面時,由第一攝影鏡頭210決定的指尖中心對應到第二攝影鏡頭220的相對位置是不會出現指尖的,故可直 接判定為指尖尚未碰觸投影面,進而加快整體運算速度。 In practice, the establishment of the stereo coordinate system program 241 has established a correspondence between the first photographic lens 210 and the second photographic lens 220 in a common plane; therefore, it is only necessary to find an operation object in the field of view of the first photographic lens 210. 102 possible positions and define possible operating points, such as fingertips, through the mapping operation (Homography) can easily determine when the fingertip touches the projection surface, the fingertips in the second photographic lens 220 will fall on these possible In the relative position of the operating point, it is only necessary to find the position of the fingertip in the position of interest in the field of view of the second photographic lens 220 and determine whether the fingertip touches the projection surface. Because the practice only cares about the touch event, when the fingertip has not touched the projection surface, the relative position of the fingertip center determined by the first photographic lens 210 corresponding to the second photographic lens 220 does not appear at the fingertip. So straight It is judged that the fingertip has not touched the projection surface, thereby speeding up the overall calculation speed.

換句話說,在整體運作上,若第一攝影鏡頭210與第二攝影鏡頭220對一物件取得之視差間距201大於一上限值,表示其無論為操作物102與否,皆尚未接觸操作平面101,故亦不需於循環時序中,再執行較耗費運算資源之以梯度辨識操作物程序242與以線形辨識操作物程序243。整體觀之,在系統架設之初,需要先透過建立立體座標系程序241、以梯度辨識操作物程序242與以線形辨識操作物程序243從第一攝影鏡頭210與第二攝影鏡頭220所取得之資訊中,辨識用以建立立體系統及校正各種參數之操作物102,藉此濾除繁冗的背景資訊;待經以視差間距確認操作物位置程序244後,確切定義操作物102位置,立體系統即呈完備。後續使用時,則先執行以視差間距確認操作物位置程序244,對第一二維平面資訊211與第二二維平面資訊221中顯然存在視差間距的物件(背景畫面的物件不會顯然存在視差間距,除非系統數據偏移),粗估其視差間距是否介於下限值與上限值之間;若其低於下限值,表示定義為背景物件或其他干擾物(類如蚊蠅飛蟲),而非操作物102;若其高於上限值,表示無論其為操作物102與否,皆尚未發生觸控事件(Event),無後續的觸摸操控功能可言。 In other words, in the overall operation, if the parallax distance 201 obtained by the first photographic lens 210 and the second photographic lens 220 for an object is greater than an upper limit value, it indicates that the operating object 102 has not touched the operation plane. Therefore, the gradient recognition operator program 242 and the line recognition operator program 243 are not required to be executed in the loop timing. As a whole, at the beginning of the system setup, it is necessary to obtain the stereo coordinate system program 241, the gradient recognition operator program 242, and the linear recognition operator program 243 from the first photographic lens 210 and the second photographic lens 220. In the information, the operator 102 for establishing the stereo system and correcting various parameters is identified, thereby filtering out the tedious background information; after confirming the operator position program 244 by the parallax spacing, the position of the operator 102 is exactly defined, and the stereo system is Complete. For subsequent use, the object position program 244 is confirmed by the parallax spacing, and the object of the first two-dimensional plane information 211 and the second two-dimensional plane information 221 obviously has a parallax distance (the object of the background image does not obviously have parallax). Spacing, unless the system data is offset, whether the parallax distance is between the lower limit and the upper limit; if it is lower than the lower limit, it is defined as a background object or other interferent (such as a mosquito fly) ), instead of the operator 102; if it is higher than the upper limit, it means that no touch event has occurred regardless of whether it is the operator 102 or not, and no subsequent touch manipulation function is available.

更具體的說明之,以一隻手臂手掌握拳伸出食指,介入已經建立好的立體操作系統為例。從第一攝影鏡頭210與第二攝影鏡頭220取得之第一二維平面資訊211與第二二維平面資訊221,可以粗略鑑別出幾個具有視差間距的物件,兩個影像間手掌或拳頭 部的視差間距會大於預設上限值,故不予處理;而兩個影像間的食指部位則應於觸摸操作平面101時,粗估其視差間距介於上限值與下限值間,故進一步啟動以梯度辨識操作物程序242與以線形辨識操作物程序243,確認其為預設之操作物102。反之,若手臂握拳,不伸出食指,落在操作平面101,則其必有至少一視差間距也會介於上限值與下限值間,故須進一步由以梯度辨識操作物程序242與以線形辨識操作物程序243,排除其為適格之操作物102。 More specifically, an arm hand grasps the fist to extend the index finger, and intervenes in the established stereo operating system as an example. The first two-dimensional plane information 211 and the second two-dimensional plane information 221 obtained from the first photographic lens 210 and the second photographic lens 220 can roughly identify several objects having parallax spacing, and the palms or fists between the two images. The parallax distance of the part is greater than the preset upper limit, so it is not processed; and the index finger between the two images should be roughly touched between the upper limit and the lower limit when the touch operation plane 101 is touched. Further, the gradient recognition operator program 242 and the line recognition operator program 243 are activated to confirm that it is the preset operator 102. On the other hand, if the arm clenches the fist, does not extend the index finger, and falls on the operation plane 101, it must have at least one parallax distance between the upper limit value and the lower limit value, so it is necessary to further identify the operator program 242 with the gradient. The operator program 243 is identified in a line shape, excluding it as an eligible operator 102.

接下來,逐一介紹前述建立立體座標系程序241、以梯度辨識操作物程序242與以線形辨識操作物程序243的技術細節。請參考第4圖,第4圖是建立立體座標系程序241的運算步驟示意圖。首先,左上小圖的第一步驟中,C點與C’點為兩台已知的攝影機中心,其代表的物理意義包括但不限於上述的第一攝影鏡頭210與第二攝影鏡頭220。平面E和平面E’分別為與C點與C’點對應的影像平面。X點為三維空間中的一個參考點,其與攝影機中心C點與C’點共同組成一個對軸平面π。而且攝影機中心C點與C’點的連線組成一條基準線,基準線分別穿過影像平面E與E’,形成對軸點e與e’。另,攝影機中心C點與C’點分別連到參考點X,各自通過影像平面E與E’,形成另一組平面參考點x與x’。對軸點e與e’各自與平面參考點x與x’形成對軸線I與I’。 Next, the technical details of the above-described stereoscopic coordinate system program 241, the gradient recognition operator program 242, and the line recognition operator program 243 are introduced one by one. Please refer to FIG. 4, which is a schematic diagram of the operation steps of establishing the stereo coordinate system program 241. First, in the first step of the upper left panel, points C and C' are two known camera centers, and the physical meanings of the representatives include, but are not limited to, the first photographic lens 210 and the second photographic lens 220 described above. The plane E and the plane E' are image planes corresponding to points C and C', respectively. The X point is a reference point in the three-dimensional space, which together with the C point and the C' point of the camera center constitutes a pair of axis plane π . Moreover, the line connecting the C point and the C' point of the camera center constitutes a reference line, and the reference line passes through the image planes E and E' respectively, forming the pair of axis points e and e'. In addition, the C and C' points of the camera center are respectively connected to the reference point X, and each of them passes through the image planes E and E' to form another set of plane reference points x and x'. The pair of pivot points e and e' form a pair of axes I and I' with the plane reference points x and x', respectively.

接下來,右上小圖的第二步驟中,當X在三維空間中移動時,對軸平面π只是繞著由攝影機中心決定的基準線做旋轉而已。對 攝影機中心C點與C’點而言,X落在E平面的x位置及E’平面的x’位置。所以,當三維空間中的X沿著cx方向移動時,對攝影機中心C點來說,參考點X還是落在相同的x位置,但對於攝影機中心C’點而言,參考點X會在對軸平面π與影像平面E’交集的對軸線I’上移動;也就是說,當三維空間中的一個點與兩個攝影機中心共同決定一個平面時,X如果在對軸平面π上移動,則只要在攝影機C的視野E決定一個點即可以在攝影機C’的視野E’對應到一條線,這樣的關係稱為對軸幾何關係。 Next, in the second step of the upper right panel, when X moves in the three-dimensional space, the axis plane π is only rotated around the reference line determined by the center of the camera. For the C and C' points of the camera center, X falls at the x position of the E plane and the x' position of the E' plane. Therefore, when X in the three-dimensional space moves along the cx direction, the reference point X still falls at the same x position for the camera center C point, but for the camera center C' point, the reference point X will Moving on the axis I' of the intersection of the axis plane π and the image plane E'; that is, when a point in the three-dimensional space and a center of the two cameras jointly determine a plane, X moves on the axis plane π Then, as long as the field of view E of the camera C determines a point, it can correspond to a line in the field of view E' of the camera C'. Such a relationship is called a pair-axis geometric relationship.

因此,假設目前只從攝影機C看到一個點x而不知道它在三維空間中真正的位置X,由對軸幾何關係可以知道x會對應到攝影機C’視野的影像平面E’中的一條線I’;接下來,立體座標系程序241利用限制X在三維空間移動的位置來建立影像平面E與E’之間點與點的對應關係。藉此便可以將第一二維平面資訊211與第二二維平面資訊221整理成一個三維立體座標。 Therefore, suppose that only a point x is seen from the camera C without knowing its true position X in the three-dimensional space, and it can be known from the axis geometry that x will correspond to a line in the image plane E' of the camera C' field of view. I'll next, the stereo coordinate system program 241 establishes the correspondence between points and points between the image planes E and E' by limiting the position where X moves in the three-dimensional space. Thereby, the first two-dimensional plane information 211 and the second two-dimensional plane information 221 can be organized into one three-dimensional coordinate.

攝影機C的視野無法分辨X沿著cx方向移動的位置,因此立體座標系程序241先利用一個不通過攝影機中心點C與C’的平面Γ切過cx。在平面的Γ限制下,當攝影機C看到了一個三維空間的點X在影像平面x的位置時,由於X被限制在平面Γ上無法沿著cx自由移動,而可以得知該點會落在攝影機C’視野中的x’而不再是一條對軸線I’。易言之,透過一個平面Γ即可建立兩台攝影機之間的點對點關係。 The field of view of the camera C cannot distinguish the position where X moves in the cx direction, so the stereo coordinate system program 241 first cuts cx by using a plane that does not pass through the camera center points C and C'. Under the Γ limit of the plane, when the camera C sees the position X of a three-dimensional space at the position of the image plane x, since X is limited to the plane Γ, it cannot move freely along cx , and it can be known that the point will be x' falling in the field of view of camera C' is no longer a pair of axes I'. In other words, a point-to-point relationship between two cameras can be established through a plane.

請再參考右下小圖的第三步驟,由三維空間中的一個平面Γ 建立兩台攝影機之間的對應,這樣的關係稱為平面對應關係。在平面對應關係中,當攝影機C的視野看到x,將射線cx延伸和平面Γ相交可以得知x是由Γ平面上的x Γ產生的;意即其可由x=H ΓC x Γ描述之。而另一方面,x Γ對攝影機C’的視野而言,是落在x’,意即其可由x'=H ΓC' x Γ描述之。結合這兩台攝影機對一共同平面的關係,即可建立兩個影像平面之間的平面對應關係為x'=H ΓC' H CΓ x=Hx;而且,因為H ΓC H ΓC' 皆是1對1映射的關係,所以H是可逆的矩陣,即x=H -1 x'Please refer to the third step of the lower right image to establish the correspondence between the two cameras from a plane in the three-dimensional space. Such a relationship is called a plane correspondence. In the plane correspondence, when the field of view of camera C sees x, the intersection of ray cx and plane Γ can be seen that x is generated by x Γ on the plane of the ;; that is, it can be x = H Γ C x description of Γ. On the other hand, x Γ falls on x' for the field of view of camera C', meaning that it can be described by x' = H Γ C' x Γ . Combining the relationship between the two cameras to a common plane, the plane correspondence between the two image planes can be established as x' = H Γ C' H C Γ x = Hx ; and, because H Γ C and H Γ C ' All are 1-to-1 mapping relationships, so H is a reversible matrix, ie x = H -1 x' .

最後,如左下小圖的第四步驟所示,藉由平面對應關係H建立兩台攝影機於三維空間中一個平面的對應關係之後,便可進一步探討當三維空間中的參考點X遠離先前所定義的平面Γ時,所產生的視差。申言之,透過平面Γ建立攝影機C與C’之間的平面對應關係H之後,對於任何一個出現在影像平面E的點x都可以透過平面對應關係H找到它在另一個影像平面E’的對應點=Hx。當X不在平面Γ上且仍然落在射線cx上,則對於C來說x和x Γ依然對應到相同的點X上。因此,透過平面對應關係H會將X映射到上;但是,對於C’而言X事實上應該落在對軸線I’的另一個點x’上,與x’之間的差距即為由平面Γ所衍生的視差。 Finally, as shown in the fourth step of the lower left panel, after the correspondence between the two cameras in one plane in three-dimensional space is established by the plane correspondence relationship H, it can be further explored that the reference point X in the three-dimensional space is far from the previous definition. The parallax produced when the plane is flawed. It is stated that after the plane correspondence H between the camera C and the C' is established through the plane ,, for any point x appearing in the image plane E, it can be found in the other image plane E' through the plane correspondence H. Corresponding point =Hx. When X is not on the plane 且 and still falls on the ray cx , x and x Γ still correspond to the same point X for C. Therefore, through the plane correspondence H will map X to Above; however, for C', X should actually fall on another point x' to axis I', The difference from x' is the parallax derived from the plane Γ.

接下來,本實施方式以第5-9圖介紹以梯度辨識操作物程序242的技術細節。請先參考第5圖,第5圖是操作物102在背景畫面的操作示意圖。從第5圖中可觀察到,操作平面101上有多媒 體背景畫面存在;此時,若多媒體背景畫面係以投影方式形成,則其勢必直接在操作物102上產生光影干擾103,影響色塊的判斷。更具體的來說,要偵測的操作物102,即手指,其受到了投影機投影的畫面所干擾,因此除了背景的變化之外,操作物102也會隨著投影畫面的內容不同而有色彩上的改變。在動態的背景以及隨背景而改變色彩的前景中,較穩定的資訊為邊緣;因此,接下來教示如何由邊緣資訊準確的找出手指的位置。 Next, the present embodiment describes the technical details of the gradient recognition operator program 242 in FIGS. 5-9. Please refer to FIG. 5 first, and FIG. 5 is a schematic diagram of the operation of the operator 102 on the background screen. It can be observed from Fig. 5 that there is a multi-media on the operation plane 101. The body background picture exists; at this time, if the multimedia background picture is formed by projection, it is bound to directly generate the light and shadow interference 103 on the operator 102, which affects the judgment of the color block. More specifically, the operator 102 to be detected, that is, the finger, is interfered by the image projected by the projector, so in addition to the change of the background, the operator 102 may have a different content depending on the projected image. The change in color. In the dynamic background and the foreground where the color changes with the background, the more stable information is the edge; therefore, the next step is how to accurately find the position of the finger from the edge information.

在實作中,除了環境光源之外,投影機也產生一個動態的光源,因此會遭遇到高動態範圍的亮度變化,為了有效的使用邊緣資訊,此處採用特徵描述算子(feature descriptor)建立方向梯度圖(Histogram of Orientation Gradients,HOG),藉以利用物體的邊緣會隨著對比的不同而在強度上有明顯差異的特徵。方向梯度算子(HOG descriptor)將影像視為許多小區域的組合並收集這些區域的特徵,透過方向的量化來處理物體產生的簡單形變,透過每一個小區域之間的正交歸化(normalization)來減輕高對比影像所產生不同邊緣的強度,詳述如下。 In practice, in addition to the ambient light source, the projector also produces a dynamic light source, so it will encounter high dynamic range brightness changes. In order to use the edge information effectively, the feature descriptor is used here. Histogram of Orientation Gradients (HOG), which is used to make the edge of an object significantly different in intensity depending on the contrast. The HOG descriptor treats the image as a combination of many small regions and collects the features of these regions. The direction is quantized to handle the simple deformation produced by the object, through normalization between each small region (normalization) To reduce the intensity of the different edges produced by high contrast images, as detailed below.

請一併參考第6圖與第7圖,其皆係方向梯度特徵描述運算的說明圖。方向梯度算子(HOG descriptor)主要是在計算影像的梯度,再針對方向量化之後統計每一個小區塊中的值,藉以透過量化及局部的統計來有效描述物體局部的外表和形狀。在方向梯度特徵描述運算中,首先利用差分濾波器(difference filters)在影像的每一個像素(pixel)都計算出梯度(gradients)。令r(x, y)和θ(x,y)分別表示強度梯度(intensity gradient)在像素pixel(x,y)的大小(magnitude)和取向(orientation)。每一個像素的梯度方向(gradient orientation)都利用下列方程式(2)或(3)量化成p個bin(此處例示p=9): 其中,B 1代表對比敏感度(contrast sensitive)的量化方式,而B 2代表對比不敏感度(contrast insensitive)的量化方式。然後,令b {0,1,...,p-1},則影像的特徵圖(feature map)由下列方程式(4)來完成(B值表示B 1B 2): Please refer to Fig. 6 and Fig. 7 together, which are explanatory diagrams of the direction gradient feature description operation. The HOG descriptor is mainly used to calculate the gradient of the image, and then counts the values in each cell block after the direction quantization, so as to effectively describe the appearance and shape of the object through quantization and local statistics. In the directional gradient characterization operation, gradients are first calculated for each pixel (pixel) of the image using differential filters. Let r(x, y) and θ (x, y) denote the magnitude and orientation of the intensity gradient at the pixel pixel(x, y), respectively. The gradient orientation of each pixel is quantized into p bins using the following equation (2) or (3) (here p=9 is illustrated): Among them, B 1 represents the quantization mode of contrast sensitivity, and B 2 represents the quantization mode of contrast insensitive. Then, let b {0,1,..., p -1}, the feature map of the image is done by the following equation (4) (B value means B 1 or B 2 ):

承上所述,函數F表示一個邊緣(edge)會被看成某一個特定的方向依照邊緣(edge)的強度voting到對應的bin。處理完每一個像素的圖之後,再統整這些資訊,將影像分割成許多小區域(cell)的組合。透過這些邊緣的統計能夠有效的描述物體的特性。舉例來說,小區域(cell)是8x8的像素,每一個區域可以看成是一個維度p的向量。為了減少物體恰巧出現在所切割的區域之間,每一個像素的邊緣強度除了貢獻給自己位置所在的區域,也會插值(interpolation)到其他鄰近的四個區域。由於梯度的值和影像的對比有很大的關係,因此區域之間還會做正交歸化(normalization),亦即一個小區域(cell)和鄰近的三個小區域看成是一個區 塊(block)。 As stated above, the function F indicates that an edge is seen as a particular direction voting to the corresponding bin according to the strength of the edge. After processing the map for each pixel, the information is then integrated to divide the image into a combination of many small cells. Statistics through these edges can effectively describe the characteristics of an object. For example, a small cell is an 8x8 pixel, and each region can be viewed as a vector of dimension p. In order to reduce the occurrence of objects happening between the cut areas, the edge intensity of each pixel is interpolated to other adjacent four areas in addition to the area where the position is contributed to itself. Since the value of the gradient has a great relationship with the contrast of the image, normalization is also performed between the regions, that is, a small cell and three adjacent small regions are regarded as one region. Block.

接下來,令δ,γ {-1,1}則正交歸化運算子(normalization factors)如下列方程式(5)所示: 其中,每一個區域和不同的幾個鄰近區域都做一次歸化(normalize),因此產生了四種歸化之後的值,如下列方程式(6)所示: 式中的T α (v)代表向量v中的每個元素和α比較之後選擇較小的值。 Next, let δ , γ {-1, 1} The normalization factors are as shown in the following equation (5): Among them, each region and several different adjacent regions are normalized, thus producing four values after naturalization, as shown in the following equation (6): T α ( v ) in the equation represents that each element in the vector v is selected to be smaller after comparison with α .

透過這些流程,對任一張輸入的影像都能得到對應的特徵圖。最後只要利用支持向量機制(Support Vector Machine,SVM)進行監督學習運算(supervised learning method)即可得到操作物102的HOG模型104,如第6圖所示。 Through these processes, the corresponding feature map can be obtained for any of the input images. Finally, the HOG model 104 of the operator 102 can be obtained by using a Support Vector Machine (SVM) to perform a supervised learning method, as shown in FIG.

最後,利用已知的HOG模型104在任何一張新輸入的影像是採用滑移視窗法(sliding-window)偵測操作物102的位置,其過程是將HOG模型104當作是一個過濾器(filter F),這個過濾器包含不同權重的向量。而偵測的過程是利用這個過濾器(filter F)在影像特徵圖H當中從左上角表示為(x,y)的子視窗(sub-window)開始,採用內積的方式計算其分數,如下列方程式(7)所示: 在計算完整張影像每一個位置的分數之後,即可透過簡單的門檻 值來找出欲偵測的物體。 Finally, using the known HOG model 104, the position of the operator 102 is detected using a sliding-window on any newly input image by treating the HOG model 104 as a filter ( Filter F), this filter contains vectors of different weights. The detection process uses the filter (filter F) to start the sub-window of the (x, y) from the upper left corner of the image feature map H, and calculate the score by using the inner product. The following equation (7) shows: After calculating the score for each position of the complete image, the object to be detected can be found through a simple threshold value.

然而,雖然HOG模型104可以偵測操作物102的位置,但是其對於物體的旋轉是相當敏感的。以第7圖為例,當背景畫面中有許多類似外型105存在時,在HOG模型104對物體旋轉敏感的特性下,都容易出現誤判。具體說明之,在將梯度取向(gradient orientation)voting到最鄰近的bin當中時,物體旋轉時產生的voting結果會和原先預期的不同。現實應用中,操作物102常會有輕微的旋轉,這些旋轉產生的特徵圖(feature map)幾乎都不同。在計算分數時,只有和HOG模型104同樣方向的物體可以得到較高的分數,進而鑑別其為操作物102。當旋轉的角度和HOG模型104差異越大,計算分數的結果也會得到越低的分數。因此只要操作物102的旋轉角度過大,計算完的分數就不足以將操作物102從背景中分離出來,進而在類似外型105處產生許多誤動作(false alarms)。 However, although the HOG model 104 can detect the position of the operator 102, it is quite sensitive to the rotation of the object. Taking Fig. 7 as an example, when there are many similar appearances 105 in the background picture, the HOG model 104 is susceptible to object rotation, and misjudgment is prone to occur. Specifically, when the gradient orientation is voted into the nearest bin, the result of the voting generated when the object is rotated may be different from what was originally expected. In real-world applications, the operators 102 often have slight rotations, and the feature maps produced by these rotations are almost all different. When calculating the score, only objects in the same direction as the HOG model 104 can obtain a higher score, thereby identifying it as the operator 102. When the angle of rotation differs from the HOG model 104, the result of calculating the score will also get a lower score. Therefore, as long as the angle of rotation of the operator 102 is too large, the calculated score is insufficient to separate the operator 102 from the background, thereby producing a number of false alarms at the similar appearance 105.

為了克服上述缺陷,本實施方式於以梯度辨識操作物程序242中進一步引入基於部分模型的HOG運算方式;細論之,由於操作物102通常是會有微幅變動的,因此除了原本的HOG模型104之外,另外加入一些可移動的部分模型,使之隨著操作物102的部分變動來改變位置。具體請參考第8圖,第8圖是基於部分模型的HOG運算方式的示意圖,其係利用原本的HOG模型104當做本過濾器106(root filter),另外加入數個部分過濾器107(part filters)來描述物體變動較大的部分,而部分過濾器107的移動並 不是毫無限制的,每個部分過濾器107有對應的變形成本108(deformation cost)。 In order to overcome the above drawbacks, the present embodiment further introduces a partial model based HOG operation method in the gradient identification operator program 242; in detail, since the operator 102 usually has a slight variation, the original HOG model is excluded. In addition to 104, some movable partial models are additionally added to change position as part of the operating object 102 changes. For details, please refer to FIG. 8. FIG. 8 is a schematic diagram of the HOG operation mode based on the partial model. The original HOG model 104 is used as the root filter, and a plurality of partial filters 107 are added. ) to describe the part where the object changes greatly, and the partial filter 107 moves and Not limited, each partial filter 107 has a corresponding deformation cost 108.

具體而言,本過濾器106(root filter)和部分過濾器107(part filters)分別在不同的解析度下計算,也就是本過濾器106(root filter)找到大概的物體位置再透過高解析度的部分過濾器107(part filters)去覆蓋物體的小部分變動。此時,雖然操作物102的外型會改變,但是可變動的部位並不是毫無限制的移動,每一個部分過濾器107(part filters)會有一個定錨(anchor)的位置來確保這些部分的移動是合乎常理的。 Specifically, the filter 106 (part filter) and the partial filter 107 (part filters) are respectively calculated at different resolutions, that is, the filter 106 (root filter) finds the approximate object position and then passes through the high resolution. Part of the filter 107 (part filters) to cover a small part of the object changes. At this time, although the appearance of the operator 102 may change, the variable portion is not unrestricted movement, and each partial filter 107 has an anchor position to ensure these portions. The movement is reasonable.

換句話說,定錨(anchor)就是可變動部位最有機會出現的位置,當部位離a定錨(anchor)越遠,對應的成本(cost)就會越大,如下列方程式(8)所述:。 In other words, the anchor is the most likely position for the variable part to appear. The farther the part is from the anchor, the higher the cost will be, as shown in the following equation (8). Said:

其中,p 0表示本過濾器106,而p 1,...,p n 表示n個部分過濾器107,函數F'.(H,p i )表示將特徵圖H中左上角為p的子視窗(sub-window)中的特徵向量排成一行;而F'則是F排成一行之後的轉置。每一個部份模型由三個部分組成,以(F i ,v i ,d i )表示,其中F i 是第i個部分的過濾器;v i 是一個二維的向量,用以表示第i個部分的定錨位置;d i 則是一個四維的向量用來表示二次函數的係數,其目的是用來表示部分離開定錨時所對應的成本函數。當d i =(0,0,1,1) 則第i個部份的變形成本只是部分和定錨之間單純二維空間的距離。 Where p 0 denotes the present filter 106, and p 1 , . . . , p n denotes n partial filters 107, function F '. ( H , p i ) denotes a feature vector in a sub-window in which the upper left corner of the feature map H is p; and F ' is a transpose after F is lined up. Each partial model consists of three parts, denoted by ( F i , v i , d i ), where F i is the filter of the i-th part; v i is a two-dimensional vector representing the i-th The fixed anchor position of the part; d i is a four-dimensional vector used to represent the coefficient of the quadratic function, and its purpose is to represent the cost function corresponding to the partial leaving anchor. When d i = (0, 0, 1, 1), the deformation cost of the i-th part is only the distance between the part and the fixed two-dimensional space between the anchors.

接下來,為了計算本體在每一個位置所對應到最好的部分位置,可以透過下列方程式(9)表示之: 剩下的步驟,就跟前述HOG運算的步驟相同,透過門檻值的選取即可偵測出物體的位置。 Next, in order to calculate the position of the body corresponding to the best part at each position, it can be expressed by the following equation (9): The remaining steps are the same as the steps of the HOG operation described above, and the position of the object can be detected by the selection of the threshold value.

最後,本實施方式以第9-11圖介紹以線形辨識操作物程序243的技術細節。承前所述,雖然採用本過濾器106(root filter)和部分過濾器107(part filters)的HOG模型可以準確找到操作物102的位置,但是以小區域(cell)為基礎的運算觀念,可能不足以將操作物102資訊處理到像素等級(pixel level)。 Finally, the present embodiment describes the technical details of the linear recognition operator program 243 in the manner of Figures 9-11. As mentioned above, although the HOG model of the filter 106 (part filter) and the partial filter 107 can accurately find the position of the operator 102, the concept of operation based on a small cell may be insufficient. The information of the operator 102 is processed to a pixel level.

以操作物102為手指作例子,指尖中心的精準度會直接影響觸控點位置的準確度及觸碰的判定;為了提高系統運作的精確度,實有將指尖中心的判斷精準度提升的需求。因此,本實施方式於以線形辨識操作物程序243中,以線形模型(shape model)來詳細描述手指的輪廓,利用手指已知的輪廓將前述包含部分模型的HOG模型104所偵測之後的誤動作(false alarms)移除,並且利用手指的輪廓來找到指尖的中心位置。線形模型(shape model)的特色是除了可以偵測物體之外,還可以將物體的外型描述出來。由於線形模型(shape model)建構在局部的輪廓特徵(pairs of adjacent segments,PAS),因此需先教示說明局部輪廓特徵(P AS)的偵測基礎原理,即輪廓網段(contour segment network,CSN)如下:首先,因為邊緣(edge)的偵測容易受到影像亮度及對比的影響,因此一條邊緣(edge)經常會被切成好幾段破碎的邊緣鍊(edge chains)。因此,連接邊緣鍊的運算方式,就是將這些被切斷的地方記錄起來。舉例來說,如果邊緣鍊C2中有一個點落在邊緣鍊C1的搜尋範圍內,則邊緣鍊C1會和邊緣鍊C2連起來;此處請參考第9圖,第9圖是兩條邊緣鍊的連接示意圖。值得注意的是,如第9圖所示,搜尋範圍109採用梯型的意義在於隨著離邊緣鍊的端點距離越遠,另一條邊緣鍊出現的位置的不確定性也隨著增加。 Taking the operator 102 as a finger as an example, the accuracy of the center of the fingertip directly affects the accuracy of the touch point position and the determination of the touch; in order to improve the accuracy of the system operation, the accuracy of the fingertip center is improved. Demand. Therefore, in the linear identification operator program 243, the present embodiment describes the outline of the finger in a line model, and uses the contour known by the finger to detect the malfunction of the HOG model 104 including the partial model. (false alarms) Remove and use the outline of the finger to find the center of the fingertip. The feature of the shape model is that in addition to detecting objects, the shape of the object can be described. Since the shape model is constructed in pairs of adjacent segments (PAS), it is necessary to teach the local contour features (P). The basic principle of AS detection is that the contour segment network (CSN) is as follows: First, because the edge detection is easily affected by image brightness and contrast, an edge is often cut. A few pieces of broken edge chains. Therefore, the way to connect the edge chain is to record these cuts. For example, if there is a point in the edge chain C2 falling within the search range of the edge chain C1, the edge chain C1 will be connected with the edge chain C2; here, refer to Figure 9, which is the two edge chains. Connection diagram. It is worth noting that, as shown in Fig. 9, the significance of the search range 109 using the ladder type is that as the distance from the end point of the edge chain is further away, the uncertainty of the position of the other edge chain also increases.

接下來,在連接邊緣鍊之後,須將邊緣鍊分割成近似於直線的輪廓段(contour segments)。此時,輪廓段的段橋(segments bridging)定義了其本身切線連續(tangent-continuous)的連接,段橋(segments bridging)可以有效的將斷掉的邊緣修復。輪廓網段(CSN)即主要建構在連接邊緣鍊(edge chain links)和輪廓段(contour segments)的利用上,其係以連接邊緣鍊(edge chain links)將可能遺失的邊緣資訊記錄下來,再基於輪廓段(contour segments)利用近似於直線的線段來重組整個網段(network)。具體請參考第10圖,第10圖是輪廓網段的運作示意圖,其係繪示下列六種修補的規則: Next, after joining the edge chains, the edge chains must be segmented into contour segments that approximate the line. At this point, the segment bridging of the contour segment defines its own tangent-continuous connection, and segment bridging can effectively repair the broken edge. The contour network segment (CSN) is mainly constructed on the use of edge chain links and contour segments, which record the missing edge information by connecting the edge chain links. The entire network segment is reorganized based on the contour segments using line segments that approximate the line. For details, please refer to Figure 10, which is a schematic diagram of the operation of the contour network segment, which depicts the following six patching rules:

(1)如第10圖中規則一(Rule 1)所示:在同一個邊緣鍊上的 其中一個線段前端點會連接到另一個線段的後端點。 (1) As shown in Rule 1 in Figure 10: on the same edge chain One of the line segment front points is connected to the back end of the other line segment.

(2)如第10圖中規則二(Rule 2)所示:當兩個邊緣鍊(C1,C2)在端點連接時,邊緣鍊C1的線段會連接到邊緣鍊C2的線段。 (2) As shown in Rule 2 in Figure 10: When two edge chains (C1, C2) are connected at the end points, the line segment of the edge chain C1 is connected to the line segment of the edge chain C2.

(3)如第10圖中規則三(Rule 3)所示:考慮T型接面的連接時,邊緣鍊C1的線段會連接到邊緣鍊C2兩個線段的最近端點。 (3) As shown in Rule 3 in Figure 10: When considering the connection of the T-junction, the line segment of the edge chain C1 is connected to the nearest end point of the two line segments of the edge chain C2.

(4)如第10圖中規則四(Rule 4)所示:令s為連接邊緣鍊C1與邊緣鍊C2的段橋,段橋s會在邊緣鍊C1線段的前端點之後,且在邊緣鍊C2線段的後端點之前。 (4) As shown in Rule 4 (Rule 4): Let s be the segment bridge connecting the edge chain C1 and the edge chain C2, the segment bridge s will be after the front end point of the edge chain C1 line segment, and at the edge chain Before the rear end of the C2 line segment.

(5)如第10圖中規則五(Rule 5)所示:同一個邊緣鍊上的兩個線段如果有連續的端點,則連接在一起。 (5) As shown in Rule 5 in Figure 10: Two line segments on the same edge chain are connected together if there are consecutive endpoints.

(6)如第10圖中規則六(Rule 6)所示:考慮一個沒有前端點連接的段橋s,因其覆蓋了邊緣鍊C2的前端點,且延續到邊緣鍊C2的後段,故若邊緣鍊C2連接到另一個邊緣鍊C3,則段橋s亦連接到邊緣鍊C3的線段。 (6) As shown in Rule 6 (Rule 6): Consider a segment bridge s without a front-end point connection, because it covers the front end point of the edge chain C2 and continues to the back segment of the edge chain C2, so The edge chain C2 is connected to the other edge chain C3, and the segment bridge s is also connected to the line segment of the edge chain C3.

承上所述,因為每一個邊緣鍊都可能連接到其他數個邊緣鍊,故輪廓網段(CSN)是一個相當複雜的網狀結構,而本體特徵(local features)可描述一對連接線段之間的關係,進而簡化整個問題的複雜度。亦即,兩兩之間的線段會被看成一個局部輪廓特徵(PAS)。接下來,說明如何利用數學方程來描述局部輪廓特徵(PAS)如下:局部輪廓特徵(PAS)P=(x,y,s,e,d)是由一個位置(x,y)、一個級距參數s、一個強度參數e及一個描述子d=(θ 1,θ 2,l 1,l 2,)所組成; 請參考第11圖,第11圖是局部輪廓特徵的運算示意圖。第11圖中,因為局部輪廓特徵(PAS)係以線段(segments)連接邊緣鍊(edge chains)之間的空隙,因此局部輪廓特徵(PAS)可以很穩定的偵測物體。由於局部輪廓特徵(PAS)只包含單獨的兩個線段,其可以有效且單純的還原物體的邊界並去除雜亂的背景。 As mentioned above, because each edge chain is likely to be connected to several other edge chains, the contoured network segment (CSN) is a rather complex mesh structure, and the local features can describe a pair of connected segments. The relationship between the two, which in turn simplifies the complexity of the entire problem. That is, the line segment between the two is considered to be a partial contour feature (PAS). Next, explain how to use the mathematical equation to describe the local contour feature (PAS) as follows: local contour feature (PAS) P = ( x , y , s , e , d ) is a position (x, y), a step The parameter s, an intensity parameter e and a descriptor d = ( θ 1 , θ 2 , l 1 , l 2 , The composition is as follows; please refer to Fig. 11, which is a schematic diagram of the operation of the local contour feature. In Fig. 11, since the local contour feature (PAS) connects the gaps between the edge chains by segments, the local contour feature (PAS) can detect the object stably. Since the local contour feature (PAS) contains only two separate line segments, it can effectively and simply restore the boundary of the object and remove the cluttered background.

局部輪廓特徵(PAS)之間的差異性透過D(P,Q)來計算,D(P,Q)表示兩個PAS描述子dp,dq之間的相異程度,如下列方程式(10): The difference between the local contour features (PAS) is calculated by D(P, Q), and D(P, Q) represents the degree of dissimilarity between the two PAS descriptors dp, dq, as shown in the following equation (10):

方程式(10)的第一項係計算兩個線段相對位置的差異,而第二項D θ [0,π/2]則係計算線段之間的取向(orientation)差異,第三項則計算他們的長度差異。舉例來說,一個操作中的手指可能會有輕微的旋轉,因此可將方程式(10)依需求改寫為下列方程式(11),以解決輕微的旋轉造成PAS之間的差異加大: 在方程式(11)中,進一步考慮了兩個PAS特徵之間的相對取向差距。 The first term of equation (10) is to calculate the difference between the relative positions of the two line segments, and the second term D θ [0, π /2] calculates the difference in orientation between line segments, and the third term calculates the difference in length. For example, a finger in an operation may have a slight rotation, so Equation (10) can be rewritten to the following equation (11) as needed to account for the slight difference between the PAS caused by a slight rotation: In equation (11), the relative orientation difference between the two PAS features is further considered.

因此,以線形辨識操作物程序243的線形模型(shape model)即是藉由事先定義一個想偵測的物體,例如操作物102,利用其所包含的PAS特徵會重複的出現在訓練資料中,透過收集這些PAS特徵來建立一個碼書(code book)C={t i }。這個碼書(code book)包含了格式T{t i }的PAS描述子,而V i 為對應到t i 的投射空間(voti ng space)。每一個投射空間(voting space)都是一個三維的空間,其中兩個維度做為位置,另外一個維度為特徵大小。每一個PAS會分配給許多格式T,只要他們的PAS差異性足夠小,如下列方程式(12)所示:γT={t j |D(d,t j )<γ} (12)而投射的權重則由PAS和該格式的差異性決定,差異越小則權重越高,如下列方程式(13)所描述:e.(1-D(d,t j )/γ) (13) Therefore, the linear shape of the operator program 243 is determined by predefining an object to be detected, such as the operator 102, and the PAS features contained therein are repeatedly displayed in the training material. A code book C={ t i } is created by collecting these PAS features. This code book contains the PAS descriptor of the format T{ t i }, and V i is the projection space corresponding to t i (voti ng space). Each voting space is a three-dimensional space in which two dimensions are used as positions and the other dimension is feature size. Each PAS is assigned to many formats T as long as their PAS variability is small enough, as shown in the following equation (12): γ : T = { t j | D ( d , t j ) < γ } (12) The weight of the projection is determined by the difference between the PAS and the format. The smaller the difference, the higher the weight, as described in the following equation (13): e . (1- D ( d , t j )/ γ ) (13)

惟,以上所述僅為本發明之各種實施方式而已,非因此即侷限本發明之專利範圍,故舉凡運用本發明說明書及圖式內容所為之簡易修飾及等效結構變化,均應同理包含於本發明之專利範圍內。 However, the above description is only for the various embodiments of the present invention, and is not intended to limit the scope of the present invention. Therefore, the simple modifications and equivalent structural changes of the present invention and the contents of the drawings should be included in the same manner. Within the scope of the patent of the present invention.

100‧‧‧桌上型光學觸控裝置 100‧‧‧Tabletop optical touch device

110‧‧‧投影機 110‧‧‧Projector

120‧‧‧紅外線接收鏡頭 120‧‧‧Infrared receiving lens

130‧‧‧紅外線發射器 130‧‧‧Infrared emitter

101‧‧‧操作平面 101‧‧‧Operational plane

102‧‧‧操作物 102‧‧‧Operator

103‧‧‧光影干擾 103‧‧‧Light and shadow interference

104‧‧‧HOG模型 104‧‧‧HOG model

105‧‧‧類似外型 105‧‧‧similar appearance

106‧‧‧本過濾器 106‧‧‧This filter

107‧‧‧部分過濾器 107‧‧‧Partial filter

108‧‧‧變形成本 108‧‧‧Transformation costs

109‧‧‧搜尋範圍 109‧‧‧Search range

200‧‧‧光學式觸控系統 200‧‧‧Optical touch system

210‧‧‧第一攝影鏡頭 210‧‧‧ first photographic lens

211‧‧‧第一二維平面資訊 211‧‧‧ first two-dimensional information

220‧‧‧第二攝影鏡頭 220‧‧‧second photographic lens

221‧‧‧第二二維平面資訊 221‧‧‧Second 2D Plane Information

201、202‧‧‧視差間距 201, 202‧‧‧ Parallax spacing

230‧‧‧處理單元 230‧‧‧Processing unit

240‧‧‧處理機制 240‧‧ ‧ treatment mechanism

241-244‧‧‧步驟 241-244‧‧‧Steps

第1圖 係為習知之桌上型光學觸控裝置的操作原理示意圖。 Figure 1 is a schematic diagram of the operation principle of a conventional desktop optical touch device.

第2圖 係為本實施方式之光學式觸控系統200的結構示意圖。 FIG. 2 is a schematic structural view of the optical touch system 200 of the present embodiment.

第3圖 係為本實施方式以視差間距確認操作物位置的原理示意圖。 Fig. 3 is a schematic view showing the principle of confirming the position of the operator by the parallax distance in the present embodiment.

第4圖 係為建立立體座標系程序241的運算步驟示意圖。 Fig. 4 is a schematic diagram showing the operation steps of the stereo coordinate system program 241.

第5圖 係為操作物102在背景畫面的操作示意圖。 Figure 5 is a schematic diagram of the operation of the operator 102 on the background screen.

第6圖 係為方向梯度特徵描述運算的說明圖。 Figure 6 is an explanatory diagram of the direction gradient feature description operation.

第7圖 係為方向梯度特徵描述運算的說明圖。 Figure 7 is an explanatory diagram of the directional gradient characterization operation.

第8圖 係為基於部分模型的HOG運算方式的示意圖。 Figure 8 is a schematic diagram of a HOG operation based on a partial model.

第9圖 係為兩條邊緣鍊的連接示意圖。 Figure 9 is a schematic diagram of the connection of two edge chains.

第10圖 係為輪廓網段的運作示意圖 Figure 10 is a schematic diagram of the operation of the contour network segment

第11圖 係為局部輪廓特徵的運算示意圖。 Figure 11 is a schematic diagram of the operation of the local contour feature.

110‧‧‧投影機 110‧‧‧Projector

101‧‧‧操作平面 101‧‧‧Operational plane

102‧‧‧操作物 102‧‧‧Operator

200‧‧‧光學式觸控系統 200‧‧‧Optical touch system

210‧‧‧第一攝影鏡頭 210‧‧‧ first photographic lens

211‧‧‧第一二維平面資訊 211‧‧‧ first two-dimensional information

220‧‧‧第二攝影鏡頭 220‧‧‧second photographic lens

221‧‧‧第二二維平面資訊 221‧‧‧Second 2D Plane Information

230‧‧‧處理單元 230‧‧‧Processing unit

240‧‧‧處理機制 240‧‧ ‧ treatment mechanism

241-244‧‧‧步驟 241-244‧‧‧Steps

Claims (10)

一種光學式觸控系統,係用以辨識一操作物於一三維操作空間中的位置,包括:一第一攝影鏡頭,係用以自該三維操作空間中,取得一第一二維平面資訊;一第二攝影鏡頭,係用以自該三維操作空間中,取得一第二二維平面資訊;以及一運算單元,係用以利用該第一二維平面資訊與該第二二維平面資訊,辨識該操作物的位置;其中,該運算單元係執行一處理機制來辨識該操作物的位置,該處理機制包括下列步驟:步驟一:利用該第一二維平面資訊與該第二二維平面資訊建立一立體座標系;步驟二:以該第一二維平面資訊與該第二二維平面資訊中的梯度變化,辨識該操作物;步驟三:以該第一二維平面資訊與該第二二維平面資訊中的線形資訊,辨識該操作物;以及步驟四:以該操作物於該第一二維平面資訊與該第二二維平面資訊中的視差間距,定義該操作物的位置。 An optical touch system for identifying a position of an operation object in a three-dimensional operation space, comprising: a first photographic lens for acquiring a first two-dimensional plane information from the three-dimensional operation space; a second photographic lens for acquiring a second two-dimensional plane information from the three-dimensional operation space; and an operation unit for utilizing the first two-dimensional plane information and the second two-dimensional plane information, Identifying the location of the operator; wherein the computing unit performs a processing mechanism to identify the location of the operator, the processing mechanism includes the following steps: Step 1: using the first two-dimensional plane information and the second two-dimensional plane The information establishes a three-dimensional coordinate system; step two: identifying the operation object by using the first two-dimensional plane information and the gradient change in the second two-dimensional plane information; and step three: using the first two-dimensional plane information and the first The linear information in the two-dimensional plane information identifies the operator; and the fourth step: defining the parallax distance between the first two-dimensional plane information and the second two-dimensional plane information The position of the operator. 如申請專利範圍第1項所述之光學式觸控系統,其中該運算單元係於執行該步驟二時,事先建立該操作物之一方向梯度圖(HOG) 模型,且將該方向梯度圖(HOG)模型當作過濾器,分別在該第一二維平面資訊與該第二二維平面資訊中辨識該操作物。 The optical touch system of claim 1, wherein the operation unit is configured to perform a step gradient map (HOG) of the operator in advance. And using the directional gradient map (HOG) model as a filter, respectively identifying the operator in the first two-dimensional plane information and the second two-dimensional plane information. 如申請專利範圍第2項所述之光學式觸控系統,其中該方向梯度圖(HOG)模型係由一本過濾器(root filter)與複數個部分過濾器(part filters)所組成。 The optical touch system of claim 2, wherein the directional gradient map (HOG) model is composed of a root filter and a plurality of partial filters. 如申請專利範圍第3項所述之光學式觸控系統,其中該複數個部分過濾器(part filters)係各自有相對應之變形成本,以定錨(anchor)在該本過濾器(root filter)之周邊。 The optical touch system of claim 3, wherein the plurality of partial filters each have a corresponding deformation cost to anchor an anchor in the filter (root filter) Around it. 如申請專利範圍第1項所述之光學式觸控系統,其中該運算單元係於執行該步驟三時,事先建立該操作物之一線形模型來描述該操作物的輪廓,進而分別在該第一二維平面資訊與該第二二維平面資訊中辨識該操作物。 The optical touch system of claim 1, wherein the computing unit is configured to perform a step of the third step to establish a linear model of the operating object to describe the contour of the operating object, and then respectively The two-dimensional plane information and the second two-dimensional plane information identify the operator. 如申請專利範圍第5項所述之光學式觸控系統,其中該運算單元在處理該第一二維平面資訊與該第二二維平面資訊中的線形資訊時,係將複數個破碎的邊緣鍊以至少一段橋,修補連接成至少一輪廓段。 The optical touch system of claim 5, wherein the computing unit is configured to process a plurality of broken edges when processing the linear information in the first two-dimensional plane information and the second two-dimensional plane information. The chain is patched into at least one contour segment with at least one section of the bridge. 如申請專利範圍第1項所述之光學式觸控系統,該處理機制更 包括:當一物件介入該第一二維平面資訊與該第二二維平面資訊,且偵測該物件之視差間距介於一上限值與一下限值間,執行該步驟二與該步驟三,以判定該物件是否為該操作物。 For example, the optical touch system described in claim 1 of the patent scope, the processing mechanism is more The method includes: when an object is involved in the first two-dimensional plane information and the second two-dimensional plane information, and detecting that the parallax distance of the object is between an upper limit value and a lower limit value, performing step 2 and step three To determine whether the object is the operator. 如申請專利範圍第7項所述之光學式觸控系統,該處理機制更包括:當該物件之視差間距低於該下限值,直接判定該物件非該操作物。 The optical touch system of claim 7, wherein the processing mechanism further comprises: determining that the object is not the operator when the parallax spacing of the object is lower than the lower limit. 如申請專利範圍第7項所述之光學式觸控系統,該處理機制更包括:當該物件之視差間距高於該下限值,直接判定該物件未觸發一觸控事件。 The optical touch system of claim 7, wherein the processing mechanism further comprises: when the parallax spacing of the object is higher than the lower limit, directly determining that the object does not trigger a touch event. 一種光學式觸控系統,係用以辨識一操作物於一三維操作空間中的位置,包括:一第一攝影鏡頭,係用以自該三維操作空間中,取得一第一二維平面資訊;一第二攝影鏡頭,係用以自該三維操作空間中,取得一第二二維平面資訊;以及一運算單元,係用以利用該第一二維平面資訊與該第二二維平 面資訊,辨識該操作物的位置;其中,該運算單元係執行一處理機制來辨識該操作物的位置,該處理機制包括下列步驟:步驟一:利用該第一二維平面資訊與該第二二維平面資訊建立一立體座標系;步驟二:事先建立該操作物之一方向梯度圖(HOG)模型,且將該方向梯度圖(HOG)模型當作過濾器,分別在該第一二維平面資訊與該第二二維平面資訊中辨識該操作物;步驟三:事先建立該操作物之一線形模型來描述該操作物的輪廓,進而分別在該第一二維平面資訊與該第二二維平面資訊中辨識該操作物;步驟四:以該操作物於該第一二維平面資訊與該第二二維平面資訊中的視差間距,定義該操作物的位置;步驟五:當一物件介入該第一二維平面資訊與該第二二維平面資訊,且偵測該物件之視差間距介於一上限值與一下限值間,執行該步驟二與該步驟三,以判定該物件是否為該操作物。 An optical touch system for identifying a position of an operation object in a three-dimensional operation space, comprising: a first photographic lens for acquiring a first two-dimensional plane information from the three-dimensional operation space; a second photographic lens for acquiring a second two-dimensional plane information from the three-dimensional operation space; and an operation unit for utilizing the first two-dimensional plane information and the second two-dimensional plane The surface information identifies the location of the operator; wherein the computing unit performs a processing mechanism to identify the location of the operator, the processing mechanism includes the following steps: Step 1: using the first two-dimensional plane information and the second The two-dimensional plane information establishes a three-dimensional coordinate system; step two: pre-establishing a direction gradient map (HOG) model of the operator, and using the direction gradient map (HOG) model as a filter, respectively in the first two-dimensional Identifying the operator in the plane information and the second two-dimensional plane information; Step three: pre-establishing a linear model of the operator to describe the contour of the operator, and then respectively in the first two-dimensional plane information and the second Identifying the operator in the two-dimensional plane information; Step 4: defining the position of the operator by the parallax distance between the first two-dimensional plane information and the second two-dimensional plane information; Step 5: When one The object intervenes the first two-dimensional plane information and the second two-dimensional plane information, and detects that the parallax distance of the object is between an upper limit value and a lower limit value, and performs step two and the third step to determine Determine if the object is the operator.
TW101150858A 2012-12-28 2012-12-28 Optical touch system TW201426466A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
TW101150858A TW201426466A (en) 2012-12-28 2012-12-28 Optical touch system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
TW101150858A TW201426466A (en) 2012-12-28 2012-12-28 Optical touch system

Publications (1)

Publication Number Publication Date
TW201426466A true TW201426466A (en) 2014-07-01

Family

ID=51725549

Family Applications (1)

Application Number Title Priority Date Filing Date
TW101150858A TW201426466A (en) 2012-12-28 2012-12-28 Optical touch system

Country Status (1)

Country Link
TW (1) TW201426466A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
TWI574196B (en) * 2016-01-21 2017-03-11 緯創資通股份有限公司 Optical touch apparatus, a method for determining a position of a touch indicator point and an optical touch system

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
TWI574196B (en) * 2016-01-21 2017-03-11 緯創資通股份有限公司 Optical touch apparatus, a method for determining a position of a touch indicator point and an optical touch system

Similar Documents

Publication Publication Date Title
JP6079832B2 (en) Human computer interaction system, hand-to-hand pointing point positioning method, and finger gesture determination method
US9519968B2 (en) Calibrating visual sensors using homography operators
CN103984928B (en) Finger gesture recognition methods based on depth image
CN110443154B (en) Three-dimensional coordinate positioning method and device of key point, electronic equipment and storage medium
Prisacariu et al. 3D hand tracking for human computer interaction
TW201120681A (en) Method and system for operating electric apparatus
CN104317391A (en) Stereoscopic vision-based three-dimensional palm posture recognition interactive method and system
CN110825234A (en) Projection type augmented reality tracking display method and system for industrial scene
TWI428804B (en) Optical screen touch system and method thereof
US10165168B2 (en) Model-based classification of ambiguous depth image data
WO2021036373A1 (en) Target tracking method and device, and computer readable storage medium
JP6487642B2 (en) A method of detecting a finger shape, a program thereof, a storage medium of the program, and a system for detecting a shape of a finger.
TW201322058A (en) Gesture recognition system and method
US20190266798A1 (en) Apparatus and method for performing real object detection and control using a virtual reality head mounted display system
CN103761011B (en) A kind of method of virtual touch screen, system and the equipment of calculating
CN112657176A (en) Binocular projection man-machine interaction method combined with portrait behavior information
Kerdvibulvech Hand tracking by extending distance transform and hand model in real-time
CN107797648A (en) Virtual touch system, image recognition positioning method and computer-readable storage medium
US9122346B2 (en) Methods for input-output calibration and image rendering
WO2021004413A1 (en) Handheld input device and blanking control method and apparatus for indication icon of handheld input device
US20230224576A1 (en) System for generating a three-dimensional scene of a physical environment
TW201426466A (en) Optical touch system
US20220050528A1 (en) Electronic device for simulating a mouse
CN113807191B (en) Non-invasive visual test script automatic recording method
KR20140046197A (en) An apparatus and method for providing gesture recognition and computer-readable medium having thereon program