TW460813B - Apparatus and method for context-based indexing and retrieval of image sequences - Google Patents

Apparatus and method for context-based indexing and retrieval of image sequences Download PDF

Info

Publication number
TW460813B
TW460813B TW89101505A TW89101505A TW460813B TW 460813 B TW460813 B TW 460813B TW 89101505 A TW89101505 A TW 89101505A TW 89101505 A TW89101505 A TW 89101505A TW 460813 B TW460813 B TW 460813B
Authority
TW
Taiwan
Prior art keywords
frame
objects
movement
main object
sub
Prior art date
Application number
TW89101505A
Other languages
Chinese (zh)
Inventor
Sriram Sethuraman
Edmond Chalom
Iraj Sodagar
Original Assignee
Sarnoff Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Sarnoff Corp filed Critical Sarnoff Corp
Application granted granted Critical
Publication of TW460813B publication Critical patent/TW460813B/en

Links

Abstract

An apparatus (150) and method for implementing object motion segmentation and object trajectory segmentation for an image sequence. Specficially, block-based motion vectors for a pair of adjacent frames are used to derive optical flow, e.g., affine, motion parameters. Such optical flow motion parameters are employed to determine key objects where their motion and trajectory within a sequence of frames are calculated and stored. Such object motion information is used to improve or offer image processing functions such as context-based indexing of the input image sequence by using motion-based information.

Description

460813 五'發明說明(1) 本申請案主張在1 999年1月28日所申請的美國 60/1 1 7, 629,其在僅列出供參考。 系孤 本發明係有關將例如影像的影像順序編索引與祿 像處理。更明確而言,本發明係關於強調將具移動^ = 影像順序以本文為基礎編索引與操取之有效率纟士構 、 發明背景 、Ό 。 隨著例如視聽内容的可用多媒體内容的增加, 長與複雜資訊組織與管理的需要變成很重要。 =士, 當f媒體内容的文庫持續成長時,它便變成不容易^ 此高度複雜的資訊編索引以便在猶後幫助有效 藉著將描述多媒體内容的描述符號最 ^取^ 資料庫出現的内容便可找出,藉此使搜尋盘 此-界面標準*,而此界=(索 =2際標準提供將 與過濾媒體所使用。_G標準的此新成員丨:丄丨擎、 内容描述界面,而且代號是"MpEG_7"珉員疋〒名為多媒體 例如,一影像順序的典型 ”數個影像拍攝"而獲得。"影内 ::藉著將順序分成 件之影像剪裁中的連續圖框,象而且攝曰,疋義為描述在一事 、擦拭、或消退的一突然景象變化^如一混合、溶解 之前與之後。影像拍攝邊界的偵測2 ^殊效果景象變化 像剪裁,如此便構成朝向内 許事件隨機存取一影 。只要偵測到一影像拍攝,稱為"、選擇瀏覽的第一步驟 擷取,以補捉事件的演變,如,圖框"的代表圖框便可 、 ’主圖框可識別以表示一 460813 五、發明說明(2) 增加景象、一 處理一影像順 一些主圖框。 的景色 以操取 當提 沒有與 (例如, 。即是 組織描 能由搜 捉當作 可支援 來自單 内容低 主圖框 供一簡 移動有 除了" ’具有 述符號 尋引擎 一描述 來自不 一透視 支援大範圍的 督的應用領域 .因此,在技 像順序中的移 文為基礎之索 本發明的一 動分段裝置與 以移動為基礎 本文為基礎索 動作追蹤場面景象、一浪漫景 序的許多影像圖框的複雜問題 在例如來自靜態影像的顏色、 準位抽像化中的既有知識主體 的後設資料。 單的解決以擷取後設資料時, 關的資訊。移動資訊能相當擴 名詞"之外,詢問能具有"動詞 基於與移動資訊互有關聯的顏 的已知資訊額外情況是有利的 所使用圖景動力的更聰明描述 符號之有關物件移動資訊,該 同透視圖的景色飛躍快速分析 圖的景色及只儲存對應的後設 料想不到詢問。例如,這在例 是非常重要的,其中它非始終 藝中存在的一需要是用以擷取 動資訊之一裝置及方法,藉此 引和操取及各種不同編瑪功能 具體實施例是用以實施一影像 物件軌道分段之—裝置方 的資訊而改良或提供例如輸入 引之其他影像處理功能。更明 象等。此簡化 ,而只要處理 形狀、與組織 然後可提供, 上面的 大有關 ”)的詢 色、形 ,以傳 。有利 描述符 ,以取 資料, 如 可預期 描述便 内容 問範圍 狀、與 送有關 的是補 號最後 代分析 藉此能 性與監 詢問= 及描述在 影 改良例如以本 之影像處理。 順序的物件移 ’藉此可使用 影像順序的以 確而言,以區460813 Five 'Invention Description (1) This application claims US 60/1 1 7, 629 filed on January 28, 999, which is listed here for reference only. The present invention relates to the sequential indexing of images such as images and the processing of images. To be more specific, the present invention is about emphasizing efficient structure, background of the invention, and indexing of indexing and manipulation of moving sequences based on this article. With the increase of available multimedia content such as audiovisual content, the need for long and complex information organization and management becomes important. =, When the library of media content continues to grow, it becomes difficult ^ This highly complex information is indexed in order to help effectively in later time by taking the most descriptive symbols describing multimedia content ^ The content that appears in the database You can find out, so that the search interface is this interface standard *, and this sector = (索 = 2 international standard provides will be used with filtering media. This new member of the _G standard 丨: 丄 丨 Engine, content description interface, Moreover, the code name is "MpEG_7", and the nickname is "Multimedia. For example, a typical sequence of an image" is obtained by "shooting several images". "In-picture :: Continuous images in image trimming by dividing the sequence into pieces Frame, image and photo, meaning is to describe a sudden scene change in one thing, wipe, or fade away ^ such as a mix, before and after dissolving. Detection of the boundary of image shooting 2 ^ special effect scene change is like clipping, so it constitutes A random access to the event towards the internal event. As long as an image capture is detected, it is called " the first step of choosing to browse to capture the evolution of the event, such as the representative frame of the frame " Yes, 'The main frame can be identified to indicate a 460813. V. Description of the invention (2) Add scenes, and process an image along some of the main frames. The scenes can be manipulated without mentioning (for example, that is, the organization can describe It is regarded as supportable from a single content low-main frame for a simple movement. In addition to "" with a symbol search engine, a description is provided from a variety of perspectives to support a wide range of applications. Therefore, in the image sequence The complex problem of a moving segmentation device based on the present invention and a motion-based text-based motion tracking scene scene, a romantic scene sequence of many image frames, such as the color and level from a still image Existing knowledge subject's meta-data in abstraction. Single solution to extract meta-data when meta-data is retrieved. Mobile information can be quite extended. In addition, queries can have verb-based and mobile information. Known additional information about interrelated colors is beneficial. Uses scene dynamics to more intelligently describe symbols about object movement information. The same perspective Scenery Leap Quickly analyze the scenery of the map and store only the corresponding post-design materials unexpectedly. For example, this is very important in the example, where a need that is not always present is a device and method for capturing mobile information. In this way, specific embodiments of indexing and manipulating and various editing functions are used to implement the segmentation of an image object track—the information on the device side to improve or provide other image processing functions such as input indexing, and so on. This simplifies, but as long as the shape, organization and then can provide, the above big inquiry ", the shape, to pass. Favorable descriptors to obtain information, if the description can be expected, the content asks the scope, and is related to sending It is the last-generation analysis of the supplement number to take this capability and monitor the inquiry = and describe the improvement in the film such as the image processing of the book. Sequential Object Shift ’which allows you to use the image sequence

ί 460813 i、發明說明(3) ---. 塊為基礎之移動向量是用來取得例如仿射移動參數的 流移動參數。 干 明確而言,光學流(例如,仿射)物件移動分段的最初 行是用於一對毗連圖框。仿射移動參數然後可用來決定 識別在每個圖框内的主要物件。這些主要物件然後可在^ 像順序(亦已知為具有許多輸入影像順序圖框的一"影像^ 攝”)的某些區間上受到監督,而且他們的移動資訊;^在 些區間時間上擷取及追蹤。 5 然後,光學流(例如,仿射)軌道分段是在影像順序上執 行。明確而言,每個毗連對圖框的每個識別的主要物件 產生的仿射移動參數會在影像順序的一區間上處理,以 物件軌道分段生效。即是,例如方向、速度、與加速的移 ,軌道可在某些圖框區間上推測每個主要物件,藉此提供 月b由δ旬問所利用的移動資訊.之另一觀點。 、 圖式之簡蕈說明 本發明的描述可藉由考慮下列的詳細描述及附圖而 ,其中: 肝 圖1描述本發明的一本文伺服器方塊圖; 圖2描述本發明的以本文為基礎之索引程式方塊圖; 圖3描述藉由使用以移動為基礎之資訊而實施一 像順序的以本文為基礎之索引的方法流圖; , 圖4描述用以實施光學流(例如,仿射)物件移動 一方法流圖; 圖5描述用以實施光學流(例如,仿射)物件移動分段的ί 460813 i. Description of the invention (3) ---. Block-based motion vectors are used to obtain, for example, affine motion parameters for flow motion parameters. To be clear, the first rows of optical flow (for example, affine) object moving segments are for a pair of contiguous frames. The affine movement parameters can then be used to determine the main objects identified in each frame. These main objects can then be supervised over certain intervals in the sequence of images (also known as "images" with many input image sequence frames), and their movement information; at intervals of time Acquisition and tracking. 5 Then, optical flow (eg, affine) orbit segmentation is performed on the image sequence. Specifically, the affine movement parameters generated by each adjoining pair of main objects identified by the frame will be It is processed in a section of the image sequence, and it takes effect in segments of object orbits. That is, for example, the movement of direction, velocity, and acceleration, the orbit can infer each major object in some frame intervals, thereby providing the month b δ Xunwen uses another aspect of mobile information. Schematic description of the invention The description of the present invention can be considered by considering the following detailed description and drawings, of which: Figure 1 depicts a text server of the present invention. Block diagram; Figure 2 depicts a text-based indexing program block diagram of the present invention; Figure 3 depicts a text-based indexing method that implements an image sequence by using mobile-based information Flow diagram; Figure 4 depicts a method flow diagram to implement optical flow (eg, affine) object movement; Figure 5 depicts a method to implement optical flow (eg, affine) object movement segmentation

460813 五、發明說明(4) 一更詳細方法流圖;及 圖6描述具有複數連接區域之一 為了要幫助了解,使用相框。 數字是在圖式中表示相同的::參考數字,其中這些參考 圖1描述本發明的一 施例中,本文飼服器100是藉由 方塊圖。在一具體實 。因此,描述的本文伺服器-般目的電腦而實施 器)140、例如隨機存取記憶體一處理器(中央處理 士七也甘a ^仔取忑隱體(RAM)的一記憶體120、一以 本文為基礎之索引程式150、一選擇性編碼器⑴ 不同的輸入/輸出裝置130、(例如 各一種 聲音記錄器、-照相機、-手提攝影機、一影像監督器一_、 任何數目的影像裝置、或儲存裝置,其包括但未限制於一 磁帶機、一軟式磁碟機、一硬式磁碟機、或一雷射光碟 機)。 應了解到編碼器110與以本文為基礎之索引程式15 〇能共 同或個別實施。即是’編碼器11 0與以本文為基礎之索引八 程式15〇可以疋藉著一通彳§通道而耦.合至中央處理器1.的 實際裝置。或者,編碼器110與以本文為基礎之索引程式 150能由一或多個軟體應用(或甚至例如藉由使用特殊應用 積體電路(ASIC)的軟體與硬體的組合)表示,其令軟體是 從一儲存媒體载入,(例如,一磁性或光碟機或磁碟月), 並立f由在電滕記憶體120的中央處理器所操作。同樣地 ,編碼器110與本發明的以本文為基礎之索引程式15〇(包460813 V. Description of the invention (4) A more detailed method flow diagram; and FIG. 6 depicts one of a plurality of connected regions. To help understand, use a photo frame. The numbers are the same in the drawings:: Reference numerals, where these references are shown in Figure 1. In one embodiment of the present invention, the feeder 100 herein is illustrated by means of a block diagram. On a concrete level. Therefore, the server described in this article is a general purpose computer and implemented) 140, such as a random access memory-processor (Central Processing Unit Qi Yegan a) a memory 120, a RAM, Indexing program 150 based on this article, a selective encoder ⑴ different input / output devices 130, (e.g. each kind of sound recorder,-camera,-hand-held camera, a video monitor-any number of video devices Or storage devices, including but not limited to a tape drive, a floppy disk drive, a hard disk drive, or a laser drive.) It should be understood that the encoder 110 and the indexing program 15 based on this article 〇 Can be implemented jointly or individually. That is, the "encoder 110" and the index-based program 15 based on this article can be coupled to the actual device of the central processing unit 1 through a communication channel. Or, encoding The device 110 and the indexing program 150 based on this text can be represented by one or more software applications (or even, for example, by using a combination of software and hardware of a special application integrated circuit (ASIC)), which makes the software from a The storage medium is loaded (for example, a magnetic or optical disk drive or a magnetic disk drive), and the parallelism is operated by the central processing unit in the electric memory 120. Similarly, the encoder 110 and the present invention are based on this document. Index program 15〇 (package

括結合的資料結構)可儲存在一電腦可讀媒體上,例如 機存取記憶體、磁性或光碟機、或磁碟等。 在實施方面,各種不同的多媒體資訊是在路徑105上 收’而且儲存在本文伺服器1〇〇的一儲存裝置13〇内。多 體資訊可包括但未局制於例如完整電影、電影剪、 拍攝、廣告剪裁、音樂影像等的各種不同影像順序= J序可或不能包括聲頻流或資料流,例如完成的加上^幕 由於可用多媒體内容的增加及他們較大的大小,在路 丄05上的輸入資訊可經由插述為一或多値選擇性編碼器ιι〇 = 該等編碼器U〇包含例如類似MPEG編碼器 的景J象與耷s編碼器,而這些類似的肝跖編碼器在叶 <減少”與暫時的冗贅 '然而,任何數目的壓縮方法皆 =可明未局限於任何特殊的方法。既然輸入 的資ν 了ι已在本文伺服器的外部進行各種不同的壓縮 是選擇性,其中輸入的流已是在屋縮 格式。在此實施方面’編碼器U〇可被省略。 ,:=ί=Πΐί150是採用來分析輸八的資訊 的大量而時常複雜的多媒體;;在:ίί:上i3°所儲存 礎之索!丨程式丨5G是要提供允本文為基 引方法及結合的資料结構,/二·、冑效率方法之索 在路徑195上快速地操取。更明後確允雜的/媒體内容能 礙之索引程式15。採用移動匕確,目前以本文為基 勒貢訊’以允許採用”動詞”(例如.(Including combined data structures) can be stored on a computer-readable medium, such as a machine access memory, magnetic or optical drive, or magnetic disk. In terms of implementation, various multimedia information is received on the path 105 'and stored in a storage device 13 of the server 100 of this document. Multi-body information may include, but is not limited to, a variety of different video sequences such as complete movies, film cuts, shootings, advertising clips, music videos, etc. = sequence may or may not include audio or data streams, such as completed plus ^ screen Due to the increase in available multimedia content and their larger size, the input information on Route 05 can be interpolated as one or more selective encoders. Ι = These encoders U0 contain, for example, similar MPEG encoders. Scene J and 耷 s encoders, and these similar liver 跖 encoders in the leaf < reduce " and temporary redundancy '; however, any number of compression methods are clear and not limited to any special method. Since the input The various types of compression that have been performed outside the server of this document are optional, where the input stream is already in a contracted format. In this implementation, the 'encoder U0 can be omitted., == ί = Πΐί150 is a large and often complex multimedia used to analyze the information lost in the eighth; in: ίί: i3 ° stored basis of the code! 丨 program 丨 5G is to provide a method that allows this article to be used as a reference and a combined data structure. // 2 ·, The method of efficiency method is quickly accessed on path 195. It is more clear that the miscellaneous / media content can be hindered by the index program 15. It is confirmed by the use of mobile knives, and this article is currently based on Kirigon ' To allow the use of "verbs" (eg.

4 6〇813 -:----- 五、發明說明(6) ^件的有關移動資訊)的更複雜詢問,以取代只 岡(例如,物件的顏色)。 r ^ t ’包含例如藍色天空的一藍色背景之影像順序,門 二如:▲多明f的詢問,藉此減少詢問功能的效力。對』 ,其中-物社县以ί 景景“象順序而修改 回應:可= 速而在前景移至左•,那麼詢問的 速移動飛機的一影像m序橫過藍色天空的背景。決 及丨程式15°包含一物件移動分段器16〇 θ物件軌道刀段器170。簡單地說,物件移動分段哭1Rn 疋採用來廣泛地決定在每個圖框内的物件動段器】6° 物件軌道分段則是採用來寬泛地決定在相二移二= 的許多圖框中的物件㈣。 *景/像順序陴 =2描述包含物件移動分段器16〇與物件 7 :^之一以本文為基礎的索引程式15〇方二:: 主要物件分離器216、與一光學流(例如,仿 器:18。物件軌道分段器17〇包含一主要物件軌道分段刀器又 220及一子物件執道分段器222。由這些模組赏 功能是在圖2簡短描述.這些功能的詳細描述是仃的寬廣 圖3-6的流圖及其他圖式而提供。 在實施方面,一影像順序會在以區塊為基礎之 器210内接收,其中例如以區塊為基礎之移動向量移的移動4 6〇813-: ----- V. Description of the invention (6) More complicated inquiry about mobile information) to replace only the gang (for example, the color of the object). r ^ t ′ contains, for example, the image sequence of a blue background with a blue sky. The second door is as follows: ▲ Doming's inquiry, thereby reducing the effectiveness of the inquiry function. Right ", where-Wushe County modified the response in the order of" view ": can = speed and move to the left in the foreground, then an image m of the speed-moving plane inquired crosses the blue sky background. And the program 15 ° includes an object moving segmenter 16〇θ object track cutter 170. In short, the object moving segmentation cry 1Rn 疋 is used to widely determine the object moving segmenter in each frame] The 6 ° object track segmentation is used to broadly determine the objects 许多 in the many frames of phase two shift two =. * Scene / image order 陴 = 2 description includes object moving segmenter 16 and object 7: ^ One is an indexing program based on this article. Fifteen: The main object separator 216, and an optical stream (for example, a simulator: 18. The object track segmenter 17 includes a main object track segmentation knife. 220 and a sub-object implement the segmenter 222. The functions rewarded by these modules are briefly described in Figure 2. A detailed description of these functions is provided in the broad flow chart and other diagrams of Figure 3-6. On the one hand, an image sequence is received in the block-based device 210, of which As to the block based motion vector as the movement of the shift

460813460813

五、發明說明(7) 資訊是從每個圖框的影像順序計算。然而,如果本文祠服 器100具有一外部編碼器110 ,或輸入影像順序已包含移動 資訊,亦即,其中移動向量是以影像順序編碼,那麼以區 塊為基礎之移動估計器21 〇便可省略。即是,以區塊為某 礎之移動資訊便可從壓縮的位元流本身擷取或由本文''飼"·服 器100的其他模組所提供,藉此減少必須計算移動向量 物件移動分段器1 6 0。 依次’光學流(例如’仿射)分段器212可提供移動向量 資訊,以產生"仿射移動參數"。雖然本發明是在下面藉由 使用仿射移動模型來描述,但是應了解到其他的光學流模 型亦能採用❶仿射移動模型是由j. Niewegl〇wski等在& 1993 年8 月的"A Novel Video Coding Scheme Based 〇n5. Description of the invention (7) Information is calculated from the order of the images of each frame. However, if the server 100 in this paper has an external encoder 110, or the input image sequence already contains motion information, that is, where the motion vector is encoded in the image sequence, then the block-based motion estimator 21 can be used. Omitted. That is, block-based movement information can be retrieved from the compressed bit stream itself or provided by other modules in this article "server" 100, thereby reducing the need to calculate motion vector objects Move the segmenter 1 6 0. In turn, an 'optical flow (e.g.,' affine ') segmenter 212 can provide motion vector information to generate " affine motion parameters ". Although the present invention is described below by using an affine motion model, it should be understood that other optical flow models can also be used. The affine motion model was developed by j. Nieweglówski et al. In & August 1993 " A Novel Video Coding Scheme Based 〇n

Temporal Prediction Using Digital Image Warping", IEEE Trans.消費者電子,第39,3冊的第141_15〇頁揭露 ",其在此此列出供參考。仿射移動模型藉由運用已知為 "影像扭曲"的一幾何轉換而從先前的影像構成一預測影像 或圖框。變換指定在先前與預測影像每點之間的一空間關 係。 母 通常’使用區塊符合的移動補償可提供轉換移動的—良 好整體效率。然而,當移動包含迴轉或比例化元件(例如 ’縮放或影像一旋轉)時,區塊符合移動估計是一不良 執行器。 對照下,仿射移動模型(仿射轉換)是由6個參數至8 所定義,並且表示如下: 6Temporal Prediction Using Digital Image Warping ", IEEE Trans. Consumer Electronics, Vol. 39, Vol. 3, page 141-15, Rev. ", which is listed here for reference. The affine motion model constructs a predicted image or frame from a previous image by applying a geometric transformation known as " Image Distortion ". The transformation specifies a spatial relationship between the previous and each point of the predicted image. The parent's use of block-matched motion compensation usually provides the ability to switch movements—good overall efficiency. However, when the motion contains a rotation or scaling element (such as' zoom or image-rotation), the block conforming to the motion estimation is a bad actuator. In contrast, the affine movement model (affine transformation) is defined by 6 parameters to 8, and is expressed as follows: 6

4 608 13 五、發明說明(8) Ί a4 〇 I a5 〇 I (1) a6 0 I j 广 la! [x, y, 1] = [u, v, 1] I a2 - 1 a34 608 13 V. Description of the invention (8) Ί a4 〇 I a5 〇 I (1) a6 0 I j Guang la! [X, y, 1] = [u, v, 1] I a2-1 a3

L 其中(x ’y)是在先前圖框中的圖素坐標,而且(u,v) 是在預測圖框中的一特定圖素之坐標。在決定6個參數的 詳細討論是在J. Nieweglowski等參考中出現。仿射關係 的特徵是由6個參數表示。因此,仿射移動模型在例如轉 換、比例化、與旋轉的預測移動能更有效,而這些預測移 動不僅能以自然的順序時常觀察,而且能藉由使用數位效 果而在綜合性景色觀察 射分段器 仿射參數 動的追蹤 夠重要的 而識別, 要物件。 在背景的 要物件的 義。只要 主要物件 果每個主 的影像處 即是,仿 "主要物件" 視為他們移 是重要的足 的大小部份 物件不是主 物件,_然而 而,合格主 使用者所定 動然後可由 或者,如 目的或其他 21 2的工作是影像順序的每個圖框之 的識別、分段、與產生。主要物件可 對於編索引或其他影像處理功能目的 物件。典型上,主要物件是基於他們 亦即大物件典型是主要物件,然而小 因此,一移動交通工具典型是一主要 小移動昆蟲不是一主要物件。然 需求疋特殊應用’而且是由本發明的 主要物件被定義,這些主要物件的移 追蹤器214所追蹤。 要物件元件的移動資訊對於編戶斤引的 理功能也是很重要,額外的處理是 五、發明說明(9) 由主要物件分離器216所執行。明確而言,一主要物件可 为成子物件,而且這些子物件的移動資訊可個別追蹤。例 如’一人類的主要物件可分成6個子物件,其包含頭、身 體、與4肢。因此’一詢問現可以是技術,以搜尋與在一 主要物件内的子物件有關的"動作",例如搜尋一影像順 序’其中人的4肢會抬高超過人頭等。 雖然些主要物件能很快分成已定義的子物件,但是其 他的主要物件需要進一步處理,以識別子物件的邊界。因 此,主要物件資訊便可直接從主要物件追蹤器214傳送給 用以每個主要物件的"子物件"的識別與分段之一仿射分段 器218。因此,仿射分段器218亦能與產生子物件的仿射移 作。應注意到雖然以本文為基礎之索引程式描述 f施2器212和218,但是應了解到單一仿射分段器能 以執行仿射處理(亦即,主要物件與子物件 兩位罕” ^ 主ii件2物件追蹤器214的移動資訊會傳送給-主要物1軌道分段器220。雖然它能維持及追蹤移 ,例如每個主要物件的仿射移動參數是存 此移動資訊需要-實質儲存需求。 ::資:會傳送給主要物件軌道分段器 要物的 道資訊與區間(圖框區間)可為每個主要物件產其中移气軌 移動資訊可概括地描豸成"纟要物件軌道資訊;宜 些定義區間(在許多圖框上)方向、速度 ^某 以多媒體内容的有效率以移動為速等^此允許 土变或其他影像L where (x'y) is the pixel coordinate in the previous frame, and (u, v) is the coordinate of a specific pixel in the prediction frame. A detailed discussion of determining the six parameters appears in the reference by J. Nieweglowski et al. The feature of the affine relationship is represented by 6 parameters. Therefore, affine movement models can be more effective in predictive movements such as transformation, scaling, and rotation, and these predictive movements can be observed not only from time to time in a natural order, but also by using digital effects to observe the shooting score in a comprehensive scene. The trajectory tracking of the segmenter affine parameters is important enough to identify the object. The meaning of the object in the background. As long as the main object is located at the image of each master, imitating "main objects" considers that their movement is important. The large and small objects are not the main objects. However, the movements determined by a qualified master user can then be determined by or For example, the purpose or other work of 21 2 is the identification, segmentation, and generation of each frame of the image sequence. Primary objects can be objects for indexing or other image processing functions. Typically, the main objects are based on them. That is, large objects are typically the main objects, but small. Therefore, a mobile vehicle is typically a major small moving insect and not a major object. Of course, the requirements "special applications" are also defined by the main objects of the present invention, which are tracked by the movement tracker 214. The movement information of the required object components is also very important for the editing function of the household index. The additional processing is 5. Explanation of the invention (9) is performed by the main object separator 216. Specifically, a main object can be a child object, and the movement information of these child objects can be tracked individually. For example, a main object of a human being can be divided into 6 sub-objects, which include a head, a body, and four limbs. So 'a query can now be a technique to search for " actions " related to sub-objects within a main object, such as searching for an image sequence, where a person's four limbs will rise above the head and so on. Although these main objects can be quickly divided into defined sub-objects, other main objects need further processing to identify the boundaries of the sub-objects. Therefore, the main object information can be directly transmitted from the main object tracker 214 to one of the " sub-objects " identification and segmentation of each main object, the affine segmenter 218. Therefore, the affine segmenter 218 can also interact with the affine generating sub-objects. It should be noted that although indexers based on this document describe the applicators 212 and 218, it should be understood that a single affine segmenter can perform affine processing (ie, the main object and the sub-object are rare). ^ The movement information of the main object 2 object tracker 214 will be transmitted to the main object 1 orbit segmenter 220. Although it can maintain and track the movement, for example, the affine movement parameters of each main object are required to store this movement information-substance Storage requirements. :: Information: The track information and interval (graph frame interval) that will be sent to the main object track segmenter can be produced for each main object. The gas track movement information can be summarized as " Information on object orbits; it is better to define the direction and speed of the interval (on many frames) ^ Some are based on the efficiency of multimedia content and move as fast as possible ^ This allows soil deformation or other images

1M 第13頁 4 60813 五、發明說明(ίο) 處理)的一隔式補捉及儲存。 訊會傳送給子物件軌道分子段器件的移動資 區間(圖框區間)可為每個子物件產生。㈣軌道貝訊與 圖3描述用以實施仿射分段之一方法流 以移動為基礎的資精此使用 以本文為基礎之索;丨的其2 =例如一輸入影像順序的 方綱是在步驟30 S T 功能。更明確而言, 物件移動分段進行步驟310 ’其中仿射 ! 的許多圖框之',影像拍攝")的某 :取』::二:且f們的移動資訊可在這些區間時間被 主要物件產生在步驟310,仿射移動參數可為每個識別的 的t ^ ^320 ’每础連對圖框的每個識別主要物件所產生 ^射^動參數是在影像順序區間時間上處理,以使物件 在芊:^ 5二:是’例如方向速度與加速的移動軌道可 岣時間上推測每個主要物件,藉此提供可由 驟3°25結1束用的移動資訊的另一觀點。方法300然後會在步 ^ 4描述用以實施仿射物件移動分段的-方法3 10流圖。 I7疋、,方法310是圖3的步驟3 1()的更詳細描述。 、方法31〇疋在步驟4〇5開始,並且進行步驟a?,其中方 ^ a 了從以區塊為基礎之移動資訊產生仿射移動參數。 ,一隨機數目的區塊可被選取,其中他們的以區塊為 土之移動向量可採用,以取得如下所討論的仿射移動參1M P. 13 4 60813 V. A separate catch and storage of the invention description (ίο) processing). The mobile data interval (frame interval) transmitted to the sub-object orbital molecular segment device can be generated for each sub-object. ㈣Orbital Besson and Figure 3 describe one method used to implement affine segmentation. Flow-based assets are used in this paper. The 2 = e.g., The outline of an input image sequence is in Step 30 ST function. More specifically, the object movement segmentation is performed in step 310 'of the many frames of affine!', And one of the image capture ") :: Take :: 2: The movement information of f can be changed in these intervals. The main object is generated in step 310. The affine movement parameter can be generated for each identified t ^ ^ 320 'for each identified main object of the frame for each identified ^ shot ^ motion parameter is processed in the image sequence interval time. In order to make the object in 芊: ^ 52: Yes, for example, the direction of speed and acceleration of the moving orbit can infer each major object in time, thereby providing another view of the movement information that can be used in 3 ° 25 knot . The method 300 will then be described in step ^ 4 to implement the affine object movement segmentation-method 3 to 10 flow diagram. I7. Method 310 is a more detailed description of step 31 () of FIG. 3. Method 31〇 疋 starts at step 405, and proceeds to step a ?, where ^ a generates an affine movement parameter from the block-based movement information. A random number of blocks can be selected, and their motion vectors with blocks as the soil can be used to obtain the affine mobile parameters discussed below.

460813 五、發明說明(11) 數。 在步驟410,方法310嘗試從一先前圖框來追縱一或 識別主要,件’亦即,從一先前的圖框獲得一主要物 標籤。即是’只要主要物件已識別如在步驟42〇所討論 一對她連圖框,它#雪承的 便需要重新在影像順序中提供下一圖柩 的相同偵測步驟。即是,斟於 &向』 圃框 P 對於一新圖框而言,以區塊為基 =的移動向,可㈣來很快決定該等區塊是否指向 標示的主要物件。如罢如… , 无刖 相η沾护滋加 果如此,此區塊將保留與先前的圖框 相同的標籤。例如,如果四(4) 連圖框,而且如果下一圖框η要:件已決定-對晚 /用來決定在目前圖框的5個連接區域的Ϊ 令4個是否與4個識別的主要 逆按匕埤的其 廢口右膝忽沾留、* 1 I物件互有關聯。如果如此,那 、、 、接區域是根據步驟420而測試,以決 J它是否為-主要物件。追蹤方法可明顯減少的二決 亦即主要物件是在影像順序中 的負何, 行,例如物件已變成太小岑耔 牛不再追蹤而 件可識別,步驟41。便會跳越,s此情況是方法 一新的影像拍攝。 次」1 ϋ會處理 在步驟4 2 0,方法31 〇 ·?τ钱: 圖框内的主要物件。在一:像順序的-對础連 決定-區塊是否在仿射物件流内而識別。要物件了藉由 在步驟430,方法310可選擇性合併識 的,要物件可能太小而不能明顯用於;Π』 的,亦即,在-特殊應用的索引,例如單獨的:1 = 460813 五、發明說明(12) 圖框内與其他的鳥合併,以形成一群鳥的主要物件。 在步驟44 0,方法310可能選擇性識別在主要物件内的子 物件。在與步驟4 3 0對照下,識別的主要物件結合具有用 以在一特殊應用編索引之元件(子物件)的重要移動資訊。 例如,由人類所構成的一識別的主要物件包含亦即人四肢 的子物件,其中子物件的有關移動資訊對於將影像拍攝編 索引是很重要。 .在步驟4 5 0,方法3 1 0會詢問額外的圖框是否結合目前的 π影像拍攝"。如果詢問是否定地回答,那麼方法3 1 0便會 在步驟4 5 5結束。如果詢問是肯定地回答,那麼方法3 1 0便 進行步驟407,其中仿射移動參數會在下一對連續圖框產 生。即是,仿射分段必須在連續對圖框之間執行。理由是 仿射移動參數在模型化執道的每個瞬間上是需要的,1且 亦處理新的物件/閉塞。當所有的圖框已處理時,方法3 1 0 然後會在步驟4 5 5結束。 圖5描述用以實施仿射物件移動分段的一方法5 0 0流圖。 更明確而言,方法5 0 0是圖4的步驟4 1 0和4 2 0的一更詳細描 述。 方法5 0 0是在步驟5 0 5開始,並且進行步驟5 1 0,其中主 要物件將會決定。應注意到各種不同方法的存在是用以決 定在一圖框内的區域移動資訊,例如,用以計算每個圖素 移動的光學流技術,其中這些技術通常是計算複雜。此外 ,如果順序已編碼(例如,當作u 1 Μ P E G - 2位元流),光學 流計算將需要位元流的完全解碼。因此,本本明中,仿射460813 V. Description of invention (11). At step 410, the method 310 attempts to trace or identify the main item from a previous frame, that is, to obtain a main object label from a previous frame. That is, as long as the main object has been identified as a pair of her connected frames as discussed in step 42, it will need to provide the same detection steps for the next image in the image sequence again. That is to say, considering & direction ", for a new frame, the moving direction based on the block = can quickly determine whether these blocks point to the main objects marked. If you want to ..., there is no way to protect it. If so, this block will retain the same label as the previous frame. For example, if there are four (4) consecutive frames, and if the next frame n is required: the item has been decided-to be late / is used to determine whether the 4 commands in the 5 connected areas of the current frame are identified with 4 Mainly against the left knee of the dagger, the * 1 I objects are related to each other. If so, the,,, and access fields are tested according to step 420 to determine if it is a -main object. The tracking method can significantly reduce the second decision, that is, the main object is the negative effect of the image sequence, for example, the object has become too small, the census is no longer tracked, and the item can be identified, step 41. It will skip, s this situation is the method of a new image capture. The time "1" will be processed in step 4 2 0, method 31 ○ · τ money: the main object in the frame. In one: the order-like basis determines whether or not the block is identified within the affine object stream. If the object is identified by the method 310 in step 430, the object may be too small to be used obviously; Π ′, that is, the index of the special application, for example: 1 = 460813 V. Description of the invention (12) The frame is merged with other birds to form the main object of a group of birds. At step 440, the method 310 may selectively identify child objects within the main object. In contrast to step 430, the identified main object combines important movement information with components (sub-objects) used to index in a particular application. For example, a main object identified by a human being includes sub-objects, that is, human limbs, and the movement information of the sub-objects is important for indexing the image capture. In step 4 50, method 3 1 0 will ask whether the extra frame is combined with the current π image capture ". If the query is answered negatively, method 3 1 0 will end at step 4 5 5. If the query is answered affirmatively, then method 3 10 proceeds to step 407, where the affine movement parameters are generated in the next pair of consecutive frames. That is, affine segmentation must be performed between successive pairs of frames. The reason is that the affine movement parameters are needed at every instant of modeled execution, 1 and also deal with new objects / occlusions. When all the frames have been processed, method 3 1 0 then ends at steps 4 5 5. FIG. 5 depicts a 500 flow diagram of a method for implementing affine object moving segmentation. More specifically, method 5 0 0 is a more detailed description of steps 4 1 0 and 4 2 0 of FIG. 4. Method 5 0 0 starts at step 5 0 5 and proceeds to step 5 1 0 where the main object will be determined. It should be noted that various methods exist to determine the area movement information within a frame, for example, optical flow techniques used to calculate each pixel movement, where these techniques are usually computationally complex. In addition, if the order is already coded (for example, as u 1 MP PEG-2 bit stream), the optical stream calculation will require full decoding of the bit stream. Therefore, in this book, affine

O:\62\62535.ptd 第16頁 4 6 0 813 五、發明說明(13) 移動參數可在壓縮領域藉由使用符合絕對差(在硬體或軟 體平台上)加總的簡單區塊來計算,或從P圖框的移動向量 獲得。 在一具體實施例中,仿射移動參數可藉由使用與一非常 複雜的徹底搜尋相比較能提供迅速方法識別潛在主要物件 候選之隨機取樣一致(RANSAC)方法而獲得。該方法使用在 亦即影像拍攝的一影像順序或部分的每兩個連續圖框之間 的仿射流(換句話說,一仿射移動分段)而計算區域。即是 ,在步驟5 1 0,方法5 0 0會先嘗試找出具一圖框的所有''連 接區域"。任何數目的方法可採用來決定例如一巨區塊的 一區塊圖素是否為一"連接區域"的部份。描述區塊連接性 的一範例已在_____所申請的美國專案號"SAR 1 3 0 7 6 ”名稱 "Method And Apparatus For Generic Shape Coding"— 中 階露,其在此僅列出供參考。例如,在一形狀邊界内的巨 區塊能以一連接的區域看出。 在步驟515,3個隨機巨區塊是從這組連接的巨區塊選取 ,亦即從每個連接的區域内。圖6描述具有複數連接區域 6 1 0 -6 4 0之一圖框。例如,連接區域6 1 0和6 3 0分別為背景 與前景,然而連接區域6 2 0和6 4 0是分別為一帆船及一旗幟 。為了要描述,在步驟5 1 5,由一些X所描述的3個區塊是 連接區域610的任意選取,而由0所描述的3個區塊是連接 區域620的任意選取等。 在步驟5 2 0,方法5 0 0能計算3個任意選取區塊的仿射移 動參數。明確而言,這些3個選擇區塊的以區塊為基礎的O: \ 62 \ 62535.ptd Page 16 4 6 0 813 V. Description of the invention (13) The movement parameters can be used in the compression field by using simple blocks that are summed up in absolute difference (on hardware or software platforms) Calculated, or obtained from the motion vector of the P frame. In a specific embodiment, the affine movement parameters can be obtained by using a Random Sampling Consistency (RANSAC) method that provides a rapid method to identify potential primary object candidates compared to a very complex thorough search. This method uses the affine stream (in other words, an affine moving segment) between every two consecutive frames of an image sequence or part of the image capture to calculate the area. That is, in step 5 1 0, the method 5 0 0 will first try to find all the "connection areas" with a frame. Any number of methods can be used to determine, for example, whether a block pixel of a giant block is part of a " connection area ". An example describing the connectivity of a block has been applied for the US project number " SAR 1 3 0 7 6 '' Name " Method And Apparatus For Generic Shape Coding " for _____ — Zhonglulu, which is only listed here For reference. For example, a giant block within a shape boundary can be seen as a connected area. In step 515, three random giant blocks are selected from this set of connected giant blocks, that is, from each connection Figure 6 depicts a frame with a plurality of connected regions 6 1 0-6 4 0. For example, the connected regions 6 1 0 and 6 3 0 are the background and foreground respectively, while the connected regions 6 2 0 and 6 4 0 Are a sailing boat and a flag respectively. In order to describe, in step 5 1 5, the three blocks described by some X are arbitrarily selected from the connection area 610, and the three blocks described by 0 are the connection area. Arbitrary selection of 620, etc. At step 5 2 0, method 5 0 0 can calculate the affine movement parameters of 3 randomly selected blocks. Specifically, these 3 selected blocks are block-based

O:\62\62535.ptd 第17頁 460813 五、發明說明(14) 移動向量足以推測6個仿射移動參數。既然3個區塊是產生 6個仿射移動參數所需的最低限度數目的區塊,所以應了 鮮到從連接區域超過3個區塊的移動向量亦能使用。應注 意到本假設在於仿射移動參數是從3個巨區塊計算。在由3 個區塊所形成的三角形頂點的特定移動,6個仿射移動參 數可被計算。即是,該等仿射移動參數在由三個巨區塊所 形成的三角形片段内每點上提供換置。 在步驟5 2 5,方法5 0 0可在整個圖框内的每個區塊移動的 逐一區塊基礎上詢問一區塊的移動是否在由仿射移動參數 所定義的仿射流内。換句話說,將仿射移動參數提供給在 圖框的每個巨區塊,並且計算在此位置與由區塊的移動向 量所提供位置之間的距離。如果距離小於一臨界 n e T h r e s hπ ,那麼巨區塊便會加入區塊的清單,當作連接 區域的部份,亦即,巨區塊貼標示為主要物件的部份。因 此,如果詢問是肯定回答,那麼區塊會在步驟5 3 5認為是 落在一仿射流物件内。如果詢問是否定回答,那麼區塊會 在步驟5 3 0認為落在一仿射流物件的外面。即是,一主要 物件可能由超過臨界的全部巨區塊所識別或標示。 在步驟54 0,方法5 0 0詢問包括的區塊是否大於許多區塊 "Reg MBs ize"(例如,10個巨區塊)的一臨界。此臨界"Reg Μ B s i z e "是例如決定在輸入圖框的大小與解析度的特殊應 用》如果詢問是肯定回答,那麼方法5 0 0便進行步驟5 5 0, 其中選定或包括的區塊會認為是一潛在的"主要物件"。如 果詢問是否定回答,那麼選定或包括的區塊便認為太小而O: \ 62 \ 62535.ptd Page 17 460813 V. Description of the invention (14) The motion vector is sufficient to infer 6 affine motion parameters. Since 3 blocks are the minimum number of blocks required to generate 6 affine motion parameters, it should be possible to use motion vectors that exceed 3 blocks from the connection area. It should be noted that this assumption is that the affine movement parameters are calculated from 3 giant blocks. At a particular movement of a triangle vertex formed by 3 blocks, 6 affine movement parameters can be calculated. That is, the affine movement parameters provide a transposition at each point within a triangle segment formed by three giant blocks. In step 5 25, the method 5 0 0 can ask whether the movement of a block is within the affine stream defined by the affine movement parameter on a block-by-block basis for each block movement in the entire frame. In other words, an affine movement parameter is provided to each giant block in the frame, and the distance between this position and the position provided by the block's movement vector is calculated. If the distance is less than a critical n e T h r e s hπ, then the giant block is added to the list of blocks as part of the connection area, that is, the part of the giant block labeled as the main object. Therefore, if the question is answered affirmatively, then the block is considered to fall in an affine stream object at step 5 35. If the query is answered in the negative, the block is considered to fall outside an affine stream object at step 530. That is, a major object may be identified or marked by all macroblocks exceeding the threshold. At step 540, the method 50 asks whether the included block is larger than a threshold of many blocks " Reg MBs ize " (e.g., 10 giant blocks). This critical " Reg Μ B size " is, for example, a special application for determining the size and resolution of the input frame. If the query is affirmative, then method 5 0 0 proceeds to step 5 5 0, where the selected or included area A block would be considered a potential "main item". If the question is answered in the negative, the selected or included block is considered too small and

O:\62\62535.ptd 第18頁 4 60813 五、發明說明(15) 無法追蹤。即是’如果具相同仿射參數的巨區塊總數目大 於某預设數目RegMBsize,那麼落在這些參數内的所有巨 ,塊可提供一通常的標籤,例如主要物件1等。具相同標 籤的所有巨區塊構成—新的移動區域,亦即聚集成一主要 物件。此步驟可確保3個頂點確實屬於一實際的物件,而 且該物件不會太小。 >在f ϊ I!5,方法5 0 〇詢問是否有額外連接的區域。如果 ,問;^青定回答,那麼方法5 〇 〇便回到步驟5丨5,其中3個 區塊是在連接區域任意選取。步驟515_55〇然後會在 下一連接區域重複。如果詢問是否定回答,然後方法5 〇 〇 會在步驟5 6 0開始合併操作。應注意到先前標示為一主要 物,部=的區塊不再於步驟5丨5 _ 5 5 〇的隨後執行中評估。 即疋’只要在一圖框内的一區塊認為是一主要物件的乘份 ,此區塊便是從其他主要物件的隨後偵測考量而移除,亦 減少區塊數目可用於隨後的評估。因此,如可能,目 則的主要物件分段程序藉由選取來自最大連接區域的3個 區塊開始會較.佳。 在步驟56 0 ’ 一合併操作會接著進行,其中兩實際連接 =主要物件可被選取,目的是要運用測試,以決定兩選定 區域的移動 > 讯是否接近,以致於可將他們認為是一整體 主要物件’以取代兩分開的主要物件。 π Ϊ ί 新組的3個區塊可從兩選定連接區域組合 斤$義的,域選取。例如,如果連接區域62〇和是一合 併操作的》平估,那麼3個任意選定區塊便可由參考數字6 5 〇O: \ 62 \ 62535.ptd Page 18 4 60813 V. Description of the invention (15) Untraceable. That is, 'if the total number of giant blocks with the same affine parameter is greater than a certain preset number of RegMBsize, then all giant blocks that fall within these parameters can provide a common label, such as the main object 1 and so on. All giant blocks with the same label constitute a new mobile area, that is, aggregated into a main object. This step ensures that the 3 vertices really belong to an actual object and that the object is not too small. > At f ϊ I! 5, method 50 asks if there are additional connected areas. If, ask; ^ Qingding answer, then the method 500 returns to step 5 丨 5 of which 3 blocks are arbitrarily selected in the connection area. Steps 515_55 are then repeated in the next connection area. If the query is answered negatively, then method 5 00 will begin the merge operation at step 5 6 0. It should be noted that previously marked as a major, the block of Ministry = is no longer evaluated in the subsequent execution of step 5 丨 5 _ 5 5 〇. That is, as long as a block in a frame is considered to be a multiplier of a main object, this block is removed from subsequent detection and consideration of other main objects, and the number of blocks can be reduced for subsequent evaluation. . Therefore, if possible, it is better to start the main object segmentation process by selecting 3 blocks from the largest connected area. At step 56 0 'a merge operation will proceed, where the two actual connections = the main object can be selected, the purpose is to use a test to determine whether the movement of the two selected areas is close, so that they can be considered as one Integral main object 'to replace two separate main objects. π Ϊ ί The three blocks of the new group can be combined from two selected connection regions, and the domains are selected. For example, if the connection area 62 and the sum are combined and estimated, then three arbitrary selected blocks can be referenced by the reference number 65.

O:\62\62535.ptdO: \ 62 \ 62535.ptd

第19頁 460813 五、發明說明(16) 描述。類似步驟5 2 0,一新組仿射移動參數是從3個區塊的 嶄新選取組的移動向量產生。 類似步驟5 2 5,在步驟5 7 0,方法5 0 0可在兩主要物件内 的每個區塊移動的逐一區塊基礎上詢問該區塊移動是否在 由新產生的仿射移動參數所定義的仿射流内。換句話說, 兩連接的主要物件是與一通常適當的仿射相比較,並且決 定在兩物件中的某些百分比巨區塊具有低於eThresh的誤 差距離。如果百分比超過,那麼兩主要物件便合併成單一 連接的主要物件。即是,如果詢問是肯定回答(誤差減少) ,那麼多數區塊便認為是落在新形成仿射流物件内。因此 ,兩區域是在步驟575合併。 如果詢問是否定回答(誤差增加),那麼多數區塊便認為 是落在新形成的仿射流物件外。因此,兩區域不會在步_驟 5 8 0合併。 在步驟5 8 5,方法5 0 0詢問是否有額外的主要物件認為是 要合併。如果詢問是肯定回答,那麼方法5 0 0便回到步驟 5 6 0,其中兩連接的主要物件可被選擇,亦即,步驟5 6 0 -5 8 0然後會重複,直到所有的物件認為要合併為止。如果 詢問是否定回答,那麼方法5 0 0會在步驟5 9 0結束。 此外,應注意到圖5的相關步驟適合取得如參考圖2上述 的識別主要物件的子物件之仿射移動參數。明確而言,一 主要物件可分成子物件,其中方法5 0 0然後可產生仿射流 子物件。即是,方法5 0 0可在識別主要物件的其中每個的 子物件位準上提供,亦即以"子物件”取代在圖5的"物件"Page 19 460813 V. Description of Invention (16) Description. Similar to step 5 2 0, a new set of affine motion parameters is generated from the motion vectors of the newly selected set of 3 blocks. Similar to step 5 2 5, at step 5 7 0, method 5 0 0 can query whether the block movement is based on the newly generated affine movement parameters on a block-by-block basis for each block movement in the two main objects. Within the defined affine stream. In other words, the two connected main objects are compared to a generally appropriate affine, and it is determined that some percentage of the giant blocks in the two objects have error distances below eThresh. If the percentage is exceeded, the two main objects are merged into a single connected main object. That is, if the query is affirmative (reduced error), then most of the blocks are considered to fall within the newly formed affine stream object. Therefore, the two regions are merged in step 575. If the query is answered negatively (increased error), then most blocks are considered to fall outside the newly formed affine stream object. Therefore, the two regions will not merge in step 580. At step 585, method 5 0 0 asks if there are additional primary items considered to be merged. If the question is answered affirmatively, then method 5 0 returns to step 5 6 0, in which the two connected main objects can be selected, that is, steps 5 6 0-5 8 0 are then repeated until all the objects consider it necessary. Until the merger. If the query is answered negatively, then method 5 0 0 ends at step 5 9 0. In addition, it should be noted that the relevant steps of FIG. 5 are suitable for obtaining the affine movement parameters of the sub-objects identifying the main object as described above with reference to FIG. 2. Specifically, a main object can be divided into sub-objects, where method 500 can then generate affine flow sub-objects. That is, the method 5 0 0 can be provided at the level of the sub-objects that identify each of the main objects, that is, the "sub-objects" in Fig. 5 are replaced with "quote-sub-objects"

O:\62\62535.ptd 第 20 頁 4 60813 五、發明說明(17) 用語。 在每個主要物件内的階層組織,區域分離可根據在子區 間的區域移動而完成。例如,在將當作一主要物件的整個 人體分段之後,例如臉、四肢等的人體部份可分成子區域 或子物件的一連接組。來自區塊移動向量、追蹤、與移動 軌道模型程序的仿射靠模型方法亦可運用在這些子區域。 因此,每個子區域在區間時間上亦是片段模型化。 •假使兩或多個子區域屬於相同的主要物件,他們的軌道 便能用來識別由該主要物件所執行的動作。明確而言’如 果在一主要物件與他的子區域上的進一步本文是藉由手動 裝置或他們形狀、顏色、組織等的分析而提供,那麼在子 區域之間的一組織關係上的此知識條件可發展。因此,藉 由分析子區域移動執道,由主要物件所執行的特殊動作-便 可推測。 例如,如果結果判明主要物件是人(例如,經由皮膚色 澤偵測與形狀),而且子區域符合此人的四肢,那麼從四 肢的不同接合處軌道,例如跑、散步、舉起等(看見在為 下面動作的一本詳細目錄)的特殊移動便可識別。 在不同圖框中物件之間的對應是藉由追蹤在第一圖框的 明顯大小物件而執行,直到它不再追蹤為止,例如直到物 件移出景色或永久封閉為止。在先前圖框中沒有對應的未 來圖框中明顯大小的物件是在剪纔内所提供的新物件標籤 ,並且從該圖框追蹤到未來。既然區塊相配已執行,追蹤 可藉著將在前圖框的一區域指定給在先前圖框的區域追蹤O: \ 62 \ 62535.ptd Page 20 4 60813 V. Description of the invention (17) Terms. Hierarchical organization within each main object, zone separation can be done based on zone movement between sub-zones. For example, after segmenting the entire human body as a main object, human body parts such as faces, limbs, etc. can be divided into sub-regions or a connected group of sub-objects. Affine model methods from block motion vectors, tracking, and moving orbit model programs can also be applied to these sub-regions. Therefore, each sub-region is also segment-modeled in interval time. • If two or more sub-areas belong to the same main object, their orbits can be used to identify the actions performed by that main object. Specifically, 'if further text on a main object and his subarea is provided by a manual device or analysis of their shape, color, organization, etc., then this knowledge of an organizational relationship between subareas Conditions can develop. Therefore, by analyzing the movement of the sub-area, the special action performed by the main object can be speculated. For example, if the result determines that the main object is a person (for example, through skin color detection and shape), and the sub-regions match the person's limbs, then track from different joints of the limbs, such as running, walking, lifting, etc. (see in This is a detailed list of actions below). The correspondence between objects in different frames is performed by tracking objects of obvious size in the first frame until it is no longer tracked, for example, until the object moves out of view or is permanently closed. Objects that did not have a corresponding size in the previous frame in the previous frame are the new object tags provided in the cutout, and are tracked from the frame to the future. Now that block matching has been performed, tracking can be performed by assigning an area in the previous frame to the area in the previous frame.

O:\62\62535.ptd 第 21 頁 4 60813 五、發明說明(18) 組中的§亥區域之1 一間.早方式達成’而此習圖框是大部份巨 區塊移動向量所指向的。 上述方法主要是用來擷取在一"影像拍攝"内大移動物件 。換句話說,此階段的臨界會選取,如此一物件的移動能 以整體補捉,而不是在它之内的較小變化。既然所有的物 件可追蹤,所以在圖框之間的最初巨區塊成員與仿射移動 參數便足以獲得每個主要物件移動軌道的一粗劣合成。移 *軌道的平穩片段可.藉由使用時間多項式而模型化。用以 將移動軌道模型化之方法是由物件執道分段器1 7 0執行, 其在下面會簡短描述。事實上’新物件執道分段器1 7 〇的 詳細描述已在美國專利案號SAR 1 342 7的名稱"Apparatus And Method For Describing The Motion Parameters Of An Object In An Image Sequence"中描述,其在此僅列 出供參考,而且同時在此申請。 物件軌道分段器1 70可用來減少在影像順序的每個圖框 ^ 仿射參數的需要。執道能以片段方式藉由使用適當 二Ξ ΐ係數表ΐ二e粗劣物件執道在有關整體物件移動 2回答拘問是相當有用,例如快速移動、向上、 向 左、向右、對角線、旋轉、進入z方向、移 .典型上,最大連接的主要物件將會是背 ° =動分析可自動提供例如搖鏡、縮放责傾=^ 外歎=些可以疋由於或攝影機移動或由 外,*攝影機所追蹤的主要物件可由此 :2 : 置上具有最少變化的事實而識別。 中在&間時間位O: \ 62 \ 62535.ptd Page 21 4 60813 V. Invention description (18) One of the §11 areas in the group. Earlier method was achieved ', and this exercise frame is where most of the giant block movement vectors are Pointing. The above method is mainly used to capture large moving objects within an "Image Shooting". In other words, the criticality at this stage is chosen so that the movement of an object can be captured as a whole, rather than a small change within it. Since all objects are trackable, the initial giant block members and affine movement parameters between the frames are sufficient to obtain a crude composition of the movement track of each main object. Smooth segments of shifted * tracks can be modeled by using time polynomials. The method used to model the moving trajectory is performed by the object track segmenter 170, which is described briefly below. In fact, the detailed description of the "new object execution segmenter 1 70" has been described in the name " Apparatus And Method For Describing The Motion Parameters Of An Object In An Image Sequence " of U.S. Patent No. SAR 1 342 7. It is listed here for reference only and is also applied here. Object track segmenter 1 70 can be used to reduce the need for affine parameters in each frame of the image sequence. It can be fragmented by using the appropriate Ξ ΐ coefficient table ΐ 2 e. Poor objects. It is very useful to answer questions about the overall object movement. 2 For example, move quickly, up, left, right, diagonal. , Rotation, enter the z direction, and shift. Typically, the main object of the largest connection will be the back. Dynamic analysis can automatically provide for example panning, zooming, etc. ^ Wai sigh = some can be caused by either camera movement or external , * The main objects tracked by the camera can be identified by this: 2: Put on the fact that has the least change. Between & Time

五、發明說明(19) 在一具體實施例中,主要物件可藉著使用橢圓形而模型 化。藉由估計來自軌道資訊、閉塞與主要物件交談的橢圓 形重疊而識別。閉塞可藉著移動平穩持續通過閉塞點的事 實而從交談區別,然而物件交談通常會造成移動的變化傾 向。物件交談可以是例如相接、碰撞等的情況,其本自是 重要的後設資料(當作想要的暫時點)。閉塞可使用在圖框 的選擇,而這些圖框可用於空間分析(以擷取顏色、形狀 、織法後設資料)。典型上,在主要物件(除了背景之外) 之間的最小重疊圖框是理想用於此目的,所以在每個主要 物件上的適當空間後設資料便可擷取。 應注意到追蹤步驟在已分段的影像是不需要執行,例如 一MPEG-4影像物件。然而,一仿射模型轉換的巨區塊移動 向量的程序與用以估計主要物件的一片段移動軌道的方_法 仍可應用。 當推論可汲時或有多少推論細節將決定在應用領域。例 如,若要有助於快速搜尋與擷取,一搜尋引擎要計算”動 作π ,並且根據他們而將順序分類。若要允許在不同應用 與相互間操作的搜尋,如果移動轨道描述符號連同一些開 始位置及大小描述符號是每個主要物件/區域/子區域的 標準化 ',它會是足夠的。這些描述符號提供一小型然強有 力能根據動作而執行搜尋/過濾。這些描述符號亦可順利 地補助其他空間的描述符號與來自這些描述符號的影響力 知識,以改良擷取的精確性。結構本質可使它導致一階層 描述方法。應注意到雖然目前的物件移動分段已使用仿射V. Description of the Invention (19) In a specific embodiment, the main object can be modeled by using an oval shape. Recognized by estimating the elliptical overlap from orbital information, occlusion, and conversation with the main object. Occlusion can be distinguished from the conversation by the fact that the movement continues smoothly through the occlusion point, but object conversations often cause a change in movement. Object conversations can be situations such as connection, collision, etc., which are inherently important meta data (as a temporary point of interest). Occlusion can be used in the selection of frames, and these frames can be used for spatial analysis (to capture color, shape, and post-weave data). Typically, the smallest overlapping frame between the main objects (except the background) is ideal for this purpose, so the data can be retrieved by setting the data after an appropriate space on each main object. It should be noted that the tracking step need not be performed on segmented images, such as an MPEG-4 image object. However, the procedure of a huge block moving vector transformed by an affine model and the method of estimating a fragment's moving trajectory of the main object are still applicable. When inferences can be drawn or how many details of inferences will determine the field of application. For example, to facilitate fast search and retrieval, a search engine calculates "actions π" and sorts them according to their order. To allow searches in different applications and interoperability, if the moving track description symbol is accompanied by some The start position and size descriptors are standardized for each main object / area / sub-area, and it will be sufficient. These descriptors provide a small but powerful search / filter based on the action. These descriptors also work smoothly To supplement the descriptive symbols of other spaces and the knowledge of the influence from these descriptive symbols to improve the accuracy of extraction. The structural nature can lead to a one-level description method. It should be noted that although the current object segmentation has used affine

O:\62\62535.ptd 第23頁 4 60813 五、發明說明(20) 模型來描述,但是目前的新結構亦能容易地擴展成其他的 移動模型。例如,目前的物件移動分段可藉由使用"轉換 移動"模型、”等方性移動”模型(一定的比例,以X , y +轉 換旋轉)、與"透視移動”模型而改造。 在一具體實施例中,物件軌道分段器1 7 0採用一區域幾 何意義的二次方程(有關時間)移動模型之一分裂與合併方 法,該區域的幾何意義可歸納成如下所示的隨機移動軌道 模型:假使一區域的逐一圖框移動資訊發展一表示式以預 測在區間時間的物件上之所有點的位置,並且在根據與原 始位置相比較預測的每個時間瞬間上,於區域每點上選取 測量距離偏離總和的一誤差公制。然後如果該誤差公制超 過某一實驗臨界T 1且在偏移最小的瞬間上,此區間便可分 開。如果在兩區間上的共同預測表示產生小於T 1或另外一實 驗臨界的一錯誤,合併程序將可合併兩毗連區間。 仿射移動模型的範例如下所示: (a) 在毗連圖框之間取得仿射移動參數,以便在每個時 間瞬間的一區域上描述每點的位置。 (b) 決定於例如二次方程式需要至少兩資料組等適合順 序,在選定區間的一子取樣組圖框之間取得仿射移動參數 〇 (C)將仿射移動參數分解成它的元件,即是比例、旋轉 、剪裁、與轉換。假設每個元件的不同暫時模型是決定在 他的本質。例如,轉換可藉由使用時間二次方程式而模型 化。比例可隨著時間線性變化而模型化。旋轉可藉由使用O: \ 62 \ 62535.ptd Page 23 4 60813 V. Description of the invention (20) model to describe, but the current new structure can also be easily extended to other mobile models. For example, the current segmentation of object movement can be modified by using the "transformation movement" model, "isotropic movement" model (certain proportions, X, y + transformation rotation), and "perspective movement" model. In a specific embodiment, the object track segmenter 170 uses a method of splitting and merging a quadratic equation (relevant time) movement model of a region's geometric meaning, and the geometric meaning of the region can be summarized as follows Random moving orbit model: Suppose that one area of frame-by-frame movement information of an area develops an expression to predict the position of all points on the object at the interval time, and at each time instant predicted based on the comparison with the original position, the area At each point, an error metric is selected that deviates from the sum of the measurement distance. Then if the error metric exceeds a certain experimental threshold T 1 and the instant at which the offset is the smallest, this interval can be separated. If a common prediction representation on both intervals results An error less than T 1 or another experimental criticality, the merge procedure will merge two contiguous intervals. An example of an affine movement model is shown below: (a) Obtain affine movement parameters between adjacent frames in order to describe the position of each point on a region at each time instant. (b) Determined by, for example, a quadratic equation that requires at least two data sets in a suitable order. Obtain affine movement parameters between the frames of a sub-sampling group in the selected interval. (C) Decompose the affine movement parameters into its components, namely scale, rotation, clipping, and transformation. Assume that each component has a different temporal The model is determined by his essence. For example, transformations can be modeled by using a quadratic equation of time. Proportions can be modeled linearly with time. Rotation can be modeled by using

O:\62\62535.ptd 第24頁 <80813 五、發明說明(21) 一固定的角速度假設而模型化。 (d) 隨著使用子取要的仿射移動,獲得選定預測模型的 係數。 (e) 在每個圖框上應用預測模型,並且在每個圖框的所 有點上計算偏離總和。將圖框錯誤加總,將他們常態比較 一預設臨界T1。如果誤差〉T1,在最大圖框誤差位置上將 區間分成2。 (f )重複步驟(b ) - ( e)有關每個區間重複5直到在所有區 間中的誤差低於ΤΓ為止。 (g) 現在如果兩毗連區間的共同預測模型造成低於臨界 T 2的正常誤差,現便可合併區間。 (h) 對於在任何進一步合併的上述T 2合併區間中增加正 常化誤差之前,於所有的遞迴區間重複(g )。 - 資料結構 為了此揭露的目的,應了解到一影像順序可分成11影像 拍攝"。每個影像拍攝是藉著在影像拍攝内的不同區間時 間上使用有效的影像拍攝描述方法的一連結清單(亦即, 一新的資料結構)而表示。影像拍攝資料結構(DS)根據顏 色、形狀、組織、與移動而具有子描述方式或描述符號。 一移動資料連同一移動描述符號是下面揭露。資料結構具 有 global_motion、object_motion、與 region_niotion 描 述的一階層結構。一連結清單可用來指向下一區間時間。 移動軌道描述方法 D S描述方法O: \ 62 \ 62535.ptd Page 24 < 80813 V. Description of the invention (21) A fixed angular velocity is assumed and modeled. (d) Obtain the coefficients of the selected prediction model with the affine movement required by the sub-fetch. (e) Apply the prediction model on each frame and calculate the sum of the deviations on all points of each frame. Sum the frame errors and compare them to a normal threshold T1. If the error is> T1, divide the interval into 2 at the position of the maximum frame error. (f) Repeat steps (b)-(e) Repeat 5 for each interval until the error in all intervals is lower than TΓ. (g) Now if the common prediction model of two contiguous intervals causes a normal error below the critical T 2, the intervals can now be merged. (h) Repeat (g) for all recursive intervals before adding normalization errors to any of the further T 2 merged intervals described above. -Data Structure For the purpose of this disclosure, it should be understood that an image sequence can be divided into 11 image captures ". Each image capture is represented by a linked list (ie, a new data structure) using effective image capture description methods at different intervals in the image capture. The image capture data structure (DS) has sub-description methods or description symbols according to color, shape, organization, and movement. A mobile data with the same mobile description symbol is disclosed below. The data structure has a hierarchical structure described by global_motion, object_motion, and region_niotion. A linked list can be used to point to the next interval. Moving track description method DS description method

O:\62\62535.ptd 第25頁 46〇8i3O: \ 62 \ 62535.ptd Page 25 46〇8i3

五、發明說明(22) 一連結清單資料結構可被假設(例如,在一影像拍攝内 的主要物件描述方法可藉著從最初指標器開始及存取下一 連結的.指標器而存取)。在階級中的描述是以鋸齒狀顯示 ° (開始時間、區間時間)對是始終使用,以表示時間支援 °時間可以是媒體時間(在本文上的時間參考)或實際時 間。 shot_DS{ - start一time, time—interval (Time and duration in media/actual time)V. Description of the invention (22) A link list data structure can be assumed (for example, the main object description method in an image capture can be accessed by starting from the initial indicator and accessing the next indicator.) . The description in the class is shown in a zigzag pattern. ° (start time, interval time) is always used to indicate time support. ° Time can be media time (time reference in this article) or actual time. shot_DS {-start_time, time—interval (Time and duration in media / actual time)

Number of key objects (First level key objects in a shot)Number of key objects (First level key objects in a shot)

Global—motion—DS (Background key object)Global—motion—DS (Background key object)

Pointer to Key 一 object JDS array (Pointer to first entry in a linked list of key object DSs)Pointer to Key-object JDS array (Pointer to first entry in a linked list of key object DSs)

Key一object—interaction一times DS (DS containing the instants within the shot when key objects interacted) - } GIobal_motion_DS{ start—time,time_intervalKey_object_interaction_times DS (DS containing the instants within the shot when key objects interacted)-} GIobal_motion_DS {start_time, time_interval

Pointer to first link of Global_motion_subinterval 一 trajectory JDS }Pointer to first link of Global_motion_subinterval a trajectory JDS}

Sub_interval_global_motion_trajectory_DS{ start」ime,time」ntervalSub_interval_global_motion_trajectory_DS {start ”ime, time” nterval

Model_class」d for motion_trajectory—descriptor (Trajectory description may use a predefined set of models (e.g., quadratic in position of object centroid) - the model—clas一id is the index of the chosen model in the set))Model_class ”d for motion_trajectory—descriptor (Trajectory description may use a predefined set of models (e.g., quadratic in position of object centroid)-the model—clas-id is the index of the chosen model in the set))

Global motion—trajectory—descriptor ' (descriptor parameters for global motion trajectory based on the model—class_id)Global motion_trajectory_descriptor '(descriptor parameters for global motion trajectory based on the model_class_id)

Pointer to next_interval_Pointer to next_interval_

Sub—interval_globaljmotion_trajectory—DSSub—interval_globaljmotion_trajectory—DS

Key_obj ect—interaction 一 times 一 DS {Key_obj ect—interaction one times one DS {

Pointer to first like of key_object—interaction time descriptorPointer to first like of key_object—interaction time descriptor

O:\62\62535.ptd 第26頁 460813 五、發明說明(23)O: \ 62 \ 62535.ptd Page 26 460813 V. Description of the invention (23)

Key„obj ect_DS {Key „obj ect_DS {

Key一object_ID (Label or index of key object) start一time,time一interval (start time and time interval in termsKey_object_ID (Label or index of key object) start_time, time_interval (start time and time interval in terms

of frames, i.e., at what frame is object detected and for how long) Key_object 一 first__frame„geometry_DSof frames, i.e., at what frame is object detected and for how long) Key_object a first__frame „geometry_DS

(Used to define the initial coarse shape) Key_object_motion_trajectory_DS(Used to define the initial coarse shape) Key_object_motion_trajectory_DS

Number of sub_regions (Number of sub-objects within the object)Number of sub_regions (Number of sub-objects within the object)

Pointer to sub_region„DS arrayPointer to sub_region „DS array

Key一object一spatial_DS (Spatial description of key objects (such as color,shape,texture, etc·)) }Key_object_spatial_DS (Spatial description of key objects (such as color, shape, texture, etc.))}

Key„object_motion_trajectory_DS{Key „object_motion_trajectory_DS {

Pointer to first link of key—object—sub_interval一motion_trajectory_DS (Linked—list of key一object一sub一interval—motion一trajectory—DS) }Pointer to first link of key_object_sub_interval_motion_trajectory_DS (Linked_list of key_object_sub_interval_motion_trajectory_DS)}

Key_object_sub—interval— motion—trajectory 一 DS{ start_tim6, timejnterval Model—class一id for motion—trajectory一descriptor Key一object motion_trajectory_descriptor Pointer to sub一region—DS array Pointer to next_interval key—object一sub-interval一motion 一trajectory—DS }Key_object_sub_interval_ motion_trajectory-DS {start_tim6, timejnterval Model_class-id for motion_trajectory-descriptor Key-object motion_trajectory_descriptor Pointer to sub-region-DS array Pointer to next_interval key-object-sub-interval-motion-trajectory —DS}

Sub_Region_DS{Sub_Region_DS {

Sub_Region_ID — start_tixne, time_intervalSub_Region_ID — start_tixne, time_interval

SuB—region一motion—trajectory 一DS Sub—region 一 spatial JDS }SuB_region_motion_trajectory_DS Sub_region_spatial JDS}

Sub„region_motion_trajectory_DS{Sub „region_motion_trajectory_DS {

Pointer to first link of sub„region_sub_interval 一 motion_trajectory_DS (Linked—list of sub一region一sub 一 interval一motion—trajectory—DS) ΙΗΠ 第27頁 460813 五、發明說明(24)Pointer to first link of sub „region_sub_interval_motion_trajectory_DS (Linked_list of sub_region_sub_ interval_motion_trajectory_DS) ΙΗΠ page 27 460813 V. Description of the invention (24)

Sub_region_sub_interval_ motion_trajectory_DS{ start_time,time」ntervalSub_region_sub_interval_ motion_trajectory_DS {start_time, time "nterval

Model_class_id for motion_trajectory_descriptor Sub_region motion一trajectory—descriptor Pointer to next_intervalModel_class_id for motion_trajectory_descriptor Sub_region motion_trajectory—descriptor Pointer to next_interval

sub_re^ion_sub_interval._m〇tion_trajectory_DS 插 ifi· 1 ' ------ ......... ., ....... .... .... _____„ ________..... .··..............................------------------------- motion_trajectory„descriptor{ float ax, ay? vx, vy, px, py; (for quadratic model) )sub_re ^ ion_sub_interval._m〇tion_trajectory_DS Insert ifi · 1 '------ .........., ....... .... .... _____ „________... .............................................. ---------- motion_trajectory „descriptor {float ax, ay? vx, vy, px, py; (for quadratic model))

Model_class_ID descriptor for motion trajectory {Model_class_ID descriptor for motion trajectory {

Model class ID of the motion trajectory model usedModel class ID of the motion trajectory model used

Key一object一interaction_time—descriptor { time reference (Time with respect to start of shot) number of key_objects (involved in the interaction) key一object—id of the objects interacting at that time (indices of key objects) 一 根據一物件時間幾何平均位置的二次方程模型之移動軌 道描述符號只作為一描述提供。 鑑於本移動分段與軌道分段,複數新詢問能以如下所示 實施: 只根據移動的可能詢問 a) 通用的移動: i ) 一影像拍攝,其中攝影機正在擺鏡、傾斜、滚動、縮 放、或追蹤(背景移動)。 ii)通用移動(表示在影像拍攝的動作範圍)的範圍 b) 區域物件/區域移動: i )在影像拍攝中明顯物件數目Key_object_interaction_time_descriptor {time reference (Time with respect to start of shot) number of key_objects (involved in the interaction) key_object_id of the objects interacting at that time (indices of key objects) The moving orbit description symbol of the quadratic equation model of the time geometric mean position is provided as a description only. In view of this moving segment and orbit segment, a plurality of new queries can be implemented as follows: Only possible queries based on movement a) Universal motion: i) An image capture in which the camera is tilting, tilting, rolling, zooming , Or tracking (background moving). ii) The range of general movement (indicating the range of action in image shooting) b) Area object / area movement: i) The number of obvious objects in image shooting

O:\62\62535.ptd 第28頁 4 6 0 81 3 五、發明說明(25) i i )移動(快/慢)的範圍 i i i )由攝影機所追蹤的物件 i v)突然非連續的位置 v)激烈動作 v i)識別橫過順序與時間上的類似追蹤 vii)内接觸與外面接觸移動的比較 根據在其他描述符號上的條件而可能的詢問 人類的動作:散步、跑、跳越、彎曲、坐著、站立、倒 下、旋轉、跳舞、姿態、握手 '投擲、捕捉、舉起、打擊 、溜冰鞋、跪、踏板、以拳重擊、切割、用籬笆圍住、點 頭、保持、寫、類型 車輛動作:飛、移動、停止、航行、巡航、轉彎、潛 水、旋轉 - 其他生活物件動作:動物(鳥等)奔跑、飛躍、游泳、潛 水的動作 任意動作:樹葉、河流 物件交談:在動作/與無動作物件上的人類動作 用以尋找上面每個詢問相配的類似公制 一般的解決是在藉由使用序列資料而在移動模型參數空 間建立每個動作的可能密度功能(p d f )。然後,一特定取 樣移動模型參數組可藉由使用具當作公制的最大可能(或 記錄可能)測試Μ - a r y假設而分成該等動作的其中一個。模 型參數必須正常適合於p d f的建立與隨後的分類,如此可 以是物件的大小和位置與分析的圖框率無關的。O: \ 62 \ 62535.ptd Page 28 4 6 0 81 3 V. Description of the invention (25) ii) Range of movement (fast / slow) iii) Objects tracked by the camera iv) Sudden discontinuous positions v) Vigorous movements vi) Recognition of traversing sequence and similar tracking in time vii) Comparison of internal and external contact movements Possible human inquiries based on conditions on other descriptors: walking, running, jumping, bending, sitting Standing, standing, falling, spinning, dancing, stance, shaking hands, throwing, catching, raising, hitting, skates, kneeling, pedaling, punching, cutting, fenced, nodding, holding, writing, type Vehicle movements: fly, move, stop, sail, cruise, turn, dive, spin-other living objects. Action: animal (bird, etc.) running, leap, swimming, diving. Any action: leaves, river objects. Talking: in action / A similar metric-like general solution to human actions on non-action objects to find each query above is to create a feasible model for each action in the moving model parameter space by using sequence data. Energy density function (p d f). Then, a specific set of sampled moving model parameters can be divided into one of these actions by using the maximum possible (or recorded possible) test of the metric to test the M-a y hypothesis. The model parameters must be normally suitable for the establishment of p d f and subsequent classification, so that the size and position of the objects are independent of the frame rate of the analysis.

O:\62\62535.ptd 第29頁 4 60 81 3 五、發明說明(26) 雖然本物件移動分段與物件移動執道分段在上述是用以 改良影像順序的索引與擷取,但是應了解到本發明可用來 改良或提供其他的影像處理功能。例如,本發明能採用在 編碼功能,例如,識別的主要物件可認為是特殊重要區 域,或取回通用的移動參數。 此外,雖然本發明在上面是以物件描述,但是應了解到 一物件是廣泛定義為視應用而定的變化大小之特定區域。 同樣地,雖然本發明在上面是以例如巨區塊之區塊描述, 但是應了解到一區塊是寬廣定義成決定在特殊應用的一改 變大小區塊。 雖然結合本發明闡述的各種不同具體實施例已在此詳細 顯示及描述,但是習於此技者仍然可結合這些闡述而 成 其他不同的具體實施例。 一O: \ 62 \ 62535.ptd Page 29 4 60 81 3 V. Description of the Invention (26) Although the object movement segmentation and object movement instruction segmentation are used to improve the order and retrieval of the image sequence, but It should be understood that the present invention can be used to improve or provide other image processing functions. For example, the present invention can be used in a coding function, for example, the main object identified can be considered as a particularly important area, or general motion parameters can be retrieved. In addition, although the invention is described above as an object, it should be understood that an object is a specific area that is broadly defined as a variable size depending on the application. Similarly, although the present invention is described above as a block such as a giant block, it should be understood that a block is a block that is broadly defined to determine a change in size for a particular application. Although various specific embodiments described in connection with the present invention have been shown and described in detail herein, those skilled in the art can still combine these descriptions to form other different specific embodiments. One

O:\62\62535.ptd 第30頁O: \ 62 \ 62535.ptd Page 30

Claims (1)

/、、申請專利範圍 h 種用以執行具有複數圖樞 刀奴之方法,該方法包含下列步騍.影像順序的物件移動 a) 決定在該影像順序的其中〜願 域; 圖框内之至少一連接區 :及 b) 應用以區塊為基礎之移動 蕙產生光學流移動參數 c) 使用該產生的光學流移動 2. 如申請專利範圍第丨項之方法数以識別一主要物件。 該至少比連接區域的至少3個區’、其中該步驟b)可提供 向量’以產生該出光學流移動參的以區塊為基礎之移動 3. 如申請專利範圍第1項之方法 驟: ’其進一步包含下列步 d) 決定在步驟㈧上發現的任 - 單一主要物件。 4兩識別主要物件是否為 4. 如申請專利範圍第1項之方法,*、 驟: ’其進一步包含下列步 發現之一識別的主要物件;及 -目前圖框的移動向量資訊而追蹤在 一區域鱼一杏^ = Γ不主要物件,以使在該目前圖框中的 °° ,、 刚圖框的標不主要物件形成相互關聯。 5·如申請專利範圍第丨項之方法,其進一列步 驟: d)根據來自步驟c)的任何識別主要物件的該光學流移 動參數而將該影像順序予以索引。 r 46081 3 六、申請專利範圍 驟6:.如申請專利範圍第!項之方法,其進一步包含下列步 d) 將該識別的主要物件分離成複數子物件;及 e) 提供來自該等子物件的其中每一子物件的至 ^ = = = Π向量,以產生該等子物件:其 子物件之組新的光學流移動參數。 、 與軌道:二:具:框的-影像順序的移動分段 退々段之方法該方法包含下列步驟: 圖框2 Ϊ用光學流物件移動分段’以獲得該影像順序的-圖框的至少一主要物件之移動資訊;及 的 圖樞t) Ϊ用光學流物件軌道分段’以獲得該影像順序的-&間的該至少一主要物件之軌道資訊。 與軌:ί用以執行具有複數圖框的一影像順序之移動分段 興軌^分段之裝置(15〇),該裝置包含: 刀奴 序的移動分段器(16〇),用以獲得該影像順 町一圖框的至少一主要物件之移動資訊;及 一光學流物件軌道分段器(17〇),用以獲得該影像順 Q 圖框區間的該至少一主要物件之軌道資訊。 指人—種具有儲夸複數指令之電腦可讀媒體(130),複數 包括當由一處理器執行時,造成該處理器執行下列步 域a)決定在該影像順序的其中一圖框内至少一連接的區 b)應用以區塊為基礎之移動向量產生光學流移動參數 4 80813_ 六、申請專利範圍 ;及 C)使用該所產生的光學流移動參數以識別一主要物件 〇 10. —種具有儲存複數指令之電腦可讀媒體(130),複數 指令包括當由一處理器執行時,造成該處理器執行下列步 驟: a )應用光學流物件移動分段,以獲得該影像順序的 一圖框的至少一主要物件之移動資訊;及 b )應用光學流物件軌道分段,以獲得該影像順序的 一圖框區間的該至少一主要物件之軌道資訊。/ 、, Patent application scope h Methods for performing a method with plural figure pivot knife slaves, the method includes the following steps: image sequence of object movement a) decide in the image sequence of ~ ~ field; at least in the picture frame; A connection area: and b) applying block-based movements to generate optical flow movement parameters c) using the generated optical flow movements 2. The number of methods such as the scope of the patent application to identify a main object. The at least three areas that are more than the connection area ', wherein the step b) can provide a vector' to generate the block-based movement of the optical flow movement parameter 3. As the method step of the scope of patent application: 'It further includes the following step d) to decide on any of the items found in step--a single primary object. 4 Identifies whether the main object is 4. If the method of the scope of the patent application is the first, *, step: 'It further includes the main object identified by one of the following steps; and-the current vector of the frame is tracked in one Regional fish and apricot ^ = Γ are not the main objects, so that in the current frame, °, the main objects of the rigid frame are related to each other. 5. The method according to item 丨 of the scope of patent application, which further includes a series of steps: d) indexing the image sequence according to any optical flow movement parameter identifying the main object from step c). r 46081 3 VI. Scope of Patent Application Step 6: If the method of the scope of patent application is applied, it further includes the following steps d) separating the identified main object into plural sub-objects; and e) providing from these sub-objects Up to ^ = = = Π vector of each of the sub-objects to generate such sub-objects: a new set of optical flow movement parameters for its sub-objects. , And track: two: with: frame-moving segmentation method of image segment retreat segment method The method includes the following steps: Frame 2: Use optical flow objects to move segments' to obtain the image sequence-frame-frame Movement information of at least one main object; and Figure t) (i) Use optical flow object track segmentation to obtain track information of the at least one main object between-& of the image sequence. And rail: a device (15) for performing moving segmentation and segmentation of an image sequence with a plurality of frames. The device includes: a knife-sequential mobile segmenter (16) to obtain the The moving information of at least one main object of a frame of the image along the frame; and an optical flow object track segmenter (17) to obtain the orbit information of the at least one main object of the image along the frame of the Q frame. Refers to a person—a computer-readable medium (130) with a plurality of stored instructions. The plural includes when executed by a processor, causing the processor to perform the following steps: a) Decide at least in one of the frames of the image sequence. A connected area b) applying block-based motion vectors to generate optical flow movement parameters 4 80813_ VI. Patent application scope; and C) using the generated optical flow movement parameters to identify a major object A computer-readable medium (130) having a plurality of instructions stored therein. When executed by a processor, the processor causes the processor to perform the following steps: a) Applying an optical stream object to move segments to obtain a picture of the image sequence Movement information of at least one main object of the frame; and b) applying optical flow object track segmentation to obtain track information of the at least one main object in a frame interval of the image sequence. 第33頁Page 33
TW89101505A 1999-01-28 2000-01-28 Apparatus and method for context-based indexing and retrieval of image sequences TW460813B (en)

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US11762999P 1999-01-28 1999-01-28

Publications (1)

Publication Number Publication Date
TW460813B true TW460813B (en) 2001-10-21

Family

ID=22373951

Family Applications (1)

Application Number Title Priority Date Filing Date
TW89101505A TW460813B (en) 1999-01-28 2000-01-28 Apparatus and method for context-based indexing and retrieval of image sequences

Country Status (1)

Country Link
TW (1) TW460813B (en)

Similar Documents

Publication Publication Date Title
US6643387B1 (en) Apparatus and method for context-based indexing and retrieval of image sequences
TW454158B (en) Apparatus and method for describing the motion parameters of an object in an image sequence
Siarohin et al. First order motion model for image animation
Ngo et al. Motion analysis and segmentation through spatio-temporal slices processing
JP5355422B2 (en) Method and system for video indexing and video synopsis
CN107707931B (en) Method and device for generating interpretation data according to video data, method and device for synthesizing data and electronic equipment
Dang et al. RPCA-KFE: Key frame extraction for video using robust principal component analysis
US20190066732A1 (en) Video Skimming Methods and Systems
Pritch et al. Webcam synopsis: Peeking around the world
JP4973188B2 (en) Video classification device, video classification program, video search device, and video search program
Kim et al. Object-based video abstraction for video surveillance systems
US6956573B1 (en) Method and apparatus for efficiently representing storing and accessing video information
Tompkin et al. Videoscapes: exploring sparse, unstructured video collections
JP2001526859A (en) Instruction and editing method of compressed image on world wide web and architecture
Assa et al. The virtual director: a correlation‐based online viewing of human motion
JP2006244074A (en) Moving object close-up frame detection method and program, storage medium storing program, moving object close-up shot detection method, moving object close-up frame or shot detection method and program, and storage medium storing program
JP4792471B2 (en) Information element extraction method for image sequence data search and recording medium recording the method
WO1999005865A1 (en) Content-based video access
TW460813B (en) Apparatus and method for context-based indexing and retrieval of image sequences
Gebali et al. Detection of salient events in large datasets of underwater video
Kim et al. Keyframe selection for motion capture using motion activity analysis
Kaup et al. Video analysis for universal multimedia messaging
Gandhi et al. Object-based surveillance video synopsis using genetic algorithm
Ciampi et al. Mc-gta: A synthetic benchmark for multi-camera vehicle tracking
Zeppelzauer et al. Retrieval of motion composition in film

Legal Events

Date Code Title Description
GD4A Issue of patent certificate for granted invention patent
MK4A Expiration of patent term of an invention patent