TW201143452A - Image encoding device, image decoding device, image encoding method and image decoding method - Google Patents

Image encoding device, image decoding device, image encoding method and image decoding method Download PDF

Info

Publication number
TW201143452A
TW201143452A TW100104089A TW100104089A TW201143452A TW 201143452 A TW201143452 A TW 201143452A TW 100104089 A TW100104089 A TW 100104089A TW 100104089 A TW100104089 A TW 100104089A TW 201143452 A TW201143452 A TW 201143452A
Authority
TW
Taiwan
Prior art keywords
image
vector
direct
time
tile
Prior art date
Application number
TW100104089A
Other languages
Chinese (zh)
Inventor
Yusuke Itani
Shunichi Sekiguchi
Kazuo Sugimoto
Yoshihisa Yamada
Original Assignee
Mitsubishi Electric Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Mitsubishi Electric Corp filed Critical Mitsubishi Electric Corp
Publication of TW201143452A publication Critical patent/TW201143452A/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • H04N19/105Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/176Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/577Motion compensation with bidirectional frame interpolation, i.e. using B-pictures
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • H04N19/61Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

Time direct vector generating part 13 specifies a macro block located spatially at a same position with a macro block of an encoding subject from macro blocks constituting encoded pictures timely close to the macro block of the encoding subject, and generates time direct vectors of a time direct mode from moving vectors of four rectangular blocks located in the center of said macro block.

Description

201143452 * 六、發明說明: 【發明所屬之技術領域】 本發明係有關使用於影像壓縮編碼技術或壓縮影像資 料傳送技術等的,讀編碼裝置、影像解碼裝置、影像編碼 方法、以及影像解碣方法。 【先前技術】 「例如,在 MP_oving Picture 細⑶忭 Gr〇up)或 ITU TU. 26χ」等國際標準影像編碼方式中,係以將輝度 (brightness)訊號16xl6像素(pixel)、與對應於該輝度訊 號訊號16x16像素的色差訊號8χ8像素集合而成的圖塊資 料〇31〇4(18七3,以下稱「圖塊集(贴(^〇131〇(:1〇」)作為一 單位,而採用根據動作補償技術或正交變換/變換係數量子 化技術而壓縮的方法。 在衫像編碼裝置以及影像解碼裝置的動作補償處理 中’係參照前方或後方的圖像(piC1:ure),而以圖塊集單位 進行動作向量(vector )的偵測和預測影像的產生。 此時’僅參照1張圖像而進行晝面間預測編碼者稱為 P圖像,同時參照2張圖像而進行畫面間預測編碼者則稱 為B圖像。 於為國際標準方式的AVC/H.264CIS0/IEC 14496-101 ITU-T H.264)中,將B圖像編碼之際,可選擇被稱為直接 模式(direct mode)的編碼模式(參照例如非專利文獻1)。 亦即,可選擇於編碼對象的圖塊集中不具有動作向量 之編碼資料,而藉由使用已編碼的其他圖像之圖塊集的動 3 322764 201143452 作向量、和空間上為周圍的已編碼圖塊集之動作向量的預 定演算而產生編碼對象之圖塊集之動作向量的編碼模式。 於該直接模式中,係存有時間直接模式與空間直接模 式的兩種模式。 時間直接模式,係參照已編碼的其他圖像之動作向 量*因應已編碼圖像與編碼對象圖像間之時間差而進行動 作向量之縮放(scaling)處理,藉此產生編碼對象之圖塊集 的動作向量。 空間直接模式,係參照位於編碼對象之圖塊集周圍的 至少1個以上之已編碼圖塊集之動作向量,並從該等動作 向量中產生編碼對象的圖塊集之動作向量。 於此直接模式中,藉由使用設定於切片標頭(si ice header)之旗標的 “direct_spatial_mv_pred_flag” ,而 可以切片為單位地選擇時間直接模式或空間直接模式其中 任一方。 在此,第12圖為示有在時間直接模式下產生動作向量 的方法的示意圖。 於第12圖中,「P」表示P圖像,「B」表示B圖像。 另外,數字0-3係表示圖像之顯示順序,表示其為時 間ΤΟ、ΤΙ、T2、T3之顯示影像。 圖像之編碼處理係依P0、P3、Bl、B2之順序進行。 例如,假設為將圖像B2之中的圖塊集MB1以時間直接 模式編碼的情形。 此時,在位於圖像B2之時間軸上後方的已編碼圖像之 4 322764 201143452 中,使用為最接近圖像B2的圖像P3之動作向量的位於與 圖塊集MB1在空間上相同位置的圖塊集MB2的動作向量MV。 該動作向量MV係參照圖像P0,於將圖塊集MB1編竭 時所使用的動作向量MVL0、MVL1,係可由以下式(1)求出。 MVLO = MVL\ 二 T2-T0 Γ3-ΓΟ Γ2-Γ3 T3-T0201143452 * VI. Description of the Invention: [Technical Field] The present invention relates to a video encoding and decoding technology, a compressed image data transmission technology, a read encoding device, a video decoding device, a video encoding method, and an image decoding method. . [Prior Art] "For example, in MP_oving Picture Fine (3) 忭 Gr 〇up) or ITU TU. 26 χ", the international standard image coding method is to use a brightness signal of 16 x 16 pixels (pixel), and corresponding to the luminance Signal signal 16x16 pixel color difference signal 8χ8 pixels aggregated block data 〇31〇4 (18 7 3, hereinafter referred to as "tile set (paste (^〇131〇 (:1〇)) as a unit, and A method of compressing according to a motion compensation technique or an orthogonal transform/transform coefficient quantization technique. In the motion compensation processing of the shirt image encoding device and the video decoding device, the image is referred to in front or rear (piC1: ure), and In the block set unit, the motion vector (vector) is detected and the predicted image is generated. At this time, the reference image is referred to as a P image by referring to only one image, and the two images are simultaneously referred to. The inter-picture predictive coder is called a B picture. In the AVC/H.264 CIS0/IEC 14496-101 ITU-T H.264, which is an international standard method, when B picture is coded, it can be called Direct mode encoding mode For example, Non-Patent Document 1). That is, it is possible to select a coded material that does not have an action vector in a tile set to be encoded, and to use a motion of a tile set of other images that have been encoded as a vector, And an encoding mode of the motion vector of the tile set of the encoding object generated spatially for the predetermined calculation of the motion vector of the surrounding coded tile set. In the direct mode, there are two modes of the time direct mode and the spatial direct mode. The time direct mode refers to the motion vector of the other images that have been encoded. * The scaling of the motion vector is performed in response to the time difference between the encoded image and the image to be encoded, thereby generating a map of the encoded object. An action vector of a block set. The spatial direct mode refers to an action vector of at least one or more coded tile sets located around a tile set of the encoding object, and generates a tile set of the encoding object from the motion vectors. Motion vector. In this direct mode, by using the "direct_spatial_mv_pred_flag" set to the flag of the si ice header, Alternatively, one of the time direct mode or the spatial direct mode may be selected in units of slices. Here, Fig. 12 is a schematic diagram showing a method of generating an action vector in the temporal direct mode. In Fig. 12, "P" indicates P image, "B" indicates B image. In addition, the number 0-3 indicates the display order of the image, indicating that it is a display image of time ΤΟ, ΤΙ, T2, T3. The image encoding processing is based on P0, The order of P3, Bl, and B2 is performed. For example, assume a case where the tile set MB1 among the images B2 is encoded in the temporal direct mode. At this time, in 4 322764 201143452 of the encoded image located behind the time axis of the image B2, the motion vector using the image P3 closest to the image B2 is located in the same spatial position as the tile set MB1. The action vector MV of the tile set MB2. The motion vector MV is the reference image P0, and the motion vectors MVL0 and MVL1 used when the tile set MB1 is compiled are obtained by the following equation (1). MVLO = MVL\ II T2-T0 Γ3-ΓΟ Γ2-Γ3 T3-T0

xMV χλ/V (1) 但,於AVC/H. 264中係規定,當圖像Ρ3的圖塊集ΜΒ2 被分割成2個以上的矩形圖塊時,如第13圖所示,將使用 包含圖塊集ΜΒ2之左上像素的矩形圖塊(block)之動作向 量作為動作向量MV。 (先前技術文獻) (非專利文獻) [非專利文獻 1] MPEG-4 AVCCIS0/IEC 14496-10)/ ITU-T H· 264 規格 【發明内容】 (發明所欲解決的課題) 由於習知技術之影像編碼裝置係採用如上所述之構 成,故恆常使用包含圖塊集MB2之左上像素的矩形圖塊之 動作向篁作為動作向量MV。因此,當圖像P3的團塊集MB2 被分=為2個以上之矩形圖塊時,所使用的動作向量MV並 不一疋與編碼對象之團塊集MB1展現出空間上相同的像素 位置之動作’而存有時間直接向量之精度下降,編碼效率xMV χλ/V (1) However, in AVC/H.264, when the tile set ΜΒ2 of the image Ρ3 is divided into two or more rectangular tiles, as shown in Fig. 13, the inclusion is included. The motion vector of the rectangular block of the upper left pixel of the tile set ΜΒ2 is used as the motion vector MV. (Prior Art Document) (Non-Patent Document) [Non-Patent Document 1] MPEG-4 AVCCIS0/IEC 14496-10) / ITU-T H·264 Specifications [Disclosure] (Problems to be Solved by the Invention) Due to the conventional technology Since the video encoding apparatus is configured as described above, the motion vector MV of the rectangular block including the upper left pixel of the tile set MB2 is constantly used. Therefore, when the cluster set MB2 of the image P3 is divided into two or more rectangular tiles, the motion vector MV used does not exhibit spatially the same pixel position as the cluster set MB1 of the encoding object. The action 'has the time precision of the direct vector, the coding efficiency

322764 S 201143452 惡化的問題。 本發明係為了鯉、玉L、 可以產生接近―對裹述課題而研發者,其目的為獲得 提高編碼效率的影像2塊集之動作的時間直接向量, 另外,本”tr 轉編碼方法。 手段產生時間向量7為獲讀由以與編竭裝置相同的 影像解碼裝置及影像解:::碼效率高的編碼串流解碼的 (解決課題的手段) 本發明之影像編瑪裝置,係 成時間性上位於編 、 11里產生手段從構 中,特定出與一 圖塊附近的已編像之圖塊 塊,並由位於上述圖塊象:灿 而產生時間直接的複數個分割區域之動作向量 而產生時間直接模式之時間直接向量。 (發明效果) 依據本發明,由於 時間性上位於編碼對象肖量產生手段係構成為從構成 特定出與編碼對象圖近的已編碼圖像之圖塊中’ 由位於上述圖塊中心的^間上位於相同位置的圖塊,並 時間直接模式之時間直;:=域之動作向量而產生 碼對象之圖塊動作的時、;故可產生接近於編 果,即具有可提昇蝙$ =接模式之時間直接向量,其結 【實施方式】 對象圖塊之編竭效率的效果。 以下,參照圖式, (^1 W(«) 心說明本發明之實施形態。 322764 6 201143452 " 於此第1實施形態中’說明以藉由將構成輸入影像的 影像圖框(frame)均等分割為16x16像素之矩形區域(圖塊 集)為單位,進行封閉於圖框内之編碼的影像編碼裝置,以 及對應於該影像編碼裝置的影像解碼裝置。 另外,於本第1實施形態中,影像編瑪裝置以及影像 解碼裝置係以使用於AVC/H. 264規格中採用的編鳴方式為 例進行說明。 ^ 但,於本第1實施形態中,影像編竭褒置及影像解碼 裝置於使用直接模式時’雖設為使用時間直接模气,伸亦 可視需要而使用空間直接模式。 一 第1圖為示有本發明第1實施形態之影像編碼裝 構成圖。 於第1圖中,動作補償預測部i係從錯存於圖框記憶 體8的1圖框以上之動作補償預測用參照影像中選擇^ ^ 框的參照影像,以構成輸入影像的圖塊集(或將圖塊集分割 而成的副(sub)圖塊集)為單位而執行動作補償預測處理, 產生該圖塊集(編碼對象之圖塊集)的動作向量而產 影像,分別於每個圖塊執行將所選擇的參照圖像之識= 號、動作向量、以及預測影像等輸出的處理。 β 、 但,在此為了說明的方便,係假設以圖塊集為單 生動作向量而產生預測影像。 亦即,動作補償預測部1係安裝有將構成 或將圖塊集分割成的副圖塊集)之動作向量儲 動作向1記憶體u(參照第2圖),時間直接模式時,係對 322764 7 201143452 於構成輸入影像的每個圖塊集 間附近的已編蝎圖像的圖塊集,從構成位於該圖塊集的時 於空間上相同位置的圖塊集,、,中,特定出與該圖塊集位 例如4個®塊(分割區域)之置於該圖塊集之中心的 憶體11的動作向量)中產生肖置(儲存於動作向量記 量,且藉由使用該時間直接^直接模式的時間直接^ 理,而實施產生預測影像的處理里而實施動作補償預剛處 減法器2係算出藉由動 : 影像與輸人影像之差分影像=制部1而產生的預測 測差分訊號輸出至編碼模式列定:將表示其差分影像的預 編碼模式判定部3係評價從減 訊號之預測效率,而於從減法器2 2輪出的預測差分 預測差分訊號之中,選擇預測效率=出的至少1以上的 並於動作補償預測部i將有關該預預測差分訊號, 的產生所使用的動作向量、圖塊集類,〗分訊號的預測影像 如在該圖塊集中所使用的編碼模式為圖塊集類型(例 直接模式的任一者之資訊)及參^影表示框間模式或 碼模式資訊^輸出至可變長度編焉部9: 編號作為編 率最高的預測差分信號輸出至壓縮二4的施將預測效 壓縮部4係藉由實施對於從編碼模式判^ 的預測差分訊號的DCT(離散餘弦(c〇sine) 。3所輪出 出DCT係數,並且將該DCT係數量子化,實,)處理而算 後的DCT係數之>1縮資料(量子化係數)輪a,置子化 5以及可變長度編碣部9的處理。 局。卩解碼部 322764 8 201143452 又,減法H 2、編碼模式判定部3、以及壓 成量子化手段。 *你構 局部解碼部5係將從壓縮部4所輸出的壓縮資料 子化而求出DCT係數,藉由實施對於該DCT係數的反 離散餘弦變換)處理’而實施算出相當於從編碼模式判 3所輸出的制差分訊號的賴誤差訊號的處理。° 加法器6係藉由將由局部解碼部5所算出 訊號與動作補償預測部1所產生的表示預測影像之預^ 號之處理。 “局部解郷㈣局我碼影像訊 迴圈遽波器(i〇Qp filter)7係將從加法器6輪 邰解碼衫像錢中所包含的編碼失真予以補償,實施將補 償編碼失真後的局部解郷像訊號所表示的局部解碼影 作為參照影像而向圖框記憶體8輸出的處理。 ’ 圖框記憶體8為儲存從迴圈遽波器7所輸出的參昭馬 像的RAM(random access memory,隨機存取記憶體)等 錄媒體。 吞己 可變長度編碼部9係將從壓縮部4所輸出的 以及從動作補償預測部1所輸出的編碼模式資訊(圖塊隼 類型/副圖塊集類型、動作向量、參照影像之識別編綱 (entr〇Py)編碼,產生表示該編碼結果的位元串流(編^ 料)’實施將該位元串流(bit stream)輸出的處理。.、、 又,可變長度編碼部9係構成可變長度編碼手段。 i_對於動作向量資訊,亦可直接地進行編碼,例如, 322764 9 201143452 亦可採用如H. 264/AVC的方式,使用已編碼圖塊集之動作 向量而產生預測向量,且將其與預測向量間的差分予以編 碼0 當預測非矩形圖塊之動作向量時,例如,亦可考慮如 第11圖之方法。於第11圖中,箭號係表示利用於預測向 量之導出的週邊之動作向量。 以〇所圍起的3條動作向量所指示的分割區域之預測 向量,係由以〇所圍起的3條動作向量之中央值(median) 而求得。 第2圖為表示本發明第1實施形態的影像編碼裝置之 動作補償預測部1的構成圖。 於第2圖中,動作向量記憶體11係儲存有構成已編碼 圖像的圖塊集(或將圖塊集分割而成的副圖塊集)之動作向 量的例如RAM等紀錄媒體。 動作向量探索部12若接收到表示編碼化模式為框間 模式(inter mode)的資訊(例如,從外部接收表示使用框間 模式的資訊),則以框間模式探索最適合的動作向量,實施 將該動作向量向動作補償處理部14以及動作向量記憶體 11輸出的處理。 時間直接向量產生部13若接收到表示編碼模式為時 間直接模式的資訊,則對於每個編碼對象之圖塊集,從構 成位於該圖塊集之時間附近的已編碼圖像的圖塊集之中, 特定出與該圖塊集位於空間上相同位置的圖塊集,且從位 置於該圖塊集之中心的例如4個圖塊(分割區域)之動作向 10 322764 201143452 楔式的直作接向向量量記.:1^1的動作向量)產生時間直接 ^實施輪h動料動作向量 處理。 及動作向罝記憶體11的 段。又,時間直接向量產生部13係構成直接向量產生手 間作向量探—時 憶艘8的1圖框之參照影像,而實存於圖插記 藉此而實施產生預測影像的處動=:測處理’ 係構成預測影像產生手段。 x動作補仏處理部14 構成=。3圖為表林發明第1實施形態之影像解碼裝置的 解:’可變長度解碼部21係將從第1圖之影像 … 斤輪出的位元串流(編碼資料)輪入,W 流中將壓縮資料(量子化係數 模’該位元串 類型/副圖塊集類型、動作向量、參圖塊集 解碼’將該*縮資料向預測誤差解二;象輪之號)滴 該编瑪模式資訊向動作補償賴部23輸_^且^ 變長度解辦21係構成可變長度解碼手段的處理。又’可 預測#差解碼部22係將從可變長度 ==子化而求出DCT係數,藉由實施對= '、反DCT處理,而實施算出表示差分影像 訊號(相當於從第1圖之編碼模式判定部3:=: 322764 11 201143452 分訊號的預測誤差訊號)的處理β 係構成反量子化手段_。 又,預測誤差解碼部322764 S 201143452 The problem of deterioration. The present invention has been developed for the purpose of producing a close-to-be-wrapped subject, and the object of the present invention is to obtain a temporal direct vector of the operation of the video 2 block set for improving coding efficiency, and the present "tr-transcoding method". The generation time vector 7 is decoded by the same image decoding device and image decoding solution that is more efficient than the editing device. (The means for solving the problem) The image encoding device of the present invention is time-based. Sexually located in the composition, 11 in the means of construction, specific to a tiled block near the tile, and the action vector of the plurality of segmentation regions directly at the time of the tile image: The time direct vector of the time direct mode is generated. (Effect of the Invention) According to the present invention, the means for generating the object to be encoded temporally is configured to be from the block constituting the coded image which is specific to the image to be encoded. ' A tile located at the same position between the centers of the above-mentioned tiles, and the time of the direct mode is straight; := the motion vector of the domain generates the tile action of the code object Time, so it can produce close to the editing result, that is, it has the time to increase the bat $ = connected mode direct vector, and its effect [embodiment] The effect of the editing efficiency of the object block. Hereinafter, refer to the figure, (^1 An embodiment of the present invention will be described with reference to W («). 322764 6 201143452 " In the first embodiment, a description will be given of a rectangular region in which an image frame constituting an input image is equally divided into 16x16 pixels (Fig. A video encoding device that performs encoding in a frame and a video decoding device corresponding to the video encoding device in units of blocks. In addition, in the first embodiment, the video encoding device and the video decoding device are The syllabic method used in the AVC/H.264 specification will be described as an example. However, in the first embodiment, the video editing device and the video decoding device are used when the direct mode is used. The direct mode is used for the time, and the spatial direct mode can be used as needed. Fig. 1 is a view showing a configuration of a video encoding device according to the first embodiment of the present invention. In Fig. 1, the motion compensation predicting unit i Selecting a reference image of the ^ ^ frame from the motion compensation prediction reference image that is misplaced in the frame of the frame memory 8 to form a tile set of the input image (or a pair of the tile set) (sub) tile set) performs motion compensation prediction processing for each unit, generates an action vector of the tile set (a tile set of the encoding target), and generates an image, and executes the selected reference image for each tile. The processing of the output such as the number, the motion vector, and the predicted image. β , However, for the convenience of explanation, it is assumed that the predicted image is generated by using the tile set as a single motion vector. That is, the motion compensation prediction unit In the 1st system, the motion vector storage operation of the sub-block set that constitutes or divides the tile set is attached to the 1 memory u (see FIG. 2), and in the time direct mode, the pair 322764 7 201143452 constitutes the input image. a set of tiles of the edited image near each tile set, from the set of tiles constituting the same position in space at the time of the tile set, and, specifically, the tile set For example 4 ® blocks (segmented areas) A motion vector is generated in the motion vector of the memory layer 11 placed at the center of the tile set (the motion vector vector is stored, and the processing for generating the predicted image is performed by directly controlling the time of the direct mode using the time) The motion compensation pre-fighting subtractor 2 is used to calculate the difference: the difference between the image and the input image = the predicted difference signal generated by the system 1 is output to the coding mode: the pre-representation of the differential image is indicated The coding mode determination unit 3 evaluates the prediction efficiency from the subtraction signal, and selects at least one of the prediction efficiency = out of the prediction difference prediction difference signal that is rotated from the subtractor 2 2 in the operation compensation prediction unit i. The motion vector used for the generation of the pre-predicted differential signal, the tile set class, and the predicted image of the divided signal, such as the coding mode used in the tile set, is the tile set type (for example, any of the direct modes) Information) and the parameter indicates the inter-frame mode or code mode information ^ output to the variable length editing part 9: The number is the highest predicted prediction difference signal output to the compression of the second prediction 4 embodiment by a compression system for the determination of the prediction difference signal from the ^ coding mode DCT (Discrete Cosine (c〇sine). 3 rounds out the DCT coefficient, and quantizes the DCT coefficient, and then calculates the calculated DCT coefficient >1 reduction data (quantization coefficient) round a, set 5 and variable length coding The processing of the section 9. Bureau.卩 Decoding unit 322764 8 201143452 Further, subtraction H 2, coding mode determination unit 3, and compression quantization means. * The local decoding unit 5 determines that the DCT coefficients are obtained by subdividing the compressed data output from the compression unit 4 and performing inverse discrete cosine transform processing on the DCT coefficients. The processing of the output error signals of the three output differential signals. The adder 6 is a process of expressing the predicted image of the predicted video generated by the signal calculated by the local decoding unit 5 and the motion compensation predicting unit 1. "Partial solution (4) Bureau I code video loop chopper (i〇Qp filter) 7 system will compensate the encoding distortion contained in the adder 6 rim decoding shirt image money, implement the distortion of the compensation code The partial decoded image indicated by the partial image signal is output as a reference image to the frame memory 8. The frame memory 8 is a RAM for storing the reference horse image output from the loop chopper 7 ( A recording medium such as a random access memory or a random access memory. The encoding mode information output from the compression unit 4 and output from the motion compensation prediction unit 1 (block type/ The sub-block set type, the motion vector, and the reference image identification (entr〇Py) code, and generate a bit stream (program) indicating the result of the encoding to implement the bit stream output. The processing of the variable length coding unit 9 constitutes a variable length coding means. i_ can also directly encode the motion vector information. For example, 322764 9 201143452 can also be used as H.264/AVC. Way to use a coded tile set The motion vector generates a prediction vector and encodes the difference between the prediction vector and the prediction vector. When predicting the motion vector of the non-rectangular tile, for example, the method as shown in Fig. 11 can also be considered. In Fig. 11, the arrow The number indicates the motion vector used for the derivation of the prediction vector. The prediction vector of the segmentation region indicated by the three motion vectors enclosed by 〇 is the central value of the three motion vectors enclosed by 〇 ( Fig. 2 is a view showing the configuration of the operation compensation predicting unit 1 of the video encoding apparatus according to the first embodiment of the present invention. In Fig. 2, the motion vector memory 11 stores the encoded image. The recording vector of the motion vector of the tile set (or the sub-block set obtained by dividing the tile set) is, for example, a recording medium such as a RAM. The motion vector search unit 12 receives the inter-frame mode (inter mode) when the coding mode is received. Information (for example, receiving information indicating the use of the inter-frame mode from the outside), and searching for the most suitable motion vector in the inter-frame mode, and performing the motion vector to the motion compensation processing unit 14 and the motion vector memory When the time direct vector generation unit 13 receives the information indicating that the encoding mode is the time direct mode, the encoded image of the block set of each encoding target is formed near the time of the set of the tile set. Among the set of tiles, a set of tiles that are spatially identical to the set of tiles is specified, and actions from, for example, four tiles (partitioned regions) located at the center of the tile set are directed to 10 322764 201143452 The wedge-shaped straight-forward vector vector.: 1^1 motion vector) generates the time directly to implement the round-h motion vector processing and the motion to the memory 11. Further, the time direct vector generation unit 13 The direct vector is used to generate the reference image of the 1 frame of the ship 8 and the actual image is inserted into the image to generate the predicted image. . The x motion compensation processing unit 14 has a configuration =. 3 is a solution of the video decoding device according to the first embodiment of the invention, in which the variable length decoding unit 21 rotates the bit stream (encoded data) that is output from the image of Fig. 1 . The compressed data (quantization coefficient module 'the bit string type / sub-block set type, motion vector, reference block set decoding 'the * shrink data to the prediction error solution two; the image of the wheel) The horse mode information is processed by the motion compensation compensation unit 23 to form a variable length decoding means. Further, the 'predictable# difference decoding unit 22 obtains the DCT coefficients from the variable length == sub-ization, and performs the calculation of the differential image signal by performing the pairing of '' and the inverse DCT processing (equivalent to the first image). The coding mode determination unit 3:=: 322764 11 201143452 The prediction error signal of the sub-signal) constitutes an inverse quantization means_. Prediction error decoding unit

動作補償_邹23係減存於圖框記憶體2 框以上之參照影像中讀取從可變長度解碼部21 圖 識別編號表*的參考雜,且從可變長度解如=的 的圖塊集類型/副圖塊集類型係表示使用框間模二出 時,藉由使用從可變長度解碼部21所輸出的動作涵 述參照影像而實施動作補償預測處理,藉此實扩^ 〇上 影像的處理。 &產生預測 另 万面,S從可變長度解碣部21所輸出 型/副圖塊集類型係表示使用直接模式的意涵時,藉 1圖之影像編碼|置的動作補償預測部i相同地^生^ 直接向量’並使㈣時間直接向量與上述參昭奋B 動作補償刪處理,祕實麵生關影_1。 加法器24係將由動作補償預測部23所產生的 22 差刀衫像相加,而實施產生表示相當於從第i :: = r輸出的局部解碼影像的解‘解碼 迴圈遽波器25係將由加法器24所產生的解竭影像訊 號中所包含的編碼失真Μ補償,並將補償編碼失真後的 解碼影像城所衫_碼料作為傾f彡像而儲存於圖 棍記憶體26 ’並且實鱗該解碼影像向外部輸出的處理。 又’加法器24及迴圈攄波器25係構成影像加法手段。 322764 12 201143452 圖框錢'體26係儲存從迴圈丨慮波器25輸出的參昭與 像的RAM等紀錄媒體。 蚤“、、〜 第4圖為 巧表不本發明第1實施形態之影像解碼裝 動作補償預測部23的構成圖。 裝置之 於第4圖中,叙於a 動作向1記憶體31係將構成已解m圄德 的圖塊集(或將圖塊隹八幻丄 鮮碼圖像 的例如RAM等紀錄媒體。 里保孖 時間直接向量產生部犯係於當從可變長 輸出的圖塊集類型/副圍说上 1 i/Μ圖塊類型係表示使用直接模式 涵時,對於每個為解碼對 . ^ Τ象的圖塊集,從構成位於該圖塊 =:::::=_中,與該圖塊 丨J位置的圖塊集,從位置於該圖塊隼 中心的例如4個圖堍r八 國塊(分割區域)之動作向量(儲 向量記憶體31的動作& θ^ ^ θ 勒作向篁)產生時間直接模式的時間直接 向置’將該時間直接向旦从从Motion compensation_Zou 23 is stored in the reference image above the frame memory 2 frame and reads the reference miscellaneous portion from the variable length decoding unit 21, and the block from the variable length such as = The set type/sub-block set type indicates that when the inter-frame mode is used, the motion compensation prediction process is performed by using the operation output from the variable length decoding unit 21 to describe the reference picture, thereby realizing the expansion. Image processing. & generates a prediction, and the output type/sub-block set type of the S variable from the variable length decoding unit 21 indicates that the direct mode is used, and the motion compensation prediction unit i is set by the image coding of the image. The same ^ ^ ^ directly vector ' and (4) time direct vector and the above-mentioned Shen Zhao F B action compensation deletion processing, secret face ying _1. The adder 24 adds the 22-difference knife image generated by the motion compensation prediction unit 23, and generates a solution 'decoding loop chopper 25 indicating a local decoded image equivalent to the output from the i: ::= r. The coding distortion Μ contained in the decommissioned video signal generated by the adder 24 is compensated, and the decoded image decoded by the compensation code is stored as a dump image and stored in the stick memory 26' and Real scale The processing of decoding the image to the outside. Further, the adder 24 and the loop chopper 25 constitute an image adding means. 322764 12 201143452 The frame money 'body 26 series stores the recording medium such as the RAM and the image output from the loop filter. 、", 〜4 Figure 4 is a configuration diagram of the video decoding device operation compensating prediction unit 23 according to the first embodiment of the present invention. The device is shown in Fig. 4, and the operation is performed in the memory 31. A set of tiles constituting the decoded m 圄 ( (or a recording medium such as RAM, which is a block 隹 丄 丄 。 。 。 。 。 。 。 。 。 。 。 。 。 。 。 。 。 。 。 。 。 直接 直接 直接 直接 直接 直接 直接 直接 直接 直接Set type/sub-circumference on 1 i/Μ tile type indicates that when using direct mode culvert, for each block set that is a decoding pair. ^ Τ, from the block located in the block =:::::= In the _, the tile set at the position of the tile 丨J, from the motion vector of the four 堍r 八八块 (divided region) located at the center of the tile ( (the action of the memory vector memory 31 & θ ^ ^ θ 勒 篁 篁 产生 产生 产生 产生 产生 产生 产生 产生 产生 产生 产生 产生 产生 产生 产生 产生 产生 产生 产生 产生 产生 产生

〇里作為動作向量,實施對於動作補 償處理部33以及動作& β & F 功作向篁記憶體31輸出的處理。 又’時間直接向量產生部Μ係構成直接向量產生手 段。 、動作補償處理部33係從儲存於圖框記憶體別的i圖 參4像中’讀取從可變長度解碼部21所輸出的 識別編號表示的參昭旦: “、、〜像,且圖塊集類型/副團塊集類梨表 Μ “ 函時’係使用從可變長度解鳴部21所 輸出的動作向量與上流会 L ^ . 述參照影像,實施動作補償預測處理In the case of the motion vector, the motion compensation processing unit 33 and the motion & β & F function are output to the memory 31. Further, the time direct vector generation unit constitutes a direct vector generation means. The motion compensation processing unit 33 reads the reference code indicated by the identification number output from the variable length decoding unit 21 from the i-picture 4 image stored in the frame memory: ", ~, and The tile set type/sub-block set type Pear Table Μ "Letter" uses the motion vector output from the variable length damper unit 21 and the upstream stream L ^ .

而藉此產生預測影像;I w豕’田圖塊集類型/副團塊集類型表示使 13 322764 201143452 用直接模式的意涵時’係使用從時間直接向量產生部32所 輸出的動作向量與上述參照影像實施動作補償預測處理而 藉此實施產生預測影像的處理。又,動作補償處理部33係 構成預測影像產生手段。 ' 於第1圖中,雖假設為影像編碼裝置之構成要件的動 作補償預剛部1、減法器2、編碼模式判定部3、壓縮部4、 局部解碼部5、加法器6、迴圈濾波器7、以及可變長度鴆 譌部9係各自由專用的硬體(例如,安裝有CPU的半導體積 醴電路’或單晶片微電腦等)構成,但當影像編碼裝置由電 腦構成時’亦可採用於該電腦之記憶體儲存記述有動作補 償頓測部1、減法器2、編碼模式判定部3、壓縮部4、局 部解碼部5、加法器6、迴圈濾波器7、以及可變長度編碼 部9之處理内容的程式,並由該電腦之cpu執行儲存於該 記憶體的程式的方式。 第5圖為表示本發明第1實施形態之影像編碼裝置的 處理内容的流程圖。 於第3圖中,雖假設為影像解碼裝置之構成要件的可 變長度解碼部21、預測誤差解碼部22、動作補償預測部 23、加法部24、以及迴圈濾波器25係各自由專用的硬體(例 如,安裝有CPU的半導體積體電路,或單晶片微電腦等) 構成’但^影像編碼裝置由電腦構成時,亦可採用於該電 腦之記憶體儲存記述有可變長度解碼部21、預測誤差解碼 邡22、動作補償預測部23、加法部24、以及迴圈濾波器 25之處理内容的程式,由該電腦之CPU執行儲存於該記憶 322764 14 201143452 體的程式的方式。 第6圖為表示本發明第1實施形態之影像解碼裝置的 處理内容的流程圖。 接著,對於動作進行說明。 首先,說明第1圖之影像編碼裝置的處理内容。 動作補償預測部1當輸入表示輸入影像的動畫影像訊 號時,係將該動晝影像訊號之各圖框分割為圖塊集單位(或 副圖塊集單位)。 動作補償預測部1係於當將動晝影像訊號分割為圖塊 集單位(或副圖塊集單位)時,從儲存於圖框記憶體8的1 圖框以上之動作補償預測用參照影像中選擇1圖框的參照 影像,以圖塊集為單位(或副圖塊集為單位)對於每個色成 分執行動作補償預測處理,產生編碼對象的圖塊集(或副圖 塊集)之動作向量而產生預測影像。 動作補償預測部1係於當產生編碼對象之圖塊集(或 副圖塊集)之動作向量而產生預測影像時,則將該預測影像 向減法器2輸出,並且將使用於該預測影像之產生的動作 向量、圖塊集類型/副圖塊集類型(例如,含有表示於該圖 塊集(或副圖塊集)中所使用的編碼模式為框間模式或直接 模式之任一者的資訊)、以及參照影像的識別編號向編碼模 式判定部3輸出。 以下,具體說明動作補償預測部1之處理内容。 但,在此,為了說明上的方便,係假設以圖塊集為單 位產生動作向量而產生預測影像。 15 322764 201143452 全動作補偽預測部1之動作向量探索部12係當接收表示 、爲碼模式為框間模式的意涵之資訊(例如,從 使用框間模式㈣涵之資訊)時(第5圖之步驟ST1),^ =模式探索4最適當的動作向 量,將該動作向量向動作補 償處Z 14輸出(步驟ST2)。 n 模式中探索最適當之動作向量的處理本身係為 %知技術’故在此省略其詳細說明。 動^補償預測部1之時間直接向量產生部㈣ 接模式的意涵之資訊一), 間上附近的已:圖塊集,係從構成位於該圖塊集之時 在空間上^ 1圖像的圖塊集之中,特定出與該圖塊集 工 立於相同位置的圖塊集(步驟ST3)。 ^ p,., ^田編蝎對象之圖塊集如第12圖所示為圖像B2 已編圓MB1時’從構成時間上位於圖塊集MB1之附近的 位;π像P3的圖塊集中,特定出與圖塊集腦在空間上 位於相同位置的圖塊集職。 時間直接向旦 集顧在空間生部13係於當將與編碼對象之圖塊 j 於相同位置的圖塊集MB2特定時,如第 之中,^從儲存於動作向量記憶體11的已編碼動作向量 純於該圖塊集觀之中心的4個矩形圖塊(分割 叫之動作向量MVi(i叫,2,3,4)(步驟⑽。 Λ μ/第7圖之例中’係將左上之矩形圖塊之動作向量設 為,右下之矩形圖塊之動作向量設為·,依照光栅掃 為(raster scan)之順序分配編號。 322764 16 201143452 時間直接向量產生部13當取得4個矩形圖塊之動作向 量MVi(i = 1,2,3, 4)時,則如下述式(2)所示,藉由求出4 個動作向量MVi的加法平均而產生編碼對象之圖塊集MB1 的時間直接模式之時間直接向量MVdiwt(步驟ST5)。Thereby, the predicted image is generated; the I w豕 'field tile set type/sub-block set type representation makes 13 322764 201143452 use the direct mode meaning time' to use the action vector output from the time direct vector generation unit 32 The reference image is subjected to motion compensation prediction processing, thereby performing processing for generating a predicted image. Further, the motion compensation processing unit 33 constitutes a predictive video generating means. In the first drawing, the operation compensation pre-compensation unit 1, the subtractor 2, the encoding mode determination unit 3, the compression unit 4, the local decoding unit 5, the adder 6, and the loop filter are assumed to be constituent elements of the video encoding device. The device 7 and the variable length 9 9 are each composed of a dedicated hardware (for example, a semiconductor semiconductor circuit with a CPU mounted or a single-chip microcomputer), but when the image coding device is composed of a computer, The memory storage used in the computer includes a motion compensation detection unit 1, a subtractor 2, an encoding mode determination unit 3, a compression unit 4, a local decoding unit 5, an adder 6, a loop filter 7, and a variable length. The program for processing the content of the encoding unit 9 and the program stored in the memory by the CPU of the computer. Fig. 5 is a flow chart showing the processing contents of the video encoding apparatus according to the first embodiment of the present invention. In the third diagram, the variable length decoding unit 21, the prediction error decoding unit 22, the motion compensation prediction unit 23, the addition unit 24, and the loop filter 25, which are constituent elements of the video decoding device, are each dedicated. The hardware (for example, a semiconductor integrated circuit mounted with a CPU, or a single-chip microcomputer or the like) constitutes a 'however' the image encoding device is constituted by a computer, and the memory storage of the computer may be described as having a variable length decoding unit 21 The program of the prediction error decoding 邡22, the motion compensation prediction unit 23, the addition unit 24, and the loop filter 25 is executed by the CPU of the computer to execute the program stored in the memory 322764 14 201143452. Fig. 6 is a flow chart showing the processing contents of the video decoding device according to the first embodiment of the present invention. Next, the operation will be described. First, the processing contents of the video encoding apparatus of Fig. 1 will be described. When the motion compensation video signal indicating the input image is input, the motion compensation prediction unit 1 divides each frame of the motion picture signal into a tile set unit (or a sub-block set unit). The motion compensation prediction unit 1 is configured to be used in the motion compensation prediction reference image stored in the frame of the frame memory 8 when the motion picture signal is divided into tile set units (or sub-block unit units). Select the reference image of the 1 frame, and perform motion compensation prediction processing for each color component in units of tile sets (or sub-block sets) to generate the action of the tile set (or sub-block set) of the encoding target. The vector produces a predicted image. The motion compensation prediction unit 1 outputs a predicted image to the subtractor 2 when the motion vector of the tile set (or the sub-block set) to be encoded is generated to generate the predicted image, and is used for the predicted image. The generated motion vector, tile set type/sub-block set type (for example, containing the coding mode used in the tile set (or sub-block set) is either an inter-frame mode or a direct mode The information) and the identification number of the reference image are output to the encoding mode determination unit 3. Hereinafter, the processing content of the motion compensation prediction unit 1 will be specifically described. However, here, for convenience of explanation, it is assumed that a motion vector is generated in units of a tile set to generate a predicted image. 15 322764 201143452 The motion vector search unit 12 of the full-motion complementation prediction unit 1 receives information indicating that the code pattern is the inter-frame mode (for example, information from the use of the inter-frame mode (4)) (5th) In step ST1) of the figure, ^ = mode search 4 is the most appropriate motion vector, and the motion vector is output to the motion compensation portion Z 14 (step ST2). The process of exploring the most appropriate motion vector in the n mode is itself a known technique, and thus a detailed description thereof will be omitted herein. The information of the direct vector generation unit of the motion compensation prediction unit 1 (4) The information of the meaning of the connection mode 1), the vicinity of the existing: tile set, which is spatially located when the block set is located Among the tile sets, a tile set that is at the same position as the tile set is specified (step ST3). ^ p,., ^ The tile set of the field editing object is shown in Fig. 12 as the image B2 has been circled MB1 'from the time of the composition block MB1 near the block set MB1; π image P3 block Concentration, specific tile placements that are spatially co-located with the tile set brain. The time is directly assigned to the spatial group 13 in the case where the tile set MB2 at the same position as the tile j of the encoding object is specified, as in the middle, from the encoded code stored in the motion vector memory 11. The motion vector is pure to the four rectangular tiles in the center of the tile set view (the segmentation is called the motion vector MVi (i, 2, 3, 4) (step (10). Λ μ / in the example of Figure 7 The motion vector of the rectangular block on the upper left side is set such that the motion vector of the rectangular block in the lower right is set to ·, and the number is assigned in the order of raster scan. 322764 16 201143452 The time direct vector generation unit 13 obtains four When the motion vector MVi (i = 1, 2, 3, 4) of the rectangular block is as shown in the following formula (2), the set of the coding target is generated by obtaining the addition average of the four motion vectors MVi. The time direct mode of the MB1 direct vector MVdiwt (step ST5).

MVdircct = — (2) 時間直接向量產生部13當產生時間直接模式的時間 直接向量MVdirect時,則將該時間直接向量MVdirect作為動作 向量而向動作補償處理部14以及動作向量記憶體11輸出。 動作補償預測部1之動作補償處理部14係於當編碼模 式為框間模式時,若從動作向量探索部12接收到動作向 量,則使用該動作向量與儲存於圖框記憶體8的1圖框參 照影像而實施動作補償預測處理,藉此產生預測影像(步驟 ST6)。 另一方面,當編碼模式為直接模式時,若從時間直接 向量產生部13接受時間直接向量MVdirm作為動作向量時, 使用該時間直接向量MVdirect與儲存於圖框記憶體8的1圖 框的參照影像實施動作補償預測處理,藉此產生預測影像 (步驟ST6)。 又,動作補償處理部14的動作補償預測處理係周知技 術,故省略詳細說明。 減法器2係於當動作補償預測部1產生預測影像時, 則算出該預測影像與輸入影像之差分影像,將表示該差分 影像的預測差分訊號向編碼模式判定部3輸出(步驟ST7)。 17 322764 201143452 編碼模式判定部3係於每次從減法器2接收預測差分 訊號時,評價該預測差分訊號之預測效率,從由減法器2 輸出的至少1個以上的預測差分訊號中,選擇預測效率最 高的預測差分訊號。 編碼模式判定部3的評價預測差分訊號之預測效率的 處理本身係為習知技術,故在此省略詳細說明。 編碼模式判定部3係於選擇出預測效率最高的預測差 分訊號後,將於動作補償預測部1中含有在有關於該預測 差分訊號的預測影像之產生中所使用的動作向量、圖塊集 類型/副圖塊集類型(例如,含有表示於該圖塊集中使用的 編碼模式為框間模式或直接模式之何一者的資訊)、以及參 照影像之識別編號的編碼模式資訊向可變長度編碼部9輸 出。 另外,編碼模式判定部3係將預測效率最高的預測差 分訊號輸出至壓縮部。 惟,編碼模式判定部3於當編碼模式為框間模式時, 將使用於預測影像之產生的動作向量包含於編碼模式資 訊,而將含有該動作向量的編碼模式資訊向可變長度編碼 部9輸出,但當編碼模式為直接模式時,則不於編碼模式 資訊包含使用於預測影像之產生的動作向量,而將未包含 該動作向量的編碼模式資訊向可變長度編碼部9輸出。 壓縮部4係於從編碼模式判定部3接收預測差分訊號 時,對於該預測差分訊號實施DCT處理,藉此算出DCT係 數,將該DCT係數量子化(步驟ST8)。 18 322764 201143452 壓縮部4係將為量子化後的DCT係數之壓縮資料向局 部解碼部5以及可變長度編碼部9輸出。 ▲局。p解碼5係當從壓縮部4接收到壓縮資料時,將 ,壓縮資料反量子化而求得DCT係數,實施對於該DCT係 :反DCT處理,軌而算出相當於從編碼模式判定部3 輸出的預測差分訊號的_誤差訊號。 6 Μ局部解碼5算出預測誤 :測誤差訊號與表示由動作補償娜"所產生的預: =預測訊號予《加上,藉此產生表示 邛解碼影像訊號。 馬〜像的局 迴圈濾波器7係iπ # τ A 從加法器6所輸出的:人之’碼處理作準備’而將 真補償,且將補償碼影像訊號巾所含有的編媽失 局部解碼喊作為“ 料碼影像訊號所示之 可變長度編物==:記憶體8。 :將該壓縮資料與從動作補償預測部1 :料’則 =ΓΓ圖塊集類型、動作向量(當== ,照4之識別編號)熵編碼,產 流,將該位元串流輪出(步物广亥 ' W第3圖之影像解碼裝置的處理 可變長度解媽部21係當輸入從第!圖之 輸出的位元串列輸人時,從該位元“中將馬裝置 ::模式資訊(圖塊集類型/副圖塊集類型、動作向及 為框龍柄)、參照影像之朗錢)_碼= 322764 19 201143452 誤差解碼部22輸出,將該編竭模式資w 動作補㈣測部23輪出(第6圖之步驟ST11)貝況向 預測誤差解石馬部22當從可變長度解 縮資料時,㈣該壓崎解.^ 21接收到壓 對該DCT係數實施反DCT ^ *>CT係數,且 的預測誤差訊號(相當於從^ ’猎此而异出表示差分影像 的預測差分訊號的_料圖之編碼模式判定部3輪出 動作補償預(步驟ST12)。 照影像之識_號時,從 &長度解碼部21接收到參 上之參照影像中,進行:===,圖框以 另外,動作補償預測部 ;:像的讀取。 收到圖塊集類型/副圖塊集類型;從:=:碼部21接 副圖塊集類型而判別第] 、’、參'、、、該圖塊集類型/ 式或直接模柄為編顯像^;裝置錢用框間模 ^ ra1 ^ ^ ^ ^ 11 ^ f ^ ® ^ ^ ^ ^ ^ ^ 4 ^ ^ 的動作向量與上述參照影像21輪出 此產生預測影像。 也動作補彳員預測處理,轎 另一方面,當第1圖之旦 為編碼模式時,與第丨圖之:‘以置使用直接模式作 部1相同地赵相直接=、、碼裝置的動作補償預蜊 動作補償預測處理二:_時間直接向量實施 動I,二體說明動作補償預測部23之處理内容。 動作補_測部23之時間直接向量產生二從 322764 20 201143452 可變長度解碼部21輸出的圖塊集類型/副圖塊集類型係表 示使用直接模式的意涵時,對於每個解碼對象之圖塊集, 從構成位於該圖塊集之時間上附近的已解碼圖像的圖塊集 之中,特定出在空間上與該圖塊集位於相同位置的圖塊集 (步驟 ST14)。 例如,當解碼對象之圖塊集如第12圖所示為圖像B2 之圖塊集MB1時,係從構成時間上位於圖塊集MB1附近的 已解碼圖像P3的圖塊集中,特定出與圖塊集MB1在空間上 位於相同位置之圖塊集MB2。 時間直接向量產生部32若特定出與解碼對象之圖塊 集MB1位於空間上相同位置的圖塊集MB2,則如第7圖所 示,從儲存於動作向量記憶體31的已解碼之動作向量中, 取得位於該圖塊集MB2之中心的4個矩形圖塊(分割區域) 之動作向量1^丨(丨=1,2,3,4)(步驟81'15)。 時間直接向量產生部3 2若取得4個矩形圖塊之動作向 量MVi(i = 1, 2, 3, 4),則如上述之式(2)所示,藉由求得4 個動作向量MVi之加法平均,而產生解碼對象的圖塊集MB1 的時間直接模式之時間直接向量MVdirect(步驟ST16)。 時間直接向量產生部32係於當產生時間直接模式之 時間直接向量MVdirect時,以該時間直接向量MVdirect作為動 作向量,而向動作補償處理部33以及動作向量記憶體31 輸出。 動作補償預測部23之動作補償處理部33當從可變長 度解碼部21接收到參照影像之識別編號時,則進行從儲存 21 322764 201143452 於圖框記憶體26的1圖框以上之參照影像中讀取該識 號所示之參照影像。 、扁 動作補償處理部33係於當從可變長度解碼部21所輪 出的圖塊集類型/副圖塊集類型係表示使用框間模式的二 涵時,使用從可變長度解碼部21輸出的動作向量與上述^ 照影像實施動作補償預測處理藉此而產生預測影二 ST17) 〇 (步顿 另一方面,當從可變長度解碼部21所輸出的圖塊 型/副圖塊集類型係表示使用直接模式的意涵時,若從時颉 =接向量產生部32接收到時間直接向量MVdi⑽作為動^間 量則使用§亥時間直接向量MVdirect與上述參照影彳象實扩° 作補償預測處理而藉此產生預測影像(步驟ST17)。知動 知技:T動作補償處理部33之動作補償預測處理為習 技術,故在此省略詳細說明。 加法器24係於當動作補償預測部23產生出 轉航韻誤差科部22 與後差》影像相加’而產生表示相當於從第1圖之 I:碼ΐ置之加法器6所輸出的局部解碼影像的解碼影 的解蝎影像訊號(步驟ST18)。 私 迴圈濾波器25係於當加法器24產生出缺版办 時’補償於該解碼f彡像減巾所含有 t〜減 ,失真後的解碼影像訊號所表示 :=:)圖框記憶體26 ’並且將該解瑪影像向外部輸MVdircct = - (2) When the time direct vector generation unit 13 generates the time direct vector MVdirect, the time direct vector MVdirect is output as the motion vector to the motion compensation processing unit 14 and the motion vector memory 11. The operation compensation processing unit 14 of the motion compensation prediction unit 1 is configured to use the motion vector and the image stored in the frame memory 8 when the motion vector is received from the motion vector search unit 12 when the coding mode is the inter-frame mode. The frame performs motion compensation prediction processing with reference to the video, thereby generating a predicted image (step ST6). On the other hand, when the encoding mode is the direct mode, when the time direct vector generating unit 13 receives the time direct vector MVdirm as the motion vector, the time direct vector MVdirect and the reference of the 1 frame stored in the frame memory 8 are used. The image performs motion compensation prediction processing, thereby generating a predicted image (step ST6). Further, since the motion compensation prediction processing of the motion compensation processing unit 14 is well known, the detailed description thereof will be omitted. When the motion compensation prediction unit 1 generates a predicted image, the subtracter 2 calculates a difference image between the predicted image and the input image, and outputs a predicted difference signal indicating the difference image to the encoding mode determining unit 3 (step ST7). 17 322764 201143452 The coding mode determination unit 3 evaluates the prediction efficiency of the prediction difference signal every time the prediction difference signal is received from the subtractor 2, and selects the prediction from at least one or more prediction difference signals output from the subtractor 2. The most efficient predictive differential signal. The process of evaluating the prediction efficiency of the differential signal by the coding mode determination unit 3 is a conventional technique, and thus detailed description thereof will be omitted. The coding mode determination unit 3 selects the action vector and the tile set type used in the generation of the predicted image with respect to the predicted difference signal, after selecting the prediction difference signal having the highest prediction efficiency. /Sub-block set type (for example, information indicating whether the encoding mode used in the tile set is the inter-frame mode or the direct mode), and the encoding mode information of the reference image identification number to the variable length encoding Part 9 output. Further, the coding mode determination unit 3 outputs the prediction difference signal having the highest prediction efficiency to the compression unit. However, when the encoding mode is the inter-frame mode, the encoding mode determining unit 3 includes the operation vector for generating the predicted image in the encoding mode information, and the encoding mode information including the motion vector is transmitted to the variable length encoding unit 9. When the encoding mode is the direct mode, the encoding mode information is not included in the encoding mode information, and the encoding mode information not including the motion vector is output to the variable length encoding unit 9. When receiving the prediction difference signal from the coding mode determination unit 3, the compression unit 4 performs DCT processing on the prediction difference signal, thereby calculating the DCT coefficient and quantizing the DCT coefficient (step ST8). 18 322764 201143452 The compression unit 4 outputs the compressed data of the quantized DCT coefficients to the local decoding unit 5 and the variable length coding unit 9. ▲ Bureau. When the compressed data is received from the compression unit 4, the p-decode 5 is obtained by dequantizing the compressed data to obtain a DCT coefficient, and performing the DCT-based: inverse DCT processing on the track to calculate the output from the coding mode determining unit 3. The _ error signal of the predicted differential signal. 6 Μ Local Decoding 5 Calculate the prediction error: The measurement error signal and the pre-: = prediction signal generated by the motion compensation compensation are added to generate a representation 邛 decoded video signal. The horse's local loop filter 7 is iπ # τ A is output from the adder 6: the person's 'code processing is prepared' and will be compensated, and the compensation code image signal towel contains the missing part of the mother. The decoding is called "variable length editing shown by the material code image signal ==: memory 8. The compression data and the slave motion compensation prediction unit 1: material 'is = ΓΓ tile set type, motion vector (when = = , according to the identification number of 4) Entropy coding, production flow, the bit stream is rounded out (steps of the wide sea' W 3rd image decoding device processing variable length solution Ma Department 21 when input from the first When the bit of the output of the graph is input, the device is: from the bit: the mode information (the tile set type/the sub-block set type, the action direction is the frame handle), and the reference image朗 钱 _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ When the variable length is decompressed, (4) the pressure solution is solved. ^ 21 receives the inverse DCT ^ * > CT coefficient of the DCT coefficient, and the pre- The error signal (corresponding to the coding mode determination unit 3 that outputs the prediction difference signal indicating the differential image from the above-mentioned code), the operation mode compensation unit 3 is rotated (step ST12). The length decoding unit 21 receives the reference video received on the target, and performs: ===, the frame is additionally operated, and the prediction unit is operated; the image is read. The received tile set type/sub-block set type; :=: The code part 21 is connected to the sub-block set type and discriminates the], ', 参', ,, the tile set type/form or the direct mold handle to compose the image ^; the device money uses the inter-module ^ ra1 ^ ^ ^ ^ 11 ^ f ^ ® ^ ^ ^ ^ ^ ^ ^ ^ ^ ^ The action vector and the above reference image 21 round out to produce the predicted image. Also the action complementer predictive processing, the car on the other hand, when the first picture When it is in the coding mode, it is the same as the first diagram: 'The direct use of the direct mode is the same as the first phase, and the motion compensation compensation prediction process 2 of the code device: _ time direct vector implementation I, The processing content of the motion compensation prediction unit 23 will be described in two parts. The time vector of the motion compensation unit 23 is generated from two 322,764 2 . 0 201143452 The tile set type/sub-block set type output by the variable length decoding unit 21 indicates the time when the direct mode is used, and the block set for each decoding object is composed at the time of the tile set. Among the tile sets of the decoded image in the vicinity, a tile set spatially located at the same position as the tile set is specified (step ST14). For example, when the tile set of the decoding target is as shown in FIG. When it is the tile set MB1 of the image B2, it is specified from the tile set of the decoded image P3 which is temporally located near the tile set MB1, and the tile set which is spatially identical to the tile set MB1 is specified. MB2. When the temporal direct vector generation unit 32 specifies the tile set MB2 that is spatially identical to the tile set MB1 to be decoded, as shown in FIG. 7, the decoded motion vector stored in the motion vector memory 31 is decoded. The motion vector 1 丨 (丨 = 1, 2, 3, 4) of the four rectangular tiles (divided regions) located at the center of the tile set MB2 is obtained (step 81 '15). When the time direct vector generation unit 3 2 obtains the motion vector MVi (i = 1, 2, 3, 4) of the four rectangular blocks, as shown by the above equation (2), four motion vectors MVi are obtained. The addition is averaged, and the time direct mode MVdirect of the time direct mode of the tile set MB1 of the decoding target is generated (step ST16). The time direct vector generation unit 32 is connected to the motion compensation processing unit 33 and the motion vector memory 31 when the time direct vector MVdirect is generated as the motion vector. When receiving the identification number of the reference video from the variable length decoding unit 21, the operation compensation processing unit 33 of the motion compensation prediction unit 23 performs the storage of the reference image from the frame 1 of the frame memory 26 stored in 21 322764 201143452. Read the reference image shown by the identification number. The flat motion compensation processing unit 33 uses the slave variable length decoding unit 21 when the tile set type/subblock set type that is rotated from the variable length decoding unit 21 indicates the use of the two culverts of the inter-frame mode. The output motion vector and the motion compensation prediction processing described above generate a prediction shadow ST17). (On the other hand, when the tile type/sub-block set is output from the variable length decoding unit 21 When the type indicates the meaning of using the direct mode, if the time direct vector MVdi(10) is received from the time 接=the vector generation unit 32 as the dynamic amount, the direct vector MVdirect and the reference image are used as the dynamic expansion. The predicted image is generated by the compensation prediction process (step ST17). The motion recognition technique: the motion compensation prediction process of the T motion compensation processing unit 33 is a conventional technique, and thus detailed description thereof will be omitted. The adder 24 is used when the motion compensation prediction is performed. The portion 23 generates a translation error image section 22 which adds the image to the posterior difference image and generates a decoded image corresponding to the decoded image of the local decoded image outputted from the adder 6 of the I: code set of Fig. 1 . Image signal Step ST18) The private loop filter 25 is used to compensate for the t-subtraction of the decoded image when the adder 24 generates the missing version, and the decoded image signal after the distortion is represented by: =:) The frame memory 26' and the image is output to the outside

S 322764 22 201143452 二Γ上所述:依據本第1實施形態,時間直接向量產 、斤的Ρ絶係構成為從構成時間上位於編碼對象之圖塊集附 位於Γ像之圖塊集中,特定出與編碼對象之圖塊集 “固矩形圖塊之動作向=拉從位於該圖塊集之中心的 詈,故可本, 里產生時間直接模式之時間直接向 接模式之時間直碼::之圖塊集動作向量的時間直 象之圖r碼效率的影二可提昇編瑪對 係構成為從構成時L ㈣,時間直接向量產生部32 ^ @ ^ 、 立於解碼對象圖塊集之附近的已解 :=塊集中’特定出軸對象之圖塊集在空間上 矩形圖’並從位於該圖塊集之中心的4個 故可^技作向置產生時間直接模式之時間直接向量, 接近似於解碼對象之圖塊集動作向量的時間直接 圖直接向量’結果’即達到獲得可將解碼對象之 :^ 編碼效率高的影像編碼裝置所已編碼的串流予 以解碼的f彡像解料置的效果。 f外,於本第1實施形態中,雖例示為時間直接向量 ^部13、32係從位於圖塊集中心的4個矩形圖塊之動作 =里ΜΜι - 1,2’ 3, 4)產生時間直接模式的時間直接向量 ’但位於圖塊集中心的圖塊並不限於4個,只要位於 的圖塊為3個以下或5個以上,則亦可從3個 旦上的㈣中產生時間直接模式之時間直接向 M MVdirect ° 322764 23 201143452 另外,於此第1實施形熊 ― 向量作為動作向量,但亦可fτ,雖例示為產生時間直接 (第2實施形態) 用該向量作為預測向量。 於上逑第1實施形態令 部13、32係、藉由求得位於输疏雖例示為時間直接向量產生 心的形方塊之動作(解码對象)之圖塊集中 均’而產切間直接模 1 ^12,3, 4)的加法平 對應於編褐對象(解碼對象)=直接向量,-,但亦可 之圖塊大小,將複數個分自圖塊集内的複數個分割區域 加,藉此而產生時間直接模式加權相 ㈣間直接向量-二 第8圖為不有以對應 m次 間直接向量I之產生方 同地,對於每個係與上述第1實施形態相 上位於該圖塊集之附近的已^對象)圖塊集,從構成時間 集之中,特定出與該圖塊集在已解碼圖像)的圖塊 集。 仗二間上位於相同位置的圖塊 為圖Z當編碼對象(解喝對象)之圖塊集如第12圖所-為圖像B2之圖塊集MM 第圖所不 MB1之附近的已編碼 可從構成時間上位於圖塊集 定出與圖塊集MB1位於1 7解碼圖像P3)的圖塊集中特 時間直接向量產生部^相同位置的圖塊集腿。 (解碼對象)之圖塊集MB1 、^於當特定出與編碼對象 二間上位於相同位置的圖塊集 322764 24 201143452 MB2時,則如第8圖所示,從儲存於動作向量記憶體11、 31的已編碼動作向量(已解瑪動作向量)之中取得位於該圖 塊集MB2之中心的4個圖塊(分割區域)之動作向量MVi(i =1,2,3,4)。 於第8圖之例中,係將左上圖塊之動作向量設為MVi, 右下圖塊之動作向量設為MV4,依照光柵掃描之順序分配編 號。 另外,時間直接向量產生部13、32係取得從圖塊集 MB2之中心至4個圖塊之中心為止的距離di(i = 1,2, 3, 4)。 又,從圖塊集MB2之中心至4個圖塊之中心為止的距 離ώ係與該圖塊之大小成比例。 當時間直接向量產生部13、32取得4個圖塊之動作向 量MVi與從圖塊集MB2之中心至4個圖塊之中心為止的距離 ώ時,則如下述式(3)所示,藉由對應至4個圖塊之中心為 止的距離di而將4個動作向量MVi的加權相加,藉此產生 時間直接模式之時間直接向量MVdirect。 MVdircct = 1 —, (3)S 322764 22 201143452 In the first embodiment, according to the first embodiment, the time direct vector production and the weighting are configured to be attached to the image set of the image from the block set that is temporally located in the encoding target. The set of tiles corresponding to the encoding object "the action direction of the solid rectangular block = pull from the center of the set of the tile set, so the time of the direct mode can be generated directly in the time of the direct connection mode: : The time series of the motion vector of the block set motion vector is the image of the r code efficiency, which can improve the composition of the code, and the composition of the code is L (four), the time direct vector generation unit 32 ^ @ ^, and the set of decoding object blocks. The solution in the vicinity: = the set of blocks in the block set 'specifically the axis object in the space rectangle' and the time from the 4 in the center of the set of tiles to generate the time direct mode directly Vector, which is similar to the time of the block set motion vector of the decoding object. The direct vector direct result 'results' is obtained by decoding the stream encoded by the image encoding apparatus capable of decoding the object: Like the effect of the solution In addition, in the first embodiment, it is exemplified that the time direct vector parts 13 and 32 are generated from the actions of four rectangular blocks located at the center of the tile set = ΜΜι - 1, 2' 3, 4) The time direct vector of the time direct mode 'but the block located at the center of the tile set is not limited to four. As long as the number of tiles located is less than 3 or more than 5, the time can also be generated from the (four) of the three deniers. The time of the direct mode is directly to M MVdirect ° 322764 23 201143452. In addition, although the first embodiment of the bear-vector is used as the motion vector, but fτ can be exemplified as the generation time directly (second embodiment), the vector is used as the prediction vector. In the first embodiment of the first embodiment, the parts 13 and 32 are obtained by directly determining the operation of the block (decoding target) in which the square is generated as a time direct vector. The addition of modulo 1 ^12, 3, 4) corresponds to the browning object (decoding object) = direct vector, -, but also the size of the tile, which is divided into a plurality of divided regions within the tile set. , thereby generating a time direct mode weighted phase (four) The vector-two figure 8 shows that there is no corresponding square of the direct vector I corresponding to the m-th order, and each of the blocks is located near the tile set in the first embodiment. Set, from the set of time sets, the set of tiles that are associated with the tile set in the decoded image. The tile at the same position on the second frame is the image of the coded object (the object to be dehydrated) The block set is as shown in Fig. 12 - the block set MM of the image B2. The coded near the MB1 of the picture can be located at the block set from the block set and the block set MB1 is located at the decoded image P3. The tile set is a set of legs at the same position as the direct vector generation unit. The tile set MB1 of the (decoding target) is stored in the motion vector memory 11 as shown in FIG. 8 when the tile set 322764 24 201143452 MB2 located at the same position as the encoding object 2 is specified. Among the encoded motion vectors (31), the motion vector MVi (i = 1, 2, 3, 4) of the four tiles (divided regions) located at the center of the tile set MB2 is obtained. In the example of Fig. 8, the motion vector of the upper left tile is set to MVi, and the motion vector of the lower right tile is set to MV4, and the numbers are assigned in the order of raster scan. Further, the time direct vector generation units 13 and 32 acquire the distance di (i = 1, 2, 3, 4) from the center of the tile set MB2 to the center of the four tiles. Further, the distance from the center of the tile set MB2 to the center of the four tiles is proportional to the size of the tile. When the time direct vector generation units 13 and 32 acquire the distance ώ from the center of the tile set MB2 to the center of the four tiles, the time vector generation unit 13 and 32 obtain the distance ώ from the center of the tile set MB2 to the center of the four tiles. The weights of the four motion vectors MVi are added by the distance di corresponding to the center of the four tiles, thereby generating the temporal direct vector MVdirect of the temporal direct mode. MVdircct = 1 —, (3)

(4) 由以上可知,依據本第2實施形態,由於時間直接向 量產生部13、32係構成為對應複數個分割區域之圖塊大小 而將複數個分割區域之動作向量MVi加權相加,藉此產生時 25 322764 201143452 :开直ϊΓί之時間直接向量MVdi—故相較於上述第i實 MV加;抑位於圖塊集之中心的4個矩形圖塊之動作向量 ML加法平均而產生時間直接向 = …解碼對象)之_二二= ="達到可提昇時間直接向量一精密ί (第3實施形態) 於上述第1實施形態中 ㈣训藉由求得位於編碼接向量產生 ::置的:個矩形方塊,向量=?2)= 平均,而產生時間直接模 ,2,3, 4)的加法 可對應於編竭對象(解碼=向二I’但亦 域之差㈣作向量,魏個圖料數個分割區 加權相加,藉此而產生 ° °°域之動作向量MVi予以 MV_。 時間直接模式之時間直接向量 以下,具體說明時間直 第9圖為示有以對應差分動的產生方法。 時間直接向量MVd夕洋 卞问夏的加權相加進行的 時間直接向量的說明圖。 同地’對於每個編觸象第1彻態相 上位於該圖塊集之附近 《囫塊集,從構成時間 集之中,特定出與該圖塊已解像)的圖塊 集。 4間上位於相同位置的圖塊 例如,當編喝對象(解瑪對象 )之圖塊集如第12圖所示 322764 26 201143452 為圖像B2之圖塊集觀時,可從構成時間上 MB1之附近的已編碼圖像P3(已解碼圖 塊= 時間直接向量產生部13、32於當 (解碼對象)之圖塊集ΜΒ1在空間上位 、編碼對象 ΜΒ2時,則如第9圖所示,從儲存於動置的圖塊集 31的已編碼動作向量(已解碼動作向量 ° °己憶體11、 塊集ΜΒ2之中心的4個圖塊(分之中取得位於該圖 = 1,2,3,4)0 #⑹之動作向量MVi(i 於第9圖之例中,係將左上圖塊 右下圖塊之動作向量設為,依照 σ置5免為MVl, 號。 “、、柵掃描之順序分配編 另外,時間直接向量產生部13、扣, MB2之中心的4個圖塊之差分動作向θ係取得位於圖塊集 當時間直接向量產生部13、32^^MVdl( 1 = 1,2, 3, 4)。 量,與4個圖塊之差分動作向4mv=4個圖塊之動作向 所示,藉由對應4個圖塊之差分動作向日則如下述式(5) 作向量MVi的加權相加,藉此產生 Q里MVdi而將4個動 向'5 MVdirect。 曰直接模式之時間直接 ’則給予固定值Α作 但,當差分動作向量MVdi為〇時 為加權。 ' Σ^Μνΐ (Srect(4) As described above, according to the second embodiment, the temporal direct vector generation units 13 and 32 are configured to weight-add the motion vector MVi of the plurality of divided regions in accordance with the tile size of the plurality of divided regions. When this occurs, 25 322764 201143452: the direct vector MVdi of the straight ϊΓ — is compared with the above ith real MV plus; the motion vector ML of the four rectangular blocks located at the center of the tile set is averaged and the time is directly generated. _2 ==" to reach the boostable time direct vector-precision ί (third embodiment) In the above-described first embodiment, (4) training is obtained by finding the coded vector: : a rectangular square, vector =? 2) = average, and the addition of the temporal direct mode, 2, 3, 4) can correspond to the compiled object (decoding = to the second I' but also the difference between the fields (four) as a vector, Wei is divided into several partitions, and the motion vector MVi of the °° domain is generated by MV_. The time direct mode is below the direct vector, and the time is straight. Method of production. Time directly MVd 卞 卞 卞 卞 卞 的 的 的 的 的 的 的 的 的 说明 说明 说明 说明 说明 说明 说明 说明 说明 说明 说明 说明 说明 说明 说明 说明 说明 说明 说明 说明 MV 说明 MV MV MV MV MV MV MV MV MV MV MV MV In the set, the set of tiles that have been resolved with the tile is specified. 4 blocks located at the same position, for example, when the tile set of the object (the object to be solved) is 322764 26 201143452 as shown in Fig. 12 as the block set of the image B2, the MB1 can be constructed from the time. The coded image P3 in the vicinity (the decoded tile = time direct vector generation sections 13 and 32) when the tile set ΜΒ1 of the (decoding target) is spatially upper and the coding target ΜΒ2 is as shown in FIG. From the coded motion vector stored in the set of motion blocks 31 (the decoded motion vector, the four tiles in the center of the block 11 and the block set ( 2 are obtained in the figure = 1, 2, 3,4)0 #(6) The motion vector MVi(i) In the example of Fig. 9, the motion vector of the lower right block of the upper left block is set to MV1, according to σ set 5, ". In addition, the difference between the four blocks of the center of the time direct vector generation unit 13, the buckle, and the MB2 is obtained in the θ system. The time vector generation unit 13 and 32^^MVdl are located in the tile set (1 = 1,2, 3, 4). The difference between 4 and 4 tiles is shown in the direction of 4mv=4 tiles, corresponding to 4 tiles. The differential action to the sun is weighted by the vector MVi as shown in the following equation (5), thereby generating Q MVdi and 4 movements to '5 MVdirect. 曰 Direct mode time directly' gives a fixed value, but when When the differential action vector MVdi is 〇, it is weighted. ' Σ^Μνΐ (Srect

R (5) 322764 27 (6) 201143452 (6) 201143452 &lt;Pi MVd; 差分二=向量_l之大小係表示向量之孤立度, 加權之叶丄:之大小越大’則孤立度可能越大,故於 加榷之计异式的式(6), 的倒數,而可⑹Γ刀動作向量讀1之大小 稚小孤立的向量之加權。 產生=二依據本第3實施形態,由於時間直接向量 曰™ 構成為對應複數個分割區域之差分動作向 =MVdi而將複數個分割區域之動作向量點加權相加,藉 時間直接模式之時間直接向量MWt,故相較於上 述第1實㈣態’彳達到提昇時間直接向量MVdi⑽之精密 度的效果。 (第4實施形態) 於上述第2實施形態中,雖例示為對應複數個分割區 域之圖塊大小而將複數個分割區域之動作向量Ml加權相 加;於上述第3實施形態中,雖例示為對應複數個分割區 域之差分動作向量MVdi ’將複數個分割區域之動作向量MVi 予以加權相加,但亦可對應複數個分割區域之編碼模式, 而將複數個分割區域之動作向量MVi加權相加。 例如,當屬於已編碼圖塊(已解碼圖塊)的分割區域為 内部圖塊(intra b 1 ock)時,亦可考慮將該分割區域從計算 對象中剔除而產生時間直接向量MVdirect的方法。 具體而言,第8圖中,例如,當左上之圖塊(分割區域) 為内部圖塊時(第8圖中雖於左上圖塊標示有動作向量 322764 28 201143452 MVi,但當其為内部圖塊時,則不具有動作向量之資訊), 藉由將左上的圖塊從計算對象剔除,而將剩下3個圖塊之 動作向量MVi(i = 2, 3, 4)加權相加,而產生時間直接模式之 時間直接向量MVd irect ° 由於内部圖塊不具有動作向量之資訊,故藉由將内部 圖塊從計算對象剔除,則可避免無謂地使時間直接向量成 為零向量,而提昇時間直接向量之精密度。 (第5實施形態) 於上述第1至3實施形態中,雖例示為使用4個圖塊 之向量而產生時間直接向量MVdirect,但亦可使用圖塊集内 全部圖塊之動作向量而產生時間直接向量MVdirect。 第10圖為示有使用圖塊集内全部圖塊之動作向量而 產生時間直接向量MVdir^t之方法的說明圖。 若假設與編碼對象(解碼對象)之圖塊集MB1在空間上 位於相同位置的圖塊集MB2内之k個圖塊之動作向量為MVi (OSiSk),k個圖塊之大小為Si(OSiSk),則藉由演算下 述式(7),即可產生時間直接模式之時間直接向量MVdirect。 Σ奶MV〖 - ~~(7) Ψι =Si (8) (產業上的可利用性) 本發明係適用於將影像壓縮編碼而傳送壓縮影像之 際,要求高編碼效率的影像編碼裝置。 29 322764 201143452 【圖式簡單說明】 第1圖係表示太^ 構成圖。 明第1實施形態之影像編碼裳置的 第2圖係示有本秣 動作補償預剩部12明第1實施形態之影像編碼裝置的 第3_係表j成圖。 構成圖。 下本發明第1實施形態之影像解喝農置的 第4圖係 動作補償_部2 ^第1實施形態之影像解喝裝置的 進,第1實施形態之影像編, 進行5:===1實施形― 為村日㈣直接模狀㈣直 王方法的說明圖。 1 MVdirect之 時間2二'示有藉由對應圖塊大小的加權相加而進行的 接向icMVdirect之產生方法的說明圖。 仃的 =9圖為示有藉由對應差分動作向量的加 、:間直接向量MWt之產生方法的說明圖。’進 而產生i〇門圖古為示有曰使用圖塊集内之全部圖塊的動作向量 ' 4直接向直MVdirect之方法的說明圖。 ^ 11圖為不有非矩形圖塊之動作1^7吾 說明圖。 鬼之動作向里的預測處理的 第12圖為示有以時間直接模式產生動作向量之方法 322764 30 201143452 的示意圖。 第13圖為示有當圖像p3的圖塊集mb2被分割為2個 以上的矩形圖塊時所使用的動作向量MV的說明圖。 【主要元件符號說明】 1 動作補償預測部 2 減法器(量子化手段) 3 編碼模式判定部(量子化手段) 4 壓縮部(量子化手段) 5 局部解碼部 6 加法器 7 迴圈濾波器 8 圖框記憶體 9 可變長度編碼部(可變長度編碼手段) 11 動作向量記憶體 12 動作向量探索部 13時間直接向量產生部(直接向量產生手段) 1動作補償處理部(預測影像產生手段) 可憂長度解竭部(可變長度解碼手段) 預測誤差解碣部(反量子化手段) 23 動作補償預測部 24加法器(影像相加手段) 25迴圈遽波器(影像相加手段) 26 圖框記憶體 動作向量記憶體 322764 31 31 201143452 32 時間直接向量產生部(直接向量產生手段) 33 動作補償處理部(預測影像產生手段) 32 322764R (5) 322764 27 (6) 201143452 (6) 201143452 &lt;Pi MVd; The difference between the difference ==vector_l indicates the degree of isolation of the vector, and the weighted leaf 丄: the larger the size, the greater the isolation may be Therefore, the reciprocal of the formula (6), which is the equivalent of the crowning method, can be weighted by the (6) sickle motion vector read 1 size and the isolated vector. According to the third embodiment, the time direct vector 曰TM is configured to weight-add the motion vector points of the plurality of divided regions corresponding to the difference operation direction of the plurality of divided regions, and the time of the direct mode is directly The vector MWt is compared to the above-described first real (four) state '彳 to achieve the effect of improving the precision of the direct vector MVdi(10). (Fourth Embodiment) In the second embodiment, the motion vector M1 of the plurality of divided regions is weighted and added in accordance with the tile size of the plurality of divided regions. In the third embodiment, the third embodiment is illustrated. The motion vector MVi of the plurality of divided regions is weighted and added for the differential motion vector MVdi ' of the plurality of divided regions, but may also be corresponding to the coding mode of the plurality of divided regions, and the motion vector MVi of the plurality of divided regions is weighted plus. For example, when the divided area belonging to the encoded tile (decoded tile) is an intra b ock, a method of culling the segmented region from the calculation object to generate the temporal direct vector MVdirect may also be considered. Specifically, in FIG. 8, for example, when the upper left tile (divided area) is an internal tile (in the eighth figure, although the upper left tile is marked with an action vector 322764 28 201143452 MVi, but it is an internal image In the case of a block, there is no information of the motion vector. By subtracting the upper left tile from the calculation object, the motion vectors MVi(i = 2, 3, 4) of the remaining three tiles are weighted and added. The time direct vector MVd irect is generated. Since the inner block does not have the information of the motion vector, by eliminating the inner block from the calculation object, it is possible to avoid unnecessarily making the time direct vector a zero vector and raising the time. The precision of the direct vector. (Fifth Embodiment) In the first to third embodiments described above, the time direct vector MVdirect is generated by using vectors of four tiles, but the motion vectors of all the tiles in the tile set may be used to generate time. Direct vector MVdirect. Fig. 10 is an explanatory diagram showing a method of generating a temporal direct vector MVdir^t using motion vectors of all tiles in a tile set. If it is assumed that the motion vector of k tiles in the tile set MB2 which is spatially located at the same position as the tile set MB1 of the encoding object (decoding object) is MVi (OSiSk), the size of k tiles is Si (OSiSk) ), by calculating the following formula (7), the time direct vector MVdirect of the time direct mode can be generated. Σ MV MV = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = 29 322764 201143452 [Simple description of the diagram] Figure 1 shows the composition of the figure. The second embodiment of the video encoding device according to the first embodiment of the present invention is shown in the third embodiment of the video encoding device according to the first embodiment. Make up the picture. Fig. 4 of the first embodiment of the present invention is an operation of the image decommissioning device according to the first embodiment of the present invention. The video editing device according to the first embodiment performs 5:=== 1 Implementation form - an illustration of the method of direct (4) direct king method for the village day (four). 1 MVdirect time 2'' shows an explanatory diagram of a method of generating an icMVdirect by weighted addition of the corresponding tile size. The =9 figure is an explanatory diagram showing a method of generating the direct vector MWt by the addition of the corresponding differential action vector. </ RTI> </ RTI> </ RTI> </ RTI> </ RTI> </ RTI> </ RTI> </ RTI> </ RTI> </ RTI> ^ 11 The picture shows the action of non-rectangular blocks. Figure 12 of the prediction process of the ghost action inward is a schematic diagram showing a method 322764 30 201143452 for generating an action vector in a time direct mode. Fig. 13 is an explanatory diagram showing an operation vector MV used when the tile set mb2 of the image p3 is divided into two or more rectangular tiles. [Description of main component symbols] 1 Motion compensation prediction unit 2 Subtractor (quantization means) 3 Encoding mode determination unit (quantization means) 4 Compression unit (quantization means) 5 Local decoding section 6 Adder 7 Loop filter 8 Frame memory 9 variable length coding unit (variable length coding means) 11 motion vector memory 12 motion vector search unit 13 temporal direct vector generation unit (direct vector generation means) 1 motion compensation processing unit (predictive video generation means) Worry length depletion unit (variable length decoding means) Predictive error decoding unit (anti-quantization means) 23 Motion compensation prediction unit 24 adder (image addition means) 25 loop chopper (image addition means) 26 Frame memory motion vector memory 322764 31 31 201143452 32 Time direct vector generation unit (direct vector generation means) 33 Motion compensation processing unit (predictive video generation means) 32 322764

Claims (1)

201143452 ^ 七、申請專利範固: 1. -種影像編碼裝置,係具有:直接向量產生手段,對於 每個構成輸人影像_塊,㈣雜上位於該 圖塊之附 近的已編碼圖像之動作向量而產生時間直接模式之時 間,接向1,預測影像錢手段,藉由使用*上述直接 向里產生手丰又所產生的時間直接向量實施動作補償預 :處理,而產生預測影像;量子化手段,將由上述預測 影像產生手段所產生的預測影像與上述輸入影像之差 刀办像予以里子化,輸出上述差分影像之量子化係數; =及可變長度編蜗手段’將由上述量子化手段所輸出的 子2係數予以可變長度編碼,輸出上述量子化係數之 編馬貝料’並且’上述直接向量產生手段係從構成時間 =上位於、、扁碼對象圖塊附近的已編碼圖像之圖塊中,特 疋出與柄對象圖塊在空間上位於相同位置的圖塊,由 位於上述圖塊中心的複數個分割區域之動作向量而產 生時間直接模式之時間直接向量。 2·如申^專利範圍第w所述之影像編碼裝置,其中,直 接=量產生手段係對應複數個分割區域之圖塊大小而 將複數個分割區域之動作向量加權相加,藉此產生時間 直接模式之時間直接向量。 3·如申請專利範圍第丨項所述之影像編碼裝置,其中,直 ^向量產生手段係對應複數個分割區域之差分動作向 量而將複數個分割區域之動作向量加權相加,藉此產生 時間直接模式之時間直接向量。 322764 1 201143452 4. 如申請專利範圍第1項所述之影像編碼裝置,i中,直 接向量產生手段係對應複數個分割區域之編碼模式而 將複數個分龍域之動作向量加權相力 直接模式之時間直接向量。 $ 5. -種f彡像解碼裝置,係具有:可變長度 碼資料解碼出量子化係數;反量子化手段,將由上述Ϊ 變長訪碼手段解碼的量子化係數反量子化;直接向量 產^手段’由時間性上位於解碼對象圖塊附近的已解碼 圖像之動作向量而產生時間直接模式之時間直接向 篁,^測㈣產生手段,藉由使用由上述直接向量產生 手#斤產生的時間直接向量實施動作補償預測處理,而 像;以及影像相加手段,將由上述預測影像 產:又所產生的預測影像與表示上述反量子化手段 :t2、°果的差分影像相加’而得到相當於影像編 碼裝置輸人影像的解碼影像;並且,上述直接向量產 生手段係從構成時間性上位於解碼對象圖塊附近的已 解碼圖像之圖塊中’特定出與解碼對象圖塊在空間上位 ,目同::的圖塊’由位於上述圖塊中心的複數個分割 區=之動作向量而產生時間直接模式之時間直接向量。 6·如申,專利範固第5項所述之影像解料置,其中,直 2手段係對應複數個分割區域之圖塊大小而 古Ϊ描/刀割區域之動作尚量加權相加,藉此產生時間 直接模式之時間直接向量。 7·如申請專利範園第5項所述之影像解料置,其中,直 322764 2 201143452 2向量產生手段係對應複數個分割區域之差分動作向 t而將複數個分割區域之動作向量加權相加 時間直接模式之時間直接向量。 8·如申請專利範圍帛5項所述之影像解石馬裝置,立中,直 = 量產生手段係對應複數個分割區域之編碼模式而 f複數個分割區域之動作向量加權相加,藉此產生時間 直接模式之時間直接向量。 9. -種影像編碼方法,係具有:直接向量產生處理步驟, 係直接向量產生手段躲每㈣成輸 時間性上位於該圖塊之附近的已編碼圖像之動作向= 而^生時間直接模式之㈣直接向量;預 =步驟,係_影像產生手段藉由使用由上述直接向量 理生產生的時間直接向量實施動作補償預測處 ,、生預測影像;量子化處理步驟,係量子化手段 上作述1員測衫像產生處理步驟所產生的預測影像與 == 象之差分影像予以量子化,輪出上述差分影 像之直子化係數;以及可變長度編碼處理步驟’係可變 手1又將由上述量子化處理步驟所輸出的量子 可變長度叫輸出上述量子化係數之編碼 處理+驟夕J&quot;述直接向量產生手段實施直接向量產生 近二編碼Γ係從構成時間性上位於編馬對象圖塊附 間上位於i像之圖塊中’特定出與編碼對象圖塊在空 個八宝位置的圖塊’由位於上述圖射心的複數 刀π域之動作向量而產生時間直接模式之時間直 322764 3 201143452 接向量。 ίο. —種影像解碼方法,係具有:可變長度解碼處理步驟, 係可變長度解碼手段從編碼資料解碼出量子化係數;反 量子化處理步驟,係反量子化手段將由上述可變長度解 碼處理步驟解碼的量子化係數反量子化;直接向量產生 處理步驟,係直接向量產生手段由時間性上位於解碼對 象圖塊附近的已解碼圖像之動作向量而產生時間直接 =式之時間直接向量;預測影像產生處理步驟,係預測 衫像產生手段藉由使用由上述直接向量產生處理步驟 的時間直接向量實施動作補償預測處理,而產生 由上述預_#^ 步驟,係影像相加手段將 示上述反量生處理步驟所產生的預測影像與表 相加,而得至Γ相步驟之反量子化結果的差分影像 像;並且’上诚:二影像編碼裝置之輸入影像的解碼影 5 ^ ’I直接向量產生手段實施直接向量產生声 v之際’係從構成時間性处 的已解碼圖像之圖秘 解碼對象圖壤附近 上位於相同位置=中,肢出與解韻象圖塊在空間 分割區域之:由位於上述圖塊中心的複_ 向量。 °直而產生時間直接模式之時間直接 322764 4201143452 ^ VII. Application for patents: 1. Image coding device, which has: direct vector generation means, for each of the input image_blocks, (4) mixed with the coded image located near the block The action vector generates the time of the direct mode, and the method of predicting the image money is performed by using the time direct vector generated by directly generating the hand-draw and the motion compensation pre-processing to generate the predicted image; a method for lining up the difference between the predicted image generated by the predicted image generating means and the input image, and outputting the quantized coefficient of the difference image; and the variable length worming means 'to be quantized by the above means The output sub-coefficient is variable-length coded, and the coded output of the above-mentioned quantized coefficient is outputted, and the above-mentioned direct vector generation means is a coded image from the constituent time=upper and the flat code object block. In the tile, the tile corresponding to the handle object tile is spatially located at the same position, and the plurality of tiles are located at the center of the tile. Cutting motion vector generation area and the time of the temporal direct mode, the direct vectors. 2. The video coding apparatus according to claim w, wherein the direct=quantization generating means weights the motion vectors of the plurality of divided regions corresponding to the tile size of the plurality of divided regions, thereby generating time Direct vector time direct vector. 3. The image coding apparatus according to claim 2, wherein the direct vector generation means weight-adds the motion vectors of the plurality of divided regions corresponding to the differential motion vectors of the plurality of divided regions, thereby generating the time Direct vector time direct vector. 322764 1 201143452 4. The image coding apparatus according to claim 1, wherein the direct vector generation means corresponds to the coding modes of the plurality of divided regions and the action vectors of the plurality of divided domains are weighted by the direct force mode. Time direct vector. $ 5. A f-image decoding device having: variable-length code data to decode a quantization coefficient; and an anti-quantization means to dequantize a quantized coefficient decoded by the above-described Ϊ variable length access method; direct vector The method of generating the time-directed mode by the action vector of the decoded image temporally located near the decoding target tile is directly directed to the 篁, ^ (4) generating means by using the direct vector generated by the hand The generated time direct vector implements the motion compensation prediction process, and the image and the image addition means add the predicted image generated by the predicted image to the difference image indicating the anti-quantization means: t2, ° fruit' And obtaining a decoded image corresponding to the input image of the image encoding device; and the direct vector generating means is configured to 'specify and decode the target block from the block that constitutes the decoded image temporally located near the decoding target block. In space, the same block:: the tile 'generates the direct mode from the action vector of multiple partitions = at the center of the above tile Time direct vector. 6·如申, the image unwrapping device described in the fifth item of the patent Fan Gu, wherein the straight 2 means corresponds to the block size of the plurality of divided regions, and the actions of the ancient scanning/knife cutting regions are weighted and added. This produces a time direct vector of the time direct mode. 7. The image unwinding device described in claim 5, wherein the straight 322764 2 201143452 2 vector generating means weights the motion vectors of the plurality of divided regions corresponding to the differential action of the plurality of divided regions to t Add time direct mode time direct vector. 8. The image-resolving stone device according to claim 5, the centering, the direct-quantity generating means is corresponding to the coding modes of the plurality of divided regions, and the motion vectors of the plurality of divided regions are weighted and added. Generates a direct vector of time direct mode. 9. The image encoding method has the following steps: a direct vector generation processing step, and a direct vector generation means to hide the motion of the encoded image located near the tile in the time of the transmission time. (4) direct vector; pre-step, the system _ image generation means by using the time direct vector generated by the direct vector rationalization to implement the motion compensation prediction, the raw prediction image; the quantization processing step, is based on the quantization means The predicted image generated by the one-step tying image generation processing step and the differential image of the == image are quantized, and the straightening coefficient of the differential image is rotated; and the variable length encoding processing step 'the variable hand 1 is again The quantum variable length outputted by the above quantization processing step is called the encoding processing for outputting the above-mentioned quantization coefficient. The direct vector generation means implements the direct vector generation of the near-two coding system, which is temporally located in the horse-drawn object map. In the block of the i-picture, the block that is specific to the block of the encoding object is in the space of the eight treasures. The complex number of the heart is the action vector of the π domain and the time of the direct mode is generated. 322764 3 201143452 The vector is connected. The image decoding method has a variable length decoding processing step, wherein the variable length decoding means decodes the quantization coefficient from the encoded data; and the inverse quantization processing step is performed by the variable length decoding method. The quantized coefficient dequantization decoded by the processing step; the direct vector generation processing step is a direct vector generation method that generates a temporal direct vector of time directly from the motion vector of the decoded image temporally located near the decoding target tile. a predictive image generation processing step of predicting a shirt image generation means by performing motion compensation prediction processing using a time direct vector generated by the direct vector generation processing step, and generating a pre-_#^ step, the image addition means will show The predicted image generated by the inverse quantization step is added to the table to obtain a differential image of the inverse quantization result of the phase step; and 'Sheng Cheng: the decoded image of the input image of the second image encoding device 5 ^ ' I direct vector generation means to implement direct vector generation of sound v FIG secret code image is located in the vicinity of the decoding target soil FIG position = the same, and the solution Yun limb like tiles in a space divided area: _ a complex vector located at the center of the tile. ° Straight and time direct mode time directly 322764 4
TW100104089A 2010-02-10 2011-02-08 Image encoding device, image decoding device, image encoding method and image decoding method TW201143452A (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2010027520A JP2013077865A (en) 2010-02-10 2010-02-10 Image encoding apparatus, image decoding apparatus, image encoding method and image decoding method
PCT/JP2011/000418 WO2011099241A1 (en) 2010-02-10 2011-01-26 Image encoding device, image decoding device, image encoding method, and image decoding method

Publications (1)

Publication Number Publication Date
TW201143452A true TW201143452A (en) 2011-12-01

Family

ID=44367530

Family Applications (1)

Application Number Title Priority Date Filing Date
TW100104089A TW201143452A (en) 2010-02-10 2011-02-08 Image encoding device, image decoding device, image encoding method and image decoding method

Country Status (3)

Country Link
JP (1) JP2013077865A (en)
TW (1) TW201143452A (en)
WO (1) WO2011099241A1 (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2012114712A1 (en) * 2011-02-25 2012-08-30 パナソニック株式会社 Image encoding method and image decoding method
CN109996075B (en) * 2017-12-29 2022-07-12 华为技术有限公司 Image decoding method and decoder

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2005236584A (en) * 2004-02-18 2005-09-02 Victor Co Of Japan Ltd Moving picture information conversion encoding apparatus

Also Published As

Publication number Publication date
JP2013077865A (en) 2013-04-25
WO2011099241A1 (en) 2011-08-18

Similar Documents

Publication Publication Date Title
US20210392340A1 (en) Video encoding apparatus, video decoding apparatus, and video decoding method for performing intra-prediction based on directionality of neighboring block
KR101418096B1 (en) Video Coding Method and Apparatus Using Weighted Prediction
JP5061179B2 (en) Illumination change compensation motion prediction encoding and decoding method and apparatus
US9736475B2 (en) Method and apparatus for generating a reference frame and method and apparatus for encoding/decoding image using the same
KR20190016056A (en) Methods of decoding using skip mode and apparatuses for using the same
TW201206203A (en) Image encoding device, image decoding device, image encoding method and image decoding method
US9485513B2 (en) Encoding device, decoding device, and program
KR101768865B1 (en) Video Coding Method and Apparatus Using Weighted Prediction
KR20150135457A (en) Method for encoding a plurality of input images and storage medium and device for storing program
TW201143452A (en) Image encoding device, image decoding device, image encoding method and image decoding method
JP2012513141A (en) Method and apparatus for predicting and encoding motion parameters of video picture sequences
TW201143455A (en) Image encoding device, image decoding device, image encoding method, image decoding method
KR101562343B1 (en) Method and apparatus for image interpolation having quarter pixel accuracy using intra prediction modes
US20190238895A1 (en) Method for local inter-layer prediction intra based
Park et al. Adaptive lossless video compression
Choi et al. Implicit line-based intra 16× 16 prediction for H. 264/AVC high-quality video coding
KR20150083828A (en) Method and apparatus for image interpolation having quarter pixel accuracy using intra prediction modes