TW201924333A - Encoding device, decoding device, encoding method, and decoding method - Google Patents
Encoding device, decoding device, encoding method, and decoding method Download PDFInfo
- Publication number
- TW201924333A TW201924333A TW107139023A TW107139023A TW201924333A TW 201924333 A TW201924333 A TW 201924333A TW 107139023 A TW107139023 A TW 107139023A TW 107139023 A TW107139023 A TW 107139023A TW 201924333 A TW201924333 A TW 201924333A
- Authority
- TW
- Taiwan
- Prior art keywords
- block
- processing
- unit
- boundary
- adjacent
- Prior art date
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/117—Filters, e.g. for pre-processing or post-processing
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/157—Assigned coding mode, i.e. the coding mode being predefined or preselected to be further used for selection of another element or parameter
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/17—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
- H04N19/176—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/80—Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation
- H04N19/82—Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation involving filtering within a prediction loop
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/85—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression
- H04N19/86—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression involving reduction of coding artifacts, e.g. of blockiness
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
Abstract
Description
發明領域
本發明是關於一種編碼裝置、解碼裝置、編碼方法及解碼方法。FIELD OF THE INVENTION The present invention relates to an encoding apparatus, a decoding apparatus, an encoding method, and a decoding method.
發明背景
以往存在H.265作為用以編碼動態圖像的規格。H.265亦稱為HEVC(High Efficiency Video Coding(高效率視訊編碼))。
先行技術文獻BACKGROUND OF THE INVENTION In the past, H.265 was used as a specification for encoding moving images. H.265 is also known as HEVC (High Efficiency Video Coding).
Advanced technical literature
非專利文獻
[非專利文獻1]H.265(ISO/IEC 23008-2 HEVC(High Efficiency Video Coding))Non-patent literature
[Non-Patent Document 1] H.265 (ISO/IEC 23008-2 HEVC (High Efficiency Video Coding))
發明概要
發明欲解決之課題
該類編碼方法及解碼方法期待可提升解碼圖像的畫質。SUMMARY OF THE INVENTION Problems to be Solved by the Invention This type of encoding method and decoding method are expected to improve the image quality of a decoded image.
本發明的目的在於提供一種可提升解碼圖像的畫質的編碼裝置、解碼裝置、編碼方法及解碼方法。An object of the present invention is to provide an encoding apparatus, a decoding apparatus, an encoding method, and a decoding method capable of improving the image quality of a decoded image.
用以解決課題之手段
本發明的一態樣的編碼裝置具備電路及記憶體;前述電路利用前述記憶體,對於對象區塊、及鄰接於前述對象區塊的鄰接區塊中至少一方的區塊,即判定用區塊適用預測圖像的亮度修正處理時,於前述對象區塊與前述鄰接區塊的邊界,適用去區塊濾波處理,對於前述判定用區塊不適用前述亮度修正處理時,不於前述邊界適用前述去區塊濾波處理。Means for Solving the Problem A coding apparatus according to an aspect of the present invention includes a circuit and a memory, and the circuit uses the memory to block at least one of a target block and an adjacent block adjacent to the target block. When the brightness correction processing of the prediction image is applied to the determination block, the deblocking filtering process is applied to the boundary between the target block and the adjacent block, and when the brightness correction processing is not applied to the determination block, The aforementioned deblocking filtering process is applied not to the aforementioned boundary.
本發明的一態樣的解碼裝置具備電路及記憶體;前述電路利用前述記憶體,對於對象區塊、及鄰接於前述對象區塊的鄰接區塊中至少一方的區塊,即判定用區塊適用預測圖像的亮度修正處理時,於前述對象區塊與前述鄰接區塊的邊界,適用去區塊濾波處理,對於前述判定用區塊不適用前述亮度修正處理時,不於前述邊界適用前述去區塊濾波處理。A decoding apparatus according to an aspect of the present invention includes a circuit and a memory, wherein the circuit uses the memory to determine at least one of a target block and an adjacent block adjacent to the target block, that is, a determination block. When the brightness correction processing of the predicted image is applied, the deblocking filtering process is applied to the boundary between the target block and the adjacent block, and when the brightness correction processing is not applied to the determination block, the foregoing is not applied to the boundary. Deblocking filtering.
再者,該等全面或具體的態樣亦得以系統、裝置、方法、積體電路、電腦程式、或電腦可讀取CD-ROM等非暫時性記錄媒體來實現,或以系統、裝置、方法、積體電路、電腦程式、及記錄媒體的任意組合來實現。Furthermore, such comprehensive or specific aspects may also be implemented by systems, devices, methods, integrated circuits, computer programs, or non-transitory recording media such as computer readable CD-ROMs, or by systems, devices, and methods. Any combination of integrated circuits, computer programs, and recording media.
發明效果
本發明可提供一種能提升解碼圖像的畫質的解碼裝置、編碼裝置、解碼方法或編碼方法。Advantageous Effects of Invention The present invention can provide a decoding apparatus, an encoding apparatus, a decoding method, or an encoding method capable of improving the image quality of a decoded image.
用以實施發明之形態
本發明的一態樣的編碼裝置具備電路及記憶體;前述電路利用前述記憶體,對於對象區塊、及鄰接於前述對象區塊的鄰接區塊中至少一方的區塊,即判定用區塊適用預測圖像的亮度修正處理時,於前述對象區塊與前述鄰接區塊的邊界,適用去區塊濾波處理,對於前述判定用區塊不適用前述亮度修正處理時,不於前述邊界適用前述去區塊濾波處理。EMBODIMENT OF THE INVENTION The coding apparatus according to one aspect of the present invention includes a circuit and a memory, and the circuit uses the memory to block at least one of a target block and an adjacent block adjacent to the target block. When the brightness correction processing of the prediction image is applied to the determination block, the deblocking filtering process is applied to the boundary between the target block and the adjacent block, and when the brightness correction processing is not applied to the determination block, The aforementioned deblocking filtering process is applied not to the aforementioned boundary.
若依據此,該編碼裝置在對於判定用區塊適用亮度修正處理時,進行去區塊濾波處理。藉此,由於可減低主觀上醒目的區塊雜訊,因此可提升解碼圖像的畫質。According to this, the encoding apparatus performs deblocking filtering processing when the brightness correction processing is applied to the determination block. Thereby, since the subjectively conspicuous block noise can be reduced, the image quality of the decoded image can be improved.
例如對於前述對象區塊及前述鄰接區塊的一方或雙方,適用前述亮度修正處理時,於前述邊界適用前述去區塊濾波處理亦可;對於前述對象區塊及前述鄰接區塊的任一方,均不適用前述亮度修正處理時,不於前述邊界適用前述去區塊濾波處理亦可。For example, when the luminance correction processing is applied to one or both of the target block and the adjacent block, the deblocking filtering process may be applied to the boundary; and for any one of the target block and the adjacent block, When the above-described brightness correction processing is not applied, the above-described deblocking filtering processing may be applied without the above boundary.
例如不論對於前述對象區塊有無適用前述亮度修正處理,對於前述鄰接區塊適用前述亮度修正處理時,均於前述邊界適用前述去區塊濾波處理,對於前述鄰接區塊不適用前述亮度修正處理時,均不於前述邊界適用前述去區塊濾波處理亦可。For example, regardless of whether or not the luminance correction processing is applied to the target block, when the luminance correction processing is applied to the adjacent block, the deblocking filtering processing is applied to the boundary, and the luminance correction processing is not applied to the adjacent block. The foregoing deblocking filtering process may not be applied to the foregoing boundary.
若依據此,該編碼裝置可抑制過度進行去區塊濾波處理。According to this, the encoding apparatus can suppress excessive deblocking filtering processing.
例如前述亮度修正處理為LIC(Local Illumination Compensation(局部照明補償))處理亦可。For example, the brightness correction processing may be LIC (Local Illumination Compensation) processing.
例如對於前述判定用區塊適用前述亮度修正處理時,藉由將表示前述邊界的邊界強度的Bs設定為0以外之值,以於前述邊界適用前述去區塊濾波處理亦可;對於前述判定用區塊不適用前述亮度修正處理時,藉由將前述Bs設定為0,以不於前述邊界適用前述去區塊濾波處理亦可。For example, when the brightness correction processing is applied to the determination block, the Bs indicating the boundary strength of the boundary is set to a value other than 0, and the deblocking filtering process may be applied to the boundary; When the block is not subjected to the above-described brightness correction processing, by setting the Bs to 0, the above-described deblocking filtering process may be applied without the above boundary.
本發明的一態樣的解碼裝置具備電路及記憶體;前述電路利用前述記憶體,對於對象區塊、及鄰接於前述對象區塊的鄰接區塊中至少一方的區塊,即判定用區塊適用預測圖像的亮度修正處理時,於前述對象區塊與前述鄰接區塊的邊界,適用去區塊濾波處理,對於前述判定用區塊不適用前述亮度修正處理時,不於前述邊界適用前述去區塊濾波處理。A decoding apparatus according to an aspect of the present invention includes a circuit and a memory, wherein the circuit uses the memory to determine at least one of a target block and an adjacent block adjacent to the target block, that is, a determination block. When the brightness correction processing of the predicted image is applied, the deblocking filtering process is applied to the boundary between the target block and the adjacent block, and when the brightness correction processing is not applied to the determination block, the foregoing is not applied to the boundary. Deblocking filtering.
若依據此,該解碼裝置在對於判定用區塊適用亮度修正處理時,進行去區塊濾波處理。藉此,由於可減低主觀上醒目的區塊雜訊,因此可提升解碼圖像的畫質。According to this, the decoding apparatus performs deblocking filtering processing when the brightness correction processing is applied to the determination block. Thereby, since the subjectively conspicuous block noise can be reduced, the image quality of the decoded image can be improved.
例如對於前述對象區塊及前述鄰接區塊的一方或雙方,適用前述亮度修正處理時,於前述邊界適用前述去區塊濾波處理亦可;對於前述對象區塊及前述鄰接區塊的任一方,均不適用前述亮度修正處理時,不於前述邊界適用前述去區塊濾波處理亦可。For example, when the luminance correction processing is applied to one or both of the target block and the adjacent block, the deblocking filtering process may be applied to the boundary; and for any one of the target block and the adjacent block, When the above-described brightness correction processing is not applied, the above-described deblocking filtering processing may be applied without the above boundary.
例如不論對於前述對象區塊有無適用前述亮度修正處理,對於前述鄰接區塊適用前述亮度修正處理時,均於前述邊界適用前述去區塊濾波處理,對於前述鄰接區塊不適用前述亮度修正處理時,均不於前述邊界適用前述去區塊濾波處理亦可。For example, regardless of whether or not the luminance correction processing is applied to the target block, when the luminance correction processing is applied to the adjacent block, the deblocking filtering processing is applied to the boundary, and the luminance correction processing is not applied to the adjacent block. The foregoing deblocking filtering process may not be applied to the foregoing boundary.
若依據此,該編碼裝置可抑制過度進行去區塊濾波處理。According to this, the encoding apparatus can suppress excessive deblocking filtering processing.
例如前述亮度修正處理為LIC(Local Illumination Compensation(局部照明補償))處理亦可。For example, the brightness correction processing may be LIC (Local Illumination Compensation) processing.
例如對於前述判定用區塊適用前述亮度修正處理時,藉由將表示前述邊界的邊界強度的Bs設定為0以外之值,以於前述邊界適用前述去區塊濾波處理亦可;對於前述判定用區塊不適用前述亮度修正處理時,藉由將前述Bs設定為0,以不於前述邊界適用前述去區塊濾波處理亦可。For example, when the brightness correction processing is applied to the determination block, the Bs indicating the boundary strength of the boundary is set to a value other than 0, and the deblocking filtering process may be applied to the boundary; When the block is not subjected to the above-described brightness correction processing, by setting the Bs to 0, the above-described deblocking filtering process may be applied without the above boundary.
本發明的一態樣的編碼方法對於對象區塊、及鄰接於前述對象區塊的鄰接區塊中至少一方的區塊,即判定用區塊適用預測圖像的亮度修正處理時,於前述對象區塊與前述鄰接區塊的邊界,適用去區塊濾波處理,對於前述判定用區塊不適用前述亮度修正處理時,不於前述邊界適用前述去區塊濾波處理。In the encoding method according to an aspect of the present invention, when the target block and the block adjacent to at least one of the adjacent blocks of the target block, that is, the brightness correction processing for applying the prediction image to the determination block, the object is The deblocking filtering process is applied to the boundary between the block and the adjacent block, and when the brightness correction processing is not applied to the determination block, the deblocking filtering process is not applied to the boundary.
若依據此,該編碼方法在對於判定用區塊適用亮度修正處理時,進行去區塊濾波處理。藉此,由於可減低主觀上醒目的區塊雜訊,因此可提升解碼圖像的畫質。According to this, the encoding method performs deblocking filtering processing when the brightness correction processing is applied to the determination block. Thereby, since the subjectively conspicuous block noise can be reduced, the image quality of the decoded image can be improved.
本發明的一態樣的解碼方法對於對象區塊、及鄰接於前述對象區塊的鄰接區塊中至少一方的區塊,即判定用區塊適用預測圖像的亮度修正處理時,於前述對象區塊與前述鄰接區塊的邊界,適用去區塊濾波處理,對於前述判定用區塊不適用前述亮度修正處理時,不於前述邊界適用前述去區塊濾波處理。In the decoding method according to an aspect of the present invention, when at least one of the target block and the adjacent block adjacent to the target block, that is, the brightness correction processing of the prediction image is applied to the determination block, the object is The deblocking filtering process is applied to the boundary between the block and the adjacent block, and when the brightness correction process is not applied to the block for determination, the deblocking filtering process is not applied to the boundary.
若依據此,該解碼方法在對於判定用區塊適用亮度修正處理時,進行去區塊濾波處理。藉此,由於可減低主觀上醒目的區塊雜訊,因此可提升解碼圖像的畫質。According to this, the decoding method performs deblocking filtering processing when the brightness correction processing is applied to the determination block. Thereby, since the subjectively conspicuous block noise can be reduced, the image quality of the decoded image can be improved.
進而言之,該等全面或具體的態樣亦得以系統、裝置、方法、積體電路、電腦程式、或電腦可讀取CD-ROM等非暫時性記錄媒體來實現,或以系統、裝置、方法、積體電路、電腦程式、及記錄媒體的任意組合來實現。In addition, such comprehensive or specific aspects can also be realized by systems, devices, methods, integrated circuits, computer programs, or non-transitory recording media such as computer readable CD-ROMs, or by systems, devices, The method, the integrated circuit, the computer program, and any combination of recording media are implemented.
以下參考圖式來具體說明實施形態。The embodiment will be specifically described below with reference to the drawings.
再者,以下所說明的實施形態均表示全面或具體的範例。以下實施形態所示數值、形狀、材料、構成要件、構成要件的配置位置及連接形態、步驟、步驟的順序等為一例,其主旨不在於限定申請專利範圍。又,關於以下實施形態的構成要件中,未記載於表示最高位概念的獨立請求項的構成要件,則作為任意的構成要件來說明。
(實施形態1)Furthermore, the embodiments described below represent comprehensive or specific examples. The numerical values, shapes, materials, constituent elements, arrangement positions and connection forms of the constituent elements, the order of the steps and the steps, and the like are exemplified in the following embodiments, and the gist of the invention is not limited to the scope of the patent application. Further, among the constituent elements of the following embodiments, the constituent elements of the independent request item indicating the highest-order concept are not described as arbitrary constituent elements.
(Embodiment 1)
首先,說明實施形態1的概要,來作為可適用後述的本發明的各態樣所說明的處理及/或構成的編碼裝置及解碼裝置的一例。但實施形態1僅為可適用本發明的各態樣所說明的處理及/或構成的編碼裝置及解碼裝置的一例,本發明的各態樣所說明的處理及/或構成,亦可於與實施形態1不同的編碼裝置及解碼裝置實施。First, an outline of the first embodiment will be described as an example of an encoding device and a decoding device which are applicable to the processing and/or configuration described in each aspect of the present invention to be described later. However, the first embodiment is only an example of an encoding device and a decoding device to which the processing and/or configuration described in each aspect of the present invention can be applied, and the processing and/or configuration described in each aspect of the present invention may be The encoding device and the decoding device are implemented in different embodiments.
對於實施形態1適用本發明的各態樣所說明的處理及/或構成時,亦可進行例如以下之任一項。
(1)對於實施形態1的編碼裝置或解碼裝置,將構成該編碼裝置或解碼裝置的複數個構成要件中之對應於本發明的各態樣所說明的構成要件之構成要件,置換成本發明的各態樣所說明的構成要件;
(2)對於實施形態1的編碼裝置或解碼裝置,針對構成該編碼裝置或解碼裝置的複數個構成要件中之一部分構成要件,施以功能或實施處理的追加、置換、刪除等任意變更後,將對應於本發明的各態樣所說明的構成要件之構成要件,置換成本發明的各態樣所說明的構成要件;
(3)對於實施形態1的編碼裝置或解碼裝置所實施的方法,針對處理的追加、及/或該方法所含的複數種處理中之一部分處理,施以置換、刪除等任意變更後,將對應於本發明的各態樣所說明的處理之處理,置換成本發明的各態樣所說明的處理;
(4)將構成實施形態1的編碼裝置或解碼裝置的複數個構成要件中之一部分構成要件,與本發明的各態樣所說明的構成要件、具備本發明的各態樣所說明的構成要件所具備的一部分功能的構成要件、或實施本發明的各態樣所說明的構成要件所實施的一部分處理的構成要件組合而實施;
(5)將具備構成實施形態1的編碼裝置或解碼裝置的複數個構成要件中之一部分構成要件所具備的一部分功能的構成要件、或實施構成實施形態1的編碼裝置或解碼裝置的複數個構成要件中之一部分構成要件所實施的一部分處理的構成要件,與本發明的各態樣所說明的構成要件、具備本發明的各態樣所說明的構成要件所具備的一部分功能的構成要件、或實施本發明的各態樣所說明的構成要件所實施的一部分處理的構成要件組合而實施;
(6)對於實施形態1的編碼裝置或解碼裝置所實施的方法,將該方法所含的複數種處理中之對應於本發明的各態樣所說明的處理之處理,置換成本發明的各態樣所說明的處理;
(7)將實施形態1的編碼裝置或解碼裝置所實施的方法所含之複數種處理中之一部分處理,與本發明的各態樣所說明的處理組合而實施。In the case where the processing and/or configuration described in each aspect of the present invention is applied to the first embodiment, for example, any of the following may be employed.
(1) In the coding apparatus or the decoding apparatus according to the first embodiment, the constituent elements corresponding to the constituent elements described in the respective aspects of the present invention among the plurality of constituent elements constituting the encoding apparatus or the decoding apparatus are replaced by the invention. The constituent elements described in each aspect;
(2) The encoding device or the decoding device according to the first embodiment is configured as a component of a plurality of constituent elements constituting the encoding device or the decoding device, and is arbitrarily changed by adding, replacing, or deleting a function or a process. The constituent elements corresponding to the constituent elements described in the various aspects of the present invention are replaced by the constituent elements described in the various aspects of the invention;
(3) The method implemented by the encoding device or the decoding device according to the first embodiment, after the processing is added, and/or one of the plurality of processes included in the method is subjected to any change such as replacement or deletion, The processing described in connection with the aspects of the present invention replaces the processing described in the various aspects of the invention;
(4) A part of a plurality of constituent elements constituting the coding apparatus or the decoding apparatus according to the first embodiment, and constituent elements described in the aspects of the present invention, and the constituent elements described in the aspects of the present invention. The constituent elements of a part of the functions provided, or the constituent elements of a part of the processing performed by the constituent elements described in the various aspects of the present invention are combined;
(5) A component constituting a part of the functions of the plurality of components constituting the coding device or the decoding device according to the first embodiment, or a plurality of components constituting the coding device or the decoding device according to the first embodiment One of the requirements constitutes a part of the processing required by the requirements, and the constituent elements described in the aspects of the present invention, and the constituent elements of the functions of the constituent elements described in the aspects of the present invention, or Carrying out a combination of constituent elements of a part of the processing performed by the constituent elements described in the aspects of the present invention;
(6) The method performed by the encoding device or the decoding device according to the first embodiment, the processing corresponding to the processing described in each aspect of the present invention among the plurality of processings included in the method is replaced with the state of the invention. The treatment described;
(7) One of a plurality of processes included in the method performed by the encoding device or the decoding device according to the first embodiment is combined with the processing described in each aspect of the present invention.
再者,本發明的各態樣所說明的處理及/或構成的實施方式,不限定於上述例子。例如在與實施形態1所揭示的動態圖像/圖像編碼裝置或動態圖像/圖像解碼裝置,以不同目的利用的裝置實施,或單獨實施各態樣所說明的處理及/或構成均可。又,亦可組合在不同態樣所說明的處理及/或構成而實施。
[編碼裝置的概要]Furthermore, the embodiments of the processes and/or configurations described in the various aspects of the present invention are not limited to the above examples. For example, in the moving image/image encoding device or the moving image/image decoding device disclosed in the first embodiment, the device is used for a different purpose, or the processing and/or configuration described in each aspect is performed separately. can. Further, it may be implemented by combining the processes and/or configurations described in different aspects.
[Summary of coding device]
首先,說明實施形態1的編碼裝置的概要。圖1是表示實施形態1之編碼裝置100的功能構成之方塊圖。編碼裝置100是以區塊為單位編碼動態圖像/圖像的動態圖像/圖像編碼裝置。First, an outline of an encoding apparatus according to the first embodiment will be described. Fig. 1 is a block diagram showing a functional configuration of an encoding apparatus 100 according to the first embodiment. The encoding device 100 is a moving image/image encoding device that encodes a moving image/image in units of blocks.
如圖1所示,編碼裝置100是以區塊為單位編碼圖像的裝置,具備分割部102、減算部104、轉換部106、量化部108、熵編碼部110、反量化部112、反轉換部114、加算部116、區塊記憶體118、迴路濾波部120、幀記憶體122、幀內預測部124、幀間預測部126及預測控制部128。As shown in FIG. 1, the encoding apparatus 100 is an apparatus for encoding an image in units of blocks, and includes a dividing unit 102, a subtracting unit 104, a converting unit 106, a quantization unit 108, an entropy encoding unit 110, an inverse quantization unit 112, and an inverse conversion. The unit 114, the adding unit 116, the block memory 118, the loop filter unit 120, the frame memory 122, the intra prediction unit 124, the inter prediction unit 126, and the prediction control unit 128.
編碼裝置100藉由例如通用處理器及記憶體來實現。此時,由處理器執行儲存於記憶體的軟體程式時,處理器是作為分割部102、減算部104、轉換部106、量化部108、熵編碼部110、反量化部112、反轉換部114、加算部116、迴路濾波部120、幀內預測部124、幀間預測部126及預測控制部128發揮功能。又,編碼裝置100亦可作為對應於分割部102、減算部104、轉換部106、量化部108、熵編碼部110、反量化部112、反轉換部114、加算部116、迴路濾波部120、幀內預測部124、幀間預測部126及預測控制部128的專用的1以上的電子電路來實現。The encoding device 100 is realized by, for example, a general purpose processor and a memory. At this time, when the processor executes the software program stored in the memory, the processor functions as the division unit 102, the subtraction unit 104, the conversion unit 106, the quantization unit 108, the entropy coding unit 110, the inverse quantization unit 112, and the inverse conversion unit 114. The addition unit 116, the loop filter unit 120, the intra prediction unit 124, the inter prediction unit 126, and the prediction control unit 128 function. Further, the encoding device 100 may correspond to the division unit 102, the subtraction unit 104, the conversion unit 106, the quantization unit 108, the entropy coding unit 110, the inverse quantization unit 112, the inverse conversion unit 114, the addition unit 116, and the loop filter unit 120. The intra prediction unit 124, the inter prediction unit 126, and the prediction control unit 128 are realized by a dedicated one or more electronic circuits.
以下說明編碼裝置100所含的各構成要件。
[分割部]Each constituent element included in the encoding device 100 will be described below.
[Division Department]
分割部102將輸入動態圖像所含的各圖片,分割為複數個區塊,將各區塊輸出至減算部104。例如分割部102首先將圖片分割為固定大小(例如128´128)的區塊。該固定大小的區塊有時稱為編碼樹單元(CTU)。然後,分割部102根據遞迴的四元樹(quadtree)及/或二元樹(binary tree)區塊分割,將固定大小的區塊各個分割為可變大小(例如64´64)的區塊。該可變大小的區塊有時稱為編碼單元(CU)、預測單元(PU)或轉換單元(TU)。再者,於本實施形態無須區分CU、PU及TU,圖片內的一部分或所有區塊為CU、PU、TU的處理單位亦可。The division unit 102 divides each picture included in the input moving image into a plurality of blocks, and outputs each block to the subtraction unit 104. For example, the dividing unit 102 first divides the picture into blocks of a fixed size (for example, 128 ́128). This fixed size block is sometimes referred to as a coding tree unit (CTU). Then, the segmentation unit 102 divides the blocks of fixed size into blocks of variable size (for example, 64 ́64) according to the recursive quadtree and/or binary tree segmentation. . The variable size block is sometimes referred to as a coding unit (CU), a prediction unit (PU), or a conversion unit (TU). Furthermore, in this embodiment, it is not necessary to distinguish between CU, PU, and TU, and some or all of the blocks in the picture may be processing units of CU, PU, and TU.
圖2是表示實施形態1的區塊分割的一例的圖。於圖2,實線表示四元樹區塊分割的區塊邊界,虛線表示二元樹區塊分割的區塊邊界。Fig. 2 is a view showing an example of block division in the first embodiment; In FIG. 2, the solid line indicates the block boundary of the quaternary tree block division, and the broken line indicates the block boundary of the binary tree block division.
於此,區塊10為128´128像素的正方形區塊(128´128區塊)。該128´128區塊10首先分割為4個正方形的64´64區塊(四元樹區塊分割)。Here, block 10 is a 128 ́128 pixel square block (128 ́128 block). The 128 ́128 block 10 is first partitioned into 4 square 64 ́64 blocks (quaternary tree block partitioning).
左上64´64區塊進一步垂直分割為2個矩形的32´64區塊,左32´64區塊進一步垂直分割為2個矩形的16´64區塊(二元樹區塊分割)。其結果,左上64´64區塊分割為2個16´64區塊11、12及32´64區塊13。The upper left 64 ́64 block is further vertically divided into two rectangular 32 ́64 blocks, and the left 32 ́64 block is further vertically divided into two rectangular 16 ́64 blocks (binary tree block segmentation). As a result, the upper left 64 ́64 block is divided into two 16 ́64 blocks 11, 12 and 32 ́64 blocks 13.
右上64´64區塊水平分割為2個矩形的64´32區塊14、15(二元樹區塊分割)。The upper right 64 ́64 block is horizontally divided into two rectangular 64 ́32 blocks 14, 15 (binary tree block segmentation).
左下64´64區塊分割為4個正方形的32´32區塊(四元樹區塊分割)。4個32´32區塊中,左上區塊及右下區塊進一步分割。左上32´32區塊垂直分割為2個矩形的16´32區塊,右16´32區塊進一步水平分割為2個16´16區塊(二元樹區塊分割)。右下32´32區塊水平分割為2個32´16區塊(二元樹區塊分割)。其結果,左下64´64區塊分割為16´32區塊16、2個16´16區塊17、18、2個32´32區塊19、20及2個32´16區塊21、22。The lower left 64 ́64 block is divided into 4 square 32 ́32 blocks (quaternary tree block partitioning). Among the four 32 ́32 blocks, the upper left block and the lower right block are further divided. The upper left 32 ́32 block is vertically divided into two rectangular 16 ́32 blocks, and the right 16 ́32 block is further horizontally divided into two 16 ́16 blocks (binary tree block segmentation). The lower right 32 ́32 block is horizontally divided into two 32 ́16 blocks (binary tree block partitioning). As a result, the lower left 64 ́64 block is divided into 16 ́32 blocks 16, two 16 ́16 blocks 17, 18, two 32 ́32 blocks 19, 20 and two 32 ́16 blocks 21, 22 .
右下64´64區塊23未分割。The lower right 64 ́64 block 23 is undivided.
如以上,於圖2,區塊10根據遞迴的四元樹及二元樹區塊分割,分割為13個可變大小的區塊11~23。此類分割有時稱為QTBT(quad-tree plus binary tree(四元樹加二元樹))分割。As above, in FIG. 2, the block 10 is divided into 13 variable-sized blocks 11 to 23 according to the recursive quadtree and binary tree block division. Such segmentation is sometimes referred to as QTBT (quad-tree plus binary tree) segmentation.
再者,於圖2,1個區塊分割為4個或2個區塊(四元樹或二元樹區塊分割),但分割不限定於此。例如1個區塊亦可分割為3個區塊(三元樹區塊分割)。該類包含三元樹區塊分割在內的分割有時稱為MBT(multi type tree(多型樹))分割。
[減算部]Furthermore, in FIG. 2, one block is divided into four or two blocks (quaternary tree or binary tree block division), but the division is not limited thereto. For example, one block can also be divided into three blocks (three-dimensional tree block division). This type of partitioning including ternary tree block partitioning is sometimes referred to as MBT (multi-type tree) partitioning.
[Reduction Department]
減算部104是以分割部102所分割的區塊單位,從原訊號(原樣本)減算預測訊號(預測樣本)。總言之,減算部104算出編碼對象區塊(以下稱為目前區塊)的預測誤差(亦稱為殘差)。然後,減算部104將算出的預測誤差輸出至轉換部106。The subtraction unit 104 subtracts the prediction signal (predicted sample) from the original signal (original sample) in the block unit divided by the division unit 102. In summary, the subtraction unit 104 calculates a prediction error (also referred to as a residual) of a coding target block (hereinafter referred to as a current block). Then, the subtraction unit 104 outputs the calculated prediction error to the conversion unit 106.
原訊號為編碼裝置100的輸入訊號,其為表示構成動態圖像的各圖片之圖像的訊號(例如亮度(luma)訊號及2個色差(chroma)訊號)。以下有時亦將表示圖像的訊號稱為樣本。
[轉換部]The original signal is an input signal of the encoding device 100, which is a signal (for example, a luma signal and two chroma signals) indicating images of the respective pictures constituting the moving image. The signal representing the image is sometimes referred to as a sample below.
[conversion department]
轉換部106將空間域的預測誤差,轉換成頻率域的轉換係數,將轉換係數輸出至量化部108。具體而言,轉換部106對於例如空間域的預測誤差,進行預先決定的離散餘弦變換(DCT)或離散正弦變換(DST)。The conversion unit 106 converts the prediction error in the spatial domain into a conversion coefficient in the frequency domain, and outputs the conversion coefficient to the quantization unit 108. Specifically, the conversion unit 106 performs a predetermined discrete cosine transform (DCT) or discrete sine transform (DST) on, for example, a prediction error in the spatial domain.
再者,轉換部106亦可從複數種轉換類型中,適應性地選擇轉換類型,利用與選擇的轉換類型相對應的轉換基函數(transform basis function),將預測誤差轉換成轉換係數。該類轉換有時稱為EMT(explicit multiple core transform(顯式多重核心轉換))或AMT(adaptive multiple transform(適應性多重轉換))。Furthermore, the conversion unit 106 may adaptively select a conversion type from a plurality of types of conversions, and convert the prediction error into a conversion coefficient by using a transform basis function corresponding to the selected conversion type. This type of conversion is sometimes called EMT (explicit multiple core transform) or AMT (adaptive multiple transform).
複數種轉換類型包含例如DCT-II、DCT-V、DCT-VIII、DST-I及DST-VII。圖3是表示對應於各轉換類型的基函數的表。於圖3,N表示輸入像素數。從該等複數種轉換類型中選擇轉換類型時,可取決於例如預測的種類(幀內預測及幀間預測),或取決於幀內預測模式。A plurality of types of conversion include, for example, DCT-II, DCT-V, DCT-VIII, DST-I, and DST-VII. FIG. 3 is a table showing basis functions corresponding to respective conversion types. In Fig. 3, N represents the number of input pixels. The choice of the type of conversion from among the plurality of types of conversion may depend, for example, on the type of prediction (intra prediction and inter prediction) or on the intra prediction mode.
該類表示適用EMT或AMT與否的資訊(例如稱為AMT旗標)、及表示選擇的轉換類型的資訊是以CU級別來訊號化。再者,該等資訊的訊號化無須限定在CU級別,亦可為其他級別(例如序列級別、圖片級別、切片級別、方塊級別或CTU級別)。This class indicates information applicable to EMT or AMT (for example, called AMT flag), and information indicating the type of conversion selected is signaled at the CU level. Furthermore, the signalization of such information need not be limited to the CU level, but may be other levels (such as sequence level, picture level, slice level, block level or CTU level).
又,轉換部106亦可將轉換係數(轉換結果)予以再轉換。該類再轉換有時稱為AST(adaptive secondary transform(適應性第二轉換))或NSST(non-separable secondary transform(不可分第二轉換))。例如轉換部106就對應於幀內預測誤差的轉換係數的區塊所含子區塊(例如4´4子區塊),逐一進行再轉換。表示適用NSST與否的資訊、及用於NSST的轉換矩陣的相關資訊是以CU級別來訊號化。再者,該等資訊的訊號化無須限定在CU級別,亦可為其他級別(例如序列級別、圖片級別、切片級別、方塊級別或CTU級別)。Further, the conversion unit 106 may reconvert the conversion coefficient (conversion result). This type of retransformation is sometimes referred to as AST (adaptive secondary transform) or NSST (non-separable secondary transform). For example, the conversion unit 106 performs retransformation one by one by sub-blocks (for example, 4 ́4 sub-blocks) included in the block corresponding to the conversion coefficient of the intra prediction error. Information indicating whether NSST is applicable or not, and information about the conversion matrix used for NSST is signaled at the CU level. Furthermore, the signalization of such information need not be limited to the CU level, but may be other levels (such as sequence level, picture level, slice level, block level or CTU level).
於此,Separable(可分離)轉換是指就各方向,僅分離輸入的維數而進行複數次轉換的方式,Non-Separable(不可分離)轉換是指在輸入為多維時,將2個以上的維度統整視為1維而一次進行轉換的方式。Here, Separable conversion refers to a method of performing multiple conversions in which only the dimensions of the input are separated in each direction, and Non-Separable conversion means that when the input is multi-dimensional, two or more are used. The dimension is treated as a one-dimensional and one-time conversion.
例如作為Non-Separable轉換的一例,可舉出在輸入為4´4的區塊時,將其視為具有16個要件的1個陣列,對於該陣列,以16´16的轉換矩陣進行轉換處理。For example, as an example of the Non-Separable conversion, when a block of 4 ́4 is input, it is regarded as an array having 16 elements, and for the array, conversion processing is performed with a conversion matrix of 16 ́16. .
又,同樣地將4´4的輸入區塊視為具有16個要件的1個陣列後,對於該陣列,進行複數次Givens旋轉(Hypercube Givens Transform(超立方體Givens轉換))的轉換,亦為Non-Separable轉換例。
[量化部]Similarly, after the 4 ́4 input block is regarded as one array having 16 elements, a plurality of Givens rotation (Hypercube Givens Transform) conversion is performed for the array, which is also Non. -Separable conversion example.
[Quantization Department]
量化部108量化從轉換部106輸出的轉換係數。具體而言,量化部108以預定的掃描順序掃描目前區塊的轉換係數,根據對應於經掃描的轉換係數的量化參數(QP),來量化該轉換係數。然後,量化部108將目前區塊的經量化的轉換係數(以下稱為量化係數),輸出至熵編碼部110及反量化部112。The quantization unit 108 quantizes the conversion coefficients output from the conversion unit 106. Specifically, the quantization section 108 scans the conversion coefficient of the current block in a predetermined scanning order, and quantizes the conversion coefficient according to a quantization parameter (QP) corresponding to the scanned conversion coefficient. Then, the quantization unit 108 outputs the quantized conversion coefficients (hereinafter referred to as quantized coefficients) of the current block to the entropy coding unit 110 and the inverse quantization unit 112.
預定的順序是轉換係數的量化/反量化用順序。例如預定的掃描順序是以頻率的升序(從低頻往高頻的順序)或降序(從高頻往低頻的順序)定義。The predetermined order is the order of quantization/dequantization of the conversion coefficients. For example, the predetermined scanning order is defined in ascending order of frequency (in order from low frequency to high frequency) or descending order (in order from high frequency to low frequency).
量化參數是定義量化步階(量化寬)的參數。例如若增加量化參數值,則量化步階亦增加。總言之,若量化參數值增加,則量化誤差增大。
[熵編碼部]The quantization parameter is a parameter that defines the quantization step (quantization width). For example, if the quantization parameter value is increased, the quantization step is also increased. In summary, if the quantization parameter value increases, the quantization error increases.
[Entropy coding unit]
熵編碼部110藉由將從量化部108輸入的量化係數,予以可變長度編碼,來生成編碼訊號(編碼位元串流)。具體而言,熵編碼部110例如將量化係數二值化,將二值訊號予以算術編碼。
[反量化部]The entropy coding unit 110 generates a coded signal (coded bit stream) by variable length coding of the quantized coefficients input from the quantization unit 108. Specifically, the entropy coding unit 110 binarizes the quantized coefficients, for example, and arithmetically encodes the binary signals.
[Anti-quantization department]
反量化部112將來自量化部108的輸入即量化係數,予以反量化。具體而言,反量化部112以預定的掃描順序,將目前區塊的量化係數予以反量化。然後,反量化部112將目前區塊之已被反量化的轉換係數,輸出至反轉換部114。
[反轉換部]The inverse quantization unit 112 inversely quantizes the quantized coefficient, which is an input from the quantization unit 108. Specifically, the inverse quantization unit 112 inversely quantizes the quantized coefficients of the current block in a predetermined scanning order. Then, the inverse quantization unit 112 outputs the converted coefficients of the current block that have been inversely quantized to the inverse conversion unit 114.
[Anti-conversion department]
反轉換部114藉由將來自反量化部112的輸入即轉換係數,予以反轉換,來復原預測誤差。具體而言,反轉換部114藉由對於轉換係數,進行與轉換部106的轉換相對應的反轉換,來復原目前區塊的預測誤差。然後,反轉換部114將復原的預測誤差輸出至加算部116。The inverse conversion unit 114 inversely converts the conversion coefficient, which is an input from the inverse quantization unit 112, to restore the prediction error. Specifically, the inverse conversion unit 114 restores the prediction error of the current block by performing inverse conversion corresponding to the conversion of the conversion unit 106 with respect to the conversion coefficient. Then, the inverse conversion unit 114 outputs the restored prediction error to the addition unit 116.
再者,由於復原的預測誤差是因量化而喪失資訊,因此不會與減算部104所算出的預測誤差一致。亦即,於復原的預測誤差,包含有量化誤差。
[加算部]Furthermore, since the predicted prediction error is lost due to quantization, it does not match the prediction error calculated by the subtraction unit 104. That is, the prediction error in the restoration includes a quantization error.
[Additional Department]
加算部116藉由加算從反轉換部114的輸入即預測誤差與從預測控制部128的輸入即預測樣本,來重構目前區塊。然後,加算部116將重構的區塊輸出至區塊記憶體118及迴路濾波部120。重構區塊有時亦稱為局部解碼區塊。
[區塊記憶體]The adding unit 116 reconstructs the current block by adding the prediction error, which is an input from the inverse conversion unit 114, and the prediction sample input from the prediction control unit 128. Then, the adding unit 116 outputs the reconstructed block to the block memory 118 and the loop filter unit 120. Reconstructed blocks are sometimes referred to as local decoded blocks.
[block memory]
區塊記憶體118是用以儲存幀內預測所參考的區塊,且為編碼對象圖片(以下稱為目前圖片)內的區塊的記憶部。具體而言,區塊記憶體118儲存從加算部116輸出的重構區塊。
[迴路濾波部]The block memory 118 is a memory for storing a block to which intra prediction is referred, and is a block in a picture to be encoded (hereinafter referred to as a current picture). Specifically, the tile memory 118 stores the reconstructed block output from the addition unit 116.
[loop filter unit]
迴路濾波部120對於由加算部116重構的區塊,施以迴路濾波器,將已濾波的重構區塊輸出至幀記憶體122。迴路濾波器是在編碼迴路內使用的濾波器(迴路內濾波器),包含例如去區塊濾波器(DF)、樣本適用性偏移(SAO)及適應性迴路濾波器(ALF)等。The loop filter unit 120 applies a loop filter to the block reconstructed by the addition unit 116, and outputs the filtered reconstructed block to the frame memory 122. The loop filter is a filter (intra-loop filter) used in the coding loop, and includes, for example, a deblocking filter (DF), a sample applicability offset (SAO), and an adaptive loop filter (ALF).
ALF適用用以去除編碼失真的最小平方誤差濾波器,例如就目前區塊內的2´2子區塊,逐一適用根據局部梯度(gradient)的方向及活性度(activity)而從複數個濾波器中選擇的1個濾波器。ALF is applicable to the least square error filter used to remove coding distortion. For example, for the 2 ́2 sub-blocks in the current block, the filter is applied one by one according to the direction and activity of the local gradient. 1 filter selected in .
具體而言,首先子區塊(例如2´2子區塊)分類為複數個組別(例如15或25組)。子區塊的分類是根據梯度的方向及活性度來進行。例如利用梯度的方向值D(例如0~2或0~4)及梯度的活性值A(例如0~4),來算出分類值C(例如C=5D+A)。然後,根據分類值C,將子區塊分類為複數個組別(例如15或25組)。Specifically, first, sub-blocks (for example, 2 ́2 sub-blocks) are classified into a plurality of groups (for example, 15 or 25 groups). The classification of sub-blocks is based on the direction and activity of the gradient. For example, the classification value C (for example, C=5D+A) is calculated by using the gradient direction value D (for example, 0 to 2 or 0 to 4) and the gradient activity value A (for example, 0 to 4). Then, based on the classification value C, the sub-blocks are classified into a plurality of groups (for example, 15 or 25 groups).
梯度的方向值D是藉由例如比較複數個方向(例如水平、垂直及2個對角方向)的梯度來導出。又,梯度的活性值A是藉由例如加算複數個方向的梯度,量化加算結果來導出。The direction value D of the gradient is derived by, for example, comparing gradients in a plurality of directions (eg, horizontal, vertical, and 2 diagonal directions). Further, the activity value A of the gradient is derived by, for example, adding a gradient in a plurality of directions and quantizing the addition result.
根據此類分類的結果,從複數個濾波器中決定子區塊用的濾波器。Based on the results of such classification, the filters for the sub-blocks are determined from a plurality of filters.
ALF所用濾波器的形狀可利用例如圓對稱形狀。圖4A~圖4C是表示ALF所用濾波器的形狀的複數例之圖。圖4A表示5´5鑽石形狀濾波器,圖4B表示7´7鑽石形狀濾波器,圖4C表示9´9鑽石形狀濾波器。表示濾波器形狀的資訊是以圖片級別訊號化。再者,表示濾波器形狀的資訊的訊號化無須限定在圖片級別,亦可為其他級別(例如序列級別、切片級別、方塊級別、CTU級別或CU級別)。The shape of the filter used in the ALF can be, for example, a circularly symmetrical shape. 4A to 4C are diagrams showing a plurality of examples of shapes of filters used for ALF. Fig. 4A shows a 5 ́5 diamond shape filter, Fig. 4B shows a 7 ́7 diamond shape filter, and Fig. 4C shows a 9 ́9 diamond shape filter. The information representing the shape of the filter is signalized at the picture level. Furthermore, the signalization of the information indicating the shape of the filter need not be limited to the picture level, and may be other levels (for example, sequence level, slice level, block level, CTU level or CU level).
ALF的開啟/關閉是以例如圖片級別或CU級別來決定。例如就亮度而言,以CU級別決定是否適用ALF,就色差而言,以圖片級別決定是否適用ALF。表示ALF的開啟/關閉的資訊是以圖片級別或CU級別來訊號化。再者,表示ALF的開啟/關閉的資訊的訊號化無須限定在圖片級別或CU級別,亦可為其他級別(例如序列級別、切片級別、方塊級別或CTU級別)。The ALF on/off is determined by, for example, a picture level or a CU level. For example, in terms of brightness, whether or not ALF is applied is determined by the CU level, and in terms of color difference, whether or not ALF is applied is determined at the picture level. Information indicating that ALF is turned on/off is signalized at the picture level or CU level. Furthermore, the signalization of the information indicating that the ALF is turned on/off is not limited to the picture level or the CU level, and may be other levels (for example, sequence level, slice level, block level, or CTU level).
可選擇的複數個濾波器(例如從15到25的濾波器)的係數集合是以圖片級別訊號化。再者,係數集合的訊號化無須限定在圖片級別,亦可為其他級別(例如序列級別、切片級別、方塊級別、CTU級別、CU級別或子區塊級別)。
[幀記憶體]The set of coefficients for a selectable plurality of filters (e.g., filters from 15 to 25) is signaled at the picture level. Furthermore, the signalization of the coefficient set need not be limited to the picture level, and may be other levels (such as sequence level, slice level, block level, CTU level, CU level or sub-block level).
[frame memory]
幀記憶體122是用以儲存幀間預測所用的參考圖片的記憶部,有時亦稱為幀緩衝器。具體而言,幀記憶體122儲存由迴路濾波部120所濾波的重構區塊。
[幀內預測部]The frame memory 122 is a memory for storing reference pictures used for inter prediction, and is sometimes referred to as a frame buffer. Specifically, the frame memory 122 stores the reconstructed block filtered by the loop filter unit 120.
[Intra Prediction Unit]
幀內預測部124參考儲存於區塊記憶體118的目前圖片內的區塊,來進行目前區塊的幀內預測(亦稱為畫面內預測),藉此生成預測訊號(幀內預測訊號)。具體而言,幀內預測部124參考鄰接於目前區塊的區塊的樣本(例如亮度值、色差值),來進行幀內預測,藉此生成幀內預測訊號,將幀內預測訊號輸出至預測控制部128。The intra prediction unit 124 refers to the block stored in the current picture of the block memory 118 to perform intra prediction (also referred to as intra prediction) of the current block, thereby generating a prediction signal (intra prediction signal). . Specifically, the intra prediction unit 124 performs intra prediction by referring to samples (for example, luminance values and color difference values) of the blocks adjacent to the current block, thereby generating an intra prediction signal and outputting the intra prediction signal. To the prediction control unit 128.
例如幀內預測部124利用預先規定的複數種幀內預測模式中之一種,進行幀內預測。複數種幀內預測模式包含1以上的非方向性預測模式及複數種方向性預測模式。For example, the intra prediction unit 124 performs intra prediction using one of a plurality of predetermined intra prediction modes. The plurality of intra prediction modes include one or more non-directional prediction modes and a plurality of directional prediction modes.
1以上的非方向性預測模式包含例如H.265/HEVC(High-Efficiency Video Coding(高效率視訊編碼))規格(非專利文獻1)所規定的Planar(平面)預測模式及DC預測模式。The non-directional prediction mode of 1 or more includes, for example, a Planar prediction mode and a DC prediction mode defined by the H.265/HEVC (High-Efficiency Video Coding) specification (Non-Patent Document 1).
複數種方向性預測模式包含例如H.265/HEVC規格所規定的33方向的預測模式。再者,複數種方向性預測模式除了33方向以外,亦可進一步包含32方向的預測模式(合計65個方向性預測模式)。圖5A是表示幀內預測之67個幀內預測模式(2個非方向性預測模式及65個方向性預測模式)之圖。實線箭頭表示H.265/HEVC規格所規定的33方向,虛線箭頭表示追加的32方向。The plurality of directional prediction modes include, for example, a prediction mode of 33 directions specified by the H.265/HEVC specification. Further, the plurality of directional prediction modes may further include a prediction mode of 32 directions (a total of 65 directional prediction modes) in addition to the 33 direction. 5A is a diagram showing 67 intra prediction modes (two non-directional prediction modes and 65 directional prediction modes) of intra prediction. The solid arrows indicate the 33 directions defined by the H.265/HEVC specifications, and the dotted arrows indicate the additional 32 directions.
再者,於色差區塊的幀內預測中,亦可參考亮度區塊。總言之,亦可根據目前區塊的亮度成分,來預測目前區塊的色差成分。該類幀內預測有時稱為CCLM(cross-component linear model(跨組件線性模式))預測。該類參考亮度區塊的色差區塊的幀內預測模式(例如稱為CCLM模式)亦可加入作為色差區塊的幀內預測模式之一。Furthermore, in the intra prediction of the color difference block, the luminance block can also be referred to. In summary, the color difference component of the current block can also be predicted based on the brightness component of the current block. This type of intra prediction is sometimes referred to as CCLM (cross-component linear model) prediction. The intra prediction mode (for example, referred to as CCLM mode) of the color difference block of the reference luminance block may also be added as one of the intra prediction modes as the color difference block.
幀內預測部124亦可根據水平/垂直方向的參考像素的梯度,來修正幀內預測後的像素值。伴隨有該類修正的幀內預測有時稱為PDPC(position dependent intra prediction combination(位置相依之幀內預測組合))。表示PDPC有無適用的資訊(稱為例如PDPC旗標)是以CU級別來訊號化。再者,該資訊的訊號化無須限定在CU級別,亦可為其他級別(例如序列級別、圖片級別、切片級別、方塊級別或CTU級別)。
[幀間預測部]The intra prediction unit 124 can also correct the intra-predicted pixel value based on the gradient of the reference pixels in the horizontal/vertical direction. The intra prediction associated with this type of correction is sometimes referred to as PDPC (position dependent intra prediction combination). Information indicating whether or not the PDPC is applicable (referred to as a PDPC flag, for example) is signaled at the CU level. Furthermore, the signalization of the information need not be limited to the CU level, but may be other levels (such as sequence level, picture level, slice level, block level or CTU level).
[Inter Prediction Unit]
幀間預測部126參考儲存於幀記憶體122的參考圖片,且與目前圖片不同的參考圖片,來進行目前區塊的幀間預測(亦稱為畫面間預測),藉此生成預測訊號(幀間預測訊號)。幀間預測是以目前區塊或目前區塊內的子區塊(例如4´4區塊)的單位進行。例如幀間預測部126針對目前區塊或子區塊,進行參考圖片內移動估計(motion estimation)。然後,幀間預測部126利用藉由移動估計所獲得的移動資訊(例如移動向量),來進行移動補償,藉此生成目前區塊或子區塊的幀間預測訊號。然後,幀間預測部126將生成的幀間預測訊號輸出至預測控制部128。The inter prediction unit 126 refers to the reference picture stored in the frame memory 122 and the reference picture different from the current picture to perform inter prediction (also referred to as inter-picture prediction) of the current block, thereby generating a prediction signal (frame). Inter-predictive signal). Inter prediction is performed in units of current blocks or sub-blocks within the current block (eg, 4 ́4 blocks). For example, the inter prediction unit 126 performs reference intra-picture motion estimation for the current block or sub-block. Then, the inter prediction unit 126 performs motion compensation using the motion information (for example, motion vector) obtained by the motion estimation, thereby generating an inter prediction signal of the current block or sub-block. Then, the inter prediction unit 126 outputs the generated inter prediction signal to the prediction control unit 128.
用於移動補償的移動資訊被訊號化。移動向量的訊號化亦可利用預測移動向量(motion vector predictor)。總言之,亦可將移動向量與預測移動向量之間的差距訊號化。Mobile information for motion compensation is signaled. The signal vectorization of the motion vector can also utilize a motion vector predictor. In summary, the difference between the motion vector and the predicted motion vector can also be signaled.
再者,不僅可利用藉由移動估計所獲得的目前區塊的移動資訊,亦可利用鄰接區塊的移動資訊,來生成幀間預測訊號。具體而言,亦可將根據藉由移動估計所獲得的移動資訊的預測訊號、與根據鄰接區塊的移動資訊的預測訊號,予以加算權重,藉此以目前區塊內的子區塊為單位來生成幀間預測訊號。該類幀間預測(移動補償)有時稱為OBMC(overlapped block motion compensation(重疊區塊移動補償))。Furthermore, not only the movement information of the current block obtained by the motion estimation but also the movement information of the adjacent block can be used to generate the inter prediction signal. Specifically, the prediction signal according to the movement information obtained by the motion estimation and the prediction signal according to the movement information of the adjacent block may be added to the weight, thereby using the sub-block in the current block as a unit. To generate an inter prediction signal. This type of inter prediction (motion compensation) is sometimes referred to as OBMC (overlapped block motion compensation).
於該類OBMC模式,表示OBMC用子區塊的大小的資訊(例如稱為OBMC區塊大小)是以序列級別訊號化。又,表示適用OBMC模式與否的資訊(例如稱為OBMC旗標)是以CU級別來訊號化。再者,該等資訊的訊號化級別無須限定在序列級別及CU級別,亦可為其他級別(例如圖片級別、切片級別、方塊級別、CTU級別或子區塊級別)。In this type of OBMC mode, information indicating the size of the OBMC sub-blocks (for example, called OBMC block size) is signaled at the sequence level. Also, information indicating whether or not the OBMC mode is applied (for example, referred to as an OBMC flag) is signaled at the CU level. Furthermore, the level of signalization of such information need not be limited to the sequence level and CU level, and may be other levels (such as picture level, slice level, block level, CTU level or sub-block level).
更具體說明OBMC模式。圖5B及圖5C是用以說明OBMC處理的預測圖像修正處理的概要的流程圖及概念圖。More specifically describes the OBMC mode. 5B and 5C are a flowchart and a conceptual diagram for explaining an outline of a predicted image correction process of the OBMC process.
首先,利用分派給編碼對象區塊的移動向量(MV),取得一般的移動補償的預測圖像(Pred)。First, a general motion compensated predicted image (Pred) is obtained by using a motion vector (MV) assigned to a coding target block.
接著,將編碼完畢的左鄰接區塊的移動向量(MV_L),適用於編碼對象區塊,取得預測圖像(Pred_L),將前述預測圖像與Pred_L賦予權重並重疊,藉此進行預測圖像的第1次修正。Next, the motion vector (MV_L) of the encoded left adjacent block is applied to the coding target block, the predicted image (Pred_L) is obtained, and the predicted image and the Pred_L are weighted and overlapped, thereby performing the predicted image. The first revision.
同樣地,將編碼完畢的上鄰接區塊的移動向量(MV_U),適用於編碼對象區塊,取得預測圖像(Pred_U),將已進行前述第1次修正的預測圖像與Pred_U賦予權重並重疊,藉此進行預測圖像的第2次修正,將其作為最終的預測圖像。Similarly, the motion vector (MV_U) of the encoded upper adjacent block is applied to the coding target block, the predicted image (Pred_U) is obtained, and the predicted image and the Pred_U that have been subjected to the first correction are given weights. By overlapping, the second correction of the predicted image is performed, and this is used as the final predicted image.
再者,於此說明了利用左鄰接區塊及上鄰接區塊的2階段修正的方法,但亦可採用利用右鄰接區塊或下鄰接區塊,進行次數多於2階段的修正的構成。Furthermore, the method of two-stage correction using the left adjacent block and the upper adjacent block has been described here, but it is also possible to adopt a configuration in which the right adjacent block or the lower adjacent block is used to perform correction with more than two stages.
再者,進行重疊的區域非區塊全體的像素區域,亦可僅為區塊邊界附近的一部分區域。Furthermore, the pixel area in which the entire area of the overlapping area is not overlapped may be only a part of the area near the boundary of the block.
再者,於此說明了從1張參考圖片進行的預測圖像修正處理,但從複數張參考圖片修正預測圖像的情況亦同樣如此,取得從各個參考圖片修正的預測圖像後,進一步重疊獲得的預測圖像以作為最終的預測圖像。Furthermore, although the predicted image correction processing from one reference picture has been described here, the same applies to the case where the predicted image is corrected from the plurality of reference pictures, and the predicted images corrected from the respective reference pictures are acquired, and further overlapped. The obtained predicted image is taken as the final predicted image.
再者,前述處理對象區塊以預測區塊為單位,或以進一步分割預測區塊而成的子區塊為單位均可。Furthermore, the processing target block may be in units of prediction blocks, or may be in units of sub-blocks further divided into prediction blocks.
作為判定是否適用OBMC處理的方法,包括例如利用表示是否適用OBMC處理的訊號obmc_flag的方法。具體例是於編碼裝置,判定編碼對象區塊是否屬於移動複雜的區域,若屬於移動複雜的區域時,obmc_flag設定值1,適用OBMC處理而進行編碼,不屬於移動複雜的區域時,obmc_flag設定值0,不適用OBMC處理而進行編碼。另,於解碼裝置,藉由解碼串流所描述的obmc_flag,因應其值來切換是否適用OBMC處理而進行解碼。As a method of determining whether or not the OBMC processing is applied, for example, a method of indicating whether or not the OBMC-processed signal obmc_flag is applied is used. The specific example is that the encoding device determines whether the encoding target block belongs to a mobile complex region, and if it belongs to a mobile complex region, the obmc_flag setting value is 1 and is encoded by the OBMC processing, and does not belong to the mobile complex region, the obmc_flag setting value. 0, not encoded for OBMC processing. Further, in the decoding apparatus, by decoding the obmc_flag described by the stream, it is switched according to the value of whether or not the OBMC processing is applied.
再者,移動資訊未訊號化,從解碼裝置側導出亦可。例如亦可利用H.265/HEVC規格所規定的合併模式。又,例如亦可藉由在解碼裝置側進行移動估計,來導出移動資訊。此時,不利用目前區塊的像素值而進行移動估計。Furthermore, the mobile information is not signalized and may be derived from the decoding device side. For example, the merge mode specified in the H.265/HEVC specification can also be utilized. Further, for example, the mobile information may be derived by performing motion estimation on the decoding device side. At this time, the motion estimation is performed without using the pixel value of the current block.
於此,說明有關在解碼裝置側進行移動估計的模式。該在解碼裝置側進行移動估計的模式有時稱為PMMVD(pattern matched motion vector derivation(模式匹配移動向量導出))模式或FRUC(frame rate up-conversion(幀率提升轉換))模式。Here, a mode for performing motion estimation on the decoding apparatus side will be described. The mode in which motion estimation is performed on the decoding device side is sometimes referred to as a PMMVD (pattern matched motion vector derivation) mode or a FRUC (frame rate up-conversion) mode.
於圖5D表示FRUC處理的一例。首先,參考與目前區塊在空間上或時間上鄰接的編碼完畢區塊的移動向量,生成各個具有預測移動向量的複數個候補清單(亦可與合併清單共通)。接著,從登錄於候補清單的複數個候補MV之中,選擇最佳候補MV。例如算出候補清單所含的各候補的評估值,根據評估值選擇1個候補。An example of the FRUC process is shown in FIG. 5D. First, a plurality of candidate lists having predicted motion vectors (which may also be common to the merge list) are generated with reference to the motion vectors of the coded blocks that are spatially or temporally adjacent to the current block. Next, the optimal candidate MV is selected from among the plurality of candidate MVs registered in the candidate list. For example, an evaluation value of each candidate included in the candidate list is calculated, and one candidate is selected based on the evaluation value.
然後,根據選擇的候補移動向量,導出目前區塊用的移動向量。具體而言,例如選擇的候補移動向量(最佳候補MV)直接被導出作為目前區塊用的移動向量。又,例如在對應於選擇的候補移動向量的參考圖片內的位置的周邊區域,進行模式匹配,藉此導出目前區塊用的移動向量亦可。亦即,對於最佳候補MV的周邊區域,以同樣的方法進行估計,若有評估值為更佳數值的MV時,將最佳候補MV更新為前述MV,將其作為目前區塊的最終MV亦可。再者,亦可採用不實施該處理的構成。Then, based on the selected candidate motion vector, the motion vector for the current block is derived. Specifically, for example, the selected candidate motion vector (best candidate MV) is directly derived as the motion vector for the current block. Further, for example, pattern matching may be performed in the peripheral region of the position in the reference picture corresponding to the selected candidate motion vector, thereby deriving the motion vector for the current block. That is, for the peripheral region of the optimal candidate MV, the estimation is performed in the same manner. If there is an MV whose evaluation value is a better value, the optimal candidate MV is updated to the aforementioned MV as the final MV of the current block. Also. Further, a configuration in which the processing is not performed may be employed.
以子區塊為單位進行處理時,亦可採用完全同樣的處理。When processing in units of sub-blocks, the same processing can be used.
再者,評估值是藉由對應於移動向量的參考圖片內的區域、與預定區域之間的圖案匹配,求出重構圖像的差分值來算出。再者,除了差分值以外,亦可利用其以外的資訊來算出評估值。Furthermore, the evaluation value is obtained by obtaining a difference value of the reconstructed image by matching the region in the reference picture corresponding to the motion vector with the predetermined region. Furthermore, in addition to the difference value, the evaluation value can be calculated using information other than the difference value.
模式匹配利用第1模式匹配及第2模式匹配。第1模式匹配及第2模式匹配有時分別稱為雙向匹配(bilateral matching)及模板匹配(template matching)。Pattern matching utilizes the first mode matching and the second mode matching. The first mode matching and the second mode matching are sometimes referred to as bidirectional matching and template matching, respectively.
於第1模式匹配,在不同的2個參考圖片內的2個區塊,且沿著目前區塊的移動軌道(motion trajectory)的2個區塊之間,進行模式匹配。因此,於第1模式匹配,利用沿著目前區塊的移動軌道之其他參考圖片內的區域,來作為上述候補評估值算出用的預定區域。In the first pattern matching, pattern matching is performed between two blocks in different two reference pictures and between two blocks of the motion trajectory of the current block. Therefore, in the first pattern matching, the region in the other reference picture along the moving track of the current block is used as the predetermined region for calculating the candidate evaluation value.
圖6是用以說明沿著移動軌道的2個區塊間之模式匹配(雙向匹配)之一例的圖。如圖6所示,於第1模式匹配,藉由估計沿著目前區塊(Cur block)的移動軌道的2個區塊,且不同的2個參考圖片(Ref0,Ref1)內的2個區塊的配對中最匹配的配對,來導出2個移動向量(MV0,MV1)。具體而言,對於目前區塊,導出由候補MV指定的第1編碼完畢參考圖片(Ref0)內的指定位置的重構圖像、與由對稱MV指定的第2編碼完畢參考圖片(Ref1)內的指定位置的重構圖像的差分,利用獲得的差分值來算出評估值,而前述對稱MV是將前述候補MV經以顯示時間間隔進行縮放。選擇複數個候補MV之中評估值為最佳值的候補MV,來作為最終MV。Fig. 6 is a view for explaining an example of pattern matching (bidirectional matching) between two blocks along a moving track. As shown in FIG. 6, in the first mode matching, two blocks in the moving track along the current block (Cur block) are estimated, and two regions in the two different reference pictures (Ref0, Ref1) The best matching pair in the pair of blocks is used to derive 2 motion vectors (MV0, MV1). Specifically, for the current block, the reconstructed image at the specified position in the first encoded reference picture (Ref0) specified by the candidate MV and the second encoded reference picture (Ref1) specified by the symmetric MV are derived. The difference of the reconstructed image at the specified position is calculated using the obtained difference value, and the symmetric MV is scaled by the candidate MV at the display time interval. The candidate MV whose evaluation value is the best value among the plurality of candidate MVs is selected as the final MV.
假定是連續的移動軌道,指示2個參考區塊的移動向量(MV0,MV1)會與目前圖片(Cur Pic)與2個參考圖片(Ref0,Ref1)之間的時間距離(TD0,TD1)成比例。例如目前圖片在時間上位於2個參考圖片之間,從目前圖片到2個參考圖片的時間距離相等時,於第1模式匹配,導出反射對稱的雙向移動向量。Assume that it is a continuous moving track, indicating that the motion vectors (MV0, MV1) of the two reference blocks are equal to the time distance (TD0, TD1) between the current picture (Cur Pic) and the two reference pictures (Ref0, Ref1). proportion. For example, when the current picture is located between two reference pictures in time, when the time distance from the current picture to the two reference pictures is equal, the first mode matching is performed, and the bidirectional motion vector with reflection symmetry is derived.
於第2模式匹配,在目前圖片內的模板(在目前圖片內,鄰接於目前區塊的區塊(例如上及/或左鄰接區塊))與參考圖片內的區塊之間,進行模式匹配。因此,於第2模式匹配,利用目前圖片內鄰接於目前區塊的區塊,來作為上述候補評估值算出用的預定區域。In the second pattern matching, the template in the current picture (in the current picture, adjacent to the block of the current block (for example, the upper and/or left adjacent block)) and the block in the reference picture, the mode is performed. match. Therefore, in the second pattern matching, the block adjacent to the current block in the current picture is used as the predetermined area for calculating the candidate evaluation value.
圖7是用以說明目前圖片內的模板與參考圖片內的區塊之間的模式匹配(模板匹配)之一例的圖。如圖7所示,於第2模式匹配,藉由於參考圖片(Ref0)內,估計與目前圖片(Cur Pic)內鄰接於目前區塊(Cur block)的區塊最匹配的區塊,來導出目前區塊的移動向量。具體而言,對於目前區塊,導出左鄰接及上鄰接之雙方或某一方的編碼完畢區域的重構圖像、與由候補MV指定的編碼完畢參考圖片(Ref0)內的同等位置的重構圖像的差分,利用獲得的差分值算出評估值,選擇複數個MV候補之中評估值為最佳值的候補MV,來作為最佳候補MV即可。FIG. 7 is a diagram for explaining an example of pattern matching (template matching) between a template in a current picture and a block in a reference picture. As shown in FIG. 7, in the second mode matching, by estimating the block in the reference picture (Ref0) that best matches the block in the current picture (Cur Pic) adjacent to the current block (Cur block), the block is derived. The current vector of the block. Specifically, for the current block, the reconstructed image of the coded region of both the left adjacent and the upper adjacent or the adjacent one is derived, and the reconstructed image in the encoded reference picture (Ref0) specified by the candidate MV is reconstructed. The difference between the images is calculated using the obtained difference value, and the candidate MV whose evaluation value is the optimum value among the plurality of MV candidates is selected as the optimal candidate MV.
表示適用該類FRUC模式與否的資訊(例如稱為FRUC旗標)是以CU級別來訊號化。又,適用FRUC模式時(例如FRUC旗標為真時),表示模式匹配方法(第1模式匹配或第2模式匹配)的資訊(例如稱為FRUC模式旗標)是以CU級別來訊號化。再者,該等資訊的訊號化無須限定在CU級別,亦可為其他級別(例如序列級別、圖片級別、切片級別、方塊級別、CTU級別或子區塊級別)。Information indicating whether or not this type of FRUC mode is applied (for example, called the FRUC flag) is signaled at the CU level. Further, when the FRUC mode is applied (for example, when the FRUC flag is true), information indicating the pattern matching method (first pattern matching or second pattern matching) (for example, referred to as FRUC mode flag) is signaled at the CU level. Furthermore, the signalization of such information need not be limited to the CU level, but may be other levels (such as sequence level, picture level, slice level, block level, CTU level or sub-block level).
於此,說明根據假定等速直線運動的模型,來導出移動向量的模式。該模式有時稱為BIO(bi-directional optical flow(雙向光流))模式。Here, a mode in which a motion vector is derived based on a model of assuming constant-speed linear motion will be described. This mode is sometimes referred to as BIO (bi-directional optical flow) mode.
圖8是用以說明假定等速直線運動的模型之圖。於圖8,(vx ,vy )表示速度向量,t0 、t1 分別表示目前圖片(Cur Pic)與2個參考圖片(Ref0 ,Ref1 )之間的時間距離。(MVx0 ,MVy0 )表示對應於參考圖片Ref0 的移動向量,(MVx1 ,MVy1 )表示對應於參考圖片Ref1 的移動向量。Fig. 8 is a view for explaining a model assuming constant-speed linear motion. In Fig. 8, (v x , v y ) represents a velocity vector, and t 0 and t 1 respectively represent the temporal distance between the current picture (Cur Pic) and two reference pictures (Ref 0 , Ref 1 ). (MVx 0 , MVy 0 ) represents a motion vector corresponding to the reference picture Ref 0 , and (MVx 1 , MVy 1 ) represents a motion vector corresponding to the reference picture Ref 1 .
此時,在速度向量(vx
,vy
)的等度直線運動的假定下,(MVx0
,MVy0
)及(MVx1
,MVy1
)分別表示為(vx
τ0
,vy
τ0
)及(-vx
τ1
,-vy
τ1
),以下光流等式(1)成立。
[數1]
At this time, under the assumption of the iso-linear motion of the velocity vector (v x , v y ), (MVx 0 , MVy 0 ) and (MVx 1 , MVy 1 ) are expressed as (v x τ 0 , v y τ 0 , respectively). And (-v x τ 1 , -v y τ 1 ), the following optical flow equation (1) holds.
[Number 1]
於此,I(k) 表示移動補償後的參考圖像k(k=0,1)的亮度值。該光流等式表示(i)、(ii)與(iii)的和等於零,其中(i)亮度值的時間微分,(ii)水平方向的速度及參考圖像的空間梯度的水平分量的積,(iii)垂直方向的速度及參考圖像的空間梯度的垂直分量的積。根據該光流等式與赫米特內插法(Hermite interpolation)的組合,以像素單位修正從合併清單等所獲得的區塊單位的移動向量。Here, I (k) represents the luminance value of the reference image k (k = 0, 1) after the motion compensation. The optical flow equation represents the sum of (i), (ii) and (iii) equal to zero, where (i) the time differential of the luminance value, (ii) the product of the horizontal velocity and the horizontal component of the spatial gradient of the reference image. (iii) the product of the velocity in the vertical direction and the vertical component of the spatial gradient of the reference image. According to the combination of the optical flow equation and the Hermitian interpolation, the motion vector of the block unit obtained from the merge list or the like is corrected in units of pixels.
再者,採用與根據假定等速直線運動的模型的移動向量導出不同的方法,在解碼裝置側導出移動向量亦可。例如根據複數個鄰接區塊的移動向量,以子區塊為單位來導出移動向量亦可。Furthermore, it is also possible to derive a motion vector on the decoding device side by using a different method from the motion vector derivation of the model based on the assumed constant velocity linear motion. For example, according to the motion vector of a plurality of adjacent blocks, the motion vector may be derived in units of sub-blocks.
於此,說明根據複數個鄰接區塊的移動向量,以子區塊為單位來導出移動向量之模式。該模式有時稱為仿射移動補償預測(affine motion compensation prediction)模式。Here, a mode in which a motion vector is derived in units of sub-blocks according to a motion vector of a plurality of adjacent blocks will be described. This mode is sometimes referred to as an affine motion compensation prediction mode.
圖9A是用以說明根據複數個鄰接區塊的移動向量來導出子區塊單位的移動向量之圖。於圖9A,目前區塊包含16之4´4子區塊。於此,根據鄰接區塊的移動向量,來導出目前區塊左上角控制點的移動向量v0
,根據鄰接子區塊的移動向量,來導出目前區塊右上角控制點的移動向量v1
。然後,利用2個移動向量v0
及v1
,藉由下式(2)來導出目前區塊內的各子區塊的移動向量(vx
,vy
)。
[數2]
FIG. 9A is a diagram for explaining a motion vector of a sub-block unit derived from a motion vector of a plurality of adjacent blocks. FIG. In Figure 9A, the current block contains 16 4 ́4 sub-blocks. Here, the motion vector v 0 of the control point of the upper left corner of the current block is derived according to the motion vector of the adjacent block, and the motion vector v 1 of the control point of the upper right corner of the current block is derived according to the motion vector of the adjacent sub-block. Then, using two motion vectors v 0 and v 1 , the motion vectors (v x , v y ) of the respective sub-blocks in the current block are derived by the following equation (2).
[Number 2]
於此,x及y分別表示子區塊的水平位置及垂直位置,w表示預先決定的加權係數。Here, x and y respectively indicate the horizontal position and the vertical position of the sub-block, and w represents a predetermined weighting coefficient.
於該類仿射移動補償預測模式,亦可包含左上及右上角控制點的移動向量導出方法不同的數種模式。表示該類仿射移動補償預測模式的資訊(例如稱為仿射旗標)是以CU級別來訊號化。再者,表示該仿射移動補償預測模式的資訊的訊號化無須限定在CU級別,亦可為其他級別(例如序列級別、圖片級別、切片級別、方塊級別、CTU級別或子區塊級別)。
[預測控制部]The affine motion compensation prediction mode may also include several modes in which the motion vector derivation methods of the upper left and upper right control points are different. Information indicating this type of affine motion compensated prediction mode (for example, called an affine flag) is signaled at the CU level. Furthermore, the signalization of the information indicating the affine motion compensation prediction mode need not be limited to the CU level, and may be other levels (eg, sequence level, picture level, slice level, block level, CTU level, or sub-block level).
[Predictive Control Department]
預測控制部128選擇幀內預測訊號及幀間預測訊號的任一者,將選擇的訊號作為預測訊號而輸出至減算部104及加算部116。The prediction control unit 128 selects any one of the intra prediction signal and the inter prediction signal, and outputs the selected signal as a prediction signal to the subtraction unit 104 and the addition unit 116.
於此,說明藉由合併模式來導出編碼對象圖片的移動向量之例。圖9B是用以說明利用合併模式的移動向量導出處理的概要的圖。Here, an example of deriving a motion vector of a coding target picture by a merge mode will be described. FIG. 9B is a diagram for explaining an outline of a motion vector derivation process using a merge mode.
首先,生成登錄有預測MV的候補的預測MV清單。預測MV的候補包括:空間鄰接預測MV,其為空間上位於編碼對象區塊周邊的複數個編碼完畢區塊所具有的MV;時間鄰接MV,其為投影有編碼完畢參考圖片的編碼完畢區塊的位置附近的區塊所具有的MV;結合預測MV,其為組合空間鄰接預測MV與時間鄰接預測MV的MV值而生成的MV;及值為零的MV,即零預測MV等。First, a list of predicted MVs to which candidates for the predicted MV are registered is generated. The candidate for predicting the MV includes: a spatial adjacency prediction MV, which is an MV of a plurality of coded blocks spatially located around the block of the coding target; a temporally adjacent MV, which is a coded block in which the encoded reference picture is projected. The MV of the block near the location; combined with the prediction MV, which is the MV generated by combining the MV value of the spatial neighbor prediction MV with the temporal neighbor prediction MV; and the MV with zero value, that is, the zero prediction MV.
接著,藉由從登錄於預測MV清單的複數個預測MV之中,選擇1個預測MV,來決定作為編碼對象區塊的MV。Next, by selecting one prediction MV from among a plurality of prediction MVs registered in the prediction MV list, the MV as the coding target block is determined.
進而於可變長度編碼部,將表示選擇的預測MV為何的訊號merge_idx,描述於串流而編碼。Further, the variable length coding unit encodes the signal merge_idx indicating the selected prediction MV in the stream.
再者,圖9B所說明的登錄於預測MV清單的預測MV為一例,亦可個數與圖中的個數不同,亦或為不包含圖中的預測MV的一部分種類的構成,亦或為追加有圖中的預測MV的種類以外的預測MV的構成。Further, the prediction MV registered in the prediction MV list described in FIG. 9B is an example, and the number may be different from the number in the figure, or may be a configuration that does not include a part of the prediction MV in the figure, or The configuration of the prediction MV other than the type of the prediction MV in the figure is added.
再者,亦可藉由利用合併模式導出的編碼對象區塊的MV,進行後述的DMVR處理,來決定最終的MV。Furthermore, the final MV may be determined by performing DMVR processing, which will be described later, by using the MV of the encoding target block derived by the merge mode.
於此,說明利用DMVR處理來決定MV之例。Here, an example in which MV is determined by DMVR processing will be described.
圖9C是用以說明DMVR處理的概要的概念圖。9C is a conceptual diagram for explaining an outline of DMVR processing.
首先,將設定於處理對象區塊的最佳MVP作為候補MV,按照前述候補MV,從L0方向的處理完畢圖片之第1參考圖片、及從L1方向的處理完畢圖片之第2參考圖片,分別取得參考像素,取得各參考像素的平均來生成模板。First, the optimal MVP set in the processing target block is used as the candidate MV, and the first reference picture of the processed picture in the L0 direction and the second reference picture of the processed picture in the L1 direction are respectively selected according to the candidate MV. A reference pixel is obtained, and an average of each reference pixel is obtained to generate a template.
接著,利用前述模板,分別估計第1參考圖片及第2參考圖片的候補MV的周邊區域,決定成本最小的MV來作為最終的MV。再者,成本值是利用模板的各像素值與估計區域的各像素值的差分值及MV值來算出。Next, using the template, the peripheral regions of the candidate MVs of the first reference picture and the second reference picture are respectively estimated, and the MV with the lowest cost is determined as the final MV. Further, the cost value is calculated by using a difference value and an MV value between each pixel value of the template and each pixel value of the estimated region.
再者,於編碼裝置及解碼裝置,於此所說明的處理的概要基本上是共通。Furthermore, the outlines of the processes described herein are basically common to both the encoding device and the decoding device.
再者,不利用於此說明的處理本身,亦可利用其他處理,只要是可估計候補MV的周邊並導出最終的MV的處理均可。Further, it is disadvantageous that the processing itself used in the above description may be performed by any other processing as long as it is a process in which the periphery of the candidate MV can be estimated and the final MV is derived.
於此,說明利用LIC處理來生成預測圖像的模式。Here, a mode in which a predicted image is generated by the LIC processing will be described.
圖9D是用以說明利用LIC處理的亮度修正處理的預測圖像生成方法的概要的圖。9D is a diagram for explaining an outline of a predicted image generating method of brightness correction processing by LIC processing.
首先,從編碼完畢圖片之參考圖片,導出用以取得對應於編碼對象區塊的參考圖像的MV。First, an MV for obtaining a reference image corresponding to the block of the encoding target is derived from the reference picture of the encoded picture.
接著,對於編碼對象區塊,利用左鄰接及上鄰接的編碼完畢周邊參考區域的亮度像素值、及由MV指定的參考圖片內的同等位置的亮度像素值,擷取表示亮度值在參考圖片及編碼對象圖片如何變化的資訊,算出亮度修正參數。Next, for the coding target block, the luminance pixel value of the coded peripheral reference region adjacent to the left adjacent and the upper adjacent and the luminance pixel value of the equivalent position in the reference picture specified by the MV are used to extract the luminance value in the reference picture and The information on how the image of the encoding object changes is calculated, and the brightness correction parameter is calculated.
對於由MV指定的參考圖片內的參考圖像,利用前述亮度修正參數進行亮度修正處理,藉此生成對於編碼對象區塊的預測圖像。For the reference image in the reference picture specified by the MV, the brightness correction processing is performed using the aforementioned brightness correction parameter, thereby generating a predicted image for the encoding target block.
再者,圖9D的前述周邊參考區域的形狀為一例,亦可利用該形狀以外的形狀。Further, the shape of the peripheral reference region in FIG. 9D is an example, and a shape other than the shape may be used.
又,於此說明了從1張參考圖片生成預測圖像的處理,但從複數張參考圖片生成預測圖像的情況亦同樣如此,對從各個參考圖片取得的參考圖像,以同樣方法進行亮度修正處理後,生成預測圖像。Further, although the process of generating a predicted image from one reference picture has been described here, the same applies to the case of generating a predicted image from a plurality of reference pictures, and the brightness is performed in the same manner for the reference picture obtained from each reference picture. After the correction process, a predicted image is generated.
作為判定是否適用LIC處理的方法,包括例如利用表示是否適用LIC處理的訊號lic_flag的方法。具體例是於編碼裝置,判定編碼對象區塊是否屬於發生亮度變化的區域,屬於發生亮度變化的區域時,lic_flag設定值1,適用LIC處理而進行編碼,不屬於發生亮度變化的區域時,lic_flag設定值0,不適用LIC處理而進行編碼。另,於解碼裝置,藉由解碼串流所描述的lic_flag,因應其值來切換是否適用LIC處理而進行解碼。As a method of determining whether or not the LIC processing is applied, for example, a method of indicating whether or not the LIC processing signal lic_flag is applied is used. In the specific example, the encoding device determines whether or not the encoding target block belongs to the region where the luminance change occurs, and belongs to the region where the luminance change occurs, and the lic_flag is set to 1 and is encoded by the LIC processing, and is not included in the region where the luminance change occurs, lic_flag The setting value is 0, and encoding is not applied to LIC processing. Further, in the decoding apparatus, by decoding the lic_flag described by the stream, it is switched according to the value of whether or not the LIC processing is applied.
作為判定是否適用LIC處理的方法,亦包括例如按照周邊區塊是否適用LIC處理來判定的方法。作為具體一例,編碼對象區塊為合併模式時,判定在合併處理的MV導出時所選擇的周邊的編碼完畢區塊,是否適用了LIC處理而編碼,因應其結果來切換是否適用LIC處理而進行編碼。再者,此例的情況下,解碼的處理亦完全同樣。
[解碼裝置的概要]The method of determining whether or not the LIC processing is applied also includes, for example, a method of determining whether or not the peripheral block is LIC-processed. As a specific example, when the encoding target block is in the merge mode, it is determined whether or not the surrounding coded block selected at the time of MV export of the merge process is encoded by the LIC process, and the result is switched whether or not the LIC process is applied. coding. Furthermore, in the case of this example, the decoding process is also identical.
[Summary of decoding device]
接著,說明可解碼從上述編碼裝置100輸出的編碼訊號(編碼位元串流)的解碼裝置的概要。圖10是表示實施形態1之解碼裝置200的功能構成之方塊圖。解碼裝置200是以區塊為單位解碼動態圖像/圖像的動態圖像/圖像解碼裝置。Next, an outline of a decoding device that can decode the encoded signal (encoded bit stream) output from the encoding device 100 will be described. FIG. 10 is a block diagram showing a functional configuration of a decoding device 200 according to the first embodiment. The decoding device 200 is a moving image/image decoding device that decodes a moving image/image in units of blocks.
如圖10所示,解碼裝置200具備熵解碼部202、反量化部204、反轉換部206、加算部208、區塊記憶體210、迴路濾波部212、幀記憶體214、幀內預測部216、幀間預測部218及預測控制部220。As shown in FIG. 10, the decoding apparatus 200 includes an entropy decoding unit 202, an inverse quantization unit 204, an inverse conversion unit 206, an addition unit 208, a block memory 210, a loop filter unit 212, a frame memory 214, and an intra prediction unit 216. The inter prediction unit 218 and the prediction control unit 220.
解碼裝置200藉由例如通用處理器及記憶體來實現。此時,由處理器執行儲存於記憶體的軟體程式時,處理器是作為熵解碼部202、反量化部204、反轉換部206、加算部208、迴路濾波部212、幀內預測部216、幀間預測部218及預測控制部220發揮功能。又,解碼裝置200亦可作為對應於熵解碼部202、反量化部204、反轉換部206、加算部208、迴路濾波部212、幀內預測部216、幀間預測部218及預測控制部220的專用的1以上的電子電路來實現。The decoding device 200 is realized by, for example, a general purpose processor and a memory. At this time, when the processor executes the software program stored in the memory, the processor functions as the entropy decoding unit 202, the inverse quantization unit 204, the inverse conversion unit 206, the addition unit 208, the loop filter unit 212, and the intra prediction unit 216. The inter prediction unit 218 and the prediction control unit 220 function. Further, the decoding device 200 may correspond to the entropy decoding unit 202, the inverse quantization unit 204, the inverse conversion unit 206, the addition unit 208, the loop filter unit 212, the intra prediction unit 216, the inter prediction unit 218, and the prediction control unit 220. A dedicated one or more electronic circuits are implemented.
以下說明解碼裝置200所含的各構成要件。
[熵解碼部]Hereinafter, each constituent element included in the decoding device 200 will be described.
[Entropy decoding unit]
熵解碼部202將編碼位元串流予以熵解碼。具體而言,熵解碼部202例如從編碼位元串流算數解碼為二值訊號。然後,熵解碼部202將二值訊號予以多值化(debinarize)。藉此,熵解碼部202以區塊為單位,將量化係數輸出至反量化部204。
[反量化部]The entropy decoding unit 202 entropy decodes the encoded bit stream. Specifically, the entropy decoding unit 202 decodes, for example, a coded bit stream number into a binary signal. Then, the entropy decoding unit 202 debinarizes the binary signal. Thereby, the entropy decoding unit 202 outputs the quantized coefficients to the inverse quantization unit 204 in units of blocks.
[Anti-quantization department]
反量化部204將來自熵解碼部202的輸入,即解碼對象區塊(以下稱為目前區塊)的量化係數,予以反量化。具體而言,反量化部204針對目前區塊的各個量化係數,根據對應於該量化係數的量化參數,將該量化係數予以反量化。然後,反量化部204將目前區塊之已被反量化的量化係數(亦即轉換係數),輸出至反轉換部206。
[反轉換部]The inverse quantization unit 204 inversely quantizes the input coefficient from the entropy decoding unit 202, that is, the quantization coefficient of the decoding target block (hereinafter referred to as the current block). Specifically, the inverse quantization unit 204 inversely quantizes the quantized coefficients according to the quantization parameters corresponding to the quantized coefficients for the respective quantized coefficients of the current block. Then, the inverse quantization unit 204 outputs the quantized coefficients (i.e., conversion coefficients) of the current block that have been inversely quantized to the inverse conversion unit 206.
[Anti-conversion department]
反轉換部206藉由將來自反量化部204的輸入即轉換係數,予以反轉換,來復原預測誤差。The inverse conversion unit 206 inversely converts the conversion coefficient, which is an input from the inverse quantization unit 204, to restore the prediction error.
例如表示從編碼位元串流解讀的資訊適用EMT或AMT時(例如AMT旗標為真),反轉換部206根據表示解讀的轉換類型的資訊,來將目前區塊的轉換係數予以反轉換。For example, when the information interpreted from the encoded bit stream is applied to EMT or AMT (for example, the AMT flag is true), the inverse conversion unit 206 inversely converts the conversion coefficient of the current block based on the information indicating the type of conversion of the interpretation.
又,例如表示從編碼位元串流解讀的資訊適用NSST時,反轉換部206對轉換係數適用反再轉換。
[加算部]Further, for example, when the information interpreted from the encoded bit stream is applied to the NSST, the inverse conversion unit 206 applies inverse retransformation to the conversion coefficient.
[Additional Department]
加算部208藉由加算從反轉換部206輸入的預測誤差與從預測控制部220輸入的預測樣本,來重構目前區塊。然後,加算部208將重構的區塊輸出至區塊記憶體210及迴路濾波部212。
[區塊記憶體]The addition unit 208 reconstructs the current block by adding the prediction error input from the inverse conversion unit 206 and the prediction sample input from the prediction control unit 220. Then, the adding unit 208 outputs the reconstructed block to the block memory 210 and the loop filter unit 212.
[block memory]
區塊記憶體210是用以儲存幀內預測所參考的區塊,且為解碼對象圖片(以下稱為目前圖片)內的區塊的記憶部。具體而言,區塊記憶體210儲存從加算部208輸出的重構區塊。
[迴路濾波部]The tile memory 210 is a memory for storing a block to which intra prediction is referred, and is a block in a decoding target picture (hereinafter referred to as a current picture). Specifically, the tile memory 210 stores the reconstructed block output from the addition unit 208.
[loop filter unit]
迴路濾波部212對於由加算部208重構的區塊,施以迴路濾波器,將已濾波的重構區塊輸出至幀記憶體214及顯示裝置等。The loop filter unit 212 applies a loop filter to the block reconstructed by the adding unit 208, and outputs the filtered reconstructed block to the frame memory 214, the display device, and the like.
表示從編碼位元串流解讀的ALF之開啟/關閉的資訊,表示ALF開啟時,根據局部梯度的方向及活性度而從複數個濾波器中選擇的1個濾波器,選擇的濾波器適用於重構區塊。
[幀記憶體]The information indicating the on/off of the ALF read from the encoded bit stream indicates that one filter is selected from the plurality of filters according to the direction and the degree of activity of the local gradient when the ALF is turned on, and the selected filter is applied to Reconstruct the block.
[frame memory]
幀記憶體214是用以儲存幀間預測所用的參考圖片的記憶部,有時亦稱為幀緩衝器。具體而言,幀記憶體214儲存由迴路濾波部212所濾波的重構區塊。
[幀內預測部]The frame memory 214 is a memory for storing reference pictures used for inter prediction, and is sometimes referred to as a frame buffer. Specifically, the frame memory 214 stores the reconstructed block filtered by the loop filter unit 212.
[Intra Prediction Unit]
幀內預測部216根據從編碼位元串流解讀的幀內預測模式,參考儲存於區塊記憶體210的目前圖片內的區塊,來進行幀內預測,藉此生成預測訊號(幀內預測訊號)。具體而言,幀內預測部216參考鄰接於目前區塊的區塊的樣本(例如亮度值、色差值),來進行幀內預測,藉此生成幀內預測訊號,將幀內預測訊號輸出至預測控制部220。The intra prediction unit 216 performs intra prediction by referring to a block stored in the current picture of the block memory 210 based on the intra prediction mode interpreted from the encoded bit stream, thereby generating a prediction signal (intra prediction) Signal)). Specifically, the intra prediction unit 216 performs intra prediction by referring to samples (for example, luminance values and color difference values) of the blocks adjacent to the current block, thereby generating an intra prediction signal and outputting the intra prediction signal. To the prediction control unit 220.
再者,於色差區塊的幀內預測中,選擇參考亮度區塊的幀內預測模式時,幀內預測部216根據目前區塊的亮度成分,來預測目前區塊的色差成分亦可。Furthermore, when the intra prediction mode of the reference luminance block is selected in the intra prediction of the color difference block, the intra prediction unit 216 predicts the color difference component of the current block based on the luminance component of the current block.
又,從編碼位元串流解讀的資訊表示適用PDPC時,幀內預測部216根據水平/垂直方向的參考像素的梯度,來修正幀內預測後的像素值。
[幀間預測部]Further, when the information interpreted from the encoded bit stream indicates that the PDPC is applied, the intra prediction unit 216 corrects the intra-predicted pixel value based on the gradient of the reference pixels in the horizontal/vertical direction.
[Inter Prediction Unit]
幀間預測部218參考儲存於幀記憶體214的參考圖片,預測目前區塊。預測是以目前區塊或目前區塊內的子區塊(例如4´4區塊)的單位進行。例如幀間預測部218利用從編碼位元串流解讀的移動資訊(例如移動向量),來進行移動補償,藉此生成目前區塊或子區塊的幀間預測訊號,並將幀間預測訊號輸出至預測控制部220。The inter prediction section 218 refers to the reference picture stored in the frame memory 214 to predict the current block. The prediction is made in units of the current block or sub-blocks within the current block (for example, 4 ́4 blocks). For example, the inter prediction unit 218 performs motion compensation using motion information (for example, motion vector) interpreted from the encoded bit stream, thereby generating an inter prediction signal of the current block or sub-block, and inter-predicting the signal. The output is output to the prediction control unit 220.
再者,從編碼位元串流解讀的資訊表示適用OBMC模式時,幀間預測部218不僅可利用藉由移動估計所獲得的目前區塊的移動資訊,亦可利用鄰接區塊的移動資訊,來生成幀間預測訊號。Furthermore, when the information interpreted from the encoded bit stream indicates that the OBMC mode is applied, the inter prediction unit 218 can use not only the mobile information of the current block obtained by the motion estimation but also the mobile information of the adjacent block. To generate an inter prediction signal.
又,從編碼位元串流解讀的資訊表示適用FRUC模式時,幀間預測部218按照從編碼串流解讀的模式匹配的方法(雙向匹配或模板匹配),來進行移動估計,藉此導出移動資訊。然後,幀間預測部218利用導出的移動資訊來進行移動補償。Further, when the information interpreted from the encoded bit stream indicates that the FRUC mode is applied, the inter predicting unit 218 performs motion estimation in accordance with the pattern matching method (bidirectional matching or template matching) interpreted from the encoded stream, thereby deriving the movement. News. Then, the inter predicting unit 218 performs motion compensation using the derived moving information.
又,幀間預測部218在適用BIO模式時,根據假定等速直線運動的模型,來導出移動向量。又,從編碼位元串流解讀的資訊表示適用仿射移動補償預測模式時,幀間預測部218根據複數個鄰接區塊的移動向量,以子區塊為單位來導出移動向量。
[預測控制部]Further, when the BIO mode is applied, the inter predicting unit 218 derives a motion vector based on a model that assumes a constant-speed linear motion. Further, when the information interpreted from the encoded bit stream indicates that the affine motion compensation prediction mode is applied, the inter prediction unit 218 derives the motion vector in units of sub-blocks based on the motion vectors of the plurality of adjacent blocks.
[Predictive Control Department]
預測控制部220選擇幀內預測訊號及幀間預測訊號的任一者,將選擇的訊號作為預測訊號而輸出至加算部208。
[去區塊濾波處理的第1態樣]The prediction control unit 220 selects any one of the intra prediction signal and the inter prediction signal, and outputs the selected signal as a prediction signal to the addition unit 208.
[The first aspect of deblocking filtering processing]
圖11是由編碼裝置100所含的迴路濾波部120進行的去區塊濾波處理的第1態樣的流程圖。圖11所示的處理是就鄰接的2個區塊的每區塊邊界進行。再者,以下主要說明編碼裝置100所含的迴路濾波部120的動作,但解碼裝置200所含的迴路濾波部212的動作亦同。FIG. 11 is a flowchart showing a first aspect of the deblocking filtering process performed by the loop filter unit 120 included in the encoding device 100. The processing shown in Fig. 11 is performed for each block boundary of two adjacent blocks. In the following, the operation of the loop filter unit 120 included in the encoding device 100 will be mainly described, but the operation of the loop filter unit 212 included in the decoding device 200 is also the same.
首先,迴路濾波部120是於去區塊濾波處理中,利用位於處理對象的區塊邊界,即對象邊界兩側的區塊的資訊,算出稱為Bs之值(S201)。接著,迴路濾波部120因應Bs之值,決定是否於對象邊界進行去區塊濾波處理(S202及S203)。First, in the deblocking filtering process, the loop filter unit 120 calculates a value called Bs by using the block boundary located at the processing target, that is, the information of the block on both sides of the object boundary (S201). Next, the loop filter unit 120 determines whether or not to perform deblocking filtering processing on the target boundary in accordance with the value of Bs (S202 and S203).
具體而言,迴路濾波部120是於Bs=2時(S202,YES(是)且S203,YES),對於亮度及色差雙方進行去區塊濾波處理(S204及S205)。又,迴路濾波部120是於Bs=1時(S202,YES且S203,NO(否)),對於亮度進行去區塊濾波處理(S206),對於色差不進行去區塊濾波處理。又,迴路濾波部120是於Bs=0時(S202,NO),對於亮度及色差的任一者均不進行去區塊濾波處理。Specifically, when the loop filter unit 120 is Bs=2 (S202, YES (YES) and S203, YES), both the luminance and the chrominance are subjected to deblocking processing (S204 and S205). Further, when the loop filter unit 120 is Bs=1 (S202, YES, S203, NO), the luminance is subjected to deblocking processing (S206), and the chrominance is not subjected to deblocking processing. Further, when the loop filter unit 120 is Bs=0 (S202, NO), the deblocking filtering process is not performed on any of luminance and chrominance.
又,圖12是表示算出Bs的方法例的圖。於HEVC的去區塊濾波處理的Bs算出中,採用如下4個條件:(1)跨越邊界相鄰接的2個區塊中至少一區塊為幀內預測;(2)跨越邊界相鄰接的2個區塊中至少一區塊包含優勢的DCT係數;(3)跨越邊界相鄰接的2個區塊的移動向量的差分為臨限值以上;(4)於跨越邊界相鄰接的2個區塊,移動向量(MV)的數目或參考圖像不同。相對於此,於第1態樣,加入條件(5)於跨越邊界相鄰接的2個區塊中至少一區塊,進行LIC處理。再者,(5)於跨越邊界相鄰接的2個區塊中至少一區塊,進行LIC處理時,Bs值若非0即可,1或2均可。FIG. 12 is a diagram showing an example of a method of calculating Bs. In the Bs calculation of the deblocking filtering process of HEVC, the following four conditions are adopted: (1) at least one of the two blocks adjacent to the boundary is intra prediction; (2) adjacent to the boundary At least one of the two blocks contains dominant DCT coefficients; (3) the difference between the motion vectors of the two blocks adjacent to the boundary is above the threshold; (4) adjacent to the boundary crossing Two blocks, the number of moving vectors (MV) or the reference image is different. On the other hand, in the first aspect, the condition (5) is added to at least one of the two blocks adjacent to each other across the boundary, and the LIC processing is performed. Furthermore, (5) at least one of the two blocks adjacent to the boundary, when performing LIC processing, the Bs value may be 0 or 1 or 2.
圖13是於符合條件(5)時,設定為Bs=1的情況下的Bs算出處理(S201)的流程圖。圖14表示算出此時的Bs的方法例的圖。FIG. 13 is a flowchart of Bs calculation processing (S201) when Bs=1 is set when the condition (5) is satisfied. FIG. 14 is a view showing an example of a method of calculating Bs at this time.
首先,迴路濾波部120判定是否符合條件(1)(S211)。符合條件(1)時(S211,YES),迴路濾波部120設定為Bs=2(S213)。First, the loop filter unit 120 determines whether or not the condition (1) is satisfied (S211). When the condition (1) is satisfied (S211, YES), the loop filter unit 120 is set to Bs=2 (S213).
不符合條件(1)時(S211,NO),迴路濾波部120判定是否符合條件(2)~(5)中至少一條件(S212)。符合條件(2)~(5)中至少一條件時(S212,YES),迴路濾波部120設定為Bs=1(S214)。不符合條件(2)~(5)中任一條件時(S212,NO),迴路濾波部120設定為Bs=0(S215)。When the condition (1) is not satisfied (S211, NO), the loop filter unit 120 determines whether or not at least one of the conditions (2) to (5) is satisfied (S212). When at least one of the conditions (2) to (5) is satisfied (S212, YES), the loop filter unit 120 sets Bs=1 (S214). When any of the conditions (2) to (5) is not satisfied (S212, NO), the loop filter unit 120 sets Bs = 0 (S215).
圖15是於符合條件(5)時,設定為Bs=2的情況下的Bs算出處理(S201)的流程圖。圖16表示算出此時的Bs的方法例的圖。FIG. 15 is a flowchart of Bs calculation processing (S201) when Bs=2 is set when the condition (5) is satisfied. FIG. 16 is a view showing an example of a method of calculating Bs at this time.
首先,迴路濾波部120判定是否符合條件(1)及(5)的至少一方(S211A)。符合條件(1)及(5)的至少一方時(S211A,YES),迴路濾波部120設定為Bs=2(S213)。First, the loop filter unit 120 determines whether or not at least one of the conditions (1) and (5) is satisfied (S211A). When at least one of the conditions (1) and (5) is satisfied (S211A, YES), the loop filter unit 120 is set to Bs=2 (S213).
不符合條件(1)及(5)的任一條件時(S211A,NO),迴路濾波部120判定是否符合條件(2)~(4)中至少一條件(S212A)。符合條件(2)~(4)中至少一條件時(S212A,YES),迴路濾波部120設定為Bs=1(S214)。不符合條件(2)~(4)中任一條件時(S212A,NO),迴路濾波部120設定為Bs=0(S215)。
[LIC處理]When any of the conditions (1) and (5) is not satisfied (S211A, NO), the loop filter unit 120 determines whether or not at least one of the conditions (2) to (4) is satisfied (S212A). When at least one of the conditions (2) to (4) is satisfied (S212A, YES), the loop filter unit 120 sets Bs=1 (S214). When any of the conditions (2) to (4) is not satisfied (S212A, NO), the loop filter unit 120 sets Bs=0 (S215).
[LIC processing]
已利用圖9D說明了LIC處理(亮度修正處理),以下說明其詳細。The LIC processing (brightness correction processing) has been described using FIG. 9D, and the details thereof will be described below.
首先,幀間預測部126從編碼完畢圖片即參考圖片,導出用以取得對應於編碼對象區塊的參考圖像的移動向量。First, the inter predicting unit 126 derives a motion vector for acquiring a reference image corresponding to the encoding target block from the reference picture which is the encoded picture.
接著,幀間預測部126對於編碼對象區塊,利用左鄰接及上鄰接的編碼完畢周邊參考區域的亮度像素值、及由移動向量指定的參考圖片內的同等位置的亮度像素值,擷取表示亮度值在參考圖片及編碼對象圖片如何變化的資訊,算出亮度修正參數。例如編碼對象圖片內的周邊參考區域內的某像素的亮度像素值設為p0,與該像素同等位置的參考圖片內的周邊參考區域內的像素的亮度像素值設為p1。幀間預測部126對於周邊參考區域內的複數個像素,算出使A×p1+B=p0最佳化的係數A及B,來作為亮度修正參數。Next, the inter prediction unit 126 uses the luma pixel values of the left adjacent and upper adjacent coded peripheral reference regions and the luminance pixel values of the equivalent positions in the reference picture specified by the motion vector for the encoding target block. The brightness correction parameter is calculated based on the information of how the reference picture and the coding target picture change. For example, the luminance pixel value of a certain pixel in the peripheral reference region in the encoding target picture is p0, and the luminance pixel value of the pixel in the peripheral reference region in the reference picture at the same position as the pixel is set to p1. The inter predicting unit 126 calculates coefficients A and B which optimize A × p1 + B = p0 for a plurality of pixels in the peripheral reference region as luminance correction parameters.
接著,幀間預測部126對於由移動向量指定的參考圖片內的參考圖像,利用亮度修正參數進行亮度修正處理,藉此生成對於編碼對象區塊的預測圖像。例如參考圖像內的亮度像素值設為p2,亮度修正處理後的預測圖像的亮度像素值設為p3。幀間預測部126對於參考圖像內的各像素,算出A×p2+B=p3,藉此生成亮度修正處理後的預測圖像。Next, the inter predicting unit 126 performs luminance correction processing on the reference image in the reference picture specified by the motion vector using the luminance correction parameter, thereby generating a predicted image for the encoding target block. For example, the luminance pixel value in the reference image is set to p2, and the luminance pixel value of the predicted image after the luminance correction processing is set to p3. The inter prediction unit 126 calculates A×p2+B=p3 for each pixel in the reference image, thereby generating a predicted image after the luminance correction processing.
再者,圖9D的周邊參考區域的形狀為一例,亦可採用其他的形狀。又,亦可採用圖9D所示的周邊參考區域的一部分。又,周邊參考區域不限於鄰接於編碼對象區塊的區域,不鄰接於編碼對象區塊的區域亦可。又,於圖9D所示之例,參考圖片內的周邊參考區域是從編碼對象圖片內的周邊參考區域,由編碼對象圖片的移動向量指定的區域,但由其他移動向量指定區域亦可。例如該其他移動向量亦可為編碼對象圖片內的周邊參考區域的移動向量。Furthermore, the shape of the peripheral reference region of FIG. 9D is an example, and other shapes may be employed. Also, a part of the peripheral reference area shown in Fig. 9D may be employed. Further, the peripheral reference region is not limited to the region adjacent to the encoding target block, and may not be adjacent to the region of the encoding target block. Further, in the example shown in FIG. 9D, the peripheral reference area in the reference picture is an area specified by the motion vector of the encoding target picture from the peripheral reference area in the encoding target picture, but may be specified by other motion vectors. For example, the other motion vector may also be a motion vector of a peripheral reference region within the encoding target picture.
再者,於此說明了編碼裝置100的動作,但編碼裝置200的動作亦同。
[第1態樣的效果]Although the operation of the encoding device 100 has been described here, the operation of the encoding device 200 is also the same.
[Effect of the first aspect]
若依據第1態樣的構成,可能可減低主觀上醒目的區塊雜訊。於此,由於若進行LIC處理,預測圖像的亮度值會變化,因此於採用重構圖像與原圖的殘差訊號未完全發送的lossy編碼時,主觀上區塊雜訊可能容易變得醒目。然而,假定進行與HEVC同樣的Bs算出處理時,即使進行LIC處理,有時仍成為Bs=0,存在未進行去區塊濾波處理的情況。依據本態樣,可抑制該狀況。According to the composition of the first aspect, it is possible to reduce the subjectively blocky block noise. Here, since the luminance value of the predicted image changes when the LIC processing is performed, the subjective block noise may easily become used when the reconstructed image and the residual signal of the original image are not completely transmitted. Eye-catching. However, when the Bs calculation processing similar to HEVC is performed, even if the LIC processing is performed, Bs=0 may be obtained, and the deblocking filtering processing may not be performed. According to this aspect, this condition can be suppressed.
再者,解碼裝置200所含的迴路濾波部212實施與編碼裝置100所含的迴路濾波部120的上述去區塊濾波處理同樣的處理。Further, the loop filter unit 212 included in the decoding device 200 performs the same processing as the above-described deblocking filter processing of the loop filter unit 120 included in the encoding device 100.
又,並非總是需要第1態樣所記載的所有處理,亦可實施第1態樣所記載的一部分處理。
[去區塊濾波處理的第2態樣]Further, not all the processes described in the first aspect are always required, and a part of the processes described in the first aspect may be performed.
[Second aspect of deblocking filtering processing]
表示第2態樣的編碼裝置100所含的迴路濾波部120的去區塊濾波處理的流程圖,與圖11所示的第1態樣的去區塊濾波處理的流程圖同樣。The flowchart of the deblocking filtering process of the loop filter unit 120 included in the encoding apparatus 100 of the second aspect is the same as the flowchart of the deblocking filtering process of the first aspect shown in FIG.
於第2態樣,Bs判定處理(S201)的內容與第1態樣不同。於第1態樣,條件(5)是於跨越邊界相鄰接的2個區塊中至少一區塊,進行LIC處理,但於第2態樣考慮到區塊的位置關係。具體而言,迴路濾波部120是於隔著對象邊界的2個區塊(目前區塊及鄰接區塊)中之鄰接區塊,進行LIC處理時,設定為Bs≠0。例如迴路濾波部120是於如圖17所示,於位於左右的區塊中之左區塊(鄰接區塊),進行LIC處理時,設定為Bs≠0。又,迴路濾波部120是於如圖18所示,於位於上下的區塊中之上區塊(鄰接區塊),進行LIC處理時,設定為Bs≠0。圖19是表示算出此情況的Bs的方法例的圖。再者,(5A)於位於上下(左右)的區塊中之上(左)區塊,進行LIC處理時,Bs值若非0即可,1或2均可。In the second aspect, the content of the Bs determination process (S201) is different from that of the first aspect. In the first aspect, the condition (5) is that at least one of the two blocks adjacent to the boundary is subjected to LIC processing, but the positional relationship of the block is considered in the second aspect. Specifically, the loop filter unit 120 sets Bs ≠ 0 when performing LIC processing on adjacent blocks in two blocks (current block and adjacent block) across the boundary of the object. For example, the loop filter unit 120 is set to Bs ≠ 0 when the LIC process is performed on the left block (adjacent block) in the left and right blocks as shown in FIG. 17 . Further, as shown in FIG. 18, the loop filter unit 120 sets Bs≠0 when performing LIC processing on the upper block (adjacent block) in the upper and lower blocks. FIG. 19 is a view showing an example of a method of calculating Bs in this case. Furthermore, (5A) in the upper (left) block located in the upper and lower (left and right) blocks, when the LIC processing is performed, the Bs value may be 0 or 1 or 2.
圖20是於符合條件(5A)時,設定為Bs=1的情況下的Bs算出處理(S201)的流程圖。圖21表示算出此時的Bs的方法例的圖。FIG. 20 is a flowchart of Bs calculation processing (S201) when Bs=1 is set when the condition (5A) is satisfied. FIG. 21 is a view showing an example of a method of calculating Bs at this time.
首先,迴路濾波部120判定是否符合條件(1)(S221)。符合條件(1)時(S221,YES),迴路濾波部120設定為Bs=2(S223)。First, the loop filter unit 120 determines whether or not the condition (1) is satisfied (S221). When the condition (1) is satisfied (S221, YES), the loop filter unit 120 sets Bs=2 (S223).
不符合條件(1)時(S221,NO),迴路濾波部120判定是否符合條件(2)~(5A)中至少一條件(S222)。符合條件(2)~(5A)中至少一條件時(S222,YES),迴路濾波部120設定為Bs=1(S224)。不符合條件(2)~(5A)中任一條件時(S222,NO),迴路濾波部120設定為Bs=0(S225)。When the condition (1) is not satisfied (S221, NO), the loop filter unit 120 determines whether or not at least one of the conditions (2) to (5A) is satisfied (S222). When at least one of the conditions (2) to (5A) is satisfied (S222, YES), the loop filter unit 120 sets Bs=1 (S224). When any of the conditions (2) to (5A) is not satisfied (S222, NO), the loop filter unit 120 sets Bs = 0 (S225).
圖22是於條件(5A)時,設定為Bs=2的情況下的Bs算出處理(S201)的流程圖。圖23表示算出此時的Bs的方法例的圖。FIG. 22 is a flowchart of Bs calculation processing (S201) when Bs=2 is set in the condition (5A). FIG. 23 is a view showing an example of a method of calculating Bs at this time.
首先,迴路濾波部120判定是否符合條件(1)及(5A)的至少一方(S221A)。符合條件(1)及(5A)的至少一方時(S221A,YES),迴路濾波部120設定為Bs=2(S223)。First, the loop filter unit 120 determines whether or not at least one of the conditions (1) and (5A) is satisfied (S221A). When at least one of the conditions (1) and (5A) is satisfied (S221A, YES), the loop filter unit 120 is set to Bs=2 (S223).
不符合條件(1)及(5A)的任一條件時(S221A,NO),迴路濾波部120判定是否符合條件(2)~(4)中至少一條件(S222A)。符合條件(2)~(4)中至少一條件時(S222A,YES),迴路濾波部120設定為Bs=1(S224)。不符合條件(2)~(4)中任一條件時(S222A,NO),迴路濾波部120設定為Bs=0(S225)。
[第2態樣的效果]When any of the conditions (1) and (5A) is not satisfied (S221A, NO), the loop filter unit 120 determines whether or not at least one of the conditions (2) to (4) is satisfied (S222A). When at least one of the conditions (2) to (4) is satisfied (S222A, YES), the loop filter unit 120 sets Bs=1 (S224). When any of the conditions (2) to (4) is not satisfied (S222A, NO), the loop filter unit 120 sets Bs = 0 (S225).
[Effect of the second aspect]
若依據第2態樣的構成,相較於第1態樣,可能可減低過度的平滑化。例如將隔著對象邊界而鄰接的2個區塊以外的區域,作為LIC亮度修正用周邊參考區域利用而進行LIC處理時,區塊邊界主觀上容易變得醒目。此情況是於鄰接區塊進行LIC處理的情況。其他的情況下,例如若於將隔著對象邊界而鄰接的2個區塊內的區域,作為LIC亮度修正用周邊參考區域利用而進行LIC處理時,亦使用去區塊濾波器,則可能進行過度的平滑化。此情況是指於對象區塊進行LIC處理的情況。故,藉由第2態樣的構成,可能可解決該問題。According to the configuration of the second aspect, excessive smoothing may be reduced as compared with the first aspect. For example, when the LIC processing is performed using the area other than the two blocks adjacent to the boundary of the object as the LIC brightness correction peripheral reference area, the block boundary is subjectively conspicuous. This case is the case where the LIC processing is performed in the adjacent block. In other cases, for example, when the LIC processing is performed as the LIC brightness correction peripheral reference area by using the area in the two blocks adjacent to the target boundary, the deblocking filter may be used. Excessive smoothing. This case refers to the case where the LIC processing is performed on the target block. Therefore, this problem may be solved by the configuration of the second aspect.
再者,解碼裝置200所含的迴路濾波部212實施與編碼裝置100所含的迴路濾波部120的上述去區塊濾波處理同樣的處理。Further, the loop filter unit 212 included in the decoding device 200 performs the same processing as the above-described deblocking filter processing of the loop filter unit 120 included in the encoding device 100.
又,並非總是需要第2態樣所記載的所有處理,亦可實施第2態樣所記載的一部分處理。
[變形例]Further, not all the processes described in the second aspect are always required, and a part of the processes described in the second aspect may be performed.
[Modification]
迴路濾波部120以切片為單位切換進行或不進行上述處理亦可。迴路濾波部120以方塊為單位切換進行或不進行上述處理亦可。迴路濾波部120以CTU為單位切換進行或不進行上述處理亦可。迴路濾波部120以CU為單位切換進行或不進行上述處理亦可。迴路濾波部120因應處理對象的幀的幀種類(P幀、B幀等),來切換進行或不進行上述處理亦可。The loop filter unit 120 may or may not perform the above-described processing in units of slices. The loop filter unit 120 may or may not perform the above processing in units of blocks. The loop filter unit 120 may or may not perform the above processing in units of CTUs. The loop filter unit 120 may or may not perform the above processing in units of CU. The loop filter unit 120 may or may not perform the above-described processing depending on the frame type (P frame, B frame, etc.) of the frame to be processed.
又,於上述實施形態,表示LIC亮度修正用周邊參考區域為區塊的左鄰接及上鄰接的編碼完畢周邊參考區域的亮度像素值之例,但不限於此。例如作為LIC亮度修正用周邊參考區域,編碼裝置100利用上鄰接的編碼完畢周邊參考區域的亮度像素值,不利用左鄰接的編碼完畢周邊參考區域的亮度像素值亦可。又,編碼裝置100亦可因應條件來變化LIC亮度修正用周邊參考區域。例如編碼裝置100亦可將移動向量的差分小的區塊,利用作為LIC亮度修正用周邊參考區域。Further, in the above-described embodiment, the LIC brightness correction peripheral reference region is an example in which the luminance pixel values of the left adjacent and adjacent coded peripheral reference regions of the block are adjacent to each other, but the present invention is not limited thereto. For example, as the LIC brightness correction peripheral reference region, the encoding device 100 may use the luminance pixel value of the adjacent coded peripheral reference region, and may not use the luminance pixel value of the left adjacent coded peripheral reference region. Further, the encoding device 100 can also change the LIC brightness correction peripheral reference region depending on the conditions. For example, the encoding device 100 may use a block having a small difference in motion vectors as a peripheral reference region for LIC luminance correction.
又,於上述實施形態,表示Bs值設定0~2之值的例,但Bs值可取定的值不限於該例,亦可採用其他值。例如Bs值可利用3以上之值。此時,迴路濾波部120亦可於符合上述條件(5)或(5A)時,Bs值亦可設定3以上之值。Further, in the above embodiment, an example in which the Bs value is set to a value of 0 to 2 is shown. However, the value at which the Bs value can be determined is not limited to this example, and other values may be employed. For example, the Bs value can take a value of 3 or more. At this time, when the loop filter unit 120 satisfies the above condition (5) or (5A), the Bs value may be set to a value of three or more.
又,迴路濾波部120參考表示於各區塊是否使用LIC處理的LIC旗標,或不利用LIC旗標,來進行於2個區塊的各個區塊是否使用LIC處理的判定均可。例如制訂於周邊區塊使用LIC處理時,於對象區塊亦使用LIC處理,於周邊區塊不使用LIC處理時,於對象區塊亦不使用LIC處理等規則時,迴路濾波部120亦可利用周邊區塊的LIC使用狀況,來判定於對象區塊是否使用LIC處理。Further, the loop filter unit 120 refers to the LIC flag indicating whether or not each block uses the LIC process, or does not use the LIC flag to determine whether or not the LIC process is used for each block of the two blocks. For example, when the LIC processing is used in the peripheral block, the LIC processing is also used in the target block, and when the LIC processing is not used in the peripheral block, the loop filter unit 120 can also be used when the target block does not use the LIC processing or the like. The LIC usage status of the neighboring block determines whether the target block uses LIC processing.
又,於上述說明,說明因應是否適用LIC處理,來決定是否進行去區塊濾波處理之例,但編碼裝置100亦可因應是否適用LIC處理,來變更去區塊濾波處理的濾波強度。例如編碼裝置100亦可於適用LIC處理時,使去區塊濾波處理的濾波強度比未適用LIC處理時強。Further, in the above description, an example of determining whether or not to perform deblocking filtering processing is determined depending on whether or not the LIC processing is applied. However, the encoding apparatus 100 may change the filtering strength of the deblocking filtering processing depending on whether or not the LIC processing is applied. For example, the encoding device 100 can also make the filtering strength of the deblocking filtering process stronger than when the LIC processing is not applied when the LIC processing is applied.
再者,解碼裝置200所含的迴路濾波部212亦同。Furthermore, the loop filter unit 212 included in the decoding device 200 is also the same.
如以上,本實施形態的編碼裝置100及解碼裝置200進行圖24所示的處理。As described above, the encoding device 100 and the decoding device 200 of the present embodiment perform the processing shown in FIG.
編碼裝置100判定對於對象區塊、及鄰接於對象區塊的鄰接區塊中至少一區塊,即判定用區塊,是否適用預測圖像的亮度修正處理(例如LIC處理)(S231)。The encoding device 100 determines whether or not the brightness correction processing (for example, LIC processing) of the predicted image is applied to at least one of the target block and the adjacent block adjacent to the target block, that is, the determination block (S231).
編碼裝置100對於判定用區塊適用亮度修正處理時(S231,YES),於對象區塊與鄰接區塊的邊界,適用去區塊濾波處理(S232)。另,編碼裝置100對於判定用區塊不適用亮度修正處理時(S231,NO),不於前述邊界適用去區塊濾波處理(S233)。When the encoding device 100 applies the brightness correction processing to the determination block (S231, YES), the deblocking filtering process is applied to the boundary between the target block and the adjacent block (S232). When the encoding apparatus 100 does not apply the brightness correction processing to the determination block (S231, NO), the deblocking processing is not applied to the boundary (S233).
於此,於上述第1態樣,判定用區塊是指對象區塊及鄰接區塊。總言之,編碼裝置100對於對象區塊及鄰接區塊的一方或雙方,適用亮度修正處理時,於前述邊界適用去區塊濾波處理。編碼裝置100對於對象區塊及鄰接區塊的任一方,均不適用亮度修正處理時,不於前述邊界適用去區塊濾波處理。Here, in the first aspect described above, the determination block refers to the target block and the adjacent block. In short, when the encoding device 100 applies the brightness correction processing to one or both of the target block and the adjacent block, the deblocking filtering process is applied to the boundary. When the encoding apparatus 100 does not apply the luminance correction processing to any of the target block and the adjacent block, the deblocking filtering process is not applied to the boundary.
又,於上述第2態樣,判定用區塊是指鄰接區塊。總言之,編碼裝置100不論對於對象區塊有無適用亮度修正處理,對於鄰接區塊適用亮度修正處理時,均於前述邊界適用去區塊濾波處理。編碼裝置100不論對於對象區塊有無適用亮度修正處理,對於鄰接區塊不適用前述亮度修正處理時,均不於前述邊界適用去區塊濾波處理。Further, in the second aspect described above, the determination block refers to the adjacent block. In summary, the encoding apparatus 100 applies the deblocking processing to the boundary regardless of whether or not the luminance correction processing is applied to the target block, and when the luminance correction processing is applied to the adjacent block. The encoding apparatus 100 applies the luminance correction processing to the target block, and does not apply the deblocking processing to the boundary when the luminance correction processing is not applied to the adjacent block.
例如編碼裝置100對於判定用區塊適用亮度修正處理時,藉由將表示前述邊界的邊界強度的Bs設定為0以外之值,以於前述邊界適用去區塊濾波處理。編碼裝置100對於判定用區塊不適用亮度修正處理時,藉由將Bs設定為0,以不於前述邊界適用去區塊濾波處理。For example, when the encoding device 100 applies the brightness correction processing to the determination block, the de-block filtering process is applied to the boundary by setting Bs indicating the boundary strength of the boundary to a value other than 0. When the encoding apparatus 100 does not apply the luminance correction processing to the determination block, by setting Bs to 0, the deblocking filtering processing is applied without the boundary.
又,本實施形態的解碼裝置200判定對於對象區塊、及鄰接於對象區塊的鄰接區塊中至少一區塊,即判定用區塊,是否適用預測圖像的亮度修正處理(例如LIC處理)(S231)。Further, the decoding apparatus 200 according to the present embodiment determines whether or not the brightness correction processing (for example, LIC processing) of the predicted image is applied to at least one of the target block and the adjacent block adjacent to the target block, that is, the determination block. ) (S231).
解碼裝置200對於判定用區塊適用亮度修正處理時(S231,YES),於對象區塊與鄰接區塊的邊界,適用前述去區塊濾波處理(S232)。另,解碼裝置200對於判定用區塊不適用亮度修正處理時(S231,NO),不於前述邊界適用去區塊濾波處理(S233)。When the decoding device 200 applies the brightness correction processing to the determination block (S231, YES), the deblocking filtering process is applied to the boundary between the target block and the adjacent block (S232). When the decoding device 200 does not apply the brightness correction processing to the determination block (S231, NO), the deblocking processing is not applied to the boundary (S233).
於此,於上述第1態樣,判定用區塊是指對象區塊及鄰接區塊。總言之,解碼裝置200對於對象區塊及鄰接區塊的一方或雙方,適用亮度修正處理時,於前述邊界適用去區塊濾波處理。解碼裝置200對於對象區塊及鄰接區塊的任一方,均不適用亮度修正處理時,不於前述邊界適用去區塊濾波處理。Here, in the first aspect described above, the determination block refers to the target block and the adjacent block. In short, when the decoding device 200 applies the luminance correction processing to one or both of the target block and the adjacent block, the deblocking filtering process is applied to the boundary. When the decoding device 200 does not apply the luminance correction processing to any of the target block and the adjacent block, the deblocking filtering process is not applied to the boundary.
又,於上述第2態樣,判定用區塊是指鄰接區塊。總言之,解碼裝置200不論對於對象區塊有無適用亮度修正處理,對於鄰接區塊適用亮度修正處理時,均於前述邊界適用去區塊濾波處理。解碼裝置200不論對於對象區塊有無適用亮度修正處理,對於鄰接區塊不適用前述亮度修正處理時,均不於前述邊界適用去區塊濾波處理。Further, in the second aspect described above, the determination block refers to the adjacent block. In summary, the decoding apparatus 200 applies the deblocking processing to the boundary regardless of whether or not the luminance correction processing is applied to the target block and the luminance correction processing is applied to the adjacent block. The decoding apparatus 200 applies the luminance correction processing to the target block, and does not apply the deblocking processing to the boundary when the luminance correction processing is not applied to the adjacent block.
例如解碼裝置200對於判定用區塊適用亮度修正處理時,藉由將表示前述邊界的邊界強度的Bs設定為0以外之值,以於前述邊界適用去區塊濾波處理。解碼裝置200對於判定用區塊不適用亮度修正處理時,藉由將Bs設定為0,以不於前述邊界適用去區塊濾波處理。For example, when the decoding device 200 applies the luminance correction processing to the determination block, the de-block filtering process is applied to the boundary by setting Bs indicating the boundary strength of the boundary to a value other than 0. When the decoding device 200 does not apply the luminance correction processing to the determination block, by setting Bs to 0, the deblocking filtering process is applied without the boundary.
又,本實施形態的編碼裝置100具備:分割部102,將圖像分割為複數個區塊;幀內預測部124,利用包含於前述圖像的參考圖片,來預測包含於前述圖像的區塊;幀間預測部126,利用包含於與前述圖像不同的其他圖像的參考區塊,來預測包含於前述圖像的區塊;迴路濾波部120,對包含於前述圖像的區塊適用濾波器;轉換部106,轉換由前述幀內預測部124或前述幀間預測部126所生成的預測訊號與原訊號的預測誤差,生成轉換係數;量化部108,量化前述轉換係數,生成量化係數;及熵編碼部110,藉由將前述量化係數進行可變長度編碼,生成編碼位元串流。迴路濾波部120判定對於對象區塊、及鄰接於對象區塊的鄰接區塊中至少一區塊,即判定用區塊,是否適用預測圖像的亮度修正處理(例如LIC處理)(S231)。迴路濾波部120對於判定用區塊適用亮度修正處理時(S231,YES),於對象區塊與鄰接區塊的邊界,適用去區塊濾波處理(S232)。另,迴路濾波部120對於判定用區塊不適用亮度修正處理時(S231,NO),不於前述邊界適用去區塊濾波處理(S233)。Further, the encoding apparatus 100 of the present embodiment includes a dividing unit 102 that divides an image into a plurality of blocks, and the intra prediction unit 124 predicts a region included in the image by using a reference picture included in the image. a block; inter prediction unit 126 predicts a block included in the image by using a reference block included in another image different from the image; and a loop filter unit 120 for the block included in the image The filter unit 106 converts the prediction error of the prediction signal generated by the intra prediction unit 124 or the inter prediction unit 126 and the original signal to generate a conversion coefficient, and the quantization unit 108 quantizes the conversion coefficient to generate a quantization. The coefficient and the entropy coding unit 110 generate a coded bit stream by performing variable length coding on the quantized coefficients. The loop filter unit 120 determines whether or not the brightness correction processing (for example, LIC processing) of the predicted image is applied to at least one of the target block and the adjacent block adjacent to the target block, that is, the determination block (S231). When the loop filter unit 120 applies the brightness correction processing to the determination block (S231, YES), the deblocking filter processing is applied to the boundary between the target block and the adjacent block (S232). When the loop filter unit 120 does not apply the brightness correction processing to the determination block (S231, NO), the deblocking processing is not applied to the boundary (S233).
又,本實施形態的解碼裝置200具備:解碼部(熵解碼部202),解碼編碼位元串流,輸出量化係數;反量化部204,將前述量化係數予以反量化,輸出轉換係數;反轉換部206,將前述轉換係數予以反轉換,輸出預測誤差;幀內預測部216,利用包含於圖像的參考圖片,來預測包含於前述圖像的區塊;幀間預測部218,利用包含於與前述圖像不同的其他圖像的參考區塊,來預測包含於前述圖像的區塊;及迴路濾波部212,對包含於前述圖像的區塊適用濾波器。迴路濾波部212判定對於對象區塊、及鄰接於對象區塊的鄰接區塊中至少一區塊,即判定用區塊,是否適用預測圖像的亮度修正處理(例如LIC處理)(S231)。迴路濾波部212對於判定用區塊適用亮度修正處理時(S231,YES),於對象區塊與鄰接區塊的邊界,適用去區塊濾波處理(S232)。另,迴路濾波部212對於判定用區塊不適用亮度修正處理時(S231,NO),不於前述邊界適用去區塊濾波處理(S233)。
[編碼裝置的實作例]Further, the decoding apparatus 200 of the present embodiment includes a decoding unit (entropy decoding unit 202) that decodes a coded bit stream and outputs a quantized coefficient, and an inverse quantization unit 204 that inversely quantizes the quantized coefficient to output a conversion coefficient; The unit 206 inversely converts the conversion coefficient to output a prediction error, and the intra prediction unit 216 predicts a block included in the image by using a reference picture included in the image; the inter prediction unit 218 is included in A reference block of another image different from the image is used to predict a block included in the image; and a loop filter unit 212 applies a filter to the block included in the image. The loop filter unit 212 determines whether or not the brightness correction processing (for example, LIC processing) of the predicted image is applied to at least one of the target block and the adjacent block adjacent to the target block, that is, the determination block (S231). When the loop filter unit 212 applies the brightness correction processing to the determination block (S231, YES), the deblocking filter processing is applied to the boundary between the target block and the adjacent block (S232). When the loop filter unit 212 does not apply the brightness correction processing to the determination block (S231, NO), the deblocking processing is not applied to the boundary (S233).
[Example of coding device]
圖25是表示實施形態1的編碼裝置100的實作例的方塊圖。編碼裝置100具備電路160及記憶體162。例如圖1所示的編碼裝置100的複數個構成要件,是藉由圖25所示的電路160及記憶體162來實作。Fig. 25 is a block diagram showing a practical example of the coding apparatus 100 according to the first embodiment. The encoding device 100 includes a circuit 160 and a memory 162. For example, a plurality of components of the encoding device 100 shown in FIG. 1 are implemented by the circuit 160 and the memory 162 shown in FIG.
電路160是進行資訊處理的電路,可於記憶體162存取。例如電路160是編碼動態圖像的專用或通用電子電路。電路160可為諸如CPU的處理器。又,電路160亦可為複數個電子電路的集合體。又,電路160亦可發揮圖1等所示的編碼裝置100的複數個構成要件中,用以記憶資訊的構成要件除外的複數個構成要件的作用。The circuit 160 is a circuit for performing information processing and is accessible to the memory 162. For example, circuit 160 is a dedicated or general purpose electronic circuit that encodes a dynamic image. Circuitry 160 can be a processor such as a CPU. Moreover, circuit 160 can also be a collection of a plurality of electronic circuits. Further, the circuit 160 can also function as a plurality of constituent elements excluding the constituent elements for storing information among the plurality of constituent elements of the encoding apparatus 100 shown in FIG. 1 and the like.
記憶體162是記憶電路160用以編碼動態圖像的資訊之通用或專用記憶體。記憶體162為電子電路或連接於電路160均可。又,記憶體162包含於電路160亦可。又,記憶體162亦可為複數個電子電路的集合體。又,記憶體162亦可為磁碟片或光碟片等,或亦可表現為儲存器或記錄媒體等。又,記憶體162為非揮發性記憶體或揮發性記憶體均可。The memory 162 is a general purpose or special memory used by the memory circuit 160 to encode information of a moving image. The memory 162 may be an electronic circuit or may be connected to the circuit 160. Further, the memory 162 may be included in the circuit 160. Further, the memory 162 may be an aggregate of a plurality of electronic circuits. Moreover, the memory 162 may be a magnetic disk or a compact disk, or may be embodied as a storage or a recording medium. Further, the memory 162 may be either a non-volatile memory or a volatile memory.
又,例如於記憶體162,記憶編碼的動態圖像,或對應於已編碼的動態圖像的位元串均可。又,於記憶體162,亦可記憶電路160用以編碼動態圖像的程式。Also, for example, in the memory 162, a coded moving image or a bit string corresponding to the encoded moving image may be used. Moreover, in the memory 162, the program used by the memory circuit 160 to encode a moving image can also be used.
又,例如記憶體162亦可發揮圖1等所示的編碼裝置100的複數個構成要件中,用以記憶資訊的構成要件的作用。具體而言,記憶體162亦可發揮圖1所示的區塊記憶體118及幀記憶體122的作用。更具體而言,於記憶體162,亦可記憶重構完畢區塊及重構完畢圖片等。Further, for example, the memory 162 can also function as a constituent element for storing information in a plurality of constituent elements of the encoding device 100 shown in FIG. 1 and the like. Specifically, the memory 162 can also function as the block memory 118 and the frame memory 122 shown in FIG. More specifically, in the memory 162, the reconstructed block, the reconstructed picture, and the like can also be memorized.
再者,於編碼裝置100,未實作圖1等所示複數個構成要件的全部,或未進行上述複數種處理的全部均可。圖1等所示的複數個構成要件的一部分,包含於其他裝置,或由其他裝置來執行上述複數種處理的一部分均可。然後,於編碼裝置100,可實作圖1等所示的複數個構成要件中的一部分,藉由進行上述複數種處理的一部分,有效率地進行移動補償。
[解碼裝置的實作例]Further, in the encoding apparatus 100, all of the plurality of constituent elements shown in FIG. 1 and the like may not be realized, or all of the above-described plural kinds of processing may not be performed. A part of the plurality of constituent elements shown in FIG. 1 and the like may be included in other devices, or a part of the plurality of processes may be executed by other devices. Then, in the encoding apparatus 100, part of the plurality of constituent elements shown in FIG. 1 and the like can be realized, and by performing a part of the plurality of types of processing described above, the movement compensation can be efficiently performed.
[Example of decoding device]
圖26是表示實施形態1的解碼裝置200的實作例的方塊圖。解碼裝置200具備電路260及記憶體262。例如圖10所示的解碼裝置200的複數個構成要件,是藉由圖26所示的電路260及記憶體262來實作。Fig. 26 is a block diagram showing a practical example of the decoding device 200 of the first embodiment. The decoding device 200 includes a circuit 260 and a memory 262. For example, a plurality of components of the decoding device 200 shown in FIG. 10 are implemented by the circuit 260 and the memory 262 shown in FIG.
電路260是進行資訊處理的電路,可於記憶體262存取。例如電路260是解碼動態圖像的專用或通用電子電路。電路260可為諸如CPU的處理器。又,電路260亦可為複數個電子電路的集合體。又,例如電路260亦可發揮圖10等所示的解碼裝置200的複數個構成要件中,用以記憶資訊的構成要件除外的複數個構成要件的作用。The circuit 260 is a circuit for performing information processing and is accessible to the memory 262. For example, circuit 260 is a dedicated or general purpose electronic circuit that decodes dynamic images. Circuitry 260 can be a processor such as a CPU. Also, circuit 260 can be an aggregate of a plurality of electronic circuits. Further, for example, the circuit 260 may function as a plurality of constituent elements excluding the constituent elements for storing information among the plurality of constituent elements of the decoding apparatus 200 shown in FIG. 10 and the like.
記憶體262是記憶電路260用以解碼動態圖像的資訊之通用或專用記憶體。記憶體262為電子電路或連接於電路260均可。又,記憶體262亦可包含於電路260。又,記憶體262亦可為複數個電子電路的集合體。又,記憶體262亦可為磁碟片或光碟片等,或亦可表現為儲存器或記錄媒體等。又,記憶體262為非揮發性記憶體或揮發性記憶體均可。The memory 262 is a general-purpose or dedicated memory used by the memory circuit 260 to decode information of a moving image. The memory 262 may be an electronic circuit or may be connected to the circuit 260. Also, the memory 262 may be included in the circuit 260. Further, the memory 262 may be an aggregate of a plurality of electronic circuits. Moreover, the memory 262 may be a magnetic disk or a compact disk, or may be embodied as a memory or a recording medium. Further, the memory 262 may be either a non-volatile memory or a volatile memory.
例如於記憶體262,記憶對應於已編碼的動態圖像的位元串,或記憶對應於已解碼的位元串的動態圖像均可。又,於記憶體262,亦可記憶電路260用以解碼動態圖像的程式。For example, in the memory 262, a bit string corresponding to the encoded moving picture is memorized, or a moving picture corresponding to the decoded bit string can be memorized. Further, in the memory 262, a program for decoding a moving image by the memory circuit 260 is also possible.
又,例如記憶體262亦可發揮圖10等所示的解碼裝置200的複數個構成要件中,用以記憶資訊的構成要件的作用。具體而言,記憶體262亦可發揮圖10所示的區塊記憶體210及幀記憶體214的作用。更具體而言,於記憶體262,亦可記憶重構完畢區塊及重構完畢圖片等。Further, for example, the memory 262 can also function as a constituent element for storing information in a plurality of constituent elements of the decoding device 200 shown in FIG. Specifically, the memory 262 can also function as the block memory 210 and the frame memory 214 shown in FIG. More specifically, in the memory 262, the reconstructed block and the reconstructed picture can also be memorized.
再者,於解碼裝置200,未實作圖10等所示的複數個構成要件的全部,或未進行上述複數種處理的全部均可。圖10等所示的複數個構成要件的一部分,包含於其他裝置,或由其他裝置來執行上述複數種處理的一部分均可。然後,於解碼裝置200,可實作圖10等所示的複數個構成要件中的一部分,藉由進行上述複數種處理的一部分,有效率地進行移動補償。
[補充]Further, in the decoding device 200, all of the plurality of constituent elements shown in FIG. 10 and the like may not be implemented, or all of the above-described plural kinds of processing may not be performed. A part of the plurality of constituent elements shown in FIG. 10 and the like may be included in other devices, or a part of the above-described plurality of processes may be executed by other devices. Then, in the decoding device 200, a part of the plurality of constituent elements shown in FIG. 10 and the like can be realized, and the movement compensation can be efficiently performed by performing a part of the plurality of types of processing described above.
[supplement]
本實施形態的編碼裝置100及解碼裝置200分別利用作為圖像編碼裝置及圖像解碼裝置,或分別利用作為動態圖像編碼裝置及動態圖像解碼裝置均可。或者,編碼裝置100及解碼裝置200分別可利用作為迴路濾波裝置。The encoding device 100 and the decoding device 200 according to the present embodiment are each used as an image encoding device and an image decoding device, or may be used as a moving image encoding device and a moving image decoding device, respectively. Alternatively, the encoding device 100 and the decoding device 200 can be utilized as loop filtering devices, respectively.
亦即,編碼裝置100及解碼裝置200亦可分別僅對應於迴路濾波部120及迴路濾波部212。然後,轉換部106及反轉換部206等其他構成要件亦可包含於其他裝置。In other words, the encoding device 100 and the decoding device 200 may correspond to only the loop filter unit 120 and the loop filter unit 212, respectively. Then, other components such as the conversion unit 106 and the inverse conversion unit 206 may be included in other devices.
又,於本實施形態,各構成要件亦可由專用的硬體來構成,或藉由執行適於各構成要件的軟體程式來實現。各構成要件亦可由CPU或處理器等之程式執行部,讀出並執行記錄在硬碟或半導體記憶體等記錄媒體的軟體程式來實現。Further, in the present embodiment, each constituent element may be constituted by a dedicated hardware or by executing a software program suitable for each constituent element. Each component can be realized by a program execution unit such as a CPU or a processor, which reads and executes a software program recorded on a recording medium such as a hard disk or a semiconductor memory.
具體而言,編碼裝置100及解碼裝置200亦可分別具備:處理電路(Processing Circuitry);及記憶裝置(Storage),其電連接於該處理電路,可從該處理電路進行存取。例如處理電路對應於電路160或260,記憶裝置對應於記憶體162或262。Specifically, the encoding device 100 and the decoding device 200 may each include a processing circuit (Processing Circuitry) and a memory device (Storage) electrically connected to the processing circuit and accessible from the processing circuit. For example, the processing circuit corresponds to circuit 160 or 260, and the memory device corresponds to memory 162 or 262.
處理電路包含專用的硬體及程式執行部的至少一方,利用記憶裝置來執行處理。又,在處理電路包含程式執行部時,記憶裝置記憶由該程式執行部所執行的軟體程式。The processing circuit includes at least one of a dedicated hardware and a program execution unit, and performs processing using a memory device. Further, when the processing circuit includes the program execution unit, the memory device memorizes the software program executed by the program execution unit.
於此,實現本實施形態的編碼裝置100及解碼裝置200等之軟體是如下程式。Here, the software that implements the encoding device 100 and the decoding device 200 of the present embodiment is the following program.
又,如上述,各構成要件亦可為電路。該等電路作為全體構成1個電路,或各自為不同電路均可。又,各構成要件以通用處理器來實現,或以專用處理器來實現均可。Further, as described above, each constituent element may be an electric circuit. These circuits may constitute one circuit as a whole or may be different circuits. Further, each component may be implemented by a general purpose processor or by a dedicated processor.
又,其他構成要件來執行特定構成要件所執行的處理亦可。又,變更執行處理的順序,或複數種處理同時進行均可。又,編碼解碼裝置亦可具備編碼裝置100及解碼裝置200。Further, other components may be used to execute the processing performed by the specific constituent elements. Further, the order of the execution processing may be changed, or a plurality of types of processing may be performed simultaneously. Further, the codec device may include the encoding device 100 and the decoding device 200.
以上根據實施形態,說明了編碼裝置100及解碼裝置200的態樣,但編碼裝置100及解碼裝置200的態樣不限定於該實施形態。只要不脫離本發明之旨趣,熟悉該技藝人士所想到的各種變形施行於本實施形態、或組合不同實施形態的構成要件予以建構的形態,亦可包含於編碼裝置100及解碼裝置200的態樣的範圍內。Although the aspects of the encoding device 100 and the decoding device 200 have been described above based on the embodiment, the aspects of the encoding device 100 and the decoding device 200 are not limited to the embodiment. Various modifications, which are contemplated by those skilled in the art, may be made in the present embodiment or in combination with the constituent elements of the different embodiments, and may be included in the aspects of the encoding device 100 and the decoding device 200, without departing from the scope of the present invention. In the range.
亦可將本態樣,與本發明的其他態樣的至少一部分組合而實施。又,亦可將本態樣的流程圖所記載的一部分處理、裝置的一部分構成、語法的一部分等,與其他態樣組合而實施。
(實施形態2)This aspect can also be implemented in combination with at least a portion of other aspects of the invention. Further, a part of the processing described in the flowchart of the present aspect, a part of the configuration of the device, a part of the grammar, and the like may be combined with other aspects.
(Embodiment 2)
於以上各實施形態,功能方塊的各個一般可藉由MPU及記憶體等來實現。又,功能方塊的各個的處理一般藉由處理器等之程式執行部,讀出並執行記錄於ROM等記錄媒體的軟體(程式)來實現。該軟體藉由下載等來分發,或記錄於半導體記憶體等記錄媒體來分發均可。再者,當然亦可藉由硬體(專用電路)來實現各功能方塊。In each of the above embodiments, each of the functional blocks can be generally implemented by an MPU, a memory, or the like. Further, each processing of the function block is generally realized by a program execution unit such as a processor reading and executing a software (program) recorded on a recording medium such as a ROM. The software can be distributed by downloading or the like, or can be distributed on a recording medium such as a semiconductor memory. Furthermore, it is of course also possible to implement the functional blocks by hardware (dedicated circuit).
又,於各實施形態所說明的處理藉由利用單一裝置(系統)集中處理,或利用複數個裝置分散處理來實現均可。又,執行上述程式的處理器為單一數目或複數個均可。亦即,集中處理或分散處理均可。Further, the processing described in each embodiment may be realized by centralized processing by a single device (system) or by a plurality of device distributed processing. Further, the processor for executing the above program may be a single number or a plurality of processors. That is, centralized processing or distributed processing is possible.
本發明的態樣不限定於以上實施形態,可予以各種變更,該等變更亦包含於本發明的範圍內。The aspects of the present invention are not limited to the above embodiments, and various modifications can be made thereto, and such modifications are also included in the scope of the present invention.
進一步在此說明上述各實施形態所示之動態圖像編碼方法(圖像編碼方法)或動態圖像解碼方法(圖像解碼方法)之應用例、與利用其之系統。該系統的特徵在於具有利用圖像編碼方法的圖像編碼裝置、利用圖像解碼方法的圖像解碼裝置、及具備雙方的圖像編碼解碼裝置。關於系統的其他構成,可因應情況適當地變更。
[使用例]Further, an application example of the moving image encoding method (image encoding method) or the moving image decoding method (image decoding method) described in each of the above embodiments and a system using the same will be described. This system is characterized in that it includes an image coding device using an image coding method, an image decoding device using an image decoding method, and an image coding and decoding device including both. The other components of the system can be appropriately changed depending on the situation.
[usage]
圖27是表示實現內容發布服務之內容供給系統ex100的全體構成圖。將通訊服務之提供區分割為所需大小,於各蜂巢(cell)內分別設置固定無線台即基地局ex106、ex107、ex108、ex109、ex110。FIG. 27 is a view showing the overall configuration of a content supply system ex100 that realizes a content distribution service. The communication service providing area is divided into required sizes, and fixed base stations ex106, ex107, ex108, ex109, and ex110 are respectively set in each cell.
於該內容供給系統ex100,經由網際網路服務提供者ex102及通訊網ex104、及基地局ex106~ex110,將電腦ex111、遊戲機ex112、攝影機ex113、家電ex114及智慧型手機ex115等各機器,連接於網際網路ex101。該內容供給系統ex100亦可組合上述任一要件而連接。不經由固定無線台即基地局ex106~ex110,各機器經由電話網或近距離無線等而直接或間接地相互連接亦可。又,串流化伺服器ex103經由網際網路ex101等來與電腦ex111、遊戲機ex112、攝影機ex113、家電ex114及智慧型手機ex115等各機器連接。又,串流化伺服器ex103經由衛星ex116來與飛機ex117內的熱點內的終端裝置等連接。The content supply system ex100 connects the computers ex111, the game machine ex112, the camera ex113, the home appliance ex114, and the smart phone ex115 to each other via the Internet service provider ex102 and the communication network ex104, and the base stations ex106 to ex110. Internet ex101. The content supply system ex100 can also be connected in combination with any of the above requirements. The devices may be directly or indirectly connected to each other via a telephone network or short-range wireless without passing through the fixed wireless stations, that is, the base stations ex106 to ex110. Further, the streaming server ex103 is connected to each of the devices such as the computer ex111, the game machine ex112, the camera ex113, the home appliance ex114, and the smartphone ex115 via the Internet ex101 or the like. Further, the streaming server ex103 is connected to a terminal device or the like in the hot spot in the aircraft ex117 via the satellite ex116.
再者,亦可利用無線存取點或熱點等,來取代基地局ex106~ex110。又,串流化伺服器ex103不經由網際網路ex101或網際網路服務提供者ex102而直接與通訊網ex104連接,或不經由衛星ex116而直接與飛機ex117連接均可。Furthermore, it is also possible to replace the base stations ex106 to ex110 by using wireless access points or hotspots. Further, the streaming server ex103 may be directly connected to the communication network ex104 via the Internet ex101 or the Internet service provider ex102, or may be directly connected to the aircraft ex117 via the satellite ex116.
攝影機ex113是數位攝影機等可拍攝靜止圖及動畫的機器。又,智慧型手機ex115是一般支援2G、3G、3.9G、4G,以及今後稱為5G的移動通訊系統方式的智慧型機、行動電話或PHS(Personal Handyphone System(個人手持電話系統))等。The camera ex113 is a machine that can capture still images and animations, such as a digital camera. In addition, the smart phone ex115 is a smart phone, a mobile phone, or a PHS (Personal Handyphone System) that generally supports 2G, 3G, 3.9G, 4G, and a mobile communication system called 5G in the future.
家電ex118為冰箱或家庭用燃料電池汽電共生系統所含的機器等。The appliance ex118 is a machine included in a refrigerator or a household fuel cell cogeneration system.
於內容供給系統ex100,具有攝影功能的終端裝置藉由透過基地局ex106等,連接於串流化伺服器ex103,可實現實況發布等。於實況發布,終端裝置(電腦ex111、遊戲機ex112、攝影機ex113、家電ex114、智慧型手機ex115及飛機ex117內的終端裝置等)對於使用者利用該終端裝置所拍攝的靜止圖或動畫內容,進行上述各實施形態所說明的編碼處理,將藉由編碼所獲得的影像資料、及對應於影像的聲音經編碼的聲音資料,予以多工,將獲得的資料發送至串流化伺服器ex103。亦即,各終端裝置作為本發明一態樣的圖像編碼裝置而發揮功能。In the content supply system ex100, the terminal device having the photographing function is connected to the streaming server ex103 via the base station ex106 or the like, thereby realizing live delivery and the like. In the live release, the terminal device (the computer ex111, the game machine ex112, the camera ex113, the home appliance ex114, the smart phone ex115, and the terminal device in the aircraft ex117) performs the still image or the animation content captured by the user using the terminal device. In the encoding process described in each of the above embodiments, the video data obtained by the encoding and the audio data encoded by the audio corresponding to the video are multiplexed, and the obtained data is transmitted to the streaming server ex103. That is, each terminal device functions as an image coding device according to an aspect of the present invention.
另,串流化伺服器ex103將對於有需求的客戶所發送內容資料,進行串流發布。客戶是可將上述經編碼處理的資料解碼的電腦ex111、遊戲機ex112、攝影機ex113、家電ex114、智慧型手機ex115及飛機ex117內的終端裝置等。接收到發布資料的各機器將接收資料予以解碼處理並再生。亦即,各機器作為本發明一態樣的圖像解碼裝置而發揮功能。
[分散處理]In addition, the streaming server ex103 distributes the content data sent by the client in need thereof. The client is a computer ex111, a game machine ex112, a camera ex113, a home appliance ex114, a smart phone ex115, and a terminal device in the aircraft ex117, which can decode the encoded data. Each machine that receives the published material decodes the received data and reproduces it. That is, each machine functions as an image decoding device of one aspect of the present invention.
[Distributed processing]
又,串流化伺服器ex103亦可為複數台伺服器或複數台電腦,分散處理、記錄或發布資料。例如串流化伺服器ex103亦可藉由CDN(Contents Delivery Network(內容發布網路))來實現,藉由連結分散在全世界的許多邊緣伺服器與邊緣伺服器間的網路,來實現內容發布。於CDN,因應客戶而動態地分派物理上接近的邊緣伺服器。然後,藉由對該邊緣伺服器快取及發布內容,可減少延遲。又,由於在發生某種錯誤時,或通訊狀態因流量增加等而改變時,以複數台邊緣伺服器分散處理,或將發布主體切換為其他邊緣伺服器,可繞過發生障礙的的網路部分來繼續發布,因此可實現高速且穩定的發布。Moreover, the serialization server ex103 can also be a plurality of servers or a plurality of computers to process, record, or distribute data. For example, the streaming server ex103 can also be implemented by a CDN (Contents Delivery Network), which realizes content by connecting a network distributed between many edge servers and edge servers all over the world. release. At CDN, physically close edge servers are dynamically dispatched in response to the customer. Then, by caching and publishing content to the edge server, latency can be reduced. In addition, when a certain error occurs, or the communication status changes due to an increase in traffic, etc., the distributed edge server is distributed, or the distribution body is switched to another edge server, thereby bypassing the network that is in trouble. Partly to continue to release, so high-speed and stable release can be achieved.
又,不僅止於發布本身的分散處理,拍攝資料的編碼處理在各終端裝置或於伺服器側進行均可,或亦可互相分擔進行。作為一例,編碼處理一般進行2次處理迴路。於第1次迴路,檢出幀或場景單位的畫面複雜度或碼量。又,於第2次迴路,進行維持畫質或使編碼效率提升的處理。例如終端裝置進行第1次的編碼裝置,收到內容的伺服器側進行第2次編碼處理,藉此可減少各終端裝置的處理負載,同時可使內容的品質及效率提升。此時,若要求幾乎即時接收並解碼時,可將終端裝置進行的第1次編碼完成的資料,在其他終端裝置接收並再生,因此亦可實現更靈活的即時發布。Further, not only the distributed processing of the distribution itself but also the encoding processing of the captured data may be performed on each terminal device or on the server side, or may be performed separately. As an example, the encoding process generally performs a processing loop twice. In the first loop, the picture complexity or code size of the frame or scene unit is detected. Further, in the second loop, processing for maintaining image quality or improving encoding efficiency is performed. For example, the terminal device performs the first encoding device, and the server side that receives the content performs the second encoding process, thereby reducing the processing load of each terminal device and improving the quality and efficiency of the content. In this case, if it is required to receive and decode almost instantaneously, the data of the first encoding performed by the terminal device can be received and reproduced by the other terminal device, so that more flexible instant distribution can be realized.
作為其他例,攝影機ex113等從圖像進行特徵量擷取,將有關特徵量的資料作為元資料壓縮,發送至伺服器。伺服器因應圖像的意義來進行壓縮,例如從特徵量,判斷物件的重要性而切換量子精度等。特徵量資料尤其對於在伺服器再度壓縮時,提升移動向量預測的精度及效率有效。又,於終端裝置進行VLC(可變長度編碼)等之簡易編碼,於伺服器進行CABAC(上下文適應型二值算術編碼方式)等處理負載大的編碼亦可。As another example, the camera ex113 or the like extracts the feature amount from the image, compresses the data on the feature amount as metadata, and transmits it to the server. The server compresses according to the meaning of the image, for example, switching the quantum precision from the feature quantity, judging the importance of the object. The feature quantity data is especially effective for improving the accuracy and efficiency of motion vector prediction when the server is recompressed. Further, the terminal device performs simple coding such as VLC (variable length coding), and the server may perform coding with a large load such as CABAC (Context-Adapted Binary Arithmetic Coding).
進而言之,作為其他例,於體育館、購物中心或工廠等,有時存在藉由複數個終端裝置,拍攝大致同一場景的複數個影像資料。此時,利用進行攝影的複數個終端裝置、與因應需要未拍攝的其他終端裝置及伺服器,以例如GOP(Group of Picture(圖片群組))單位、圖片單位或圖片經分割的方塊單位等,分別分派編碼處理來分散處理。藉此可減少延遲,更實現即時性。Further, as another example, in a gymnasium, a shopping mall, a factory, or the like, a plurality of video files of substantially the same scene may be captured by a plurality of terminal devices. In this case, a plurality of terminal devices for photographing and other terminal devices and servers that are not photographed as needed are used, for example, a GOP (Group of Picture) unit, a picture unit, or a divided square unit of a picture. , respectively, assigning encoding processing to distribute processing. This can reduce latency and achieve immediacy.
又,由於複數個影像資料大致為同一場景,因此以伺服器管理及/或指示互相參考各終端裝置拍攝的影像資料亦可。又,伺服器接收來自各終端裝置的編碼完畢資料,於複數個資料間變更參考關係,或修正或更換圖片本身,予以重新編碼亦可。藉此,可生成提高每1個資料的品質及效率的串流。Moreover, since the plurality of video data are substantially in the same scene, it is also possible to manage and/or instruct the video data captured by each terminal device with the server. Moreover, the server receives the encoded data from each terminal device, changes the reference relationship between the plurality of data, or corrects or replaces the image itself, and re-encodes it. Thereby, it is possible to generate a stream that improves the quality and efficiency of each piece of data.
又,伺服器亦可進行變更影像資料編碼方式的轉碼後,再發布影像資料。例如伺服器將MPEG系統的編碼方式,轉換成VP系統,或將H.264轉換成H.265均可。In addition, the server can also perform image transcoding after changing the encoding method of the image data, and then release the image data. For example, the server converts the encoding method of the MPEG system into a VP system, or converts H.264 into H.265.
如此,編碼處理可藉由終端裝置或1以上的伺服器來進行。故,以下作為進行處理的主體,雖採用「伺服器」或「終端」等記載,但由伺服器進行的處理的一部分或全部,亦可由終端裝置來進行,或由終端裝置進行的處理的一部分或全部,亦可由伺服器來進行。又,關於該等,就解碼處理而言亦同理。
[3D、多角度]In this way, the encoding process can be performed by the terminal device or a server of one or more. Therefore, the following is a description of the "server" or "terminal" used as the main body for processing. However, part or all of the processing performed by the server may be performed by the terminal device or part of the processing performed by the terminal device. Or all, can also be carried out by the server. Also, regarding these, the same is true for the decoding process.
[3D, multi-angle]
近年來,越來越多將由互相大致同步的複數個攝影機ex113及/或智慧型手機ex115等終端裝置所拍攝的不同場景、或從不同角度拍攝同一場景的圖像或影像予以整合利用。各終端裝置所拍攝的影像根據另外取得的終端裝置間的相對位置關係、或影像所含特徵點一致的區域等來整合。In recent years, more and more images captured by a plurality of cameras ex113 and/or smart phones ex115, which are substantially synchronized with each other, or images or images of the same scene are captured from different angles. The images captured by the respective terminal devices are integrated based on the relative positional relationship between the acquired terminal devices, the regions in which the feature points included in the images match, and the like.
伺服器不僅編碼二維的動態圖像,亦可根據動態圖像的場景分析等,自動或於使用者所指定的時刻,編碼靜止圖,並發送至接收終端裝置。伺服器進一步在可取得攝影終端裝置間的相對位置關係時,不僅根據二維的動態圖像,亦可根據從不同角度拍攝同一場景的影像,來生成該場景的三維形狀。再者,伺服器另外編碼點雲等所生成的三維資料,或利用三維資料來辨識人物或物件,或者根據追蹤的結果,從複數個終端裝置拍攝的影像,選擇或重構要發送至接收終端裝置的影像均可。The server not only encodes the two-dimensional moving image, but also encodes the still image automatically or at the time specified by the user according to the scene analysis of the moving image, and transmits it to the receiving terminal device. Further, when the server can acquire the relative positional relationship between the photographing terminal devices, the three-dimensional shape of the scene can be generated based on not only the two-dimensional moving image but also the image of the same scene from different angles. Furthermore, the server additionally encodes the three-dimensional data generated by the point cloud or the like, or uses the three-dimensional data to identify the person or the object, or selects or reconstructs the image captured from the plurality of terminal devices according to the tracking result to be transmitted to the receiving terminal. The image of the device can be.
如此,使用者可任意選擇對應於各攝影終端裝置的各影像,欣賞場景,亦可欣賞從利用複數個圖像或影像重構的三維資料,切出任意視點的影像的內容。進而言之,與影像相同,聲音亦可從複數個不同角度來收音,伺服器配合影像,將來自特定角度或空間的聲音,與影像進行多工並發送。In this way, the user can arbitrarily select each image corresponding to each of the photographing terminal devices, and enjoy the scene, and can also view the content of the image of any viewpoint from the three-dimensional data reconstructed from the plurality of images or images. In other words, like the image, the sound can be collected from a plurality of different angles, and the server cooperates with the image to multiplex and transmit the sound from a specific angle or space.
又,近年來Virtual Reality(虛擬實境)(VR)及Augmented Reality(擴增實境)(AR)等使現實世界與虛擬世界相對應的內容亦日益普及。VR圖像時,伺服器分別製作右眼用及左眼用的視點圖像,藉由Multi-View Coding(多視角編碼)(MVC)等,進行各視點影像間容許參考的編碼,或互相不參考,作為不同的串流來編碼均可。於不同的串流解碼時,因應使用者的視點,使其互相同步再生,以重現虛擬的三維空間即可。In addition, in recent years, content such as Virtual Reality (VR) and Augmented Reality (AR) that make the real world correspond to the virtual world has become increasingly popular. In the VR image, the server creates the viewpoint images for the right eye and the left eye, and performs the encoding of the allowable reference between the respective viewpoint images by Multi-View Coding (MVC) or the like, or does not mutually For reference, it can be encoded as a different stream. In the case of different stream decoding, in response to the user's viewpoint, they can be synchronized with each other to reproduce the virtual three-dimensional space.
AR圖像時,伺服器對現實空間的攝影機資訊,根據三維位置或使用者的視點移動,來重疊虛擬空間上的虛擬物體資訊。解碼裝置亦可取得或保持虛擬物體資訊及三維資料,因應使用者的視點移動,來生成二維圖像,平滑地接合以製作重疊資料。又,解碼裝置除了發送虛擬物體資訊的請求以外,還將使用者的視點移動發送至伺服器,伺服器配合從保持於伺服器的三維資料所接收的視點移動,製作重疊資料,編碼重疊資料並發布至解碼裝置亦可。再者,重疊資料除了RGB以外,還具有表示穿透度的a值,伺服器亦可於從三維資料製作的物件以外部分的a值設定為0等,在該部分會穿透的狀態下編碼。或者,伺服器亦可如色度鍵,於背景設定預定值的RGB值,物件以外的部分設為背景色而生成資料。In the AR image, the server superimposes the virtual object information on the virtual space according to the three-dimensional position or the user's viewpoint movement. The decoding device can also acquire or maintain virtual object information and three-dimensional data, generate a two-dimensional image in response to the user's viewpoint movement, and smoothly join to create overlapping data. Moreover, in addition to the request for transmitting the virtual object information, the decoding device transmits the viewpoint movement of the user to the server, and the server cooperates with the viewpoint received from the three-dimensional data held by the server to generate overlapping data, and encodes the overlapping data. It can also be released to the decoding device. Furthermore, the superimposed data has a value indicating the degree of penetration in addition to RGB, and the server can also set the value of a other than the object created from the three-dimensional data to 0, etc., and encode the state in which the part is penetrated. . Alternatively, the server may also set the RGB value of the predetermined value in the background as the chroma key, and the part other than the object is set as the background color to generate the data.
同樣地,發布資料的解碼處理在作為客戶的各終端裝置或於伺服器側進行均可,或亦可互相分擔進行。作為一例,某終端裝置亦可一旦對伺服器發送接收要求,於其他終端裝置接收因應該要求的內容,進行解碼處理,對具有顯示器的裝置,發送解碼完畢的訊號。不受可通訊的終端裝置本身的性能影響,將處理分散,選擇適當的內容,藉此可再生畫質良好的資料。又,作為其他例,亦可於TV等接收大尺寸的圖像資料,同時於觀賞者的個人終端裝置,解碼圖片經分割的方塊等一部分區域而顯示。藉此,可共有全體圖像,同時在手邊確認自身的負責領域或欲更詳細確認的區域。Similarly, the decoding process of the distribution data may be performed on each terminal device as a client or on the server side, or may be shared with each other. As an example, when a certain terminal device transmits and receives a request to the server, the other terminal device receives the content requested, performs decoding processing, and transmits the decoded signal to the device having the display. It is not affected by the performance of the communication-enabled terminal device itself, and the processing is dispersed, and appropriate content is selected, thereby reproducing data with good image quality. Further, as another example, it is also possible to receive a large-sized image data on a TV or the like, and to display a partial region such as a divided square in a viewer's personal terminal device. Thereby, the entire image can be shared, and at the same time, the area of responsibility of the user or the area to be confirmed in more detail can be confirmed at hand.
又,今後預料在不受屋內外的影響,可使用複數種近距離、中距離或長距離的無線通訊的狀況下,利用MPEG-DASH等發布系統規格,對於連接中的通訊一面切換適當的資料,一面無縫地接收內容。藉此,使用者不侷限於自身的終端裝置,可自由地選擇設置於屋內外的顯示器等之解碼裝置或顯示裝置,一面即時地切換。又,可根據自身的位置資訊等,一面切換解碼的終端裝置及顯示的終端裝置,一面解碼。藉此,亦可於往目的地移動中,一面使嵌入可顯示器件的相鄰建築物的壁面或地面的一部分,顯示地圖資訊,一面移動。又,亦可根據在網路上對編碼資料的存取容易度,諸如編碼資料於可從接收終端裝置短時間存取的伺服器進行快取,或複製於內容發布服務的邊緣伺服器等,來切換接收資料的位元率。
[可適性編碼]In addition, in the future, it is expected that the use of multiple types of short-range, medium-range or long-distance wireless communication can be used to switch the appropriate data to the communication during the connection using the MPEG-DASH and other wireless communication systems. , receiving content seamlessly. Thereby, the user can switch to the decoding device or the display device such as a display installed indoors or outdoors without being limited to the terminal device of the home. Further, it is possible to decode the decoded terminal device and the displayed terminal device based on the position information of the user or the like. Thereby, it is also possible to move the map information while displaying the map information on the wall surface or a part of the floor of the adjacent building in which the display device is embedded while moving to the destination. Moreover, according to the ease of accessing the encoded data on the network, such as the encoded data being cached by a server that can be accessed from the receiving terminal device for a short time, or copied to an edge server of the content publishing service, etc. Switch the bit rate of the received data.
[Adaptability coding]
關於內容切換,利用圖28所示,應用上述各實施形態所示動態圖像編碼方法所壓縮編碼的可適性串流來說明。伺服器具有複數種內容相同、質不同的串流作為個別串流亦無妨,但亦可如圖示,活用藉由分層進行編碼所實現的時間性/空間性可適性串流的特徵,來切換內容而構成亦可。總言之,解碼側因應性能之內在要因與通訊頻帶狀態等外在要因,來決定解碼所至之層,解碼側藉此可自由切換低解像度的內容與高解像度的內容而解碼。例如移動中以智慧型手機ex115視聽的影像的後續,回家後想在網際網路TV等機器視聽時,該機器只要將相同串流解碼至不同層即可,因此可減輕使用者側的負擔。Regarding the content switching, the adaptive stream of the compression encoding by the moving image encoding method shown in each of the above embodiments will be described with reference to FIG. The server has a plurality of streams of the same content and different quality as individual streams, but it is also possible to use the characteristics of temporal/temporal adaptability realized by layered coding as shown in the figure. It is also possible to switch contents. In summary, the decoding side determines the layer to be decoded due to the external factors such as the internal factors of the performance and the state of the communication band, and the decoding side can freely switch the content of the low resolution and the content of the high resolution to decode. For example, in the case of moving the video viewed by the smart phone ex115, when the user wants to view the Internet TV or the like after returning home, the device can decode the same stream to different layers, thereby reducing the burden on the user side. .
進而言之,除了如上述,實現就每層編碼圖片,於基礎層的高位存在有增強層的可適性的構成以外,增強層包含基於圖像統計資訊等之元資訊,解碼側根據元資訊,來將基礎層的圖片進行超解像,藉此生成高畫質化的內容亦可。超解像亦可指同一解像度的SN比提升及解像度擴大的任一者。元資訊包含用以特定使用於超解像處理的線性或非線性濾波器係數的資訊,或特定使用於超解像處理的濾波器處理、機械學習或最小平方運算的參數值的資訊等。Further, in addition to the above, the implementation of the code for each layer, the presence of the enhancement layer in the upper layer of the base layer, the enhancement layer includes meta-information based on image statistics and information, and the decoding side is based on the meta-information. It is also possible to super-resolution images of the base layer to generate high-quality images. Super-resolution can also refer to any of the SN ratio enhancement and resolution expansion of the same resolution. The meta information includes information for specifying linear or nonlinear filter coefficients used for super-resolution processing, or information for parameter values specific to filter processing, mechanical learning, or least squares operations for super-resolution processing.
又,亦可構成如:因應圖像內的物件等的語意,將圖片分割為方塊等,解碼側選擇解碼的方塊,藉此僅解碼一部分區域。又,將物件屬性(人物、車、球等)及影像內位置(同一圖像內的座標位置等)作為元資訊儲存,藉此解碼側可根據元資訊,特定所需物件的位置,決定包含該物件的方塊。例如如圖29所示,元資訊利用HEVC的SEI訊息等,與像素資料不同的資料儲存構造來儲存。該元資訊表示例如主物件的位置、大小或色彩等。Further, for example, the picture may be divided into squares or the like in accordance with the semantics of the object or the like in the image, and the decoding side selects the decoded block, thereby decoding only a part of the area. Moreover, the object attributes (person, car, ball, etc.) and the position within the image (coordinate position in the same image, etc.) are stored as meta information, whereby the decoding side can determine the position of the desired object based on the meta information, and decide to include The square of the object. For example, as shown in FIG. 29, the meta information is stored in a material storage structure different from the pixel material by using the SEI message of HEVC or the like. The meta information indicates, for example, the position, size, or color of the main object.
又,以串流、序列、隨機存取單位等,以由複數個圖片所構成的單位,來儲存元資訊亦可。藉此,解碼側可取得特定人物出現在影像內的時刻等,藉由配合圖片單位的資訊,可特定物件所存在的圖片及圖片內物件的位置。
[網頁最佳化]Further, the metadata may be stored in units of a plurality of pictures in a stream, a sequence, a random access unit, or the like. Thereby, the decoding side can obtain the time when the specific person appears in the image, and the information of the picture unit can be used to specify the picture existing in the object and the position of the object in the picture.
[Webpage optimization]
圖30是表示電腦ex111等之網頁的顯示畫面例的圖。圖31是表示智慧型手機ex115等之網頁的顯示畫面例的圖。如圖30及圖31所示,網頁有時包含複數個對圖像內容的連結的連結圖像,依瀏覽的器件,其觀看方式會不同。畫面上可看到複數個連結圖像時,顯示裝置(解碼裝置)顯示各內容所具有的靜止圖或I圖片作為連結圖像,或顯示複數個靜止圖或I圖像等諸如gif動畫的影像,亦或僅接收基礎層,解碼及顯示影像,直到使用者明確地選擇連結圖像,或連結圖像靠近圖像中央附近,或者連結圖像全體進入畫面內為止。FIG. 30 is a view showing an example of a display screen of a web page such as the computer ex111. FIG. 31 is a view showing an example of a display screen of a web page such as the smartphone ex115. As shown in FIG. 30 and FIG. 31, a web page sometimes includes a plurality of connected images of links to image content, and the viewing mode may be different depending on the device being browsed. When a plurality of connected images are visible on the screen, the display device (decoding device) displays a still image or an I picture of each content as a connected image, or displays a plurality of images such as gif animation such as still images or I images. Or only receive the base layer, decode and display the image until the user explicitly selects the connected image, or the connected image is near the center of the image, or the connected image is all entered into the screen.
由使用者選擇了連結圖像時,顯示裝置將基礎層最優先解碼。再者,構成網頁的HTML具有表示可適性內容的資訊時,顯示裝置解碼至增強層亦可。又,為了保證即時性,在選擇前或通訊頻帶非常嚴苛時,顯示裝置僅解碼及顯示前方參考的圖片(I圖片、P圖片、僅前方參考的B圖片),藉此可減低開頭圖片的解碼時刻與顯示時刻之間的延遲(從內容開始解碼到開始顯示的延遲)。又,顯示裝置亦可特意忽視圖片的參考關係,進行前方參考,粗略地解碼所有B圖片及P圖片,隨著時間經過,接收的圖片增加而進行正常解碼。
[自動行駛]When the user selects the connected image, the display device decodes the base layer first. Furthermore, when the HTML constituting the web page has information indicating the content of the adaptability, the display device may decode the enhancement layer. Moreover, in order to ensure immediacy, before the selection or the communication band is very strict, the display device only decodes and displays the picture referenced in front (I picture, P picture, B picture only referenced in front), thereby reducing the initial picture. The delay between the decoding time and the display time (the delay from the start of decoding of the content to the start of display). Moreover, the display device can also deliberately ignore the reference relationship of the picture, perform forward reference, and roughly decode all B pictures and P pictures. As time passes, the received pictures increase and perform normal decoding.
[Automatic driving]
又,為了車輛自動行駛或支援行駛而接收二維或三維地圖資訊等靜止圖或影像資料時,接收終端裝置亦可除了接收屬於1以上的層的圖像資料以外,亦接收天候或施工資訊等作為元資訊,使該等相對應而解碼。再者,元資訊屬於層,或單純與圖像資料進行多工均可。In addition, when receiving still images or video data such as two-dimensional or three-dimensional map information for automatic vehicle driving or support driving, the receiving terminal device may receive weather information or construction information in addition to image data belonging to one or more layers. As meta information, the corresponding decoding is performed. Furthermore, the meta information belongs to the layer, or it can be multiplexed with the image data alone.
此時,由於包含接收終端裝置的車輛、無人飛行載具或飛機等進行移動,因此接收終端裝置在接收要求時,發送該接收終端裝置的位置資訊,藉此可一面切換基地局ex106~ex110,一面實現無縫的接收及解碼。又,接收終端裝置可因應使用者的選擇、使用者的狀況或通訊頻帶的狀態,動態切換元資訊的接收程度或地圖資訊的更新程度。At this time, since the vehicle including the receiving terminal device, the unmanned aerial vehicle, or the airplane moves, the receiving terminal device transmits the location information of the receiving terminal device when receiving the request, thereby switching the base stations ex106 to ex110. Seamless reception and decoding on one side. Moreover, the receiving terminal device can dynamically switch the degree of reception of the meta-information or the degree of update of the map information in response to the user's selection, the state of the user, or the state of the communication band.
如以上,於內容供給系統ex100,客戶可即時接收由使用者發送的經編碼的資訊,予以解碼並再生。
[個人內容發布]As described above, in the content supply system ex100, the client can immediately receive the encoded information transmitted by the user, decode it, and reproduce it.
[Personal Content Release]
又,於內容供給系統ex100,不僅可發布來自影像發布業者的高畫質、長時間的內容,亦可進行來自個人的低畫質、短時間的內容的單播或組播發布。又,該類個人內容今後應會日益增加。為了使個人內容成為更優質的內容,伺服器進行編輯處理後再進行編碼處理亦可。此可由例如以下構成來實現。Further, in the content supply system ex100, not only high-quality, long-time content from the image distribution company but also unicast or multicast distribution of low-quality, short-time content from the individual can be performed. Also, such personal content should increase in the future. In order to make personal content into better content, the server may perform editing processing and then perform encoding processing. This can be achieved, for example, by the following constitution.
攝影時,伺服器即時或積存而在攝影後,從原圖像或編碼完畢的資料,進行攝影錯誤、場景估計、語意分析及物件檢出等辨識處理。然後,伺服器根據辨識結果,進行如下編輯:採手動或自動修正失焦或手震等,刪除亮度比其他圖片低、焦點未對準的場景等重要性低的場景,強調物件的邊緣,或使色調變化等。伺服器根據編輯結果來編碼編輯後的資料。又,已知若攝影時間過長,收視率會下降,伺服器亦可根據圖像處理結果,因應攝影時間,不僅自動剪輯重要性低的場景,亦剪輯移動少的場景等,以使內容維持在特定時間範圍內。又,伺服器亦可根據場景的語意分析的結果,生成並編碼摘要。During photography, the server immediately or accumulates the image, and after the shooting, performs recognition processing such as shooting error, scene estimation, semantic analysis, and object detection from the original image or the encoded data. Then, according to the identification result, the server performs the following editing: manually or automatically corrects out-of-focus or hand-shock, etc., deletes a scene with low importance such as a scene with lower brightness than other pictures and a misaligned focus, and emphasizes the edge of the object, or Make the color tone change, etc. The server encodes the edited data based on the edited result. Moreover, it is known that if the shooting time is too long, the viewing rate will decrease, and the server can also not only automatically edit scenes with low importance, but also clip scenes with less movement, depending on the image processing result, in order to maintain the content. Within a specific time frame. In addition, the server can generate and encode a digest based on the semantic analysis of the scene.
再者,於個人內容亦出現直接拍到侵害著作權、著作者人權或肖像權等之內容的個案,亦出現共有的範圍超越了意圖的範圍等,對個人為不便的情況。故,例如伺服器亦可將圖像變更為特意使焦點,不對準畫面周邊部的人的臉部或家中等而編碼。又,伺服器亦可辨識編碼對象圖像內,是否拍到與預先登錄的人物不同的人物的臉部,拍到時,亦可進行對臉部部分,加上馬賽克等之處理。又,作為編碼前的預處理或後處理,亦可基於著作權等的觀點,指定使用者欲進行圖像加工的人物或背景區域,伺服器進行將指定的區域置換為其他影像,或模糊焦點等處理。若是人物,亦可一面於動態圖像追蹤人物,一面置換臉部部分的影像。In addition, in the case of personal content, cases in which the copyright infringement, the human rights of the author, or the right to portrait are directly photographed are also found, and the scope of the sharing exceeds the scope of the intention, which is inconvenient for the individual. Therefore, for example, the server can also change the image to specifically focus on the image, and the code is not aligned with the face or home of the person in the peripheral portion of the screen. Further, the server can recognize whether or not the face of the person different from the person registered in advance is captured in the image to be encoded, and when the photograph is taken, the face portion can be subjected to processing such as mosaic. Further, as the pre-processing or post-processing before encoding, a person or a background area in which the user wants to perform image processing may be specified based on the viewpoint of copyright, etc., and the server may replace the designated area with another image, or blur the focus, etc. deal with. If it is a character, you can also replace the image of the face while tracking the character on the moving image.
又,資料量少的個人內容的視聽,強烈要求即時性,因此雖依頻帶區而不同,解碼裝置首先最優先接收基礎層,進行解碼及再生。解碼裝置在此期間接收增強層,於再生循環時等再生2次以上時,包含增強層在內而再生高畫質影像亦可。如此,若是進行可適性編碼的串流,可提供如下體驗:在未選擇時或開始觀看的階段,動畫雖粗略,但隨著串流智慧化,圖像改善。除了可適性編碼以外,將第1次所再生的粗略串流、及參考第1次動畫而編碼的第2次串流,作為1個串流而構成,亦可提供相同的體驗。
[其他使用例]Further, since the viewing of personal content with a small amount of data strongly requires immediacy, the decoding device first receives the base layer first and performs decoding and reproduction, depending on the frequency band region. The decoding device receives the enhancement layer during this period, and may reproduce the high-quality image including the enhancement layer when the reproduction layer is reproduced two or more times during the reproduction cycle. In this way, if the stream is adaptively encoded, the following experience can be provided: the animation is rough at the stage of not selecting or starting to watch, but the image is improved as the stream is intelligent. In addition to the adaptive coding, the rough stream that is reproduced for the first time and the second stream that is encoded with reference to the first animation are configured as one stream, and the same experience can be provided.
[Other use cases]
又,該等編碼或解碼處理一般在各終端裝置所具有的LSIex500處理。LSIex500為單晶片或由複數個晶片所組成的構成均可。再者,將動態圖像編碼或解碼用的軟體,組入電腦ex111等可讀取之某種記錄媒體(CD-ROM、軟碟、或硬碟等),利用該軟體進行編碼或解碼處理亦可。進而言之,智慧型手機ex115附有攝影機時,發送由該攝影機所取得的動畫資料亦可。此時的動畫資料是經智慧型手機ex115所具有的LSIex500進行編碼處理的資料。Further, these encoding or decoding processes are generally processed by the LSI ex500 included in each terminal device. The LSI ex500 is a single wafer or a composite of a plurality of wafers. Furthermore, the software for encoding or decoding moving pictures is incorporated into a recording medium (CD-ROM, floppy disk, hard disk, etc.) readable by a computer such as ex111, and is also encoded or decoded by the software. can. In other words, when the smart phone ex115 is attached to a camera, the animation data obtained by the camera can also be sent. The animation data at this time is data that is encoded by the LSI ex500 included in the smart phone ex115.
再者,LSIex500亦可為下載應用軟體並啟用的構成。此時,終端裝置首先判斷該終端裝置是否支援內容的編碼方式,或是否具有特定服務的執行能力。終端裝置不支援內容的編碼方式時,或不具有特定服務的執行能力時,終端裝置下載內容或應用軟體,其後取得並再生內容。Furthermore, the LSIex500 can also be configured to download application software and enable it. At this time, the terminal device first determines whether the terminal device supports the encoding method of the content or whether it has the execution capability of the specific service. When the terminal device does not support the content encoding method or does not have the execution capability of the specific service, the terminal device downloads the content or the application software, and then acquires and reproduces the content.
又,不限於經由網際網路ex101的內容供給系統ex100,於數位播放用系統,亦可組入上述各實施形態的至少任一種動態圖像編碼裝置(圖像編碼裝置)或動態圖像解碼裝置(圖像解碼裝置)。由於利用衛星等,使播放用電波承載影像及聲音已被多工的多工資料來收發,因此相對於內容供給系統ex100容易進行單播的構成,其差異在於適合組播,但關於編碼處理及解碼處理,可進行同樣的應用。
[硬體構成]Further, the present invention is not limited to the content supply system ex100 via the Internet ex101, and may be incorporated in at least one of the above-described embodiments of the video encoding device (image encoding device) or the moving image decoding device. (Image decoding device). Since the broadcast radio-capable video and the multiplexed multiplexed data are transmitted and received by the satellite or the like, the unicast configuration is easily performed with respect to the content supply system ex100, and the difference is that it is suitable for multicast, but the encoding process and The same application can be performed by decoding processing.
[Hardware composition]
圖32是表示智慧型手機ex115的一例的圖。又,圖33是表示智慧型手機ex115的構成例的圖。智慧型手機ex115具有:天線ex450,用以與基地局ex110之間收發電波;攝影機部ex465,可拍攝影像及靜止圖;及顯示部ex458,顯示由攝影機部ex465所拍攝的影像、及由天線ex450所接收的影像等經解碼的資料。智慧型手機ex115進一步具備:操作部ex466,其為觸控面板等;聲音輸出部ex457,用以輸出聲音或音響的揚聲器等;聲音輸入部ex456,用以進行聲音輸入的麥克風等;記憶體部ex467,可保存拍攝的影像或靜止圖、錄音的聲音、接收的影像或靜止圖、郵件等經編碼的資料,或經解碼的資料;及插槽部ex464,其是與SIMex468的介面部,該SIMex468是用以特定使用者,進行對於以網路為首之各種資料存取的認證。再者,亦可使用外接記憶體,來取代記憶體部ex467。FIG. 32 is a diagram showing an example of the smartphone ex115. In addition, FIG. 33 is a view showing an example of the configuration of the smartphone ex115. The smart phone ex115 has an antenna ex450 for transmitting and receiving radio waves with the base station ex110, a camera unit ex465 for capturing images and a still image, and a display unit ex458 for displaying images captured by the camera unit ex465 and the antenna ex450. The decoded data such as the received image. The smart phone ex115 further includes an operation unit ex466, which is a touch panel or the like, a sound output unit ex457, a speaker for outputting sound or sound, a sound input unit ex456, a microphone for inputting sound, and the like; Ex467, can save the captured image or still picture, recorded sound, received image or still picture, mailed and other encoded data, or decoded data; and slot part ex464, which is the interface with SIMex468, the SIMex468 is used for specific users to perform authentication for various data accesses such as the Internet. Furthermore, an external memory can be used instead of the memory portion ex467.
又,統籌控制顯示部ex458及操作部ex466等的主控制部ex460,經由匯流排ex470而與電源電路部ex461、操作輸入控制部ex462、影像訊號處理部ex455、攝影機介面部ex463、顯示器控制部ex459、調變/解調部ex452、多工/分離部ex453、聲音訊號處理部ex454、插槽部ex464及記憶體部ex467連接。Further, the main control unit ex460 such as the display control unit ex458 and the operation unit ex466 is integrated with the power supply circuit unit ex461, the operation input control unit ex462, the video signal processing unit ex455, the camera interface ex463, and the display control unit ex459 via the bus line ex470. The modulation/demodulation unit ex452, the multiplex/separation unit ex453, the audio signal processing unit ex454, the slot unit ex464, and the memory unit ex467 are connected.
若藉由使用者操作而使電源鍵成為開啟狀態時,電源電路部ex461從電池包對於各部供給電力,將智慧型手機ex115啟動為可動作的狀態。When the power button is turned on by the user's operation, the power supply circuit unit ex461 supplies power to each unit from the battery pack, and activates the smartphone ex115 in an operable state.
智慧型手機ex115根據由CPU、ROM及RAM等所組成的主控制部ex460的控制,進行通話及資料通訊等處理。通話時,以聲音訊號處理部ex454,將由聲音輸入部ex456所收音的聲音訊號,轉換成數位聲音訊號,以調變/解調部ex452,將其進行頻譜擴散處理,以發送/接收部ex451施行數位類比轉換處理及頻率轉換處理後,經由天線ex450發送。又,放大接收資料,施行頻率轉換處理及類比數位轉換處理,以調變/解調部ex452進行頻譜反擴散處理,以聲音訊號處理部ex454轉換成類比聲音資料後,從聲音輸出部ex457將其輸出。於資料通訊模式時,藉由主體部的操作部ex466等之操作,經由操作輸入控制部ex462,將文本、靜止圖或影像資料送出至主控制部ex460,並同樣地進行收發處理。資料通訊模式時發送影像、靜止圖或影像及聲音時,影像訊號處理部ex455將保存於記憶體部ex467的影像訊號、或從攝影機部ex465輸入的影像訊號,藉由上述各實施形態所示動態圖像編碼方法,予以壓縮編碼,將已被編碼的影像資料送出至多工/分離部ex453。又,聲音訊號處理部ex454將聲音訊號編碼,將已被編碼的聲音資料送出至多工/分離部ex453,其中該聲音訊號是在以攝影機部ex465拍攝影像或靜止圖等之同時,以聲音輸入部ex456收音。多工/分離部ex453以預定的方式,將編碼完畢影像資料及編碼完畢聲音資料予以多工,以調變/解調部(調變/解調電路部)ex452及發送/接收部ex451,施以調變處理及轉換處理,經由天線ex450來發送。The smartphone ex115 performs processing such as call and data communication under the control of the main control unit ex460 composed of a CPU, a ROM, and a RAM. At the time of the call, the audio signal processing unit ex454 converts the audio signal received by the voice input unit ex456 into a digital audio signal, and performs the spectrum diffusion processing on the modulation/demodulation unit ex452, and performs the transmission/reception unit ex451. The digital analog conversion processing and the frequency conversion processing are transmitted via the antenna ex450. Further, the received data is amplified, the frequency conversion process and the analog-to-digital conversion process are performed, and the spectrum de-diffusion process is performed by the modulation/demodulation unit ex452, and converted into analog sound data by the audio signal processing unit ex454, and then the audio output unit ex457 is used. Output. In the data communication mode, the text, the still image, or the video data is sent to the main control unit ex460 via the operation input unit ex462 by the operation unit ex466 of the main unit, and the transmission and reception processing is performed in the same manner. When transmitting a video, a still picture, or a video and audio in the data communication mode, the video signal processing unit ex455 updates the video signal stored in the memory unit ex467 or the video signal input from the camera unit ex465 by the above-described embodiments. The image encoding method compresses and encodes the image data that has been encoded to the multiplex/separation unit ex453. Further, the audio signal processing unit ex454 encodes the audio signal, and sends the encoded audio data to the multiplex/separation unit ex453, which is the audio input unit while capturing the video or the still picture by the camera unit ex465. Ex456 radio. The multiplexer/separation unit ex453 multiplexes the encoded video data and the encoded audio data in a predetermined manner, and uses a modulation/demodulation unit (modulation/demodulation circuit unit) ex452 and a transmission/reception unit ex451. The modulation processing and the conversion processing are transmitted via the antenna ex450.
接收添附於電子郵件或聊天的影像、或連結於網頁等之影像時,為了解碼經由天線ex450所接收的多工資料,多工/分離部ex453藉由分離多工資料,來將多工資料區分為影像資料的位元串流與聲音資料的位元串流,經由同步匯流排ex470,將已被編碼的影像資料供給至影像訊號處理部ex455,並且將已被編碼的聲音資料供給至聲音訊號處理部ex454。影像訊號處理部ex455藉由對應於上述各實施形態所示動態圖像編碼方法的動態圖像解碼方法,解碼影像訊號,經由顯示器控制部ex459,從顯示部ex458顯示連結的動態圖像檔所含的影像或靜止圖。又,聲音訊號處理部ex454解碼聲音訊號,從聲音輸出部ex457輸出聲音。再者,由於即時串流化已普及,因此依使用者的狀況,亦可能發生聲音的再生就社會觀點而言不妥的情況。因此,作為初始值,宜採用不再生聲音訊號,僅再生影像訊號的構成。僅於使用者進行點擊影像資料等操作時,同步再生聲音亦可。When receiving an image attached to an e-mail or a chat or an image linked to a web page or the like, in order to decode the multiplexed data received via the antenna ex450, the multiplex/separation unit ex453 separates the multiplexed data by separating the multiplexed data. For the bit stream of the video data and the bit stream of the audio data, the encoded video data is supplied to the video signal processing unit ex455 via the synchronous bus ex470, and the encoded audio data is supplied to the audio signal. Processing unit ex454. The video signal processing unit ex455 decodes the video signal by the moving picture decoding method corresponding to the moving picture coding method described in each of the above embodiments, and displays the connected video file included in the display unit ex458 via the display control unit ex459. Image or still image. Further, the audio signal processing unit ex454 decodes the audio signal and outputs the sound from the audio output unit ex457. Furthermore, since real-time streaming has become widespread, depending on the user's situation, it is also possible that the reproduction of sound is not appropriate from a social point of view. Therefore, as an initial value, it is preferable to use a non-reproduced sound signal to reproduce only the image signal. It is also possible to synchronize the reproduced sound only when the user clicks on an image data or the like.
又,於此以智慧型手機ex115為例說明,但除了作為終端裝置具有編碼器及解碼器雙方的收發型終端裝置以外,亦可考慮僅具有編碼器的發送終端裝置及僅具有解碼器的接收終端裝置等3種實作形式。進而言之,說明了於數位播放用系統,接收或發送對影像資料多工有聲音資料的多工資料,但於多工資料,除了聲音資料以外,與影像相關連的文字資料等亦可受到多工,或不接收或發送多工資料而接收或發送影像資料本身亦可。In addition, although the smart phone ex115 is described as an example here, in addition to the transceiving type terminal device having both the encoder and the decoder as the terminal device, it is also conceivable to have only the transmitting terminal device having the encoder and the receiving only with the decoder. Three implementation forms such as terminal devices. In other words, it describes the digital playback system that receives or transmits multiplexed data that has audio data for multiplexed video data. However, in addition to voice data, text data associated with images can also be received by multiplexed data. It is also possible to receive or transmit image data by multiplexing or not receiving or transmitting multiplexed data.
再者,說明了包含CPU的主控制部ex460控制編碼或解碼處理,但終端裝置亦經常具備GPU。故,亦可構成如:藉由在CPU與GPU被共通化的記憶體,或藉由位址受管理以便可共通使用的記憶體,來活用GPU的性能以統一處理大區域。藉此,可縮短編碼時間,確保即時性,實現低延遲。特言之,不採用CPU而採用GPU,以圖片等為單位,統一進行移動估計、去區塊濾波、SAO(Sample Adaptive Offset(樣本適用性偏移))及轉換‧量化的處理甚有效率。Furthermore, it has been explained that the main control unit ex460 including the CPU controls the encoding or decoding processing, but the terminal device often includes the GPU. Therefore, it is also possible to utilize the performance of the GPU to uniformly process a large area by using a memory that is common to the CPU and the GPU, or a memory that is managed by the address so as to be commonly used. Thereby, the encoding time can be shortened, the immediacy can be ensured, and low latency can be realized. In particular, the GPU is not used for the CPU, and the motion estimation, deblocking filtering, SAO (Sample Adaptive Offset), and conversion ‧ quantization processing are performed efficiently in units of pictures and the like.
亦可將本態樣,與本發明的其他態樣的至少一部分組合而實施。又,亦可將本態樣的流程圖所記載的一部分處理、裝置的一部分構成、語法的一部分等,與其他態樣組合而實施。This aspect can also be implemented in combination with at least a portion of other aspects of the invention. Further, a part of the processing described in the flowchart of the present aspect, a part of the configuration of the device, a part of the grammar, and the like may be combined with other aspects.
產業上之可利用性
本發明可利用於例如電視接收機、數位錄影機、車用導航器、行動電話、數位相機、數位攝影機、電視會議系統或電子鏡等。Industrial Applicability The present invention can be utilized, for example, in a television receiver, a digital video recorder, a car navigator, a mobile phone, a digital camera, a digital camera, a video conferencing system, or an electronic mirror.
10~23‧‧‧區塊10~23‧‧‧ Block
100‧‧‧編碼裝置 100‧‧‧ coding device
102‧‧‧分割部 102‧‧‧ Division
104‧‧‧減算部 104‧‧‧Decrease Department
106‧‧‧轉換部 106‧‧‧Transition Department
108‧‧‧量化部 108‧‧‧Quantity Department
110‧‧‧熵編碼部 110‧‧‧ Entropy Coding Department
112、204‧‧‧反量化部 112, 204‧‧‧Anti-Quantization Department
114、206‧‧‧反轉換部 114, 206‧‧‧Anti-conversion department
116、208‧‧‧加算部 116, 208‧‧‧Additional Department
118、210‧‧‧區塊記憶體 118, 210‧‧‧ Block memory
120、212‧‧‧迴路濾波部 120, 212‧‧‧Circuit Filtering Department
122、214‧‧‧幀記憶體 122, 214‧‧‧ frame memory
124、216‧‧‧幀內預測部 124, 216‧‧ Intra Prediction Department
126、218‧‧‧幀間預測部 126, 218‧‧ interframe prediction department
128、220‧‧‧預測控制部 128, 220‧‧‧Predictive Control Department
160、260‧‧‧電路 160, 260‧‧‧ circuits
162、262‧‧‧記憶體 162, 262‧‧‧ memory
200‧‧‧解碼裝置 200‧‧‧ decoding device
202‧‧‧熵解碼部 202‧‧‧ Entropy Decoding Department
ALF‧‧‧適應性迴路濾波器 ALF‧‧‧Adaptive loop filter
AMT‧‧‧適應性多重轉換 AMT‧‧‧Adaptive multiple conversion
AR‧‧‧擴增實境 AR‧‧‧Augmented Reality
AST‧‧‧適應性第二轉換 AST‧‧ ‧ adaptive second conversion
BIO‧‧‧雙向光流 BIO‧‧‧Two-way optical flow
CCLM‧‧‧跨組件線性模式 CCLM‧‧‧cross-component linear mode
CABAC‧‧‧上下文適應型二值算術編碼方式 CABAC‧‧‧Context-adaptive binary arithmetic coding
CDN‧‧‧內容發布網路 CDN‧‧‧Content Publishing Network
CTU‧‧‧編碼樹單元 CTU‧‧‧ coding tree unit
CU‧‧‧編碼單元 CU‧‧‧ coding unit
Cur block‧‧‧目前區塊 Cur block‧‧‧ current block
DCT‧‧‧離散餘弦變換 DCT‧‧ discrete cosine transform
DF‧‧‧去區塊濾波器 DF‧‧‧deblock filter
DST‧‧‧離散正弦變換 DST‧‧‧discrete sine transform
EMT‧‧‧顯式多重核心轉換 EMT‧‧‧ Explicit multiple core conversion
ex100‧‧‧內容供給系統 Ex100‧‧‧Content Supply System
ex101‧‧‧網際網路 Ex101‧‧‧Internet
ex102‧‧‧網際網路服務提供者 Ex102‧‧‧Internet Service Provider
ex103‧‧‧串流化伺服器 Ex103‧‧‧Streaming server
ex104‧‧‧通訊網 Ex104‧‧‧Communication Network
ex106~ex110‧‧‧基地局 Ex106~ex110‧‧‧base office
ex111‧‧‧電腦 Ex111‧‧‧ computer
ex112‧‧‧遊戲機 Ex112‧‧‧game machine
ex113‧‧‧攝影機 Ex113‧‧‧ camera
ex114‧‧‧家電 Ex114‧‧‧Home appliances
ex115‧‧‧智慧型手機 Ex115‧‧‧Smart mobile phone
ex116‧‧‧衛星 Ex116‧‧‧ satellite
ex117‧‧‧飛機 Ex117‧‧ aircraft
ex450‧‧‧天線 Ex450‧‧‧Antenna
ex451‧‧‧發送/接收部 Ex451‧‧‧Send/Receive Department
ex452‧‧‧調變/解調部 Ex452‧‧‧Modulation/Demodulation Department
ex453‧‧‧多工/分離部 Ex453‧‧‧Multiplex/Separation Department
ex454‧‧‧聲音訊號處理部 Ex454‧‧‧Sound Signal Processing Department
ex455‧‧‧影像訊號處理部 Ex455‧‧‧Image Signal Processing Department
ex456‧‧‧聲音輸入部 Ex456‧‧‧Sound Input Department
ex457‧‧‧聲音輸出部 Ex457‧‧‧Sound Output Department
ex458‧‧‧顯示部 Ex458‧‧‧Display Department
ex459‧‧‧顯示器控制部 Ex459‧‧‧Display Control Department
ex460‧‧‧主控制部 Ex460‧‧‧Main Control Department
ex461‧‧‧電源電路部 Ex461‧‧‧Power Circuit Department
ex462‧‧‧操作輸入控制部 Ex462‧‧‧Operation Input Control Department
ex463‧‧‧攝影機介面部 Ex463‧‧‧ camera face
ex464‧‧‧插槽部 Ex464‧‧‧Slots
ex465‧‧‧攝影機部 Ex465‧‧‧Photo Camera Department
ex466‧‧‧操作部 Ex466‧‧‧Operation Department
ex467‧‧‧記憶體部 Ex467‧‧‧ memory department
ex468‧‧‧SIM Ex468‧‧‧SIM
ex470‧‧‧匯流排、同步匯流排 Ex470‧‧‧ bus bar, synchronous bus
ex500‧‧‧LSI ex500‧‧‧LSI
FRUC‧‧‧幀率提升轉換 FRUC‧‧‧ frame rate upgrade conversion
GOP‧‧‧圖片群組 GOP‧‧‧Photo Group
HEVC‧‧‧高效率視訊編碼 HEVC‧‧‧High Efficiency Video Coding
LIC‧‧‧局部照明補償 LIC‧‧‧Local Lighting Compensation
MBT‧‧‧多型樹 MBT‧‧‧Multiple Tree
MV、MV0、MV1、MV_L、MV_U‧‧‧移動向量 MV, MV0, MV1, MV_L, MV_U‧‧‧ mobile vector
MVC‧‧‧多視角編碼 MVC‧‧‧Multi-view coding
NSST‧‧‧不可分第二轉換 NSST‧‧‧ can not be divided into second conversion
OBMC‧‧‧重疊區塊移動補償 OBMC‧‧‧Overlap block motion compensation
p0、p1、p2、p3‧‧‧亮度像素值 P0, p1, p2, p3‧‧‧ brightness pixel values
PDPC‧‧‧獨立位置幀內預測組合 PDPC‧‧‧Independent Position Intra Prediction Combination
PMMVD‧‧‧模式匹配移動向量導出 PMMVD‧‧‧ pattern matching mobile vector export
Pred、Pred_L、Pred_U‧‧‧預測圖像 Pred, Pred_L, Pred_U‧‧‧ Forecast Image
PU‧‧‧預測單元 PU‧‧‧ forecasting unit
QP‧‧‧量化參數 QP‧‧‧Quantification parameters
QTBT‧‧‧四元樹加二元樹 QTBT‧‧‧ ternary tree plus binary tree
Ref0、Ref1‧‧‧參考圖片 Ref0, Ref1‧‧‧ reference picture
S201~S206、S211~S215、S211A、S212A、S221~S225、S221A、S222A、S231~S233‧‧‧步驟 Steps S201 to S206, S211 to S215, S211A, S212A, S221 to S225, S221A, S222A, S231 to S233‧‧
SAO‧‧‧樣本適用性偏移 SAO‧‧‧ sample applicability offset
TU‧‧‧轉換單元 TU‧‧ conversion unit
v0、v1、vx、vy‧‧‧移動向量v 0 , v 1 , v x , v y ‧‧‧moving vectors
VLC‧‧‧可變長度編碼 VLC‧‧‧Variable length coding
VR‧‧‧虛擬實境 VR‧‧‧Virtual Reality
圖1是表示實施形態1的編碼裝置的功能構成的方塊圖。Fig. 1 is a block diagram showing a functional configuration of an encoding apparatus according to a first embodiment.
圖2是表示實施形態1的區塊分割的一例的圖。Fig. 2 is a view showing an example of block division in the first embodiment;
圖3是表示對應於各轉換類型的轉換基底函數的表。FIG. 3 is a table showing a conversion basis function corresponding to each conversion type.
圖4A是表示ALF所用的濾波器的形狀的一例的圖。4A is a view showing an example of a shape of a filter used for ALF.
圖4B是表示ALF所用的濾波器的形狀的其他一例的圖。4B is a view showing another example of the shape of the filter used for the ALF.
圖4C是表示ALF所用的濾波器的形狀的其他一例的圖。4C is a view showing another example of the shape of a filter used for ALF.
圖5A是表示幀內預測的67個幀內預測模式的圖。FIG. 5A is a diagram showing 67 intra prediction modes of intra prediction.
圖5B是用以說明OBMC處理的預測圖像修正處理的概要的流程圖。FIG. 5B is a flowchart for explaining an outline of a predicted image correction process of the OBMC process.
圖5C是用以說明OBMC處理的預測圖像修正處理的概要的概念圖。FIG. 5C is a conceptual diagram for explaining an outline of a predicted image correction process of the OBMC process.
圖5D是表示FRUC的一例的圖。FIG. 5D is a diagram showing an example of FRUC.
圖6是用以說明沿著移動軌道的2個區塊間的模式匹配(雙向匹配)的圖。Figure 6 is a diagram for explaining pattern matching (bidirectional matching) between two blocks along a moving track.
圖7是用以說明目前圖片內的模板與參考圖片內的區塊之間的模式匹配(模板匹配)的圖。FIG. 7 is a diagram for explaining pattern matching (template matching) between a template in a current picture and a block in a reference picture.
圖8是用以說明假定等速直線運動的模型的圖。Fig. 8 is a view for explaining a model assuming constant-speed linear motion.
圖9A是用以說明根據複數個鄰接區塊的移動向量來導出子區塊單位的移動向量的圖。9A is a diagram for explaining a motion vector of a sub-block unit derived from a motion vector of a plurality of adjacent blocks.
圖9B是用以說明利用合併模式的移動向量導出處理的概要的圖。FIG. 9B is a diagram for explaining an outline of a motion vector derivation process using a merge mode.
圖9C是用以說明DMVR處理的概要的概念圖。9C is a conceptual diagram for explaining an outline of DMVR processing.
圖9D是用以說明利用LIC處理的亮度修正處理的預測圖像生成方法的概要的圖。9D is a diagram for explaining an outline of a predicted image generating method of brightness correction processing by LIC processing.
圖10是表示實施形態1的解碼裝置的功能構成的方塊圖。Fig. 10 is a block diagram showing a functional configuration of a decoding device according to the first embodiment.
圖11是表示實施形態1的去區塊濾波處理的第1態樣的流程圖。Fig. 11 is a flowchart showing a first aspect of the deblocking filtering process in the first embodiment.
圖12是表示實施形態1的第1態樣的Bs算出方法例的圖。FIG. 12 is a view showing an example of a Bs calculation method according to the first aspect of the first embodiment.
圖13是表示實施形態1的第1態樣的Bs算出處理的第1例的流程圖。FIG. 13 is a flowchart showing a first example of Bs calculation processing in the first aspect of the first embodiment.
圖14是表示實施形態1的第1態樣的Bs算出方法的第1例的圖。FIG. 14 is a view showing a first example of the Bs calculation method according to the first aspect of the first embodiment.
圖15是表示實施形態1的第1態樣的Bs算出處理的第2例的流程圖。Fig. 15 is a flowchart showing a second example of Bs calculation processing in the first aspect of the first embodiment.
圖16是表示實施形態1的第1態樣的Bs算出方法的第2例的圖。FIG. 16 is a view showing a second example of the Bs calculation method according to the first aspect of the first embodiment.
圖17是表示實施形態1的對象區塊及鄰接區塊的狀況的圖。Fig. 17 is a view showing the state of the target block and the adjacent block in the first embodiment;
圖18是表示實施形態1的對象區塊及鄰接區塊的狀況的圖。Fig. 18 is a view showing the state of the target block and the adjacent block in the first embodiment;
圖19是表示實施形態1的第2態樣的Bs算出方法例的圖。Fig. 19 is a view showing an example of a Bs calculation method according to a second aspect of the first embodiment.
圖20是表示實施形態1的第2態樣的Bs算出處理的第1例的流程圖。FIG. 20 is a flowchart showing a first example of Bs calculation processing in the second aspect of the first embodiment.
圖21是表示實施形態1的第2態樣的Bs算出方法的第1例的圖。FIG. 21 is a view showing a first example of the Bs calculation method according to the second aspect of the first embodiment.
圖22是表示實施形態1的第2態樣的Bs算出處理的第2例的流程圖。Fig. 22 is a flowchart showing a second example of Bs calculation processing in the second aspect of the first embodiment.
圖23是表示實施形態1的第2態樣的Bs算出方法的第2例的圖。FIG. 23 is a view showing a second example of the Bs calculation method according to the second aspect of the first embodiment.
圖24是實施形態1的去區塊濾波處理的流程圖。Fig. 24 is a flowchart showing the deblocking filtering process of the first embodiment.
圖25是表示編碼裝置的實作例的方塊圖。Figure 25 is a block diagram showing a practical example of an encoding device.
圖26是表示解碼裝置的實作例的方塊圖。Fig. 26 is a block diagram showing a practical example of the decoding device.
圖27是實現內容發布服務的內容供給系統的全體構成圖。Fig. 27 is a view showing the overall configuration of a content supply system for realizing a content distribution service.
圖28是表示可適性編碼時的編碼構造的一例的圖。FIG. 28 is a diagram showing an example of a coding structure at the time of adaptive coding.
圖29是表示可適性編碼時的編碼構造的一例的圖。FIG. 29 is a diagram showing an example of a coding structure at the time of adaptive coding.
圖30是表示網頁的顯示畫面例的圖。FIG. 30 is a view showing an example of a display screen of a web page.
圖31是表示網頁的顯示畫面例的圖。31 is a diagram showing an example of a display screen of a web page.
圖32是表示智慧型手機的一例的圖。32 is a diagram showing an example of a smart phone.
圖33是表示智慧型手機的構成例的方塊圖。Fig. 33 is a block diagram showing an example of the configuration of a smart phone.
Claims (12)
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201762581984P | 2017-11-06 | 2017-11-06 | |
US62/581,984 | 2017-11-06 |
Publications (1)
Publication Number | Publication Date |
---|---|
TW201924333A true TW201924333A (en) | 2019-06-16 |
Family
ID=66331415
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
TW107139023A TW201924333A (en) | 2017-11-06 | 2018-11-02 | Encoding device, decoding device, encoding method, and decoding method |
Country Status (2)
Country | Link |
---|---|
TW (1) | TW201924333A (en) |
WO (1) | WO2019088162A1 (en) |
Family Cites Families (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP2116061A2 (en) * | 2007-01-04 | 2009-11-11 | Thomson Licensing | Methods and apparatus for reducing coding artifacts for illumination compensation and/or color compensation in multi-view coded video |
WO2008084996A1 (en) * | 2007-01-11 | 2008-07-17 | Samsung Electronics Co, . Ltd. | Method and apparatus for deblocking-filtering video data |
-
2018
- 2018-10-31 WO PCT/JP2018/040490 patent/WO2019088162A1/en active Application Filing
- 2018-11-02 TW TW107139023A patent/TW201924333A/en unknown
Also Published As
Publication number | Publication date |
---|---|
WO2019088162A1 (en) | 2019-05-09 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
KR102608008B1 (en) | Encoding device, decoding device, encoding method and decoding method | |
JP7422811B2 (en) | non-transitory storage medium | |
TWI812414B (en) | Coding device, decoding device and recording medium | |
TW202002648A (en) | Encoding device, decoding device, encoding method, and decoding method | |
TW201941603A (en) | Coding device, decoding device, coding method, and decoding method | |
JP7408862B2 (en) | Encoding device and decoding device | |
JPWO2018186433A1 (en) | Encoding device, decoding device, encoding method and decoding method | |
WO2018097115A1 (en) | Encoding device, decoding device, encoding method, and decoding method | |
TW201933869A (en) | Encoding device, decoding device, encoding method, and decoding method | |
TW201924341A (en) | Coding device, decoding device, coding method and decoding method | |
JP7442607B2 (en) | Encoding method and decoding method | |
JP7364768B2 (en) | Encoding device and decoding device | |
JPWO2018186430A1 (en) | Encoding device, decoding device, encoding method and decoding method | |
TW201904293A (en) | Encoding device, encoding method, decoding device and decoding method | |
TW201911865A (en) | Encoding device, encoding method, decoding device and decoding method | |
JPWO2018186429A1 (en) | Encoding device, decoding device, encoding method and decoding method | |
WO2018097078A1 (en) | Encoding device, decoding device, encoding method, and decoding method | |
WO2018097117A1 (en) | Encoding device, decoding device, encoding method, and decoding method | |
WO2019031369A1 (en) | Coding device, decoding device, coding method and decoding method | |
WO2019021803A1 (en) | Encoding device, decoding device, encoding method, and decoding method | |
CN112136326B (en) | Encoding device, decoding device, encoding method, and decoding method | |
WO2019244809A1 (en) | Coding device, decoding device, coding method, and decoding method | |
WO2019146718A1 (en) | Coding device, decoding device, coding method, and decoding method | |
WO2019098092A1 (en) | Encoding device, decoding device, encoding method, and decoding method | |
WO2018097116A1 (en) | Encoding device, decoding device, encoding method, and decoding method |