WO2013113274A1 - Method and apparatus for loop filtering cross tile or slice boundaries - Google Patents

Method and apparatus for loop filtering cross tile or slice boundaries Download PDF

Info

Publication number
WO2013113274A1
WO2013113274A1 PCT/CN2013/071108 CN2013071108W WO2013113274A1 WO 2013113274 A1 WO2013113274 A1 WO 2013113274A1 CN 2013071108 W CN2013071108 W CN 2013071108W WO 2013113274 A1 WO2013113274 A1 WO 2013113274A1
Authority
WO
WIPO (PCT)
Prior art keywords
tile
slice
loop filter
pixels
boundary
Prior art date
Application number
PCT/CN2013/071108
Other languages
French (fr)
Inventor
Chih-Wei Hsu
Chia-Yang Tsai
Yu-Wen Huang
Original Assignee
Mediatek Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Mediatek Inc. filed Critical Mediatek Inc.
Priority to NZ616358A priority Critical patent/NZ616358A/en
Priority to US14/374,918 priority patent/US9967563B2/en
Priority to BR112013027865-0A priority patent/BR112013027865B1/en
Priority to EP13743612.7A priority patent/EP2735162A4/en
Priority to CN201380001869.1A priority patent/CN103636227B/en
Publication of WO2013113274A1 publication Critical patent/WO2013113274A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/124Quantisation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/117Filters, e.g. for pre-processing or post-processing
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/136Incoming video signal characteristics or properties
    • H04N19/14Coding unit complexity, e.g. amount of activity or edge presence estimation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/174Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a slice, e.g. a line of blocks or a group of blocks
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/176Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/80Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation
    • H04N19/82Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation involving filtering within a prediction loop
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/85Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression
    • H04N19/86Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression involving reduction of coding artifacts, e.g. of blockiness
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/154Measured or subjectively estimated visual quality after decoding, e.g. measurement of distortion

Definitions

  • the present invention relates to video coding.
  • the present invention relates to video coding techniques associated with loop filter processing at tile or slice boundaries.
  • Motion estimation is an effective inter-frame coding technique to exploit temporal redundancy in video sequences.
  • Motion-compensated inter- frame coding has been widely used in various international video coding standards.
  • the motion estimation adopted in various coding standards is often a block-based technique, where motion information such as coding mode and motion vector is determined for each macroblock or similar block configuration.
  • intra-coding is also adaptively applied, where the picture is processed without reference to any other picture.
  • the inter-predicted or intra-predicted residues are usually further processed by transformation, quantization, and entropy coding to generate a compressed video bitstream.
  • coding artifacts are introduced, particularly by the quantization process.
  • additional processing is often applied to the reconstructed video to enhance picture quality.
  • the additional processing is configured in an in-loop operation so that the encoder and decoder may derive the same reference pictures for proper operation.
  • Fig. 1 illustrates an exemplary adaptive inter/intra video coding system incorporating in-loop filter processing.
  • Motion Estimation (ME)/Motion Compensation (MC) 112 is used to provide prediction data based on video data from other picture or pictures.
  • Switch 114 selects Intra Prediction 110 or inter-prediction data from ME/MC 112 and the selected prediction data is supplied to Adder 116 to form prediction errors, also called prediction residues or residues.
  • the prediction error is then processed by Transformation (T) 118 followed by Quantization (Q) 120.
  • T Transformation
  • Q Quantization
  • the transformed and quantized residues are then coded by Entropy Encoder 122 to form a video bitstream corresponding to the compressed video data.
  • the bitstream associated with the transform coefficients is then packed with side information such as motion, mode, and other information associated with the image unit.
  • the side information may also be processed by entropy coding to reduce required bandwidth. Accordingly, the side information data is also provided to Entropy Encoder 122 as shown in Fig. 1 (the motion/mode paths to Entropy Encoder 122 are not shown).
  • a reconstruction loop is used to generate reconstructed pictures at the encoder end. Consequently, the transformed and quantized residues are processed by Inverse Quantization (IQ) 124 and Inverse Transformation (IT) 126 to recover the processed residues.
  • the processed residues are then added back to prediction data 136 by Reconstruction (REC) 128 to reconstruct the video data.
  • the reconstructed video data may be stored in Reference Picture Buffer 134 and used for prediction of other frames.
  • incoming video data undergoes a series of processing in the encoding system.
  • the reconstructed video data from REC 128 may be subject to various impairments due to the series of processing. Accordingly, various loop processing is applied to the reconstructed video data before the reconstructed video data is used as prediction data in order to improve video quality.
  • HEVC High Efficiency Video Coding
  • Deblocking Filter (DF) 130 Sample Adaptive Offset (SAO) 131 and Adaptive Loop Filter (ALF) 132 have been developed to enhance picture quality.
  • the Deblocking Filter (DF) 130 is applied to boundary pixels and the DF processing is dependent on the underlying pixel data and coding information associated with the corresponding blocks.
  • the SAO and ALF processing are adaptive, where filter information such as filter parameters and filter type may be dynamically changed according to the underlying video data. Therefore, filter information associated with SAO and ALF is incorporated in the video bitstream so that a decoder can properly recover the required information. Therefore, filter information from SAO and ALF is provided to Entropy Encoder 122 for incorporation into the bitstream.
  • DF 130 is applied to the reconstructed video first; SAO 131 is then applied to DF-processed video; and ALF 132 is applied to SAO-processed video.
  • the processing order among DF, SAO and ALF may be re-arranged.
  • the loop filter processing only includes DF.
  • the loop filter processing includes DF and SAO.
  • HM-5.0 HEVC Test Model Version 5.0
  • a picture is partitioned into multiple slices or tiles, where the image unit partition is often aligned with boundaries of Largest Coding Units (LCUs).
  • LCUs Largest Coding Units
  • the processing sequence of the tiles may be according to raster scan order.
  • Slices and tiles can be configured independently. Therefore, one slice may run across multiple tiles, and one tile may also run across multiple slices.
  • Loop filter processing such as DF, SAO and ALF may depend on neighboring pixels. Accordingly, the loop filter processing for a current tile or slice may have to wait until some neighboring slices or tiles become available.
  • Slices or independent tiles are mainly designed for parallel processing, where reconstructing LCUs (e.g. MV prediction, intra prediction, entropy coding) within one slice or tile does not need any data from other tiles.
  • LCUs e.g. MV prediction, intra prediction, entropy coding
  • loop filtering can be applied across slice or tile boundaries or restricted to within a slice or a tile according to a flag. While in the mode that coding process is applied within the slice or independent tile, the loop filter processing may still need some information from a neighboring slice or tile. For example, when DF is applied across a block boundary, the DF process depends on the QP (quantization parameter) values from both sides of the block boundary.
  • QP quantization parameter
  • the DF process depends on the QP values on both sides of the slice or tile. This will be a problem if the coding process is intended for not crossing the slice or tile boundary. Accordingly, it is desirable to develop a method and apparatus that can overcome dependency on parameter across a slice or tile boundary.
  • a method and apparatus for loop filter processing of boundary pixels across a block boundary are disclosed.
  • the block boundary has first pixels on a first side and second pixels on a second side of the block boundary.
  • the loop filter processing is applied to the boundary pixels across the block boundary depends on a first parameter for the first pixels on the first side and a second parameter for the second pixels on the second side.
  • Embodiments according to the present invention adaptively apply loop filter processing across slice or tile boundaries according to a flag indicating whether cross slice or tile loop filter processing is allowed or not.
  • the method comprises receiving reconstructed video data associated with a picture; determining a value of a cross-slice loop filter flag or a cross-tile loop filter flag corresponding to the slice or the tile respectively; and applying said loop filter processing to the boundary pixels across a slice or tile boundary aligning a block boundary according to the value of the cross-slice loop filter flag or the cross-tile loop filter flag, wherein said loop filter processing for the boundary pixels of the slice or tile boundary depends on both the first parameter for the first pixels inside the slice or tile and the second parameter for the second pixels outside the slice or tile.
  • the first parameter and the second parameter correspond to quantization parameters (QPs) on both sides of the block boundary.
  • QPs quantization parameters
  • For a slice or tile if the value of the cross-slice loop filter flag or the cross-tile loop filter flag indicates that no cross-slice or cross-tile filter loop processing is allowed, the quantization parameter corresponding to a neighboring slice or tile is set unavailable for the loop filter processing.
  • the quantization parameter corresponding to a neighboring slice or tile is provided for the loop filter processing. In this case, the loop filter processing is applied to the boundary pixels of the slice or tile using an average quantization parameter of the first quantization parameter and the second quantization parameter.
  • FIG. 1 illustrates an exemplary video coding system using Inter/Intra prediction, where loop filter processing including Deblocking Filter (DF), Sample Adaptive Offset (SAO) and Adaptive Loop Filter (ALF) is incorporated.
  • loop filter processing including Deblocking Filter (DF), Sample Adaptive Offset (SAO) and Adaptive Loop Filter (ALF) is incorporated.
  • DF Deblocking Filter
  • SAO Sample Adaptive Offset
  • ALF Adaptive Loop Filter
  • Fig. 2 illustrates pixels on both sides of a vertical boundary involved in Deblocking Filter.
  • FIG. 3 illustrates an exemplary flowchart of cross slice or tile loop filter processing incorporating an embodiment of the present invention.
  • the coding process in HEVC is applied to Largest Coding Units (LCU) of an image.
  • LCU Largest Coding Units
  • the LCU is adaptively partitioned into coding units using quadtree.
  • DF is performed for each 8x8 block in HEVC Test Model Version 5.0 (HM- 5.0).
  • HM- 5.0 HEVC Test Model Version 5.0
  • the DF is applied to 8x8 block boundaries.
  • horizontal filtering across vertical block boundaries is first applied, and then vertical filtering across horizontal block boundaries is applied.
  • Fig. 2 shows an example of a vertical boundary to be filtered between two blocks 210 and 220 shown as thick boxes in Fig. 2, where each block consists of 8x8 pixels.
  • Eight pixel lines associated with the two neighboring blocks are labeled from 231 through 238 as shown in Fig. 2.
  • block 210 and block 220 can be two prediction units (PUs) or transform units (TUs).
  • the deblocking process includes steps of determining filter ON/OFF, determining filter strength and applying deblocking filter.
  • the filter ON/OFF decision checks if the transition at the boundary is a natural edge or is caused by coding artifacts. If it is a natural edge, the filter is turned OFF to preserve the sharpness of the picture associated with the respective boundary. Otherwise, the deblocking filter is turned ON to reduce the artifacts.
  • the filter ON/OFF decision is first performed for all block boundaries of the respective picture area to be filtered. If a boundary is to be filtered, filter strength decision, i.e., selecting a strong or weak filter, will be determined. Subsequently, a deblocking filter with the determined filter strength is applied to the boundary to be filtered.
  • deblocking filter in the field of video coding.
  • the filter used for deblocking is also called deblocking filter. Therefore, the term deblocking filter may refer to the deblocking process or the filter used for deblocking depending on the context.
  • the filter ON/OFF decision according to conventional HEVC is determined based on only two pixel lines. For example, in HM-5.0, line 2 (corresponding to pixel line 233 in Fig. 2) and line 5 (corresponding to pixel line 236 in Fig. 2) are used. Edge activity measure, d is computed based on pixels in lines 2 and 5:
  • the Edge activity measure d is smaller than a pre-defined threshold ⁇ , the corresponding block boundary will be filtered.
  • the pre-defined threshold, ⁇ is related to quantization parameter (QP). If a block boundary is determined to be filtered, the weak/strong filter decision is then performed line by line according to the conditions: d ⁇ ⁇ » 2 ,
  • tc is another pre-defined threshold related to QP to avoid over- filtering pixels. If all three conditions above are satisfied, a strong filer is selected. Otherwise a weak filter is selected. While the above example describes DF processing for a vertical block boundary, the DF processing for a horizontal block boundary can be derived similarly.
  • the slice or tile boundary can be aligned with a block boundary.
  • the vertical block 210 is also a vertical slice/tile boundary.
  • the two sides of a block boundary are referred as a first side and a second side.
  • the boundary pixels on the first side are referred as first pixels and the boundary pixels on the second side are referred as second pixels.
  • boundary pixels on one side of the slice- or tile-aligned block boundary are within the given slice or tile, and boundary pixels on the other side of the slice- or tile-aligned block boundary are outside the given slice or tile.
  • the boundary pixels (qOf, qlf, q2f, q3f) are designated as the first pixels and the corresponding side is designated as the first side.
  • the boundary pixels (p3f, p2f, plf, pOf) are designated as the second pixels and the corresponding side is designated as the second side.
  • the first pixels may be designated as the boundary pixels being processed by DF and the first pixels are located inside a current slice or a tile. Therefore, the second pixels on the second side of the block boundary are located in a neighboring slice or tile, i.e., outside the current slice or tile.
  • both thresholds tc and ⁇ are dependent on the quantization parameter QP.
  • the QP values QPQ and QPp associated with the respective first pixels and second pixel of the block boundary may differ. Therefore, in HM-5.0, the final QP is formed by averaging QPQ and QPp and the averaged QP is used to derive tc and ⁇ thresholds.
  • QPQ and QPp are associated with two neighboring slices or tiles. If intra slice or tile processing is selected, the QP value from the other slice or tile (QPp or QPQ) will not be available.
  • each slice may be independently processed without any dependency on any other slices.
  • One flag, NonCrossSliceLoopFilterFlag is used to indicate whether the loop filter processing is allowed to apply across the slice boundary.
  • both independent and dependent tiles are supported, where each independent tile is processed without dependency on any other tiles.
  • One flag, NonCrossTileLoopFilterFlag is used to indicate whether the loop filter processing is allowed to apply across the tile boundary.
  • Embodiments according to the present invention use the parameter associated with a neighboring slice according to the cross-slice loop filter flag. For example, in the case of a slice boundary, QPp (i.e., quantization parameter for the second pixels outside the slice) is set to unavailable for loop filter processing on pixels within the slice if the non-cross-slice loop filter flag, NonCrossSliceLoopFilterFlag is true (i.e., no cross-slice loop filter processing is allowed). In this case, loop filter processing may not be applied to boundary pixels across the slice boundary.
  • QPp i.e., quantization parameter for the second pixels outside the slice
  • NonCrossSliceLoopFilterFlag is true (i.e., no cross-slice loop filter processing is allowed).
  • loop filter processing may not be applied to boundary pixels across the slice boundary.
  • NonCrossTileLoopFilterFlag is used to indicate whether the loop filter processing is applied across the tile boundary or not.
  • Embodiments according to the present invention also use the parameter associated with a neighboring tile according to the cross-tile loop filter flag. For example, in the case of an independent tile boundary for a tile, QPp (i.e., quantization parameter for the second pixels outside the tile) is set to unavailable for loop filter processing applied inside the tile if the non-cross-tile loop filter flag, NonCrossTileLoopFilterFlag is true (i.e., no cross-tile loop filter processing is allowed).
  • QPp i.e., quantization parameter for the second pixels outside the tile
  • loop filter processing may not be applied to boundary pixels across the tile boundary.
  • FIG. 3 illustrates an exemplary flowchart of an encoder or a decoder incorporating an embodiment of the present invention.
  • the reconstructed video data associated with a picture is received from a media or a processor as shown in step 310, wherein the reconstructed video data is partitioned into one or more slices or tiles, wherein a slice boundary or a tile boundary of a slice or tile is aligned with one block boundary.
  • the reconstructed video data may be retrieved from storage such as a computer memory of buffer (RAM or DRAM).
  • the reconstructed video data may also be received from a processor such as a controller, a central processing unit or a digital signal processor that reconstructs the video data.
  • a processor such as a controller, a central processing unit or a digital signal processor that reconstructs the video data.
  • the value of the cross-slice loop filter flag or the cross-tile loop filter flag corresponding to the slice or the tile is determined respectively in step 320.
  • the loop filter processing is then applied to the boundary pixels of the slice or tile boundary according to the value of the cross-slice loop filter flag or the cross-tile loop filter flag as shown in step 330, wherein the loop filter processing for the boundary pixels depends on a first parameter of the first pixels inside the slice or tile and a second parameter of the second pixels outside the slice or tile.
  • Embodiment of the present invention as described above may be implemented in various hardware, software codes, or a combination of both.
  • an embodiment of the present invention can be a circuit integrated into a video compression chip or program code integrated into video compression software to perform the processing described herein.
  • An embodiment of the present invention may also be program code to be executed on a Digital Signal Processor (DSP) to perform the processing described herein.
  • DSP Digital Signal Processor
  • the invention may also involve a number of functions to be performed by a computer processor, a digital signal processor, a microprocessor, or field programmable gate array (FPGA). These processors can be configured to perform particular tasks according to the invention, by executing machine-readable software code or firmware code that defines the particular methods embodied by the invention.
  • the software code or firmware code may be developed in different programming languages and different formats or styles.
  • the software code may also be compiled for different target platforms.
  • different code formats, styles and languages of software codes and other means of configuring code to perform the tasks in accordance with the invention will not depart from the spirit and scope of the invention.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)
  • Image Processing (AREA)

Abstract

A method and apparatus for loop filter processing of boundary pixels across a block boundary aligned with a slice or tile boundary is disclosed. Embodiments according to the present invention use a parameter of a neighboring slice or tile for loop filter processing across slice or tile boundaries according to a flag indicating whether cross slice or tile loop filter processing is allowed or not. According to one embodiment of the present invention, the parameter is a quantization parameter corresponding to a neighboring slice or tile, and the quantization parameter is used for filter decision in deblocking filter.

Description

METHOD AND APPARATUS FOR LOOP FILTERING
CROSS TILE OR SLICE BOUNDARIES
CROSS REFERENCE TO RELATED APPLICATIONS
[0001] The present invention claims priority to U.S. Provisional Patent Application, No. 61/594,457, filed on February 3, 2012, entitled "Constrain QP availability at slice or tile boundaries". The U.S. Provisional Patent Application is hereby incorporated by reference in its entirety.
TECHNICAL FIELD
[0002] The present invention relates to video coding. In particular, the present invention relates to video coding techniques associated with loop filter processing at tile or slice boundaries.
BACKGROUND
[0003] Motion estimation is an effective inter-frame coding technique to exploit temporal redundancy in video sequences. Motion-compensated inter- frame coding has been widely used in various international video coding standards. The motion estimation adopted in various coding standards is often a block-based technique, where motion information such as coding mode and motion vector is determined for each macroblock or similar block configuration. In addition, intra-coding is also adaptively applied, where the picture is processed without reference to any other picture. The inter-predicted or intra-predicted residues are usually further processed by transformation, quantization, and entropy coding to generate a compressed video bitstream. During the encoding process, coding artifacts are introduced, particularly by the quantization process. In order to alleviate the coding artifacts, in newer coding systems, additional processing is often applied to the reconstructed video to enhance picture quality. The additional processing is configured in an in-loop operation so that the encoder and decoder may derive the same reference pictures for proper operation.
[0004] Fig. 1 illustrates an exemplary adaptive inter/intra video coding system incorporating in-loop filter processing. For inter-prediction, Motion Estimation (ME)/Motion Compensation (MC) 112 is used to provide prediction data based on video data from other picture or pictures. Switch 114 selects Intra Prediction 110 or inter-prediction data from ME/MC 112 and the selected prediction data is supplied to Adder 116 to form prediction errors, also called prediction residues or residues. The prediction error is then processed by Transformation (T) 118 followed by Quantization (Q) 120. The transformed and quantized residues are then coded by Entropy Encoder 122 to form a video bitstream corresponding to the compressed video data. The bitstream associated with the transform coefficients is then packed with side information such as motion, mode, and other information associated with the image unit. The side information may also be processed by entropy coding to reduce required bandwidth. Accordingly, the side information data is also provided to Entropy Encoder 122 as shown in Fig. 1 (the motion/mode paths to Entropy Encoder 122 are not shown). When the inter-prediction mode is used, a previously reconstructed reference picture or pictures have to be used to form prediction residues. Therefore, a reconstruction loop is used to generate reconstructed pictures at the encoder end. Consequently, the transformed and quantized residues are processed by Inverse Quantization (IQ) 124 and Inverse Transformation (IT) 126 to recover the processed residues. The processed residues are then added back to prediction data 136 by Reconstruction (REC) 128 to reconstruct the video data. The reconstructed video data may be stored in Reference Picture Buffer 134 and used for prediction of other frames.
[0005] As shown in Fig. 1, incoming video data undergoes a series of processing in the encoding system. The reconstructed video data from REC 128 may be subject to various impairments due to the series of processing. Accordingly, various loop processing is applied to the reconstructed video data before the reconstructed video data is used as prediction data in order to improve video quality. In the High Efficiency Video Coding (HEVC) standard being developed, Deblocking Filter (DF) 130, Sample Adaptive Offset (SAO) 131 and Adaptive Loop Filter (ALF) 132 have been developed to enhance picture quality. The Deblocking Filter (DF) 130 is applied to boundary pixels and the DF processing is dependent on the underlying pixel data and coding information associated with the corresponding blocks. There is no DF- specific side information needs to be incorporated in the video bitstream. On the other hand, the SAO and ALF processing are adaptive, where filter information such as filter parameters and filter type may be dynamically changed according to the underlying video data. Therefore, filter information associated with SAO and ALF is incorporated in the video bitstream so that a decoder can properly recover the required information. Therefore, filter information from SAO and ALF is provided to Entropy Encoder 122 for incorporation into the bitstream. In Fig. 1, DF 130 is applied to the reconstructed video first; SAO 131 is then applied to DF-processed video; and ALF 132 is applied to SAO-processed video. However, the processing order among DF, SAO and ALF may be re-arranged. In H.264/AVC video standard, the loop filter processing only includes DF. In the High Efficiency Video Coding (HEVC) video standard being developed, the loop filter processing includes DF and SAO.
[0006] In HEVC Test Model Version 5.0 (HM-5.0), different image unit structures, including slice and tile, are introduced. A picture is partitioned into multiple slices or tiles, where the image unit partition is often aligned with boundaries of Largest Coding Units (LCUs). Within each picture, the processing sequence of the tiles may be according to raster scan order. Slices and tiles can be configured independently. Therefore, one slice may run across multiple tiles, and one tile may also run across multiple slices. Loop filter processing such as DF, SAO and ALF may depend on neighboring pixels. Accordingly, the loop filter processing for a current tile or slice may have to wait until some neighboring slices or tiles become available. There are two types of tiles: independent tiles and dependent tiles. Slices or independent tiles are mainly designed for parallel processing, where reconstructing LCUs (e.g. MV prediction, intra prediction, entropy coding) within one slice or tile does not need any data from other tiles. However, loop filtering can be applied across slice or tile boundaries or restricted to within a slice or a tile according to a flag. While in the mode that coding process is applied within the slice or independent tile, the loop filter processing may still need some information from a neighboring slice or tile. For example, when DF is applied across a block boundary, the DF process depends on the QP (quantization parameter) values from both sides of the block boundary. When DF is applied to a slice or tile boundary aligned with a block boundary, the DF process depends on the QP values on both sides of the slice or tile. This will be a problem if the coding process is intended for not crossing the slice or tile boundary. Accordingly, it is desirable to develop a method and apparatus that can overcome dependency on parameter across a slice or tile boundary.
SUMMARY
[0007] A method and apparatus for loop filter processing of boundary pixels across a block boundary are disclosed. The block boundary has first pixels on a first side and second pixels on a second side of the block boundary. The loop filter processing is applied to the boundary pixels across the block boundary depends on a first parameter for the first pixels on the first side and a second parameter for the second pixels on the second side. Embodiments according to the present invention adaptively apply loop filter processing across slice or tile boundaries according to a flag indicating whether cross slice or tile loop filter processing is allowed or not. According to one embodiment of the present invention, the method comprises receiving reconstructed video data associated with a picture; determining a value of a cross-slice loop filter flag or a cross-tile loop filter flag corresponding to the slice or the tile respectively; and applying said loop filter processing to the boundary pixels across a slice or tile boundary aligning a block boundary according to the value of the cross-slice loop filter flag or the cross-tile loop filter flag, wherein said loop filter processing for the boundary pixels of the slice or tile boundary depends on both the first parameter for the first pixels inside the slice or tile and the second parameter for the second pixels outside the slice or tile.
[0008] The first parameter and the second parameter correspond to quantization parameters (QPs) on both sides of the block boundary. For a slice or tile, if the value of the cross-slice loop filter flag or the cross-tile loop filter flag indicates that no cross-slice or cross-tile filter loop processing is allowed, the quantization parameter corresponding to a neighboring slice or tile is set unavailable for the loop filter processing. For a slice or tile, if the value of the cross-slice loop filter flag or the cross-tile loop filter flag indicates that cross-slice or cross-tile filter loop processing is allowed, the quantization parameter corresponding to a neighboring slice or tile is provided for the loop filter processing. In this case, the loop filter processing is applied to the boundary pixels of the slice or tile using an average quantization parameter of the first quantization parameter and the second quantization parameter. BRIEF DESCRIPTION OF DRAWINGS
[0009] Fig. 1 illustrates an exemplary video coding system using Inter/Intra prediction, where loop filter processing including Deblocking Filter (DF), Sample Adaptive Offset (SAO) and Adaptive Loop Filter (ALF) is incorporated.
[0010] Fig. 2 illustrates pixels on both sides of a vertical boundary involved in Deblocking Filter.
[0011] Fig. 3 illustrates an exemplary flowchart of cross slice or tile loop filter processing incorporating an embodiment of the present invention.
DETAILED DESCRIPTION
[0012] The coding process in HEVC is applied to Largest Coding Units (LCU) of an image. The LCU is adaptively partitioned into coding units using quadtree. In each leaf CU, DF is performed for each 8x8 block in HEVC Test Model Version 5.0 (HM- 5.0). The DF is applied to 8x8 block boundaries. For each 8x8 block, horizontal filtering across vertical block boundaries is first applied, and then vertical filtering across horizontal block boundaries is applied.
[0013] Fig. 2 shows an example of a vertical boundary to be filtered between two blocks 210 and 220 shown as thick boxes in Fig. 2, where each block consists of 8x8 pixels. Eight pixel lines associated with the two neighboring blocks are labeled from 231 through 238 as shown in Fig. 2. Four pixels on each side of the vertical boundary are labeled as (p3i, p2i, pli, pOi, qOi, qli, q2i, q3i), where i is the index for the pixel lines and i=0,...,7. In this example, block 210 and block 220 can be two prediction units (PUs) or transform units (TUs). The deblocking process includes steps of determining filter ON/OFF, determining filter strength and applying deblocking filter. The filter ON/OFF decision checks if the transition at the boundary is a natural edge or is caused by coding artifacts. If it is a natural edge, the filter is turned OFF to preserve the sharpness of the picture associated with the respective boundary. Otherwise, the deblocking filter is turned ON to reduce the artifacts. The filter ON/OFF decision is first performed for all block boundaries of the respective picture area to be filtered. If a boundary is to be filtered, filter strength decision, i.e., selecting a strong or weak filter, will be determined. Subsequently, a deblocking filter with the determined filter strength is applied to the boundary to be filtered. The deblocking process is termed as deblocking for convenience in this disclosure. The deblocking process is also called deblocking filter in the field of video coding. The filter used for deblocking is also called deblocking filter. Therefore, the term deblocking filter may refer to the deblocking process or the filter used for deblocking depending on the context.
[0014] In order to keep the computational complexity low, the filter ON/OFF decision according to conventional HEVC is determined based on only two pixel lines. For example, in HM-5.0, line 2 (corresponding to pixel line 233 in Fig. 2) and line 5 (corresponding to pixel line 236 in Fig. 2) are used. Edge activity measure, d is computed based on pixels in lines 2 and 5:
d= I p -ip 1 +112! -¾¾ + ι I +1 P ~2P +p 1 +1 -¾¾ +Φ51.
[0015] If the Edge activity measure d is smaller than a pre-defined threshold β, the corresponding block boundary will be filtered. The pre-defined threshold, β is related to quantization parameter (QP). If a block boundary is determined to be filtered, the weak/strong filter decision is then performed line by line according to the conditions: d < β » 2 ,
\ p\ -pQl \ + \ qhl -qQl \< P » , aaA
I p0t - q0t |< 5 · tc + 1 ,
where tc is another pre-defined threshold related to QP to avoid over- filtering pixels. If all three conditions above are satisfied, a strong filer is selected. Otherwise a weak filter is selected. While the above example describes DF processing for a vertical block boundary, the DF processing for a horizontal block boundary can be derived similarly.
[0016] The slice or tile boundary can be aligned with a block boundary. In this case, the vertical block 210 is also a vertical slice/tile boundary. The two sides of a block boundary are referred as a first side and a second side. The boundary pixels on the first side are referred as first pixels and the boundary pixels on the second side are referred as second pixels. For a given slice or tile, boundary pixels on one side of the slice- or tile-aligned block boundary are within the given slice or tile, and boundary pixels on the other side of the slice- or tile-aligned block boundary are outside the given slice or tile. For convenience, the boundary pixels (qOf, qlf, q2f, q3f) are designated as the first pixels and the corresponding side is designated as the first side.
Similarly, the boundary pixels (p3f, p2f, plf, pOf) are designated as the second pixels and the corresponding side is designated as the second side. Further, the first pixels may be designated as the boundary pixels being processed by DF and the first pixels are located inside a current slice or a tile. Therefore, the second pixels on the second side of the block boundary are located in a neighboring slice or tile, i.e., outside the current slice or tile.
[0017] As mentioned above, both thresholds tc and β are dependent on the quantization parameter QP. The QP values QPQ and QPp associated with the respective first pixels and second pixel of the block boundary may differ. Therefore, in HM-5.0, the final QP is formed by averaging QPQ and QPp and the averaged QP is used to derive tc and β thresholds. At a slice or tile boundary, QPQ and QPp are associated with two neighboring slices or tiles. If intra slice or tile processing is selected, the QP value from the other slice or tile (QPp or QPQ) will not be available.
[0018] In HM-5.0, each slice may be independently processed without any dependency on any other slices. One flag, NonCrossSliceLoopFilterFlag, is used to indicate whether the loop filter processing is allowed to apply across the slice boundary. Similarly, both independent and dependent tiles are supported, where each independent tile is processed without dependency on any other tiles. One flag, NonCrossTileLoopFilterFlag, is used to indicate whether the loop filter processing is allowed to apply across the tile boundary. When non-cross slice or tile operations are specified, the usage of data, including the quantization parameter QP from other slice or tile is prohibited. Therefore, the QP of pixels outside the slice or tile is unavailable for deriving the average QP for DF processing at a slice or tile boundary.
[0019] Embodiments according to the present invention use the parameter associated with a neighboring slice according to the cross-slice loop filter flag. For example, in the case of a slice boundary, QPp (i.e., quantization parameter for the second pixels outside the slice) is set to unavailable for loop filter processing on pixels within the slice if the non-cross-slice loop filter flag, NonCrossSliceLoopFilterFlag is true (i.e., no cross-slice loop filter processing is allowed). In this case, loop filter processing may not be applied to boundary pixels across the slice boundary. The quantization parameter QPQ for the first pixels on the first side of the block boundary (inside the slice) may be used as QP for loop filter processing applied to the first pixels, i.e., QP= QPQ. If the non-cross-slice loop filter flag, NonCrossSliceLoopFilterFlag is false (i.e., cross-slice loop filter processing is allowed), QPp is set to available. In this case, the average of the first quantization parameter and the second quantization parameter may be used as the quantization parameter for loop filter processing on boundary pixels of the slice boundary, i.e., QP=1/2(QPP + QPQ).
[0020] For tiles, there are independent tiles and dependent tiles. For independent tiles, a loop filter flag, NonCrossTileLoopFilterFlag is used to indicate whether the loop filter processing is applied across the tile boundary or not. Embodiments according to the present invention also use the parameter associated with a neighboring tile according to the cross-tile loop filter flag. For example, in the case of an independent tile boundary for a tile, QPp (i.e., quantization parameter for the second pixels outside the tile) is set to unavailable for loop filter processing applied inside the tile if the non-cross-tile loop filter flag, NonCrossTileLoopFilterFlag is true (i.e., no cross-tile loop filter processing is allowed). In this case, loop filter processing may not be applied to boundary pixels across the tile boundary. The quantization parameter QPQ for the first pixels inside the tile may be used as QP for loop filter processing applied to the first pixels, i.e., QP= QPQ. If the non-cross-tile loop filter flag, NonCrossTileLoopFilterFlag is false (i.e., cross-tile loop filter processing is allowed), QPp is set to available. In this case, the average of the first quantization parameter and the second quantization parameter may be used as the quantization parameter for loop filter processing, i.e., QP=l/2(QPp + QPQ).
[0021] When dependent tiles are used, information from a previous tile is allowed for loop filter processing of the current tile. Therefore, QPp is always set to available regardless of whether NonCrossTileLoopFilterFlag is true or false.
[0022] The loop filter processing method which uses the quantization parameter associated with a neighboring slice or tile according to the cross-slice or cross-tile loop filter flag described above can be used in a video encoder as well as a video decoder. Fig. 3 illustrates an exemplary flowchart of an encoder or a decoder incorporating an embodiment of the present invention. The reconstructed video data associated with a picture is received from a media or a processor as shown in step 310, wherein the reconstructed video data is partitioned into one or more slices or tiles, wherein a slice boundary or a tile boundary of a slice or tile is aligned with one block boundary. The reconstructed video data may be retrieved from storage such as a computer memory of buffer (RAM or DRAM). The reconstructed video data may also be received from a processor such as a controller, a central processing unit or a digital signal processor that reconstructs the video data. The value of the cross-slice loop filter flag or the cross-tile loop filter flag corresponding to the slice or the tile is determined respectively in step 320. The loop filter processing is then applied to the boundary pixels of the slice or tile boundary according to the value of the cross-slice loop filter flag or the cross-tile loop filter flag as shown in step 330, wherein the loop filter processing for the boundary pixels depends on a first parameter of the first pixels inside the slice or tile and a second parameter of the second pixels outside the slice or tile.
[0023] The flowchart shown above is intended to illustrate an example of a loop filtering processing method at slice or tile boundaries for a video encoder and a decoder incorporating an embodiment of the present invention. A person skilled in the art may modify each step, re-arranges the steps, split a step, or combine the steps to practice the present invention without departing from the spirit of the present invention.
[0024] The above description is presented to enable a person of ordinary skill in the art to practice the present invention as provided in the context of a particular application and its requirement. Various modifications to the described embodiments will be apparent to those with skill in the art, and the general principles defined herein may be applied to other embodiments. Therefore, the present invention is not intended to be limited to the particular embodiments shown and described, but is to be accorded the widest scope consistent with the principles and novel features herein disclosed. In the above detailed description, various specific details are illustrated in order to provide a thorough understanding of the present invention. Nevertheless, it will be understood by those skilled in the art that the present invention may be practiced.
[0025] Embodiment of the present invention as described above may be implemented in various hardware, software codes, or a combination of both. For example, an embodiment of the present invention can be a circuit integrated into a video compression chip or program code integrated into video compression software to perform the processing described herein. An embodiment of the present invention may also be program code to be executed on a Digital Signal Processor (DSP) to perform the processing described herein. The invention may also involve a number of functions to be performed by a computer processor, a digital signal processor, a microprocessor, or field programmable gate array (FPGA). These processors can be configured to perform particular tasks according to the invention, by executing machine-readable software code or firmware code that defines the particular methods embodied by the invention. The software code or firmware code may be developed in different programming languages and different formats or styles. The software code may also be compiled for different target platforms. However, different code formats, styles and languages of software codes and other means of configuring code to perform the tasks in accordance with the invention will not depart from the spirit and scope of the invention.
[0026] The invention may be embodied in other specific forms without departing from its spirit or essential characteristics. The described examples are to be considered in all respects only as illustrative and not restrictive. The scope of the invention is therefore, indicated by the appended claims rather than by the foregoing description. All changes which come within the meaning and range of equivalency of the claims are to be embraced within their scope.

Claims

1. A method of loop filter processing for boundary pixels across a block boundary, wherein the boundary pixels comprises first pixels on a first side of the block boundary and second pixels on a second side of the block boundary, the method comprising:
receiving reconstructed video data associated with a picture from a media or a processor, wherein the reconstructed video data is partitioned into slices or tiles, and a slice boundary or a tile boundary of a slice or tile is aligned with one block boundary; determining a value of a cross-slice loop filter flag or a cross-tile loop filter flag corresponding to the slice or tile respectively; and
applying said loop filter processing to the boundary pixels across the slice boundary or tile boundary according to the value of the cross-slice loop filter flag or the cross-tile loop filter flag, wherein said loop filter processing depends on a first parameter for the first pixels inside the slice or tile and a second parameter for the second pixels outside the slice or tile.
2. The method of Claim 1, wherein the first parameter for the first pixels corresponds to first quantization parameter (QP) for the first pixels and the second parameter for the second pixels corresponds to second quantization parameter (QP) for the second pixels.
3. The method of Claim 1, wherein an average of the first parameter and the second parameter is provided for said applying said loop filter processing to the boundary pixels if the cross-slice loop filter flag or the cross-tile loop filter flag indicates that said loop filter processing is applied across the slice or tile respectively.
4. The method of Claim 1, wherein the cross-slice loop filter flag or the cross- tile loop filter flag is incorporated in a picture level of compressed video bitstream.
5. The method of Claim 1, wherein said loop filter processing corresponds to deblocking filter, and filter decision depends on the first parameter and the second parameter.
6. An apparatus of loop filter processing for boundary pixels across a block boundary, wherein the boundary pixels comprises first pixels on a first side of the block boundary and second pixels on a second side of the block boundary, the apparatus comprising:
means for receiving reconstructed video data associated with a picture from a media or a processor, wherein the reconstructed video data is partitioned into slices or tiles, wherein a slice boundary or a tile boundary of a slice or tile is aligned with one block boundary;
means for determining a value of a cross-slice loop filter flag or a cross-tile loop filter flag corresponding to the slice or tile respectively; and
means for applying said loop filter processing to the boundary pixels across the slice boundary or tile boundary according to the value of the cross-slice loop filter flag or the cross-tile loop filter flag, wherein said loop filter processing depends on a first parameter for the first pixels inside the slice or tile and a second parameter for the second pixels outside the slice or tile.
7. The apparatus of Claim 6, wherein the first parameter for the first pixels corresponds to first quantization parameter (QP) for the first pixels and the second parameter for the second pixels corresponds to second quantization parameter (QP) for the second pixels.
8. The apparatus of Claim 6, wherein an average of the first parameter and the second parameter is provided for said applying said loop filter processing to the boundary pixels if the cross-slice loop filter flag or the cross-tile loop filter flag indicates that said loop filter processing is applied across the slice or tile respectively.
9. The apparatus of Claim 6, wherein the cross-slice loop filter flag or the cross- tile loop filter flag is incorporated in a picture level of compressed video bitstream.
10. The apparatus of Claim 6, wherein said loop filter processing corresponds to deblocking filter, and filter decision depends on the first parameter and the second parameter.
PCT/CN2013/071108 2012-02-03 2013-01-30 Method and apparatus for loop filtering cross tile or slice boundaries WO2013113274A1 (en)

Priority Applications (5)

Application Number Priority Date Filing Date Title
NZ616358A NZ616358A (en) 2012-02-03 2013-01-30 Method and apparatus for loop filtering cross tile or slice boundaries
US14/374,918 US9967563B2 (en) 2012-02-03 2013-01-30 Method and apparatus for loop filtering cross tile or slice boundaries
BR112013027865-0A BR112013027865B1 (en) 2012-02-03 2013-01-30 METHOD AND APPARATUS OF LOOP FILTER PROCESSING FOR BORDER PIXELS ACROSS A BLOCK BORDER
EP13743612.7A EP2735162A4 (en) 2012-02-03 2013-01-30 Method and apparatus for loop filtering cross tile or slice boundaries
CN201380001869.1A CN103636227B (en) 2012-02-03 2013-01-30 Through segment or the loop circuit filtering method and its device of band boundaries

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US201261594457P 2012-02-03 2012-02-03
US61/594,457 2012-02-03

Publications (1)

Publication Number Publication Date
WO2013113274A1 true WO2013113274A1 (en) 2013-08-08

Family

ID=48904413

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2013/071108 WO2013113274A1 (en) 2012-02-03 2013-01-30 Method and apparatus for loop filtering cross tile or slice boundaries

Country Status (6)

Country Link
US (1) US9967563B2 (en)
EP (1) EP2735162A4 (en)
CN (2) CN107197259B (en)
BR (1) BR112013027865B1 (en)
NZ (1) NZ616358A (en)
WO (1) WO2013113274A1 (en)

Families Citing this family (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9060174B2 (en) * 2010-12-28 2015-06-16 Fish Dive, Inc. Method and system for selectively breaking prediction in video coding
US9749627B2 (en) 2013-04-08 2017-08-29 Microsoft Technology Licensing, Llc Control data for motion-constrained tile set
JP6150011B2 (en) * 2013-07-15 2017-06-21 ソニー株式会社 Extension of motion constrained tileset SEI messages for interactivity
KR102299573B1 (en) * 2014-10-22 2021-09-07 삼성전자주식회사 Application processor for performing real time in-loop filtering, method thereof, and system including the same
US9998745B2 (en) * 2015-10-29 2018-06-12 Microsoft Technology Licensing, Llc Transforming video bit streams for parallel processing
WO2017127838A1 (en) 2016-01-22 2017-07-27 Nuvasive, Inc. Systems and methods for facilitating spine surgery
US20180054613A1 (en) * 2016-08-22 2018-02-22 Mediatek Inc. Video encoding method and apparatus with in-loop filtering process not applied to reconstructed blocks located at image content discontinuity edge and associated video decoding method and apparatus
US10999602B2 (en) 2016-12-23 2021-05-04 Apple Inc. Sphere projected motion estimation/compensation and mode decision
US11259046B2 (en) 2017-02-15 2022-02-22 Apple Inc. Processing of equirectangular object data to compensate for distortion by spherical projections
US10924747B2 (en) 2017-02-27 2021-02-16 Apple Inc. Video coding techniques for multi-view video
US11093752B2 (en) 2017-06-02 2021-08-17 Apple Inc. Object tracking in multi-view video
US10754242B2 (en) 2017-06-30 2020-08-25 Apple Inc. Adaptive resolution and projection format in multi-direction video
US20190005709A1 (en) * 2017-06-30 2019-01-03 Apple Inc. Techniques for Correction of Visual Artifacts in Multi-View Images
WO2019026721A1 (en) * 2017-08-01 2019-02-07 Sharp Kabushiki Kaisha Systems and methods for filtering reconstructed video data using adaptive loop filtering techniques
SG11202109980WA (en) * 2019-03-11 2021-10-28 Huawei Tech Co Ltd Sub-picture level filtering in video coding
WO2021027773A1 (en) 2019-08-10 2021-02-18 Beijing Bytedance Network Technology Co., Ltd. Subpicture size definition in video processing
BR112022006387A2 (en) 2019-10-02 2022-07-26 Beijing Bytedance Network Tech Co Ltd VIDEO PROCESSING METHOD AND EQUIPMENT, AND COMPUTER READable MEDIUM
EP4032290A4 (en) 2019-10-18 2022-11-30 Beijing Bytedance Network Technology Co., Ltd. Syntax constraints in parameter set signaling of subpictures
WO2021196035A1 (en) * 2020-03-31 2021-10-07 深圳市大疆创新科技有限公司 Video coding method and apparatus

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2011013580A1 (en) * 2009-07-31 2011-02-03 ソニー株式会社 Image processing apparatus and method
CN102165780A (en) * 2008-09-25 2011-08-24 联发科技股份有限公司 Adaptive filter
WO2011140960A1 (en) * 2010-05-10 2011-11-17 Mediatek Inc. Method and apparatus for adaptive loop filtering

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7227901B2 (en) * 2002-11-21 2007-06-05 Ub Video Inc. Low-complexity deblocking filter
US7460596B2 (en) * 2004-04-29 2008-12-02 Mediatek Incorporation Adaptive de-blocking filtering apparatus and method for MPEG video decoder
US20090316793A1 (en) * 2008-06-20 2009-12-24 Yang Zhijie Michael Method and system for adaptive deblocking for avs1-p2
CN101321290B (en) * 2008-07-17 2010-12-15 北京数码视讯科技股份有限公司 Block-removing filtering method based on digital signal processor
CN101883276B (en) * 2009-05-06 2012-11-21 中国科学院微电子研究所 Multi-format HD video decoder structure capable of decoding by combining software and hardware for decoding
CN102223538A (en) * 2011-06-17 2011-10-19 中兴通讯股份有限公司 Parallel filtering method and device

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102165780A (en) * 2008-09-25 2011-08-24 联发科技股份有限公司 Adaptive filter
WO2011013580A1 (en) * 2009-07-31 2011-02-03 ソニー株式会社 Image processing apparatus and method
WO2011140960A1 (en) * 2010-05-10 2011-11-17 Mediatek Inc. Method and apparatus for adaptive loop filtering

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See also references of EP2735162A4 *

Also Published As

Publication number Publication date
US9967563B2 (en) 2018-05-08
BR112013027865A2 (en) 2017-01-03
CN103636227A (en) 2014-03-12
BR112013027865A8 (en) 2017-07-11
CN107197259A (en) 2017-09-22
US20150010091A1 (en) 2015-01-08
CN103636227B (en) 2017-07-07
CN107197259B (en) 2019-11-19
EP2735162A4 (en) 2015-03-18
BR112013027865B1 (en) 2023-03-21
EP2735162A1 (en) 2014-05-28
NZ616358A (en) 2015-03-27

Similar Documents

Publication Publication Date Title
US9967563B2 (en) Method and apparatus for loop filtering cross tile or slice boundaries
US11297352B2 (en) Method of deblocking for intra block copy in video coding
EP2708027B1 (en) Method and apparatus for reduction of in-loop filter buffer
AU2012327672B2 (en) Method and apparatus for non-cross-tile loop filtering
US10003798B2 (en) Method and apparatus for reduction of deblocking filter
EP2697973B1 (en) Method and apparatus for loop filtering across slice or tile boundaries
US11303900B2 (en) Method and apparatus for motion boundary processing
US9872015B2 (en) Method and apparatus for improved in-loop filtering
KR101752612B1 (en) Method of sample adaptive offset processing for video coding
US8913656B2 (en) Method and apparatus for in-loop filtering
US9729879B2 (en) Method and apparatus of deblocking filter
WO2018068744A1 (en) Method and apparatus of smoothing filter for ringing artefact removal
US9832461B2 (en) Method and apparatus of deblocking filter with simplified boundary strength decision

Legal Events

Date Code Title Description
REEP Request for entry into the european phase

Ref document number: 2013743612

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 2013743612

Country of ref document: EP

121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 13743612

Country of ref document: EP

Kind code of ref document: A1

REG Reference to national code

Ref country code: BR

Ref legal event code: B01A

Ref document number: 112013027865

Country of ref document: BR

WWE Wipo information: entry into national phase

Ref document number: 14374918

Country of ref document: US

NENP Non-entry into the national phase

Ref country code: DE

ENP Entry into the national phase

Ref document number: 112013027865

Country of ref document: BR

Kind code of ref document: A2

Effective date: 20131029