US20120251012A1 - Image filter, encoding device, decoding device, and data structure - Google Patents

Image filter, encoding device, decoding device, and data structure Download PDF

Info

Publication number
US20120251012A1
US20120251012A1 US13/515,957 US201013515957A US2012251012A1 US 20120251012 A1 US20120251012 A1 US 20120251012A1 US 201013515957 A US201013515957 A US 201013515957A US 2012251012 A1 US2012251012 A1 US 2012251012A1
Authority
US
United States
Prior art keywords
filter
image
coefficient
filter coefficient
control variable
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US13/515,957
Inventor
Tomohiro Ikai
Yukinobu Yasugi
Tomoko Aono
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Sharp Corp
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Assigned to SHARP KABUSHIKI KAISHA reassignment SHARP KABUSHIKI KAISHA ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: AONO, TOMOKO, IKAI, TOMOHIRO, YASUGI, YUKINOBU
Publication of US20120251012A1 publication Critical patent/US20120251012A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • H04N19/61Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/20Image enhancement or restoration by the use of local operators
    • G06T5/70
    • G06T5/73
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/117Filters, e.g. for pre-processing or post-processing
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/44Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/80Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/80Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation
    • H04N19/82Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation involving filtering within a prediction loop
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/85Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression
    • H04N19/86Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression involving reduction of coding artifacts, e.g. of blockiness
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20004Adaptive image processing
    • G06T2207/20012Locally adaptive
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20172Image enhancement details
    • G06T2207/20192Edge enhancement; Edge preservation

Definitions

  • the present inventions relates to an image filter for filtering an image. Moreover, the present invention relates to an encoding device and a decoding device, each of which is provided with the image filter. Furthermore, the present invention relates to a data structure of an encoded data generated by the encoding device and referred to by the decoding device.
  • a moving image encoding device is used to efficiently transmit or record a moving image.
  • moving image encoding techniques H.264/MPEG-4 AVC and one adopted in KTA software may be exemplified as specific examples.
  • pictures constituting the moving image are managed in the form of a hierarchic structure constituted by slices (obtained by diving a picture), macro blocks (obtained by dividing a slice), and blocks (obtained by dividing the macro blocks).
  • a moving image is encoded on block basis.
  • Non-Patent Literature 1 discloses an ALF (Adaptive Loop Filter) for filtering a decoded image.
  • ALF performs filtering in which slices in a decoded image are unit areas. For respective unit areas, filter coefficients to minimize an error between an original image and an ALF-treated decoded image are determined. The filtering of ALF is performed on unit area basis with the filter coefficients.
  • the filtering with ALF is performed with a invariable filter coefficient for each respective unit area.
  • ALF invariable filter coefficient
  • An object of the present invention is to realize an image filter capable of performing more appropriate filtering even for a unit area whose feature is not uniform, by use of filter coefficients predetermined for respective unit areas.
  • an image filter for calculating a pixel value of target pixel in an output image from a pixel value of each of a pixel or pixels belonging to a target area in an input image by using a filter coefficient vector, including: filter coefficient vector changing means for changing the filter coefficient vector according to at least either where the target area is in the input image, or where the target pixel is in the output image.
  • the filter coefficient vector changing means for changing the filter coefficient vector according to at least either where the target area is in the input image, or where the target pixel is in the output image is provided, it is possible to perform filtering more appropriately adaptive for every target area according to at least either where the target area is in the input image, or where the target pixel is in the output image.
  • an image filter is an image filter for calculating a pixel value of target pixel in an output image from a pixel value of each of a pixel or pixels belonging to a target area in an input image by using a filter coefficient vector, including: filter coefficient vector changing means for changing the filter coefficient vector according to at least either where the target area is in the input image, or where the target pixel is in the output image. Therefore, it is possible to perform filtering more appropriately adaptive for every target area according to at least either where the target area is in the input image, or where the target pixel is in the output image.
  • FIG. 1 is a block diagram illustrating a configuration of an adaptive filter according to one embodiment.
  • FIG. 2 is a view for explaining an effect of the adaptive filter according to the present embodiment, and illustrates a unit area including a plurality of target areas.
  • FIG. 3 is a flowchart illustrating one example of a flow of a filter coefficient determining process for the adaptive filter of the present embodiment.
  • FIG. 4 is a flowchart illustrating another one example of a flow of a filter coefficient determining process for the adaptive filter of the present embodiment.
  • FIG. 5 is a block diagram illustrating a moving image encoding device including the adaptive filter according to the embodiment.
  • FIG. 6 is a view illustrating configurations of filter coefficient information to be outputted from the adaptive filter according the embodiment.
  • ( a ) of FIG. 6 illustrates a first example of filter coefficient information.
  • ( b ) of FIG. 6 illustrates a second example of filter coefficient information.
  • ( c ) of FIG. 6 illustrates a third example of filter coefficient information.
  • FIG. 7 is a block diagram illustrating an inter-prediction image generating section provided to the moving image encoding device according to the present embodiment, the inter-prediction image generating section being provided with the adaptive filter according to the embodiment.
  • FIG. 8 is a block diagram illustrating a moving image decoding device provided with the adaptive filter according to the embodiment.
  • FIG. 9 is a block diagram illustrating an inter-prediction image generating section provided to the moving image decoding device according to the present embodiment, the inter-prediction image generating section being provided with the adaptive filter according to the embodiment.
  • FIG. 10 is a view illustrating a bit stream of encoded data generated by the moving image encoding device according to the embodiment, so as to be inputted to the moving image decoding device according to the embodiment.
  • FIG. 11 is a view more specifically illustrating a configuration of the filter coefficient information included in the encoded data generated by the moving image encoding device according to the embodiment, so as to be inputted to the moving image decoding device according to the embodiment.
  • a configuration of the adaptive filter 100 according to one embodiment is explained, referring to FIG. 1 .
  • FIG. 1 is a block diagram illustrating a configuration of the adaptive filter 100 .
  • the adaptive filter 100 includes a filter section 110 and a control variable determining section 120 .
  • the adaptive filter 100 is a image filter capable for filtering image data.
  • the adaptive filter 100 can perform filtering for unit areas in an image of image data on unit area basis by taking the unit area as a target of the filtering, wherein the adaptive filter 100 performs the filtering adjusted adaptively for respective areas (filtering target area T) obtained by dividing the unit area.
  • the adaptive filter 100 is suitable applicable for a moving image encoding device for encoding moving image data, or a moving image decoding device for decoding such encoded moving image data, for example.
  • the adaptive filter 100 generates and outputs output image data # 110 a by obtaining weighted linear sum, based on a filter coefficient, of pixel values contained in a certain area, among pixel values in inputted image data # 100 .
  • the adaptive filter 100 receives teacher data # 200 .
  • the teacher data # 200 is image data indicative of teacher image, which is referred to when the filter coefficient is determined as described later.
  • One specific example of the teacher data # 200 is input image data inputted in the image encoding device provided with the adaptive filter 100 .
  • the adaptive filter 100 outputs filter coefficient information # 110 b together with the output image data # 110 a .
  • the filter coefficient information # 110 b is information regarding a filter coefficient used for filtering.
  • the adaptive filter 100 obtains the weighted linear sum based on the filter coefficient that is adjusted per area according to the image data # 100 and to side information # 100 a of the image data # 100 .
  • the side information # 100 a of the image data # 100 is information obtained from encoded data necessary for generating the image data # 100 . That is, side information # 100 a of the image data # 100 includes all of transformation coefficients necessary for generating prediction residual, a size of frequency transformation, information of quantum parameter, information necessary for generating predicted image (i.e., prediction mode, a size of motion compensation, motion vector, information on reference frame etc., weight information for weighted prediction, information on block and positions of block boarders).
  • the filter coefficient information # 110 b is information including the filter coefficient and information necessary for generating the filter coefficient. Specific examples of the filter coefficient information # 110 b will be given later.
  • the filter target area T has a rectangular shape of MT ⁇ NT.
  • the filter target area T may be a single pixel, or a block or macro block.
  • the filter target area T is not limited to a rectangular shape, and may be rhombus, circular, a line segment (vertical, horizontal, or diagonal line segment), or another shape.
  • the size of the filter target area T may or may not be fixed; thus may be variable. For example, if the size of the filter target area T is variable block size variable according to transformation or motion compensation, it is appropriate that the area T is the variable block.
  • the control variable determining section 120 receives (i) the image data # 100 and (ii) the size information # 100 a of the image data # 100 , and supplies control variable # 120 a and control base # 120 b to a filter section 110 .
  • the filter section 110 calculates out output image data # 110 a for respective pixel values of the image data # 100 by performing filter coefficient-based weighted liner summation and offset value addition for pixel values included in a certain area.
  • the filter section 110 performs weight linear summation as expressed in Equation (1), thereby to find a pixel value SO (x′, y′) at coordinates (x′, y′) in the output image data # 110 a .
  • R is an area constituted with pixels with which the weighted linear summation is performed (hereinafter, this area is called a filter reference area). More specifically, R is a set of relative coordinates of pixels with which the weighted linear summation is performed.
  • R ⁇ ( ⁇ 1, ⁇ 1), ( ⁇ 1,0), ( ⁇ 1,1), (0, ⁇ 1), (0,0), (0,1), (1, ⁇ 1), (1,0), (1,1) ⁇ .
  • hoffset is the offset value to be added to the pixel value.
  • the coordinates (x′, y′) and the coordinates (x, y) may be identical with each other, or may be different from each other provided that they are one-to-one correspondence with each other.
  • the present invention is not limited as to how to associate the coordinates (x′, y′) and the coordinates (x,y) with each other. How to associate the coordinates (x′, y′) and the coordinates (x,y) with each other is determined depending on a concrete configuration of an image processing device to which the adaptive filter 100 is provided.
  • M ⁇ N tapped filtering can be defined by a filter coefficient matrix H, which is a M ⁇ N matrix whose components are filter coefficients h (i, j), and the offset hoffset.
  • the present embodiment is explained based on an example in which the filter reference area is a rectangular M ⁇ N tapped area. It should be noted that the present invention is not limited to this example, and is applicable to filter reference areas R having a rhombus shape, a circular shape, a line segment (vertical, horizontal, or diagonal line segment), or another shape.
  • the filter coefficient and the offset are expressed by vector notation in which the components of the filter coefficient matrix H and the offset hoffset are mapped into vectors of M ⁇ N+1 dimension.
  • V ( h 11 ,h 12 , . . . ,h 1N ,h 21 ,h 22 , . . . ,h 2N , . . . ,h M1 ,h M2 , . . . ,h MN ,h offset ) (3)
  • 1st to M ⁇ N-th components in the filter coefficient vector V represent filter coefficient h (i, j).
  • the last component (hereinafter, referred to as offset component) in the filter coefficient vector V represents the offset h offset.
  • the components of the filter coefficient vector V are expressed as v(K) with a 1 dimensional suffix K.
  • Equation (1) can be expressed as Equation (4):
  • SO (X′) and SI (X+K) are 1-dimensional notations corresponding to SO (x′, y′) and SI (x+i, y+j).
  • summation with respect to K covers not only the area R but also the offset component. In other words, it is assumed that the summation over the area R′ includes not only 1-dimensional summation to the filter reference area R but also summation to the offset component.
  • the 1st to M ⁇ N-th components in the filter coefficient vector may or may not be independent from each other.
  • the filter coefficient vector V has been conventionally restricted to be in vertically or horizontally bilateral symmetry or point symmetry.
  • the filter coefficient vector V can be expressed as below:
  • w 1 , w 2 , and w 3 are elements of filter coefficient vector W.
  • the 4-dimensional filter coefficient vector V are expressed by using 3-dimensional filter coefficient vector W, thereby reducing its degree of freedom by 1.
  • Expression capacity of the filter is reduced as much as the degree of freedom is reduced. Meanwhile, the reduction in degree of freedom leads to reduction in parameters necessary for expressing the filter (hereinafter, the parameters are referred to as filter parameters.
  • the parameters are referred to as filter parameters.
  • the components of the filter coefficient vector W By this, code quantity can be reduced. For example point symmetry has been widely adopted for this purpose.
  • a function for calculating out the filter coefficient vector V from the filter parameters is referred to as a filter calculating function F.
  • Equation (5) is one example of such a filter calculating function F.
  • the base is not limited to symmetric bases as described above, and may be another bases.
  • the present invention is not limited to the aforementioned explanation in which the weighted linear summation based on the filter coefficient vector V (or the filter coefficient matrix H) is discussed as one example.
  • the filtering of the adaptive filter 100 according to the present invention is not limited to the weighted linear summation and may adopt weighted non-linear summation using the component of the filter coefficient vector V.
  • the adaptive filter 100 can switch over the filter coefficient vector V according to the side information # 100 a , the pixel values included in the image data # 100 , positions of the pixels included in the image data # 100 , or functions thereof.
  • the control variable determining section 120 sets a control variable v_skip to 1, wherein the control variable v_skip is either 0 or 1 depending on whether or not a pixel at the coordinates (x, y) in the image data # 100 belongs to a macro block to which the skip mode is applied. Then, the control variable determining section 120 outputs the control variable v_skip to a filter section 110 .
  • control variable determining section 120 outputs a filter base b_skip to the filter section 110 .
  • the filter base b_skip is associated with information (hereinafter, referred to as skip information) regarding whether or not the pixel at the coordinates (x, y) in the image data # 100 belongs to the macro block to which the skip mode is applied.
  • the filter base b_skip is a base vector constituted from base components corresponding to the components of the filter coefficient vector V. That is, the filter base b_skip is a M ⁇ N+1 dimensional vector in the case of the M ⁇ N-tapped filtering.
  • the filter base b_skip may be, for example, edge-stress/edge-blurred filter base, preferably.
  • the edge-stress/edge-blurred filter base is a filter base effective to stress an edge or to blur the edge.
  • Examples of the edge-stress/edge-blurred filter base encompass a base for double differentiating filter.
  • the edge-stress/edge-blurred filter base may be abbreviated simply as an edge-stress filter base.
  • the edge-stress filter base has a blurring effect as well.
  • edge-stress filter base encompass, for the case of 3 ⁇ 3 tapped filtering, a horizontal edge stress base (0, 0, 0, ⁇ 1, 2, ⁇ 1, 0, 0,0), a vertical edge stress base (0, ⁇ 1, 0, 0, 2, 0, ⁇ 1, 0, 0), a standard Laplacian base ( ⁇ 1, ⁇ 1, ⁇ 1, ⁇ 1, 8, ⁇ 1, ⁇ 1, ⁇ 1, ⁇ 1, 0), a modified Laplacian base ( ⁇ 1, ⁇ 2, ⁇ 1, ⁇ 2, 12, ⁇ 2, ⁇ 1, ⁇ 2, ⁇ 1, 0), a first diagonal Laplacian base ( ⁇ 1, ⁇ 2, 0, ⁇ 2, 6, 0, ⁇ 1, 0, 0, 0), a second diagonal Laplacian (0, 0, ⁇ 1, 0, 6, ⁇ 2, 0, ⁇ 2, ⁇ 1, 0), and the like.
  • the control variable determining section 120 outputs, to the filter section 110 , an edge-stress filter base or its linear sum thereof as the filter base b_skip.
  • the filter section 110 determines the filter coefficient vector V, slice by slice or frame by frame.
  • the components v(k) of the filter coefficient vector V are expressed by Equation (6) below:
  • v ( K ) w ( K )+ a _skip ⁇ v _skip ⁇ b _skip( K ) (6)
  • w(K) is a component of the filter coefficient vector W (fixed coefficient vector)
  • b_skip (K) is a component of the above-mentioned filter base (fixed vector) b_skip.
  • a_skip a coefficient (fixed coefficient) associated to the control variable v_skip.
  • fixed in the fixed vector and fixed coefficient means that the value does not vary depending on the position of the pixel.
  • variable in the fixed vector and fixed coefficient means that the value varies depending on the position of the pixel.
  • Equation (6) is one example of the filter calculation function F using the control variable, and the filter parameters in the function F are the fixed coefficient vector, fixed coefficient, and control variable.
  • the filter section 110 determines fixed ones (hereinafter, referred to as fixed filter parameter) common to each area among the filter parameters.
  • the fixed ones among the filter parameter thus determined are the components w(K) of the fixed coefficient vector W and the fixed coefficient a_skip.
  • the filter section 110 can determine the fixed filter parameter by, for example, statistical approach.
  • determining the fixed filter parameter is to determine the fixed filter parameter that can minimizes a square error E from the pixel values in the teacher data # 200 , the square error E being expressed by Equation (7) below.
  • S(X) is a pixel value at coordinates (x, y) in the teacher data # 200 .
  • Sum with respect to X is sum of all the pixels included in the unit area. That is, in case where the fixed filter parameter is determined per slice, the sum with respect to X is sum for all the pixels included in a slice in question. Even though this example is configured such that the sum with respect to X is the sum for all the pixels included in the unit area, it may be so configured that the sum with respect to X is sum for some of the pixels included in the unit area.
  • the filter section 110 is used as an adaptive de-blocking filter, it is appropriate to determine the fixed filter parameter only from a processing area restricted to pixels distanced from a block border by a certain distance or a distance shorter than that.
  • an area not to be subjected to the filtering is provided inside a unit area to be subjected to the filtering, and information regarding whether to perform the filtering is encoded, it is appropriate to determine the fixed filter parameter from pixels in the area to be subjected to the filtering.
  • the filter parameter V in Equation (7) is transformed into a function of the filter parameters by using the filter calculating function F.
  • an equation thus obtained is partially differentiated with the fixed parameters, and then is set to be zero, thereby obtaining an equation generally called normal equation.
  • the normal equation is solved by a process of Gaussian elimination etc., so as to calculate out the fixed parameters.
  • S, SI, and the control variable here, v_skip
  • the introduction of the control variable v_skip can be expressed as M ⁇ N+2 tapped, like the introduction of the offset is expressed as M ⁇ N+1 tapped filter.
  • One other specific example of the method for determining the filter coefficient vector V is to determine a filter coefficient vector that minimizes an absolute value error obtained by replacing the square calculation in the square error E with absolute value calculation.
  • the present invention is not limited to the statistical approach for determining the filter coefficient vector.
  • the adaptive filter 100 may be configured such that the statistical approach is used with the control variable being constantly 0, so that the vector coefficient vector V for the case where the changing the filter coefficient vector with the control variable is not performed is determined.
  • control variable v_skip thus set is supplied to the filter section 110 .
  • the filter section 110 calculates the pixel values of the output image data # 110 a by calculating out a weighted average of the pixel values in the image data # 100 according to the components of the filter coefficient vector V using the control variable v_skip supplied from the control variable determining section 120 . More specifically, the filter section 110 calculates out a pixel value SO (X′) by applying the components of the filter coefficient vector V to Equation (4).
  • the adaptive filter 100 repeats the steps S 3 an S 4 .
  • the adaptive filter 100 can perform the filtering with different filter coefficient vectors, depending on whether or not the pixel included in the image data # 100 belongs to a macro block to which the skip mode is applied.
  • the adaptive filter 100 performs filtering having the edge-stress effect when a pixel at coordinates (x, y) in the image data # 100 belongs to a macro block to which skip mode is applied, whereas the adaptive filter 100 performs filtering having no edge-stress effect, when a pixel at coordinates (x, y) in the image data # 100 belongs to a macro block to which skip mode is not applied.
  • edge-stressing is preferable for an image in the macro block to which the skip mode is applied, compared with the macro block to which the skip mode is not applied.
  • the use of the adaptive filter 100 with the above configuration makes it possible to perform filtering in such a way that the edge stressing is applied for the macro block to which the skip mode is applied, but it is avoided to apply excessive edge stressing to the macro block to which the skip mode is not applied.
  • the use of the adaptive filter 100 with the above configuration makes it possible to perform more effective filtering, compared with the conventional configuration in which the filter coefficient is not switched over.
  • the adaptive filter 100 with the above configuration can perform optimal blurring processing by determining a_skip as a negative value.
  • the adaptive filter 100 can switch over the filter coefficient vector V by setting the value of the control variable v_skip to 1 or 0. That is, the adaptive filter 100 can switch over the filter coefficient per macro block without calculating out the filter coefficient per macro block. This makes it possible to perform effective filtering while reducing the calculation load necessary to calculate out the filter coefficient.
  • skip information assigned to one of the macro blocks may be used.
  • the image filter (adaptive filter 100 ) is an image filter for calculating a pixel value of a target pixel in an output image (output image data # 110 a ) from a pixel value of each of a pixel or pixels belonging to a target area (filter reference area R) in an input image (image data # 100 ) and from a filter coefficient vector (filter coefficient vector V), the image filter comprising: filter coefficient vector changing means (control variable determining section 120 ) for changing the filter coefficient vector according to at least either where the target area is positioned in the input image or where the target pixel is positioned in the output image.
  • the filter coefficient vector changing means (control variable determining section 120 ) for changing the filter coefficient vector according to at least either where the target area is positioned in the input image or where the target pixel is positioned in the output image is provided, it is possible to perform filtering as appropriate for individual target areas according to at least one of the position of the target area and the position of the target area in the output image.
  • the image filter (adaptive filter 100 ) is configured such that the filter coefficient vector (filter coefficient vector V) can be divided into a sum of fixed filter coefficient vectors and a variable filter coefficient vector, wherein the fixed filter coefficient vectors (filter coefficient vectors calculated out from the fixed coefficient vector W) are predetermined per unit area constituting the output image, and the variable filter coefficient vector is changed by the filter coefficient vector changing means (control variable determining section 120 ) according to at least either where the target area (filter target area T) is positioned in the input image (image data # 100 ) or where the target pixel is positioned in the output image.
  • the filter coefficient vector can be divided into the sum of the fixed filter coefficient vectors and the variable filter coefficient vector, wherein the fixed filter coefficient vectors are predetermined per unit area constituting the output image, and the filter coefficient vector changing means can change the variable filter coefficient vector according to at least either where the target area is positioned in the input image or where the target pixel is positioned in the output image.
  • the unit area may be a slice, a macro block, or a block.
  • the fixed filter coefficient vector and the variable filter coefficient vectors may be defined to include an offset component.
  • the image filter (adaptive filter 100 ) is further configured such that the filter coefficient vector can be divided into a fixed coefficient, a variable coefficient, and a fixed vector, wherein the fixed coefficient (for example, a_skip) is predetermined per unit area constituting the output image, the variable coefficient (control variable, for example, v_skip) is changed by the filter coefficient vector changing means according to either where the target area (filter target area T) is positioned in the input image (image data # 100 ) or where the target pixel is positioned in the output image, and the fixed vector (filter base, for example, b_skip is predetermined.
  • the fixed coefficient for example, a_skip
  • control variable for example, v_skip
  • variable filter coefficient vector is further configured that the variable filter coefficient vector can be divided into the fixed coefficient determined per unit area constituting the output image, the variable coefficient, and the fixed vector.
  • the fixed vector is predetermined and the filter coefficient vector changing means can change the variable coefficient according to either where the target area (filter target area T) is positioned in the input image (image data # 100 ) or where the target pixel is positioned in the output image. Therefore, with this configuration, the filtering can be performed more appropriately for the feature of the image in the target area.
  • FIG. 2 is a view illustrating a case where a Q number (Q ⁇ 2) of target areas PR 1 to PRQ are present on an unit area UR.
  • the configuration according to the present invention it is possible to perform different types filtering adaptively for the individual target areas by using a smaller number of filter coefficients, compared with a configuration in which the filter coefficient vector is calculated out for every target areas individually.
  • the 1st to 3rd, and 7th to 10th components in the filter coefficient vector V are not dependent on the control variable v_skip when b_skip is horizontal edge stress base (0, 0, 0, ⁇ 1, 2, ⁇ 1, 0, 0, 0, 0). That is, regardless of whether in the skip mode or not, the 1st to 3rd, and 7th to 10th components in the filter coefficient vector V are constant values predetermined by the statistical approach.
  • Equation (6) a filter coefficient derivation process of Equation (6) can be obtained separately for a component changing according to the control variable and for a component not changed according to the control variable.
  • the filter coefficient derivation process performed by the adaptive filter 100 is explained referring to FIG. 3 , based on an example in which b_skip is horizontal edge stress base (0, 0, 0, ⁇ 1, 2, ⁇ 1, 0, 0, 0, 0).
  • FIG. 3 is a view illustrating one example of an operation flow for the filtering process.
  • the adaptive filter 100 calculates out w′ (i) by using Equation (6) with the filter coefficient vector changing according to the control variable, and keeps w′ (i) thus calculated out.
  • w ′ ⁇ ( 5 ) w ⁇ ( 5 ) + a_skip ⁇ v_skip ⁇ 2
  • w ′ ⁇ ( 6 ) w ⁇ ( 6 ) + a_skip ⁇ v_skip ⁇ - 1
  • Step S 101 Loop 1 : for every filter target area T, processes S 102 to S 109 are performed for the whole unit area constituting the output image.
  • Step S 102 Calculate out a control variable v_skip for a filter target area T.
  • Step S 103 Loop 2 : S 104 to S 108 are performed until a loop variable i becomes 10 from 1.
  • Step S 108 End loop 2 .
  • Step S 109 Perform filtering expressed as Equation (4), by using the filter coefficient vector V thus derived.
  • the derivation of the filter coefficient vector V is carried out by branching, but not by using the summation with the component changing according to the control variable.
  • the process using the summation and the process using branching seem to be different processes, but both the processes produce the same result, and the present invention encompasses both the processes.
  • the control variable for use in the branching is not limited to the above, and may be any control variable including the following example.
  • FIG. 4 is a view illustrating another example of the operation flow for the filtering.
  • FIG. 4 illustrates a case where the calculation of the filter coefficient vector V is carried out by using the summation with the component changing according to the control variable.
  • the processes in S 103 to S 107 in FIG. 3 are replaced with the following step S 154 using Equation (6).
  • Step S 154 Calculate out the element v(i) of the filter coefficient vector from Equation (6) by using the control variable v_skip.
  • the values of the filter coefficient has been stored in a 2-dimensional table v 2 (i, v) predetermined according to the positions of the components of the filter coefficient and the values of the control variable, the values of the filter coefficient being stored in the 2-dimensional table v 2 (i, v) in such a way that the values of the filter coefficient are associated with the positions of the components of the filter coefficient and the values of the control variable, and (ii) the filtering is performed with reference to the 2-dimensional table v 2 (i, v).
  • i is a position of the component of the filter coefficient
  • v is a value of the control variable. Note that, if the number of the control variable is 1, the table is a 2-dimensional table, and if the number of the control variable is Nv, the table is a Nv+1 dimensional table.
  • Equation (6) is one example of the filter calculating function F.
  • the fixed coefficient vector wi (1 ⁇ i ⁇ Nw, where Nw is the number of dimensions for the fixed coefficient vector) the control variable vk (1 ⁇ k ⁇ Nv, where Nv is a total number of the control variables)
  • the fixed coefficient akj (k is as described above, 1 ⁇ j ⁇ Nak, where Nak is a total number of the fixed coefficients regarding a control variable Vk
  • the function F is generally expressed as in Equation (8).
  • F ( w 1 , . . . ,wNw,a 11 , . . . ,a 1 — Na 1 , . . . ,aNv 1 , . . . ,aNv — NaNv,v 1 , . . . ,vNv ) F 1( w 1 , . . . ,wNw )+ F 2( Na 1 , . . . ,aNv 1 , . . . ,aNv — NaNv,v 1 , . . . ,vNv ) (8)
  • the filter calculating function F has a structure that can be divided into the fixed filter coefficient vector in the 1st term and the variable filter coefficient vector in the 2nd term.
  • F 1 is a function of the fixed coefficient vector
  • F 2 is a function of the control variable and the fixed coefficient.
  • the control variable and fixed coefficient may be plural. If the control variable is plural, the number of the fixed coefficients may be changed according to the control variable.
  • b_wi is a Nw number of bases regarding wi, and may be a unit base, symmetric base, or the like. In the present Specification, b_wi is a unit base, but the present invention is not limited to this. Further, b_vkj is a Nak number of bases prepared for each of the control variables vk, and may be the edge-stress filter baser or the like. In case the b_wi is a symmetric base, it is preferable that b_vkj is a base having similar symmetry. Moreover, the left-hand side of Equation (9) can be divided into the fixed filter coefficient vector in the 1st term and the variable filter coefficient vector in the 2nd term as shown in its right-hand side. Moreover, the fixed filter coefficient vector can be divided into the fixed coefficient vector (element wi) and base (b_wi). The variable filter coefficient vector can be divided into the fixed coefficient (akj), the control variable (vk), and the fixed vector (b_vkj).
  • Equation (10) One concrete example where the filter is 4-dimensional filter is shown in Equation (10).
  • the number of the fixed filter coefficients is 3 (w 1 , w 2 , w 3 ), and the number of the fixed coefficients corresponding to the control variable v_skip is 2 (a_skip 1 and a_skip 2 ), and the number of the fixed coefficient corresponding to the control variable v_flat is 1 (a_flat).
  • the meaning of the control variable v_flat will be later explained.
  • the fixed filter parameter calculation is carried out by solving a normal equation obtained by zeroing an equation obtained by partially differentiating an equation obtained by transforming the filter parameter V of Equation (7) into a function of the filter parameter by using the filter calculating function F.
  • control variable determining section 120 may be configured to output to the filter section 110 a plurality of filter bases as the filter bases associated with the skip information, the plurality of filter bases being having different effects.
  • the filter section 110 determines, with reference to the teacher data # 200 , the coefficients of the filter coefficient vector V having the components (K) expressed by the following Equation (11).
  • v ( K ) w ( K )+ a _skip ⁇ v _skip ⁇ b _skip( K )+ a — dc ⁇ v — dc ⁇ b — dc ( K ) (11)
  • a_dc is the DC offset coefficient (fixed coefficient) corresponding to the DC offset base.
  • the filter section 110 determines the components w(K) of the fixed coefficient vector W, the edge-stress coefficient a_skip, and the DC offset coefficient a_dc, for example by the aforementioned statistical approach.
  • the control variable determining section 120 For a pixel belonging to a macro block in which the skip mode is applied, the control variable determining section 120 outputs the control variable v_skip of 1 to the filter section 110 . For a pixel belonging to a macro block in which the skip mode is not applied, the control variable determining section 120 outputs the control variable v_skip of 0 to the filter section 110 .
  • the filtering section 110 performs filtering based on the filter coefficient vector V using the control variable v_skip supplied from the control variable determining section 120 .
  • the adaptive filter 100 can perform edge-stressing filtering and filtering for changing luminance by the DC offset.
  • control variable determining section 120 may output to the filter section 110 another edge-stress base b_skip 1 together with a certain edge-stress base b_skip 0 , as the filter base associated with the skip information.
  • b_skip 0 is a horizontal edge-stress base (0, 0, 0, ⁇ 1, 2, ⁇ 1, 0, 0, 0, 0)
  • b_skip 1 (0, ⁇ 1, 0, 0, 2, 0, ⁇ 1, 0, 0, 0).
  • a base mainly for blurring such as a widthwise blurring base (0, 0, 0, 1, 2, 1, 0, 0, 0), a lengthwise blurring base (0, 1, 0, 0, 2, 0, 0, 1, 0, 0), or a omnidirectional blurring base (1, 1, 1, 1, 2, 1, 1, 1, 1, 0) may be used.
  • the filter section 110 determines, with reference to the teacher data # 200 , the coefficients in the filter coefficient vector V having the components v(K) expressed by the following Equation (12).
  • v ( K ) w ( K )+ a _skip0 ⁇ v _skip ⁇ b_skip0( K )+ a _skip1 ⁇ v _skip ⁇ b _skip1( K ) (12)
  • a_skip 0 is an edge-stress coefficient (fixed coefficient) corresponding to the edge-stress base b_skip 0 (K)
  • a_skip 1 is an edge-stress coefficient (fixed coefficient) corresponding to the edge-stress base b_skip 1 (K).
  • the filter section 110 determines, for example by the aforementioned approach, the components w(K) of the fixed coefficient vector, and the edge-stress coefficients a_skip 0 and a_skip 1 per slice or per frame.
  • control variable determining section 120 For a pixel belonging to a macro block to which the skip mode is applied, the control variable determining section 120 outputs the control variable v_skip whose value is 1. For a pixel belonging to a macro block to which the skip mode is not applied, the control variable determining section 120 outputs the control variable v_skip whose value is 0.
  • the filter section 110 performs the filtering based on the filter coefficient vector V using the control variable v_skip supplied from the control variable determining section 120 .
  • the adaptive filter 100 can perform the filter coefficient-based filtering having both the edge stress effect and the blurring effect for the macro block to which the skip-mode is applied.
  • the adaptive filter 100 can perform the filter coefficient-based filtering having none of the edge stress effect and the blurring effect.
  • the present invention is not limited to the above example in which the adaptive filter 100 switches over the filter coefficient vector V depending on whether or not the pixel included in the image data # 100 belongs to a macro block to which the skip mode is applied.
  • the filter coefficient is switched over based on the image data # 100 and the side information # 100 a.
  • the degree of flatness in the filter target area T is an index that is set according to how much the pixel values of the pixels included in the filter target area T are uneven. When the pixel values of the pixels included in the filter target area T are less uneven, the degree of flatness of the image in the area is higher. When the pixel values of the pixels included in the filter target area T are more uneven, the degree of flatness of the image in the area is lower. In order to determine the control variable for the filter target area T, it is not necessary to use all the pixels in the filter target area T (thereinafter, the same is applied).
  • the degree of flatness may be determined from a rectangular area MV x NV, which is smaller than the area T.
  • the control variable determining section 120 sets a value of a control variable v_flat to 1 and outputs the control variable v_flat to the filter section 110 , wherein the control variable v_flat is set to a value of 0 or 1 depending on the degree of flatness in the filter target area T.
  • control variable determining section 120 outputs, to the filter section 110 , the filter base b_flat associated with the degree of flatness in the filter target area T.
  • the filter base b_flat may be, for example, the aforementioned edge-stress filter base, 5 ⁇ 5 tap-expanded Laplacian base (1, 1, 1, 1, 1, 1, 0, 0, 0, 1, 1, 0, ⁇ 16, 0, 1, 1, 0, 0, 0, 1, 1, 1, 1, 1, 1, 0), or a linear sum (or weighted linear sum) of them.
  • the filter section 110 determines, with reference to the teacher data # 200 , the coefficients of the filter coefficient vector V having the components v(K) expressed by the following Equation (13).
  • a_flat is a coefficient (fixed coefficient) corresponding to the filter base b_flat.
  • the filter section 110 determines the components w(K) of the filter coefficient vector and the coefficient a_flat per slice or per frame by for example the aforementioned statistical approach.
  • the control variable determining section 120 calculates out the unevenness of the pixel values of the pixels in the filter target area T for every pixel included in the image data # 100 . More specifically, for example, the control variable determining section 120 calculates out dispersion of the pixel values of the pixels in the filter target area T per pixel included in the image data # 100 .
  • control variable determining section 120 refers to the pixel values of the pixels in the image data # 100 individually, and calculates out the unevenness of the pixel value of each pixel in the filter target area T. More specifically, the control variable determining section 120 refers to the pixel values of the pixels in the image data # 100 individually, and calculates out the dispersion of the pixel value of each pixel in the filter target area T.
  • the control variable determining section 120 sets the control variable v_flat to 0 for a pixel whose pixel value has dispersion equal to or greater than a predetermined threshold.
  • the control variable determining section 120 sets the control variable v_flat to 1 for a pixel whose pixel value has dispersion less than the predetermined threshold. Then, the control variable determining section 120 outputs the control variables v_flat to the filter section 110 per pixel in the image data # 100 individually.
  • control variable determining section 120 may be configured to output the control variable v_flat to the filter section 110 per filter target area T.
  • the filter section 110 performs the filtering based on the filter variable vector V using the control variable supplied from the control variable determining section 120 .
  • the adaptive filter 100 can perform filtering having a blurring effect for a filter target area T having a large degree of flatness, whereas the adaptive filter 100 can perform filtering having no blurring effect for a filter target area T having a small degree of flatness.
  • the adaptive filter 100 can perform effective filtering adaptively for slices or frame having different degrees of flatness.
  • the use of the filter base and filter coefficient corresponding to the DC offset makes it possible to perform filtering of luminance adjustment using the DC offset.
  • it may be configured such that the degree of flatness in the image in the filter target area T is judged according to how large an AC component associated with the block to which the pixel at the coordinates (x, y) in the image data # 100.
  • w′(i) in the (preliminary step) of the aforementioned ⁇ Filter Coefficient Calculation Process> may be calculated out by the following Equation, and stored.
  • w ′ ⁇ ( 5 ) w ⁇ ( 5 ) + a_flat ⁇ v_flat ⁇ 2
  • w ′ ⁇ ( 6 ) w ⁇ ( 6 ) + a_flat ⁇ v_flat ⁇ - 1
  • control variable v_skip should be read as the control variable v_flat.
  • the control variable determining section 120 the following control variables to 1 and outputs the following control variables to the filter section 110 .
  • control variable determining section 120 outputs the following filter bases to the filter section 12 .
  • b_edge_ 0 (0, ⁇ 1, 0, 0, 2, 0, 0, ⁇ 1, 0, 0)
  • b_edge_ 90 (0, 0, 0, ⁇ 1, 2, ⁇ 1, 0, 0, 0, 0)
  • b_edge_ 45 (0, 0, ⁇ 1, 0, 2, 0, ⁇ 1, 0, 0, 0)
  • b_edge_ 135 ( ⁇ 1, 0, 0, 0, 2, 0, 0, 0, ⁇ 1, 0)
  • the filter section 110 determines, with reference to the teacher data # 200 , the coefficients of the filter coefficient vectors V having the components v (K) expressed by Equation (14) below.
  • v ( K ) w ( K )+ ⁇ a _edge — n ⁇ v _edge — n ⁇ b _edge — n ( K ) (14)
  • the adaptive filer 100 can perform filtering using different filter coefficients depending on the edge angle including in the target area T. Thus, it is possible to perform effective filtering for slices or frames having various edge angles.
  • the adaptive filter 100 may be configured such that the control variable determining section 120 outputs the following filter base to the filter section 110 .
  • the filter section 110 determines, with reference to the teacher data # 200 , the coefficients of the filter coefficient vector V having the components v(K) expressed by the following Equation (15).
  • v ( K ) w ( K )+ ⁇ a _edge — n — 0 ⁇ v _edge — n ⁇ b _edge — n — 0( K )+ ⁇ a _edge — n — 1 ⁇ v _edge — n ⁇ b _edge — n — 1( K ) (15)
  • the adaptive filter 100 can perform the filtering by using different filter coefficient depending on the edge angle of edges in the target area T.
  • it is possible to perform effective filtering for slices or frames having various edge angles.
  • the adaptive filter 100 having the above configuration can perform effective filtering not only for a direction parallel to the edge, but also for a direction orthogonal to the edge.
  • one specific example of the filter coefficient calculation process performed by the adaptive filter 100 in the present operation example is as follows. For example, assuming that the filter base b_edge_n is b_edge_ 90 , w′(i) in the (preliminary step) of the aforementioned ⁇ Filter Coefficient Calculation Process> may be calculated out by the following Equation, and stored.
  • control variable v_skip should be read as the control variable v_edge_ 90 .
  • equation should be read in the similar manner in case where the filter base b_edge_n is another base.
  • Equation (14) can be easily generalized in case n is summed.
  • the filter coefficient is switched over according to an average of luminance in the image in the filter target area T in the image data # 100 is described.
  • the average of luminance in the filter target area T is an average of luminance values of pixels included in the filter target area T.
  • the luminance values of the pixels can be obtained as a function of the pixel values of the pixels.
  • the control variable determining section 120 sets a control variable v_lum to 0, 1, or 2 according to the average of luminance in the image in the filter target area T, and outputs the control variable v_lum to the filter section 110 .
  • control variable determining section 120 outputs, to the filter section 110 , the filter base b_lum associated with the average luminance of the image in the filter target area T.
  • the filter base b_lum may be, for example, the edge-stress filter base or its linear sum (or its weighted linear sum).
  • the filter section 110 determines, with reference to the teacher data # 200 , the coefficients of the filter coefficient vector V having the components v (K) expressed by Equation (16).
  • a_lum is a coefficient (fixed coefficient) corresponding to the filter base b_lum.
  • the filter section 110 determines the components w(K) of the fixed coefficient vector and the coefficient a_lum per slice or per frame, for example, by the aforementioned statistical approach.
  • the control variable determining section 120 calculates out the average of the luminance values associated with the pixels included in the area, for every filter target area T.
  • the control variable determining section 120 sets the value of the control variable v_lum to 2, for example when the average is less than 48.
  • the control variable determining section 120 sets the value of the control variable v_lum to 1, for example when the average is not less than 48 but less than 96.
  • the control variable determining section 120 sets the value of the control variable v_lum to 0, for example when the average is not less than 96.
  • the average can be in a range of 0 to 255.
  • control variable determining section 120 outputs the control variable v_lum to the filter section 110 per filter target area T.
  • the filter section 110 performs the filtering based on the filter coefficient vector V using the control variable v_lum supplied from the control variable determining section 120 for every filter target area T.
  • w′(i) in the (preliminary step) of the aforementioned ⁇ Filter Coefficient Calculation Process> may be calculated out by the following Equation, and stored.
  • w ′ ⁇ ( 5 ) w ⁇ ( 5 ) + a_lum ⁇ v_lum ⁇ 2
  • w ′ ⁇ ( 6 ) w ⁇ ( 6 ) + a_lum ⁇ v_lum ⁇ - 1
  • control variable v_skip should be read as the control variable v_lum.
  • equation should be read in the similar manner in case where the filter base b_lum is another base.
  • the filter coefficient is switched over according to colors in the images in the filter target area T in the image data # 100 is described. Especially, an example whether or not the filter coefficient is switched over according to whether or not the colors in the pixels in the filter target area T are skin colors or not.
  • the case where the colors in the pixels in the filter target area T are skin colors is for example, an average of the colors included in the filter target area T is 133 ⁇ V ⁇ 177, and 77 ⁇ U ⁇ 127, and 60 ⁇ Y ⁇ 145 according YUV component expression.
  • Y indicates a size of the brightness component
  • U is a size of a color component defined by a product of a constant and a difference between a blue component and the luminance component
  • V is a size of a color component defined by a product of a constant and a difference between a red component and the luminance component.
  • Y, U, and V are in a range of 0 to 255.
  • the control variable determining section 120 sets the control variable v_skin to 1 according to whether the pixel at coordinates (x, y) in the image data # 100 is in a skin color, wherein the control variable v_skin can be 0 or 1.
  • the control variable determining section 120 outputs the control variable v_skin to the filter section 110 .
  • the control variable determining section 120 outputs, to the filter section 110 , the filter base b_skin associated with whether or not the pixel at coordinates (x, y) in the image data # 100 is in a skin color.
  • the filter base b_skin may be, for example, the aforementioned edge-stress filter base or its linear sum (or weighted linear sum).
  • the filter section 110 determines, with reference to the teach data # 200 , the coefficients of the filter coefficient vector V having the components v(K) expressed by Equation (17).
  • v ( K ) w ( K )+ a _skin ⁇ v _skin ⁇ b _skin( K ) (17)
  • a_skin is a coefficient (fixed coefficient) corresponding to the filter base b_skin.
  • the filter section 110 determines the components v(K) of the filter coefficient vector and the coefficient a_skin per slice or per frame, for example, by the aforementioned statistical approach.
  • the control variable determining section 120 determines whether or not the pixel at coordinate (x, y) in the image data # 100 is in the skin color. More specifically, the control variable determining section 120 determines whether or not the color of the pixel at the coordinates (x, y) in the image data # 100 is within the following ranges expressed by the YUV components: 133 ⁇ V ⁇ 177, 77 ⁇ U ⁇ 127, and 60 ⁇ Y ⁇ 145.
  • control variable v_skin is set to 1, and outputs the control variable v_skin to the filter section 110 .
  • the control variable v_skin is set to 0, and outputs the control variable v_skin to the filter section 110 .
  • the filter section 110 performs the filtering per filter target area T based on the filter coefficient vector V using the control variable v_skin supplied from the control variable determining section 120 .
  • w′(i) in the (preliminary step) of the aforementioned ⁇ Filter Coefficient Calculation Process> may be calculated out by the following Equation, and stored.
  • w ′ ⁇ ( 5 ) w ⁇ ( 5 ) + a_skin ⁇ v_skin ⁇ 2
  • w ′ ⁇ ( 6 ) w ⁇ ( 6 ) + a_skin ⁇ v_skin ⁇ - 1
  • control variable v_skip should be read as the control variable v_skin.
  • equation should be read in the similar manner in case where the filter base b_skin is another base.
  • the block is, for example, a unit of a change or a unit of motion compensation in the H.264/AC standard. In general, there is tendency that a block noise easily occurs at the block border.
  • the control variable determining section 120 sets the value of the control variable v_block to 1, wherein the value of the control variable v_block may be 0 or 1 according to whether or not the filter target area T contains a block border therein. Then, the control variable determining section 120 outputs the control variable v_block to the filter section 110 .
  • the control variable determining section 120 outputs to the filter section 110 a filter base b_block associated with whether or not the filter target area T contains a block border.
  • the filter base b_block may be, for example, the edge-stress filter base or its linear sum (or weighted linear sum).
  • the filter section 110 determines, with reference to the teacher data # 200 , the filter coefficient vector V having the components v(K) expressed by the following Equation (18).
  • v ( K ) w ( K )+ a _block ⁇ v _block ⁇ b _block( K ) (18)
  • a_block is a coefficient (fixed coefficient) corresponding to the filter base b_block.
  • the filter section 110 determines the components w(K) of the fixed coefficient vector and the coefficient a_block per slice or per frame, for example, by the aforementioned statistical approach.
  • the control variable determining section 120 determines whether or not the filter target area T contains a block border. If the filter target area T contains a block border, the control variable determining section 120 sets the value of the control variable v_block to 1. If not, the control variable determining section 120 sets the value of control variable v_block to 0. Moreover, the control variable determining section 120 outputs the control variable v_block to the filter section 110 .
  • the filter section 110 For every filter target area T, the filter section 110 performs the filtering based on the filter coefficient vector V using the control variable v_block supplied from the control variable determining section 120 .
  • w′(i) in the (preliminary step) of the aforementioned ⁇ Filter Coefficient Calculation Process> may be calculated out by the following Equation, and stored.
  • w ′ ⁇ ( 5 ) w ⁇ ( 5 ) + a_block ⁇ v_block ⁇ 2
  • w ′ ⁇ ( 6 ) w ⁇ ( 6 ) + a_block ⁇ v_block ⁇ - 1
  • control variable v_skip should be read as the control variable v_block.
  • equation should be read in the similar manner in case where the filter base b_block is another base.
  • the filter coefficient is switched over according to side information associated with a block to which a pixel at coordinates (x, y) in the image data # 100 belongs is described.
  • an example in which the filter coefficient is switched over according to the block to which the pixel at coordinates (x, y) in the image data # 100 belongs is a block generated according to a unidirectional prediction or bidirectional prediction under the H.264/AC standard.
  • the unidirectional prediction under the H.264/AC standard is a case where the prediction of a block is carried out by using one frame encoded or decoded before a frame the block belongs.
  • the bidirectional prediction under the H.264/AC standard is a case where the prediction of a block is carried out by using two frames encoded or decoded before a frame the block belongs.
  • the number of frames used for the prediction is one or two in general, but may be greater than one or two.
  • the control variable determining section 120 sets the value of the control variable v_pred to 1, wherein the value of the control variable v_pred is set to 0 or 1 according to whether the block to which the pixel at coordinates (x, y) in the image data # 100 is a block generated based on the unidirectional prediction or a block generated based on the bidirectional prediction. Then, the control variable determining section 120 outputs the control variable v_pred to the filter section 110 .
  • control variable determining section 120 outputs to the filter section 110 a filter base b_pred associated with whether the block to which the pixel at coordinates (x, y) in the image data # 100 is a block generated based on the unidirectional prediction or a block generated based on the bidirectional prediction.
  • the filer base b_pred may be for example, the edge-stress filter base or its linear sum (or weighted linear sum).
  • the filter section 110 determines, with reference to the teacher data # 20 , the coefficients of the filter coefficient vector V having the components v(K) expressed by the following equation (19).
  • a_pred is a coefficient (fixed coefficient) corresponding to the filter base b_pred.
  • the filter section 110 determines the components w(K) of the fixed coefficient vector and the coefficient a_pred per slice or per frame, for example by the aforementioned statistical approach.
  • the control variable determining section 120 sets the control variable v_pred to 0 if the block to which the pixel at the coordinates (x, y) in the image data # 100 is a block generated based on the unidirectional prediction, whereas The control variable determining section 120 sets the control variable v_pred to 1 if the block to which the pixel at the coordinates (x, y) in the image data # 100 is a block generated based on the bidirectional prediction. Moreover, the control variable determining section 120 outputs the control variable v_pred to the filter section 110 .
  • the filter section 110 For every filter target area T, the filter section 110 performs the filtering based on the filter coefficient vector V using the control variable v_pred supplied from the control variable determining section 120 .
  • w′(i) in the (preliminary step) of the aforementioned ⁇ Filter Coefficient Calculation Process> may be calculated out by the following Equation, and stored.
  • w ′ ⁇ ( 5 ) w ⁇ ( 5 ) + a_pred ⁇ v_pred ⁇ 2
  • w ′ ⁇ ( 6 ) w ⁇ ( 6 ) + a_pred ⁇ v_pred ⁇ - 1
  • control variable v_skip should be read as the control variable v_pred.
  • equation should be read in the similar manner in case where the filter base b_pred is another base.
  • a block to which a pixel at coordinates (x, y) in the image data # 100 belongs is a block predicted by intra prediction under the H.264/AV standard, or a block predicted by inter prediction under the H.264/AV standard, is described below.
  • the control variable determining section 120 sets a value of a control variable v_int to 1, wherein the value of the control variable v_int is set to 0 or 1 according to whether the block to which the pixel at coordinates (x, y) in the image data # 100 belongs is a block predicted by intra prediction or a block predicted by inter prediction. Then, the control variable determining section 120 outputs the control variable v_int o the filter section 110 .
  • the control variable determining section 120 outputs to the output section 110 a filter base b_int associated whether the block to which the pixel at coordinates (x, y) in the image data # 100 belongs is a block predicted by intra prediction or a block predicted by inter prediction.
  • the filter base b_int may be, for example, the edge stress filter base or its linear sum (or weighted linear sum).
  • the filter section 110 determines, with reference to the teacher data # 200 , the coefficients of the filer coefficient vector V having the components expressed by the following Equation (20).
  • a_int is a coefficient (fixed coefficient) corresponding to the filter base b_int.
  • the filter section 110 determines the components w(K) of the fixed coefficient vector and the coefficient a_int per slice or per frame, for example by the aforementioned statistical approach.
  • control variable determining section 120 sets the value of the control variable v_int to 1 for the block, whereas if the block to which the pixel at the coordinates (x, y) in the image data # 100 belongs is one predicted by the inter prediction, the control variable determining section 120 sets the value of the control variable v_int to 0 for the block. Then the control variable determining section 120 outputs the control variable v_into to the filter section 110 .
  • the filter section 110 For every filer target area T, the filter section 110 performs the filtering based on the filter coefficient vector V using the control variable v_pred supplied from the control variable determining section 120 .
  • edges in the block predicted by the intra prediction tends to be easily blurred.
  • w′(i) in the (preliminary step) of the aforementioned ⁇ Filter Coefficient Calculation Process> may be calculated out by the following Equation, and stored.
  • w ′ ⁇ ( 5 ) w ⁇ ( 5 ) + a_int ⁇ v_int ⁇ 2
  • w ′ ⁇ ( 6 ) w ⁇ ( 6 ) + a_int ⁇ v_int ⁇ - 1
  • control variable v_skip should be read as the control variable v_int.
  • equation should be read in the similar manner in case where the filter base b_int is another base.
  • the control variable determining section 120 sets a value of a control variable v_dist to 1, wherein the value of the control variable v_dist is 0 or 1 according to a distance between a pixel at coordinates (x, y) in the image data # 100 , and a block border of a block to which the pixel belongs. Then, the control variable determining section 120 outputs the control variable v_dist to the filter section 110 .
  • control variable determining section 120 outputs to the filter section 110 , a filter base b_dist associated with the distance between the pixel at the pixel at the coordinates (x, y) in the image data # 100 , and the block border of the block to which the pixel belongs.
  • the filter section 110 determines, with reference to the teacher data # 200 , the coefficients of the filter coefficient vector V having the components v(K) expressed by the following Equation (21).
  • a_dist is a coefficient (fixed coefficient) corresponding to the filter base b_dist.
  • the filter section 110 determines the components w(K) of the fixed coefficient vector and the coefficient a_dist per slice or per frame, for example by the aforementioned statistical approach.
  • the control variable determining section 120 calculates out the distance between the pixel at the coordinates (x, y) in the image data # 100 , and the block border of the block to which the pixel belongs. If the distance is smaller than a predetermined first threshold, the control variable determining section 120 sets the control variable v_dist to 2. Ff the distance is not less than the first threshold but less than a second threshold greater than the first threshold, the control variable determining section 120 sets the control variable v_dist to 1. If the distance is not les than the second threshold, the control variable determining section 120 sets the control variable v_dist to 0. Moreover, the control variable determining section 120 outputs the control variable v_dist to the filter section 110 .
  • the filter section 110 For every filter target area T, the filter section 110 performs the filtering based on the filter coefficient vector V using the control variable v_dist supplied from the control variable determining section 120 .
  • the filter base is changed according to whether the block border is a vertical border or a horizontal border. That is, it may be configured such that if the block border is a vertical border, the filter coefficient b_dist is the horizontal edge-stress base, and if the block border is a horizontal border, the filter coefficient b_dist is a vertical edge-stress base.
  • control vertical determining section 120 may be configured to output a control variable adjusted according to a distance from a vertical block border and a control variable adjusted according to a distance from a horizontal block border, thereby being able to perform filtering having different edge-stress effect, according to the distance from the vertical block border and the distance from the horizontal block border.
  • w′(i) in the (preliminary step) of the aforementioned ⁇ Filter Coefficient Calculation Process> may be calculated out by the following Equation, and stored.
  • w ′ ⁇ ( 5 ) w ⁇ ( 5 ) + a_dist ⁇ v_dist ⁇ 2
  • w ′ ⁇ ( 6 ) w ⁇ ( 6 ) + a_dist ⁇ v_dist ⁇ - 1
  • control variable v_skip should be read as the control variable v_dist.
  • equation should be read in the similar manner in case where the filter base b_dist is another base.
  • the control variable determining section 120 sets a value of a control variable v_xy to 1, wherein the control variable v_xy is 0 or 1 according to where a pixel at coordinates (x, y) in the image data # 100 is located in a frame. Then, the control variable determining section 120 outputs the control variable v_xy to the output filter section 110 .
  • control variable determining section 120 outputs, to the filter section 110 , a filter base b_xy associated to the position where the pixel at the coordinates (x,y) in the image data # 100 is located in the frame.
  • the filter base b xy may be, for example, the edge-stress base or the DC offset base.
  • the filter section 110 determines, with reference to the teacher data # 200 , the coefficients of the filter coefficient vector V having the components v(K) expressed by the following Equation (22).
  • a_xy is a coefficient (fixed coefficient) corresponding to the base filter b_xy.
  • the filter section 110 determines the components w(K) of the fixed coefficient vector and the coefficient a_xy per slice or per frame, for example by the aforementioned statistical approach.
  • the control variable determining section 120 sets the control variable v_xy to 1 if the pixel at the coordinates (x, y) in the image data # 100 is located in an upper half of the frame, whereas the control variable determining section 120 sets the control variable v_xy to o if the pixel at the coordinates (x, y) in the image data # 100 is located in a lower half of the frame. Moreover, the control variable determining section 120 outputs the control variable v_xy to the filter section 110 .
  • the filter section 110 For every filter target area T, the filter section 110 performs filtering based on the filter coefficient vector V using the control variable v_xy supplied from the control variable determining section 120 .
  • the present invention is not limited to the above-described example in which the filter base b_xy is a DC offset base.
  • the filter base b_xy may be the edge-stress base described above.
  • w′(i) in the (preliminary step) of the aforementioned ⁇ Filter Coefficient Calculation Process> may be calculated out by the following Equation, and stored.
  • w ′ ⁇ ( 5 ) w ⁇ ( 5 ) + a_xy ⁇ v_xy ⁇ 2
  • w ′ ⁇ ( 6 ) w ⁇ ( 6 ) + a_xy ⁇ v_xy ⁇ - 1
  • control variable v_skip should be read as the control variable v_xy.
  • equation should be read in the similar manner in case where the filter base b_xy is another base.
  • the quantized parameter QP is a parameter associated with a quantizing step for example in MPEG-2 and the H.264/AVC standard, and assigned to every block individually.
  • the control variable determining section 120 sets a control variable v_qp to 1 wherein the control variable v_qp is 0 or 1 according to the size of a quantized parameter associated to a block to which a pixel at coordinates (x, y) in the image data # 100 . Then the control variable determining section 120 outputs the control variable v_qp to the filter section 110 .
  • control variable determining section 120 outputs, to the filter section 110 , a filter base b_qp associated with a size of the quantized parameter QP associated with the block to which the pixel at the coordinates (x, y) in the image data # 100 belongs.
  • the filter base b_qp may be, for example, the edge-stress base.
  • the filter section 110 determines, with reference to teacher data # 200 , the coefficients of the filter coefficient vector V having the components v(K) expressed by the following Equation (23).
  • a_qp is a coefficient (fixed coefficient) corresponding to the filter base b_qp.
  • the filter section 110 determines the components w(K) of the fixed coefficient vector and the coefficient a_qp per slice or per frame by for example the aforementioned statistical approach.
  • the control variable determining section 120 sets the control variable qp to 1 when a value of the quantized parameter QP associated with the block to which the pixel at the coordinate (x, y) in the image data # 100 is equal to or greater than the predetermined threshold, whereas the control variable determining section 120 sets the control variable qp to 0 when a value of the quantized parameter QP associated with the block to which the pixel at the coordinate (x, y) in the image data # 100 is smaller than the predetermined threshold. Moreover, the control variable determining section 120 outputs the control variable v_qp to the filter section 110 .
  • the filter section 110 For every filter target area T, the filter section 110 performs filtering based on the filter coefficient vector V using the control variable v_qp supplied from the control variable determining section 120 .
  • w′(i) in the (preliminary step) of the aforementioned ⁇ Filter Coefficient Calculation Process> may be calculated out by the following Equation, and stored.
  • w ′ ⁇ ( 5 ) w ⁇ ( 5 ) + a_qp ⁇ v_qp ⁇ 2
  • w ′ ⁇ ( 6 ) w ⁇ ( 6 ) + a_qp ⁇ v_qp ⁇ - 1
  • control variable v_skip should be read as the control variable v_qp.
  • equation should be read in the similar manner in case where the filter base b_qp is another base.
  • the DC component is, for example, a coefficient corresponding to a direct current component among the coefficient after frequency transformation to the pixel values included in the block under the H.264/AVC standard.
  • the control variable determining section 120 sets a value of a control variable v_dc′ to 1, wherein the control variable v_dc′ is 0 or 1 according to the size of the DC component associated with the block to which the pixel at coordinates (x, y) in the image data # 100 belongs.
  • control variable determining section 120 outputs, to the filter section 110 , a filter base b_dc′ associated with the size of the DC component associated with the block to which the pixel at the coordinates (x, y) in the image data # 100 belongs.
  • the filter base b_dc′ may be, for example, the aforementioned edge-stress base.
  • the filter section 110 determines, with reference to the teacher data # 200 , the coefficients of the filter coefficients vector V having the components v(K) expressed by the following Equation (24).
  • a_dc′ is a coefficient (fixed coefficient) corresponding to the filter base b_dc′.
  • the filter section 110 determines the components w(K 9 of the fixed coefficient vector, and the coefficients a_dc′ per slice or per frame, for example by the aforementioned statistical approach.
  • the control variable determining section 120 sets the value of the control variable v_dc′ to 0 when the value of the DC component associated with the block to which the pixel at the coordinates (x, y) in the image data # 100 is equal to or greater than a predetermined threshold, whereas the control variable determining section 120 sets the value of the control variable v_dc′ to 1 when the value of the DC component is less than the threshold. Moreover, the control variable determining section 120 outputs the control variable v_dc′ to the filter section 110 .
  • the filter section 110 For every filter target area T, the filter section 110 performs filtering based on the filter coefficient vector V using the control variable v_dc′ supplied from the control variable determining section 120 .
  • the filter coefficient is switched over according to the value of the DC parameter as described above, it is also possible to perform the filtering having the luminance change effect due to the DC offset, by using the filter base and the filter coefficient, each of which corresponds to the DC offset.
  • w′(i) in the (preliminary step) of the aforementioned ⁇ Filter Coefficient Calculation Process > may be calculated out by the following Equation, and stored.
  • w ′ ⁇ ( 5 ) w ⁇ ( 5 ) + a_dc ′ ⁇ v_dc ′ ⁇ 2
  • w ′ ⁇ ( 6 ) w ⁇ ( 6 ) + a_dc ′ ⁇ v_dc ′ ⁇ - 1
  • control variable v_skip should be read as the control variable v_dc′.
  • equation should be read in the similar manner in case where the filter base b_dc′ is another base.
  • the motion vector is, for example, a motion vector for use for inter-frame prediction under H.264/AVC.
  • the control variable determining section 120 sets a value of a control variable v_mv to 1, wherein the control variable v_mv is 0 or 1 according to the size of the vector associated with the block to which the pixel at the coordinates (x, y) in the image data # 100 belongs. Then, the control variable determining section 120 outputs the control variable v_mv to the filter section 110 .
  • control variable determining section 120 outputs, to the filter section 110 , a filter base b_mv associated with a size of the motion vector associated with the block to which the pixel at the coordinates (x, y) in the image data # 100 belongs.
  • the filter base b_mv may be, for example, the edge-stress base.
  • the filter section 110 determines, with reference to the teacher data # 200 , the coefficients of the filter coefficient vector V having the components v(K) expressed by the following Equation (25).
  • a_mv is a coefficient (fixed coefficient) corresponding to the filter base b_mv.
  • the filter section 110 determines the components w(K) of the fixed coefficient vector and the coefficient a_mv per slice or per frame by, for example, the aforementioned statistical approach.
  • the control variable determining section 120 sets the value of the control variable v_mv to 1 when the size of the motion vector associated with the block to which the pixel at the coordinates (x, y) in the image data # 100 belongs is equal to or greater than a predetermined threshold, whereas the control variable determining section 120 sets the value of the control variable v_mv to 1 when the size of the motion vector is less than the threshold. Moreover, the control variable determining section 120 outputs the control variable v_mv to the filter section 110 .
  • the filter section 110 For every filter target area T, the filter section 110 performs the filtering based on the filter coefficient vector V using the control variable v_mv supplied from the control variable determining section 120 .
  • the filtering can be performed such that filtering having a blurring effect is performed when the size of the motion vector is equal to or greater than the threshold value, whereas filtering having no blurring effect is performed when the size of the motion vector is less than the threshold value.
  • the filter coefficient is switched over according to the size of the motion vector as described above, it is also possible to perform the filtering having the luminance change effect due to the DC offset, by using a filter base and filter coefficient, each of which corresponds to the DC offset described above.
  • the filter coefficient is switched over according to the size of a differential motion vector, instead of the size of the motion vector, wherein the differential motion vector is a difference between the motion vector and a predicted vector.
  • the present invention is not limited to the examples mentioned above.
  • the present invention encompasses cases where the filter coefficient is switched over in a way in which the above examples are combined as appropriate. That is, the control variable determining section 120 may be configured to output some of the control variables and some corresponding ones of the control bases, which are described above. With such a configuration, it is possible to perform more effective filtering.
  • w′(i) in the (preliminary step) of the aforementioned ⁇ Filter Coefficient Calculation Process> may be calculated out by the following Equation, and stored.
  • w ′ ⁇ ( 5 ) w ⁇ ( 5 ) + a_mv ⁇ v_mv ⁇ 2
  • w ′ ⁇ ( 6 ) w ⁇ ( 6 ) + a_mv ⁇ v_mv ⁇ - 1
  • control variable v_skip should be read as the control variable v_mv.
  • equation should be read in the similar manner in case where the filter base b_mv is another base.
  • control variable determining section 120 outputs a plurality of control variables
  • the control variable determining section 120 determines a plurality of control variables whose contribution to the filtering is small, a new control variable is formed from a linear sum of the plurality of control variables whose contribution to the filtering is small, and the filter coefficient is switched over by switching over the value of the new control variable.
  • the present invention is not limited to a particular method for determining which control variables are small in contribution to the filtering. For example, filtering is performed with different values for different control variables so as to find control variables with which a difference between a pixel value before the filtering and a pixel value after the filtering is small, and identify such control variables are control variables whose contribution to the filtering is small.
  • the present invention is not limited to the above-explained examples in which the coefficient for the filter base outputted from the control variable determining section 120 is determine by the statistical approach.
  • control variable determining section 120 may be configured such that a filter base optimal for the coefficient determined by the statistical approach is determined.
  • the step of calculating the coefficient for a fixed filter base and the step of determining a filter coefficient optimal for a fixed coefficient may be repeated so as to find the filter coefficient and the filter base, which are optimal for the image data # 100 .
  • a moving image encoding device and a moving image and a moving image decoding device and a moving image, each of which is provided with the adaptive filter 100 are described below as specific application examples of the adaptive filter 100 , referring to FIGS. 5 to 9 .
  • a moving image encoding device 1 provided with the adaptive filter 100 is described below, referring to FIG. 4 .
  • the moving image encoding device 1 is a moving image encoding device in part of which the H.264/AV standard and a technique adopted in the KTA software are used.
  • FIG. 5 is a block diagram illustrating a configuration of the moving image encoding device 1 .
  • the moving image encoding device 1 includes a transforming/quantizing section 11 , a variable length encoding section 12 , an inverse transforming/inverse quantizing section 13 , a buffer memory 14 , a intra predicted image generating section 15 , an inter predicted image generating section 16 , a prediction method controlling section 18 , a motion vector redundancy reducing section 19 , an adder 21 , a subtractor 22 , a deblocking filter 50 , and the adaptive filter 100 .
  • the moving image encoding device 1 receives an input image # 1 , which have been divided into block images (hereinafter, referred to as “macro blocks”) constituted from adjacent pixels.
  • macro blocks block images
  • the moving image encoding device 1 encodes the input image # 1 , and outputs encoded data # 2 .
  • the transforming/quantizing section 11 performs DCT (Discrete Cosine Transform) transformation of differential image # 22 between (i) the input image # 1 divided into the macro blocks and (ii) predicted image # 18 outputted from the later-described prediction method controlling section 18 , thereby transforming the differential image # 22 into frequency components. Then, the transforming/quantizing section 11 quantizes the frequency components, thereby generating quantized prediction residual data # 11 .
  • the quantization is calculation to associate the frequency component to integer value.
  • the DCT transformation and the quantization are performed per block obtained by dividing the macro blocks.
  • target macro blocks the macro blocks to be subjected to the above treatment
  • targets to be subjected to the above treatment is referred to as “target blocks”.
  • the inverse quantizing/inverse transforming section 13 decodes the quantized prediction residual data # 11 , thereby generating the prediction residual # 13 . More specifically, the inverse quantizing/inverse transforming section 13 performs inverse quantization of the quantized prediction residual data # 11 , (i.e., associating (i) the integral values constituting the quantized prediction residual data # 11 to (ii) the frequency components) and inverse DCT transformation of the frequency component (i.e., inverse transformation of the target macro block into pixel components on the basis of the frequency components), thereby generating the prediction residual # 13 .
  • inverse quantization of the quantized prediction residual data # 11 i.e., associating (i) the integral values constituting the quantized prediction residual data # 11 to (ii) the frequency components
  • inverse DCT transformation of the frequency component i.e., inverse transformation of the target macro block into pixel components on the basis of the frequency components
  • the adder 21 adds the prediction residual # 13 and a predicted image # 18 a , thereby obtaining a decoded image # 21 .
  • the decoded image # 21 thus generated is supplied to the deblocking filter 50 .
  • the deblocking filter 50 performs deblocking process to the block border or the macro block border in the decoded image # 21 .
  • the image having been subjected to the deblocking is outputted as a deblocked image # 50 .
  • the adaptive filter 100 performs the filtering explained in the ⁇ Examples of Filtering Performed by the Adaptive Filter 100 >. Then, the adaptive filter 100 outputs an output image data # 110 a to the buffer memory 14 .
  • the deblocked image # 50 corresponds to the image data # 100 .
  • the adaptive filter 100 receives the input image # 1 , the decoded image # 21 , and side information associated with the decoded image # 21 .
  • the input image # 1 corresponds to the teacher data # 200
  • the side information corresponds to the side information # 100 a.
  • the adaptive filter 100 outputs the filter coefficient information # 101 , which is information on the filter coefficient used in the filtering.
  • the filter coefficient information # 101 corresponds to the filter coefficient information # 110 b . That is, the filter coefficient information # 101 includes the components w(K) of the fixed coefficient vector W, and the fixed coefficient used in the filtering.
  • the filter coefficient information # 101 includes the components w(K) of the fixed coefficient vector W, and the fixed coefficient a_skip. The same is true for the other operation examples.
  • the adaptive filter 100 in the present application example corresponds to ALF (Adaptive Loop Filter) in KTA, substantially. That is, the adaptive filter 100 in the present application example receives a pixel value of a pixel at integer coordinates (hereinafter, the pixel is referred to as integer pixel), and outputs the pixel value of the integer pixel. This corresponds to a case where x, y, x′ and y′ in Equation (1) are all integers.
  • ALF Adaptive Loop Filter
  • the moving image encoding device 1 may be configured such that it does not include the deblocking filter 50 .
  • the adaptive filter 100 performs the filtering directly on the decoded image # 21 , not the deblocked image # 50 .
  • the adaptive filter 100 can be able to reduce the block noise, like the deblocking filter. In such a case, the throughput can be reduced due to the elimination of the process performed by the deblocking filter 50 .
  • an adaptive filter specialized as a deblocking filter it is possible to obtain a greater effect than the use of a conventional deblocking filter (for example, it is possible to use a filter that changes its blurring effect according to the luminance value).
  • the moving image encoding device 1 may be configured to include (i) both of the deblocking filter and the adaptive filter, (ii) only the adaptive filter, but not the deblocking filter, or (iii) the adaptive filter serving as a deblocking filter, and the adaptive filter.
  • the intra predicted image generating section 15 extracts a local decoded image # 14 a (decoded area in the same frame of the target macro block) from the decoded image # 21 stored in the buffer memory 14 , and performs intra-frame prediction on the basis of the local decoded image # 14 a , thereby generating an intra predicted image # 15 .
  • the inter predicted image generating section 16 calculates a motion vector # 17 by using a reference image # 14 b , whose whole frame has been decoded and which is stored in the buffer memory 14 . Then, inter predicted image generating section 16 allocates the motion vector to the target block on the input image # 1 . The calculated motion vector # 17 is outputted to the predicted image generating section 16 and the motion vector redundancy reducing section 19 , and stored in the buffer memory 14 . Moreover, the inter predicted image generating section 16 performs motion compensation with respect to the reference image on the basis of the motion vector # 17 per block, thereby generating an inter predicted image # 16 .
  • the prediction method controlling section 18 compares the intra predicted image # 15 , the inter predicted image # 16 , and the input image “1 for every block, so as to select and output one of the intra predicted image # 15 and the inter predicted image # 16 as a predicted image # 18 . Moreover, the prediction method controlling section 18 also outputs a prediction mode # 18 b , which is information indicating which one of the intra predicted image # 15 and the inter predicted image # 16 is selected. The prediction image # 18 a is inputted in the subtractor 22 .
  • the prediction mode # 18 b is stored in the buffer memory 14 and inputted in the variable length encoding section 12 .
  • the motion vector redundancy reducing section 19 calculates out a predicted vector after the motion vector # 17 is allotted to the target block by the inter predicted image generating section 1 , the motion vector redundancy reducing section 19 calculating out the predicted vector based on a motion vector group 14 c stored in the buffer memory 14 , the motion vector group 14 c having been allotted to other blocks than the target block to which the motion vector # 17 is allotted by the inter predicted image generating section 16 . Moreover, the motion vector redundancy reducing section 19 determines a difference between the predicted vector and the motion vector # 17 , thereby generating the differential motion vector # 19 . The differential motion vector # 19 thus generated is supplied to the variable length encoding section 12 .
  • variable length encoding section 12 performs variable length encoding process to the quantized prediction residual data # 11 , the differential motion vector # 19 , the prediction mode # 18 b , and the coefficients included in the filter coefficient information # 101 , thereby generating the encoded data # 2 .
  • the subtractor 22 determines a difference between the input image # 1 and the predicted image # 18 a for the target macro block, and outputs a differential image # 22 .
  • FIG. 6 are views illustrating configurations of encoded data of the filter coefficient information # 101 .
  • encoded data of the filer coefficient information # 101 is constituted from the fixed filter parameter.
  • the values of the filter coefficient information # 101 have higher possibility to have values closer to 0, except a center element.
  • the values of the filter coefficient information # 101 are divided into signs and absolute values.
  • the signs are encoded by 1 bit, and the absolute values are encoded by shorter codes as the absolute values are closer to 0.
  • the method of encoding is not limited to this and arithmetic encoding may be adopted. The encoding is performed, considering that the center element has a value closer to 1.
  • a predicted value of the filter coefficient information # 101 is calculated out, and a difference (prediction difference) between the filter coefficient information # 101 and the predicted value is encoded.
  • the predicted value may be such that a predicted value for w( 5 ) corresponding to a center is 0.8 and predicted values for w(i) around the center are 0.1, considering that a center value is closer to 1 and the other elements are closer to 0, for example.
  • the prediction method is not limited to this.
  • the encoded data is constituted from (1) filter coefficients regarding components invariable according to the control variable, and (2) filter coefficients regarding components variable according to the control variable, and being associated with the values of the control variables.
  • the filter coefficient v(i) as such is encoded for the component invariable according to the control variable.
  • the encoded data includes (1) filter coefficients (hereinafter, referred to as basic coefficients) regarding various components invariable according to the control variables, and (2) differences between the basic coefficients and the filter coefficients regarding the components variable according to the control variables.
  • the number of the filter coefficient information to be encoded is equal between the encoding each of the components variable and invariable according to the filter coefficients and the encoding only the filter coefficient information regarding the control variable.
  • an encoding device (moving image encoding device 1 ) according to the present invention comprises the image filter (adaptive filter 100 ) and causes the image filter to process a decoded image.
  • the encoding device according to the present invention with the aforementioned configuration can process the decoded image with the filtering performed by the image filtering, the decoded image being obtained by decoding an encoded image having been encoded by the encoding device.
  • the encoding device according to the present invention has an effect similar to that of the image filter.
  • the encoding device with the above configuration can perform filtering more suitable for the respective target areas (filter target area T) by using (i) the components of the fixed filter coefficient vector (fixed coefficient vector calculated out from the fixed coefficient vector W), and (ii) the fixed coefficient.
  • the encoding device with the above configuration, it is possible to reduce the amount of encoding the filter coefficient vector, compared with the case where the filter coefficient vector having different components for different target areas is used.
  • the adaptive filter 100 according to the present invention is applicable to the inter predicted image generating section 16 in the moving image encoding device 1 .
  • FIG. 7 is a block diagram illustrating a configuration of the inter predicted image generating section 16 provided with the adaptive filer 100 .
  • the adaptive filter 100 in the present application example may be referred to as an adaptive filter 100 ′.
  • the inter predicted image generating section 16 includes a predicted image generating section 16 a , a motion vector estimating section 17 , and an adaptive filter 100 ′.
  • the adaptive filter 100 ′ corresponds to AIF (Adaptive Interpolation Filter) in KTA, substantially.
  • the reference image # 14 b stored in the buffer memory 14 is supplied to the adaptive filter 100 ′.
  • the reference signal # 14 b is referred to as first reference signal # 14 b.
  • the adaptive filter 100 ′ performs filtering to the first reference image # 14 b on the basis of a predetermined filter coefficient, thereby outputting first output image data # 110 a ′. That is, the adaptive filter 100 ′ in this step operates as a fixed filter for filtering based on the predetermined filter coefficient.
  • the predetermined filter coefficient may be, for example, a filter coefficient used in H.264/AVC.
  • an image signal whose pixel precision is not more than integer pixels is generated, by interpolation, from the pixel values of the integer pixels contained in the first reference image # 14 b .
  • the first output image data 100 ′ is image data including (i) the pixel values of the integer pixels and (ii) an interpolation signal whose pixel precision is not more than integer pixels.
  • the motion vector estimating section 17 generates the first motion vector # 17 ′ on the basis of the first output image data # 110 a ′ and the input image # 1 .
  • the first motion vector # 17 ′ is prepared for 1 slice or for 1 frame.
  • the adaptive filter 100 ′ determines a filter coefficient by processing the first reference image # 14 b with the process described in ⁇ Examples of Filtering Performed by Adaptive Filter 100 >.
  • the image generated by filtering performed by the adaptive filter 100 ′ is referred to as second output image data # 110 A′′.
  • the first reference image # 14 b corresponds to the image data # 100 above
  • the second output image data 110 a ′′ corresponds to the output image data # 110 a above.
  • the input image # 1 is supplied to the adaptive filter 100 ′ as the teacher data # 220 .
  • the adaptive filter 100 ′ generates an image signal whose pixel precision is not more than integer pixel, by interpolation, from the pixel values of the integer pixels contained in the first reference image # 14 b . This corresponds to a case where x′ and y′ in Equation (1) are numbers other than integers. Therefore, the second output image data 110 a ′ is image data including (i) the pixel values of the integer pixel and (ii) an interpolation signal whose pixel precision is not more than integer pixel.
  • the motion vector estimating section 17 generates the second motion vector # 17 ′′ on the basis of the second output image data # 110 a ′′, and the input image # 1 .
  • the adaptive filter 100 ′ performs filtering to the first reference image # 14 b by using the filter coefficient thus obtained, thereby generating a motion compensated image based on the second motion vector # 17 ′′ so as to generate and output the inter predicted image # 16 .
  • the adaptive filter 100 ′ may be configured to generate the second output image data # 110 a ′ by using control variables and filter bases, which are set according to the locations of the pixels in the output image data # 110 a′.
  • the moving image decoding device 2 provided with the adaptive filter 100 is described, referring to FIG. 8 .
  • the moving image decoding device 2 a moving image decoding device in part of which the H.264/AV standard and a technique adopted in the KTA software are used.
  • FIG. 8 is a block diagram illustrating a configuration of the moving image decoding device 2 .
  • the moving image decoding device 2 includes a variable length decoding section 23 , a motion vector reconstituting section 24 , a buffer memory 25 , an inter predicted image generating section 26 , an intra predicted image generating section 27 , a prediction method determining section 28 , an inverse quantizing/inverse transforming section 29 , an adder 30 , a deblocking filter 50 , and the adaptive filter 100 .
  • the moving image decoding device 2 receives an encoded data # 2 and outputs an output image # 3 .
  • variable length decoding section 23 performs variable length decoding of the encoded data # 2 , and outputs a differential motion vector # 23 a , side information # 23 b , a quantized prediction residual data # 23 c , and filter coefficient information # 23 d.
  • the motion vector reconstituting section 24 reconstitutes a motion vector # 24 of a target partition from the differential motion vector # 23 a and the motion vector 25 a , which have been decoded and stored in the buffer memory 25 .
  • the buffer memory 25 stores therein the output image data # 110 a outputted from the adaptive filter 100 , the motion vector # 24 , and the side information # 23 b.
  • the inter predicted image generated section 26 generates an inter predicted image # 26 on the basis of (i) the motion vector # 24 , which has been decoded by the motion vector reconstituting section 24 and supplied thereto via the buffer memory 25 , and (ii) the reference image # 25 d stored in the buffer memory 25 .
  • a motion vector # 25 c includes a motion vector identical with the motion vector # 24 .
  • the reference image # 25 d corresponds to the later-described output image data # 110 a outputted from the adaptive filter 100 .
  • the intra predicted image generating section 27 generates an intra predicted image # 27 from a local decoded image # 25 b in the same image as the target macro block, the intra predicted image # 27 being stored in the buffer memory 25 .
  • the prediction method determining section 28 selects and outputs, as a predicted image # 28 , one of the intra predicted image # 27 and the inter predicted image # 26 .
  • the inverse quantizing/inverse transforming section 29 performs inverse quantization and inverse DCT transformation to the quantized prediction residual data # 23 c , thereby outputting prediction residual # 29 .
  • the adder 30 adds the prediction residual # 29 and the predicted image # 28 , and outputs the resultant image as the decoded image # 3 . Moreover, the decoded image # 3 thus outputted is supplied to the deblocking filter 50 .
  • the deblocking filter 50 performs deblocking process to a block border or a macro block border in the decoded image # 3 , if a difference between pixels adjacent with each other across the block border or the macro block border in the decoded image # 3 is smaller than a predetermined threshold.
  • the image data having been subjected to the deblocking process is outputted as deblocked image # 50 .
  • the adaptive filter 100 subjects the deblocked image # 50 to the filtering described in ⁇ Examples of Filtering Performed by the Adaptive Filter 100 >.
  • the deblocked image # 50 corresponds to the image data # 100 described above.
  • the adaptive filter 100 in the present application example calculates out the filter coefficient on the basis of the filter coefficient information # 23 d decoded from the encoded data # 2 , and performs the filtering by using the filter coefficient thus calculated out.
  • the adaptive filter 100 in the present application example is configured such that the adaptive filter 100 calculates out the filter coefficient based on the filter coefficient information # 23 d decoded from the encoded data # 2 , rather than determining the filter coefficient by the statistical approach as described in ⁇ Examples of Filtering Performed by the Adaptive Filter 100 >, and the adaptive filter 100 performs the filtering by using the filter coefficient thus calculated out.
  • the switching-over of the filter coefficient in the adaptive filter 100 has been described in ⁇ Examples of Filtering Performed by the Adaptive Filter 100 >
  • the adaptive filter 100 outputs to the buffer memory 25 the output image data # 110 a thus generated by filtering the deblocking image # 50 .
  • the adaptive filter 100 receives the side information # 23 b .
  • the side information # 23 b includes (i) information regarding whether the target block is double direction predicted block or single direction predicted block, (ii) information regarding whether the skip mode is applied to the target macro block or not, (iii) prediction mode information regarding whether the target macro block is intra predicted macro block or inter predicted macro block, and (iv) quantized parameter QP associated with the target block.
  • the side information # 23 b corresponds to the side information # 100 a described above.
  • the adaptive filter 100 in the present application example corresponds to ALF (Adaptive Loop Filter) in KTA, substantially. That is, the adaptive filter 100 in the application example receives and outputs pixels values of integer pixels. This corresponds to cases where x, y, x′ and y′ in Equation (1) are all integers.
  • ALF Adaptive Loop Filter
  • the moving image decoding device 2 may be provided with no deblocking filter 50 .
  • the adaptive filter 100 performs the filtering directly to the decoded image # 3 , rather than the deblocked image # 50 . This configuration can reduce the throughput, as in the case of the moving image encoding device 1 .
  • the moving image decoding device 2 may be configured to include (i) both of the deblocking filter and the adaptive filter, (ii) only the adaptive filter, but not the deblocking filter, or (iii) the adaptive filter serving as a deblocking filter, and the adaptive filter.
  • variable length decoding section 23 decodes a fixed filer parameter as the filter coefficient information # 23 d .
  • w( 1 ) to w(M ⁇ N+1) and a_skip are decoded.
  • the adaptive filter 100 receives the fixed filter parameter and the control variable into the filter calculating function F whose general forms are shown in Equations 8 and 9, and derives the filter coefficient.
  • adaptive filter 100 receives w( 1 ) to w(M ⁇ N+1) and a_skip decoded in Equation (6) are inputted, and derives the filter coefficient vector V.
  • the variable length decoding section 23 decodes the filter coefficient as the filter coefficient information # 23 d .
  • one filter coefficient is encode for the component in variable according to the control variable, whereas a plurality of filter coefficients are encoded for respective values of the control variable for the components variable according to the control variable.
  • the variable length decoding section 23 can derive a filter coefficient by branching the process depending on whether the coefficient is variable or invariable according to the control variable.
  • variable length decoding section 23 decodes the basic filter coefficient as the filter coefficient information # 23 d , then decodes a difference from the basic filter coefficient.
  • the filter coefficient can be derived by branching for every component depending on whether or not being variable to the control variable.
  • the adaptive filter 100 performs the operation as described above referring to FIGS. 3 and 4 .
  • the process is carried out as shown in the flow of FIG. 3 .
  • the filter coefficient vector is derived by summation such as Equation (9)
  • the process is carried out as shown in the flow of FIG. 4 .
  • the process may be carried out as shown in the flow of FIG. 3 , by using the preliminary step, or the process may be carried out by referring to a table specified according to positions of the components of the filter coefficient and the values of the control variable are specified.
  • the filter coefficient, the difference of the filter coefficient, or a degree of increase in the filter coefficient are decoded as the filter coefficient information # 23 d to be decoded.
  • the present invention may be configured such that predicted value of the filter coefficient, the difference of the filter coefficient, or a degree of increase in the filter coefficient is derived and a difference from the predicted value is decoded.
  • a decoding device (moving image decoding device 2 ) according to the present invention comprises the image filter (adaptive filter 100 ) and causes the image filter to process the decoded image.
  • the decoding device according to the present invention with the above configuration can attain the same effect as the image filter. Moreover, the decoding device with the above configuration can perform filtering more suitable for the respective target areas (filter target area T) by using (i) the components of the fixed filter coefficient vector (fixed coefficient vector calculated out from the fixed coefficient vector W), and (ii) the fixed coefficient. Thus, it is possible to perform the filtering by using a smaller number of components of the filter coefficient vector, compared with the case where the filter coefficient vector having different components for different target areas is used.
  • the decoding device with the above configuration, it is possible to reduce the amount of encoding the filter coefficient vector, compared with the case where the filter coefficient vector having different components for different target areas is used.
  • the adaptive filter 100 according to the present invention is applicable to the inter predicted image generating section 26 in the moving image decoding device 2 .
  • FIG. 9 is a block diagram illustrating a configuration of the inter predicted image generating section 26 provided with the adaptive filter 100 .
  • the adaptive filter 100 in the present application example is referred to as an adaptive filter 100 ′′.
  • the inter predicted image generating section 26 includes a predicted image generating section 26 a and the adaptive filter 100 ′′.
  • the adaptive filter 100 ′′ performs filtering to the reference image # 25 d stored in the buffer memory 25 .
  • the reference image # 25 d corresponds to the image data # 100 described above.
  • the adaptive filter 100 ′′ in the application example performs filtering based on the filter coefficient information # 23 d decoded from the encoded data # 2 .
  • the adaptive filter 100 ′′ in the present application example is configured such that the adaptive filter 100 ′′ performs the filtering based on the filter coefficient information # 23 d decoded from the encoded data # 2 , rather than determining the filter coefficient by the statistical approach as described in ⁇ Examples of Filtering Performed by the Adaptive Filter 100 >, and the adaptive filter 100 performs the filtering by using the filter coefficient thus calculated out.
  • the switching-over of the filter coefficient in the adaptive filter 100 has been described in
  • the adaptive filter 100 ′′ outputs, to the predicted image generating section 26 a , the output image data # 110 a ′′′ thus generated by the filtering the reference image # 25 d.
  • the adaptive filter 100 ′′ receives the side information # 23 b .
  • the side information # 23 b includes (i) information regarding whether the target block is double direction predicted block or single direction predicted block, (ii) information regarding whether the skip mode is applied to the target macro block or not, (iii) prediction mode information regarding whether the target macro block is intra predicted macro block or inter predicted macro block, and (iv) quantized parameter QP associated with the target block.
  • the side information # 23 b corresponds to the side information # 100 a described above.
  • the adaptive filter 100 ′′ in the application example corresponds to the AIF (Adaptive Interpolation Filter) in KTA, substantially. That is, the adaptive filter 100 ′′ in the application example receives pixel values of integer pixels and outputs pixels values whose pixel precision is not more than the integer pixel. This corresponds to cases where x and y in Equation (1) are integer, and x′ and y′ in Equation (1) are an integer or a decimal.
  • AIF Adaptive Interpolation Filter
  • the predicted image generating section 26 a performs motion compensation to the output image data # 110 a ′′′ by using the motion vector # 25 c , thereby generating and outputting the inter predicted image # 26 .
  • FIG. 10 is a view illustrating a bit stream #BS, which is generated for every slice of the encoded data # 2 by the moving image encoding device 1 so as to be referred to for the slice of the encoded data # 2 by the moving image decoding device 2 .
  • the bit stream #BS contains filter coefficient information FC, and macro block information MB 1 to MBN.
  • FIG. 11 is a view illustrating one example of the filter coefficient information FC.
  • the filter coefficient information FC is information including the filter coefficient information # 23 d generated by the adaptive filter 100 , like the filter coefficient information # 101 described above referring to FIG. 6 .
  • the moving image decoding device 2 may be configured to decode information indicating which kind of control variable the control variable to use is, in addition to the filter coefficient information # 23 d .
  • 8-bit information from 0 to 255 is decoded, and the kind of the control variable to use is judged according to the bit number, for instance, 1 bit indicates the degree of flatness as a control variable, 2 bits indicate the edge angle as a control variable, 4 bits indicate the luminance as a control variable, 8 bits indicate color as a control variable, 16 bits indicate, as a control variable, whether an edge is included or not, 32 bits indicate the prediction direction as a control variable, 64 bits indicate, as a control variable, whether or not skip mode is applied, and 128 bits indicate, as a control variable, whether the inter or intra is used. Note that it is possible to select a plurality of control variables in free combinations by encoding such encoded data.
  • the optional control variables may be these described in the present Specification, or those not described herein.
  • the kind of the base to use for the control variable may be decoded. For example, 8-bit information from 0 to 255 is decoded, and the kind of the base to use is judged according to the bit number, for instance, 1 bit indicates the horizontal edge-stress base as a control variable, 2 bits indicate the vertical edge-stress base as a control variable, 4 bits indicate the standard Laplacian base as a control variable, 8 bits indicate first diagonal direction Laplacian base as a control variable, 16 bits indicate the second diagonal Laplacian base as a control variable, 32 bits indicate the DC offset base as a control variable, . . . .
  • the moving image decoding device 2 can select the kinds of the bases to use for individual control variables by decoding the kind of the based for every control variable.
  • FIG. 11 an example in which bases are designated for respective two control variables # 1 and # 2 .
  • the optional bases may be these described in the present Specification, or those not described herein.
  • the moving image decoding device 2 may be configured to decode the information as to the number of tapping in the filter, for example, the decoding may be such that 0 is decoded for 3 ⁇ 3 filter, 1 is decoded for 5 ⁇ 5 filter, 2 is decoded for 7 ⁇ 7 filter, and 3 is decoded for 9 ⁇ 9 filter. In this way, it is possible to change the number of the filter coefficient vectors according to the filter effect, thereby improving encoding efficiency.
  • the moving image decoding device 2 may or may not be configured to change the number of tapping in the base according the number of tapping in the decoded filter. Therefore, there may be such a case that the number of tapping in the filter is 5 ⁇ 5 while that of the base is 3 ⁇ 3.
  • the value of the filter coefficient information # 23 d for a position out of the range of the base is decoded as the value of the filter coefficient vector.
  • the value of the filter coefficient information # 23 d for a position out of the range of the base is a component invariable according to the control variable.
  • the value of the filter coefficient vector is a component variable according to the control variable and can be defined even out of the range of the tapping in the filter, but the filtering is carried out only with filter coefficients as many as the number of taps in the filter.
  • the moving image decoding device 2 may be configured to decode information indicating the format of the encoded data of the filter. For example, it may be such that 0 is decoded for the format of ( a ) of FIG. 6 , 1 is decoded for the format of ( b ) of FIG. 6 , and 2 is decoded for the format of ( c ) of FIG. 6 . Even in an encoding device having a difficulty in determining the filter coefficient information # 23 d in the format of ( a ) of FIG. 6 because of the algorithm for determining the filter coefficient vector, it is possible to improve the encoding efficiency by changing, depending on the control variable, the filter coefficient vector to used.
  • the moving image decoding device 2 may be configured to decode information indicating bit depth of the filter coefficient information. For example, whether the bit number is 8, 10, or 12 bits is decoded.
  • the bit depth is equivalent to roughness of the quantization of the filter coefficient information.
  • the moving image decoding device 2 may be configured to decode the range of the values of the control variable. For example, whether the range of the valuables of the control variable is 0 and 1, or 0, 1, and 2 is decoded. Moreover, information indicating the threshold value for use in determining the control variable from the image data # 100 and the side information thereof may be decoded. For example, the threshold information for determining the value of the control variable from the luminance or the degree of flatness may be decoded. Moreover, filtering unit may be decoded. The filtering unit may be pixel, block, a group of block, or the other. Moreover, in case where an area not to be subjected to the filtering is included in the unit area to be subjected to the filtering, information for specifying whether there is a filtering area may be included.
  • filter coefficient information controlling information These kinds of information are referred to as filter coefficient information controlling information.
  • the filter coefficient vector can be switched over to be optimal for sequence, picture, or slice by decoding the filter coefficient information controlling information such as control variable, base, the number of tapping, or the like as described above. This improves the encoding efficiency. Moreover, the filter coefficient vector can be switched over appropriately for computing power of the encoding device. This gives the encoding device a greater degree of freedom. For example, if the capacity of the encoding device is large, the encoding efficiency can be increased by increasing the number of the control variables or the number of the bases. On the other hand, if the capacity of the encoding device is small, it may be possible to minimize the number of the control variables or the number of the bases, or not to use the control variable or the base.
  • the encoding device may be such that the degree of flatness is used as the control variable.
  • degree of freedom is not beneficial to the encoding device, but also for the decoding device. That is, if it can be assumed that the decoding device is small in capacity, the encoding can be performed in such a way that the number of the control variables to use or the number of bases to use is minimized.
  • filter coefficient information format controlling information indicating which information is included as the filter coefficient information controlling information may be decoded.
  • the filter coefficient information 4 b , the filter coefficient information controlling information, and the filter coefficient information format controlling information may be encoded in different positions on the encoded data.
  • the filter coefficient information controlling information is encoded in a slice header
  • the filter coefficient information format controlling information is encoded in a picture header
  • the filter coefficient information format controlling information is encoded in a sequence header.
  • the encoding may be to create encoded data corresponding to a picture parameter set or a sequence parameter set in the H.254/AVC.
  • the macro block information MB 1 to MBN are information regarding the macro blocks included in the slice in interest, and includes a prediction mode # 18 b of the macro block, and block dividing information, etc.
  • n is the number of the macro blocks included in the slice in interest.
  • the data structure of the encoded data according to the present invention is a data structure of an encoded data, which a decoding device refers to, the decoding device (i) including an image filter for calculating a pixel value of target pixel in an output image from a pixel value of each of a pixel or pixels belonging to a target area in an input image by using a filter coefficient vector, and (ii) causing the image filter to process a decoded image, the data structure comprising: a fixed coefficient, with which the decoding device multiplies a product of a predetermined fixed vector and a variable coefficient variable according to at least either where the target area is in the input image, or where the target pixel is in the output image; and components of a fixed filter coefficient vector to which the decoding device adds a product of the variable coefficient, the fixed vector, and the fixed coefficient, the fixed filter coefficient vector being predetermined for every unit area constituting the output image.
  • the decoding device for decoding an image based on the encoded data can perform filtering by using the components of the fixed coefficient vector, and the fixed coefficient contained in the encoded data, thereby performing filtering more appropriately adaptive for every target area by using a smaller number of filter coefficient, compared with the case where the encoded data contains different filter coefficients for the target area, respectively.
  • an image filter is an image filter for calculating a pixel value of target pixel in an output image from a pixel value of each of a pixel or pixels belonging to a target area in an input image and from a filter coefficient vector, comprising: filter coefficient vector changing means for changing the filter coefficient vector according to at least either where the target area is in the input image, or where the target pixel is in the output image.
  • the filter coefficient vector changing means for changing the filter coefficient vector according to at least either where the target area is in the input image, or where the target pixel is in the output image is provided, it is possible to perform filtering more appropriately adaptive for every target area according to at least either where the target area is in the input image, or where the target pixel is in the output image.
  • the filter coefficient vector is dividable into a sum of (i) a fixed filter coefficient vector predetermined for every unit area constituting the output image, and (ii) a variable filter coefficient vector variable by the filter coefficient vector changing means according to at least either where the target area is in the input image, or where the target pixel is in the output image.
  • the filter coefficient vector is dividable into the sum of (i) the fixed filter coefficient vector predetermined for every unit area constituting the output image, and (ii) the variable filter coefficient vector, and the filter coefficient vector changing means can change the variable filter coefficient vector according to at least either where the target area is in the input image, or where the target pixel is in the output image.
  • filtering can be more appropriately adaptive for every target area (i.e., more suitably for image characteristics in the individual target areas) by using the filter coefficient predetermined per unit area.
  • the unit area may be a slice, may be a macro block, or may be a block.
  • the fixed filter coefficient vector and the variable filter vector may be defined to include an offset component.
  • the fixed filter coefficient vector is dividable into (i) a fixed coefficient vector predetermined for every unit area constituting the output image, and (ii) a base.
  • variable filter coefficient vector is further dividable into (i) a fixed coefficient predetermined for every unit area constituting the output image, (ii) a variable coefficient variable by the filter coefficient vector changing means according to at least either where the target area is in the input image, or where the target pixel is in the output image, and a predetermined fixed vector.
  • variable filter coefficient vector is further dividable into (i) a fixed coefficient predetermined for every unit area constituting the output image, (ii) a variable coefficient, and (iii) a predetermined fixed vector.
  • the filter coefficient vector changing means can change the variable filter coefficient vector according to at least either where the target area is in the input image, or where the target pixel is in the output image. Therefore, it is possible to further attain such an effect that filtering can be more appropriately adaptive for image characteristics in the individual target areas by using the filter coefficient predetermined per unit area.
  • the image filter is supplied with the fixed coefficient vector (or fixed filter coefficient vector) determined for every unit area constituting the output image, and the fixed coefficient.
  • the above configuration makes it possible to perform filtering of desired types appropriate for 10 target areas individually by supplying the image filter with the P+1 number of coefficient data in total.
  • the fixed coefficient vectors are calculated out respectively for the 10 target areas, it becomes necessary to have the 10 ⁇ P number of coefficient data. Therefore, the above configuration makes it possible to perform filtering of types more appropriate for every target area by using a smaller number of coefficient data, compared with the case the fixed coefficient vectors (or filter coefficient vectors) are calculated out for the target areas, respectively.
  • variable coefficient is changed according to a degree of flatness in an image in the target area by the filter coefficient vector changing means; and the fixed vector is a filter base having at least one of edge-stress effect and a blurring effect.
  • the fixed vector is a filter base having at least one of edge-stress effect and a blurring effect
  • the filter coefficient vector changing means can change the variable coefficient according to a degree of flatness in an image in the target area, wherein the degree of flatness is varied according to the position of the target area. Therefore, with this configuration, in which at least one of the edge-stress effect or the blurring effect is changed according to the degree of flatness, it is possible to further attain such an effect that more appropriately adaptive filtering can be performed.
  • the filter base having at least one of the edge-stress effect or the blurring effect can provide edge-burring effect instead of edge-stress effect, depending on whether the fixed coefficient is positive or negative (the same is true hereinafter). Therefore, with this configuration, it is possible to further attain such an effect that more appropriately adaptive filtering by changing the edge-blurring effect according to the degree of flatness in the image in the target area.
  • variable coefficient is changed according to an edge angle of an edge in an image in the target area by the filter coefficient vector changing means; and the fixed vector is a filter base having at least one of edge-stress effect and a blurring effect.
  • the fixed vector is a filter base having at least one of edge-stress effect and a blurring effect
  • the filter coefficient vector changing means can change the variable coefficient according to an edge angle of an edge in an image in the target area. Therefore, with this configuration, it is possible to further attain such an effect that more appropriately adaptive filtering by changing at least one of the edge-stress effect and the blurring effect by changing the variable coefficient according the edge angle of the edge in the image in the target area.
  • variable coefficient is changed according to luminance of an image in the target area by the filter coefficient vector changing means; and the fixed vector is a filter base having at least one of edge-stress effect and a blurring effect.
  • the fixed vector is a filter base having at least one of edge-stress effect and a blurring effect
  • the filter coefficient vector changing means can change the variable coefficient according to luminance of an image in the target area, wherein the luminance is variable according to the position of the target area. Therefore, with this configuration, it is possible to further attain such an effect that more appropriately adaptive filtering by changing at least one of the edge-stress effect and the blurring effect according the edge angle of the edge in the image in the target area.
  • variable coefficient is changed according to a color of an image in the target area by the filter coefficient vector changing means; and the fixed vector is a filter base having at least one of edge-stress effect and a blurring effect.
  • the fixed vector is a filter base having at least one of edge-stress effect and a blurring effect
  • the filter coefficient vector changing means can change the variable coefficient according to a color of an image in the target area. Therefore, with this configuration, it is possible to further attain such an effect that more appropriately adaptive filtering by changing at least one of the edge-stress effect and the blurring effect according to the color of the image in the target area.
  • variable coefficient is changed according to whether the target area contains a block border or not by the filter coefficient vector changing means; and the fixed vector is a filter base having at least one of edge-stress effect and a blurring effect.
  • the fixed vector is a filter base having at least one of edge-stress effect and a blurring effect
  • the filter coefficient vector changing means can change the variable coefficient according to whether the target area contains a block border or not. Therefore, with this configuration, it is possible to further attain such an effect that more appropriately adaptive filtering by changing at least one of the edge-stress effect and the blurring effect according to whether the target area contains a block border or not.
  • variable coefficient is changed by the filter coefficient vector changing means according to whether the block containing the target area is a block subjected to unidirectional prediction or a block subjected to bidirectional prediction; and the fixed vector is a filter base having at least one of edge-stress effect and a blurring effect.
  • the fixed vector is a filter base having at least one of edge-stress effect and a blurring effect
  • the filter coefficient vector changing means can change the variable coefficient according to whether the block containing the target area is a block subjected to unidirectional prediction or a block subjected to bidirectional prediction. Therefore, with this configuration, it is possible to further attain such an effect that more appropriately adaptive filtering by changing at least one of the edge-stress effect and the blurring effect according to whether the block containing the target area is a block subjected to unidirectional prediction or a block subjected to bidirectional prediction.
  • variable coefficient is changed by the filter coefficient vector changing means according to whether or not a block including the target area is a block to which a skip mode is applied; and the fixed vector is a filter base having at least one of edge-stress effect and a blurring effect.
  • the fixed vector is a filter base having at least one of edge-stress effect and a blurring effect
  • the filter coefficient vector changing means can change the variable coefficient is changed according to whether or not a block including the target area is a block to which a skip mode is applied. Therefore, with this configuration, it is possible to further attain such an effect that more appropriately adaptive filtering by changing at least one of the edge-stress effect and the blurring effect according to whether or not a block including the target area is a block to which a skip mode is applied.
  • variable coefficient is changed by the filter coefficient vector changing means according to whether a block including the target area is an inter predicted block or an intra predicted block; and the fixed vector is a filter base having at least one of edge-stress effect and a blurring effect.
  • the fixed vector is a filter base having at least one of edge-stress effect and a blurring effect
  • the filter coefficient vector changing means can change the variable coefficient according to whether a block including the target area is an inter predicted block or an intra predicted block. Therefore, with this configuration, it is possible to further attain such an effect that more appropriately adaptive filtering by changing at least one of the edge-stress effect and the blurring effect according to whether a block including the target area is an inter predicted block or an intra predicted block.
  • components of the fixed filter coefficient vector, and the fixed coefficient are defined to minimize a difference between a certain image and the output image, where the input image is a decoded image obtained by encoding and decoding the certain image.
  • the components of the fixed filter coefficient vector, and the fixed coefficient are defined to minimize the difference between a certain image and the output image, where the input image is a decoded image obtained by encoding and decoding the certain image. Therefore, with this configuration, it is possible to further attain such an effect that the filtering can be performed by using a variable coefficient vector calculated out by using the components of the fixed filter coefficient vector, and the fixed coefficient, which have been thus optimized.
  • the fixed filter coefficient vector and the variable filter coefficient vector are orthogonal to each other.
  • this configuration in which the fixed filter coefficient vector and the variable filter coefficient vector are orthogonal to each other, it is possible to reduce the degree of freedom in the fixed filter coefficient vector, compared with a case where the fixed filter coefficient vector and the variable filter coefficient vector are not orthogonal to each other. Therefore, this configuration makes it possible to perform filtering more appropriately adaptive for every target area by using a smaller number of filtering coefficient.
  • the fixed filter coefficient vector and the variable filter coefficient vector may or may not be orthogonal to each other in the image filter according to the present invention.
  • a decoding device is configured to comprise the aforementioned image filter, and cause the image filter to process a decoded image.
  • the decoding device according to the present invention with this configuration can bring about the same effect as the image filter. Moreover, the decoding device with this configuration can perform filtering more appropriately adaptive for every target area by using the components of the fixed coefficient vector (or the fixed coefficient vector) predetermined per unit area, and the fixed coefficient predetermined per unit area. Therefore, the decoding device can obtain a filter coefficient vector by using a smaller number of coefficient data, compared with the case where a fixed coefficient vector (or a fixed coefficient vector) having different components for every target area.
  • the decoding device with this configuration, it is possible to reduce code quantity for encoding the coefficient data (filter coefficient vector), compared with the case where a fixed coefficient vector (or a fixed coefficient vector) having different components for every target area.
  • an encoding device is configured to comprise the aforementioned image filter, and cause the image filter to process a decoded image.
  • the encoding device according to the present invention with this configuration can bring about the same effect as the image filter. Moreover, the encoding device with this configuration can perform filtering more appropriately adaptive for every target area by using the components of the fixed coefficient vector (or the fixed coefficient vector) predetermined per unit area, and the fixed coefficient predetermined per unit area. Therefore, the encoding device can obtain a filter coefficient vector by using a smaller number of coefficient data, compared with the case where a fixed coefficient vector (or a fixed coefficient vector) having different components for every target area.
  • a data structure according to the present invention is a data structure of an encoded data, which a decoding device refers to, the decoding device (i) including an image filter for calculating a pixel value of target pixel in an output image from a pixel value of each of a pixel or pixels belonging to a target area in an input image by using a filter coefficient vector, and (ii) causing the image filter to process a decoded image, the data structure comprising: a fixed coefficient, with which the decoding device multiplies a product of a predetermined fixed vector and a variable coefficient variable according to at least either where the target area is in the input image, or where the target pixel is in the output image; and components of a fixed filter coefficient vector to which the decoding device adds a product of the variable coefficient, the fixed vector, and the fixed coefficient, the fixed filter coefficient vector being predetermined for every unit area constituting the output image.
  • the decoding device for decoding an image based on the encoded data can perform filtering by using the components of the fixed coefficient vector, and the fixed coefficient contained in the encoded data, thereby performing filtering more appropriately adaptive for every target area by using a smaller number of filter coefficient, compared with the case where the encoded data contains different filter coefficients for the target area, respectively.
  • the present invention is suitably applicable to an image filter for filtering image data. Moreover, the present invention is also suitably applicable to a broadcast receiving terminal configured to decode encoded image data, a HDD recorder configured to encode image data and record the encoded image data in a recording medium, and to decode image data recorded in a recording medium.

Abstract

An image filter (100) for calculating a pixel value of target pixel in an output image from a pixel value of each of a pixel or pixels belonging to a filter area in an input image by using a filter coefficient vector V, is configured to include a filter coefficient vector changing section (120) for changing the filter coefficient vector V according to at least either where the target area is in the input image, or where the target pixel is in the output image.

Description

    TECHNICAL FIELD
  • The present inventions relates to an image filter for filtering an image. Moreover, the present invention relates to an encoding device and a decoding device, each of which is provided with the image filter. Furthermore, the present invention relates to a data structure of an encoded data generated by the encoding device and referred to by the decoding device.
  • BACKGROUND ART
  • A moving image encoding device is used to efficiently transmit or record a moving image. As to moving image encoding techniques, H.264/MPEG-4 AVC and one adopted in KTA software may be exemplified as specific examples.
  • In such moving image encoding techniques, pictures constituting the moving image are managed in the form of a hierarchic structure constituted by slices (obtained by diving a picture), macro blocks (obtained by dividing a slice), and blocks (obtained by dividing the macro blocks). In general, a moving image is encoded on block basis.
  • Non-Patent Literature 1 discloses an ALF (Adaptive Loop Filter) for filtering a decoded image. ALF performs filtering in which slices in a decoded image are unit areas. For respective unit areas, filter coefficients to minimize an error between an original image and an ALF-treated decoded image are determined. The filtering of ALF is performed on unit area basis with the filter coefficients.
  • CITATION LIST Non-Patent Literature
    • Non-Patent Literature 1
    • ITU-T Telecommunication Standardization Sector VCEG-AI18, 2008 (disclosed in July 2008)
    SUMMARY OF INVENTION Technical Problem
  • However, the filtering with ALF is performed with a invariable filter coefficient for each respective unit area. Thus, if a feature of the image is not uniform within a unit area, it is difficult to perform appropriate filtering by using ALF.
  • The present invention was accomplished in view of the aforementioned problem. An object of the present invention is to realize an image filter capable of performing more appropriate filtering even for a unit area whose feature is not uniform, by use of filter coefficients predetermined for respective unit areas.
  • Solution to Problem
  • In order to solve the aforementioned problem, an image filter according to the present invention is an image filter for calculating a pixel value of target pixel in an output image from a pixel value of each of a pixel or pixels belonging to a target area in an input image by using a filter coefficient vector, including: filter coefficient vector changing means for changing the filter coefficient vector according to at least either where the target area is in the input image, or where the target pixel is in the output image.
  • With this configuration, in which the filter coefficient vector changing means for changing the filter coefficient vector according to at least either where the target area is in the input image, or where the target pixel is in the output image is provided, it is possible to perform filtering more appropriately adaptive for every target area according to at least either where the target area is in the input image, or where the target pixel is in the output image.
  • Advantageous Effects of Invention
  • As described above, an image filter according to the present invention is an image filter for calculating a pixel value of target pixel in an output image from a pixel value of each of a pixel or pixels belonging to a target area in an input image by using a filter coefficient vector, including: filter coefficient vector changing means for changing the filter coefficient vector according to at least either where the target area is in the input image, or where the target pixel is in the output image. Therefore, it is possible to perform filtering more appropriately adaptive for every target area according to at least either where the target area is in the input image, or where the target pixel is in the output image.
  • BRIEF DESCRIPTION OF DRAWINGS
  • FIG. 1 is a block diagram illustrating a configuration of an adaptive filter according to one embodiment.
  • FIG. 2 is a view for explaining an effect of the adaptive filter according to the present embodiment, and illustrates a unit area including a plurality of target areas.
  • FIG. 3 is a flowchart illustrating one example of a flow of a filter coefficient determining process for the adaptive filter of the present embodiment.
  • FIG. 4 is a flowchart illustrating another one example of a flow of a filter coefficient determining process for the adaptive filter of the present embodiment.
  • FIG. 5 is a block diagram illustrating a moving image encoding device including the adaptive filter according to the embodiment.
  • FIG. 6 is a view illustrating configurations of filter coefficient information to be outputted from the adaptive filter according the embodiment. (a) of FIG. 6 illustrates a first example of filter coefficient information. (b) of FIG. 6 illustrates a second example of filter coefficient information. (c) of FIG. 6 illustrates a third example of filter coefficient information.
  • FIG. 7 is a block diagram illustrating an inter-prediction image generating section provided to the moving image encoding device according to the present embodiment, the inter-prediction image generating section being provided with the adaptive filter according to the embodiment.
  • FIG. 8 is a block diagram illustrating a moving image decoding device provided with the adaptive filter according to the embodiment.
  • FIG. 9 is a block diagram illustrating an inter-prediction image generating section provided to the moving image decoding device according to the present embodiment, the inter-prediction image generating section being provided with the adaptive filter according to the embodiment.
  • FIG. 10 is a view illustrating a bit stream of encoded data generated by the moving image encoding device according to the embodiment, so as to be inputted to the moving image decoding device according to the embodiment.
  • FIG. 11 is a view more specifically illustrating a configuration of the filter coefficient information included in the encoded data generated by the moving image encoding device according to the embodiment, so as to be inputted to the moving image decoding device according to the embodiment.
  • DESCRIPTION OF EMBODIMENTS
  • A configuration of the adaptive filter 100 according to one embodiment is explained, referring to FIG. 1.
  • FIG. 1 is a block diagram illustrating a configuration of the adaptive filter 100. As illustrated in FIG. 1, the adaptive filter 100 includes a filter section 110 and a control variable determining section 120.
  • The adaptive filter 100 is a image filter capable for filtering image data. The adaptive filter 100 can perform filtering for unit areas in an image of image data on unit area basis by taking the unit area as a target of the filtering, wherein the adaptive filter 100 performs the filtering adjusted adaptively for respective areas (filtering target area T) obtained by dividing the unit area.
  • The adaptive filter 100 is suitable applicable for a moving image encoding device for encoding moving image data, or a moving image decoding device for decoding such encoded moving image data, for example.
  • The adaptive filter 100 generates and outputs output image data # 110 a by obtaining weighted linear sum, based on a filter coefficient, of pixel values contained in a certain area, among pixel values in inputted image data # 100.
  • Moreover, the adaptive filter 100 receives teacher data # 200. The teacher data # 200 is image data indicative of teacher image, which is referred to when the filter coefficient is determined as described later. One specific example of the teacher data # 200 is input image data inputted in the image encoding device provided with the adaptive filter 100.
  • Moreover, the adaptive filter 100 outputs filter coefficient information # 110 b together with the output image data # 110 a. The filter coefficient information # 110 b is information regarding a filter coefficient used for filtering.
  • The adaptive filter 100 obtains the weighted linear sum based on the filter coefficient that is adjusted per area according to the image data # 100 and to side information # 100 a of the image data # 100.
  • The side information # 100 a of the image data # 100 is information obtained from encoded data necessary for generating the image data # 100. That is, side information # 100 a of the image data # 100 includes all of transformation coefficients necessary for generating prediction residual, a size of frequency transformation, information of quantum parameter, information necessary for generating predicted image (i.e., prediction mode, a size of motion compensation, motion vector, information on reference frame etc., weight information for weighted prediction, information on block and positions of block boarders).
  • The filter coefficient information # 110 b is information including the filter coefficient and information necessary for generating the filter coefficient. Specific examples of the filter coefficient information # 110 b will be given later.
  • The present invention is not limited to the configuration described herein in which the filter target area T has a rectangular shape of MT×NT. For example, the filter target area T may be a single pixel, or a block or macro block. Moreover, the filter target area T is not limited to a rectangular shape, and may be rhombus, circular, a line segment (vertical, horizontal, or diagonal line segment), or another shape. Moreover, the size of the filter target area T may or may not be fixed; thus may be variable. For example, if the size of the filter target area T is variable block size variable according to transformation or motion compensation, it is appropriate that the area T is the variable block.
  • Hereinafter, sections of the adaptive filter 100 are explained.
  • The control variable determining section 120 receives (i) the image data # 100 and (ii) the size information #100 a of the image data # 100, and supplies control variable #120 a and control base # 120 b to a filter section 110.
  • The filter section 110 calculates out output image data #110 a for respective pixel values of the image data # 100 by performing filter coefficient-based weighted liner summation and offset value addition for pixel values included in a certain area.
  • More specifically, the filter section 110 performs weight linear summation as expressed in Equation (1), thereby to find a pixel value SO (x′, y′) at coordinates (x′, y′) in the output image data #110 a.
  • [ Math . 1 ] S O ( x , y ) = ( ( i , j ) R h ( i , j ) S l ( x + i , y + j ) ) + h offset ( 1 )
  • where SI (x, y) is a pixel value at the coordinates (x, y) of the image data # 100, h (i, j) is a filter coefficient with which the pixel value SI (x+i, y+j) is multiplied, R is an area constituted with pixels with which the weighted linear summation is performed (hereinafter, this area is called a filter reference area). More specifically, R is a set of relative coordinates of pixels with which the weighted linear summation is performed. For instance, in case of 3×3-tapped filtering around a pixel of coordinates (x, y), R={(−1,−1), (−1,0), (−1,1), (0,−1), (0,0), (0,1), (1,−1), (1,0), (1,1)}. Moreover, hoffset is the offset value to be added to the pixel value. Here, the coordinates (x′, y′) and the coordinates (x, y) may be identical with each other, or may be different from each other provided that they are one-to-one correspondence with each other. Moreover, the present invention is not limited as to how to associate the coordinates (x′, y′) and the coordinates (x,y) with each other. How to associate the coordinates (x′, y′) and the coordinates (x,y) with each other is determined depending on a concrete configuration of an image processing device to which the adaptive filter 100 is provided.
  • In general, M×N tapped filtering can be defined by a filter coefficient matrix H, which is a M×N matrix whose components are filter coefficients h (i, j), and the offset hoffset.
  • In the following, the present embodiment is explained based on an example in which the filter reference area is a rectangular M×N tapped area. It should be noted that the present invention is not limited to this example, and is applicable to filter reference areas R having a rhombus shape, a circular shape, a line segment (vertical, horizontal, or diagonal line segment), or another shape.
  • Hereinafter, the filter coefficient and the offset are expressed by vector notation in which the components of the filter coefficient matrix H and the offset hoffset are mapped into vectors of M×N+1 dimension.
  • More specifically, the filter coefficient matrix H expressed by the following Equation (2),
  • [ Math . 2 ] H = ( h 11 h 12 h 1 N h 21 h 22 h 2 N h M 1 h M 2 h MN ) ( 2 )
  • and the offset hoffset are mapped into the filter coefficient vector V of M×N+1 defined in Equation (3) below.
  • [Math. 3]

  • V=(h 11 ,h 12 , . . . ,h 1N ,h 21 ,h 22 , . . . ,h 2N , . . . ,h M1 ,h M2 , . . . ,h MN ,h offset)  (3)
  • As shown in Equation (3), 1st to M×N-th components in the filter coefficient vector V represent filter coefficient h (i, j). The last component (hereinafter, referred to as offset component) in the filter coefficient vector V represents the offset h offset.
  • The use of such notation makes it possible to express the filter coefficient h (i, j), and the offset hoffset in a unified manner.
  • Hereinafter, the components of the filter coefficient vector V are expressed as v(K) with a 1 dimensional suffix K. For instance, the components of the filter coefficient vector V in the above example are expressed as follows: v(1)=h(1, 1), v(2)=h(1,2), . . . , v(M×N)=v(M, N), and v(M×N+1)=hoffset.
  • In general, the relationship between the 1-dimensional suffix K and the 2-dimensional suffix (i, j) may be expressed as i=K % M+1, and j=K/M+1, where % is reminder. That is, x % y is a reminder remained after dividing x with y. Moreover, x/y is a result of rounding out fractional part of the result of dividing x with y (that is, x/y is a result of operation of Floor function).
  • With these notations, Equation (1) can be expressed as Equation (4):
  • [ Math . 4 ] S O ( X ) = K R v ( K ) S I ( X + K ) ( 4 )
  • where SO (X′) and SI (X+K) are 1-dimensional notations corresponding to SO (x′, y′) and SI (x+i, y+j). Moreover, the offset component for SI (X+K) is defined as 1. That is, it is defined for the M×N tapped filtering that SI (X+K)=1 when K=M×N+1. Furthermore, it is assumed that summation with respect to K covers not only the area R but also the offset component. In other words, it is assumed that the summation over the area R′ includes not only 1-dimensional summation to the filter reference area R but also summation to the offset component.
  • Moreover, the 1st to M×N-th components in the filter coefficient vector may or may not be independent from each other. For example, the filter coefficient vector V has been conventionally restricted to be in vertically or horizontally bilateral symmetry or point symmetry. For example, in case of horizontal bilateral symmetry for 3×1 tapped filter (4 tapped including the offset), the filter coefficient vector V can be expressed as below:
  • [ Math . 5 ] ( v 1 v 2 v 3 v 4 ) = w 1 ( 0 1 0 0 ) + w 2 ( 1 0 1 0 ) + w 3 ( 0 0 0 1 ) ( 5 )
  • where w1, w2, and w3 are elements of filter coefficient vector W. In this example, the 4-dimensional filter coefficient vector V are expressed by using 3-dimensional filter coefficient vector W, thereby reducing its degree of freedom by 1. Here, vectors for w1, w2, and w3 are called “bases”. In case where the bases are unit bases, W=V.
  • Expression capacity of the filter is reduced as much as the degree of freedom is reduced. Meanwhile, the reduction in degree of freedom leads to reduction in parameters necessary for expressing the filter (hereinafter, the parameters are referred to as filter parameters. In this example, the components of the filter coefficient vector W). By this, code quantity can be reduced. For example point symmetry has been widely adopted for this purpose.
  • In this Specification, a function for calculating out the filter coefficient vector V from the filter parameters is referred to as a filter calculating function F. Equation (5) is one example of such a filter calculating function F. It should be noted that the base is not limited to symmetric bases as described above, and may be another bases.
  • Moreover, the present invention is not limited to the aforementioned explanation in which the weighted linear summation based on the filter coefficient vector V (or the filter coefficient matrix H) is discussed as one example. In other words, the filtering of the adaptive filter 100 according to the present invention is not limited to the weighted linear summation and may adopt weighted non-linear summation using the component of the filter coefficient vector V.
  • The adaptive filter 100 according to the present embodiment can switch over the filter coefficient vector V according to the side information #100 a, the pixel values included in the image data # 100, positions of the pixels included in the image data # 100, or functions thereof.
  • <Examples of Filtering Process Performed by Adaptive Filter 100>
  • <Example in which the Filter Coefficient is Switched Over Depending on Whether to Perform the Filtering in a Skip Mode or not)
  • In the following, one example of filter coefficient vector calculation of the adaptive filter 100, and switching-over of the filter coefficient vector, based on the way of notation explained above. Especially, the following explanation is based on a case where the adaptive filter 100 switches over the filter coefficient vector, depending on whether or not a pixel at the coordinates (x, y) in the image data # 100 belongs to a macro block to which the skip mode in H.246/AVC standard is applied.
  • (Step S1)
  • The control variable determining section 120 sets a control variable v_skip to 1, wherein the control variable v_skip is either 0 or 1 depending on whether or not a pixel at the coordinates (x, y) in the image data # 100 belongs to a macro block to which the skip mode is applied. Then, the control variable determining section 120 outputs the control variable v_skip to a filter section 110.
  • Moreover, the control variable determining section 120 outputs a filter base b_skip to the filter section 110. The filter base b_skip is associated with information (hereinafter, referred to as skip information) regarding whether or not the pixel at the coordinates (x, y) in the image data # 100 belongs to the macro block to which the skip mode is applied.
  • Here, the filter base b_skip is a base vector constituted from base components corresponding to the components of the filter coefficient vector V. That is, the filter base b_skip is a M×N+1 dimensional vector in the case of the M×N-tapped filtering.
  • The filter base b_skip may be, for example, edge-stress/edge-blurred filter base, preferably. The edge-stress/edge-blurred filter base is a filter base effective to stress an edge or to blur the edge. Examples of the edge-stress/edge-blurred filter base encompass a base for double differentiating filter. Hereinafter, the edge-stress/edge-blurred filter base may be abbreviated simply as an edge-stress filter base. However, it should be noted that the edge-stress filter base has a blurring effect as well.
  • Examples of the edge-stress filter base encompass, for the case of 3×3 tapped filtering, a horizontal edge stress base (0, 0, 0, −1, 2, −1, 0, 0, 0,0), a vertical edge stress base (0, −1, 0, 0, 2, 0, 0, −1, 0, 0), a standard Laplacian base (−1, −1, −1, −1, 8, −1, −1, −1, −1, 0), a modified Laplacian base (−1, −2, −1, −2, 12, −2, −1, −2, −1, 0), a first diagonal Laplacian base (−1, −2, 0, −2, 6, 0, −1, 0, 0, 0), a second diagonal Laplacian (0, 0, −1, 0, 6, −2, 0, −2, −1, 0), and the like.
  • The control variable determining section 120 outputs, to the filter section 110, an edge-stress filter base or its linear sum thereof as the filter base b_skip.
  • (Step S2)
  • The filter section 110 determines the filter coefficient vector V, slice by slice or frame by frame. Here, the components v(k) of the filter coefficient vector V are expressed by Equation (6) below:

  • v(K)=w(K)+a_skip×v_skip×b_skip(K)  (6)
  • where w(K) is a component of the filter coefficient vector W (fixed coefficient vector), b_skip (K) is a component of the above-mentioned filter base (fixed vector) b_skip. Moreover, v_skip is a control variable (variable coefficient) described above, and is a control variable that has been set as v_skip=1 at Step S1. Furthermore, a_skip a coefficient (fixed coefficient) associated to the control variable v_skip.
  • The term “fixed” in the fixed vector and fixed coefficient means that the value does not vary depending on the position of the pixel. The term “variable” in the fixed vector and fixed coefficient means that the value varies depending on the position of the pixel.
  • Equation (6) is one example of the filter calculation function F using the control variable, and the filter parameters in the function F are the fixed coefficient vector, fixed coefficient, and control variable.
  • For each of the unit areas such as slices or frames, the filter section 110 determines fixed ones (hereinafter, referred to as fixed filter parameter) common to each area among the filter parameters. Here, the fixed ones among the filter parameter thus determined are the components w(K) of the fixed coefficient vector W and the fixed coefficient a_skip.
  • When the coefficient concerning the edge-stress base b_skip(K) is positive in Equation (6) (in this example, when a_skip is positive), it is possible to obtain edge stressing effect in an area in which the control variable v_skip=1. On the contrary, when the coefficient concerning the edge-stress base b_skip(K) is negative in Equation (6) (in this example, when a_skip is negative), it is possible to obtain edge blurring effect in an area in which the control variable v_skip=1.
  • The filter section 110 can determine the fixed filter parameter by, for example, statistical approach.
  • One specific example of determining the fixed filter parameter is to determine the fixed filter parameter that can minimizes a square error E from the pixel values in the teacher data # 200, the square error E being expressed by Equation (7) below.
  • [ Math . 6 ] E = X ( S ( X ) - K R v ( K ) S I ( X + K ) ) 2 ( 7 )
  • where S(X) is a pixel value at coordinates (x, y) in the teacher data # 200. Sum with respect to X is sum of all the pixels included in the unit area. That is, in case where the fixed filter parameter is determined per slice, the sum with respect to X is sum for all the pixels included in a slice in question. Even though this example is configured such that the sum with respect to X is the sum for all the pixels included in the unit area, it may be so configured that the sum with respect to X is sum for some of the pixels included in the unit area. For example, in case where the filter section 110 is used as an adaptive de-blocking filter, it is appropriate to determine the fixed filter parameter only from a processing area restricted to pixels distanced from a block border by a certain distance or a distance shorter than that. Moreover, in case where an area not to be subjected to the filtering is provided inside a unit area to be subjected to the filtering, and information regarding whether to perform the filtering is encoded, it is appropriate to determine the fixed filter parameter from pixels in the area to be subjected to the filtering.
  • More specifically, the filter parameter V in Equation (7) is transformed into a function of the filter parameters by using the filter calculating function F. Next, an equation thus obtained is partially differentiated with the fixed parameters, and then is set to be zero, thereby obtaining an equation generally called normal equation. The normal equation is solved by a process of Gaussian elimination etc., so as to calculate out the fixed parameters. In the calculation of the fixed parameters, S, SI, and the control variable (here, v_skip), which each pixel or each area uniquely has, are substituted in the calculation.
  • The introduction of the control variable v_skip can be expressed as M×N+2 tapped, like the introduction of the offset is expressed as M×N+1 tapped filter. In this case, the dimensions of V is expanded to M×N+2 dimensions by putting the M×N+2-th term in v as v(M×N+2)=a_skip, and the dimensions of SI is expanded by putting the M×N+2-th term in SI as SI(M×N+2)=Σv_skip×b_skip(K)×SI(K). By using such expression, to determine the M×N+2 numbers of v values, which can minimize Equation (7) becomes the question.
  • One other specific example of the method for determining the filter coefficient vector V is to determine a filter coefficient vector that minimizes an absolute value error obtained by replacing the square calculation in the square error E with absolute value calculation. Moreover, it is possible to use, as an object function, a function obtaining adding the square error E and a value obtained by multiplying square of each terms in the filter coefficient vector with a certain constant (this is so-called ridge regression). In this way, it is possible to reduce the square error E while keeping the filter coefficient vector at a small value. In this case, the square error E becomes slightly larger than its minimal value, but the filter coefficient vector becomes small, thereby avoiding an increase in the code quantity. Note that the present invention is not limited to the statistical approach for determining the filter coefficient vector.
  • The adaptive filter 100 may be configured such that the statistical approach is used with the control variable being constantly 0, so that the vector coefficient vector V for the case where the changing the filter coefficient vector with the control variable is not performed is determined.
  • The filter coefficient vector V thus determined is such that a filter having a filter-stress effect due to contribution from the second term in Equation (6) when v_skip=1. On the other hand, when v_skip=0, the second term in Equation (6) does not make contribution, whereby the filter coefficient vector V can also express a filter having no filter-stress effect. In the following, the filter coefficient vector V for v_skip=1 is referred to as the filter coefficient vector V (v_skip=1), and the filter coefficient vector V for v_skip=0 is referred to as the filter coefficient vector V (v_skip=0).
  • (Step S3)
  • The control variable determining section 120 sets the value of the control variable v_skip depending on whether or not the pixel at coordinates (x, y) in the image data # 100 belongs to a macro block to which the skip mode is applied. More specifically, the control variable determining section 120 sets the control variable v_skip as v_skip=1 when the pixel at coordinates (x, y) in the image data # 100 belongs to a macro block to which the skip mode is applied, whereas the control variable determining section 120 sets the control variable v_skip as v_skip=0 when the pixel at coordinates (x, y) in the image data # 100 belongs to a macro block to which the skip mode is not applied.
  • For the respective pixels included in the image data # 100, the control variable v_skip thus set is supplied to the filter section 110.
  • (Step S4)
  • The filter section 110 calculates the pixel values of the output image data #110 a by calculating out a weighted average of the pixel values in the image data # 100 according to the components of the filter coefficient vector V using the control variable v_skip supplied from the control variable determining section 120. More specifically, the filter section 110 calculates out a pixel value SO (X′) by applying the components of the filter coefficient vector V to Equation (4).
  • The adaptive filter 100 repeats the steps S3 an S4. Thus, the adaptive filter 100 can perform the filtering with different filter coefficient vectors, depending on whether or not the pixel included in the image data # 100 belongs to a macro block to which the skip mode is applied. To be more exact, the adaptive filter 100 performs filtering based on the filter coefficient vector V (v_skip=1) when a pixel at coordinates (x, y) in the image data # 100 belongs to a macro block to which skip mode is applied, whereas the adaptive filter 100 performs filtering based on the filter coefficient vector V (v_skip=0) when a pixel at coordinates (x, y) in the image data # 100 belongs to a macro block to which skip mode is not applied.
  • Therefore, the adaptive filter 100 performs filtering having the edge-stress effect when a pixel at coordinates (x, y) in the image data # 100 belongs to a macro block to which skip mode is applied, whereas the adaptive filter 100 performs filtering having no edge-stress effect, when a pixel at coordinates (x, y) in the image data # 100 belongs to a macro block to which skip mode is not applied.
  • In general, there is a tendency that edge-stressing is preferable for an image in the macro block to which the skip mode is applied, compared with the macro block to which the skip mode is not applied.
  • Therefore, the use of the adaptive filter 100 with the above configuration makes it possible to perform filtering in such a way that the edge stressing is applied for the macro block to which the skip mode is applied, but it is avoided to apply excessive edge stressing to the macro block to which the skip mode is not applied. Thus, the use of the adaptive filter 100 with the above configuration makes it possible to perform more effective filtering, compared with the conventional configuration in which the filter coefficient is not switched over.
  • Depending on sequences, there are some cases that blurring is preferably in the skip mode. In such a case, the adaptive filter 100 with the above configuration can perform optimal blurring processing by determining a_skip as a negative value.
  • Moreover, the adaptive filter 100 can switch over the filter coefficient vector V by setting the value of the control variable v_skip to 1 or 0. That is, the adaptive filter 100 can switch over the filter coefficient per macro block without calculating out the filter coefficient per macro block. This makes it possible to perform effective filtering while reducing the calculation load necessary to calculate out the filter coefficient.
  • In case where a plurality of macro blocks are including in the filter target area T, skip information assigned to one of the macro blocks may be used.
  • As described above, the image filter (adaptive filter 100) according to the present embodiment is an image filter for calculating a pixel value of a target pixel in an output image (output image data #110 a) from a pixel value of each of a pixel or pixels belonging to a target area (filter reference area R) in an input image (image data #100) and from a filter coefficient vector (filter coefficient vector V), the image filter comprising: filter coefficient vector changing means (control variable determining section 120) for changing the filter coefficient vector according to at least either where the target area is positioned in the input image or where the target pixel is positioned in the output image.
  • With this configuration, in which the filter coefficient vector changing means (control variable determining section 120) for changing the filter coefficient vector according to at least either where the target area is positioned in the input image or where the target pixel is positioned in the output image is provided, it is possible to perform filtering as appropriate for individual target areas according to at least one of the position of the target area and the position of the target area in the output image.
  • Moreover, the image filter (adaptive filter 100) according to the present embodiment is configured such that the filter coefficient vector (filter coefficient vector V) can be divided into a sum of fixed filter coefficient vectors and a variable filter coefficient vector, wherein the fixed filter coefficient vectors (filter coefficient vectors calculated out from the fixed coefficient vector W) are predetermined per unit area constituting the output image, and the variable filter coefficient vector is changed by the filter coefficient vector changing means (control variable determining section 120) according to at least either where the target area (filter target area T) is positioned in the input image (image data #100) or where the target pixel is positioned in the output image.
  • With this configuration, the filter coefficient vector can be divided into the sum of the fixed filter coefficient vectors and the variable filter coefficient vector, wherein the fixed filter coefficient vectors are predetermined per unit area constituting the output image, and the filter coefficient vector changing means can change the variable filter coefficient vector according to at least either where the target area is positioned in the input image or where the target pixel is positioned in the output image.
  • Therefore, with this configuration, it is possible to perform filtering as appropriate for individual target areas (i.e. filtering suitable for features of image in the individual target areas) by using a filter coefficient predetermined per unit area.
  • If the present invention is applied to H.264, the unit area may be a slice, a macro block, or a block.
  • Moreover, the fixed filter coefficient vector and the variable filter coefficient vectors may be defined to include an offset component.
  • Moreover, the image filter (adaptive filter 100) according to the present embodiment is further configured such that the filter coefficient vector can be divided into a fixed coefficient, a variable coefficient, and a fixed vector, wherein the fixed coefficient (for example, a_skip) is predetermined per unit area constituting the output image, the variable coefficient (control variable, for example, v_skip) is changed by the filter coefficient vector changing means according to either where the target area (filter target area T) is positioned in the input image (image data #100) or where the target pixel is positioned in the output image, and the fixed vector (filter base, for example, b_skip is predetermined.
  • With this configuration, the variable filter coefficient vector is further configured that the variable filter coefficient vector can be divided into the fixed coefficient determined per unit area constituting the output image, the variable coefficient, and the fixed vector. Moreover, the fixed vector is predetermined and the filter coefficient vector changing means can change the variable coefficient according to either where the target area (filter target area T) is positioned in the input image (image data #100) or where the target pixel is positioned in the output image. Therefore, with this configuration, the filtering can be performed more appropriately for the feature of the image in the target area.
  • Moreover, with this configuration, it is possible to perform adaptive filtering of different types as desired for the individual target area by supplying the image filter with the fixed coefficient vector and the fixed coefficient determined per unit area constituting the output image.
  • The effect of the above configuration according to the present invention is explained below, referring to FIG. 2.
  • FIG. 2 is a view illustrating a case where a Q number (Q≧2) of target areas PR1 to PRQ are present on an unit area UR.
  • Consider a case where the Q number of target areas PR1 to PRQ on the unit area UR as illustrated in FIG. 2 is to be filtered with the above configuration. Assuming that the number of the component of the fixed coefficient vector W is P (P≧2), and the number of the fixed coefficient is 1, this configuration can make it possible to perform the filtering adaptively for all the individual Q number of target areas PR1 to PRQ by using the P+1 number of filter coefficients in total.
  • On the other hand, according to the conventional method in which the fixed coefficient vector W is calculated out for every target areas PRn (1≦n≦Q) in case a Q number of target areas PR1 to PRQ are present on an unit area UR as illustrated in FIG. 2, a P×Q number of filter coefficients are necessary for performing such adaptive filtering.
  • Therefore, according to the configuration according to the present invention, it is possible to perform different types filtering adaptively for the individual target areas by using a smaller number of filter coefficients, compared with a configuration in which the filter coefficient vector is calculated out for every target areas individually.
  • Note that the Q number of target areas PR1 to PRQ overlap with one another in general, even though FIG. 2 illustrates them as areas not overlapping each other.
  • While the above explanation describes the feature of the present invention by using the vector expression for the filter coefficients, the present invention can be expresses as operations on the components in the filter coefficient.
  • For instance, consider a case of a 3×3 tapped filtering wherein b_skip is horizontal edge stress base (0, 0, 0, −1, 2, −1, 0, 0, 0, 0). In this case, the components in the filter coefficient vector V can be expressed as below (see Equation (6)).

  • (w(1),w(2),w(3),w(4)-a_skip×v_skip,w(5)+2×a_skip×v_skip,w(6)-a_skip×v_skip,w(7),w(8),w(9),w(10))
  • As understood from the above equation, the 1st to 3rd, and 7th to 10th components in the filter coefficient vector V are not dependent on the control variable v_skip when b_skip is horizontal edge stress base (0, 0, 0, −1, 2, −1, 0, 0, 0, 0). That is, regardless of whether in the skip mode or not, the 1st to 3rd, and 7th to 10th components in the filter coefficient vector V are constant values predetermined by the statistical approach.
  • On the other hand, when b_skip is horizontal edge stress base (0, 0, 0, −1, 2, −1, 0, 0, 0, 0), the 4th to 6th components in the filter coefficient vector V are dependent on the control variable v_skip. Moreover, how much the 4th to 6th components in the filter coefficient vector V are changed in with respect to the control variable v_skip is specified by the filter base b_skip.
  • In other words, when b_skip is horizontal edge stress base (0, 0, 0, −1, 2, −1, 0, 0, 0, 0), certain components (4th to 6th components) in the filter coefficient vector V are changed according to the value the control variable v_skip, and the other components (1st to 3rd, and 7th to 10th components) other than the certain components are determined independent of the value of the control variable v_skip.
  • Therefore, in general, which one of the components of the filter coefficient vector V is changed depending on the control variable and which one of the filter coefficient vector V is changed independently of the control variable are determined according to the filter base outputted from the control variable determining section 120. Moreover, how much the component dependent on the control variable is changed with respect to the control variable is determined by the filter base outputted from the control variable determining section 120.
  • <Filter Coefficient Derivation Process>
  • Moreover, a filter coefficient derivation process of Equation (6) can be obtained separately for a component changing according to the control variable and for a component not changed according to the control variable. In the following, the filter coefficient derivation process performed by the adaptive filter 100 is explained referring to FIG. 3, based on an example in which b_skip is horizontal edge stress base (0, 0, 0, −1, 2, −1, 0, 0, 0, 0).
  • FIG. 3 is a view illustrating one example of an operation flow for the filtering process.
  • (Preliminary Step)
  • In advance, the adaptive filter 100 calculates out w′ (i) by using Equation (6) with the filter coefficient vector changing according to the control variable, and keeps w′ (i) thus calculated out.
  • w ( 4 ) = w ( 4 ) + a_skip × v_skip × b_skip ( 4 ) = w ( 4 ) + a_skip × v_skip × - 1 w ( 5 ) = w ( 5 ) + a_skip × v_skip × 2 w ( 6 ) = w ( 6 ) + a_skip × v_skip × - 1
  • (Filtering Step)
  • (Step S101) Loop 1: for every filter target area T, processes S102 to S109 are performed for the whole unit area constituting the output image.
  • (Step S102) Calculate out a control variable v_skip for a filter target area T.
  • (Step S103) Loop 2: S104 to S108 are performed until a loop variable i becomes 10 from 1.
  • (Step S104) If the component v(i) of the filter coefficient vector for the loop variable i is not changed according to the control variable, the process goes to S105. That is, when i=1, 2, 3, 7, 8, 9, 10, the process goes to S105. Otherwise, the process goes to S106.
  • (Step S105) Determine that v(i)=w(i). The process then goes to S108.
  • (Step S106) If the control variable v_skip=0, goes to S105. If the control variable v_skip=1, goes to S107.
  • (Step S107) Determine that v (i)=w′(i). Then the process goes to S108.
  • (Step S108) End loop 2.
  • (Step S109) Perform filtering expressed as Equation (4), by using the filter coefficient vector V thus derived.
  • (Step S110) End Loop 1
  • In the above the derivation of the filter coefficient vector V is carried out by branching, but not by using the summation with the component changing according to the control variable. The process using the summation and the process using branching seem to be different processes, but both the processes produce the same result, and the present invention encompasses both the processes. The control variable for use in the branching is not limited to the above, and may be any control variable including the following example.
  • FIG. 4 is a view illustrating another example of the operation flow for the filtering. FIG. 4 illustrates a case where the calculation of the filter coefficient vector V is carried out by using the summation with the component changing according to the control variable. In this example, the processes in S103 to S107 in FIG. 3 are replaced with the following step S154 using Equation (6).
  • (Step S154) Calculate out the element v(i) of the filter coefficient vector from Equation (6) by using the control variable v_skip.
  • It may be preferable that (i) before the filtering, the values of the filter coefficient has been stored in a 2-dimensional table v2 (i, v) predetermined according to the positions of the components of the filter coefficient and the values of the control variable, the values of the filter coefficient being stored in the 2-dimensional table v2 (i, v) in such a way that the values of the filter coefficient are associated with the positions of the components of the filter coefficient and the values of the control variable, and (ii) the filtering is performed with reference to the 2-dimensional table v2 (i, v). Here, i is a position of the component of the filter coefficient, and v is a value of the control variable. Note that, if the number of the control variable is 1, the table is a 2-dimensional table, and if the number of the control variable is Nv, the table is a Nv+1 dimensional table.
  • So far, the expression for the function for the components of the filter coefficient vector V according to the present invention is explained.
  • <General Example of the Filter Calculating Function>
  • Equation (6) is one example of the filter calculating function F. Where the fixed coefficient vector wi (1≦i≦Nw, where Nw is the number of dimensions for the fixed coefficient vector), the control variable vk (1≦k≦Nv, where Nv is a total number of the control variables), the fixed coefficient akj (k is as described above, 1≦j≦Nak, where Nak is a total number of the fixed coefficients regarding a control variable Vk, the function F is generally expressed as in Equation (8).

  • F(w1, . . . ,wNw,a11, . . . ,a1 Na1, . . . ,aNv1, . . . ,aNv NaNv,v1, . . . ,vNv)=F1(w1, . . . ,wNw)+F2(Na1, . . . ,aNv1, . . . ,aNv NaNv,v1, . . . ,vNv)  (8)
  • Here, the filter calculating function F has a structure that can be divided into the fixed filter coefficient vector in the 1st term and the variable filter coefficient vector in the 2nd term. F1 is a function of the fixed coefficient vector and F2 is a function of the control variable and the fixed coefficient. Here, the control variable and fixed coefficient may be plural. If the control variable is plural, the number of the fixed coefficients may be changed according to the control variable.
  • Assuming that all the functions F are linear operations, they can be expressed as Equation (9).
  • [ Math . 7 ] V = i = 1 Nw wi × b_wi + k = 1 Nv j = 1 Nak akj × vk × b_vkj ( 9 )
  • Here, b_wi is a Nw number of bases regarding wi, and may be a unit base, symmetric base, or the like. In the present Specification, b_wi is a unit base, but the present invention is not limited to this. Further, b_vkj is a Nak number of bases prepared for each of the control variables vk, and may be the edge-stress filter baser or the like. In case the b_wi is a symmetric base, it is preferable that b_vkj is a base having similar symmetry. Moreover, the left-hand side of Equation (9) can be divided into the fixed filter coefficient vector in the 1st term and the variable filter coefficient vector in the 2nd term as shown in its right-hand side. Moreover, the fixed filter coefficient vector can be divided into the fixed coefficient vector (element wi) and base (b_wi). The variable filter coefficient vector can be divided into the fixed coefficient (akj), the control variable (vk), and the fixed vector (b_vkj).
  • One concrete example where the filter is 4-dimensional filter is shown in Equation (10).
  • [ Math . 8 ] ( v 1 v 2 v 3 v 4 ) = w 1 ( 0 1 0 0 ) + w 2 ( 1 0 1 0 ) + w 3 ( 0 0 0 1 ) + a_skip 1 × v_skip × ( - 1 2 - 1 0 ) + a_skip 2 × v_skip × ( 0 0 0 1 ) + a_flat × v_flat × ( 0 0 0 1 ) ( 10 )
  • In this example, the number of the fixed filter coefficients is 3 (w1, w2, w3), and the number of the fixed coefficients corresponding to the control variable v_skip is 2 (a_skip 1 and a_skip 2), and the number of the fixed coefficient corresponding to the control variable v_flat is 1 (a_flat). The meaning of the control variable v_flat will be later explained.
  • <General Example of the Fixed Filter Parameter Calculation>
  • As explained above, the fixed filter parameter calculation is carried out by solving a normal equation obtained by zeroing an equation obtained by partially differentiating an equation obtained by transforming the filter parameter V of Equation (7) into a function of the filter parameter by using the filter calculating function F.
  • The following describes an example in which another base is used as the base regarding the control variable, and an example in which another variable is used as the control variable, and effects of these examples.
  • Note that the present invention is not limited to the above Example described based on the case where the filter base b_skip is an edge-stress filter base. For example, the control variable determining section 120 may be configured to output to the filter section 110 a plurality of filter bases as the filter bases associated with the skip information, the plurality of filter bases being having different effects.
  • For example, the control variable determining section 120 may be configured to output to the filter section 110 an edge-stress base b_skip 0 and DC offset base b_dc=(0, 0, 0, 0, 0, 0, 0, 0, 0, 1) as the filter bases associated with the skip information.
  • In case where the control variable determining section 120 outputs the edge-stress base b_skip and the DC offset base b_dc, the filter section 110 determines, with reference to the teacher data # 200, the coefficients of the filter coefficient vector V having the components (K) expressed by the following Equation (11).

  • v(K)=w(K)+a_skip×v_skip×b_skip(K)+a dc×v dc×b dc(K)  (11)
  • where a_dc is the DC offset coefficient (fixed coefficient) corresponding to the DC offset base.
  • Per slice or per frame, the filter section 110 determines the components w(K) of the fixed coefficient vector W, the edge-stress coefficient a_skip, and the DC offset coefficient a_dc, for example by the aforementioned statistical approach.
  • For a pixel belonging to a macro block in which the skip mode is applied, the control variable determining section 120 outputs the control variable v_skip of 1 to the filter section 110. For a pixel belonging to a macro block in which the skip mode is not applied, the control variable determining section 120 outputs the control variable v_skip of 0 to the filter section 110.
  • The filtering section 110 performs filtering based on the filter coefficient vector V using the control variable v_skip supplied from the control variable determining section 120.
  • For the macro block in which the skip mode is applied, the adaptive filter 100 with such an arrangement can perform edge-stressing filtering and filtering for changing luminance by the DC offset.
  • It is also possible to perform adaptive filtering for pixel values of pixels whose luminance change individually, such as fading, by using the DC offset base and a center element base b_c=(0, 0, 0, 0, 1, 0, 0, 0, 0, 0) in which only center element is 1.
  • Moreover, the control variable determining section 120 may output to the filter section 110 another edge-stress base b_skip 1 together with a certain edge-stress base b_skip0, as the filter base associated with the skip information. For example, it may be such that b_skip0 is a horizontal edge-stress base (0, 0, 0, −1, 2, −1, 0, 0, 0, 0), and b_skip1=(0, −1, 0, 0, 2, 0, −1, 0, 0, 0). Moreover, as an edge stress/blurring base, a base mainly for blurring such as a widthwise blurring base (0, 0, 0, 1, 2, 1, 0, 0, 0, 0), a lengthwise blurring base (0, 1, 0, 0, 2, 0, 0, 1, 0, 0), or a omnidirectional blurring base (1, 1, 1, 1, 2, 1, 1, 1, 1, 0) may be used.
  • In case where the control variable determining section 120 outputs the edge-stress base b_skip0 and b_skip 1, the filter section 110 determines, with reference to the teacher data # 200, the coefficients in the filter coefficient vector V having the components v(K) expressed by the following Equation (12).

  • v(K)=w(K)+a_skip0×v_skip×b_skip0(K)+a_skip1×v_skip×b_skip1(K)  (12)
  • Here, a_skip0 is an edge-stress coefficient (fixed coefficient) corresponding to the edge-stress base b_skip0(K), and a_skip1 is an edge-stress coefficient (fixed coefficient) corresponding to the edge-stress base b_skip1(K).
  • The filter section 110 determines, for example by the aforementioned approach, the components w(K) of the fixed coefficient vector, and the edge-stress coefficients a_skip0 and a_skip1 per slice or per frame.
  • For a pixel belonging to a macro block to which the skip mode is applied, the control variable determining section 120 outputs the control variable v_skip whose value is 1. For a pixel belonging to a macro block to which the skip mode is not applied, the control variable determining section 120 outputs the control variable v_skip whose value is 0.
  • The filter section 110 performs the filtering based on the filter coefficient vector V using the control variable v_skip supplied from the control variable determining section 120.
  • With this configuration, the adaptive filter 100 can perform the filter coefficient-based filtering having both the edge stress effect and the blurring effect for the macro block to which the skip-mode is applied. For the macro block to which the skip-mode is not applied, the adaptive filter 100 can perform the filter coefficient-based filtering having none of the edge stress effect and the blurring effect.
  • Moreover, the present invention is not limited to the above example in which the adaptive filter 100 switches over the filter coefficient vector V depending on whether or not the pixel included in the image data # 100 belongs to a macro block to which the skip mode is applied.
  • In the following, the filter coefficient is switched over based on the image data # 100 and the side information #100 a.
  • (Example in which the Filter Coefficient is Switched Over Depending on Degree of Flatness in the Image)
  • Firstly, an example in which the filter coefficient is switched over depending on degree of flatness in the filter target area T in the image is described here. The degree of flatness in the filter target area T is an index that is set according to how much the pixel values of the pixels included in the filter target area T are uneven. When the pixel values of the pixels included in the filter target area T are less uneven, the degree of flatness of the image in the area is higher. When the pixel values of the pixels included in the filter target area T are more uneven, the degree of flatness of the image in the area is lower. In order to determine the control variable for the filter target area T, it is not necessary to use all the pixels in the filter target area T (thereinafter, the same is applied). The degree of flatness may be determined from a rectangular area MV x NV, which is smaller than the area T.
  • The control variable determining section 120 sets a value of a control variable v_flat to 1 and outputs the control variable v_flat to the filter section 110, wherein the control variable v_flat is set to a value of 0 or 1 depending on the degree of flatness in the filter target area T.
  • Moreover, the control variable determining section 120 outputs, to the filter section 110, the filter base b_flat associated with the degree of flatness in the filter target area T. The filter base b_flat may be, for example, the aforementioned edge-stress filter base, 5×5 tap-expanded Laplacian base (1, 1, 1, 1, 1, 1, 0, 0, 0, 1, 1, 0, −16, 0, 1, 1, 0, 0, 0, 1, 1, 1, 1, 1, 1, 0), or a linear sum (or weighted linear sum) of them.
  • The filter section 110 determines, with reference to the teacher data # 200, the coefficients of the filter coefficient vector V having the components v(K) expressed by the following Equation (13).

  • v(K)=w(K)+a_flat×v_flat×b_flat(K)  (13)
  • where a_flat is a coefficient (fixed coefficient) corresponding to the filter base b_flat.
  • The filter section 110 determines the components w(K) of the filter coefficient vector and the coefficient a_flat per slice or per frame by for example the aforementioned statistical approach.
  • The control variable determining section 120 calculates out the unevenness of the pixel values of the pixels in the filter target area T for every pixel included in the image data # 100. More specifically, for example, the control variable determining section 120 calculates out dispersion of the pixel values of the pixels in the filter target area T per pixel included in the image data # 100.
  • In other words, the control variable determining section 120 refers to the pixel values of the pixels in the image data # 100 individually, and calculates out the unevenness of the pixel value of each pixel in the filter target area T. More specifically, the control variable determining section 120 refers to the pixel values of the pixels in the image data # 100 individually, and calculates out the dispersion of the pixel value of each pixel in the filter target area T.
  • The control variable determining section 120 sets the control variable v_flat to 0 for a pixel whose pixel value has dispersion equal to or greater than a predetermined threshold. The control variable determining section 120 sets the control variable v_flat to 1 for a pixel whose pixel value has dispersion less than the predetermined threshold. Then, the control variable determining section 120 outputs the control variables v_flat to the filter section 110 per pixel in the image data # 100 individually.
  • Moreover, the control variable determining section 120 may be configured to output the control variable v_flat to the filter section 110 per filter target area T.
  • The filter section 110 performs the filtering based on the filter variable vector V using the control variable supplied from the control variable determining section 120.
  • It is known that, generally, different filter coefficients are optimal for areas having different degrees of flatness.
  • With this configuration, the adaptive filter 100 can perform filtering having a blurring effect for a filter target area T having a large degree of flatness, whereas the adaptive filter 100 can perform filtering having no blurring effect for a filter target area T having a small degree of flatness. Thus, the adaptive filter 100 can perform effective filtering adaptively for slices or frame having different degrees of flatness.
  • Even in case where the switching-over of the filter coefficient is performed according to the degree of flatness of image, the use of the filter base and filter coefficient corresponding to the DC offset makes it possible to perform filtering of luminance adjustment using the DC offset.
  • Moreover, in the present example, it may be configured such that the degree of flatness in the image in the filter target area T is judged according to how large an AC component associated with the block to which the pixel at the coordinates (x, y) in the image data #100.
  • Moreover, one specific example of the filter coefficient calculation process performed by the adaptive filter 100 in the present operation example is as follows. For example, assuming that the filter base b_flat is 3×3 horizontal edge-stress filter base, w′(i) in the (preliminary step) of the aforementioned <Filter Coefficient Calculation Process> may be calculated out by the following Equation, and stored.
  • w ( 4 ) = w ( 4 ) + a_flat × v_flat × b_flat ( 4 ) = w ( 4 ) + a_flat × v_flat × - 1 w ( 5 ) = w ( 5 ) + a_flat × v_flat × 2 w ( 6 ) = w ( 6 ) + a_flat × v_flat × - 1
  • For the other steps in the <Filter Coefficient Calculation Process>, the control variable v_skip should be read as the control variable v_flat.
  • (Example in which the Filter Coefficient is Switched Over According to Edge Angle)
  • Next, an example in which the filter coefficient is switched over according to edge angle of an edge included in the image in the filter target area T of the image data # 100 is described.
  • The control variable determining section 120 the following control variables to 1 and outputs the following control variables to the filter section 110.
      • control variable v_edge_0, being 0 or 1 according to whether or not the filter target area T has a vertical edge,
      • control variable v_edge_90, being 0 or 1 according to whether or not the filter target area T has a horizontal edge,
      • control variable v_edge_45, being 0 or 1 according to whether or not the filter target area T has a right-up diagonal edge, and
      • control variable v_edge_135, being 0 or 1 according to whether or not the filter target area T has a right-down diagonal edge.
  • Moreover, the control variable determining section 120 outputs the following filter bases to the filter section 12.
      • filter base b_edge_0 associated with information regarding whether or not the filter target area T has a vertical edge,
      • filter base b_edge_90 associated with information regarding whether or not the filter target area T has a horizontal edge,
      • filter base b_edge_45 associated with information regarding whether or not the filter target area T has a right-up diagonal edge,
      • filter base b_edge_135 associated with information regarding whether or not the filter target area T has a right-down diagonal edge.
  • The filter base b_edge_n (N=0, 45, 90, 135) is as follows, for example in the case of 3×3 tapped filtering.
  • b_edge_0=(0, −1, 0, 0, 2, 0, 0, −1, 0, 0)
  • b_edge_90=(0, 0, 0, −1, 2, −1, 0, 0, 0, 0)
  • b_edge_45=(0, 0, −1, 0, 2, 0, −1, 0, 0, 0)
  • b_edge_135=(−1, 0, 0, 0, 2, 0, 0, 0, −1, 0)
  • The filter section 110 determines, with reference to the teacher data # 200, the coefficients of the filter coefficient vectors V having the components v (K) expressed by Equation (14) below.

  • v(K)=w(K)+Σa_edge n×v_edge n×b_edge n(K)  (14)
  • where the sign indicates sum of the suffix n (n=0, 45, 90, 135) indicating the edge angle, and a_edge_n (n=0, 45, 90, 135) is a coefficient corresponding to the filter base b_edge_n (n=0, 45, 90, 135), respectively.
  • The filter section 110 determines the components w (K) of the fixed coefficient vector, coefficient a_edge_n (n=0, 45, 90, 135) per slice or per frame, for example by the aforementioned statistical approach.
  • The control variable determining section 120 sets the value of the control variable v_edge_n (n=0, 45, 90, 135) depending on which kind of the edge the filter target area T contains, and outputs the control variable v_edge_n (n=0, 45, 90, 135) to the output section 110 for every filter target area T.
      • v_edge_0=1, if the filter target area T contains a vertical edge, and v_edge_0=0, if not,
      • v_edge_90=1, if the filter target area T contains a horizontal edge, v_edge_90=0, if not,
      • v_edge_45=1, if the filter target area T contains a right-up diagonal edge, v_edge_45=0, if not,
      • v_edge_135=1, if the filter target area T contains a right-down diagonal edge, v_edge_135=0, if not.
  • The filter section 110 performs the filtering based on the filter coefficient vector V using the control variable v_edge_n (n=0, 45, 90, 135) supplied from the control variable determining section 120.
  • With this configuration, the adaptive filer 100 can perform filtering using different filter coefficients depending on the edge angle including in the target area T. Thus, it is possible to perform effective filtering for slices or frames having various edge angles.
  • The adaptive filter 100 may be configured such that the control variable determining section 120 outputs the following filter base to the filter section 110.
      • the first filter base b_edge_0_0=b_edge_0, where the first filter base b_edge_0_0 is associated with the information regarding whether the filter target area T contains a vertical edge,
      • the second filter base b_edge_0_1=b_edge_90, where the second filter base b_edge_0_1 is associated with the information regarding whether the filter target area T contains a vertical edge,
      • the first filter base b_edge_90_0=b_edge_90, where the first filter base b_edge_90_0 is associated with the information regarding whether the filter target area T contains a horizontal edge,
      • the second filter base b_edge_90_1=b_edge_0, where the second filter base b_edge_90_1 is associated with the information regarding whether the filter target area T contains a horizontal edge,
      • the first filter base b_edge_45_0=b_edge_45, where the first filter base b_edge_45_0 is associated with the information regarding whether the filter target area T contains a right-up diagonal edge,
      • the second filter base b_edge_45_1=b_edge_135, where the second filter base b_edge_45_1 is associated with the information regarding whether the filter target area T contains a right-up diagonal edge,
      • the first filter base b_edge_135_0=b_edge_135, where the first filter base b_edge_135_0 is associated with the information regarding whether the filter target area T contains a right-down diagonal edge,
      • the second filter base b_edge_135_1=b_edge_45, where the second filter base b_edge_135_1 is associated with the information regarding whether the filter target area T contains a right-down diagonal edge.
  • In this case, the filter section 110 determines, with reference to the teacher data # 200, the coefficients of the filter coefficient vector V having the components v(K) expressed by the following Equation (15).

  • v(K)=w(K)+Σa_edge n 0×v_edge n×b_edge n 0(K)+Σa_edge n 1×v_edge n×b_edge n 1(K)  (15)
  • where the sign Σ indicates sum of the suffix n (n=0, 45, 90, 135) indicating the edge angle, and a_edge_n_0 (n=0, 45, 90, 135) and a_edge_n_1 (n=0, 45, 90, 135) are coefficients corresponding to the filter base b_edge_n_0 (n=0, 45, 90, 135) and the filter base b_edge_n_1 (n=0, 45, 90, 135), respectively.
  • The filter section 110 determines the component w(K) of the fixed coefficient vector, the coefficient a_edge_n_0 (n=0, 45, 90, 135), and coefficient a_edge_n_1 (n=0, 45, 90, 135) per slice or per frame, for example, the aforementioned statistical approach.
  • The values of the control variables v_edge_n (n=0, 45, 90, 135) of the control variable determining section 120 are set by the same method described above. Moreover, the filter section 110 performs the filtering base on the filer coefficient vector V using the control variable v_edge_n (n=0, 45, 90, 135) supplied from the control variable determining section 120.
  • With this configuration, the adaptive filter 100 can perform the filtering by using different filter coefficient depending on the edge angle of edges in the target area T. Thus, it is possible to perform effective filtering for slices or frames having various edge angles.
  • Moreover, with this configuration, the filter base b_edge_n_0 (n=0, 45, 90, 135), and the filter base b_edge_n_1 (n=0, 45, 90, 135) each have base a component having an edge-stress effect to a direction orthogonal to the edge, the adaptive filter 100 having the above configuration can perform effective filtering not only for a direction parallel to the edge, but also for a direction orthogonal to the edge.
  • Again in case where the filtering coefficients are switched over according to the edge angle, it is also possible to perform filtering having luminance change due to the DC offset by using the filter base and filter coefficient, each of which corresponds to the DC offset.
  • Moreover, one specific example of the filter coefficient calculation process performed by the adaptive filter 100 in the present operation example is as follows. For example, assuming that the filter base b_edge_n is b_edge_90, w′(i) in the (preliminary step) of the aforementioned <Filter Coefficient Calculation Process> may be calculated out by the following Equation, and stored.

  • w′(4)=w(4)+a_edge90×v_edge90×b_edge90(4)=w(4)+a_edge90×v_edge90×−1

  • w′(5)=w(5)+a_edge90×v_edge90×2

  • w′(6)=w(6)+a_edge90×v_edge90×−1
  • For the other steps in the <Filter Coefficient Calculation Process>, the control variable v_skip should be read as the control variable v_edge_90. Moreover, the equation should be read in the similar manner in case where the filter base b_edge_n is another base. Moreover, Equation (14) can be easily generalized in case n is summed.
  • (Example in which the Filter Coefficient is Switched Over According to Luminance)
  • Next, an example in which the filter coefficient is switched over according to an average of luminance in the image in the filter target area T in the image data # 100 is described. Here, the average of luminance in the filter target area T is an average of luminance values of pixels included in the filter target area T. Moreover, the luminance values of the pixels can be obtained as a function of the pixel values of the pixels.
  • The control variable determining section 120 sets a control variable v_lum to 0, 1, or 2 according to the average of luminance in the image in the filter target area T, and outputs the control variable v_lum to the filter section 110.
  • Moreover, the control variable determining section 120 outputs, to the filter section 110, the filter base b_lum associated with the average luminance of the image in the filter target area T. The filter base b_lum may be, for example, the edge-stress filter base or its linear sum (or its weighted linear sum).
  • The filter section 110 determines, with reference to the teacher data # 200, the coefficients of the filter coefficient vector V having the components v (K) expressed by Equation (16).

  • v(K)=w(K)+a lum×v lum×b lum(K)  (16)
  • where a_lum is a coefficient (fixed coefficient) corresponding to the filter base b_lum.
  • The filter section 110 determines the components w(K) of the fixed coefficient vector and the coefficient a_lum per slice or per frame, for example, by the aforementioned statistical approach.
  • The control variable determining section 120 calculates out the average of the luminance values associated with the pixels included in the area, for every filter target area T. The control variable determining section 120 sets the value of the control variable v_lum to 2, for example when the average is less than 48. The control variable determining section 120 sets the value of the control variable v_lum to 1, for example when the average is not less than 48 but less than 96. The control variable determining section 120 sets the value of the control variable v_lum to 0, for example when the average is not less than 96. Here, it is assumed that the average can be in a range of 0 to 255.
  • Moreover, the control variable determining section 120 outputs the control variable v_lum to the filter section 110 per filter target area T.
  • The filter section 110 performs the filtering based on the filter coefficient vector V using the control variable v_lum supplied from the control variable determining section 120 for every filter target area T.
  • With this configuration, it is possible to perform filtering having a more intensive blurring when the average of the luminance values of the pixels included in the filter target area T is low, whereas it is possible to perform filtering having a weaker blurring when the average of the luminance values of the pixels included in the filter target area T is high.
  • Again in case where the filtering coefficients are switched over according to the luminance, it is also possible to perform filtering having luminance change due to the DC offset by using the filter base and filter coefficient, each of which corresponds to the DC offset.
  • Moreover, one specific example of the filter coefficient calculation process performed by the adaptive filter 100 in the present operation example is as follows. For example, assuming that the filter base b_lum is a widthwise direction edge-stress base, w′(i) in the (preliminary step) of the aforementioned <Filter Coefficient Calculation Process> may be calculated out by the following Equation, and stored.
  • w ( 4 ) = w ( 4 ) + a_lum × v_lum × b_lum ( 4 ) = w ( 4 ) + a_lum × v_lum × - 1 w ( 5 ) = w ( 5 ) + a_lum × v_lum × 2 w ( 6 ) = w ( 6 ) + a_lum × v_lum × - 1
  • For the other steps in the <Filter Coefficient Calculation Process>, the control variable v_skip should be read as the control variable v_lum. Moreover, the equation should be read in the similar manner in case where the filter base b_lum is another base.
  • (Example in which the Filtering Coefficient is Switched Over According to Color)
  • Next, an example in which the filter coefficient is switched over according to colors in the images in the filter target area T in the image data # 100 is described. Especially, an example whether or not the filter coefficient is switched over according to whether or not the colors in the pixels in the filter target area T are skin colors or not. Here, the case where the colors in the pixels in the filter target area T are skin colors is for example, an average of the colors included in the filter target area T is 133≦V≦177, and 77≦U≦127, and 60≦Y≦145 according YUV component expression. Here, Y indicates a size of the brightness component, U is a size of a color component defined by a product of a constant and a difference between a blue component and the luminance component, and V is a size of a color component defined by a product of a constant and a difference between a red component and the luminance component. In the above example, Y, U, and V are in a range of 0 to 255.
  • The control variable determining section 120 sets the control variable v_skin to 1 according to whether the pixel at coordinates (x, y) in the image data # 100 is in a skin color, wherein the control variable v_skin can be 0 or 1. The control variable determining section 120 outputs the control variable v_skin to the filter section 110.
  • The control variable determining section 120 outputs, to the filter section 110, the filter base b_skin associated with whether or not the pixel at coordinates (x, y) in the image data # 100 is in a skin color. The filter base b_skin may be, for example, the aforementioned edge-stress filter base or its linear sum (or weighted linear sum).
  • The filter section 110 determines, with reference to the teach data # 200, the coefficients of the filter coefficient vector V having the components v(K) expressed by Equation (17).

  • v(K)=w(K)+a_skin×v_skin×b_skin(K)  (17)
  • where a_skin is a coefficient (fixed coefficient) corresponding to the filter base b_skin.
  • The filter section 110 determines the components v(K) of the filter coefficient vector and the coefficient a_skin per slice or per frame, for example, by the aforementioned statistical approach.
  • The control variable determining section 120 determines whether or not the pixel at coordinate (x, y) in the image data # 100 is in the skin color. More specifically, the control variable determining section 120 determines whether or not the color of the pixel at the coordinates (x, y) in the image data # 100 is within the following ranges expressed by the YUV components: 133≦V≦177, 77≦U≦127, and 60≦Y≦145.
  • Moreover, if the color of the pixel at the coordinates (x, y) in the image data # 100 is within the ranges, the control variable v_skin is set to 1, and outputs the control variable v_skin to the filter section 110. On the other hand, if the color of the pixel at the coordinates (x, y) in the image data # 100 is not within the ranges, the control variable v_skin is set to 0, and outputs the control variable v_skin to the filter section 110.
  • The filter section 110 performs the filtering per filter target area T based on the filter coefficient vector V using the control variable v_skin supplied from the control variable determining section 120.
  • With this configuration, when the color of the pixel at the coordinates (x, y) in the image data # 100 is within the ranges, it is possible to perform filtering having blurring effect, whereas if the color of the pixel at the coordinates (x, y) in the image data # 100 is not within the ranges, it is possible to perform filtering having no blurring effect.
  • Again in case where the filtering coefficients are switched over according to the color, it is also possible to perform filtering having luminance change due to the DC offset by using the filter base and filter coefficient, each of which corresponds to the DC offset.
  • Moreover, one specific example of the filter coefficient calculation process performed by the adaptive filter 100 in the present operation example is as follows. For example, assuming that the filter base b_skip_n is a widthwise direction edge-stress base, w′(i) in the (preliminary step) of the aforementioned <Filter Coefficient Calculation Process> may be calculated out by the following Equation, and stored.
  • w ( 4 ) = w ( 4 ) + a_skin × v_skin × b_skin ( 4 ) = w ( 4 ) + a_skin × v_skin × - 1 w ( 5 ) = w ( 5 ) + a_skin × v_skin × 2 w ( 6 ) = w ( 6 ) + a_skin × v_skin × - 1
  • For the other steps in the <Filter Coefficient Calculation Process>, the control variable v_skip should be read as the control variable v_skin. Moreover, the equation should be read in the similar manner in case where the filter base b_skin is another base.
  • (Example in which the Filter Coefficient is Switched Over According to Whether or not there is a Block Border)
  • Next, an example in which the filter coefficient is switched over according to whether or not there is a block border in the filter target region T is described.
  • The block is, for example, a unit of a change or a unit of motion compensation in the H.264/AC standard. In general, there is tendency that a block noise easily occurs at the block border.
  • The control variable determining section 120 sets the value of the control variable v_block to 1, wherein the value of the control variable v_block may be 0 or 1 according to whether or not the filter target area T contains a block border therein. Then, the control variable determining section 120 outputs the control variable v_block to the filter section 110.
  • The control variable determining section 120 outputs to the filter section 110 a filter base b_block associated with whether or not the filter target area T contains a block border. The filter base b_block may be, for example, the edge-stress filter base or its linear sum (or weighted linear sum).
  • The filter section 110 determines, with reference to the teacher data # 200, the filter coefficient vector V having the components v(K) expressed by the following Equation (18).

  • v(K)=w(K)+a_block×v_block×b_block(K)  (18)
  • where a_block is a coefficient (fixed coefficient) corresponding to the filter base b_block.
  • The filter section 110 determines the components w(K) of the fixed coefficient vector and the coefficient a_block per slice or per frame, for example, by the aforementioned statistical approach.
  • The control variable determining section 120 determines whether or not the filter target area T contains a block border. If the filter target area T contains a block border, the control variable determining section 120 sets the value of the control variable v_block to 1. If not, the control variable determining section 120 sets the value of control variable v_block to 0. Moreover, the control variable determining section 120 outputs the control variable v_block to the filter section 110.
  • For every filter target area T, the filter section 110 performs the filtering based on the filter coefficient vector V using the control variable v_block supplied from the control variable determining section 120.
  • With this configuration, it is possible to perform the filtering having a blurring effect for a filter target area T if it contains a block border, whereas it is possible to perform the filtering having no blurring effect for a filter target area T if it does not contain a block border.
  • Again in case where the filtering coefficients are switched over according to whether the filter target area T contains a block border, it is also possible to perform filtering having luminance change due to the DC offset by using the filter base and filter coefficient, each of which corresponds to the DC offset.
  • Moreover, one specific example of the filter coefficient calculation process performed by the adaptive filter 100 in the present operation example is as follows. For example, assuming that the filter base b_block is a widthwise direction edge-stress base, w′(i) in the (preliminary step) of the aforementioned <Filter Coefficient Calculation Process> may be calculated out by the following Equation, and stored.
  • w ( 4 ) = w ( 4 ) + a_block × v_block × b_block ( 4 ) = w ( 4 ) + a_block × v_block × - 1 w ( 5 ) = w ( 5 ) + a_block × v_block × 2 w ( 6 ) = w ( 6 ) + a_block × v_block × - 1
  • For the other steps in the <Filter Coefficient Calculation Process>, the control variable v_skip should be read as the control variable v_block. Moreover, the equation should be read in the similar manner in case where the filter base b_block is another base.
  • (Example in which the Filter Coefficient is Switched Over According to Prediction Direction)
  • Next an example in which the filter coefficient is switched over according to side information associated with a block to which a pixel at coordinates (x, y) in the image data # 100 belongs is described. Especially, an example in which the filter coefficient is switched over according to the block to which the pixel at coordinates (x, y) in the image data # 100 belongs is a block generated according to a unidirectional prediction or bidirectional prediction under the H.264/AC standard.
  • Here, the unidirectional prediction under the H.264/AC standard is a case where the prediction of a block is carried out by using one frame encoded or decoded before a frame the block belongs. The bidirectional prediction under the H.264/AC standard is a case where the prediction of a block is carried out by using two frames encoded or decoded before a frame the block belongs. The number of frames used for the prediction is one or two in general, but may be greater than one or two.
  • The control variable determining section 120 sets the value of the control variable v_pred to 1, wherein the value of the control variable v_pred is set to 0 or 1 according to whether the block to which the pixel at coordinates (x, y) in the image data # 100 is a block generated based on the unidirectional prediction or a block generated based on the bidirectional prediction. Then, the control variable determining section 120 outputs the control variable v_pred to the filter section 110.
  • Moreover, the control variable determining section 120 outputs to the filter section 110 a filter base b_pred associated with whether the block to which the pixel at coordinates (x, y) in the image data # 100 is a block generated based on the unidirectional prediction or a block generated based on the bidirectional prediction. The filer base b_pred may be for example, the edge-stress filter base or its linear sum (or weighted linear sum).
  • The filter section 110 determines, with reference to the teacher data #20, the coefficients of the filter coefficient vector V having the components v(K) expressed by the following equation (19).

  • v(K)=w(K)+a pred×v pred×b pred(K)  (19)
  • where a_pred is a coefficient (fixed coefficient) corresponding to the filter base b_pred.
  • The filter section 110 determines the components w(K) of the fixed coefficient vector and the coefficient a_pred per slice or per frame, for example by the aforementioned statistical approach.
  • The control variable determining section 120 sets the control variable v_pred to 0 if the block to which the pixel at the coordinates (x, y) in the image data # 100 is a block generated based on the unidirectional prediction, whereas The control variable determining section 120 sets the control variable v_pred to 1 if the block to which the pixel at the coordinates (x, y) in the image data # 100 is a block generated based on the bidirectional prediction. Moreover, the control variable determining section 120 outputs the control variable v_pred to the filter section 110.
  • For every filter target area T, the filter section 110 performs the filtering based on the filter coefficient vector V using the control variable v_pred supplied from the control variable determining section 120.
  • With this configuration, it is possible to perform filtering having no edge-stress or blurring effect for a block, if the block to which the pixel at the coordinates (x, y) of the image data # 100 belongs is one generated under the unidirectional prediction, whereas it is possible to perform filtering having the edge stress or blurring effect for a block, if the block to which the pixel at the coordinates (x, y) of the image data # 100 belongs is one generated under the bidirectional prediction.
  • Again in case where the filtering coefficients are switched over according to whether the unidirectional prediction or the bidirectional prediction is performed as described above, it is also possible to perform filtering having luminance change due to the DC offset by using the filter base and filter coefficient, each of which corresponds to the DC offset.
  • Moreover, in case where a plurality of blocks are contained in the filter target area T, it may be so configured that information assigned to any one of the blocks is used.
  • Moreover, one specific example of the filter coefficient calculation process performed by the adaptive filter 100 in the present operation example is as follows. For example, assuming that the filter base b_pred is a widthwise direction edge-stress base, w′(i) in the (preliminary step) of the aforementioned <Filter Coefficient Calculation Process> may be calculated out by the following Equation, and stored.
  • w ( 4 ) = w ( 4 ) + a_pred × v_pred × b_pred ( 4 ) = w ( 4 ) + a_pred × v_pred × - 1 w ( 5 ) = w ( 5 ) + a_pred × v_pred × 2 w ( 6 ) = w ( 6 ) + a_pred × v_pred × - 1
  • For the other steps in the <Filter Coefficient Calculation Process>, the control variable v_skip should be read as the control variable v_pred. Moreover, the equation should be read in the similar manner in case where the filter base b_pred is another base.
  • (Example in which the Filter Coefficient is Switched Over According to the Prediction Method)
  • Next, an example in which a block to which a pixel at coordinates (x, y) in the image data # 100 belongs is a block predicted by intra prediction under the H.264/AV standard, or a block predicted by inter prediction under the H.264/AV standard, is described below.
  • The control variable determining section 120 sets a value of a control variable v_int to 1, wherein the value of the control variable v_int is set to 0 or 1 according to whether the block to which the pixel at coordinates (x, y) in the image data # 100 belongs is a block predicted by intra prediction or a block predicted by inter prediction. Then, the control variable determining section 120 outputs the control variable v_int o the filter section 110.
  • The control variable determining section 120 outputs to the output section 110 a filter base b_int associated whether the block to which the pixel at coordinates (x, y) in the image data # 100 belongs is a block predicted by intra prediction or a block predicted by inter prediction. The filter base b_int may be, for example, the edge stress filter base or its linear sum (or weighted linear sum).
  • The filter section 110 determines, with reference to the teacher data # 200, the coefficients of the filer coefficient vector V having the components expressed by the following Equation (20).

  • v(K)=w(K)+a int×v int×b int(K)  (20)
  • where a_int is a coefficient (fixed coefficient) corresponding to the filter base b_int.
  • The filter section 110 determines the components w(K) of the fixed coefficient vector and the coefficient a_int per slice or per frame, for example by the aforementioned statistical approach.
  • If the block to which the pixel at the coordinates (x, y) in the image data # 100 belongs is one predicted by the intra prediction, the control variable determining section 120 sets the value of the control variable v_int to 1 for the block, whereas if the block to which the pixel at the coordinates (x, y) in the image data # 100 belongs is one predicted by the inter prediction, the control variable determining section 120 sets the value of the control variable v_int to 0 for the block. Then the control variable determining section 120 outputs the control variable v_into to the filter section 110.
  • For every filer target area T, the filter section 110 performs the filtering based on the filter coefficient vector V using the control variable v_pred supplied from the control variable determining section 120.
  • In general, edges in the block predicted by the intra prediction tends to be easily blurred.
  • With this configuration, it is possible to perform the filtering in such a way that if the block to which the pixel at the coordinates (x, y) in the image data # 100 is one predicted by the intra prediction, the filtering having the edge-stress effect is performed, whereas if the block to which the pixel at the coordinates (x, y) in the image data # 100 is one predicted by the inter prediction, the filtering having no edge-stress effect is performed. Therefore, with this configuration, it is possible to perform effective filtering for the block to which the pixel at the coordinates (x, y) in the image data # 100, regardless of whether the block is one predicted by the intra prediction or one predicted by the inter prediction.
  • Again in case where the filtering coefficients are switched over according to whether the inter prediction or the intra prediction is performed as described above, it is also possible to perform filtering having luminance change due to the DC offset by using the filter base and filter coefficient, each of which corresponds to the DC offset.
  • Moreover, in case where a plurality of blocks are contained in the filter target area T, it may be so configured that information assigned to any one of the blocks is used.
  • Moreover, one specific example of the filter coefficient calculation process performed by the adaptive filter 100 in the present operation example is as follows. For example, assuming that the filter base b_int is a widthwise direction edge-stress base, w′(i) in the (preliminary step) of the aforementioned <Filter Coefficient Calculation Process> may be calculated out by the following Equation, and stored.
  • w ( 4 ) = w ( 4 ) + a_int × v_int × b_int ( 4 ) = w ( 4 ) + a_int × v_int × - 1 w ( 5 ) = w ( 5 ) + a_int × v_int × 2 w ( 6 ) = w ( 6 ) + a_int × v_int × - 1
  • For the other steps in the <Filter Coefficient Calculation Process>, the control variable v_skip should be read as the control variable v_int. Moreover, the equation should be read in the similar manner in case where the filter base b_int is another base.
  • (Example in which the Filter Coefficient is Switched Over According to a Distance from a Block Border)
  • Next, an example in which the filter coefficient is switched over according to a distance between a pixel at coordinates (x, y) in the image data # 100, and a block border of a block to which the pixel belongs.
  • The control variable determining section 120 sets a value of a control variable v_dist to 1, wherein the value of the control variable v_dist is 0 or 1 according to a distance between a pixel at coordinates (x, y) in the image data # 100, and a block border of a block to which the pixel belongs. Then, the control variable determining section 120 outputs the control variable v_dist to the filter section 110.
  • Moreover, the control variable determining section 120 outputs to the filter section 110, a filter base b_dist associated with the distance between the pixel at the pixel at the coordinates (x, y) in the image data # 100, and the block border of the block to which the pixel belongs.
  • The filter section 110 determines, with reference to the teacher data # 200, the coefficients of the filter coefficient vector V having the components v(K) expressed by the following Equation (21).

  • v(K)=w(K)+a dist×v dist×b dist(K)  (21)
  • where a_dist is a coefficient (fixed coefficient) corresponding to the filter base b_dist.
  • The filter section 110 determines the components w(K) of the fixed coefficient vector and the coefficient a_dist per slice or per frame, for example by the aforementioned statistical approach.
  • The control variable determining section 120 calculates out the distance between the pixel at the coordinates (x, y) in the image data # 100, and the block border of the block to which the pixel belongs. If the distance is smaller than a predetermined first threshold, the control variable determining section 120 sets the control variable v_dist to 2. Ff the distance is not less than the first threshold but less than a second threshold greater than the first threshold, the control variable determining section 120 sets the control variable v_dist to 1. If the distance is not les than the second threshold, the control variable determining section 120 sets the control variable v_dist to 0. Moreover, the control variable determining section 120 outputs the control variable v_dist to the filter section 110.
  • For every filter target area T, the filter section 110 performs the filtering based on the filter coefficient vector V using the control variable v_dist supplied from the control variable determining section 120.
  • With this configuration, it is possible to perform three-leveled filtering having different edge-stress effect according to the distance between the pixel at the coordinates (x,y) in the image data # 100 and the block border of the block to which the pixel belongs. That is, if the distance between the pixel and the block border is smaller than the first threshold, filtering having the greatest edge-stress effect among the three-leveled filtering is performed. If the distance is not less than the first threshold but less than the second threshold, the filtering having the second greatest edge-effect among the three-leveled filtering is performed. If the distance is not less than the second threshold, the filtering having no edge-stress effect is performed.
  • Again in case where the filtering coefficients are switched over according to the distance from the block border, it is also possible to perform filtering having luminance change due to the DC offset by using the filter base and filter coefficient, each of which corresponds to the DC offset.
  • Moreover, it may be configured that the filter base is changed according to whether the block border is a vertical border or a horizontal border. That is, it may be configured such that if the block border is a vertical border, the filter coefficient b_dist is the horizontal edge-stress base, and if the block border is a horizontal border, the filter coefficient b_dist is a vertical edge-stress base.
  • Moreover, the control vertical determining section 120 may be configured to output a control variable adjusted according to a distance from a vertical block border and a control variable adjusted according to a distance from a horizontal block border, thereby being able to perform filtering having different edge-stress effect, according to the distance from the vertical block border and the distance from the horizontal block border.
  • Moreover, one specific example of the filter coefficient calculation process performed by the adaptive filter 100 in the present operation example is as follows. For example, assuming that the filter base b_dist is a widthwise direction edge-stress base, w′(i) in the (preliminary step) of the aforementioned <Filter Coefficient Calculation Process> may be calculated out by the following Equation, and stored.
  • w ( 4 ) = w ( 4 ) + a_dist × v_dist × b_dist ( 4 ) = w ( 4 ) + a_dist × v_dist × - 1 w ( 5 ) = w ( 5 ) + a_dist × v_dist × 2 w ( 6 ) = w ( 6 ) + a_dist × v_dist × - 1
  • For the other steps in the <Filter Coefficient Calculation Process>, the control variable v_skip should be read as the control variable v_dist. Moreover, the equation should be read in the similar manner in case where the filter base b_dist is another base.
  • (Example in which the Filter Coefficient is Switched Over According to a Position in a Frame)
  • Next, an example in which the filter coefficient is switched over according to where a pixel at coordinates (x, y) in the image data # 100 is located in a frame is described.
  • The control variable determining section 120 sets a value of a control variable v_xy to 1, wherein the control variable v_xy is 0 or 1 according to where a pixel at coordinates (x, y) in the image data # 100 is located in a frame. Then, the control variable determining section 120 outputs the control variable v_xy to the output filter section 110.
  • Moreover, the control variable determining section 120 outputs, to the filter section 110, a filter base b_xy associated to the position where the pixel at the coordinates (x,y) in the image data # 100 is located in the frame. The filter base b xy may be, for example, the edge-stress base or the DC offset base.
  • The filter section 110 determines, with reference to the teacher data # 200, the coefficients of the filter coefficient vector V having the components v(K) expressed by the following Equation (22).

  • v(K)=w(K)+a xy×v xy×b xy(K)  (22)
  • where a_xy is a coefficient (fixed coefficient) corresponding to the base filter b_xy.
  • The filter section 110 determines the components w(K) of the fixed coefficient vector and the coefficient a_xy per slice or per frame, for example by the aforementioned statistical approach.
  • The control variable determining section 120 sets the control variable v_xy to 1 if the pixel at the coordinates (x, y) in the image data # 100 is located in an upper half of the frame, whereas the control variable determining section 120 sets the control variable v_xy to o if the pixel at the coordinates (x, y) in the image data # 100 is located in a lower half of the frame. Moreover, the control variable determining section 120 outputs the control variable v_xy to the filter section 110.
  • For every filter target area T, the filter section 110 performs filtering based on the filter coefficient vector V using the control variable v_xy supplied from the control variable determining section 120.
  • With this configuration, it is possible to perform such filtering that filtering having a DC offset effect is performed for the upper half of the frame and filtering having no DC offset effect is performed for the lower half of the frame.
  • It should be noted that the present invention is not limited to the above-described example in which the filter base b_xy is a DC offset base. The filter base b_xy may be the edge-stress base described above.
  • Moreover, one specific example of the filter coefficient calculation process performed by the adaptive filter 100 in the present operation example is as follows. For example, assuming that the filter base b_xy is a widthwise direction edge-stress base, w′(i) in the (preliminary step) of the aforementioned <Filter Coefficient Calculation Process> may be calculated out by the following Equation, and stored.
  • w ( 4 ) = w ( 4 ) + a_xy × v_xy × b_xy ( 4 ) = w ( 4 ) + a_xy × v_xy × - 1 w ( 5 ) = w ( 5 ) + a_xy × v_xy × 2 w ( 6 ) = w ( 6 ) + a_xy × v_xy × - 1
  • For the other steps in the <Filter Coefficient Calculation Process>, the control variable v_skip should be read as the control variable v_xy. Moreover, the equation should be read in the similar manner in case where the filter base b_xy is another base.
  • (Example in which the Filter Coefficient is Switched Over According to a Size of Quantized Parameter)
  • Next, an example in which the filter coefficient is switched over according to a quantized parameter QP associated with a block to which a pixel at coordinates (x, y) in the image data # 100.
  • Here, the quantized parameter QP is a parameter associated with a quantizing step for example in MPEG-2 and the H.264/AVC standard, and assigned to every block individually.
  • The control variable determining section 120 sets a control variable v_qp to 1 wherein the control variable v_qp is 0 or 1 according to the size of a quantized parameter associated to a block to which a pixel at coordinates (x, y) in the image data # 100. Then the control variable determining section 120 outputs the control variable v_qp to the filter section 110.
  • Moreover, the control variable determining section 120 outputs, to the filter section 110, a filter base b_qp associated with a size of the quantized parameter QP associated with the block to which the pixel at the coordinates (x, y) in the image data # 100 belongs. The filter base b_qp may be, for example, the edge-stress base.
  • The filter section 110 determines, with reference to teacher data # 200, the coefficients of the filter coefficient vector V having the components v(K) expressed by the following Equation (23).

  • v(K)=w(K)+a qp×v qp×b qp(K)  (23)
  • where a_qp is a coefficient (fixed coefficient) corresponding to the filter base b_qp.
  • The filter section 110 determines the components w(K) of the fixed coefficient vector and the coefficient a_qp per slice or per frame by for example the aforementioned statistical approach.
  • The control variable determining section 120 sets the control variable qp to 1 when a value of the quantized parameter QP associated with the block to which the pixel at the coordinate (x, y) in the image data # 100 is equal to or greater than the predetermined threshold, whereas the control variable determining section 120 sets the control variable qp to 0 when a value of the quantized parameter QP associated with the block to which the pixel at the coordinate (x, y) in the image data # 100 is smaller than the predetermined threshold. Moreover, the control variable determining section 120 outputs the control variable v_qp to the filter section 110.
  • For every filter target area T, the filter section 110 performs filtering based on the filter coefficient vector V using the control variable v_qp supplied from the control variable determining section 120.
  • In general, a block associated with a greater quantized parameter QP tends to have blurring easily.
  • With this configuration, it is possible to perform such filtering that if the quantized parameter QP is equal to or greater than the threshold, filtering having a blurring effect is performed, whereas if the quantized parameter QP is less than the threshold, filtering having no blurring effect is performed.
  • Therefore, with this configuration, it is possible to perform effective filtering even if the image data # 100 includes quantized parameters of various values.
  • Again in the case where the filter coefficient is switched over according to the quantized parameter QP as described above, it is possible to perform filtering having luminance change due to the DC offset by using a filter base and a filter coefficient, each of which corresponds to the DC offset.
  • Moreover, in case where a plurality of blocks are contained in the filter target area T, it may be so configured that information assigned to any one of the blocks is used.
  • Moreover, one specific example of the filter coefficient calculation process performed by the adaptive filter 100 in the present operation example is as follows. For example, assuming that the filter base b_qp is a widthwise direction edge-stress base, w′(i) in the (preliminary step) of the aforementioned <Filter Coefficient Calculation Process> may be calculated out by the following Equation, and stored.
  • w ( 4 ) = w ( 4 ) + a_qp × v_qp × b_qp ( 4 ) = w ( 4 ) + a_qp × v_qp × - 1 w ( 5 ) = w ( 5 ) + a_qp × v_qp × 2 w ( 6 ) = w ( 6 ) + a_qp × v_qp × - 1
  • For the other steps in the <Filter Coefficient Calculation Process>, the control variable v_skip should be read as the control variable v_qp. Moreover, the equation should be read in the similar manner in case where the filter base b_qp is another base.
  • (Example in which the Filter Coefficient is Switched Over According to a Size of the DC Component)
  • Next, an example in which the filter coefficient is switched over according to a size of the DC component associated with a block to which a pixel at coordinates (x, y) in the image data # 100 belongs.
  • Here, the DC component is, for example, a coefficient corresponding to a direct current component among the coefficient after frequency transformation to the pixel values included in the block under the H.264/AVC standard.
  • The control variable determining section 120 sets a value of a control variable v_dc′ to 1, wherein the control variable v_dc′ is 0 or 1 according to the size of the DC component associated with the block to which the pixel at coordinates (x, y) in the image data # 100 belongs.
  • Moreover, the control variable determining section 120 outputs, to the filter section 110, a filter base b_dc′ associated with the size of the DC component associated with the block to which the pixel at the coordinates (x, y) in the image data # 100 belongs. The filter base b_dc′ may be, for example, the aforementioned edge-stress base.
  • The filter section 110 determines, with reference to the teacher data # 200, the coefficients of the filter coefficients vector V having the components v(K) expressed by the following Equation (24).

  • v(K)=w(K)+a dc′×v dc′×b dc′(K)  (24)
  • where a_dc′ is a coefficient (fixed coefficient) corresponding to the filter base b_dc′.
  • The filter section 110 determines the components w(K9 of the fixed coefficient vector, and the coefficients a_dc′ per slice or per frame, for example by the aforementioned statistical approach.
  • The control variable determining section 120 sets the value of the control variable v_dc′ to 0 when the value of the DC component associated with the block to which the pixel at the coordinates (x, y) in the image data # 100 is equal to or greater than a predetermined threshold, whereas the control variable determining section 120 sets the value of the control variable v_dc′ to 1 when the value of the DC component is less than the threshold. Moreover, the control variable determining section 120 outputs the control variable v_dc′ to the filter section 110.
  • For every filter target area T, the filter section 110 performs filtering based on the filter coefficient vector V using the control variable v_dc′ supplied from the control variable determining section 120.
  • In general, a block associated with a smaller DC components tends to have blurring easily.
  • With this configuration, it is possible to perform such filtering that filtering having a blurring effect is performed when the value of the DC component is smaller than the threshold value, whereas filtering having no blurring effect is performed when the value of the DC component is equal to or greater than the threshold value.
  • Therefore, with this configuration, it is possible to perform effective filtering even if the image of the image data # 100 is constituted from blocks having various DC components.
  • Again in the case where the filter coefficient is switched over according to the value of the DC parameter as described above, it is also possible to perform the filtering having the luminance change effect due to the DC offset, by using the filter base and the filter coefficient, each of which corresponds to the DC offset.
  • Moreover, in case where a plurality of blocks are contained in the filter target area T, it may be so configured that information assigned to any one of the blocks is used.
  • Moreover, one specific example of the filter coefficient calculation process performed by the adaptive filter 100 in the present operation example is as follows. For example, assuming that the filter base b_dc′ is a widthwise direction edge-stress base, w′(i) in the (preliminary step) of the aforementioned <Filter Coefficient Calculation Process> may be calculated out by the following Equation, and stored.
  • w ( 4 ) = w ( 4 ) + a_dc × v_dc × b_dc ( 4 ) = w ( 4 ) + a_dc × v_dc × - 1 w ( 5 ) = w ( 5 ) + a_dc × v_dc × 2 w ( 6 ) = w ( 6 ) + a_dc × v_dc × - 1
  • For the other steps in the <Filter Coefficient Calculation Process>, the control variable v_skip should be read as the control variable v_dc′. Moreover, the equation should be read in the similar manner in case where the filter base b_dc′ is another base.
  • (Example in which the Filter Coefficient is Switched Over According to a Size of Motion Vector)
  • Next, an in which the filter coefficient is switched over according to a size of motion vector associated with a block to which a pixel at coordinates (x, y) in the image data # 100 belongs.
  • Here, the motion vector is, for example, a motion vector for use for inter-frame prediction under H.264/AVC.
  • The control variable determining section 120 sets a value of a control variable v_mv to 1, wherein the control variable v_mv is 0 or 1 according to the size of the vector associated with the block to which the pixel at the coordinates (x, y) in the image data # 100 belongs. Then, the control variable determining section 120 outputs the control variable v_mv to the filter section 110.
  • Moreover, the control variable determining section 120 outputs, to the filter section 110, a filter base b_mv associated with a size of the motion vector associated with the block to which the pixel at the coordinates (x, y) in the image data # 100 belongs. The filter base b_mv may be, for example, the edge-stress base.
  • The filter section 110 determines, with reference to the teacher data # 200, the coefficients of the filter coefficient vector V having the components v(K) expressed by the following Equation (25).

  • w(K)+a mv×v mv×b mv(K)  (25)
  • where a_mv is a coefficient (fixed coefficient) corresponding to the filter base b_mv.
  • The filter section 110 determines the components w(K) of the fixed coefficient vector and the coefficient a_mv per slice or per frame by, for example, the aforementioned statistical approach.
  • The control variable determining section 120 sets the value of the control variable v_mv to 1 when the size of the motion vector associated with the block to which the pixel at the coordinates (x, y) in the image data # 100 belongs is equal to or greater than a predetermined threshold, whereas the control variable determining section 120 sets the value of the control variable v_mv to 1 when the size of the motion vector is less than the threshold. Moreover, the control variable determining section 120 outputs the control variable v_mv to the filter section 110.
  • For every filter target area T, the filter section 110 performs the filtering based on the filter coefficient vector V using the control variable v_mv supplied from the control variable determining section 120.
  • In general, a block associated with a greater motion vector tend to have blurring easily.
  • With this configuration, the filtering can be performed such that filtering having a blurring effect is performed when the size of the motion vector is equal to or greater than the threshold value, whereas filtering having no blurring effect is performed when the size of the motion vector is less than the threshold value.
  • Therefore, with this configuration, it is possible to perform effective filtering even if the image of the image data # 100 is constituted from blocks associated with motion vectors having various sizes.
  • In the case where the filter coefficient is switched over according to the size of the motion vector as described above, it is also possible to perform the filtering having the luminance change effect due to the DC offset, by using a filter base and filter coefficient, each of which corresponds to the DC offset described above.
  • Moreover, in case where a plurality of blocks are contained in the filter target area T, it may be so configured that information assigned to any one of the blocks is used.
  • Moreover, it may be so configured that the filter coefficient is switched over according to the size of a differential motion vector, instead of the size of the motion vector, wherein the differential motion vector is a difference between the motion vector and a predicted vector.
  • So far, the examples of switching over the filter coefficients by the adaptive filter 100 according to the present embodiment. It should be noted that the present invention is not limited to the examples mentioned above. For example, the present invention encompasses cases where the filter coefficient is switched over in a way in which the above examples are combined as appropriate. That is, the control variable determining section 120 may be configured to output some of the control variables and some corresponding ones of the control bases, which are described above. With such a configuration, it is possible to perform more effective filtering.
  • Moreover, one specific example of the filter coefficient calculating process performed by the adaptive filter 100 in the present operation example is as follows. For example, assuming that the filter base b_mv is a horizontal edge-stress base, w′(i) in the (preliminary step) of the aforementioned <Filter Coefficient Calculation Process> may be calculated out by the following Equation, and stored.
  • w ( 4 ) = w ( 4 ) + a_mv × v_mv × b_mv ( 4 ) = w ( 4 ) + a_mv × v_mv × - 1 w ( 5 ) = w ( 5 ) + a_mv × v_mv × 2 w ( 6 ) = w ( 6 ) + a_mv × v_mv × - 1
  • For the other steps in the <Filter Coefficient Calculation Process>, the control variable v_skip should be read as the control variable v_mv. Moreover, the equation should be read in the similar manner in case where the filter base b_mv is another base.
  • Moreover, in case where the control variable determining section 120 outputs a plurality of control variables, it may be configured that the control variable determining section 120 determines a plurality of control variables whose contribution to the filtering is small, a new control variable is formed from a linear sum of the plurality of control variables whose contribution to the filtering is small, and the filter coefficient is switched over by switching over the value of the new control variable.
  • For example, in case a control variable v 0 and a control variable v_2 among control variables v0, v 1, and v_2 make small contributions to the filtering, the control variable determining section 120 sets a new control variable v02 where v02=v 0+v 2, and switches over a value of the control variable v02, so as to switch over the filter coefficient.
  • With this configuration, it is possible to reduce the number of the control variables whose contributions to the filtering are small. Thereby, it is possible to reduce calculation load in the adaptive filter 100.
  • The present invention is not limited to a particular method for determining which control variables are small in contribution to the filtering. For example, filtering is performed with different values for different control variables so as to find control variables with which a difference between a pixel value before the filtering and a pixel value after the filtering is small, and identify such control variables are control variables whose contribution to the filtering is small.
  • Moreover, the present invention is not limited to the above-explained examples in which the coefficient for the filter base outputted from the control variable determining section 120 is determine by the statistical approach.
  • For example, the control variable determining section 120 may be configured such that a filter base optimal for the coefficient determined by the statistical approach is determined.
  • Moreover, the step of calculating the coefficient for a fixed filter base and the step of determining a filter coefficient optimal for a fixed coefficient may be repeated so as to find the filter coefficient and the filter base, which are optimal for the image data # 100.
  • So far, the adaptive filter 100 according to the present embodiment is explained.
  • In the following, a moving image encoding device and a moving image and a moving image decoding device and a moving image, each of which is provided with the adaptive filter 100, are described below as specific application examples of the adaptive filter 100, referring to FIGS. 5 to 9.
  • (Application Example to Moving Image Encoding Device)
  • To begin with, a moving image encoding device 1 provided with the adaptive filter 100 is described below, referring to FIG. 4.
  • The moving image encoding device 1 is a moving image encoding device in part of which the H.264/AV standard and a technique adopted in the KTA software are used.
  • FIG. 5 is a block diagram illustrating a configuration of the moving image encoding device 1.
  • As illustrated in FIG. 5, the moving image encoding device 1 includes a transforming/quantizing section 11, a variable length encoding section 12, an inverse transforming/inverse quantizing section 13, a buffer memory 14, a intra predicted image generating section 15, an inter predicted image generating section 16, a prediction method controlling section 18, a motion vector redundancy reducing section 19, an adder 21, a subtractor 22, a deblocking filter 50, and the adaptive filter 100.
  • The moving image encoding device 1 receives an input image # 1, which have been divided into block images (hereinafter, referred to as “macro blocks”) constituted from adjacent pixels.
  • The moving image encoding device 1 encodes the input image # 1, and outputs encoded data # 2.
  • The transforming/quantizing section 11 performs DCT (Discrete Cosine Transform) transformation of differential image # 22 between (i) the input image # 1 divided into the macro blocks and (ii) predicted image # 18 outputted from the later-described prediction method controlling section 18, thereby transforming the differential image # 22 into frequency components. Then, the transforming/quantizing section 11 quantizes the frequency components, thereby generating quantized prediction residual data # 11. Here, the quantization is calculation to associate the frequency component to integer value. Moreover, the DCT transformation and the quantization are performed per block obtained by dividing the macro blocks. In the following, the macro blocks to be subjected to the above treatment is referred to as “target macro blocks”, and blocks to be subjected to the above treatment is referred to as “target blocks”.
  • The inverse quantizing/inverse transforming section 13 decodes the quantized prediction residual data # 11, thereby generating the prediction residual # 13. More specifically, the inverse quantizing/inverse transforming section 13 performs inverse quantization of the quantized prediction residual data # 11, (i.e., associating (i) the integral values constituting the quantized prediction residual data # 11 to (ii) the frequency components) and inverse DCT transformation of the frequency component (i.e., inverse transformation of the target macro block into pixel components on the basis of the frequency components), thereby generating the prediction residual # 13.
  • The adder 21 adds the prediction residual # 13 and a predicted image #18 a, thereby obtaining a decoded image # 21. The decoded image # 21 thus generated is supplied to the deblocking filter 50.
  • If a difference between pixel values of pixels adjacent with each other via a block border or via a macro block border in the decoded image # 21 is smaller than a predetermined threshold, the deblocking filter 50 performs deblocking process to the block border or the macro block border in the decoded image # 21. The image having been subjected to the deblocking is outputted as a deblocked image # 50.
  • To the deblocked image # 50, the adaptive filter 100 performs the filtering explained in the <Examples of Filtering Performed by the Adaptive Filter 100>. Then, the adaptive filter 100 outputs an output image data #110 a to the buffer memory 14. In this explanation, the deblocked image # 50 corresponds to the image data # 100.
  • The adaptive filter 100 receives the input image # 1, the decoded image # 21, and side information associated with the decoded image # 21. In this explanation, the input image # 1 corresponds to the teacher data # 200, and the side information corresponds to the side information #100 a.
  • Moreover, the adaptive filter 100 outputs the filter coefficient information # 101, which is information on the filter coefficient used in the filtering. In this explanation, the filter coefficient information #101 corresponds to the filter coefficient information #110 b. That is, the filter coefficient information #101 includes the components w(K) of the fixed coefficient vector W, and the fixed coefficient used in the filtering. For example, in case where, as in the operation example described above, the adaptive filter 100 switches over the filter coefficient according to whether the skip mode is applied or not, the filter coefficient information #101 includes the components w(K) of the fixed coefficient vector W, and the fixed coefficient a_skip. The same is true for the other operation examples.
  • The adaptive filter 100 in the present application example corresponds to ALF (Adaptive Loop Filter) in KTA, substantially. That is, the adaptive filter 100 in the present application example receives a pixel value of a pixel at integer coordinates (hereinafter, the pixel is referred to as integer pixel), and outputs the pixel value of the integer pixel. This corresponds to a case where x, y, x′ and y′ in Equation (1) are all integers.
  • The moving image encoding device 1 may be configured such that it does not include the deblocking filter 50. In this case, the adaptive filter 100 performs the filtering directly on the decoded image # 21, not the deblocked image # 50. Especially, in case where the control variable is whether the block border is contained or how much the distance from the block border is, the adaptive filter 100 can be able to reduce the block noise, like the deblocking filter. In such a case, the throughput can be reduced due to the elimination of the process performed by the deblocking filter 50. Further, by using an adaptive filter specialized as a deblocking filter, it is possible to obtain a greater effect than the use of a conventional deblocking filter (for example, it is possible to use a filter that changes its blurring effect according to the luminance value).
  • The moving image encoding device 1 may be configured to include (i) both of the deblocking filter and the adaptive filter, (ii) only the adaptive filter, but not the deblocking filter, or (iii) the adaptive filter serving as a deblocking filter, and the adaptive filter.
  • The intra predicted image generating section 15 extracts a local decoded image #14 a (decoded area in the same frame of the target macro block) from the decoded image # 21 stored in the buffer memory 14, and performs intra-frame prediction on the basis of the local decoded image #14 a, thereby generating an intra predicted image # 15.
  • The inter predicted image generating section 16 calculates a motion vector # 17 by using a reference image # 14 b, whose whole frame has been decoded and which is stored in the buffer memory 14. Then, inter predicted image generating section 16 allocates the motion vector to the target block on the input image # 1. The calculated motion vector # 17 is outputted to the predicted image generating section 16 and the motion vector redundancy reducing section 19, and stored in the buffer memory 14. Moreover, the inter predicted image generating section 16 performs motion compensation with respect to the reference image on the basis of the motion vector #17 per block, thereby generating an inter predicted image # 16.
  • The prediction method controlling section 18 compares the intra predicted image # 15, the inter predicted image # 16, and the input image “1 for every block, so as to select and output one of the intra predicted image # 15 and the inter predicted image # 16 as a predicted image # 18. Moreover, the prediction method controlling section 18 also outputs a prediction mode # 18 b, which is information indicating which one of the intra predicted image # 15 and the inter predicted image # 16 is selected. The prediction image #18 a is inputted in the subtractor 22.
  • The prediction mode # 18 b is stored in the buffer memory 14 and inputted in the variable length encoding section 12.
  • The motion vector redundancy reducing section 19 calculates out a predicted vector after the motion vector # 17 is allotted to the target block by the inter predicted image generating section 1, the motion vector redundancy reducing section 19 calculating out the predicted vector based on a motion vector group 14 c stored in the buffer memory 14, the motion vector group 14 c having been allotted to other blocks than the target block to which the motion vector # 17 is allotted by the inter predicted image generating section 16. Moreover, the motion vector redundancy reducing section 19 determines a difference between the predicted vector and the motion vector # 17, thereby generating the differential motion vector # 19. The differential motion vector # 19 thus generated is supplied to the variable length encoding section 12.
  • The variable length encoding section 12 performs variable length encoding process to the quantized prediction residual data # 11, the differential motion vector # 19, the prediction mode # 18 b, and the coefficients included in the filter coefficient information # 101, thereby generating the encoded data # 2.
  • The subtractor 22 determines a difference between the input image # 1 and the predicted image #18 a for the target macro block, and outputs a differential image # 22.
  • (Encoding of the Filter Coefficient Information #101)
  • Here, encoding of the filter coefficient information #101 is described specifically, referring to (a) to (c) of FIG. 6. (a) to (c) of FIG. 6 are views illustrating configurations of encoded data of the filter coefficient information # 101.
  • The filter coefficient information #101 is filter coefficient itself or information necessary for calculating out the filter coefficient. For example, if the filter coefficient calculating equation F is Equation (9), The filter coefficient information #101 is the fixed coefficient vector wi the fixed coefficient akj. If the filter coefficient calculating equation F is Equation (6), the filter coefficient information #101 is w(i) where i=1 to 10, and a_skip.
  • In the example illustrated in (a) of FIG. 6, encoded data of the filer coefficient information #101 is constituted from the fixed filter parameter. The values of the filter coefficient information #101 have higher possibility to have values closer to 0, except a center element. Thus, the values of the filter coefficient information #101 are divided into signs and absolute values. The signs are encoded by 1 bit, and the absolute values are encoded by shorter codes as the absolute values are closer to 0. The method of encoding is not limited to this and arithmetic encoding may be adopted. The encoding is performed, considering that the center element has a value closer to 1.
  • Moreover, it is also appropriate to adopt encoding in which a predicted value of the filter coefficient information #101 is calculated out, and a difference (prediction difference) between the filter coefficient information #101 and the predicted value is encoded. The predicted value may be such that a predicted value for w(5) corresponding to a center is 0.8 and predicted values for w(i) around the center are 0.1, considering that a center value is closer to 1 and the other elements are closer to 0, for example. The prediction method is not limited to this.
  • As described above referring to FIG. 3, it is possible to divide the filter coefficient vector into components variable according to the control variable and components invariable according to the control variable. The encoding may be performed by utilizing this feature. (b) and (c) of FIG. 6 illustrates two examples of this encoding.
  • The example illustrated in (b) of FIG. 6, the encoded data is constituted from (1) filter coefficients regarding components invariable according to the control variable, and (2) filter coefficients regarding components variable according to the control variable, and being associated with the values of the control variables. In this example, the filter coefficient v(i) as such is encoded for the component invariable according to the control variable. As to the component variable according to the control variable, the filter coefficient v(i) where i=4, 5, 6 for the case where the control variable v_skip=0, and the filter coefficient v′ (i) where i=4, 5, 6 for the case where the control variable v_skip=1 are encoded.
  • In the example illustrated in (c) of FIG. 6, the encoded data includes (1) filter coefficients (hereinafter, referred to as basic coefficients) regarding various components invariable according to the control variables, and (2) differences between the basic coefficients and the filter coefficients regarding the components variable according to the control variables. In this example, to begin with, the filter coefficient vector V for control variable v_skip=0, v(i) where i=1 to 10. Next, for the elements variable according to the control variables among the elements of the filter coefficient vectors, differences dv′(i) where i=4, 5, 6 between the filter coefficients v(i) for the case where the control variable v_skip=0, and the filter coefficients v′(i) for control variable v_skip=1 are encoded.
  • Again in the examples illustrated in (b) and (c) of FIG. 6, if the filter coefficients are symmetric (the bases are symmetric), it is possible to encode only one of the identical filter coefficients by being symmetric.
  • In the examples, there are three components variable according to the control variable v_skip. Thus, it is preferable to adopt the example in (a) of FIG. 6 rather than those in (b) and (c) of FIG. 6, because the example in (a) of FIG. 6, which encodes only the filter coefficient information (here, a_skip) regarding the control variable, encodes less number of the filter coefficient information than those in (b) and (c) of FIG. 6, in which each of the components variable and invariable according to the filter coefficients are encoded. However, if there is only one component variable according to the control variable, like in case where the DC offset base b_dc is used as the base corresponding to the control variable, the number of the filter coefficient information to be encoded is equal between the encoding each of the components variable and invariable according to the filter coefficients and the encoding only the filter coefficient information regarding the control variable.
  • As describe above, an encoding device (moving image encoding device 1) according to the present invention comprises the image filter (adaptive filter 100) and causes the image filter to process a decoded image.
  • The encoding device according to the present invention with the aforementioned configuration can process the decoded image with the filtering performed by the image filtering, the decoded image being obtained by decoding an encoded image having been encoded by the encoding device. Thus, the encoding device according to the present invention has an effect similar to that of the image filter. Moreover, the encoding device with the above configuration can perform filtering more suitable for the respective target areas (filter target area T) by using (i) the components of the fixed filter coefficient vector (fixed coefficient vector calculated out from the fixed coefficient vector W), and (ii) the fixed coefficient. Thus, it is possible to perform the filtering by using a smaller number of components of the filter coefficient vector, compared with the case where the filter coefficient vector having different components for different target areas is used.
  • Moreover, by using the encoding device with the above configuration, it is possible to reduce the amount of encoding the filter coefficient vector, compared with the case where the filter coefficient vector having different components for different target areas is used.
  • (Application Example in which the Inter Predicted Image Generating Section)
  • Moreover, the adaptive filter 100 according to the present invention is applicable to the inter predicted image generating section 16 in the moving image encoding device 1.
  • FIG. 7 is a block diagram illustrating a configuration of the inter predicted image generating section 16 provided with the adaptive filer 100. In the following, the adaptive filter 100 in the present application example may be referred to as an adaptive filter 100′.
  • As illustrated in FIG. 7, the inter predicted image generating section 16 includes a predicted image generating section 16 a, a motion vector estimating section 17, and an adaptive filter 100′. In the present application example, the adaptive filter 100′ corresponds to AIF (Adaptive Interpolation Filter) in KTA, substantially.
  • In the following, operations of sections in the inter predicted image generating section 16 are described below.
  • (Step 101)
  • The reference image # 14 b stored in the buffer memory 14 is supplied to the adaptive filter 100′. In the following, the reference signal # 14 b is referred to as first reference signal # 14 b.
  • The adaptive filter 100′ performs filtering to the first reference image # 14 b on the basis of a predetermined filter coefficient, thereby outputting first output image data #110 a′. That is, the adaptive filter 100′ in this step operates as a fixed filter for filtering based on the predetermined filter coefficient.
  • Here, the predetermined filter coefficient may be, for example, a filter coefficient used in H.264/AVC. Moreover, due to the filtering performed by the adaptive filter 100′, an image signal whose pixel precision is not more than integer pixels is generated, by interpolation, from the pixel values of the integer pixels contained in the first reference image # 14 b. This corresponds to a case where x′ and y′ in Equation (1) are numbers other than integers. Therefore, the first output image data 100′ is image data including (i) the pixel values of the integer pixels and (ii) an interpolation signal whose pixel precision is not more than integer pixels.
  • (Step 102)
  • The motion vector estimating section 17 generates the first motion vector # 17′ on the basis of the first output image data #110 a′ and the input image # 1. The first motion vector # 17′ is prepared for 1 slice or for 1 frame.
  • (Step 103)
  • The adaptive filter 100′ determines a filter coefficient by processing the first reference image # 14 b with the process described in <Examples of Filtering Performed by Adaptive Filter 100>. The image generated by filtering performed by the adaptive filter 100′ is referred to as second output image data #110A″. Here, the first reference image # 14 b corresponds to the image data # 100 above, and the second output image data 110 a″ corresponds to the output image data #110 a above. Moreover, the input image # 1 is supplied to the adaptive filter 100′ as the teacher data #220.
  • The adaptive filter 100′ generates an image signal whose pixel precision is not more than integer pixel, by interpolation, from the pixel values of the integer pixels contained in the first reference image # 14 b. This corresponds to a case where x′ and y′ in Equation (1) are numbers other than integers. Therefore, the second output image data 110 a′ is image data including (i) the pixel values of the integer pixel and (ii) an interpolation signal whose pixel precision is not more than integer pixel.
  • (Step 104)
  • The motion vector estimating section 17 generates the second motion vector # 17″ on the basis of the second output image data #110 a″, and the input image # 1.
  • (Step 105)
  • The adaptive filter 100′ performs filtering to the first reference image # 14 b by using the filter coefficient thus obtained, thereby generating a motion compensated image based on the second motion vector # 17″ so as to generate and output the inter predicted image # 16.
  • The adaptive filter 100′ may be configured to generate the second output image data #110 a′ by using control variables and filter bases, which are set according to the locations of the pixels in the output image data #110 a′.
  • (Application Example to Moving Image Decoding Device)
  • In the following, the moving image decoding device 2 provided with the adaptive filter 100 is described, referring to FIG. 8.
  • The moving image decoding device 2 a moving image decoding device in part of which the H.264/AV standard and a technique adopted in the KTA software are used.
  • FIG. 8 is a block diagram illustrating a configuration of the moving image decoding device 2.
  • As illustrated in FIG. 8, the moving image decoding device 2 includes a variable length decoding section 23, a motion vector reconstituting section 24, a buffer memory 25, an inter predicted image generating section 26, an intra predicted image generating section 27, a prediction method determining section 28, an inverse quantizing/inverse transforming section 29, an adder 30, a deblocking filter 50, and the adaptive filter 100.
  • The moving image decoding device 2 receives an encoded data # 2 and outputs an output image # 3.
  • The variable length decoding section 23 performs variable length decoding of the encoded data # 2, and outputs a differential motion vector #23 a, side information # 23 b, a quantized prediction residual data # 23 c, and filter coefficient information # 23 d.
  • The motion vector reconstituting section 24 reconstitutes a motion vector # 24 of a target partition from the differential motion vector #23 a and the motion vector 25 a, which have been decoded and stored in the buffer memory 25.
  • The buffer memory 25 stores therein the output image data #110 a outputted from the adaptive filter 100, the motion vector # 24, and the side information # 23 b.
  • The inter predicted image generated section 26 generates an inter predicted image # 26 on the basis of (i) the motion vector # 24, which has been decoded by the motion vector reconstituting section 24 and supplied thereto via the buffer memory 25, and (ii) the reference image # 25 d stored in the buffer memory 25. Note that, a motion vector # 25 c includes a motion vector identical with the motion vector # 24. Moreover, the reference image # 25 d corresponds to the later-described output image data #110 a outputted from the adaptive filter 100.
  • The intra predicted image generating section 27 generates an intra predicted image # 27 from a local decoded image # 25 b in the same image as the target macro block, the intra predicted image # 27 being stored in the buffer memory 25.
  • Based on prediction mode information contained in the side information # 23 b, the prediction method determining section 28 selects and outputs, as a predicted image # 28, one of the intra predicted image # 27 and the inter predicted image # 26.
  • The inverse quantizing/inverse transforming section 29 performs inverse quantization and inverse DCT transformation to the quantized prediction residual data # 23 c, thereby outputting prediction residual # 29.
  • The adder 30 adds the prediction residual # 29 and the predicted image # 28, and outputs the resultant image as the decoded image # 3. Moreover, the decoded image # 3 thus outputted is supplied to the deblocking filter 50.
  • The deblocking filter 50 performs deblocking process to a block border or a macro block border in the decoded image # 3, if a difference between pixels adjacent with each other across the block border or the macro block border in the decoded image # 3 is smaller than a predetermined threshold. The image data having been subjected to the deblocking process is outputted as deblocked image # 50.
  • The adaptive filter 100 subjects the deblocked image # 50 to the filtering described in <Examples of Filtering Performed by the Adaptive Filter 100>. Here, the deblocked image # 50 corresponds to the image data # 100 described above.
  • Moreover, the adaptive filter 100 in the present application example calculates out the filter coefficient on the basis of the filter coefficient information # 23 d decoded from the encoded data # 2, and performs the filtering by using the filter coefficient thus calculated out.
  • That is, the adaptive filter 100 in the present application example is configured such that the adaptive filter 100 calculates out the filter coefficient based on the filter coefficient information # 23 d decoded from the encoded data # 2, rather than determining the filter coefficient by the statistical approach as described in <Examples of Filtering Performed by the Adaptive Filter 100>, and the adaptive filter 100 performs the filtering by using the filter coefficient thus calculated out. The switching-over of the filter coefficient in the adaptive filter 100 has been described in <Examples of Filtering Performed by the Adaptive Filter 100>
  • The adaptive filter 100 outputs to the buffer memory 25 the output image data #110 a thus generated by filtering the deblocking image # 50.
  • The adaptive filter 100 receives the side information # 23 b. The side information # 23 b includes (i) information regarding whether the target block is double direction predicted block or single direction predicted block, (ii) information regarding whether the skip mode is applied to the target macro block or not, (iii) prediction mode information regarding whether the target macro block is intra predicted macro block or inter predicted macro block, and (iv) quantized parameter QP associated with the target block. Moreover, the side information # 23 b corresponds to the side information #100 a described above.
  • The adaptive filter 100 in the present application example corresponds to ALF (Adaptive Loop Filter) in KTA, substantially. That is, the adaptive filter 100 in the application example receives and outputs pixels values of integer pixels. This corresponds to cases where x, y, x′ and y′ in Equation (1) are all integers.
  • The moving image decoding device 2 may be provided with no deblocking filter 50. In this case, the adaptive filter 100 performs the filtering directly to the decoded image # 3, rather than the deblocked image # 50. This configuration can reduce the throughput, as in the case of the moving image encoding device 1.
  • Like the moving image encoding device 1, the moving image decoding device 2 may be configured to include (i) both of the deblocking filter and the adaptive filter, (ii) only the adaptive filter, but not the deblocking filter, or (iii) the adaptive filter serving as a deblocking filter, and the adaptive filter.
  • <Decoding and Derivation of Filter Coefficient>
  • <Example Typically Exemplified in (a) of FIG. 6>
  • In case where the fixed filter parameter or its predictive differential data has been encoded, the variable length decoding section 23 decodes a fixed filer parameter as the filter coefficient information # 23 d. In (a) of FIG. 6, w(1) to w(M×N+1) and a_skip are decoded.
  • The adaptive filter 100 receives the fixed filter parameter and the control variable into the filter calculating function F whose general forms are shown in Equations 8 and 9, and derives the filter coefficient. In this example, adaptive filter 100 receives w(1) to w(M×N+1) and a_skip decoded in Equation (6) are inputted, and derives the filter coefficient vector V.
  • <Example Typically Exemplified in (b) of FIG. 6>
  • In case where the filter coefficient or the predictive differential data has been encoded, the variable length decoding section 23 decodes the filter coefficient as the filter coefficient information # 23 d. In the example in (b) of FIG. 6, one filter coefficient is encode for the component in variable according to the control variable, whereas a plurality of filter coefficients are encoded for respective values of the control variable for the components variable according to the control variable. The variable length decoding section 23 can derive a filter coefficient by branching the process depending on whether the coefficient is variable or invariable according to the control variable. In the example in (b) of FIG. 6, v(i), where i=1, 2, 3, 7, 8, 9, and 10 are decoded as the elements invariable according to the control variable among the elements of filter coefficient vector. Next, the filter coefficient v(i) where i=4, 5, and 6 for the case where the control variable v_skip=0 and the filter coefficient v′(i) where i=4, 5, 6, are decoded as the elements variable according to the control variable among the elements of filter coefficient vector. Further, it is assumed that v′(i)=v(i) for the elements invariable according to the control variable among the elements of filter coefficient vector.
  • <Example Typically Exemplified in (c) of FIG. 6>
  • Here, a basic filter coefficient and a difference from the basis filter coefficient, or a predictive differential data is encoded. The variable length decoding section 23 decodes the basic filter coefficient as the filter coefficient information # 23 d, then decodes a difference from the basic filter coefficient. In the example in (c) of FIG. 6, v(i) where i=1 to 10 is decoded as the basis filter coefficient. These become the filter coefficient vector V for the case where the control variable v_skip=0. Next, for the elements variable according to the control variable among the elements of filter coefficient vector, dV(i) where i=4, 5, 6, which is a difference from a basic difference, is decoded. For the component invariable according to the control variable, the filter coefficient v′ (i) for the case where the control variable v_skip=1 is derived from the following by using the basing filter coefficient v(i).

  • v″(i)=v(i), where i=1, 2, 3, 7, 8, 9, 10
  • For the component variable according to the control variable, the filter coefficient v′(i) for the case where the control variable v_skip=1 is derived from the following by using the basis filter coefficient v(i) and the difference dV(i) from the basic filter coefficient.

  • v′(i)=v(i)+dV(i), where i=4, 5, 6
  • As described above, the filter coefficient can be derived by branching for every component depending on whether or not being variable to the control variable.
  • <Operation of Adaptive Filter 100>
  • Again in the moving image decoding device 2, the adaptive filter 100 performs the operation as described above referring to FIGS. 3 and 4. In case where the filter coefficient vector is derived by branching, the process is carried out as shown in the flow of FIG. 3. In case where the filter coefficient vector is derived by summation such as Equation (9), the process is carried out as shown in the flow of FIG. 4. In case where the filter coefficient vector is derived by summation, the process may be carried out as shown in the flow of FIG. 3, by using the preliminary step, or the process may be carried out by referring to a table specified according to positions of the components of the filter coefficient and the values of the control variable are specified.
  • <Decoding by Using Predicted Value>
  • In the above explanation, the filter coefficient, the difference of the filter coefficient, or a degree of increase in the filter coefficient are decoded as the filter coefficient information # 23 d to be decoded. The present invention may be configured such that predicted value of the filter coefficient, the difference of the filter coefficient, or a degree of increase in the filter coefficient is derived and a difference from the predicted value is decoded.
  • As described above, a decoding device (moving image decoding device 2) according to the present invention comprises the image filter (adaptive filter 100) and causes the image filter to process the decoded image.
  • The decoding device according to the present invention with the above configuration can attain the same effect as the image filter. Moreover, the decoding device with the above configuration can perform filtering more suitable for the respective target areas (filter target area T) by using (i) the components of the fixed filter coefficient vector (fixed coefficient vector calculated out from the fixed coefficient vector W), and (ii) the fixed coefficient. Thus, it is possible to perform the filtering by using a smaller number of components of the filter coefficient vector, compared with the case where the filter coefficient vector having different components for different target areas is used.
  • Moreover, by using the decoding device with the above configuration, it is possible to reduce the amount of encoding the filter coefficient vector, compared with the case where the filter coefficient vector having different components for different target areas is used.
  • (Application Example for Application to Inter Predicted Image Generating Section)
  • Moreover, the adaptive filter 100 according to the present invention is applicable to the inter predicted image generating section 26 in the moving image decoding device 2.
  • FIG. 9 is a block diagram illustrating a configuration of the inter predicted image generating section 26 provided with the adaptive filter 100. In the following, the adaptive filter 100 in the present application example is referred to as an adaptive filter 100″.
  • As illustrated in FIG. 9, the inter predicted image generating section 26 includes a predicted image generating section 26 a and the adaptive filter 100″.
  • In the following operations of sections of the inter predicted image generating section 26 is described.
  • The adaptive filter 100″ performs filtering to the reference image # 25 d stored in the buffer memory 25. Here, the reference image # 25 d corresponds to the image data # 100 described above.
  • In the adaptive filter 100″ in the application example performs filtering based on the filter coefficient information # 23 d decoded from the encoded data # 2.
  • That is, the adaptive filter 100″ in the present application example is configured such that the adaptive filter 100″ performs the filtering based on the filter coefficient information # 23 d decoded from the encoded data # 2, rather than determining the filter coefficient by the statistical approach as described in <Examples of Filtering Performed by the Adaptive Filter 100>, and the adaptive filter 100 performs the filtering by using the filter coefficient thus calculated out. The switching-over of the filter coefficient in the adaptive filter 100 has been described in
  • <Examples of Filtering Performed by the Adaptive Filter 100>
  • The adaptive filter 100″ outputs, to the predicted image generating section 26 a, the output image data #110 a′″ thus generated by the filtering the reference image # 25 d.
  • Moreover, the adaptive filter 100″ receives the side information # 23 b. The side information # 23 b includes (i) information regarding whether the target block is double direction predicted block or single direction predicted block, (ii) information regarding whether the skip mode is applied to the target macro block or not, (iii) prediction mode information regarding whether the target macro block is intra predicted macro block or inter predicted macro block, and (iv) quantized parameter QP associated with the target block. Moreover, the side information # 23 b corresponds to the side information #100 a described above.
  • The adaptive filter 100″ in the application example corresponds to the AIF (Adaptive Interpolation Filter) in KTA, substantially. That is, the adaptive filter 100″ in the application example receives pixel values of integer pixels and outputs pixels values whose pixel precision is not more than the integer pixel. This corresponds to cases where x and y in Equation (1) are integer, and x′ and y′ in Equation (1) are an integer or a decimal.
  • The predicted image generating section 26 a performs motion compensation to the output image data #110 a′″ by using the motion vector # 25 c, thereby generating and outputting the inter predicted image # 26.
  • (Configuration of Encoded Data #2)
  • In the following, a configuration of the encoded data # 2 generated by using the moving image encoding device 1 is described, referring to FIGS. 10 and 11.
  • FIG. 10 is a view illustrating a bit stream #BS, which is generated for every slice of the encoded data # 2 by the moving image encoding device 1 so as to be referred to for the slice of the encoded data # 2 by the moving image decoding device 2. As illustrated in FIG. 10, the bit stream #BS contains filter coefficient information FC, and macro block information MB1 to MBN.
  • FIG. 11 is a view illustrating one example of the filter coefficient information FC.
  • The filter coefficient information FC is information including the filter coefficient information # 23 d generated by the adaptive filter 100, like the filter coefficient information #101 described above referring to FIG. 6. Note that the moving image decoding device 2 may be configured to decode information indicating which kind of control variable the control variable to use is, in addition to the filter coefficient information # 23 d. For example, 8-bit information from 0 to 255 is decoded, and the kind of the control variable to use is judged according to the bit number, for instance, 1 bit indicates the degree of flatness as a control variable, 2 bits indicate the edge angle as a control variable, 4 bits indicate the luminance as a control variable, 8 bits indicate color as a control variable, 16 bits indicate, as a control variable, whether an edge is included or not, 32 bits indicate the prediction direction as a control variable, 64 bits indicate, as a control variable, whether or not skip mode is applied, and 128 bits indicate, as a control variable, whether the inter or intra is used. Note that it is possible to select a plurality of control variables in free combinations by encoding such encoded data. The optional control variables may be these described in the present Specification, or those not described herein.
  • Moreover, the kind of the base to use for the control variable may be decoded. For example, 8-bit information from 0 to 255 is decoded, and the kind of the base to use is judged according to the bit number, for instance, 1 bit indicates the horizontal edge-stress base as a control variable, 2 bits indicate the vertical edge-stress base as a control variable, 4 bits indicate the standard Laplacian base as a control variable, 8 bits indicate first diagonal direction Laplacian base as a control variable, 16 bits indicate the second diagonal Laplacian base as a control variable, 32 bits indicate the DC offset base as a control variable, . . . . Moreover, the moving image decoding device 2 can select the kinds of the bases to use for individual control variables by decoding the kind of the based for every control variable. In FIG. 11, an example in which bases are designated for respective two control variables # 1 and #2. The optional bases may be these described in the present Specification, or those not described herein.
  • Moreover, the moving image decoding device 2 may be configured to decode the information as to the number of tapping in the filter, for example, the decoding may be such that 0 is decoded for 3×3 filter, 1 is decoded for 5×5 filter, 2 is decoded for 7×7 filter, and 3 is decoded for 9×9 filter. In this way, it is possible to change the number of the filter coefficient vectors according to the filter effect, thereby improving encoding efficiency. The moving image decoding device 2 may or may not be configured to change the number of tapping in the base according the number of tapping in the decoded filter. Therefore, there may be such a case that the number of tapping in the filter is 5×5 while that of the base is 3×3. If the number of tapping in the filter is grater than that of the base, the value of the filter coefficient information # 23 d for a position out of the range of the base is decoded as the value of the filter coefficient vector. In this case, the value of the filter coefficient information # 23 d for a position out of the range of the base is a component invariable according to the control variable. Moreover, if the number of tapping in the filter is smaller than that of the base, the value of the filter coefficient vector is a component variable according to the control variable and can be defined even out of the range of the tapping in the filter, but the filtering is carried out only with filter coefficients as many as the number of taps in the filter.
  • Moreover, the moving image decoding device 2 may be configured to decode information indicating the format of the encoded data of the filter. For example, it may be such that 0 is decoded for the format of (a) of FIG. 6, 1 is decoded for the format of (b) of FIG. 6, and 2 is decoded for the format of (c) of FIG. 6. Even in an encoding device having a difficulty in determining the filter coefficient information # 23 d in the format of (a) of FIG. 6 because of the algorithm for determining the filter coefficient vector, it is possible to improve the encoding efficiency by changing, depending on the control variable, the filter coefficient vector to used.
  • Moreover, the moving image decoding device 2 may be configured to decode information indicating bit depth of the filter coefficient information. For example, whether the bit number is 8, 10, or 12 bits is decoded. The bit depth is equivalent to roughness of the quantization of the filter coefficient information.
  • Even though it is not illustrated in FIG. 11, the moving image decoding device 2 may be configured to decode the range of the values of the control variable. For example, whether the range of the valuables of the control variable is 0 and 1, or 0, 1, and 2 is decoded. Moreover, information indicating the threshold value for use in determining the control variable from the image data # 100 and the side information thereof may be decoded. For example, the threshold information for determining the value of the control variable from the luminance or the degree of flatness may be decoded. Moreover, filtering unit may be decoded. The filtering unit may be pixel, block, a group of block, or the other. Moreover, in case where an area not to be subjected to the filtering is included in the unit area to be subjected to the filtering, information for specifying whether there is a filtering area may be included.
  • These kinds of information are referred to as filter coefficient information controlling information.
  • The filter coefficient vector can be switched over to be optimal for sequence, picture, or slice by decoding the filter coefficient information controlling information such as control variable, base, the number of tapping, or the like as described above. This improves the encoding efficiency. Moreover, the filter coefficient vector can be switched over appropriately for computing power of the encoding device. This gives the encoding device a greater degree of freedom. For example, if the capacity of the encoding device is large, the encoding efficiency can be increased by increasing the number of the control variables or the number of the bases. On the other hand, if the capacity of the encoding device is small, it may be possible to minimize the number of the control variables or the number of the bases, or not to use the control variable or the base. Moreover, it is possible to change the control variable or base to use, according to the software or hardware mounted on the encoding device. For example, if the encoding device is implemented with software or hardware capable of calculating the degree of flatness at high speed, the encoding device may be such that the degree of flatness is used as the control variable. Such degree of freedom is not beneficial to the encoding device, but also for the decoding device. That is, if it can be assumed that the decoding device is small in capacity, the encoding can be performed in such a way that the number of the control variables to use or the number of bases to use is minimized.
  • Moreover, information (filter coefficient information format controlling information) indicating which information is included as the filter coefficient information controlling information may be decoded. For example, it may be configured that the control variable information is included but the tap information is not included. Note that the filter coefficient information 4 b, the filter coefficient information controlling information, and the filter coefficient information format controlling information may be encoded in different positions on the encoded data. For example, it may be such that the filter coefficient information controlling information is encoded in a slice header, the filter coefficient information format controlling information is encoded in a picture header, and the filter coefficient information format controlling information is encoded in a sequence header. Moreover, the encoding may be to create encoded data corresponding to a picture parameter set or a sequence parameter set in the H.254/AVC.
  • The macro block information MB1 to MBN are information regarding the macro blocks included in the slice in interest, and includes a prediction mode # 18 b of the macro block, and block dividing information, etc. Here, n is the number of the macro blocks included in the slice in interest.
  • As described above, the data structure of the encoded data according to the present invention is a data structure of an encoded data, which a decoding device refers to, the decoding device (i) including an image filter for calculating a pixel value of target pixel in an output image from a pixel value of each of a pixel or pixels belonging to a target area in an input image by using a filter coefficient vector, and (ii) causing the image filter to process a decoded image, the data structure comprising: a fixed coefficient, with which the decoding device multiplies a product of a predetermined fixed vector and a variable coefficient variable according to at least either where the target area is in the input image, or where the target pixel is in the output image; and components of a fixed filter coefficient vector to which the decoding device adds a product of the variable coefficient, the fixed vector, and the fixed coefficient, the fixed filter coefficient vector being predetermined for every unit area constituting the output image.
  • When the encoded data has the data structure configured as above, the decoding device for decoding an image based on the encoded data can perform filtering by using the components of the fixed coefficient vector, and the fixed coefficient contained in the encoded data, thereby performing filtering more appropriately adaptive for every target area by using a smaller number of filter coefficient, compared with the case where the encoded data contains different filter coefficients for the target area, respectively.
  • (Summary)
  • As described above, an image filter according to the present invention is an image filter for calculating a pixel value of target pixel in an output image from a pixel value of each of a pixel or pixels belonging to a target area in an input image and from a filter coefficient vector, comprising: filter coefficient vector changing means for changing the filter coefficient vector according to at least either where the target area is in the input image, or where the target pixel is in the output image.
  • With this configuration, in which the filter coefficient vector changing means for changing the filter coefficient vector according to at least either where the target area is in the input image, or where the target pixel is in the output image is provided, it is possible to perform filtering more appropriately adaptive for every target area according to at least either where the target area is in the input image, or where the target pixel is in the output image.
  • Moreover, it is preferable in the image filter according to the present invention that the filter coefficient vector is dividable into a sum of (i) a fixed filter coefficient vector predetermined for every unit area constituting the output image, and (ii) a variable filter coefficient vector variable by the filter coefficient vector changing means according to at least either where the target area is in the input image, or where the target pixel is in the output image.
  • In this configuration, the filter coefficient vector is dividable into the sum of (i) the fixed filter coefficient vector predetermined for every unit area constituting the output image, and (ii) the variable filter coefficient vector, and the filter coefficient vector changing means can change the variable filter coefficient vector according to at least either where the target area is in the input image, or where the target pixel is in the output image.
  • Therefore, with this configuration, it is possible to further attain such an effect that filtering can be more appropriately adaptive for every target area (i.e., more suitably for image characteristics in the individual target areas) by using the filter coefficient predetermined per unit area.
  • In case where the present invention is applied to H.264, the unit area may be a slice, may be a macro block, or may be a block.
  • Moreover, the fixed filter coefficient vector and the variable filter vector may be defined to include an offset component.
  • Moreover, it is preferable in the image filter according to the present invention that the fixed filter coefficient vector is dividable into (i) a fixed coefficient vector predetermined for every unit area constituting the output image, and (ii) a base.
  • Moreover, it is preferable in the image filter according to the present invention that the variable filter coefficient vector is further dividable into (i) a fixed coefficient predetermined for every unit area constituting the output image, (ii) a variable coefficient variable by the filter coefficient vector changing means according to at least either where the target area is in the input image, or where the target pixel is in the output image, and a predetermined fixed vector.
  • In this configuration, the variable filter coefficient vector is further dividable into (i) a fixed coefficient predetermined for every unit area constituting the output image, (ii) a variable coefficient, and (iii) a predetermined fixed vector. Moreover, the filter coefficient vector changing means can change the variable filter coefficient vector according to at least either where the target area is in the input image, or where the target pixel is in the output image. Therefore, it is possible to further attain such an effect that filtering can be more appropriately adaptive for image characteristics in the individual target areas by using the filter coefficient predetermined per unit area.
  • With this configuration, the image filter is supplied with the fixed coefficient vector (or fixed filter coefficient vector) determined for every unit area constituting the output image, and the fixed coefficient. By this, it is possible to attain such additional effect that filtering of a desired type appropriate for every target area can be performed.
  • For example, assume that the number of the target area in the unit area is 10, the number of the components in the fixed coefficient vector (or fixed filter coefficient vector) is P, and the number of the fixed coefficient is 1. In this case, the above configuration makes it possible to perform filtering of desired types appropriate for 10 target areas individually by supplying the image filter with the P+1 number of coefficient data in total. On the other hand, if the fixed coefficient vectors are calculated out respectively for the 10 target areas, it becomes necessary to have the 10×P number of coefficient data. Therefore, the above configuration makes it possible to perform filtering of types more appropriate for every target area by using a smaller number of coefficient data, compared with the case the fixed coefficient vectors (or filter coefficient vectors) are calculated out for the target areas, respectively.
  • It is preferable in the image filter according to the present invention that the variable coefficient is changed according to a degree of flatness in an image in the target area by the filter coefficient vector changing means; and the fixed vector is a filter base having at least one of edge-stress effect and a blurring effect.
  • With this configuration, the fixed vector is a filter base having at least one of edge-stress effect and a blurring effect, and the filter coefficient vector changing means can change the variable coefficient according to a degree of flatness in an image in the target area, wherein the degree of flatness is varied according to the position of the target area. Therefore, with this configuration, in which at least one of the edge-stress effect or the blurring effect is changed according to the degree of flatness, it is possible to further attain such an effect that more appropriately adaptive filtering can be performed.
  • Moreover, the filter base having at least one of the edge-stress effect or the blurring effect can provide edge-burring effect instead of edge-stress effect, depending on whether the fixed coefficient is positive or negative (the same is true hereinafter). Therefore, with this configuration, it is possible to further attain such an effect that more appropriately adaptive filtering by changing the edge-blurring effect according to the degree of flatness in the image in the target area.
  • It is preferable in the image filter according to the present invention that the variable coefficient is changed according to an edge angle of an edge in an image in the target area by the filter coefficient vector changing means; and the fixed vector is a filter base having at least one of edge-stress effect and a blurring effect.
  • With this configuration, the fixed vector is a filter base having at least one of edge-stress effect and a blurring effect, and by the filter coefficient vector changing means can change the variable coefficient according to an edge angle of an edge in an image in the target area. Therefore, with this configuration, it is possible to further attain such an effect that more appropriately adaptive filtering by changing at least one of the edge-stress effect and the blurring effect by changing the variable coefficient according the edge angle of the edge in the image in the target area.
  • It is preferable in the image filter according to the present invention that the variable coefficient is changed according to luminance of an image in the target area by the filter coefficient vector changing means; and the fixed vector is a filter base having at least one of edge-stress effect and a blurring effect.
  • In this configuration, the fixed vector is a filter base having at least one of edge-stress effect and a blurring effect, and the filter coefficient vector changing means can change the variable coefficient according to luminance of an image in the target area, wherein the luminance is variable according to the position of the target area. Therefore, with this configuration, it is possible to further attain such an effect that more appropriately adaptive filtering by changing at least one of the edge-stress effect and the blurring effect according the edge angle of the edge in the image in the target area.
  • It is preferable in the image filter according to the present invention that the variable coefficient is changed according to a color of an image in the target area by the filter coefficient vector changing means; and the fixed vector is a filter base having at least one of edge-stress effect and a blurring effect.
  • In this configuration, the fixed vector is a filter base having at least one of edge-stress effect and a blurring effect, and the filter coefficient vector changing means can change the variable coefficient according to a color of an image in the target area. Therefore, with this configuration, it is possible to further attain such an effect that more appropriately adaptive filtering by changing at least one of the edge-stress effect and the blurring effect according to the color of the image in the target area.
  • It is preferable in the image filter according to the present invention that the variable coefficient is changed according to whether the target area contains a block border or not by the filter coefficient vector changing means; and the fixed vector is a filter base having at least one of edge-stress effect and a blurring effect.
  • In this configuration, the fixed vector is a filter base having at least one of edge-stress effect and a blurring effect, and the filter coefficient vector changing means can change the variable coefficient according to whether the target area contains a block border or not. Therefore, with this configuration, it is possible to further attain such an effect that more appropriately adaptive filtering by changing at least one of the edge-stress effect and the blurring effect according to whether the target area contains a block border or not.
  • It is preferable in the image filter according to the present invention that the variable coefficient is changed by the filter coefficient vector changing means according to whether the block containing the target area is a block subjected to unidirectional prediction or a block subjected to bidirectional prediction; and the fixed vector is a filter base having at least one of edge-stress effect and a blurring effect.
  • In this configuration, the fixed vector is a filter base having at least one of edge-stress effect and a blurring effect, and the filter coefficient vector changing means can change the variable coefficient according to whether the block containing the target area is a block subjected to unidirectional prediction or a block subjected to bidirectional prediction. Therefore, with this configuration, it is possible to further attain such an effect that more appropriately adaptive filtering by changing at least one of the edge-stress effect and the blurring effect according to whether the block containing the target area is a block subjected to unidirectional prediction or a block subjected to bidirectional prediction.
  • It is preferable in the image filter according to the present invention that the variable coefficient is changed by the filter coefficient vector changing means according to whether or not a block including the target area is a block to which a skip mode is applied; and the fixed vector is a filter base having at least one of edge-stress effect and a blurring effect.
  • In this configuration, the fixed vector is a filter base having at least one of edge-stress effect and a blurring effect, and the filter coefficient vector changing means can change the variable coefficient is changed according to whether or not a block including the target area is a block to which a skip mode is applied. Therefore, with this configuration, it is possible to further attain such an effect that more appropriately adaptive filtering by changing at least one of the edge-stress effect and the blurring effect according to whether or not a block including the target area is a block to which a skip mode is applied.
  • It is preferable in the image filter according to the present invention that the variable coefficient is changed by the filter coefficient vector changing means according to whether a block including the target area is an inter predicted block or an intra predicted block; and the fixed vector is a filter base having at least one of edge-stress effect and a blurring effect.
  • In this configuration, the fixed vector is a filter base having at least one of edge-stress effect and a blurring effect, and the filter coefficient vector changing means can change the variable coefficient according to whether a block including the target area is an inter predicted block or an intra predicted block. Therefore, with this configuration, it is possible to further attain such an effect that more appropriately adaptive filtering by changing at least one of the edge-stress effect and the blurring effect according to whether a block including the target area is an inter predicted block or an intra predicted block.
  • It is preferable in the image filter according to the present invention that components of the fixed filter coefficient vector, and the fixed coefficient are defined to minimize a difference between a certain image and the output image, where the input image is a decoded image obtained by encoding and decoding the certain image.
  • In this configuration, the components of the fixed filter coefficient vector, and the fixed coefficient are defined to minimize the difference between a certain image and the output image, where the input image is a decoded image obtained by encoding and decoding the certain image. Therefore, with this configuration, it is possible to further attain such an effect that the filtering can be performed by using a variable coefficient vector calculated out by using the components of the fixed filter coefficient vector, and the fixed coefficient, which have been thus optimized.
  • It is preferable in the image filter according to the present invention that the fixed filter coefficient vector and the variable filter coefficient vector are orthogonal to each other.
  • With this configuration, in which the fixed filter coefficient vector and the variable filter coefficient vector are orthogonal to each other, it is possible to reduce the degree of freedom in the fixed filter coefficient vector, compared with a case where the fixed filter coefficient vector and the variable filter coefficient vector are not orthogonal to each other. Therefore, this configuration makes it possible to perform filtering more appropriately adaptive for every target area by using a smaller number of filtering coefficient.
  • Moreover, it should be noted that the fixed filter coefficient vector and the variable filter coefficient vector may or may not be orthogonal to each other in the image filter according to the present invention.
  • A decoding device according to the present invention is configured to comprise the aforementioned image filter, and cause the image filter to process a decoded image.
  • The decoding device according to the present invention with this configuration can bring about the same effect as the image filter. Moreover, the decoding device with this configuration can perform filtering more appropriately adaptive for every target area by using the components of the fixed coefficient vector (or the fixed coefficient vector) predetermined per unit area, and the fixed coefficient predetermined per unit area. Therefore, the decoding device can obtain a filter coefficient vector by using a smaller number of coefficient data, compared with the case where a fixed coefficient vector (or a fixed coefficient vector) having different components for every target area.
  • Furthermore, by using the decoding device with this configuration, it is possible to reduce code quantity for encoding the coefficient data (filter coefficient vector), compared with the case where a fixed coefficient vector (or a fixed coefficient vector) having different components for every target area.
  • Moreover, an encoding device according to the present invention is configured to comprise the aforementioned image filter, and cause the image filter to process a decoded image.
  • The encoding device according to the present invention with this configuration can bring about the same effect as the image filter. Moreover, the encoding device with this configuration can perform filtering more appropriately adaptive for every target area by using the components of the fixed coefficient vector (or the fixed coefficient vector) predetermined per unit area, and the fixed coefficient predetermined per unit area. Therefore, the encoding device can obtain a filter coefficient vector by using a smaller number of coefficient data, compared with the case where a fixed coefficient vector (or a fixed coefficient vector) having different components for every target area.
  • Furthermore, by using the encoding device with this configuration, it is possible to reduce code quantity for encoding the coefficient data (filter coefficient vector), compared with the case where a fixed coefficient vector (or a fixed coefficient vector) having different components for every target area.
  • Moreover, a data structure according to the present invention is a data structure of an encoded data, which a decoding device refers to, the decoding device (i) including an image filter for calculating a pixel value of target pixel in an output image from a pixel value of each of a pixel or pixels belonging to a target area in an input image by using a filter coefficient vector, and (ii) causing the image filter to process a decoded image, the data structure comprising: a fixed coefficient, with which the decoding device multiplies a product of a predetermined fixed vector and a variable coefficient variable according to at least either where the target area is in the input image, or where the target pixel is in the output image; and components of a fixed filter coefficient vector to which the decoding device adds a product of the variable coefficient, the fixed vector, and the fixed coefficient, the fixed filter coefficient vector being predetermined for every unit area constituting the output image.
  • When the encoded data has the data structure configured as above, the decoding device for decoding an image based on the encoded data can perform filtering by using the components of the fixed coefficient vector, and the fixed coefficient contained in the encoded data, thereby performing filtering more appropriately adaptive for every target area by using a smaller number of filter coefficient, compared with the case where the encoded data contains different filter coefficients for the target area, respectively.
  • The invention being thus described, it will be obvious that the same way may be varied in many ways. Such variations are not to be regarded as a departure from the spirit and scope of the invention, and all such modifications as would be obvious to one skilled in the art are intended to be included within the scope of the following claims.
  • INDUSTRIAL APPLICABILITY
  • The present invention is suitably applicable to an image filter for filtering image data. Moreover, the present invention is also suitably applicable to a broadcast receiving terminal configured to decode encoded image data, a HDD recorder configured to encode image data and record the encoded image data in a recording medium, and to decode image data recorded in a recording medium.
  • REFERENCE SIGNS LIST
    • 100: Adaptive Filter (Image Filter)
    • 110: Filter Section
    • 120: Control Variable Determining Section (Filter Coefficient Vector Changing Means)
    • 1: Moving Image Encoding Device (Encoding Device)
    • 2: Moving Image Decoding Device (Decoding Device)
    • 16, 26: Inter Predicted Image Generating Section
    • R: Filter Reference Area
    • T: Filter Target Area

Claims (17)

1-16. (canceled)
17. An image filter for calculating a pixel value of target pixel in an output image from a pixel value of each of a pixel or pixels belonging to a target area in an input image, from a filter coefficient vector, and from an offset, comprising: filter coefficient changing means for changing the filter coefficient and the offset according to at least either where the target area is spacially in the input image, or where the target pixel is spacially in the output image.
18. The image filter as set forth in claim 17, wherein the pixel value is a luminance value, and the filter coefficient changing means classifies the target area according to the luminance value, and sets, as the offset, a value determined by a result of the classification.
19. The image filter as set forth in claim 18, wherein the target area in the input image includes only one pixel.
20. The image filter as set forth in claim 17, wherein a filter coefficient vector, whose components are the filter coefficient and the offset, is dividable into a sum of (i) a fixed filter coefficient vector predetermined for every unit area constituting the output image, and (ii) a variable filter coefficient vector variable by the filter coefficient changing means according to at least either where the target area is spacially in the input image, or where the target pixel is spacially in the output image.
21. The image filter as set forth in claim 20, wherein the variable filter coefficient vector is further dividable into (i) a fixed coefficient predetermined for every unit area constituting the output image, (ii) a variable coefficient variable by the filter coefficient changing means according to at least either where the target area is spacially in the input image, or where the target pixel is spacially in the output image, and (iii) a predetermined fixed vector.
22. The image filter as set forth in claim 21, wherein:
the variable coefficient is changed according to a degree of flatness in an image in the target area by the filter coefficient changing means; and
the fixed vector is a filter base having at least one of edge-stress effect and a blurring effect.
23. The image filter as set forth in claim 21, wherein:
the variable coefficient is changed according to an edge angle of an edge in an image in the target area by the filter coefficient changing means; and
the fixed vector is a filter base having at least one of edge-stress effect and a blurring effect.
24. The image filter as set forth in claim 21, wherein:
the variable coefficient is changed according to luminance of an image in the target area by the filter coefficient changing means; and
the fixed vector is a filter base having at least one of edge-stress effect and a blurring effect.
25. The image filter as set forth in claim 21, wherein:
the variable coefficient is changed by the filter coefficient changing means according to whether or not a block including the target area is a block to which a skip mode is applied; and
the fixed vector is a filter base having at least one of edge-stress effect and a blurring effect.
26. The image filter as set forth in claim 21, wherein:
the variable coefficient is changed by the filter coefficient changing means according to whether a block including the target area is an inter predicted block or an intra predicted block; and
the fixed vector is a filter base having at least one of edge-stress effect and a blurring effect.
27. The image filter as set forth in claim 21, wherein components of the fixed filter coefficient vector, and the fixed coefficient are defined to minimize a difference between a certain image and the output image, where the input image is a decoded image obtained by encoding and decoding the certain image.
28. The image filter as set forth in claim 20, wherein the fixed filter coefficient vector and the variable filter coefficient vector are orthogonal to each other.
29. A decoding device comprising an image filter as set forth in claim 17, and causing the image filter to process a decoded image.
30. An encoding device comprising an image filter as set forth in claim 17, and causing the image filter to process a locally decoded image.
31. A data structure of an encoded data, which a decoding device refers to, the decoding device (i) including an image filter for calculating a pixel value of target pixel in an output image from a pixel value of each of a pixel or pixels belonging to a target area in an input image by using a filter coefficient vector, and (ii) causing the image filter to process a decoded image, the data structure comprising:
a fixed coefficient, with which the decoding device multiplies a product of a predetermined fixed vector and a variable coefficient variable according to at least either where the target area is spacially in the input image, or where the target pixel is spacially in the output image; and
components of a fixed filter coefficient vector to which the decoding device adds a product of the variable coefficient, the fixed vector, and the fixed coefficient, the fixed filter coefficient vector being predetermined for every unit area constituting the output image.
32. A data structure of encoded data, which a decoding device refers to, the decoding device (i) including an image filter for calculating a luminance value of a target pixel in an output pixel from a luminance value of each of a pixel or pixels belonging to an input image, from a filter coefficient, and from an offset, and (ii) causing the image filter to process a decoded image, wherein:
the offset is set according to the luminance value; and
the encoded data includes the offset.
US13/515,957 2009-12-18 2010-11-10 Image filter, encoding device, decoding device, and data structure Abandoned US20120251012A1 (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
JP2009288448 2009-12-18
JP2009-288448 2009-12-18
PCT/JP2010/070053 WO2011074357A1 (en) 2009-12-18 2010-11-10 Image filter, encoding device, decoding device, and data structure

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2010/070053 A-371-Of-International WO2011074357A1 (en) 2009-12-18 2010-11-10 Image filter, encoding device, decoding device, and data structure

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US14/697,400 Continuation US9514519B2 (en) 2009-12-18 2015-04-27 Image filter

Publications (1)

Publication Number Publication Date
US20120251012A1 true US20120251012A1 (en) 2012-10-04

Family

ID=44167122

Family Applications (3)

Application Number Title Priority Date Filing Date
US13/515,957 Abandoned US20120251012A1 (en) 2009-12-18 2010-11-10 Image filter, encoding device, decoding device, and data structure
US14/697,400 Active US9514519B2 (en) 2009-12-18 2015-04-27 Image filter
US15/334,048 Active US9641865B2 (en) 2009-12-18 2016-10-25 Method for decoding moving images

Family Applications After (2)

Application Number Title Priority Date Filing Date
US14/697,400 Active US9514519B2 (en) 2009-12-18 2015-04-27 Image filter
US15/334,048 Active US9641865B2 (en) 2009-12-18 2016-10-25 Method for decoding moving images

Country Status (9)

Country Link
US (3) US20120251012A1 (en)
EP (2) EP3301932A1 (en)
JP (5) JPWO2011074357A1 (en)
KR (5) KR20180038076A (en)
CN (1) CN107071480A (en)
AU (1) AU2010331511B2 (en)
BR (1) BR112012014685A2 (en)
CA (2) CA3038111C (en)
WO (1) WO2011074357A1 (en)

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150172666A1 (en) * 2012-04-16 2015-06-18 Mediatek Inc. Method and apparatus for sample adaptive offset coding with separate sign and mag -nitude
US20150264406A1 (en) * 2014-03-14 2015-09-17 Qualcomm Incorporated Deblock filtering using pixel distance
US20150278156A1 (en) * 2012-11-05 2015-10-01 Denso Corporation Relatedness determination device, non-transitory tangible computer-readable medium for the same, and relatedness determination method
US20160094823A1 (en) * 2014-09-30 2016-03-31 Apple Inc. Yuv signal generation for high dynamic range video content
US9497455B2 (en) 2011-06-23 2016-11-15 Huawei Technologies Co., Ltd. Offset decoding device, offset coding device, image filtering device, and data structure
US20170162076A1 (en) * 2014-09-03 2017-06-08 Aira Tech Corporation Methods, apparatus and systems for providing remote assistance for visually-impaired users
US10346955B2 (en) * 2016-08-05 2019-07-09 Realtek Semiconductor Corporation Image-filtering method and image-filtering device
US11100613B2 (en) * 2017-01-05 2021-08-24 Zhejiang Dahua Technology Co., Ltd. Systems and methods for enhancing edges in images
US11166025B2 (en) * 2018-03-30 2021-11-02 Panasonic Intellectual Property Corporation Of America Encoder, decoder, encoding method, and decoding method

Families Citing this family (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2525576A4 (en) * 2010-01-12 2015-04-15 Sharp Kk Encoder apparatus, decoder apparatus, and data structure
JP5789172B2 (en) * 2011-10-20 2015-10-07 日本放送協会 Image processing apparatus and program
JP5856580B2 (en) * 2013-03-14 2016-02-10 日本電信電話株式会社 Signal processing apparatus and signal processing method
CN116708784A (en) * 2016-07-12 2023-09-05 韩国电子通信研究院 Image encoding/decoding method and recording medium therefor
CN115714862A (en) * 2017-05-31 2023-02-24 交互数字麦迪逊专利控股公司 Method and apparatus for picture coding and decoding
JPWO2019107182A1 (en) * 2017-12-01 2020-11-26 ソニー株式会社 Encoding device, coding method, decoding device, and decoding method
KR102028908B1 (en) 2018-03-15 2019-11-08 써멀마스터 주식회사 A Heat Exchanger Having a Built-in Electrical Heater
JP7073186B2 (en) * 2018-05-14 2022-05-23 シャープ株式会社 Image filter device
JP7332753B2 (en) 2018-05-14 2023-08-23 シャープ株式会社 image filter device

Citations (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040179743A1 (en) * 1998-03-05 2004-09-16 Hideaki Shibata Image coding method, image coding and decoding method, image coding apparatus, and image recording and reproduction apparatus
US20050265452A1 (en) * 2004-05-27 2005-12-01 Zhourong Miao Temporal classified filtering for video compression
US20060007239A1 (en) * 2004-07-06 2006-01-12 Harrison Charles F Color correction system
US20060182356A1 (en) * 2000-12-06 2006-08-17 Karl Lillevold Video compression and decompression system with postfilter to filter coding artifacts
US20080013840A1 (en) * 2006-05-01 2008-01-17 Fuji Xerox Co., Ltd. Decoding apparatus, dequantization method, computer readable medium storing program thereof, and computer data signal
US20080068682A1 (en) * 2006-09-19 2008-03-20 Canon Kabushiki Kaisha Image reading apparatus and image reading method
US20080101719A1 (en) * 2006-10-30 2008-05-01 Samsung Electronics Co., Ltd. Image enhancement method and system
US20080247467A1 (en) * 2007-01-09 2008-10-09 Nokia Corporation Adaptive interpolation filters for video coding
US20090087116A1 (en) * 2007-09-28 2009-04-02 Fujitsu Microelectronics Limited Image processing filter, image processing method of image processing filter, and image processing circuit of image processing apparatus having image processing filter
US20090257670A1 (en) * 2008-04-09 2009-10-15 Yi-Jen Chiu In-loop adaptive wiener filter for video coding and decoding
US20100021071A1 (en) * 2007-01-09 2010-01-28 Steffen Wittmann Image coding apparatus and image decoding apparatus
US20100020205A1 (en) * 2006-09-14 2010-01-28 Kozo Ishida Image processing apparatus and imaging apparatus and method
US20100098345A1 (en) * 2007-01-09 2010-04-22 Kenneth Andersson Adaptive filter representation
US20100329335A1 (en) * 2008-04-30 2010-12-30 Goki Yasuda Video encoding and decoding apparatus
US20110019096A1 (en) * 2009-07-21 2011-01-27 Louie Lee Method and system for detection and enhancement of video images
US20110043668A1 (en) * 2009-08-24 2011-02-24 Mckinnon Patrick R Detection of Defective Pixels in an Image Sensor
US20110069752A1 (en) * 2008-04-30 2011-03-24 Takashi Watanabe Moving image encoding/decoding method and apparatus with filtering function considering edges

Family Cites Families (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3288811B2 (en) * 1993-07-26 2002-06-04 日本無線株式会社 Post-processing filter control method and post-processing filter control circuit in video codec
JPH08163561A (en) * 1994-12-06 1996-06-21 Matsushita Electric Ind Co Ltd Picture data compression device
JPH0998417A (en) * 1995-09-29 1997-04-08 Matsushita Electric Ind Co Ltd Picture communication equipment and picture communication method
KR100244290B1 (en) 1997-09-09 2000-02-01 구자홍 Method for deblocking filtering for low bit rate video
JP2001275110A (en) * 2000-03-24 2001-10-05 Matsushita Electric Ind Co Ltd Method and system for dynamic loop and post filtering
EP1978748B1 (en) 2001-11-29 2014-04-16 Panasonic Corporation Coding distortion removal method
EP2938072A1 (en) 2001-11-29 2015-10-28 Godo Kaisha IP Bridge 1 Coding distortion removal method
JP2007143178A (en) * 2001-11-29 2007-06-07 Matsushita Electric Ind Co Ltd Method for removing coding distortion
BR0304545A (en) * 2002-01-14 2004-11-03 Nokia Corp Method of encoding images in a digital video sequence to provide encoded video data, video encoder, decoding method of data indicative of a digital video sequence, video decoder, and video decoding system
US7773158B2 (en) * 2005-10-12 2010-08-10 Panasonic Corporation Visual processing device, display device, and integrated circuit
EP2001239B1 (en) * 2006-03-27 2017-09-13 Sun Patent Trust Picture coding apparatus and picture decoding apparatus
JP4872862B2 (en) * 2006-09-28 2012-02-08 ソニー株式会社 Image data arithmetic device and method, program, and recording medium
KR100922275B1 (en) 2006-12-15 2009-10-15 경희대학교 산학협력단 Derivation process of a boundary filtering strength and deblocking filtering method and apparatus using the derivation process
MX2010009194A (en) * 2008-03-07 2010-09-10 Toshiba Kk Dynamic image encoding/decoding method and device.
JP5321439B2 (en) * 2009-12-15 2013-10-23 株式会社Jvcケンウッド Image encoding device, image decoding device, image encoding method, and image decoding method
US9237355B2 (en) * 2010-02-19 2016-01-12 Qualcomm Incorporated Adaptive motion resolution for video coding
KR20120005968A (en) * 2010-07-09 2012-01-17 삼성전자주식회사 Method and apparatus for video encoding using adjustable loop-filtering, method and apparatus for video dncoding using adjustable loop-filtering
PL3876536T3 (en) * 2012-09-26 2023-11-06 Sun Patent Trust Image decoding method, image decoding device, and computer-readable medium

Patent Citations (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040179743A1 (en) * 1998-03-05 2004-09-16 Hideaki Shibata Image coding method, image coding and decoding method, image coding apparatus, and image recording and reproduction apparatus
US20060182356A1 (en) * 2000-12-06 2006-08-17 Karl Lillevold Video compression and decompression system with postfilter to filter coding artifacts
US20050265452A1 (en) * 2004-05-27 2005-12-01 Zhourong Miao Temporal classified filtering for video compression
US20060007239A1 (en) * 2004-07-06 2006-01-12 Harrison Charles F Color correction system
US20080013840A1 (en) * 2006-05-01 2008-01-17 Fuji Xerox Co., Ltd. Decoding apparatus, dequantization method, computer readable medium storing program thereof, and computer data signal
US20100020205A1 (en) * 2006-09-14 2010-01-28 Kozo Ishida Image processing apparatus and imaging apparatus and method
US20080068682A1 (en) * 2006-09-19 2008-03-20 Canon Kabushiki Kaisha Image reading apparatus and image reading method
US20080101719A1 (en) * 2006-10-30 2008-05-01 Samsung Electronics Co., Ltd. Image enhancement method and system
US20100098345A1 (en) * 2007-01-09 2010-04-22 Kenneth Andersson Adaptive filter representation
US20100021071A1 (en) * 2007-01-09 2010-01-28 Steffen Wittmann Image coding apparatus and image decoding apparatus
US20080247467A1 (en) * 2007-01-09 2008-10-09 Nokia Corporation Adaptive interpolation filters for video coding
US20090087116A1 (en) * 2007-09-28 2009-04-02 Fujitsu Microelectronics Limited Image processing filter, image processing method of image processing filter, and image processing circuit of image processing apparatus having image processing filter
US20090257670A1 (en) * 2008-04-09 2009-10-15 Yi-Jen Chiu In-loop adaptive wiener filter for video coding and decoding
US20100329335A1 (en) * 2008-04-30 2010-12-30 Goki Yasuda Video encoding and decoding apparatus
US20110069752A1 (en) * 2008-04-30 2011-03-24 Takashi Watanabe Moving image encoding/decoding method and apparatus with filtering function considering edges
US20110019096A1 (en) * 2009-07-21 2011-01-27 Louie Lee Method and system for detection and enhancement of video images
US20110043668A1 (en) * 2009-08-24 2011-02-24 Mckinnon Patrick R Detection of Defective Pixels in an Image Sensor

Cited By (23)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10440362B2 (en) 2011-06-23 2019-10-08 Huawei Technologies Co., Ltd. Offset decoding device, offset coding device, image filtering device
US10477209B2 (en) 2011-06-23 2019-11-12 Huawei Technologies Co., Ltd. Offset decoding device, offset coding device, image filtering device
US20180098067A1 (en) 2011-06-23 2018-04-05 Huawei Technologies Co., Ltd. Offset decoding device, offset coding device, image filtering device
US11089302B2 (en) 2011-06-23 2021-08-10 Huawei Technologies Co., Ltd. Offset decoding device, offset coding device, image filtering device
US9497455B2 (en) 2011-06-23 2016-11-15 Huawei Technologies Co., Ltd. Offset decoding device, offset coding device, image filtering device, and data structure
US10764580B2 (en) 2011-06-23 2020-09-01 Huawei Technologies Co., Ltd. Offset decoding device, offset coding device, image filtering device
US10390012B2 (en) 2011-06-23 2019-08-20 Huawei Technologies Co., Ltd. Offset decoding device, offset coding device, image filtering device
US11765350B2 (en) 2011-06-23 2023-09-19 Huawei Technologies Co., Ltd. Offset decoding device, offset coding device, image filtering device
US9866833B2 (en) 2011-06-23 2018-01-09 Huawei Technologies Co., Ltd. Offset decoding device, offset coding device, image filtering device
US20150172666A1 (en) * 2012-04-16 2015-06-18 Mediatek Inc. Method and apparatus for sample adaptive offset coding with separate sign and mag -nitude
US20150278156A1 (en) * 2012-11-05 2015-10-01 Denso Corporation Relatedness determination device, non-transitory tangible computer-readable medium for the same, and relatedness determination method
US10409886B2 (en) * 2012-11-05 2019-09-10 Denso Corporation Relatedness determination device, non-transitory tangible computer-readable medium for the same, and relatedness determination method
US20150264406A1 (en) * 2014-03-14 2015-09-17 Qualcomm Incorporated Deblock filtering using pixel distance
US9906790B2 (en) * 2014-03-14 2018-02-27 Qualcomm Incorporated Deblock filtering using pixel distance
US9836996B2 (en) * 2014-09-03 2017-12-05 Aira Tech Corporation Methods, apparatus and systems for providing remote assistance for visually-impaired users
US20170162076A1 (en) * 2014-09-03 2017-06-08 Aira Tech Corporation Methods, apparatus and systems for providing remote assistance for visually-impaired users
US10777097B2 (en) 2014-09-03 2020-09-15 Aira Tech Corporation Media streaming methods, apparatus and systems
US10078971B2 (en) 2014-09-03 2018-09-18 Aria Tech Corporation Media streaming methods, apparatus and systems
US9716871B2 (en) * 2014-09-30 2017-07-25 Apple Inc. YUV signal generation for high dynamic range video content
US20160094823A1 (en) * 2014-09-30 2016-03-31 Apple Inc. Yuv signal generation for high dynamic range video content
US10346955B2 (en) * 2016-08-05 2019-07-09 Realtek Semiconductor Corporation Image-filtering method and image-filtering device
US11100613B2 (en) * 2017-01-05 2021-08-24 Zhejiang Dahua Technology Co., Ltd. Systems and methods for enhancing edges in images
US11166025B2 (en) * 2018-03-30 2021-11-02 Panasonic Intellectual Property Corporation Of America Encoder, decoder, encoding method, and decoding method

Also Published As

Publication number Publication date
JP6159376B2 (en) 2017-07-05
JP5834112B2 (en) 2015-12-16
KR101849891B1 (en) 2018-04-17
US20170041636A1 (en) 2017-02-09
US20150242999A1 (en) 2015-08-27
EP3301932A1 (en) 2018-04-04
BR112012014685A2 (en) 2016-04-05
CN107071480A (en) 2017-08-18
KR20190031601A (en) 2019-03-26
JP2018142978A (en) 2018-09-13
CA2784291A1 (en) 2011-06-23
EP2515541A1 (en) 2012-10-24
US9514519B2 (en) 2016-12-06
US9641865B2 (en) 2017-05-02
JP2016036169A (en) 2016-03-17
CA3038111A1 (en) 2011-06-23
CN102656888A (en) 2012-09-05
JPWO2011074357A1 (en) 2013-04-25
KR20120105509A (en) 2012-09-25
KR101729903B1 (en) 2017-04-24
KR20160102085A (en) 2016-08-26
KR20170045361A (en) 2017-04-26
WO2011074357A1 (en) 2011-06-23
AU2010331511A1 (en) 2012-07-26
JP2017200198A (en) 2017-11-02
AU2010331511B2 (en) 2015-01-15
EP2515541A4 (en) 2015-08-26
CA3038111C (en) 2022-08-09
KR102060586B1 (en) 2019-12-30
JP2014209760A (en) 2014-11-06
KR20180038076A (en) 2018-04-13
JP6335365B2 (en) 2018-05-30

Similar Documents

Publication Publication Date Title
US9641865B2 (en) Method for decoding moving images
US11758153B2 (en) Weighted angular prediction for intra coding
US20210227242A1 (en) Unequal weight planar prediction
US20220417531A1 (en) System and method of implementing multiple prediction models for local illumination compensation
CN111279703A (en) Post-filtering for weighted angle prediction
US20120300849A1 (en) Encoder apparatus, decoder apparatus, and data structure
WO2019074985A1 (en) Adaptive unequal weight planar prediction
US11595698B2 (en) System and method for deblocking HDR content
US11956477B2 (en) System and method for deblocking HDR content
WO2021040037A1 (en) Encoding apparatus, decoding apparatus, and program

Legal Events

Date Code Title Description
AS Assignment

Owner name: SHARP KABUSHIKI KAISHA, JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:IKAI, TOMOHIRO;YASUGI, YUKINOBU;AONO, TOMOKO;REEL/FRAME:028386/0264

Effective date: 20120604

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION