WO1996018264A1 - Digital image anti-aliasing - Google Patents

Digital image anti-aliasing Download PDF

Info

Publication number
WO1996018264A1
WO1996018264A1 PCT/IB1995/001050 IB9501050W WO9618264A1 WO 1996018264 A1 WO1996018264 A1 WO 1996018264A1 IB 9501050 W IB9501050 W IB 9501050W WO 9618264 A1 WO9618264 A1 WO 9618264A1
Authority
WO
WIPO (PCT)
Prior art keywords
pixel
image
pixels
slope
plane
Prior art date
Application number
PCT/IB1995/001050
Other languages
French (fr)
Inventor
Vassilis Seferidis
Original Assignee
Philips Electronics N.V.
Philips Norden Ab
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Philips Electronics N.V., Philips Norden Ab filed Critical Philips Electronics N.V.
Publication of WO1996018264A1 publication Critical patent/WO1996018264A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/222Studio circuitry; Studio devices; Studio equipment
    • H04N5/262Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
    • H04N5/2622Signal amplitude transition in the zone between image portions, e.g. soft edges
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2340/00Aspects of display data processing
    • G09G2340/10Mixing of images, i.e. displayed pixel being the result of an operation, e.g. adding, on the corresponding input pixels

Definitions

  • the present invention relates to the coding of composite digital images and in particular to dealing with aliasing effects at the edges of foreground features overlying another object at greater image depth or a background.
  • multi-plane mixers which may, for example, combine a background video image with one or more overlaid planes containing animated graphics.
  • Such a mixer accepts various sources of video data together with their associated depth maps and produces a new picture frame based on depth and transparency values of each pixel.
  • Various coding schemes have been proposed for coding both the depth and transparency data in order to minimise the amount of data required to be stored.
  • transparency data it will be appreciated that the necessary data is greatly reduced when large areas of a picture have the same transparency or translucency value due to special effects such as fog or mist, reflections or diffractions, and fading in or out.
  • Transparency values can also be used to reduce aliasing effects by blending at the edges of overlapping shapes.
  • a problem which occurs, however, is that for complex images, a large amount of transparency information has to be encoded along with the picture data and, within the constraints of video systems, this can present a severe problem.
  • a method for display of images in the form of an image object defined in a first pixel plane overlying a background defined in a further pixel plane wherein those pixels falling wholly within the object are displayed with the pixel value defined in the first pixel plane, those pixels falling wholly outside the object are displayed with the pixel value defined in the further pixel plane, and the remaining edge pixels are displayed with a transparency value comprising contributions from the corresponding pixel in both the first and further planes, characterised in that the transparency values are derived by: identifying step discontinuities in the image edge; - applying a slope characteristic across each step discontinuity; for each pixel crossed by the slope characteristic, deriving the relative areas of the pixel above and below the slope; and setting the relative contributions of the foreground and background pixels accordingly.
  • transparencies which can be generated at display time, across the slope discontinuities (the steps on an aliased edge)
  • the effects of aliasing can be reduced without requiring the pre-calculation and storage of transparency values.
  • one or more of the transparency values may be pre-generated and stored with the image pixel data to be read from storage when the image is displayed, the remainder of the transparency values being calculated when the image is displayed.
  • This partial storage may be used where system constraints limit the facility for display time calculation of transparencies.
  • the transparency values may be pre- generated and stored for those step discontinuities having a vertical extent greater than or equal to horizontal extent, that is to say for those image edges having an edge slope of 45 degrees to greater to the horizontal.
  • the step of derivation of relative areas for a pixel may comprise integration of the slope characteristic across the pixel to produce a first area value followed by subtraction of this from the total pixel area to give a second area value.
  • the derivation may comprise dividing each pixel in a ratio given by the position of that pixel and the number of pixels across the step discontinuity.
  • the slope characteristic may be a straight line or a curve function, such as an elliptical or parabolic curve.
  • a video mixing apparatus operable to form an output image signal from first and second pixel plane signals and a matte signal identifying, for each pixel of the first plane, whether or not it comprises a part of an image object defined in the first plane to be displayed overlying a background defined in the second pixel plane
  • the apparatus comprising: an edge classification stage arranged to receive the matte signal and identify therefrom step discontinuities at the edge of the image object; first calculation means arranged to apply a predetermined slope characteristic across step discontinuities identified by the edge classification stage; second calculation means arranged to determine, for each pixel crossed by the slope characteristic, the relative areas of the pixel above and below the slope; and mixer stage coupled to receive the first and second pixel plane signals and an indication of relative areas from the second calculation means, and arranged to output a display pixel image signal in which the values of pixels falling wholly within the image object are taken from the first pixel plane signal, the values of pixels falling wholly outside the image object are taken from the second pixel plane signal, and the values of
  • the above- mentioned first calculation means may be operable to apply a selected one from a plurality of stored slope characteristics (for example elliptical or parabolic as mentioned above) with the apparatus further comprising a function store holding implementation instructions for each of the different characteristics supported.
  • a plurality of stored slope characteristics for example elliptical or parabolic as mentioned above
  • Figure 1 is a block schematic diagram of a decoder apparatus including a mixer stage embodying the present invention
  • Figure 2 is a schematic diagram of a stage of the mixer of Figure
  • Figure 3 shows a cascaded arrangement of mixer stages as in Figure 2;
  • Figure 4 shows an aliased image produced by a known technique
  • Figure 5 shows image pixels stored for use in the technique producing Figure 4;
  • Figure 6 illustrates edge pixel contributions over two horizontal scan lines
  • Figure 7 shows the image of Figure 4 to which a first anti-aliasing technique has been applied
  • Figure 8 shows the pixels stored for use in the method producing the image of Figure 7.
  • Figure 9 shows the image of Figure 4 to which a second anti- aliasing technique has been applied.
  • Figure 1 shows a general arrangement for a multi-plane video image system coupled to the host bus 10 of a CD-i (Compact Disc interactive) player.
  • the base case video which may be used to form a background to the final image, is fed to a video mixer 12, together with the video synchronising signals.
  • a graphics processor 14 and a digital video decoder 16 are also providing inputs to the video mixer.
  • the graphics processor and digital video decoder take their inputs from the host bus 10 which also provides control signals to the mixer 12.
  • the output of the mixer stage comprises digital signals for the final image colour components (R,G,B) together with the video synchronisation signals. Following conversion by a digital to analogue converter 18, the analogue video signal is output on connector 20.
  • a first stage of the video mixer 12 is shown in Figure 2 having inputs for the base case video 22 and for the video image 24 of an object to be overlaid on the background by combining in a mixer 26.
  • a matte 28 which is simple binary representation of the object, suitably using 1 to represent those pixels which are part of the image and 0 to represent all other pixels within the video frame.
  • the matte signal is used by the mixer to select either pixels from the image or pixels from the base case to make up the output picture. This technique is, however, prone to producing aliased images (that is to say the well known "staircase" effect on sloping edges) due to the whole pixel nature of the transition between the image and the background.
  • an edge classification stage 30 uses the matte signal to derive a model of the edge of the object (by noting the transition from 0 to 1 in the matte) and identifies the steps in the aliased edge to a calculation circuit 32.
  • the calculation circuit applies a function to the values of pixels forming this step, as will be described below, and passes the result of this calculation to a further calculation stage 34 which derives a transparency value to be applied to these edge pixels.
  • These transparency values are supplied to the mixer 26 which outputs a total picture 38 in which those pixels which are purely object or purely background have their original input values, and the edge pixels have a value which contains a contribution from both in amounts determined by the calculation carried out at stages 32 and 34.
  • a store of different function models 36 may be provided.
  • mixer stages as in Figure 2 are cascaded in the manner shown in Figure 3 (note that the edge classification and calculation stages of Figure 2 are omitted for reasons of clarity).
  • the background image 22 is combined through use of the appropriate matte 28 with the input image 24 having the greatest depth (Z) value to generate a composite picture.
  • This composite picture is then combined in a further mixer 26a with an object image 28a having a depth value less than that for the first image object and by use of a further matte 28a.
  • Subsequent images 24b with associated matte 28b are added in like manner such that the final picture is built up from the background and then through the image objects in descending order of depth.
  • a known anti ⁇ aliasing technique will now be described, as used to calculate the transparency value by averaging (low-pass filtering) the values of neighbouring pixels within a small window (for example 3 x 3 or 5 x 5 pixels) for each of the edge pixels.
  • the anti-aliased effect may be improved if an original higher resolution image is used: for example, an original image consisting of 720 x 576 pixels may be reduced by averaging four neighbouring pixels and then thresholding the resulting values to provide a sub-sampled version of the original image with a 360 x 288 pixel resolution.
  • An aliased picture for such a reduced resolution image is shown in Figure 4.
  • a first embodiment of the proposed method for anti-aliasing estimates the slope (i.e. the step height of the aliased edge) from two or more successive scan lines and then calculates for each pixel the area defined by this slope and the horizontal or vertical orientation.
  • the area under the slope which is used to dictate the relative contributions of the foreground and background pixels, is determined by the calculation unit 32 of Figure 2.
  • a slope characteristic other than a straight line gives improved results, for example an exponential or parabolic curve. Models of such different characteristics would be stored in the function model stage 36 of Figure 2, and the calculation stage 32 would comprise an integrator that determines the area under the particular characteristic chosen.
  • Figure 6 illustrates an implementation of this first embodiment using a straight line characteristic over two successive scan lines which differ by four pixels (that is to say a step four pixels wide appears at the edge of the aliased image).
  • the dark shaded area of Figure 6 represents the coverage of each pixel by an object with a near horizontal top edge and thus can be used to calculate its value.
  • Simple geometry shows that, given the difference N between the number of pixels across the two lines, the value of each pixel can be calculated by equation 1 below:
  • a second embodiment of the method trades a degree of accuracy for the calculation against simplicity of hardware implementation by using estimation of the required step for a pixel displacement. This is simply obtained from the ratio of pixels in a line 2N, leading to equation 2 below:
  • the slope estimation method described above is a general technique in the sense that it can be used for both near-horizontal and near-vertical edges
  • hardware and economic constraints may dictate that only a few scan lines of picture are available to the mixer at any given time.
  • the slope estimation can be realistically used only for near-horizontal edges.
  • Figure 7, for example shows the implementation of the slope estimation method on all the edges of the picture of Figure 4 but within the constraint of a maximum of two adjacent scan lines being available at any time.
  • a two-bit (four level) transparency value has been used for each anti-aliased pixel. It will be noted that although the near-horizontal lines are anti-aliased satisfactorily, this is not the case for the near vertical lines which remain aliased.
  • the preferred technique calculates the transparency value based on the contribution of pixels from a higher resolution (e.g 720 x 576 pixels) video frame to obtain the 360 x 288 standard format picture, with the transparency value for an edge pixel in the lower resolution picture being the average of the four pixels in the higher one.
  • FIG. 9 shows the final anti-aliased picture obtained from the application of the slope estimation method for the horizontal edges and the use of the stored transparency values in the vertical direction.
  • Each stored value is represented with only one bit (two transparency levels).

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Image Generation (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

A method is provided for the display of images in the form of an image object defined in a first pixel plane overlying a background defined in a further pixel plane. In order to counteract aliasing effects at display time, the edge pixels at the boundary of object and background are displayed with a transparency value comprising contributions from the corresponding pixel in both the first and further planes. Step discontinuities in the image edge have a slope characteristic applied across them and, for each pixel crossed by the slope characteristic, the relative areas of the pixel above and below the slope are calculated. The relative contributions of the image and background pixels are then set according to this relation. The slope characteristic may be a straight line (as shown) or some other suitable function such as a parabolic curve.

Description

64 PCMB95/01050
DESCRIPTION
DIGITAL IMAGE ANTI-ALIASING
The present invention relates to the coding of composite digital images and in particular to dealing with aliasing effects at the edges of foreground features overlying another object at greater image depth or a background.
One particular area where such aliasing problems occur is in multi-plane mixers which may, for example, combine a background video image with one or more overlaid planes containing animated graphics. Such a mixer accepts various sources of video data together with their associated depth maps and produces a new picture frame based on depth and transparency values of each pixel. Various coding schemes have been proposed for coding both the depth and transparency data in order to minimise the amount of data required to be stored. In terms of transparency data, it will be appreciated that the necessary data is greatly reduced when large areas of a picture have the same transparency or translucency value due to special effects such as fog or mist, reflections or diffractions, and fading in or out. Transparency values can also be used to reduce aliasing effects by blending at the edges of overlapping shapes. A problem which occurs, however, is that for complex images, a large amount of transparency information has to be encoded along with the picture data and, within the constraints of video systems, this can present a severe problem.
Accordingly, it is an object of the present invention to provide a method for anti-aliasing which reduces or removes the need for storage of transparency values with image data and which is capable of implementation at or close to real time in generating composite images for display.
In accordance with the present invention there is provided a method for display of images in the form of an image object defined in a first pixel plane overlying a background defined in a further pixel plane, wherein those pixels falling wholly within the object are displayed with the pixel value defined in the first pixel plane, those pixels falling wholly outside the object are displayed with the pixel value defined in the further pixel plane, and the remaining edge pixels are displayed with a transparency value comprising contributions from the corresponding pixel in both the first and further planes, characterised in that the transparency values are derived by: identifying step discontinuities in the image edge; - applying a slope characteristic across each step discontinuity; for each pixel crossed by the slope characteristic, deriving the relative areas of the pixel above and below the slope; and setting the relative contributions of the foreground and background pixels accordingly. By applying transparencies, which can be generated at display time, across the slope discontinuities (the steps on an aliased edge), the effects of aliasing can be reduced without requiring the pre-calculation and storage of transparency values. As will be described below, however, one or more of the transparency values may be pre-generated and stored with the image pixel data to be read from storage when the image is displayed, the remainder of the transparency values being calculated when the image is displayed. This partial storage may be used where system constraints limit the facility for display time calculation of transparencies. The transparency values may be pre- generated and stored for those step discontinuities having a vertical extent greater than or equal to horizontal extent, that is to say for those image edges having an edge slope of 45 degrees to greater to the horizontal. The step of derivation of relative areas for a pixel may comprise integration of the slope characteristic across the pixel to produce a first area value followed by subtraction of this from the total pixel area to give a second area value. Alternatively, the derivation may comprise dividing each pixel in a ratio given by the position of that pixel and the number of pixels across the step discontinuity.
The slope characteristic may be a straight line or a curve function, such as an elliptical or parabolic curve.
Also in accordance with the present invention there is provided a video mixing apparatus operable to form an output image signal from first and second pixel plane signals and a matte signal identifying, for each pixel of the first plane, whether or not it comprises a part of an image object defined in the first plane to be displayed overlying a background defined in the second pixel plane, the apparatus comprising: an edge classification stage arranged to receive the matte signal and identify therefrom step discontinuities at the edge of the image object; first calculation means arranged to apply a predetermined slope characteristic across step discontinuities identified by the edge classification stage; second calculation means arranged to determine, for each pixel crossed by the slope characteristic, the relative areas of the pixel above and below the slope; and mixer stage coupled to receive the first and second pixel plane signals and an indication of relative areas from the second calculation means, and arranged to output a display pixel image signal in which the values of pixels falling wholly within the image object are taken from the first pixel plane signal, the values of pixels falling wholly outside the image object are taken from the second pixel plane signal, and the values of those pixels identified as edge pixels by the edge classification stage are taken from both first and second pixel planes in a ratio by the said indication from the second calculation means.
In order to improve the flexibility of operation, the above- mentioned first calculation means may be operable to apply a selected one from a plurality of stored slope characteristics (for example elliptical or parabolic as mentioned above) with the apparatus further comprising a function store holding implementation instructions for each of the different characteristics supported.
Preferred embodiments of the present invention will now be described, by way of example only, and with reference to the following drawings, in which:
Figure 1 is a block schematic diagram of a decoder apparatus including a mixer stage embodying the present invention; Figure 2 is a schematic diagram of a stage of the mixer of Figure
1 ;
Figure 3 shows a cascaded arrangement of mixer stages as in Figure 2;
Figure 4 shows an aliased image produced by a known technique; Figure 5 shows image pixels stored for use in the technique producing Figure 4;
Figure 6 illustrates edge pixel contributions over two horizontal scan lines;
Figure 7 shows the image of Figure 4 to which a first anti-aliasing technique has been applied;
Figure 8 shows the pixels stored for use in the method producing the image of Figure 7; and
Figure 9 shows the image of Figure 4 to which a second anti- aliasing technique has been applied.
Figure 1 shows a general arrangement for a multi-plane video image system coupled to the host bus 10 of a CD-i (Compact Disc interactive) player. From the host bus 10 the base case video, which may be used to form a background to the final image, is fed to a video mixer 12, together with the video synchronising signals. Also providing inputs to the video mixer are a graphics processor 14 and a digital video decoder 16. The graphics processor and digital video decoder take their inputs from the host bus 10 which also provides control signals to the mixer 12. The output of the mixer stage comprises digital signals for the final image colour components (R,G,B) together with the video synchronisation signals. Following conversion by a digital to analogue converter 18, the analogue video signal is output on connector 20.
A first stage of the video mixer 12 is shown in Figure 2 having inputs for the base case video 22 and for the video image 24 of an object to be overlaid on the background by combining in a mixer 26. Accompanying the video image 24 is a matte 28 which is simple binary representation of the object, suitably using 1 to represent those pixels which are part of the image and 0 to represent all other pixels within the video frame. In conventional video mixing, the matte signal is used by the mixer to select either pixels from the image or pixels from the base case to make up the output picture. This technique is, however, prone to producing aliased images (that is to say the well known "staircase" effect on sloping edges) due to the whole pixel nature of the transition between the image and the background.
In order to reduce the aliasing effects, an edge classification stage 30 uses the matte signal to derive a model of the edge of the object (by noting the transition from 0 to 1 in the matte) and identifies the steps in the aliased edge to a calculation circuit 32. The calculation circuit applies a function to the values of pixels forming this step, as will be described below, and passes the result of this calculation to a further calculation stage 34 which derives a transparency value to be applied to these edge pixels. These transparency values are supplied to the mixer 26 which outputs a total picture 38 in which those pixels which are purely object or purely background have their original input values, and the edge pixels have a value which contains a contribution from both in amounts determined by the calculation carried out at stages 32 and 34. Where different functions may be used to obtain the edge pixel values, a store of different function models 36 may be provided.
In order to build up a picture having several overlaid images, mixer stages as in Figure 2 are cascaded in the manner shown in Figure 3 (note that the edge classification and calculation stages of Figure 2 are omitted for reasons of clarity). In the first stage, the background image 22 is combined through use of the appropriate matte 28 with the input image 24 having the greatest depth (Z) value to generate a composite picture. This composite picture is then combined in a further mixer 26a with an object image 28a having a depth value less than that for the first image object and by use of a further matte 28a. Subsequent images 24b with associated matte 28b are added in like manner such that the final picture is built up from the background and then through the image objects in descending order of depth.
For the purposes of comparison and illustration a known anti¬ aliasing technique will now be described, as used to calculate the transparency value by averaging (low-pass filtering) the values of neighbouring pixels within a small window (for example 3 x 3 or 5 x 5 pixels) for each of the edge pixels. The anti-aliased effect may be improved if an original higher resolution image is used: for example, an original image consisting of 720 x 576 pixels may be reduced by averaging four neighbouring pixels and then thresholding the resulting values to provide a sub-sampled version of the original image with a 360 x 288 pixel resolution. An aliased picture for such a reduced resolution image is shown in Figure 4. This method produces very good quality pictures and eliminates successfully the aliasing effect even for few quantisation levels (i.e four to eight levels or two to three bits). However, there is a major drawback in that the transparency value for each edge pixel has to be coded and stored for each frame. It is clear that the method requires a large overhead even for relatively simple situations: by way of example, Figure 5 shows the pixels for which a transparency value has to be stored for the picture of Figure 4. Whilst it would, of course, be possible to employ a compression coding algorithm to reduce the amount of data stored, there would be a corresponding increase in the complexity required at the decoder.
A first embodiment of the proposed method for anti-aliasing estimates the slope (i.e. the step height of the aliased edge) from two or more successive scan lines and then calculates for each pixel the area defined by this slope and the horizontal or vertical orientation. The area under the slope, which is used to dictate the relative contributions of the foreground and background pixels, is determined by the calculation unit 32 of Figure 2. For some implementations, it may be found that a slope characteristic other than a straight line gives improved results, for example an exponential or parabolic curve. Models of such different characteristics would be stored in the function model stage 36 of Figure 2, and the calculation stage 32 would comprise an integrator that determines the area under the particular characteristic chosen. Figure 6 illustrates an implementation of this first embodiment using a straight line characteristic over two successive scan lines which differ by four pixels (that is to say a step four pixels wide appears at the edge of the aliased image). The dark shaded area of Figure 6 represents the coverage of each pixel by an object with a near horizontal top edge and thus can be used to calculate its value. Simple geometry shows that, given the difference N between the number of pixels across the two lines, the value of each pixel can be calculated by equation 1 below:
value = 25δ Equation 1
Figure imgf000010_0001
where k = 1 , 3, 5 (2N- 1 ).
A second embodiment of the method trades a degree of accuracy for the calculation against simplicity of hardware implementation by using estimation of the required step for a pixel displacement. This is simply obtained from the ratio of pixels in a line 2N, leading to equation 2 below:
value = K.step = Equation 2
Figure imgf000010_0002
where K = 0, 1 , 2, 3 .... N
Experimental results have shown that the method of both equation 1 and equation 2 work reasonably well with an expected improved performance for the equation 1 method due to its accuracy. However, considering the implementation issues, the second method is preferred for ease of hardware implementation. The main advantage of both methods is that they provide the ability to calculate the coverage of each edge pixel in real time, eliminating the need to store transparency values.
Whilst the slope estimation method described above is a general technique in the sense that it can be used for both near-horizontal and near-vertical edges, hardware and economic constraints may dictate that only a few scan lines of picture are available to the mixer at any given time. In such a case, the slope estimation can be realistically used only for near-horizontal edges. Figure 7, for example, shows the implementation of the slope estimation method on all the edges of the picture of Figure 4 but within the constraint of a maximum of two adjacent scan lines being available at any time. Here a two-bit (four level) transparency value has been used for each anti-aliased pixel. It will be noted that although the near-horizontal lines are anti-aliased satisfactorily, this is not the case for the near vertical lines which remain aliased. To overcome this problem, and to secure efficient anti¬ aliasing for pixels belonging to near vertical edges, it is preferred to store their transparency values on the disc. Although this seems to be expensive concerning the storage requirements, there are two reasons that favour this choice. Firstly, the number of pixels which will have their transparency values stored on the disc is relatively small: as the example of Figure 8 shows, the pixels of this type for the test picture of Figure 4 are greatly reduced when compared to Figure 5. It is evident that, due to their vertical orientation, a simple coding scheme can be used to store them efficiently. Moreover, it has been found that the required resolution for each of these stored transparency values need not be high with only one or two bits being sufficient for most cases.
Turning now to the calculation of the transparency values to be stored for near-vertical edges, there are several possible solutions to this. The preferred technique calculates the transparency value based on the contribution of pixels from a higher resolution (e.g 720 x 576 pixels) video frame to obtain the 360 x 288 standard format picture, with the transparency value for an edge pixel in the lower resolution picture being the average of the four pixels in the higher one.
Depending on the storage constraints, a quantisation of this value may be required. For more visually pleasing results, a low-pass filter may also be applied to the higher resolution picture prior to calculation of the transparency values. Figure 9 shows the final anti-aliased picture obtained from the application of the slope estimation method for the horizontal edges and the use of the stored transparency values in the vertical direction. Each stored value is represented with only one bit (two transparency levels). For this rather extreme case of several anti-aliased vertical edges, it has been found that almost 600 bits of transparency information must be stored in addition to the depth map. As mentioned above, it would be possible to reduce this figure by adopting a coding scheme for the vertical transparency data but with a corresponding increase in complexity in decoder architecture. From reading the present disclosure, other variations will be apparent to persons skilled in the art. Such variations may involve other features which are already known in the methods and apparatuses for editing of audio and/or video signals and component parts thereof and which may be used instead of or in addition to features already described herein. Although claims have been formulated in this application to particular combinations of features, it should be understood that the scope of the disclosure of the present application also includes any novel feature or any novel combination of features disclosed herein either implicitly or explicitly or any generalisation thereof, whether or not it relates to the same invention as presently claimed in any claim and whether or not it mitigates any or all of the same technical problems as does the present invention. The applicants hereby give notice that new claims may be formulated to such features and/or combinations of such features during the prosecution of the present application or of any further application derived therefrom.

Claims

1 . A method for display of images in the form of an image object defined in a first pixel plane overlying a background defined in a further pixel plane, wherein those pixels falling wholly within the object are displayed with the pixel value defined in the first pixel plane, those pixels falling wholly outside the object are displayed with the pixel value defined in the further pixel plane, and the remaining edge pixels are displayed with a transparency value comprising contributions from the corresponding pixel in both the first and further planes, characterised in that the transparency values are derived by: identifying step discontinuities in the image edge; applying a slope characteristic across each step discontinuity; for each pixel crossed by the slope characteristic, deriving the relative areas of the pixel above and below the slope; and setting the relative contributions of the foreground and background pixels accordingly.
2. A method as claimed in Claim 1 , wherein one or more of the transparency values is pre-generated and stored with the image pixel data to be read from storage when the image is displayed, and the remainder are calculated when the image is displayed.
3. A method as claimed in Claim 2, wherein the transparency values are pre-generated and stored for those step discontinuities having a vertical extent greater than or equal to horizontal extent.
4. A method as claimed in Claim 1 , wherein the step of derivation of relative areas for a pixel comprises integration of the slope characteristic across the pixel to produce a first area value and subtracting this from the total pixel area to give a second area value.
5. A method as claimed in Claim 1 , wherein the step of derivation of the relative areas of the pixel comprises dividing each pixel in a ratio given by the position of that pixel and the number of pixels across the step discontinuity.
6. A method as claimed in Claim 1 , wherein the slope characteristic is a straight line.
7. A method as claimed in Claim 1 , wherein the slope characteristic is a curve.
8. Video mixing apparatus operable to form an output image signal from first and second pixel plane signals and a matte signal identifying, for each pixel of the first plane, whether or not it comprises a part of an image object defined in the first plane to be displayed overlying a background defined in the second pixel plane, the apparatus comprising: an edge classification stage arranged to receive the matte signal and identify therefrom step discontinuities at the edge of the image object; first calculation means arranged to apply a predetermined slope characteristic across step discontinuities identified by the edge classification stage; second calculation means arranged to determine, for each pixel crossed by the slope characteristic, the relative areas of the pixel above and below the slope; and mixer stage coupled to receive the first and second pixel plane signals and an indication of relative areas from the second calculation means, and arranged to output a display pixel image signal in which the values of pixels falling wholly within the image object are taken from the first pixel plane signal, the values of pixels falling wholly outside the image object are taken from the second pixel plane signal, and the values of those pixels identified as edge pixels by the edge classification stage are taken from both first and second pixel planes in a ratio by the said indication from the second calculation means.
9. Apparatus as claimed in Claim 8, wherein the first calculation means is operable to apply a selected one from a plurality of stored slope characteristics, the apparatus further comprising a function store holding implementation instructions for each of said slope characteristics.
PCT/IB1995/001050 1994-12-08 1995-11-24 Digital image anti-aliasing WO1996018264A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
GB9424808.5 1994-12-08
GBGB9424808.5A GB9424808D0 (en) 1994-12-08 1994-12-08 Digital image anti-aliasing

Publications (1)

Publication Number Publication Date
WO1996018264A1 true WO1996018264A1 (en) 1996-06-13

Family

ID=10765632

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/IB1995/001050 WO1996018264A1 (en) 1994-12-08 1995-11-24 Digital image anti-aliasing

Country Status (3)

Country Link
GB (1) GB9424808D0 (en)
TW (1) TW287342B (en)
WO (1) WO1996018264A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO1999031875A2 (en) * 1997-12-17 1999-06-24 Koninklijke Philips Electronics N.V. Method of chroma-keying for a digital video compression system
WO2000038171A1 (en) * 1998-12-19 2000-06-29 Powertv, Inc. Font anti-aliasing system

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4855831A (en) * 1986-10-31 1989-08-08 Victor Co. Of Japan Video signal processing apparatus
US4887159A (en) * 1987-03-26 1989-12-12 The Grass Valley Group Inc. Shadow visual effects wipe generator
EP0391513A2 (en) * 1989-04-07 1990-10-10 Sony Corporation Controlling the combining of video signals
EP0419126A2 (en) * 1989-09-22 1991-03-27 Ampex Corporation System for generating anti-aliased video signal
DE4017880A1 (en) * 1990-06-02 1991-12-05 Broadcast Television Syst Edge correction in two=dimensional pixel distribution - using edge-parallel correction lines with intermediate pixels corrected and edge shape to determine corrections

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4855831A (en) * 1986-10-31 1989-08-08 Victor Co. Of Japan Video signal processing apparatus
US4887159A (en) * 1987-03-26 1989-12-12 The Grass Valley Group Inc. Shadow visual effects wipe generator
EP0391513A2 (en) * 1989-04-07 1990-10-10 Sony Corporation Controlling the combining of video signals
EP0419126A2 (en) * 1989-09-22 1991-03-27 Ampex Corporation System for generating anti-aliased video signal
DE4017880A1 (en) * 1990-06-02 1991-12-05 Broadcast Television Syst Edge correction in two=dimensional pixel distribution - using edge-parallel correction lines with intermediate pixels corrected and edge shape to determine corrections

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO1999031875A2 (en) * 1997-12-17 1999-06-24 Koninklijke Philips Electronics N.V. Method of chroma-keying for a digital video compression system
WO1999031875A3 (en) * 1997-12-17 1999-08-19 Koninkl Philips Electronics Nv Method of chroma-keying for a digital video compression system
WO2000038171A1 (en) * 1998-12-19 2000-06-29 Powertv, Inc. Font anti-aliasing system

Also Published As

Publication number Publication date
TW287342B (en) 1996-10-01
GB9424808D0 (en) 1995-02-08

Similar Documents

Publication Publication Date Title
Wang et al. Representing moving images with layers
US6665450B1 (en) Interpolation of a sequence of images using motion analysis
US7245774B2 (en) Image processing apparatus
US6269195B1 (en) Apparatus and methods for selectively feathering a composite image
EP0358498B1 (en) Method and apparatus for generating animated images
US5090909A (en) Video graphic simulator systems
US20100002948A1 (en) Image enhancement
JP2001526477A (en) Image conversion and coding technology
US5119442A (en) Real time digital video animation using compressed pixel mappings
JPH09507935A (en) Post-processing method and apparatus for producing focus / defocus effect in computer-generated image of three-dimensional object
US20040062450A1 (en) Image processing apparatus and method and image pickup apparatus
Mitsunaga et al. Autokey: Human assisted key extraction
Adelson Layered representations for image coding
US8325196B2 (en) Up-scaling
JP2002290975A (en) Data processing method
KR100329874B1 (en) Image Processing
JP3033967B2 (en) Method and apparatus for processing image signals
US5960118A (en) Method for 2D and 3D images capturing, representation, processing and compression
US6343151B1 (en) Method and apparatus for transmitting and utilizing analog encoded information
EP0318149B1 (en) Video processing
JP3308675B2 (en) Encoding device
WO1999065224A3 (en) Creating animation from a video
US5633956A (en) Video image processing
WO1996018264A1 (en) Digital image anti-aliasing
KR20010024416A (en) Static image generation method and device

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): JP KR

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): AT BE CH DE DK ES FR GB GR IE IT LU MC NL PT SE

121 Ep: the epo has been informed by wipo that ep was designated in this application
122 Ep: pct application non-entry in european phase