EP0753183B1 - Fast perspective texture mapping for 3-d computer graphics - Google Patents

Fast perspective texture mapping for 3-d computer graphics Download PDF

Info

Publication number
EP0753183B1
EP0753183B1 EP95913283A EP95913283A EP0753183B1 EP 0753183 B1 EP0753183 B1 EP 0753183B1 EP 95913283 A EP95913283 A EP 95913283A EP 95913283 A EP95913283 A EP 95913283A EP 0753183 B1 EP0753183 B1 EP 0753183B1
Authority
EP
European Patent Office
Prior art keywords
line
texture
coordinates
lines
polygon
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Lifetime
Application number
EP95913283A
Other languages
German (de)
French (fr)
Other versions
EP0753183A1 (en
Inventor
Michael Robin Day
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Argonaut Technologies Ltd
Original Assignee
Argonaut Technologies Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Argonaut Technologies Ltd filed Critical Argonaut Technologies Ltd
Publication of EP0753183A1 publication Critical patent/EP0753183A1/en
Application granted granted Critical
Publication of EP0753183B1 publication Critical patent/EP0753183B1/en
Anticipated expiration legal-status Critical
Expired - Lifetime legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T15/003D [Three Dimensional] image rendering
    • G06T15/04Texture mapping

Definitions

  • This invention relates to planar texture mapping for 3-D computer graphics.
  • the techniques currently used can be divided into two categories: (a) those which explicitly calculate the texture coordinates for every pixel of the polygon to be rendered - this involves certain calculations to be performed at each pixel, in particular at least one divide operation; and (b) those which evaluate the texture coordinates only at the edges of each horizontal span of the polygon, and then use a linear interpolation scheme to approximate their values at the intermediate pixels (or some other form of approximation).
  • the reduced running times of techniques in category (b) are offset by the loss of accuracy incurred in the texture projection. If polygons rendered using these techniques are viewed close up, distortion of the texture will be seen, to a degree which may be unacceptable.
  • the proposed solution is to scan convert the polygon not along horizontal scanlines but along carefully chosen angled lines, representing the lines of constant z-coordinate.
  • the invention proposes a method of rendering an image on a display, comprising texture mapping of a pre-prepared texture map to flat surfaces of an object which is to be viewed in perspective on the display, characterised by scanning each surface to be texture mapped along notional lines of constant z (depth) coordinate (z-lines); and for each line of constant z, determining the object end coordinates of the lines, determining the texture values for the end coordinates, determining the texture values for the coordinates between end coordinates by interpolation, and mapping the texture values of the pre-prepared texture map to each line of constant z and thereby without modification for perspective.
  • Figure T1 shows the outline of a planar polygon which is to be rendered with texture from a texture map whose uv-coordinate space is indicated by the gridlines.
  • Such a set of z-lines is shown for the example polygon, in Figure T2.
  • the z-line can therefore be rendered using two DDAs, one to track the line in uv-space, and one to obtain the screen pixels it maps to.
  • the only texture coordinates explicitly calculated for the z-line are those at its endpoints.
  • the method for rendering the whole polygon is therefore to drop the notion of horizontal scanlines, and instead to draw parallel to the z-lines.
  • the gradient of the z-lines in view coordinates can be calculated from a knowledge of the polygon's plane equation and the geometry of the projection.
  • FIG. T3 A scheme that permits this is illustrated in Figure T3.
  • the pixels making up a representative z-line can be obtained by a standard Bresenham line draw.
  • the remaining z-lines are generated by vertically translating the representative z-line by the appropriate number of pixels if the line is at less than 45 degrees to the horizontal, and horizontally translating it if the line is at 45 degrees or more.
  • the texture coordinates must be evaluated for the start and end of each z-line drawn to. This can either be done explicitly every time new coordinates are required, or by making use of the fact that the numerator and denominator of each of u and v are linear in both screen coordinates, and can therefore be maintained incrementally as scan conversion proceeds.
  • each span consists of a contiguous block of pixels. This is not always the case for convex polygons rendered using a z-line scan.
  • Figure T4 shows the situation where a polygon edge and the z-lines have almost equal gradients - the set of pixels on a particular z-line which lie in the interior of the polygon forms a broken line. This is a result of interference between two integer approximation processes - one which generates the polygon edge, and one which generates the z-line. Two approaches may be taken:

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Graphics (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Image Generation (AREA)

Abstract

A method of processing computer graphics information for rendering an image on a display, comprising texture mapping of a pre-prepared texture map to flat surfaces of an object which is to be viewed in perspective on the display, wherein each surface to be texture mapped is scanned along notional lines of constant z (depth) coordinate (z-lines) to determine the end coordinates of those lines, to which the texture values of the pre-prepared texture map can be mapped without modification for perspective.

Description

This invention relates to planar texture mapping for 3-D computer graphics.
This is the task of applying a prestored or generated image to the flat surface of a 3-D object when the object is viewed in perspective on a display. The projected images must be drawn in perspective. It is difficult to do this in real time, i.e. so that, animations appear to happen smoothly, particularly for computer games which often do not have sufficient computing power. Details of the subject can be found in "3-D Computer Animation" by Vince, Addison-Wesley (1992).
The techniques currently used can be divided into two categories: (a) those which explicitly calculate the texture coordinates for every pixel of the polygon to be rendered - this involves certain calculations to be performed at each pixel, in particular at least one divide operation; and (b) those which evaluate the texture coordinates only at the edges of each horizontal span of the polygon, and then use a linear interpolation scheme to approximate their values at the intermediate pixels (or some other form of approximation).
Techniques that fall into category (a) above give rise to much longer execution times than those in category (b), for all but the smallest of polygons to be rendered. This makes them unsuitable for real-time rendering systems unless dedicated rendering hardware is used.
On the other hand, the reduced running times of techniques in category (b) are offset by the loss of accuracy incurred in the texture projection. If polygons rendered using these techniques are viewed close up, distortion of the texture will be seen, to a degree which may be unacceptable.
The proposed solution is to scan convert the polygon not along horizontal scanlines but along carefully chosen angled lines, representing the lines of constant z-coordinate.
Accordingly, the invention proposes a method of rendering an image on a display, comprising texture mapping of a pre-prepared texture map to flat surfaces of an object which is to be viewed in perspective on the display, characterised by scanning each surface to be texture mapped along notional lines of constant z (depth) coordinate (z-lines); and for each line of constant z, determining the object end coordinates of the lines, determining the texture values for the end coordinates, determining the texture values for the coordinates between end coordinates by interpolation, and mapping the texture values of the pre-prepared texture map to each line of constant z and thereby without modification for perspective.
To fully understand the invention, an example of the inventive technique will now be described with reference to the drawing.
Figure T1 shows the outline of a planar polygon which is to be rendered with texture from a texture map whose uv-coordinate space is indicated by the gridlines. Consider a plane of constant z, with respect to the view space coordinates. Unless the plane of the polygon itself lies parallel to this plane, the two planes will intersect in a line of constant z. A set of such planes gives rise to a set of parallel lines, referred to here as z-lines.
Such a set of z-lines is shown for the example polygon, in Figure T2. For texture which is mapped onto the polygon along a z-line, there is no perspective foreshortening as all the points have equal z-coordinate. The z-line can therefore be rendered using two DDAs, one to track the line in uv-space, and one to obtain the screen pixels it maps to. The only texture coordinates explicitly calculated for the z-line are those at its endpoints.
The method for rendering the whole polygon is therefore to drop the notion of horizontal scanlines, and instead to draw parallel to the z-lines. For a particular polygon, the gradient of the z-lines in view coordinates can be calculated from a knowledge of the polygon's plane equation and the geometry of the projection.
We must ensure that when drawing these angled lines, care is taken to write to each pixel of the polygon once and only once. A scheme that permits this is illustrated in Figure T3. The pixels making up a representative z-line can be obtained by a standard Bresenham line draw. The remaining z-lines are generated by vertically translating the representative z-line by the appropriate number of pixels if the line is at less than 45 degrees to the horizontal, and horizontally translating it if the line is at 45 degrees or more.
Finally, the texture coordinates must be evaluated for the start and end of each z-line drawn to. This can either be done explicitly every time new coordinates are required, or by making use of the fact that the numerator and denominator of each of u and v are linear in both screen coordinates, and can therefore be maintained incrementally as scan conversion proceeds.
A problem is encountered when trying to render polygons in this way. For convex polygons rendered using a simple horizontal scan, each span consists of a contiguous block of pixels. This is not always the case for convex polygons rendered using a z-line scan. Figure T4 shows the situation where a polygon edge and the z-lines have almost equal gradients - the set of pixels on a particular z-line which lie in the interior of the polygon forms a broken line. This is a result of interference between two integer approximation processes - one which generates the polygon edge, and one which generates the z-line. Two approaches may be taken:
  • (i) An approximation polygon outline can be obtained from the following observation: If all pixel columns (rows) in the z-line pattern are shifted vertically (horizontally) so that the representative z-line becomes horizontal (vertical), then the original polygon outline is mapped to a new one which is simply a vertically (horizontally) sheared version of the original, albeit with artifacts introduced by the finite resolution. Scan conversion then proceeds by notionally rendering this new sheared polygon with ordinary horizontal (vertical) scanning, and performing the inverse shift on the pixel columns (rows) at the output stage. The result is only an approximation to the desired polygon because the inverse shift introduces the same type of artifact.
  • (ii) The exact polygon outline can be obtained as follows: First a bit-per-pixel mask of the actual polygon is generated using a standard horizontal scan. The mask is then sheared vertically (horizontally) by performing column (row) shifts on the mask. The sheared mask is then used to control the output of texture along z-lines, by reading the mask in rows (columns). This produces the desired polygon at the expense of shearing and reading the bit-mask. In practice, however, it is not necessary to store the mask for the entire polygon. Assuming the z-lines lie at less than 45 degrees to the horizontal (the alternative case has a similar treatment) the scan conversion process, instead of generating the mask, is used to fill a set of buckets. Each bucket corresponds to a z-line, and contains the screen x-coordinates of all pixel columns for which the column goes from inside the polygon to out or vice versa, on that z-line. Thus a pair of entries is made for the top and bottom of the section of a pixel column that is covered by the polygon (there may be more than one pair for a concave outline). Having generated the array of buckets, the z-lines are rendered in top-to-bottom order making use of a mask which has one bit for each pixel in a z-line. It is initially zeroed, and then updated incrementally before each z-line is rendered. For each bucket entry at a given z-line, the corresponding bit in the mask is toggled, to indicate that the pixel column has just crossed the polygon boundary on the current z-line.If the exact polygon outline is generated using the bit-mask technique outlined in (ii) above, then the rendering part of the software is presumably capable of reading from a completely general mask, and there is no reason why we must restrict ourselves to (convex) polygons. Thus any planar shape can be texture mapped, such as concave polygons, disks, planar regions bounded by 2D splines, etc. provided a bit-mask can be generated of the projected interior.The solution given here combines the advantages of both the categories of technique listed in the introduction, in that it produces an accurate perspective projection of the texture, while at the same time using a linear interpolation scheme to rapidly calculate texture coordinates for interior pixels of the polygon, calculating them explicitly only on the polygon boundary.The method of scan conversion is more complicated than a standard horizontal scan. On the whole therefore, more work is required per boundary pixel. The technique will be faster than category (a) mentioned at the start only for polygons with an interior to boundary ratio higher than a certain value.
  • Claims (4)

    1. A method of rendering an image on a display, comprising texture mapping of a pre-prepared texture map to flat surfaces of an object which is to be viewed in perspective on the display, characterised by scanning each surface to be texture mapped along notional lines of constant z (depth) coordinate (z-lines); and for each line of constant z, determining the object end coordinates of the lines, determining the texture values for the end coordinates, determining the texture values for the coordinates between end coordinates by interpolation, and mapping the texture values of the pre-prepared texture map to each line of constant z and thereby without modification for perspective.
    2. A method as claimed in claim 1 wherein the pixels of a representative z-line are obtained by a Bresenham line-draw.
    3. A method as claimed in claim 1 or 2 wherein after obtaining pixel-coordinates of a representative z-line, the remaining z-lines are obtained by vertically translating the representative z-line by the appropriate number of pixels if the line is at less than 45 degrees to the horizontal, and horizontally translating it if the line is at 45 degrees or more.
    4. A method as in any preceding claim wherein for objects in which the numerator and denominator of the u and v of the (u,v) texture map are linear in both screen coordinates, the end (u,v) coordinates of the texture map line to be mapped to the object surface are explicitly calculated only once and are thereafter maintained by incremental change for succeeding scans.
    EP95913283A 1994-03-31 1995-03-31 Fast perspective texture mapping for 3-d computer graphics Expired - Lifetime EP0753183B1 (en)

    Applications Claiming Priority (3)

    Application Number Priority Date Filing Date Title
    GB9406515 1994-03-31
    GB9406515A GB9406515D0 (en) 1994-03-31 1994-03-31 Texture mapping for 3-d computer graphics
    PCT/GB1995/000751 WO1995027266A1 (en) 1994-03-31 1995-03-31 Fast perspective texture mapping for 3-d computer graphics

    Publications (2)

    Publication Number Publication Date
    EP0753183A1 EP0753183A1 (en) 1997-01-15
    EP0753183B1 true EP0753183B1 (en) 1998-11-18

    Family

    ID=10752901

    Family Applications (1)

    Application Number Title Priority Date Filing Date
    EP95913283A Expired - Lifetime EP0753183B1 (en) 1994-03-31 1995-03-31 Fast perspective texture mapping for 3-d computer graphics

    Country Status (7)

    Country Link
    US (1) US5838329A (en)
    EP (1) EP0753183B1 (en)
    JP (1) JPH09511599A (en)
    CA (1) CA2185926A1 (en)
    DE (1) DE69506092T2 (en)
    GB (1) GB9406515D0 (en)
    WO (1) WO1995027266A1 (en)

    Families Citing this family (11)

    * Cited by examiner, † Cited by third party
    Publication number Priority date Publication date Assignee Title
    JPH11509661A (en) * 1996-05-06 1999-08-24 フィリップス エレクトロニクス ネムローゼ フェンノートシャップ Method and apparatus for graphics mapping a surface to a two-dimensional image
    US6577320B1 (en) 1999-03-22 2003-06-10 Nvidia Corporation Method and apparatus for processing multiple types of pixel component representations including processes of premultiplication, postmultiplication, and colorkeying/chromakeying
    US20030063383A1 (en) * 2000-02-03 2003-04-03 Costales Bryan L. Software out-of-focus 3D method, system, and apparatus
    US6410643B1 (en) 2000-03-09 2002-06-25 Surmodics, Inc. Solid phase synthesis method and reagent
    US6828969B2 (en) * 2000-06-05 2004-12-07 Namco Ltd. Game system, program and image generating method
    WO2002013141A1 (en) * 2000-08-09 2002-02-14 Dynamic Digital Depth Research Pty Ltd Image conversion and encoding techniques
    US6853373B2 (en) * 2001-04-25 2005-02-08 Raindrop Geomagic, Inc. Methods, apparatus and computer program products for modeling three-dimensional colored objects
    US7496528B2 (en) * 2001-12-27 2009-02-24 Proto Labs, Inc. Automated quoting of molds and molded parts
    US7142211B2 (en) 2003-07-07 2006-11-28 Arcsoft, Inc. Graphic engine for fill style transferring in a resource-constrained device
    US7764287B2 (en) * 2003-07-07 2010-07-27 Arcsoft, Inc. Graphic engine for approximating a quadratic bezier curve in a resource-constrained device
    US7161597B2 (en) * 2003-07-07 2007-01-09 Arcsoft, Inc. Graphic engine for rasterizing a straight edge in a resource-constrained device

    Family Cites Families (5)

    * Cited by examiner, † Cited by third party
    Publication number Priority date Publication date Assignee Title
    US4615013A (en) * 1983-08-02 1986-09-30 The Singer Company Method and apparatus for texture generation
    US5343558A (en) * 1991-02-19 1994-08-30 Silicon Graphics, Inc. Method for scan converting shaded triangular polygons
    US5345541A (en) * 1991-12-20 1994-09-06 Apple Computer, Inc. Method and apparatus for approximating a value between two endpoint values in a three-dimensional image rendering device
    US5469535A (en) * 1992-05-04 1995-11-21 Midway Manufacturing Company Three-dimensional, texture mapping display system
    US5649082A (en) * 1995-03-20 1997-07-15 Silicon Graphics, Inc. Efficient method and apparatus for determining texture coordinates for lines and polygons

    Also Published As

    Publication number Publication date
    CA2185926A1 (en) 1995-10-12
    JPH09511599A (en) 1997-11-18
    DE69506092T2 (en) 1999-07-15
    US5838329A (en) 1998-11-17
    DE69506092D1 (en) 1998-12-24
    EP0753183A1 (en) 1997-01-15
    WO1995027266A1 (en) 1995-10-12
    GB9406515D0 (en) 1994-05-25

    Similar Documents

    Publication Publication Date Title
    US7239319B2 (en) Rendering outline fonts
    US5428718A (en) Tessellation system
    JP5188628B2 (en) Method and system for rendering a 3D model of a 3D object
    Von Herzen et al. Accurate triangulations of deformed, intersecting surfaces
    EP0638875B1 (en) A 3-dimensional animation generating apparatus and a method for generating a 3-dimensional animation
    US7034823B2 (en) 3D computer graphics processing apparatus and method
    US6239808B1 (en) Method and apparatus for determining texture values of graphical images
    US7414636B2 (en) Rendering apparatus, rendering processing method and computer program product
    Sramek et al. Fast ray-tracing of rectilinear volume data using distance transforms
    US6204857B1 (en) Method and apparatus for effective level of detail selection
    US20020094125A1 (en) Method for fast rendering of photorealistic computer graphics images
    JPH0719297B2 (en) Graphic display processing system and method
    EP0753183B1 (en) Fast perspective texture mapping for 3-d computer graphics
    US6433790B1 (en) Methods and systems for rendering line and point features for display
    Wan et al. Boundary cell-based acceleration for volume ray casting
    US6400370B1 (en) Stochastic sampling with constant density in object space for anisotropic texture mapping
    KR100453530B1 (en) Apparatus and method for drawing three dimensional graphics by converting two dimensional polygon data to three dimensional polygon data
    US20020181663A1 (en) Memory efficient shear-warp voxel projection algorithm
    US5886703A (en) Perspective correct texture mapping system and methods with intelligent subdivision
    Teitzel et al. Line integral convolution on triangulated surfaces
    US6556203B1 (en) Tile-based digital differential analyzer rasterization
    Lee et al. An efficient ray tracing method for terrain rendering
    US5821942A (en) Ray tracing through an ordered array
    US20050116951A1 (en) Using runs of cells to traverse a ray through a volume
    US20040189641A1 (en) Method and apparatus for determining intersections of a particular line with cells in a lattice

    Legal Events

    Date Code Title Description
    PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

    Free format text: ORIGINAL CODE: 0009012

    17P Request for examination filed

    Effective date: 19960923

    AK Designated contracting states

    Kind code of ref document: A1

    Designated state(s): BE DE ES FR GB IT NL SE

    GRAG Despatch of communication of intention to grant

    Free format text: ORIGINAL CODE: EPIDOS AGRA

    17Q First examination report despatched

    Effective date: 19971107

    GRAG Despatch of communication of intention to grant

    Free format text: ORIGINAL CODE: EPIDOS AGRA

    GRAH Despatch of communication of intention to grant a patent

    Free format text: ORIGINAL CODE: EPIDOS IGRA

    GRAH Despatch of communication of intention to grant a patent

    Free format text: ORIGINAL CODE: EPIDOS IGRA

    GRAA (expected) grant

    Free format text: ORIGINAL CODE: 0009210

    AK Designated contracting states

    Kind code of ref document: B1

    Designated state(s): BE DE ES FR GB IT NL SE

    PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

    Ref country code: SE

    Free format text: THE PATENT HAS BEEN ANNULLED BY A DECISION OF A NATIONAL AUTHORITY

    Effective date: 19981118

    Ref country code: NL

    Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

    Effective date: 19981118

    Ref country code: IT

    Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT;WARNING: LAPSES OF ITALIAN PATENTS WITH EFFECTIVE DATE BEFORE 2007 MAY HAVE OCCURRED AT ANY TIME BEFORE 2007. THE CORRECT EFFECTIVE DATE MAY BE DIFFERENT FROM THE ONE RECORDED.

    Effective date: 19981118

    Ref country code: ES

    Free format text: THE PATENT HAS BEEN ANNULLED BY A DECISION OF A NATIONAL AUTHORITY

    Effective date: 19981118

    REF Corresponds to:

    Ref document number: 69506092

    Country of ref document: DE

    Date of ref document: 19981224

    ET Fr: translation filed
    NLV1 Nl: lapsed or annulled due to failure to fulfill the requirements of art. 29p and 29m of the patents act
    PLBE No opposition filed within time limit

    Free format text: ORIGINAL CODE: 0009261

    STAA Information on the status of an ep patent application or granted ep patent

    Free format text: STATUS: NO OPPOSITION FILED WITHIN TIME LIMIT

    26N No opposition filed
    PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

    Ref country code: BE

    Payment date: 20000317

    Year of fee payment: 6

    PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

    Ref country code: GB

    Payment date: 20000323

    Year of fee payment: 6

    PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

    Ref country code: DE

    Payment date: 20000327

    Year of fee payment: 6

    PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

    Ref country code: FR

    Payment date: 20000329

    Year of fee payment: 6

    PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

    Ref country code: GB

    Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

    Effective date: 20010331

    Ref country code: BE

    Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

    Effective date: 20010331

    BERE Be: lapsed

    Owner name: ARGONAUT TECHNOLOGIES LTD

    Effective date: 20010331

    GBPC Gb: european patent ceased through non-payment of renewal fee

    Effective date: 20010331

    PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

    Ref country code: FR

    Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

    Effective date: 20011130

    REG Reference to a national code

    Ref country code: FR

    Ref legal event code: ST

    PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

    Ref country code: DE

    Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

    Effective date: 20020101