EP1442390A4 - System and method for environment mapping - Google Patents
System and method for environment mappingInfo
- Publication number
- EP1442390A4 EP1442390A4 EP02770564A EP02770564A EP1442390A4 EP 1442390 A4 EP1442390 A4 EP 1442390A4 EP 02770564 A EP02770564 A EP 02770564A EP 02770564 A EP02770564 A EP 02770564A EP 1442390 A4 EP1442390 A4 EP 1442390A4
- Authority
- EP
- European Patent Office
- Prior art keywords
- vector
- point
- reflection
- normal vector
- modified
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 45
- 238000013507 mapping Methods 0.000 title claims abstract description 19
- 239000013598 vector Substances 0.000 claims abstract description 122
- 238000009877 rendering Methods 0.000 claims abstract description 18
- 230000001419 dependent effect Effects 0.000 claims description 2
- 230000008569 process Effects 0.000 claims description 2
- 230000001131 transforming effect Effects 0.000 claims 2
- 238000005286 illumination Methods 0.000 description 20
- 230000003287 optical effect Effects 0.000 description 4
- 230000008859 change Effects 0.000 description 3
- 238000010586 diagram Methods 0.000 description 2
- 230000002452 interceptive effect Effects 0.000 description 2
- 230000008901 benefit Effects 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 230000005236 sound signal Effects 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
- G06T15/50—Lighting effects
- G06T15/506—Illumination models
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
- G06T15/04—Texture mapping
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2215/00—Indexing scheme for image rendering
- G06T2215/12—Shadow map, environment map
Definitions
- This invention relates generally to computer generated images and more particularly to a system and method for environment mapping.
- the illumination of a computer- generated object by discrete light sources, continuous light sources, and ambient light is described by an illumination model.
- the object is illuminated by the reflection of ambient light and the reflection of light source light from the surface of the object.
- the illumination model is a mathematical expression that operates on a set of variables to generate reflection properties, such as color and intensity of reflected light and an object's texture as viewed by an observer. Given ambient light and light sources positioned about the object, the illumination model defines the reflection properties of the object. The illumination model is considered to be accurate if the illuminated object appears realistic to an observer.
- the illumination model is incorporated in a software program executed by a vector processing unit, a central processing unit, or a rendering engine of a computer system.
- the program must be capable of computing the illumination of the object when the light sources change position with respect to the object, when the observer views the illuminated object from a different angle, or when the object is rotated.
- an efficient illumination model is needed for the processing unit to compute the illumination in real- time, for example, if the observer (i.e., a camera) is moving with respect to the object. Therefore, it is desired to incorporate terms in the illumination model that are computationally cost effective, while at the same time generating an image of the illuminated object that is aesthetically pleasing to the observer.
- FIG. 1 illustrates a prior art direct normal projection method for computing an object's texture coordinates.
- FIG. 1 includes an object's surface 105, a point P on surface 105, a normal vector n to surface 105 at point P, an observer 110a, a line-of sight 115a between observer 110a and the point P, and a projection of the normal vector n onto an x-axis 120, referred to as n x .
- a z-axis (not shown) is perpendicular to x-axis 120 and is in the plane of FIG. 1
- a y-axis (not shown) is perpendicular to x-axis 120 and the z-axis and is out of the plane of FIG. 1.
- the FIG. 1 embodiment of object's surface 105 is a line, however, surface 105 is typically any 2-D surface, and hence in general, the normal vector n may have a vector component n y along the y-axis.
- the direct normal projection method computes the projected components n x and n y of the normal vector n for each point P on object's surface 105.
- the central processing unit or vector processing unit maps (i.e., transforms) the projected components n x and n y into texture coordinates (s,t) using one or more mapping algorithms known in the' art.
- the vector processing unit uses the computed texture coordinates (s,t) for each point P, as well as other reflection variables, in an illumination model to generate a reflection pattern of object's surface 10.5.
- the direct normal projection method of the prior art may be fast, the method generates a reflection pattern that appears "painted-on" as observer 110a moves to different locations.
- the reflection pattern of object's surface 105 does not change with respect to rotation or translation of observer 110a, since the method depends upon the x and y components of the normal vector n, independent of the position of observer 110a with respect to the point P.
- the vector processing unit computes the same projected components (n x ,n y ) and texture coordinates (s,t) for an observer 110b viewing point P as observer 110a viewing point P.
- the method includes constructing a surface normal vector n at a point P on a surface of the reflective object, constructing an observation vector e from the point P to an observer, and using a modified reflection formula to compute a reflection vector r based on the surface normal vector n and the observation vector e.
- the modified reflection formula is based on reflection about a pseudo-normal vector n' at the point P on the surface.
- the pseudo-normal vector n' bisects an angle subtended by the surface normal vector n and a reference observation vector e 0 , where the reference observation vector e 0 is directed from the point P to an observer located directly in front of the point P.
- Each computed reflection vector r may be processed to generate a pair of texture coordinates (s,t).
- the reflective object is then rendered based in part on the texture coordinates (s,t) associated with each point P on the surface of the reflective object.
- the scope of the present invention covers all types of rendering schemes, such as a polygon rendering where each point P on the surface of the reflective object is located at the vertex of a polygon.
- the system includes a memory configured to store a modified reflection model, a vector processing unit configured to compute reflection vectors using the modified reflection model, and a graphics processor configured to render the reflective object in an image.
- the quality of the image is dependent upon the texture coordinates that are derived from the computed reflection vectors.
- FIG. 2 is a block diagram of one embodiment of an electronic entertainment system according to the invention.
- FIG. 3 illustrates a modified reflection projection method stored in main memory 210 of FIG. 2, according to one embodiment of the invention
- FIG. 4 is a flowchart of method steps for displaying an image of a reflective object based upon texture coordinates, according to one embodiment of the invention.
- FIG. 5 is a flowchart of method steps for step 415 of FIG. 4 to generate reflection vectors for a reflective object, according to one embodiment of the invention.
- the system and method for environment mapping described herein allow a computer-generated object's reflective appearance to change, based upon position and orientation of a camera with respect to the object's location.
- a position of the camera may be defined by a lateral location of the camera with respect to the object's location. Lateral camera movement is defined as motion to the right, left, up, or down with respect to the object's location.
- Camera orientation may be defined by rotation angles with respect to a given, fixed coordinate system.
- An exemplary embodiment of the invention is implemented as a real-time environment mapping for polygon rendering. However, the scope of the invention covers other applications, such as environment mapping for other rendering schemes. Other rendering schemes may include, but are not limited to, point-based and non-polygon volume-based primitives.
- Various embodiments of the invention may be enabled in software, hardware, or firmware.
- a central processing unit (CPU) and/ or one or more vector processing units (VPUs) use illumination models to compute reflective properties of an object.
- the object's reflective properties are associated with the objects' appearance. Reflective properties include color and intensity of light reflected by the object, and texture of the reflective object.
- the texture of an object is associated with reflective properties such as the object's shininess and overall surface appearance.
- the object's texture is specified by texture coordinates (s,t) computed by the VPU. Texture coordinates may be incorporated into a texture map which is wrapped (i.e., mapped) around the object.
- a VPU may execute environment mapping instructions that operate on variables stored in a VPU random access memory (RAM) or on variables stored in a CPU register to compute the texture coordinates.
- texture coordinates and the other computed reflective properties also referred to as illumination terms
- the GPU prepares the reflective object for display on a display device such as a computer monitor.
- FIG. 2 is a block diagram of one embodiment of an electronic entertainment system 200 according to the invention.
- System 200 includes, but is not limited to, a main memory 210, a CPU 212, a VPU 213, a GPU 214, an input/ output processor (IOP) 216, an IOP memory 218, a controller interface 220, a memory card 222, a Universal Serial Bus (USB) interface 224, and an IEEE 1394 interface 226.
- System 200 also includes an operating system read-only memory (OS ROM) 228, a sound processing unit (SPU) 232, an optical disc control unit 234, and a hard disc drive (HDD) 236, which are connected via a bus 246 to IOP 216.
- OS ROM operating system read-only memory
- SPU sound processing unit
- HDD hard disc drive
- CPU 212, VPU 213, GPU 214, and IOP 216 communicate via a system bus 244.
- CPU 212 communicates with main memory 210 via a dedicated bus 242.
- VPU 213 and GPU 214 may also communicate via a dedicated bus 240.
- CPU 212 executes programs stored in OS ROM 228 and main memory
- Main memory 210 may contain pre-stored programs and may also contain programs transferred via IOP 216 from a CD-ROM or DVD-ROM (not shown) using optical disc control unit 234.
- IOP 216 controls data exchanges between CPU 212, VPU 213, GPU 214 and other devices of system 200, such as controller interface 220.
- Main memory 210 includes, but is not limited to, a program having game instructions including an illumination -model.
- the program is preferably loaded from a DVD-ROM via optical disc control unit 234 into main memory 210.
- CPU 212 in conjunction with VPU 213, GPU 214, and SPU 232, executes game instructions and generates rendering instructions using inputs received from a user via controller interface 220. The user may also instruct CPU 212 to store certain game information on memory card 222. Other devices may be connected to system 200 via USB interface 224 and IEEE 1394 interface 226.
- VPU 213 executes instructions from CPU 212 to generate texture coordinates associated with an illuminated object by using the illumination model.
- FIG. 3 illustrates a modified environment reflection projection method stored in main memory 210 of FIG. 2, according to the present invention.
- the modified reflection projection method is used by the illumination model.
- the modified reflection projection method is incorporated into the illumination model.
- 3 includes an object's surface 305, a point P on the surface 305, a normal vector n to the surface 305 at point P (also referred to as a surface normal vector), an observer 310a, an observation vector e directed from the point P to observer 310a, and an x-axis 320.
- surface 305 is composed of polygon primitives (not shown), and at each vertex of each polygon, a point P is specified on surface 305.
- the z-axis (not shown) is perpendicular to x-axis 320 and is in the plane of FIG.
- FIG. 3 embodiment of surface 305 is a line, however, any point P on any two- dimensional surface is within the scope of the invention.
- the FIG. 3 embodiment of surface 305 may be the intersection of a two-dimensional surface (not shown) with the x-z plane, and thus the normal vector n may have a vector component n y along the y-axis.
- the modified environment reflection projection method uses a modified version of a standard reflection formula to calculate a reflection vector r for each point P on surface 305.
- the method then processes -the reflection vector r to generate texture coordinates (s,t) for each point P.
- the standard reflection formula gives a reflection vector r based upon a given observer position specified by the observation vector e.
- the standard reflection formula is a vector relationship that satisfies Snell's law of reflection, where the angle of incidence at (FIG. 3) is equal to the angle of reflection a r (FIG. 3).
- VPU 213 uses a modified version of the standard reflection formula to compute reflection vectors.
- the modified reflection projection method produces the same result as the direct normal projection method when the point P on surface 305 is located directly in front of observer 310a.
- the pseudo-normal vector n' is substituted for the normal vector n in the standard reflection formula
- n' (e + n)/(
- ) [nx, ny, nz- l]/(sqrt(nx 2 + ny 2 + (nz-1) 2 )).
- FIG. 4 is a flowchart of method steps for displaying an image of a reflective object based upon texture coordinates, according to one embodiment of the invention.
- a user loads video software into memory 210 (FIG. 2) via optical disc control unit 234 (FIG. 2), for example, and CPU 212 (FIG. 2) executes the video software.
- the video software may be an interactive or non-interactive video, and in an exemplary embodiment of the invention, the video software is a video game.
- CPU 212 generates rendering instructions for all reflective objects of a video frame. The rendering instructions may be generated in response to user input received via controller interface 220 (FIG. 2).
- VPU 213 (FIG. 2) executes the rendering instructions using an illumination model, and generates reflection vectors for each reflective object of the video frame. For example, a reflection vector is generated for each vertex point P (FIG.
- VPU 213 transforms the reflection vectors associated with each object to texture coordinates.
- the transformation may be a reflection vector mapping method, or may be configured using other known methods in the art.
- VPU 213 may compute a texture map composed of the texture coordinates for each reflective object.
- VPU 213 sends the texture coordinates and/or texture maps to GPU 214.
- GPU 214 prepares an image of each reflective object for display on a display device (not shown), based in part on the texture coordinates or texture map associated with each reflective object.
- GPU 214 may use other illumination terms generated by VPU 213 or CPU 212 in conjunction with the texture coordinates to prepare each reflective object for display as an image.
- step 435 CPU 212 determines if execution of the video game has been terminated. If execution has not terminated, then the method continues with the next video frame at step 410. However, if in step 435, CPU 212 ascertains that execution of the video game has terminated, then the method ends.
- FIG. 5 is a flowchart of method steps for generating reflection vectors for a reflective object, according to one embodiment of the invention.
- VPU 213 (FIG. 2) selects a vertex point P (FIG. 3) on surface 305 (FIG. 3) of the reflective object.
- VPU 213 obtains vector components [nx,ny,nz] of a normal vector n (FIG. 3) to the surface 305 of the reflective object at the selected vertex point P.
- the vector components are stored in registers (not shown) associated with CPU 212.
- the vector components are stored in a memory (not shown) associated with VPU 213.
- VPU 213 determines components [ex",ey",ez”] of the observation vector e" (FIG. 3), for example, directed from the point P to observer 310b (FIG. 3). VPU 213 may compute the components [ex",ey",ez”] or may receive the components from VPU 213 memory (not shown) or CPU 212 registers (not shown). Next, VPU 213 uses a modified reflection formula, the components (ex",ey",ez”) of the observation vector e", and the vector components (nx,ny,nz) of the normal vector n to compute a reflection vector r" (FIG. 3), in step 520.
- VPU 213 determines whether a reflection vector for each vertex point P associated with the reflective object has been computed. If a reflection vector for each vertex point P has not been computed, then in step 530, VPU 213 selects another vertex point P on surface 305 of the reflective object. The method then continues at step 5101 However, if in step 525, a reflection vector for each vertex point P has been computed, then the method ends.
- the modified environment reflection projection method may be executed in part or in whole by CPU 212, VPU 213, GPU 214, or a rendering engine (not shown). Or, for example, the modified environment reflection projection method may be implemented in parallel by a multiprocessor system. Additionally, the present invention may effectively be used in conjunction with systems other than those described in the embodiments above. Therefore, these and other variations upon the disclosed embodiments are intended to be covered by the present invention, which is limited only by the appended claims.
Abstract
Description
Claims
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US32849001P | 2001-10-10 | 2001-10-10 | |
US328490P | 2001-10-10 | ||
PCT/US2002/032421 WO2003032253A2 (en) | 2001-10-10 | 2002-10-08 | System and method for environment mapping |
Publications (3)
Publication Number | Publication Date |
---|---|
EP1442390A2 EP1442390A2 (en) | 2004-08-04 |
EP1442390A4 true EP1442390A4 (en) | 2007-05-16 |
EP1442390B1 EP1442390B1 (en) | 2017-09-13 |
Family
ID=23281204
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP02770564.9A Expired - Lifetime EP1442390B1 (en) | 2001-10-10 | 2002-10-08 | System and method for environment mapping |
Country Status (6)
Country | Link |
---|---|
US (3) | US7046245B2 (en) |
EP (1) | EP1442390B1 (en) |
JP (1) | JP2005506611A (en) |
AU (1) | AU2002335799A1 (en) |
TW (1) | TWI244623B (en) |
WO (1) | WO2003032253A2 (en) |
Families Citing this family (30)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP1442390B1 (en) | 2001-10-10 | 2017-09-13 | Sony Computer Entertainment America LLC | System and method for environment mapping |
US8133115B2 (en) | 2003-10-22 | 2012-03-13 | Sony Computer Entertainment America Llc | System and method for recording and displaying a graphical path in a video game |
GB2408661B (en) | 2003-11-27 | 2008-02-06 | Sony Comp Entertainment Europe | Image rendering |
US20060071933A1 (en) | 2004-10-06 | 2006-04-06 | Sony Computer Entertainment Inc. | Application binary interface for multi-pass shaders |
US7349066B2 (en) * | 2005-05-05 | 2008-03-25 | Asml Masktools B.V. | Apparatus, method and computer program product for performing a model based optical proximity correction factoring neighbor influence |
US7636126B2 (en) | 2005-06-22 | 2009-12-22 | Sony Computer Entertainment Inc. | Delay matching in audio/video systems |
US7880746B2 (en) | 2006-05-04 | 2011-02-01 | Sony Computer Entertainment Inc. | Bandwidth management through lighting control of a user environment via a display device |
US7965859B2 (en) | 2006-05-04 | 2011-06-21 | Sony Computer Entertainment Inc. | Lighting control of a user environment via a display device |
JP4031509B1 (en) * | 2006-09-21 | 2008-01-09 | 株式会社コナミデジタルエンタテインメント | Image processing apparatus, image processing apparatus control method, and program |
US10786736B2 (en) | 2010-05-11 | 2020-09-29 | Sony Interactive Entertainment LLC | Placement of user information in a game space |
US9342817B2 (en) | 2011-07-07 | 2016-05-17 | Sony Interactive Entertainment LLC | Auto-creating groups for sharing photos |
US9865074B2 (en) | 2014-04-05 | 2018-01-09 | Sony Interactive Entertainment America Llc | Method for efficient construction of high resolution display buffers |
US10783696B2 (en) | 2014-04-05 | 2020-09-22 | Sony Interactive Entertainment LLC | Gradient adjustment for texture mapping to non-orthonormal grid |
US9495790B2 (en) | 2014-04-05 | 2016-11-15 | Sony Interactive Entertainment America Llc | Gradient adjustment for texture mapping to non-orthonormal grid |
US9710881B2 (en) | 2014-04-05 | 2017-07-18 | Sony Interactive Entertainment America Llc | Varying effective resolution by screen location by altering rasterization parameters |
US10068311B2 (en) | 2014-04-05 | 2018-09-04 | Sony Interacive Entertainment LLC | Varying effective resolution by screen location by changing active color sample count within multiple render targets |
US9836816B2 (en) | 2014-04-05 | 2017-12-05 | Sony Interactive Entertainment America Llc | Varying effective resolution by screen location in graphics processing by approximating projection of vertices onto curved viewport |
US9652882B2 (en) | 2014-04-05 | 2017-05-16 | Sony Interactive Entertainment America Llc | Gradient adjustment for texture mapping for multiple render targets with resolution that varies by screen location |
US11302054B2 (en) | 2014-04-05 | 2022-04-12 | Sony Interactive Entertainment Europe Limited | Varying effective resolution by screen location by changing active color sample count within multiple render targets |
US9911395B1 (en) * | 2014-12-23 | 2018-03-06 | Amazon Technologies, Inc. | Glare correction via pixel processing |
EP3057067B1 (en) * | 2015-02-16 | 2017-08-23 | Thomson Licensing | Device and method for estimating a glossy part of radiation |
US9760113B2 (en) | 2015-02-20 | 2017-09-12 | Sony Interactive Entertainment America Llc | Backward compatibility through use of spoof clock and fine grain frequency control |
US11403099B2 (en) | 2015-07-27 | 2022-08-02 | Sony Interactive Entertainment LLC | Backward compatibility by restriction of hardware resources |
US10235219B2 (en) | 2015-07-27 | 2019-03-19 | Sony Interactive Entertainment America Llc | Backward compatibility by algorithm matching, disabling features, or throttling performance |
US9892024B2 (en) | 2015-11-02 | 2018-02-13 | Sony Interactive Entertainment America Llc | Backward compatibility testing of software in a mode that disrupts timing |
CN108885552B (en) | 2016-01-22 | 2023-03-14 | 索尼互动娱乐股份有限公司 | Spoofing CPUID for backward compatibility |
JP6728370B2 (en) | 2016-01-22 | 2020-07-22 | 株式会社ソニー・インタラクティブエンタテインメント | Simulation of legacy bus behavior for backward compatibility |
JP6453929B2 (en) * | 2017-03-15 | 2019-01-16 | 株式会社Subaru | Vehicle display system and method for controlling vehicle display system |
JP6470334B2 (en) * | 2017-03-15 | 2019-02-13 | 株式会社Subaru | Vehicle display system and method for controlling vehicle display system |
CN117870034B (en) * | 2024-03-11 | 2024-05-07 | 深圳市德尼环境技术有限公司 | Control method, device and system for environmental parameters of clean room |
Family Cites Families (78)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPS5513582A (en) | 1978-07-13 | 1980-01-30 | Sanyo Electric Co Ltd | Color television receiver |
US4241241A (en) * | 1978-11-20 | 1980-12-23 | Janice K. Scarpitti | Electromechanical record disk dynamic range expander |
US4241341A (en) | 1979-03-05 | 1980-12-23 | Thorson Mark R | Apparatus for scan conversion |
US4646075A (en) | 1983-11-03 | 1987-02-24 | Robert Bosch Corporation | System and method for a data processing pipeline |
US5555532A (en) | 1984-05-23 | 1996-09-10 | The United States Of America As Represented By The Secretary Of The Navy | Method and apparatus for target imaging with sidelooking sonar |
US4658247A (en) | 1984-07-30 | 1987-04-14 | Cornell Research Foundation, Inc. | Pipelined, line buffered real-time color graphics display system |
JPH0746391B2 (en) | 1984-09-14 | 1995-05-17 | 株式会社日立製作所 | Graphic seeding device |
US4737921A (en) | 1985-06-03 | 1988-04-12 | Dynamic Digital Displays, Inc. | Three dimensional medical image display system |
DE3650494T2 (en) | 1985-07-05 | 1996-10-02 | Dainippon Printing Co Ltd | DESIGNING METHOD AND DEVICE OF THREE-DIMENSIONAL CONTAINERS |
JPH0814854B2 (en) | 1985-10-11 | 1996-02-14 | 株式会社日立製作所 | 3D graphic display device |
US4866637A (en) | 1987-10-30 | 1989-09-12 | International Business Machines Corporation | Pipelined lighting model processing system for a graphics workstation's shading function |
US4901064A (en) * | 1987-11-04 | 1990-02-13 | Schlumberger Technologies, Inc. | Normal vector shading for 3-D graphics display system |
US5369737A (en) | 1988-03-21 | 1994-11-29 | Digital Equipment Corporation | Normalization of vectors associated with a display pixels of computer generated images |
US5448687A (en) | 1988-09-13 | 1995-09-05 | Computer Design, Inc. | Computer-assisted design system for flattening a three-dimensional surface and for wrapping a flat shape to a three-dimensional surface |
US5222203A (en) | 1989-01-20 | 1993-06-22 | Daikin Industries, Ltd. | Method and apparatus for displaying translucent surface |
US5064291A (en) | 1990-04-03 | 1991-11-12 | Hughes Aircraft Company | Method and apparatus for inspection of solder joints utilizing shape determination from shading |
US5268996A (en) | 1990-12-20 | 1993-12-07 | General Electric Company | Computer image generation method for determination of total pixel illumination due to plural light sources |
US5537638A (en) | 1991-10-25 | 1996-07-16 | Hitachi, Ltd. | Method and system for image mapping |
US5734384A (en) | 1991-11-29 | 1998-03-31 | Picker International, Inc. | Cross-referenced sectioning and reprojection of diagnostic image volumes |
JP3107452B2 (en) | 1992-04-28 | 2000-11-06 | 株式会社日立製作所 | Texture mapping method and apparatus |
IL102289A (en) | 1992-06-24 | 1997-08-14 | R Technologies Ltd B V | Method and system for processing moving images |
EP0582875B1 (en) | 1992-07-27 | 2001-10-31 | Matsushita Electric Industrial Co., Ltd. | Apparatus for parallel image generation |
US5361385A (en) | 1992-08-26 | 1994-11-01 | Reuven Bakalash | Parallel computing system for volumetric modeling, data processing and visualization |
JP2682559B2 (en) * | 1992-09-30 | 1997-11-26 | インターナショナル・ビジネス・マシーンズ・コーポレイション | Apparatus and method for displaying image of object on display device and computer graphics display system |
GB2271259A (en) | 1992-10-02 | 1994-04-06 | Canon Res Ct Europe Ltd | Processing image data |
JP3233376B2 (en) | 1993-03-15 | 2001-11-26 | 株式会社リコー | Shading processing equipment |
EP0622747B1 (en) | 1993-04-01 | 2000-05-31 | Sun Microsystems, Inc. | Method and apparatus for an adaptive texture mapping controller |
JPH0757117A (en) | 1993-07-09 | 1995-03-03 | Silicon Graphics Inc | Forming method of index to texture map and computer control display system |
US5550960A (en) | 1993-08-02 | 1996-08-27 | Sun Microsystems, Inc. | Method and apparatus for performing dynamic texture mapping for complex surfaces |
EP0641993B1 (en) | 1993-09-03 | 1999-06-30 | Canon Kabushiki Kaisha | Shape measuring apparatus |
JPH09512937A (en) | 1994-09-06 | 1997-12-22 | ザ リサーチ ファウンデーション オブ ステイト ユニヴァーシティ オブ ニューヨーク | Apparatus and method for real-time visualization of volume |
JP2642070B2 (en) | 1994-11-07 | 1997-08-20 | インターナショナル・ビジネス・マシーンズ・コーポレイション | Method and system for generating quadrilateral mesh |
GB9501832D0 (en) | 1995-01-31 | 1995-03-22 | Videologic Ltd | Texturing and shading of 3-d images |
US5704024A (en) * | 1995-07-20 | 1997-12-30 | Silicon Graphics, Inc. | Method and an apparatus for generating reflection vectors which can be unnormalized and for using these reflection vectors to index locations on an environment map |
US5870097A (en) | 1995-08-04 | 1999-02-09 | Microsoft Corporation | Method and system for improving shadowing in a graphics rendering system |
JP3203160B2 (en) | 1995-08-09 | 2001-08-27 | 三菱電機株式会社 | Volume rendering apparatus and method |
KR100261076B1 (en) | 1995-11-09 | 2000-07-01 | 윤종용 | Rendering method and apparatus of performing bump mapping and phong shading at the same time |
US6034693A (en) | 1996-05-28 | 2000-03-07 | Namco Ltd. | Image synthesizing apparatus, image synthesizing method and information storage medium |
JPH1055454A (en) | 1996-05-28 | 1998-02-24 | Namco Ltd | Picture composing device, its method and information recording medium |
US5781194A (en) | 1996-08-29 | 1998-07-14 | Animatek International, Inc. | Real-time projection of voxel-based object |
JP3358169B2 (en) * | 1996-08-30 | 2002-12-16 | インターナショナル・ビジネス・マシーンズ・コーポレーション | Mirror surface rendering method and apparatus |
JP3387750B2 (en) | 1996-09-02 | 2003-03-17 | 株式会社リコー | Shading processing equipment |
US5935198A (en) | 1996-11-22 | 1999-08-10 | S3 Incorporated | Multiplier with selectable booth encoders for performing 3D graphics interpolations with two multiplies in a single pass through the multiplier |
US5949424A (en) * | 1997-02-28 | 1999-09-07 | Silicon Graphics, Inc. | Method, system, and computer program product for bump mapping in tangent space |
JP4244391B2 (en) | 1997-04-04 | 2009-03-25 | ソニー株式会社 | Image conversion apparatus and image conversion method |
US6175367B1 (en) | 1997-04-23 | 2001-01-16 | Siligon Graphics, Inc. | Method and system for real time illumination of computer generated images |
US5912830A (en) | 1997-04-30 | 1999-06-15 | Hewlett-Packard Co. | System and method for conditionally calculating exponential values in a geometry accelerator |
US6208347B1 (en) | 1997-06-23 | 2001-03-27 | Real-Time Geometry Corporation | System and method for computer modeling of 3D objects and 2D images by mesh constructions that incorporate non-spatial data such as color or texture |
US6226006B1 (en) * | 1997-06-27 | 2001-05-01 | C-Light Partners, Inc. | Method and apparatus for providing shading in a graphic display system |
US5914724A (en) | 1997-06-30 | 1999-06-22 | Sun Microsystems, Inc | Lighting unit for a three-dimensional graphics accelerator with improved handling of incoming color values |
US6037947A (en) | 1997-10-16 | 2000-03-14 | Sun Microsystems, Inc. | Graphics accelerator with shift count generation for handling potential fixed-point numeric overflows |
ES2175326T3 (en) | 1997-12-19 | 2002-11-16 | Esab Ab | WELDING APPARATUS |
JPH11203501A (en) | 1998-01-14 | 1999-07-30 | Sega Enterp Ltd | Picture processor and picture processing method |
US6014144A (en) | 1998-02-03 | 2000-01-11 | Sun Microsystems, Inc. | Rapid computation of local eye vectors in a fixed point lighting unit |
US6577312B2 (en) | 1998-02-17 | 2003-06-10 | Sun Microsystems, Inc. | Graphics system configured to filter samples using a variable support filter |
US6850236B2 (en) | 1998-02-17 | 2005-02-01 | Sun Microsystems, Inc. | Dynamically adjusting a sample-to-pixel filter in response to user input and/or sensor input |
AU5688199A (en) | 1998-08-20 | 2000-03-14 | Raycer, Inc. | System, apparatus and method for spatially sorting image data in a three-dimensional graphics pipeline |
US6771264B1 (en) | 1998-08-20 | 2004-08-03 | Apple Computer, Inc. | Method and apparatus for performing tangent space lighting and bump mapping in a deferred shading graphics processor |
JP2000137828A (en) | 1998-10-30 | 2000-05-16 | Fujitsu Ltd | Shading arithmetic unit and image processor using the arithmetic unit |
US6342885B1 (en) | 1998-11-12 | 2002-01-29 | Tera Recon Inc. | Method and apparatus for illuminating volume data in a rendering pipeline |
US6313842B1 (en) | 1999-03-03 | 2001-11-06 | Discreet Logic Inc. | Generating image data |
US6226007B1 (en) * | 1999-05-21 | 2001-05-01 | Sun Microsystems, Inc. | Method and apparatus for modeling specular reflection |
US6717579B1 (en) * | 1999-06-10 | 2004-04-06 | Dassault Systemes | Reflection line control |
US6504538B1 (en) * | 1999-07-01 | 2003-01-07 | Microsoft Corporation | Method and system for generating light values for a set of vertices |
US6611265B1 (en) | 1999-10-18 | 2003-08-26 | S3 Graphics Co., Ltd. | Multi-stage fixed cycle pipe-lined lighting equation evaluator |
US6686924B1 (en) | 2000-02-02 | 2004-02-03 | Ati International, Srl | Method and apparatus for parallel processing of geometric aspects of video graphics data |
US6664955B1 (en) | 2000-03-15 | 2003-12-16 | Sun Microsystems, Inc. | Graphics system configured to interpolate pixel values |
US6426755B1 (en) | 2000-05-16 | 2002-07-30 | Sun Microsystems, Inc. | Graphics system using sample tags for blur |
US6825851B1 (en) | 2000-08-23 | 2004-11-30 | Nintendo Co., Ltd. | Method and apparatus for environment-mapped bump-mapping in a graphics system |
GB2367471B (en) * | 2000-09-29 | 2002-08-14 | Pixelfusion Ltd | Graphics system |
US6765573B2 (en) | 2000-10-26 | 2004-07-20 | Square Enix Co., Ltd. | Surface shading using stored texture map based on bidirectional reflectance distribution function |
US6850243B1 (en) * | 2000-12-07 | 2005-02-01 | Nvidia Corporation | System, method and computer program product for texture address operations based on computations involving other textures |
US6778181B1 (en) | 2000-12-07 | 2004-08-17 | Nvidia Corporation | Graphics processing system having a virtual texturing array |
US6639594B2 (en) | 2001-06-03 | 2003-10-28 | Microsoft Corporation | View-dependent image synthesis |
US6781594B2 (en) * | 2001-08-21 | 2004-08-24 | Sony Computer Entertainment America Inc. | Method for computing the intensity of specularly reflected light |
EP1442390B1 (en) | 2001-10-10 | 2017-09-13 | Sony Computer Entertainment America LLC | System and method for environment mapping |
US6753870B2 (en) | 2002-01-30 | 2004-06-22 | Sun Microsystems, Inc. | Graphics system configured to switch between multiple sample buffer contexts |
US7589723B2 (en) * | 2005-07-25 | 2009-09-15 | Microsoft Corporation | Real-time rendering of partially translucent objects |
-
2002
- 2002-10-08 EP EP02770564.9A patent/EP1442390B1/en not_active Expired - Lifetime
- 2002-10-08 WO PCT/US2002/032421 patent/WO2003032253A2/en active Application Filing
- 2002-10-08 AU AU2002335799A patent/AU2002335799A1/en not_active Abandoned
- 2002-10-08 US US10/267,341 patent/US7046245B2/en not_active Expired - Lifetime
- 2002-10-08 JP JP2003535147A patent/JP2005506611A/en active Pending
- 2002-10-09 TW TW091123296A patent/TWI244623B/en not_active IP Right Cessation
-
2005
- 2005-09-07 US US11/222,207 patent/US7786993B2/en active Active
-
2010
- 2010-07-23 US US12/842,353 patent/US8174527B2/en active Active
Non-Patent Citations (2)
Title |
---|
BLINN J F ET AL: "TEXTURE AND REFLECTION IN COMPUTER GENERATED IMAGES", COMMUNICATIONS OF THE ASSOCIATION FOR COMPUTING MACHINERY, ACM, NEW YORK, NY, US, vol. 19, no. 10, 1 October 1976 (1976-10-01), pages 542 - 547, XP000654139, ISSN: 0001-0782 * |
VOORHIES D ET AL: "Reflection vector shading hardware", COMPUTER GRAPHICS PROCEEDINGS. ANNUAL CONFERENCE SERIES 1994. SIGGRAPH 94 CONFERENCE PROCEEDINGS ACM NEW YORK, NY, USA, 1994, pages 163 - 166, XP002428022, ISBN: 0-89791-667-0 * |
Also Published As
Publication number | Publication date |
---|---|
TWI244623B (en) | 2005-12-01 |
AU2002335799A1 (en) | 2003-04-22 |
WO2003032253A3 (en) | 2004-03-04 |
JP2005506611A (en) | 2005-03-03 |
WO2003032253A2 (en) | 2003-04-17 |
EP1442390A2 (en) | 2004-08-04 |
US20030112238A1 (en) | 2003-06-19 |
US7046245B2 (en) | 2006-05-16 |
EP1442390B1 (en) | 2017-09-13 |
US20060001674A1 (en) | 2006-01-05 |
US8174527B2 (en) | 2012-05-08 |
US7786993B2 (en) | 2010-08-31 |
US20100283783A1 (en) | 2010-11-11 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US7786993B2 (en) | Environment mapping | |
US6825851B1 (en) | Method and apparatus for environment-mapped bump-mapping in a graphics system | |
US8390621B2 (en) | System and method for calculating multi-resolution dynamic ambient occlusion | |
US8803879B1 (en) | Omnidirectional shadow texture mapping | |
US7292242B1 (en) | Clipping with addition of vertices to existing primitives | |
US6646640B2 (en) | System and method for creating real-time shadows of complex transparent objects | |
JP2002074389A (en) | Graphics processing system and method for performing embossed bump mapping | |
US6784882B1 (en) | Methods and apparatus for rendering an image including portions seen through one or more objects of the image | |
US6529194B1 (en) | Rendering method and apparatus, game system, and computer readable program product storing program for calculating data relating to shadow of object in virtual space | |
JP3971380B2 (en) | Calculation method of specular reflection light intensity | |
JP4584956B2 (en) | Graphics processor and drawing processing method | |
US6890261B2 (en) | Game system, program and image generation method | |
JP3576126B2 (en) | Image generation system, program, and information storage medium | |
JP2001325605A (en) | Game system and information storage medium | |
JP2001286675A (en) | Game device, information recording medium and game system | |
JP2000057372A (en) | Image processor, image processing method and storage medium | |
JP2001218979A (en) | Game system and information memory medium | |
JP2002358538A (en) | Device and method for compositing three-dimensional picture, information storage medium and device and method for distributing program |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
17P | Request for examination filed |
Effective date: 20040407 |
|
AK | Designated contracting states |
Kind code of ref document: A2 Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR IE IT LI LU MC NL PT SE SK TR |
|
AX | Request for extension of the european patent |
Extension state: AL LT LV MK RO SI |
|
A4 | Supplementary search report drawn up and despatched |
Effective date: 20070418 |
|
17Q | First examination report despatched |
Effective date: 20080506 |
|
RAP1 | Party data changed (applicant data changed or rights of an application transferred) |
Owner name: SONY COMPUTER ENTERTAINMENT AMERICA LLC |
|
RAP1 | Party data changed (applicant data changed or rights of an application transferred) |
Owner name: SONY COMPUTER ENTERTAINMENT AMERICA LLC |
|
REG | Reference to a national code |
Ref country code: DE Ref legal event code: R079 Ref document number: 60249084 Country of ref document: DE Free format text: PREVIOUS MAIN CLASS: G06F0017000000 Ipc: G06T0015040000 |
|
GRAP | Despatch of communication of intention to grant a patent |
Free format text: ORIGINAL CODE: EPIDOSNIGR1 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: GRANT OF PATENT IS INTENDED |
|
RIC1 | Information provided on ipc code assigned before grant |
Ipc: G06T 15/04 20110101AFI20170309BHEP Ipc: G06T 15/50 20110101ALI20170309BHEP |
|
INTG | Intention to grant announced |
Effective date: 20170405 |
|
GRAS | Grant fee paid |
Free format text: ORIGINAL CODE: EPIDOSNIGR3 |
|
GRAA | (expected) grant |
Free format text: ORIGINAL CODE: 0009210 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: THE PATENT HAS BEEN GRANTED |
|
AK | Designated contracting states |
Kind code of ref document: B1 Designated state(s): DE FR GB |
|
REG | Reference to a national code |
Ref country code: GB Ref legal event code: FG4D |
|
REG | Reference to a national code |
Ref country code: FR Ref legal event code: PLFP Year of fee payment: 16 |
|
REG | Reference to a national code |
Ref country code: DE Ref legal event code: R096 Ref document number: 60249084 Country of ref document: DE |
|
REG | Reference to a national code |
Ref country code: DE Ref legal event code: R097 Ref document number: 60249084 Country of ref document: DE |
|
PLBE | No opposition filed within time limit |
Free format text: ORIGINAL CODE: 0009261 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: NO OPPOSITION FILED WITHIN TIME LIMIT |
|
26N | No opposition filed |
Effective date: 20180614 |
|
REG | Reference to a national code |
Ref country code: FR Ref legal event code: PLFP Year of fee payment: 17 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: DE Payment date: 20211027 Year of fee payment: 20 Ref country code: GB Payment date: 20211027 Year of fee payment: 20 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: FR Payment date: 20211025 Year of fee payment: 20 |
|
REG | Reference to a national code |
Ref country code: DE Ref legal event code: R071 Ref document number: 60249084 Country of ref document: DE |
|
REG | Reference to a national code |
Ref country code: GB Ref legal event code: PE20 Expiry date: 20221007 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: GB Free format text: LAPSE BECAUSE OF EXPIRATION OF PROTECTION Effective date: 20221007 |
|
P01 | Opt-out of the competence of the unified patent court (upc) registered |
Effective date: 20230519 |