US6466220B1 - Graphics engine architecture - Google Patents
Graphics engine architecture Download PDFInfo
- Publication number
- US6466220B1 US6466220B1 US09/263,161 US26316199A US6466220B1 US 6466220 B1 US6466220 B1 US 6466220B1 US 26316199 A US26316199 A US 26316199A US 6466220 B1 US6466220 B1 US 6466220B1
- Authority
- US
- United States
- Prior art keywords
- screen memory
- data
- regional image
- regional
- graphics
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Lifetime
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T1/00—General purpose image data processing
- G06T1/20—Processor architectures; Processor configuration, e.g. pipelining
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G5/00—Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
- G09G5/36—Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators characterised by the display of a graphic pattern, e.g. using an all-points-addressable [APA] memory
- G09G5/39—Control of the bit-mapped memory
- G09G5/393—Arrangements for updating the contents of the bit-mapped memory
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G2340/00—Aspects of display data processing
- G09G2340/04—Changes in size, position or resolution of an image
- G09G2340/0407—Resolution change, inclusive of the use of different resolutions for different screen areas
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G2340/00—Aspects of display data processing
- G09G2340/10—Mixing of images, i.e. displayed pixel being the result of an operation, e.g. adding, on the corresponding input pixels
Definitions
- the invention relates to devices related to display of graphical data. More particularly, the invention relates to a pipelined graphics engine architecture.
- graphics have been used for different purposes with different methods of distribution.
- graphics formats have been developed. For example, analog televisions receive graphical information in a different format than computer systems, which have multiple formats for graphical information.
- Multi-regional graphics are graphics that are displayed as a single image where different regions have different graphical formats (e.g., index, RGB with or without alpha blending, YUV).
- support of multi-regional graphics is the ability to simultaneously display several graphical regions with different characteristics.
- Multi-regional graphics capability is useful, for example, for a television set-top box for accessing the Internet.
- multi-regional graphics exceeds the current capabilities of typical display controllers.
- a method and apparatus for display of graphical data is described.
- a multi-regional image having at least two regions represented by at least two color formats is rendered in an off-screen memory.
- the multi-regional image is copied to an on-screen memory.
- the regions of the multi-regional image are converted to a common color format during the copy operation.
- the multi-regional image stored in the on-screen memory is displayed.
- FIG. 1 is one embodiment a system suitable for use with the invention.
- FIG. 2 illustrates a general data flow of data to be processed according to the invention.
- FIG. 3 is a conceptual diagram of data flow for synchronizing data streams according to one embodiment of the invention.
- FIG. 4 is a state diagram for synchronizing data streams according to one embodiment of the invention.
- FIG. 5 is a timing diagram showing an exemplary sequence of control circuit states for synchronizing three independent data streams according to one embodiment of the invention.
- FIG. 6 is a conceptual illustration of multiple graphics planes according to one embodiment of the invention.
- FIG. 7 is a conceptual illustration of off-screen and on-screen graphics according to one embodiment of the invention.
- FIG. 8 is a conceptual illustration of an off-screen and on-screen data path according to one embodiment of the invention.
- FIG. 9 is a flow diagram for providing multi-regional graphics according to one embodiment of the invention.
- FIG. 10 is a block diagram of a data path according to one embodiment of the invention.
- FIG. 11 a flow diagram for performing register linked list operations according to one embodiment of the invention.
- FIG. 12 illustrates parameters for a region of interest according to one embodiment of the invention.
- FIG. 13 is a conceptual example of a linked list of operations according to one embodiment of the invention.
- FIG. 14 is a block diagram of components to provide programmable alpha blending according to one embodiment of the invention.
- the invention provides an architecture for graphics processing.
- the architecture includes pipelined processing and support for multi-regional graphics.
- a graphics driver according to the invention can receive multiple (e.g., 3) independent streams of graphical data that can be in different graphical formats.
- the independent streams are synchronized and converted to a common destination format prior to being processed.
- multi-regional graphics are supported with off-screen and on-screen memory regions for processing.
- the regions of the multi-regional graphic are rendered in an off-screen memory.
- the data in the off-screen memory are converted to a common format and copied to on-screen memory.
- the data in the on-screen memory is used to generate an output image.
- Alpha blending can also be programmed to provide varying levels of transparency or other graphical features.
- graphics processing is programmable and can be paced in response to video display events.
- FIG. 1 is one embodiment of a system suitable for use with the invention.
- System 100 includes bus 105 or other communication device to communicate information and processor 110 (also referred to as a CPU in some embodiments) coupled to bus 105 to process information. While system 100 is illustrated with a single processor, system 100 can include multiple processors.
- System 100 further includes main memory 130 that can be random access memory (RAM) or other dynamic storage device, coupled to bus 105 to store information and instructions to be executed by processor 105 .
- Main memory 130 also can be used for storing temporary variables or other intermediate information during execution of instructions by processor 110 .
- RAM random access memory
- System 100 also includes read only memory (ROM) and/or other static storage device 120 coupled to bus 105 to store static information and instructions for processor 105 .
- Data storage device 180 is coupled to bus 105 to store information and instructions.
- Data storage device 180 such as a magnetic disk or optical disc and corresponding drive can be coupled to system 100 .
- Audio/visual/graphics (A/V/G) decoder 140 is coupled to bus 105 to receive A/V/G data.
- A/V decoder 140 can also receive data directly.
- A/V decoder 140 is an MPEG decoder that decodes digital A/V/G data according to one of the Motion Picture Experts Group standards (e.g., MPEG-1, MPEG-2, MPEG-4, MPEG-J, MPEG-2000).
- A/V decoder 140 can also be an analog decoder that decodes A/V/G data according to the national Television Standards Committee (NTSC) and/or Phase Alternation Line (PAL) and/or Séquentiel Couelles à Mémoire (SECAM) standards.
- NTSC National Television Standards Committee
- PAL Phase Alternation Line
- SECAM Séquentiel Couliv à Mémoire
- decoder memory 145 is coupled to A/V decoder 140 for use in decoding A/V data.
- A/V/G processor 150 is coupled to A/V decoder 140 to receive the output of A/V decoder 140 .
- A/V decoder 140 provides A/V/G processor 150 with one or more video data inputs and/or one or more audio data inputs.
- A/V/G processor 150 is also coupled to bus 105 to communicate with processor 105 and other components of system 100 .
- A/V/G processor 150 can also be coupled to multiple A/V/G decoders (not shown in FIG. 1 ).
- A/V/G memory 155 is coupled to A/V/G processor 150 .
- A/V/G memory 155 is used for A/V/G processing as described in greater detail below.
- A/V/G processor 150 uses main memory 130 for A/V/G processing rather than A/V/G memory 155 .
- Video device(s) 160 and audio device(s) 170 are coupled to A/V/G processor 150 .
- Video device(s) 160 represents one or more devices configured to display video or other graphical data output by A/V/G processor 150 .
- audio device(s) 170 represent one or more devices configured to generate audio output based on audio data generated by A/V/G processor 150 .
- A/V/G processor 150 generates two video output channels corresponding to multi-regional graphics and video in one channel and background video on a second channel; however, other configurations can also be provided.
- A/V/G processor also generates one or more audio output channels based, at least in part, on corresponding input audio channels.
- One embodiment of the present invention is related to the use of system 100 to provide processing of graphical information.
- processing of graphical information is performed by system 100 in response to processor 105 executing sequences of instructions contained in main memory 130 .
- Processing of graphical information can also be performed in response to A/V/G processor 150 executing sequences of instructions stored in main memory 130 or A/V/G memory 155 .
- main memory 130 Instructions are provided to main memory 130 from a storage device, such as magnetic disk, a ROM integrated circuit, CD-ROM, DVD, via a remote connection (e.g., over a network), etc.
- a storage device such as magnetic disk, a ROM integrated circuit, CD-ROM, DVD
- a remote connection e.g., over a network
- hard-wired circuitry can be used in place of or in combination with software instructions to implement the present invention.
- the present invention is not limited to any specific combination of hardware circuitry and software instructions.
- data input streams are scanned according to the standard progressive sequence used in NTSC and PAL encoding.
- an image is scanned starting from the pixel in the top left corner horizontally across to the pixel in the top right corner of the image. The next line down in the image is scanned from left to right. This scanning pattern is repeated until the image is completely scanned.
- the streams can have different widths in pixels; however, in one embodiment the various images start from the same pixel location (e.g., top left corner of the image).
- FIG. 2 illustrates a general data flow of data to be processed according to the invention.
- data rates are illustrated with arrow widths. The wider the arrow, the higher the data rate.
- One or more of the elements of FIG. 2 can be included in A/V/G processor 150 .
- Data sources 200 , 201 and 202 represent sources of A/V data to be processed.
- the data sources can be, for example, analog television channels, digital television channels, DVD players, VCRs.
- the data stream provided by each data source can vary from the other sources depending on, for example, data format. Varying data rates are common due to color formats having different bits per pixel. For example, 8-bit color indexed format requires and 8-bit value to represent a pixel. Thus, four pixels can be transferred through a 32-bit wide data path in a single clock cycle. However, 32-bit RGB color format requires all 32 bits to represent a single pixel. Thus, only a single pixel can be transferred through a 32-bit wide data path in a single clock cycle.
- conversion of one or more data streams to a common format can cause different latencies based on the conversions performed.
- conversion from indexed color formats to RGB color formats require retrieving a value from a look up table, the latency for which can vary depending on the location of the value in the table. The corresponding conversion latency varies in response to the look up latency.
- FIG. 2 assumes that data stream 210 is graphical data in a first format where the data rate is 1 Mbyte/sec.
- data stream 211 is graphical data in a second format where the data rate is 2 Mbyte/sec.
- data stream 212 is graphical data in a third format where the data rate is 0.3 Mbyte/sec.
- these data rates are not required by the invention.
- pipeline 220 has a longer latency (represented by a number of stages) than pipeline 221 .
- pipeline 222 has a longer latency than either pipeline 220 or 221 .
- Data streams 230 , 231 and 232 are output from pipelines 220 , 221 and 222 , respectively and provide input to pixel processing circuit 240 .
- Pixel processing circuit 240 operates on pixels received via data streams 230 , 231 and 232 . However, because data streams 230 , 231 and 232 have different data rates, the arrival of pixel data at pixel processing circuit 240 is not synchronized. In order to generate an accurate output pixel based on multiple input pixels, the pixels must, at some point in processing, be synchronized. Pixel processing circuit 240 operates on data streams 230 , 231 and 232 to synchronize the pixels received.
- Pixel processing circuit 240 performs one or more operations (e.g., boolean operations, alpha blending) on the pixels received from the pixel source buffers to generate an output pixel.
- Pixel operator 260 receives synchronized pixels from pixel control circuit 240 via pixel streams 250 , 251 and 252 . The output pixel is used to generate an output image.
- the components of FIG. 2 include pixel mirroring circuitry.
- the pixel mirroring circuitry allows pixel processing that is independent of the horizontal scanning direction.
- pixel source buffers included in pipelines 220 , 221 and 222 perform mirroring operations when necessary on data streams received.
- Pixel operator 260 reverses the mirroring operations when necessary to generate an output pixel.
- Pixel mirroring allows operations performed by pixel processing circuit 240 to be the same for images that are processed from right to left and for images that are processed from left to right.
- the use of the same operations for right to left processing and left to right processing reduces the size and complexity of pixel processing circuit 240 as compared to a circuit designed for processing images both right to left and left to right.
- the ability to perform both right to left and left to right scanning is useful, for example, when overlapping images are processed.
- mirroring is accomplished by a set of multiplexors included in the pixel source buffers of pipelines 220 , 221 and 222 ; however, mirroring can be accomplished by different circuitry.
- Pixel mirroring reverses the order of pixels received by the pixel source buffers. The reversal of pixel ordering allows right to left scanned images to be processed with the same operations as used for left to right scanned images because the scanning order is effectively reversed by the pipeline circuitry.
- pixel operator 260 includes circuitry to reverse the mirroring performed by the pipeline circuitry. If a mirrored image is desired pixel operator 260 does not reverse the mirroring performed by the pipeline circuitry.
- pixel mirroring is supported for multiple pixel widths. For example, if a 32-bit data path is communicating 1-bit color coded pixels, the order of the bits received are reversed in a bitwise manner rather than reversing the order of bytes that are received as a 32-bit word.
- FIG. 3 is a conceptual diagram of data flow for synchronizing data streams according to one embodiment of the invention.
- the flow control described with respect to FIG. 3 provides three basic functions that can be used to synchronize independent data streams.
- the flow control functions are: 1) the ability to independently halt any channel at any time; 2) the ability to identify which cycles contains a pipeline bubble; and 3) the ability to mark the boundary of a region of interest. Additional and/or different functions can also be provided for flow control or other purposes.
- function 310 operates on data received and provides the result to register 340 .
- Pipeline stage 330 corresponds to function 320 that operates on data received from register 340 to generate output data that is stored in register 350 .
- a data stream may be stalled because the stream may be ahead of other streams.
- a synchronization signal labeled PIPELINE_READY in FIG. 3, is provided to each pipeline stage processing a particular data stream.
- the PIPELINE_READY signal can be generated, for example, by pixel control circuit 240 .
- PIPELINE_READY when PIPELINE_READY is deasserted, all pipeline stages prior to the pipeline stage in which the signal is asserted are halted, unless pipeline bubbles are detected. Pipeline bubbles are described in greater detail below.
- PIPELINE_READY is asserted data from a previous pipeline stage is accepted and assumed valid.
- a pipeline bubble exists when a pipeline stage contains invalid data. In order to improve overall performance, it is desirable to reduce the number of pipeline bubbles as much as possible. Bubbles can be injected into the pipeline, for example, during an idle stage or between data streams having different data rates. Identifying and eliminating pipeline bubbles provides functional correctness and the ability to accelerate pipeline throughput.
- Elimination bubbles is the exception to the use of the PIPELINE_READY signal described above.
- all prior stages in the pipeline have the respective PIPELINE_READY signals asserted regardless of actual state. This allows data in the pipeline to advance, thereby removing the bubble from the pipeline.
- identification of pipeline bubbles is accomplished with a data validity signal, labeled DATA_VALID in FIG. 3 .
- DATA_VALID a data validity signal
- the associated pipeline stage contains a bubble.
- the DATA_VALID signal is asserted, the associated pipeline stage is processed in the normal manner.
- a region of interest is defined for processing purposes.
- the region of interest is the maximum subset area of all active images.
- the region of interest is the largest area that all images have in common.
- Other regions of interest can also be defined.
- a region of interest signal labeled DATA_LINESYNC, is used to indicate an end of an region of interest.
- the DATA_LINESYNC signal is asserted, the associated pixel is the last pixel of a horizontal line. Subsequent pixels for the line are discarded.
- a control mechanism for example, pixel control circuit 240 , uses the signals described above to control pixel flow and synchronize independent data streams. This includes the ability to discard pixels that are outside the region of interest. This also enables elimination of pipeline bubbles.
- FIG. 4 is a state diagram for synchronizing data streams according to one embodiment of the invention.
- four states are used to control and synchronize multiple independent channels.
- the embodiment described can be applied to any number of independent channels, which makes the architecture described herein advantageously scalable.
- State 400 is the initial state of a control circuit (e.g., pixel control circuit 240 ) after reset.
- a control circuit e.g., pixel control circuit 240
- all channels are synchronized and may include bubbles in the associated pipeline.
- the control circuit remains in state 400 until at least one pixel is received with the DATA_VALID signal asserted.
- the control circuit moves to state 410 .
- state 410 at least one channel has a valid pixel that is ready for processing as indicated by the DATA_VALID signal.
- the control circuit halts the one or more channels with valid pixel data by asserting the PIPELINE_READY signal for the channels having the DATA_VALID signal asserted.
- the control circuit asserts the PIPELINE_READY to the remaining channels.
- the PIPELINE_READY signal for the respective channels is asserted until the channel has valid pixel data that is ready for processing.
- the control circuit manages the channels with the PIPELINE_READY signal until the channels are synchronized. When the channels are synchronized, the system moves to state 420 .
- the channels are synchronized and have entered the region of interest.
- the synchronized pixels from the one or more channels are processed.
- one or more of the channels may have a bubble.
- state 430 at least one channel has reached the last valid pixel for a particular line.
- the control circuit stores the last pixel of a channel and asserts the PIPELINE_READY signal for the channel. Allowing the channel to advance at this stage reduces pipeline bubbles. The PIPELINE_READY signal is deasserted when the next valid pixel is available.
- the control circuit is not required to detect the DATA_LINESYNC for each channel because each channel is not required to provide the same number of valid pixels. For example, when one of the input channels provides pixel data in YUV format, four components (Y 0 ,U,Y 1 ,V ) are provided. The Y 0 and Y 1 components are the luminance components of two pixels and the two pixels share the same chrominance components, U and V. As a result, the number of YUV pixels per line is even. However, other formats, such as RGB, do not require an even number of pixels per line.
- the output image is the size of the smallest input image, only one DATA_LINESYNC from one channel is necessary. Once the last set of pixels in the region of interest is output, all channels are assumed synchronized and the control circuit returns to state 400 to process the subsequent image line.
- FIG. 5 is a timing diagram showing an exemplary sequence of control circuit states for synchronizing three independent data streams according to one embodiment of the invention.
- Channel A presents the first DATA_VALID pixel, which drives the control circuit to the SSYN state from the initial SYND state.
- the invention supports multi-regional graphics with off-screen and on-screen display buffers, which are described in greater detail below.
- the invention also uses multiple graphics planes to display multi-regional graphics.
- the off-screen display buffer is an area in memory that is not visible to a display viewer.
- the on-screen display buffer is an area in the memory that is periodically read and displayed to the viewer.
- FIG. 6 is a conceptual illustration of multiple graphics planes according to one embodiment of the invention.
- FIG. 6 illustrates four graphics planes (background, scaled video, scaled graphics, and cursor); however, a different number of graphics planes can be used. Also, certain graphics formats are described for use with the respective graphics planes, but other format configurations can also be used.
- Background plane 600 provides the background for the final image output to display 640 .
- background plane 600 can be a fixed color, a fixed pattern that can be tiled to fill background plane 600 , or a flowthough video stream.
- background plane 600 is opaque.
- Scaled video plane 610 is a video image that overlays background plane 600 .
- scaled video plane 610 can be used to provide a picture-in-picture (PIP) image on display 640 .
- PIP picture-in-picture
- scaled video plane 610 is video in the YCbCr (YUV) format; however, other formats can also be supported.
- Scaled graphics plane 620 also overlays background plane 600 .
- Scaled graphics plane 620 can also overlay scaled video plane 620 .
- scaled graphics plane 620 can be either RGB, color indexed, or YCbCr formatted data. Both scaled video plane 610 and scaled graphics plane 620 can have some degree of transparency such that other images can been seen.
- Cursor plane 630 overlays background plane 600 , scaled video plane 610 and scaled graphics plane 620 .
- cursor plane 630 is a 64-by-64 pixel bitmap image in 4-bit indexed color format; however, other sizes and formats can also be supported.
- the multiple graphical planes are combined to provide a final output image to display 640 .
- FIG. 7 is a conceptual illustration of off-screen and on-screen graphics according to one embodiment of the invention.
- the off-screen and on-screen graphics programming described with respect to FIG. 7 can be used, for example, to combine a multiple regions in different formats to generate an output image having a predetermined format.
- Other types of graphics combinations transformations can also be accomplished using the off-screen and on-screen graphics processing.
- An image stored in off-screen memory 700 is rendered with graphics blocks that represent the various regions of the multi-regional graphics.
- On-screen memory 705 stores a copy of the image in off-screen memory 700 with all of the regions converted to a single graphics format.
- the invention provides a graphics engine that performs color reduction or color expansion as necessary to match source and destination color formats.
- Off-screen memory 700 includes data for rendering a display area having one or more color formats and alpha blending factors.
- five graphical regions create image 710 .
- Region 715 is represented in four-bit indexed color format with an alpha blending factor of Alpha 1
- region 720 is represented in 16-bit RGB color format with an alpha blending factor of Alpha 4
- region 725 is represented in eight-bit indexed color format with an alpha blending factor of Alpha 2 [n]
- region 730 is represented in eight-bit indexed color format with an alpha blending factor of Alpha 3
- region 735 is represented in 24-bit RGB color format with an alpha blending factor of Alpha 5 [n].
- Image 710 stored in off-screen memory 700 is copied to on-screen memory 705 to provide image 750 .
- all regions of image 750 are represented in 24-bit RGB format with 8-bit alpha blending; however, other formats can also be used.
- image 710 having multiple graphics formats stored in off-screen memory 700 to image 750 having a single graphics format, the invention supports multi-regional graphics.
- off-screen memory 700 to on-screen memory 705 also converts the graphical information from the format in which the information was received to a predetermined format.
- on-screen memory 705 to which the images of off-screen memory 700 are copied is used to generate the scaled graphics plane described above. Before copying, the scaled graphics plane has an alpha blending value of zero which makes the plane transparent. Regions 715 , 720 , 725 , 730 and 735 are copied to on-screen memory 705 and converted to a common format. The alpha blending formats are maintained during conversion to the standard format of image 750 . Image 750 is then output to display 795 .
- FIG. 8 is a conceptual illustration of an off-screen and on-screen data path according to one embodiment of the invention.
- graphics processor 800 includes two components, block transfer engine 820 and display controller 840 , that use off-screen memory 864 and on-screen memory 865 to generate an output image.
- Graphics processor 800 also includes other components that are not shown in FIG. 8, for example, input and output buffers.
- off-screen memory 864 and on-screen memory are both included in graphics memory 860 .
- off-screen memory 864 and on-screen memory 868 can be in different memory components, or off-screen memory 864 and on-screen memory 868 can be part of a different memory component, for example, system memory (not shown in FIG. 8 ).
- off-screen memory 864 stores graphical images corresponding to multiple regions in an image, where the graphical images can be represented by different color formats.
- Block transfer engine 820 copies the data from off-screen memory 864 to on-screen memory 868 .
- block transfer engine 820 performs color expansion or color reduction as necessary while performing boolean operations on the graphical data.
- on-screen memory 868 stores the scaled graphics plane for an output image. In one embodiment, the following graphical format conversions are supported.
- the following expansion is used to convert between color formats during a copy or other operation if the destination color depth is greater than the source color depth.
- bits used to represent color data in a source value are replicated where necessary to provide color data in a different format as a destination value.
- 8-bit source value has three red (R 2 , R 1 , R 0 ) bits, three green (G 2 , G 1 , G 0 ) bits, and two blue bits (B 1 , B 0 ), where the bit labeled with a zero is the least significant bit.
- the 12-bit destination value has four red bits, four green bits, and four blue bits. The red bits are shifted to the left by one and the most significant bit of the source data (R 2 ) is replicated as the least significant bit of the destination data.
- the destination green and blue bits are similarly processed.
- Color reduction is performed if the source pixel map has greater color depth than the destination.
- two modes of color reduction linear and dither, are supported.
- the dither color reduction applies a 2 ⁇ 2 pixel error diffusion operation to the destination pixels during rounding.
- the quantization error due to rounding is weighted according to the destination pixel location with respect to origin and pitch. If the result is less than one, the destination pixel color value is increased by one.
- the following operations are used for linear color reduction.
- the following conversion is used to convert YCbCr to 16-bit RGB images.
- [ ⁇ R G B ⁇ ] [ 1.164 1.596 0 1.164 - 0.813 - 0.392 1.164 0 2.017 ] ⁇ [ ( Y - 16 ) ( Cr - 128 ) ( Cb - 128 ) ⁇ ]
- Display controller reads the scaled graphics plane from on-screen memory 868 as well as scaled video, background and cursor data from memory 860 to generate an output image.
- the output image is output to a display device (not shown in FIG. 8) to be displayed.
- FIG. 9 is a flow diagram for providing multi-regional graphics according to one embodiment of the invention.
- the multi-regional graphics are provided as part an output having multiple graphics planes.
- multiple graphics planes are not required to provide multi-regional graphics as described.
- Multi-regional graphics are rendered in off-screen memory at 910 .
- the multi-regional graphics includes one or more regions, each of which can have a different color format, with or without alpha factors.
- the scaled graphics plane in on-screen memory is initialized at 920 .
- initialization is accomplished by setting all of the pixels of the scaled graphics plane to an alpha value of zero (transparent).
- the scaled graphics plane is transparent when the regions of the multi-regional graphics are stored in off-screen memory.
- the regions of the multi-regional graphics are copied from off-screen memory to on-screen memory.
- the color format of the regions are converted, if necessary, to a common format.
- the color conversions can be accomplished as described above.
- the alpha values corresponding to the regions in off-screen memory are used to generate appropriate alpha values in on-screen memory such that the transparency of the regions is maintained.
- the scaled graphics plane stored in on-screen memory is displayed at 940 .
- the display can have multiple planes as described above, or the scaled graphics plane can be the only plane used to generate an output image.
- FIG. 10 is a block diagram of a data path according to one embodiment of the invention.
- block transfer engine 1050 copies graphics data from off-screen memory to on-screen memory.
- block transfer engine 1050 performs color reduction/expansion, a logical operation and alpha blending on one or more sets of source data.
- Registers 1040 represents a set of registers that can be used to control block transfer engine 1050 .
- registers 1040 include 24 registers; however, any number of registers can be used.
- Registers 1040 are updated by a CPU or other device (not shown in FIG. 10) or registers 1040 can be updated automatically from list 1000 in memory 1060 . Updating registers 1040 from list 1000 operates as a linked list of registers described in greater detail below. Controlling block transfer engine 1050 with a linked list of registers reduces processing overhead of a system CPU or other device.
- Source 1010 and source 1020 represent graphical regions of the same size residing in memory 1060 .
- Source 1010 and source 1020 provide input to block transfer engine 1050 .
- Block transfer engine 1050 operates on the source graphical data to generate an output that is stored in destination 1030 , which is also fed back into block transfer engine 1050 . The feed back allows block transfer engine 1050 to perform multiple operations without writing a result to memory 1060 and reading the result for the next operation.
- block transfer engine 1050 supports all 256 possible boolean raster operations for up to two source bitmaps and one destination bitmap. Different combinations of input bitmaps, output bitmaps and operations can also be supported.
- the operation codes are those used in Windows® products available from Microsoft Corporation of Redmond, Wash.; however other operation codes can also be used.
- the invention supports a linked list feature where a set of graphical data instructions are loaded into memory to be executed by the block transfer engine.
- the instructions are fetched and executed by the block transfer engine without intervention by the CPU or other processing device.
- the operations are defined by programming registers read by the block transfer engine, for example, registers 1040 of FIG. 10 . Multiple operations are linked by a special purpose register referred to as the Link Address Register.
- FIG. 11 a flow diagram for performing register linked list operations according to one embodiment of the invention.
- a Link Address Register is programmed at 1100 .
- Linked list programming is enabled at 1110 .
- a bit is set in a control register to indicate whether linked list programming is enabled.
- a Mask Register is fetched at 1120 .
- the Mask Register is used to determine a subset of registers that are accessed by a corresponding operation. By accessing a subset of registers, only the registers that are necessary for an operation are accessed, which improves performance of graphics operations.
- the Mask Register is configured as described below; however, other configurations can also be used.
- Bit 1 of the Mask Register corresponds to the Pace register.
- the Pace register is used when graphical operations are paced or triggered by an event. Pacing can be used, for example, to provide animation of graphical sequences without processor intervention. In one embodiment, when operations are paced graphics operations are suspended until the occurrence of a display event or external stimulus. Display events are, for example, the display of a specified scan line or vertical sync. External events are, for example, events that are flagged by the device writing to a control register.
- Bit 2 of the Mask Register corresponds to the Interrupt Status Register.
- the Interrupt Status Register indicates whether an interrupt is generated when a block transfer is completed, whether an interrupt is generated when a chain of block transfers is completed, whether the graphics engine is provided with sufficient memory bandwidth, and bits to mask off certain predetermined interrupts. Other configurations can also be used.
- Bit 3 of the Mask Register corresponds to the Status Register.
- the Status Register is used for controlling paced operations and indicating the status of block transfer operations. Other configurations can also be used.
- Bit 4 of the Mask Register corresponds to the Destination Channel Configuration Register.
- the Destination Channel Configuration Register indicates the format of the destination data, whether the region is scanned from right to left or left to right, the number of lines in the region, and dither information. Other information can also be included in the Destination Channel Configuration Register.
- Bit 5 of the Mask Register corresponds to the Source 1 Channel Configuration Register.
- the Source 1 Channel Configuration Register indicates the format of the source data, whether the region is scanned from right to left or left to right, the number of lines in the region, whether the source uses a global color register, and the method used to reduce color when necessary. Other information can also be included in the Source 1 Channel Configuration Register.
- Bit 6 of the Mask Register corresponds to the Source 0 Channel Configuration Register.
- the Source 0 Channel Configuration Register indicates the format of the source data, whether the region is scanned from right to left or left to right, the number of lines in the region, whether the source uses a global color register, and the method used to reduce color when necessary. Other information can also be included in the Source 0 Channel Configuration Register.
- Bit 7 corresponds to the Alpha Parameter Configuration Register.
- the Alpha Parameter Configuration Register indicates a global alpha value for a source array, a global alpha value for the destination array, and parameters for blending equations.
- the Alpha Parameter Configuration Register provides support for programmable alpha blending.
- Bit 8 corresponds to the Background Register for Source 1 .
- the Background Register for Source 1 specifies a background registers for when Source 1 is in one bit per pixel format.
- Bit 9 corresponds to the Foreground or Global Color Register for Source 1 .
- the Foreground or Global Color Register for Source 1 specifies a foreground color when Source 1 is in one bit per pixel format, or the Foreground or Global Color Register for Source 1 indicates a global color for Source 1 .
- Other formats can also be used.
- Bit 10 corresponds to the Background Register for Source 0 .
- the Background Register for Source 0 specifies a background registers for when Source 0 is in one bit per pixel format.
- Bit 11 corresponds to the Foreground or Global Color Register for Source 0 .
- the Foreground or Global Color Register for Source 0 specifies a foreground color when Source 0 is in one bit per pixel format, or the Foreground or Global Color Register for Source 0 indicates a global color for Source 0 .
- Other formats can also be used.
- Bits 12 - 21 define a region of interest for Source 0 , Source 1 and Destination.
- FIG. 12 illustrates parameters for a region of interest according to one embodiment of the invention.
- the region of interest illustrated in FIG. 12 includes N lines bounded by the height, width, left address and right address.
- the pitch is twice the width.
- the Pitch can be used to indicate the distance between the end of the region of interest and the beginning of the subsequent line.
- Bit 22 corresponds to the Link Address Register.
- the Link Address Register stores a starting address for a linked list.
- Bit 23 indicates whether the program register is to be configured.
- the Link Address Register is fetched at 1130 .
- the Link Address Register indicates the address of the starting address of the linked list.
- the starting address of the linked list stores an operation to be performed by the block transfer engine.
- the values stored in the un-masked registers as indicated by the Mask Register are fetched at 1140 .
- the un-masked registers store data to be used in the operation to be performed. By fetching only the values in the un-masked registers, only the necessary values are fetched, which improves performance as compared to fetching the values in all of the registers indicated by the Mask Register.
- the operation is executed at 1150 . If the liked list of operations is not complete at 1160 , the address of the subsequent operation is fetched. Otherwise, the process in complete. In one embodiment, the fetching and execution indicated by 1130 , 1140 and 1150 are performed for each operation in the linked list of operations.
- FIG. 13 is a conceptual example of a linked list of operations according to one embodiment of the invention.
- graphical operations are referred to as “blits” or “BLTs” (Block Transfers).
- BLTs Block Transfers
- the example of FIG. 13 describes five linked blit operations; however, any number of operations can be linked together.
- Blit 1 causes Link List Registers 1300 to be read.
- the registers indicated by the Mask Register (Blt_Mask) as described above are read to execute Blit 1 .
- Blit 1 operates on a 16-bit RGB bitmap; however, other color formats can also be used.
- Blit 2 is executed after Blit 1 is executed. In the example of FIG. 13, Blit 2 accesses data in a different set of registers indicated by the Mask Register.
- Blit 2 includes access to color look up table (CLUT) 1320 .
- Blit 3 is executed using data from Link List Registers 1310 and operates on a 4-bit indexed color bitmap.
- Blit 4 and Blit 5 are executed using data from Link List Registers 1310 or a different set of registers as indicated by the Mask Register. In the example of FIG. 13, Blit 4 and Blit 5 operates on and 8-bit indexed bitmap.
- graphical data is converted to RGB format prior to operations being performed and two alpha blending modes are supported.
- 8-bit alpha blending mode values between 128 (opaque) and 0 (transparent) are used to provide alpha blending.
- 4-bit alpha blending mode values between 15 (opaque) and 0 (transparent) are used to provide alpha blending.
- Alpha blending can be applied to each color component of each pixel or on a pixel-by-pixel basis.
- FIG. 14 is a block diagram of components to provide programmable alpha blending according to one embodiment of the invention.
- the components of FIG. 13 provide alpha blending between source and destination pixel maps.
- both 8-bit and 4-bit alpha blending is supported.
- other alpha blending modes can be supported.
- Source 0 is represented by an alpha value ( ⁇ 0 ) and an RGB value (RGB 0 ).
- Source 1 is represented by ⁇
- RGB 1 and Destination is represented by ⁇ D and RGB D .
- Source 0 , Source 1 and Destination have been converted to RGB format prior to the boolean operation described with respect to FIG. 14 .
- Alpha operator 1400 receives ⁇ 0 , ⁇ D and ⁇ D .
- Alpha operator 1400 is programmable by a control circuit or other device (not shown in FIG. 14) to select one of the alpha values for use in generating an output pixel.
- alpha operator 1400 is a multiplexor; however, other devices can be used.
- alpha operator 1400 generates an output alpha value based on the input alpha values that is not necessarily equal to one of the input alpha values.
- Alpha operator 1400 generates ⁇ S as an output alpha value.
- Boolean operator 1420 receives RGB 0 and RGB 1 as input values. Boolean operator 1420 performs one or more boolean operations on RGB 0 and RGB 1 to generate RGB S . The output values from alpha operator 1400 and boolean operator 1420 , ⁇ S and RGB S , respectively are combined to provide an input to blending unit 1440 .
- blending unit 1440 receives ⁇ S RGB S and ⁇ D RGB D as input values and performs a blending operation. For example, if ⁇ S RGB S overlays ⁇ D RGB D , blending unit 1440 generates an output pixel that has an appropriate blending of the source and destination pixels.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Computer Hardware Design (AREA)
- Image Processing (AREA)
Abstract
Description
TABLE 1 |
Color Format Conversions |
Destination |
8-bit | 4-bit | 8-bit | ||||||
alpha, | 8-bit | 15-bit | 16-bit | 24-bit | alpha, | alpha, | ||
8-bit | 8-bit | RGB | RGB | RGB | RGB | 12-bit | 24-bit | |
Source | index | index | (3:3:2) | (5:5:5) | (5:6:5) | (8:8:8) | RGB | RGB |
1-bit index | X | X | X | X | X | X | X | |
4-bit index | X | X | X | X | X | X | X | |
8-bit index | X | X | X | X | X | X | X | |
8-bit alpha, 8- | X | X | X | X | X | X | ||
bit index | ||||||||
8-bit RGB | X | X | X | X | X | |||
15-bit RGB | X | X | X | X | X | |||
16-bit RGB | X | X | X | X | X | |||
24-bit RGB | X | |||||||
4-bit alpha | X | X | X | X | X | X | ||
4-bit alpha, 12- | X | X | X | X | X | |||
bit RGB | ||||||||
8-bit alpha, 24- | X | X | X | X | X | |||
bit RGB | ||||||||
YUV (YCbCr), | X | X | X | X | X | |||
(4:2:2) | ||||||||
TABLE 2 |
Color Expansion |
DESTINA- | ||||
SOURCE | TIONS | R | G | B |
8c | 12c | R2,R1,R0, | G2,G1,G0,G2 | B1,B0,B1,B0 |
R2 | ||||
8c | 16c | R2,R1,R0, | G2,G1,G0, | B1,B0,B1, |
R2,R1 | G2,G1,G0 | B0,B1 | ||
8c | 24c | R2,R1,R0, | G2,G1,G0,G2, | B1,B0,B1,B0, |
R2,R1,R0, | G1,G0,G2,G1 | B1,B0,B1,B0 | ||
R2,R1 | ||||
12c | 16c | R3,R2,R1, | G3,G2,G1,G0, | B3,B2,B1,B0, |
R0,R3 | G3,G2 | B3 | ||
12c | 24c | R3,R2,R1, | G3,G2,G1,G0, | B3,B2,B1,B0 |
R0,R3,R2, | G3,G2,G1,G0 | B3,B2,B1,B0 | ||
R1, |
1i |
8i, 8c, 12c, | Use foreground, background register | |
16c, | ||
4i | ||
8i | Align at LSB of |
|
4i, 8i | 8c, 12c, | Use color look up table |
16c, 24c | ||
TABLE 3 |
Color reduction |
Source | Destination | Operation |
4 | 2 | (x[4] − (x[4] >> 3) + 2 {circumflex over ( )} 0) >> 1 |
4 | 2 | (x[4] − (x[4] >> 2) + 2 {circumflex over ( )} 1) >> 2 |
5 | 3 | (x[5] − (x[5] >> 3) + 2 {circumflex over ( )} 1) >> 2 |
5 | 2 | (x[5] − (x[5] >> 2) + +2 {circumflex over ( )}2) >> 3 |
5 | 4 | (x[5] − (x[5] >> 4) + 2 {circumflex over ( )} 0) >> 1 |
6 | 3 | (x[6] − (x[6] >> 3) + 2 {circumflex over ( )} 2) >> 3 |
6 | 4 | (x[6] − (x[6] >> 4) + 2 {circumflex over ( )} 1) >> 2 |
8 | 2 | (x[8] − (x[8] >> 2) + 2 {circumflex over ( )} 5) >> 6 |
8 | 3 | (x[8] − (x[8] >> 3) + 2 {circumflex over ( )} 4) >> 5 |
8 | 4 | (x[8] − (x[8] >> 4) + 2 {circumflex over ( )} 3) >> 4 |
8 | 5 | (x[8] − (x[8] >> 5) + 2 {circumflex over ( )} 2) >> 3 |
8 | 6 | (x[8] − (x[8] >> 6) + 2 {circumflex over ( )} 1) >> 2 |
TABLE 4 |
Mask Register. |
| Description | |
0 | Command register. When the command register is written, the block | |
transfer result is shared, unless suspended. | ||
1 | Program |
|
2 | Interrupt |
|
3 | |
|
4 | Destination |
|
5 | |
|
6 | |
|
7 | Alpha Parameter Configuration Register | |
8 | Background Register for |
|
9 | Foreground or Global Color Register for |
|
10 | Background Register for |
|
11 | Foreground or Global Color Register for |
|
12 | Height and Width register | |
13 | Pitch for Destination | |
14 | Pitch for |
|
15 | Pitch for |
|
16 | Destination Right Address Register | |
17 | Destination Left Address Register | |
18 | |
|
19 | |
|
20 | |
|
21 | |
|
22 | Link Address Register | |
23 | Configure the Program Register | |
Claims (15)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US09/263,161 US6466220B1 (en) | 1999-03-05 | 1999-03-05 | Graphics engine architecture |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US09/263,161 US6466220B1 (en) | 1999-03-05 | 1999-03-05 | Graphics engine architecture |
Publications (1)
Publication Number | Publication Date |
---|---|
US6466220B1 true US6466220B1 (en) | 2002-10-15 |
Family
ID=23000652
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US09/263,161 Expired - Lifetime US6466220B1 (en) | 1999-03-05 | 1999-03-05 | Graphics engine architecture |
Country Status (1)
Country | Link |
---|---|
US (1) | US6466220B1 (en) |
Cited By (40)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030151679A1 (en) * | 2002-02-08 | 2003-08-14 | Amerson Frederic C. | System and method for using multiple images in a digital image capture device |
US6614447B1 (en) * | 2000-10-04 | 2003-09-02 | Terarecon, Inc. | Method and apparatus for correcting opacity values in a rendering pipeline |
US20030164842A1 (en) * | 2002-03-04 | 2003-09-04 | Oberoi Ranjit S. | Slice blend extension for accumulation buffering |
US20040005099A1 (en) * | 2002-07-03 | 2004-01-08 | Jin-Ming Gu | Apparatus and method for alpha blending of digital images |
US6707507B2 (en) * | 2000-03-06 | 2004-03-16 | Lg Electronics Inc. | Apparatus and method for on screen display |
US20040056864A1 (en) * | 1998-11-09 | 2004-03-25 | Broadcom Corporation | Video and graphics system with MPEG specific data transfer commands |
US6753872B2 (en) * | 2000-01-14 | 2004-06-22 | Renesas Technology Corp. | Rendering processing apparatus requiring less storage capacity for memory and method therefor |
US20040224777A1 (en) * | 2001-09-28 | 2004-11-11 | Shuffle Master, Inc. | Card shuffler with reading capability integrated into multiplayer automated gaming table |
US20050066278A1 (en) * | 2003-09-19 | 2005-03-24 | Sloo David Hendler | Full scale video with overlaid graphical user interface and scaled image |
US20050122335A1 (en) * | 1998-11-09 | 2005-06-09 | Broadcom Corporation | Video, audio and graphics decode, composite and display system |
US20050164762A1 (en) * | 2004-01-26 | 2005-07-28 | Shuffle Master, Inc. | Automated multiplayer game table with unique image feed of dealer |
US20050164759A1 (en) * | 2004-01-26 | 2005-07-28 | Shuffle Master, Inc. | Electronic gaming machine with architecture supporting a virtual dealer and virtual cards |
US20050231526A1 (en) * | 1998-11-09 | 2005-10-20 | Broadcom Corporation | Graphics display system with anti-aliased text and graphics feature |
US20060044389A1 (en) * | 2004-08-27 | 2006-03-02 | Chai Sek M | Interface method and apparatus for video imaging device |
US20060050088A1 (en) * | 2003-02-28 | 2006-03-09 | Hiroshi Yahata | Recording medium, reproduction apparatus, recording method, program, and reproduction method |
US20070046687A1 (en) * | 2005-08-23 | 2007-03-01 | Atousa Soroushi | Method and Apparatus for Overlaying Reduced Color Resolution Images |
DE102006037507A1 (en) * | 2005-10-28 | 2007-05-03 | Agilent Technologies, Inc. (n.d.Ges.d.Staates Delaware), Palo Alto | Apply a variable opacity (image alpha) to power and probability distributions superimposed on cartographic displays |
US20070155462A1 (en) * | 2003-07-22 | 2007-07-05 | O'halloran Terry | Side bets in casino wagering "war" game |
US20070296713A1 (en) * | 2006-06-09 | 2007-12-27 | Samsung Electronics Co., Ltd., | Method and apparatus for driving a plurality of display devices using a single video stream |
US20080117198A1 (en) * | 2006-11-22 | 2008-05-22 | Nec Electronics Corporation | Display device and controller driver for improved FRC technique |
US20080165200A1 (en) * | 2007-01-05 | 2008-07-10 | Raymond Chow | Hardware Background Tile Generation |
CN100459678C (en) * | 2003-10-01 | 2009-02-04 | 因佛卡斯公司 | Illumination systems employing corrective optics for use in reduced tendue color video projection systems |
US7505046B1 (en) * | 2000-05-02 | 2009-03-17 | Adobe Systems Incorporated | Preserving opaque-like rendering in transparent 2D graphics using knockout groups |
US20090249393A1 (en) * | 2005-08-04 | 2009-10-01 | Nds Limited | Advanced Digital TV System |
US7729598B2 (en) | 2003-01-31 | 2010-06-01 | Panasonic Corporation | Recording medium, reproduction apparatus, recording method, program, and reproduction method |
US20110044662A1 (en) * | 2002-11-15 | 2011-02-24 | Thomson Licensing S.A. | Method and apparatus for composition of subtitles |
WO2011085029A1 (en) * | 2010-01-11 | 2011-07-14 | Apple Inc. | Parameter fifo |
US7991049B2 (en) | 1998-11-09 | 2011-08-02 | Broadcom Corporation | Video and graphics system with video scaling |
US8063916B2 (en) | 2003-10-22 | 2011-11-22 | Broadcom Corporation | Graphics layer reduction for video composition |
US20110292060A1 (en) * | 2010-06-01 | 2011-12-01 | Kno, Inc. | Frame buffer sizing to optimize the performance of on screen graphics in a digital electronic device |
US8199154B2 (en) | 1998-11-09 | 2012-06-12 | Broadcom Corporation | Low resolution graphics mode support using window descriptors |
WO2012161848A2 (en) | 2011-03-15 | 2012-11-29 | Silicon Image, Inc. | Conversion of multimedia data streams for use by connected devices |
US8475252B2 (en) | 2007-05-30 | 2013-07-02 | Shfl Entertainment, Inc. | Multi-player games with individual player decks |
CN103329194A (en) * | 2011-01-25 | 2013-09-25 | 晶像股份有限公司 | Conversion and processing of deep color video in a single clock domain |
US8724029B2 (en) * | 2011-05-26 | 2014-05-13 | Adobe Systems Incorporated | Accelerating video from an arbitrary graphical layer |
US9161006B1 (en) * | 2014-12-05 | 2015-10-13 | Kamcord, Inc. | Systems and methods for efficient screen capture |
US9462221B2 (en) | 2002-11-15 | 2016-10-04 | Thomson Licensing | Method and apparatus for composition of subtitles |
US20170177046A1 (en) * | 2015-12-16 | 2017-06-22 | Intel Corporation | Controlling Telemetry Data Communication In A Processor |
US11488349B2 (en) * | 2019-06-28 | 2022-11-01 | Ati Technologies Ulc | Method and apparatus for alpha blending images from different color formats |
US11503310B2 (en) * | 2018-10-31 | 2022-11-15 | Ati Technologies Ulc | Method and apparatus for an HDR hardware processor inline to hardware encoder and decoder |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5559954A (en) * | 1993-02-24 | 1996-09-24 | Intel Corporation | Method & apparatus for displaying pixels from a multi-format frame buffer |
US5699067A (en) * | 1996-06-28 | 1997-12-16 | Hughes Aircraft Company | Radar plot display with low CPU loading |
US5742797A (en) * | 1995-08-11 | 1998-04-21 | International Business Machines Corporation | Dynamic off-screen display memory manager |
US5745119A (en) * | 1995-12-13 | 1998-04-28 | Microsoft Corporation | Color data conversion using real and virtual address spaces |
US5751979A (en) * | 1995-05-31 | 1998-05-12 | Unisys Corporation | Video hardware for protected, multiprocessing systems |
US5936641A (en) * | 1997-06-27 | 1999-08-10 | Object Technology Licensing Corp | Graphics hardware acceleration method, computer program, and system |
-
1999
- 1999-03-05 US US09/263,161 patent/US6466220B1/en not_active Expired - Lifetime
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5559954A (en) * | 1993-02-24 | 1996-09-24 | Intel Corporation | Method & apparatus for displaying pixels from a multi-format frame buffer |
US5751979A (en) * | 1995-05-31 | 1998-05-12 | Unisys Corporation | Video hardware for protected, multiprocessing systems |
US5742797A (en) * | 1995-08-11 | 1998-04-21 | International Business Machines Corporation | Dynamic off-screen display memory manager |
US5745119A (en) * | 1995-12-13 | 1998-04-28 | Microsoft Corporation | Color data conversion using real and virtual address spaces |
US5699067A (en) * | 1996-06-28 | 1997-12-16 | Hughes Aircraft Company | Radar plot display with low CPU loading |
US5936641A (en) * | 1997-06-27 | 1999-08-10 | Object Technology Licensing Corp | Graphics hardware acceleration method, computer program, and system |
Cited By (99)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8493415B2 (en) | 1998-11-09 | 2013-07-23 | Broadcom Corporation | Graphics display system with video scaler |
US20050122335A1 (en) * | 1998-11-09 | 2005-06-09 | Broadcom Corporation | Video, audio and graphics decode, composite and display system |
US9077997B2 (en) | 1998-11-09 | 2015-07-07 | Broadcom Corporation | Graphics display system with unified memory architecture |
US7911483B1 (en) | 1998-11-09 | 2011-03-22 | Broadcom Corporation | Graphics display system with window soft horizontal scrolling mechanism |
US20040056864A1 (en) * | 1998-11-09 | 2004-03-25 | Broadcom Corporation | Video and graphics system with MPEG specific data transfer commands |
US7991049B2 (en) | 1998-11-09 | 2011-08-02 | Broadcom Corporation | Video and graphics system with video scaling |
US8199154B2 (en) | 1998-11-09 | 2012-06-12 | Broadcom Corporation | Low resolution graphics mode support using window descriptors |
US7920151B2 (en) | 1998-11-09 | 2011-04-05 | Broadcom Corporation | Graphics display system with video scaler |
US20050231526A1 (en) * | 1998-11-09 | 2005-10-20 | Broadcom Corporation | Graphics display system with anti-aliased text and graphics feature |
US8848792B2 (en) | 1998-11-09 | 2014-09-30 | Broadcom Corporation | Video and graphics system with video scaling |
US9575665B2 (en) | 1998-11-09 | 2017-02-21 | Broadcom Corporation | Graphics display system with unified memory architecture |
US7893950B2 (en) | 1999-12-22 | 2011-02-22 | Adobe Systems Incorporated | Color compositing using transparency groups |
US6753872B2 (en) * | 2000-01-14 | 2004-06-22 | Renesas Technology Corp. | Rendering processing apparatus requiring less storage capacity for memory and method therefor |
US6707507B2 (en) * | 2000-03-06 | 2004-03-16 | Lg Electronics Inc. | Apparatus and method for on screen display |
US7505046B1 (en) * | 2000-05-02 | 2009-03-17 | Adobe Systems Incorporated | Preserving opaque-like rendering in transparent 2D graphics using knockout groups |
US6614447B1 (en) * | 2000-10-04 | 2003-09-02 | Terarecon, Inc. | Method and apparatus for correcting opacity values in a rendering pipeline |
US7661676B2 (en) | 2001-09-28 | 2010-02-16 | Shuffle Master, Incorporated | Card shuffler with reading capability integrated into multiplayer automated gaming table |
US20040224777A1 (en) * | 2001-09-28 | 2004-11-11 | Shuffle Master, Inc. | Card shuffler with reading capability integrated into multiplayer automated gaming table |
US7084910B2 (en) * | 2002-02-08 | 2006-08-01 | Hewlett-Packard Development Company, L.P. | System and method for using multiple images in a digital image capture device |
US20030151679A1 (en) * | 2002-02-08 | 2003-08-14 | Amerson Frederic C. | System and method for using multiple images in a digital image capture device |
US20030164842A1 (en) * | 2002-03-04 | 2003-09-04 | Oberoi Ranjit S. | Slice blend extension for accumulation buffering |
US7095906B2 (en) * | 2002-07-03 | 2006-08-22 | Via Technologies, Inc. | Apparatus and method for alpha blending of digital images |
US20040005099A1 (en) * | 2002-07-03 | 2004-01-08 | Jin-Ming Gu | Apparatus and method for alpha blending of digital images |
US20120212670A1 (en) * | 2002-11-15 | 2012-08-23 | Thomson Licensing S.A. | Method and apparatus for composition of subtitles |
US9462221B2 (en) | 2002-11-15 | 2016-10-04 | Thomson Licensing | Method and apparatus for composition of subtitles |
US9595293B2 (en) * | 2002-11-15 | 2017-03-14 | Thomson Licensing | Method and apparatus for composition of subtitles |
US9635306B2 (en) | 2002-11-15 | 2017-04-25 | Thomson Licensing | Method and apparatus for composition of subtitles |
US9503678B2 (en) | 2002-11-15 | 2016-11-22 | Thomson Licensing | Method and apparatus for composition of subtitles |
US9749576B2 (en) | 2002-11-15 | 2017-08-29 | Thomson Licensing | Method and apparatus for composition of subtitles |
US20120219267A1 (en) * | 2002-11-15 | 2012-08-30 | Thomson Licensing S.A. | Method and apparatus for composition of subtitles |
US20120219266A1 (en) * | 2002-11-15 | 2012-08-30 | Thomson Licensing S.A. | Method and apparatus for composition of subtitles |
US8537282B2 (en) * | 2002-11-15 | 2013-09-17 | Thomson Licensing | Method and apparatus for composition of subtitles |
US8737810B2 (en) | 2002-11-15 | 2014-05-27 | Thomson Licensing | Method and apparatus for cropping of subtitle elements |
US20160211000A9 (en) * | 2002-11-15 | 2016-07-21 | Thomson Licensing | Method and apparatus for composition of subtitles |
US8363163B2 (en) * | 2002-11-15 | 2013-01-29 | Thomson Licensing | Method and apparatus for composition of subtitles |
US8373800B2 (en) * | 2002-11-15 | 2013-02-12 | Thomson Licensing | Method and apparatus for composition of subtitles |
US8531609B2 (en) * | 2002-11-15 | 2013-09-10 | Thomson Licensing | Method and apparatus for composition of subtitles |
US8432493B2 (en) * | 2002-11-15 | 2013-04-30 | Thomson Licensing | Method and apparatus for composition of subtitles |
US20110044662A1 (en) * | 2002-11-15 | 2011-02-24 | Thomson Licensing S.A. | Method and apparatus for composition of subtitles |
US8463107B2 (en) | 2003-01-31 | 2013-06-11 | Panasonic Corporation | Recording medium, reproduction apparatus, recording method, program, and reproduction method |
US7729598B2 (en) | 2003-01-31 | 2010-06-01 | Panasonic Corporation | Recording medium, reproduction apparatus, recording method, program, and reproduction method |
US7546024B2 (en) | 2003-02-28 | 2009-06-09 | Panasonic Corporation | Recording medium, reproduction apparatus, recording method, program, and reproduction method |
US20060050088A1 (en) * | 2003-02-28 | 2006-03-09 | Hiroshi Yahata | Recording medium, reproduction apparatus, recording method, program, and reproduction method |
US7962012B2 (en) | 2003-02-28 | 2011-06-14 | Panasonic Corporation | Recording medium, reproduction apparatus, recording method, program and reproduction method |
US7814422B2 (en) * | 2003-02-28 | 2010-10-12 | Panasonic Corporation | Reproduction apparatus, reproduction method and recording method |
US8676040B2 (en) | 2003-02-28 | 2014-03-18 | Panasonic Corporation | Recording medium, reproduction apparatus, and recording method |
US20070172202A1 (en) * | 2003-02-28 | 2007-07-26 | Hiroshi Yahata | Recording medium, reproduction apparatus, recording, method, program, and reproduction method |
US20080238940A1 (en) * | 2003-02-28 | 2008-10-02 | Hiroshi Yahata | Recording medium, reproduction apparatus, recording method, program and reproduction method |
US20070155462A1 (en) * | 2003-07-22 | 2007-07-05 | O'halloran Terry | Side bets in casino wagering "war" game |
US7907152B2 (en) | 2003-09-19 | 2011-03-15 | Microsoft Corporation | Full scale video with overlaid graphical user interface and scaled image |
US20050253869A1 (en) * | 2003-09-19 | 2005-11-17 | Microsoft Corporation | Full scale video with overlaid graphical user interface and scaled image |
US20050066278A1 (en) * | 2003-09-19 | 2005-03-24 | Sloo David Hendler | Full scale video with overlaid graphical user interface and scaled image |
US7705860B2 (en) | 2003-09-19 | 2010-04-27 | Microsoft Corporation | Full scale video with overlaid graphical user interface and scaled image |
US7133051B2 (en) * | 2003-09-19 | 2006-11-07 | Microsoft Corporation | Full scale video with overlaid graphical user interface and scaled image |
CN100459678C (en) * | 2003-10-01 | 2009-02-04 | 因佛卡斯公司 | Illumination systems employing corrective optics for use in reduced tendue color video projection systems |
US8063916B2 (en) | 2003-10-22 | 2011-11-22 | Broadcom Corporation | Graphics layer reduction for video composition |
US8272958B2 (en) | 2004-01-26 | 2012-09-25 | Shuffle Master, Inc. | Automated multiplayer game table with unique image feed of dealer |
US20050164759A1 (en) * | 2004-01-26 | 2005-07-28 | Shuffle Master, Inc. | Electronic gaming machine with architecture supporting a virtual dealer and virtual cards |
US20050164762A1 (en) * | 2004-01-26 | 2005-07-28 | Shuffle Master, Inc. | Automated multiplayer game table with unique image feed of dealer |
WO2006026126A3 (en) * | 2004-08-27 | 2006-05-26 | Motorola Inc | Interface method and apparatus for video imaging device |
WO2006026126A2 (en) * | 2004-08-27 | 2006-03-09 | Motorola, Inc. | Interface method and apparatus for video imaging device |
US20060044389A1 (en) * | 2004-08-27 | 2006-03-02 | Chai Sek M | Interface method and apparatus for video imaging device |
US8089510B2 (en) | 2004-08-27 | 2012-01-03 | Motorola Mobility, Inc. | Interface method and apparatus for video imaging device |
US8069466B2 (en) | 2005-08-04 | 2011-11-29 | Nds Limited | Advanced digital TV system |
US20090249393A1 (en) * | 2005-08-04 | 2009-10-01 | Nds Limited | Advanced Digital TV System |
US7557817B2 (en) * | 2005-08-23 | 2009-07-07 | Seiko Epson Corporation | Method and apparatus for overlaying reduced color resolution images |
US20070046687A1 (en) * | 2005-08-23 | 2007-03-01 | Atousa Soroushi | Method and Apparatus for Overlaying Reduced Color Resolution Images |
DE102006037507A1 (en) * | 2005-10-28 | 2007-05-03 | Agilent Technologies, Inc. (n.d.Ges.d.Staates Delaware), Palo Alto | Apply a variable opacity (image alpha) to power and probability distributions superimposed on cartographic displays |
US20070296713A1 (en) * | 2006-06-09 | 2007-12-27 | Samsung Electronics Co., Ltd., | Method and apparatus for driving a plurality of display devices using a single video stream |
US8228319B2 (en) * | 2006-11-22 | 2012-07-24 | Renesas Electronics Corporation | Display device and controller driver for improved FRC technique |
US20080117198A1 (en) * | 2006-11-22 | 2008-05-22 | Nec Electronics Corporation | Display device and controller driver for improved FRC technique |
US20080165200A1 (en) * | 2007-01-05 | 2008-07-10 | Raymond Chow | Hardware Background Tile Generation |
US8475252B2 (en) | 2007-05-30 | 2013-07-02 | Shfl Entertainment, Inc. | Multi-player games with individual player decks |
WO2011085029A1 (en) * | 2010-01-11 | 2011-07-14 | Apple Inc. | Parameter fifo |
US8749568B2 (en) | 2010-01-11 | 2014-06-10 | Apple Inc. | Parameter FIFO |
US9262798B2 (en) | 2010-01-11 | 2016-02-16 | Apple Inc. | Parameter FIFO |
US20110169848A1 (en) * | 2010-01-11 | 2011-07-14 | Bratt Joseph P | Parameter FIFO |
US20110292060A1 (en) * | 2010-06-01 | 2011-12-01 | Kno, Inc. | Frame buffer sizing to optimize the performance of on screen graphics in a digital electronic device |
US9141134B2 (en) | 2010-06-01 | 2015-09-22 | Intel Corporation | Utilization of temporal and spatial parameters to enhance the writing capability of an electronic device |
US9996227B2 (en) | 2010-06-01 | 2018-06-12 | Intel Corporation | Apparatus and method for digital content navigation |
US9037991B2 (en) | 2010-06-01 | 2015-05-19 | Intel Corporation | Apparatus and method for digital content navigation |
EP2668649A2 (en) * | 2011-01-25 | 2013-12-04 | Silicon Image, Inc. | Conversion and processing of deep color video in a single clock domain |
CN103329194B (en) * | 2011-01-25 | 2017-05-31 | 美国莱迪思半导体公司 | The conversion and treatment of deep color video in single clock zone |
EP2668649A4 (en) * | 2011-01-25 | 2014-06-25 | Silicon Image Inc | Conversion and processing of deep color video in a single clock domain |
CN103329194A (en) * | 2011-01-25 | 2013-09-25 | 晶像股份有限公司 | Conversion and processing of deep color video in a single clock domain |
US9412330B2 (en) | 2011-03-15 | 2016-08-09 | Lattice Semiconductor Corporation | Conversion of multimedia data streams for use by connected devices |
CN103430566B (en) * | 2011-03-15 | 2018-02-02 | 美国莱迪思半导体公司 | Convert multimedia data flow uses for connected device |
EP2687009A2 (en) * | 2011-03-15 | 2014-01-22 | Silicon Image, Inc. | Conversion of multimedia data streams for use by connected devices |
CN103430566A (en) * | 2011-03-15 | 2013-12-04 | 晶像股份有限公司 | Conversion of multimedia data streams for use by connected devices |
EP2687009A4 (en) * | 2011-03-15 | 2014-08-27 | Silicon Image Inc | Conversion of multimedia data streams for use by connected devices |
WO2012161848A2 (en) | 2011-03-15 | 2012-11-29 | Silicon Image, Inc. | Conversion of multimedia data streams for use by connected devices |
US8724029B2 (en) * | 2011-05-26 | 2014-05-13 | Adobe Systems Incorporated | Accelerating video from an arbitrary graphical layer |
US9161006B1 (en) * | 2014-12-05 | 2015-10-13 | Kamcord, Inc. | Systems and methods for efficient screen capture |
US20170177046A1 (en) * | 2015-12-16 | 2017-06-22 | Intel Corporation | Controlling Telemetry Data Communication In A Processor |
US9910470B2 (en) * | 2015-12-16 | 2018-03-06 | Intel Corporation | Controlling telemetry data communication in a processor |
TWI720047B (en) * | 2015-12-16 | 2021-03-01 | 美商英特爾公司 | Controlling telemetry data communication in a processor |
US11503310B2 (en) * | 2018-10-31 | 2022-11-15 | Ati Technologies Ulc | Method and apparatus for an HDR hardware processor inline to hardware encoder and decoder |
US11488349B2 (en) * | 2019-06-28 | 2022-11-01 | Ati Technologies Ulc | Method and apparatus for alpha blending images from different color formats |
US12033273B2 (en) | 2019-06-28 | 2024-07-09 | Ati Technologies Ulc | Method and apparatus for alpha blending images from different color formats |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US6466220B1 (en) | Graphics engine architecture | |
JP3442252B2 (en) | Hardware to support YUV data format conversion for software MPEG decoder | |
US5577203A (en) | Video processing methods | |
US5859651A (en) | Method and apparatus for block data transfer to reduce on-chip storage for interpolative video resizing | |
US5912710A (en) | System and method for controlling a display of graphics data pixels on a video monitor having a different display aspect ratio than the pixel aspect ratio | |
US6452641B1 (en) | Method and apparatus for providing and on-screen display with variable resolution capability | |
US8698840B2 (en) | Method and apparatus for processing video and graphics data to create a composite output image having independent and separate layers of video and graphics display planes | |
US6538656B1 (en) | Video and graphics system with a data transport processor | |
US5742272A (en) | Accelerated full screen video playback | |
JP2656737B2 (en) | Data processing device for processing video information | |
JP3268779B2 (en) | Variable pixel depth and format for video windows | |
US7672372B1 (en) | Method and system for data management in a video decoder | |
US5896140A (en) | Method and apparatus for simultaneously displaying graphics and video data on a computer display | |
JPH06303423A (en) | Coupling system for composite mode-composite signal source picture signal | |
US5345554A (en) | Visual frame buffer architecture | |
US5864345A (en) | Table-based color conversion to different RGB16 formats | |
CA2064070A1 (en) | Enhanced digital video engine | |
JP3472667B2 (en) | Video data processing device and video data display device | |
US5977947A (en) | Method and apparatus for resizing block ordered video image frames with reduced on-chip cache | |
US7623140B1 (en) | Method and apparatus for processing video and graphics data to create a composite output image having independent and separate layers of video and graphics | |
US8749566B2 (en) | System and method for an optimized on-the-fly table creation algorithm | |
EP0840276A2 (en) | Window processing in an on screen display system | |
US7414632B1 (en) | Multi-pass 4:2:0 subpicture blending | |
US7203236B2 (en) | Moving picture reproducing device and method of reproducing a moving picture | |
US20080192146A1 (en) | Video Signal Processor |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: TERALOGIC, INC., CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:CESANA, JOSEPH F.;TRAJAMAR, PETER;WANG, EDWARD;AND OTHERS;REEL/FRAME:009985/0789;SIGNING DATES FROM 19990517 TO 19990524 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
AS | Assignment |
Owner name: ZORAN CORPORATION, CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:TERALOGIC INC. BY THEIR DULY AUTHORIZED OFFICER, CHRIS DENTEN;REEL/FRAME:015394/0302 Effective date: 20041106 |
|
FPAY | Fee payment |
Year of fee payment: 4 |
|
FEPP | Fee payment procedure |
Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
AS | Assignment |
Owner name: OAK TECHNOLOGY, INC., CALIFORNIA Free format text: AGREEMENT AND PLAN OF MERGER;ASSIGNOR:TERALOGIC, INC.;REEL/FRAME:023438/0116 Effective date: 20021007 Owner name: OAK TECHNOLOGY, INC.,CALIFORNIA Free format text: AGREEMENT AND PLAN OF MERGER;ASSIGNOR:TERALOGIC, INC.;REEL/FRAME:023438/0116 Effective date: 20021007 |
|
AS | Assignment |
Owner name: TERALOGIC, INC., CALIFORNIA Free format text: MERGER;ASSIGNOR:OPTIC ACQUISITION CORP.;REEL/FRAME:023438/0579 Effective date: 20021025 Owner name: TERALOGIC, INC.,CALIFORNIA Free format text: MERGER;ASSIGNOR:OPTIC ACQUISITION CORP.;REEL/FRAME:023438/0579 Effective date: 20021025 |
|
AS | Assignment |
Owner name: ZINC ACQUISITION CORPORATION, CALIFORNIA Free format text: MERGER;ASSIGNOR:OAK TECHNOLOGY, INC.;REEL/FRAME:023449/0717 Effective date: 20030811 Owner name: ZINC ACQUISITION CORPORATION,CALIFORNIA Free format text: MERGER;ASSIGNOR:OAK TECHNOLOGY, INC.;REEL/FRAME:023449/0717 Effective date: 20030811 |
|
AS | Assignment |
Owner name: ZORAN CORPORATION, CALIFORNIA Free format text: AGREEMENT AND PLAN OF REORGANIZATION;ASSIGNOR:ZINC ACQUISITION CORPORATION;REEL/FRAME:023627/0892 Effective date: 20030504 Owner name: ZORAN CORPORATION,CALIFORNIA Free format text: AGREEMENT AND PLAN OF REORGANIZATION;ASSIGNOR:ZINC ACQUISITION CORPORATION;REEL/FRAME:023627/0892 Effective date: 20030504 |
|
FPAY | Fee payment |
Year of fee payment: 8 |
|
AS | Assignment |
Owner name: CSR TECHNOLOGY INC., CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:ZORAN CORPORATION;REEL/FRAME:027550/0695 Effective date: 20120101 |
|
FPAY | Fee payment |
Year of fee payment: 12 |
|
AS | Assignment |
Owner name: CSR TECHNOLOGY INC., CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:ZORAN CORPORATION;REEL/FRAME:036642/0395 Effective date: 20150915 |
|
FEPP | Fee payment procedure |
Free format text: PAYER NUMBER DE-ASSIGNED (ORIGINAL EVENT CODE: RMPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
AS | Assignment |
Owner name: TERALOGIC GROUP, INC., CALIFORNIA Free format text: MERGER AND CHANGE OF NAME;ASSIGNORS:OPTIC ACQUISITION CORP.;TERALOGIC, INC.;REEL/FRAME:038136/0479 Effective date: 20021025 |
|
AS | Assignment |
Owner name: ZORAN CORPORATION, CALIFORNIA Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE ASSIGNOR NAME PREVIOUSLY RECORDED AT REEL: 015394 FRAME: 0302. ASSIGNOR(S) HEREBY CONFIRMS THE ASSIGNMENT;ASSIGNOR:TERALOGIC GROUP, INC.;REEL/FRAME:038363/0675 Effective date: 20041106 |