WO1997013220A1 - Graphics image manipulation - Google Patents

Graphics image manipulation Download PDF

Info

Publication number
WO1997013220A1
WO1997013220A1 PCT/IB1996/000966 IB9600966W WO9713220A1 WO 1997013220 A1 WO1997013220 A1 WO 1997013220A1 IB 9600966 W IB9600966 W IB 9600966W WO 9713220 A1 WO9713220 A1 WO 9713220A1
Authority
WO
WIPO (PCT)
Prior art keywords
encoder
image
macroblock
lines
pixel data
Prior art date
Application number
PCT/IB1996/000966
Other languages
French (fr)
Inventor
Richard David Gallery
Octavius John Morris
Edward Stretton Eilley
David Edward Penna
Original Assignee
Philips Electronics N.V.
Philips Norden Ab
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Philips Electronics N.V., Philips Norden Ab filed Critical Philips Electronics N.V.
Priority to EP96929474A priority Critical patent/EP0800689B1/en
Priority to DE69619220T priority patent/DE69619220T2/en
Priority to JP51409397A priority patent/JP3884076B2/en
Publication of WO1997013220A1 publication Critical patent/WO1997013220A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • A63F13/12
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/30Interconnection arrangements between game servers and game devices; Interconnection arrangements between game devices; Interconnection arrangements between game servers
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/50Controlling the output signals based on the game progress
    • A63F13/52Controlling the output signals based on the game progress involving aspects of the displayed game scene
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/527Global motion vector estimation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/16Analogue secrecy systems; Analogue subscription systems
    • H04N7/173Analogue secrecy systems; Analogue subscription systems with two-way working, e.g. subscriber sending a programme selection signal
    • H04N7/17309Transmission or handling of upstream communications
    • H04N7/17318Direct or substantially direct transmission and handling of requests

Definitions

  • the present invention relates to the manipulation and coding of video images and particularly, but not exclusively, to the interactive manipulation and predictive coding of computer graphics images for transmission over a network to one or more remote users.
  • the minimum delay in a decoder is around 80 ms (40 ms minimum buffer delay and around 40 ms decoding time) and operation at this extreme produces low quality pictures, assuming a network interface transfer rate of around 1.5 Mb/s. With higher speeds the quality may be restored or the latency may be reduced, by reducing the buffer delay, but the higher speed capability would produce penalties in terms of the cost and complexity of the equipment required.
  • the commensurate figure for the encoder is around 60 ms, of which around 20 ms is the time to grab one field and 40 ms is the encoding time. Again, at this extreme, the quality would be poor.
  • the delay attributable to the network back channel, passing control messages from user to server is about 15 ms, typically.
  • the minimum round trip delay in a remote game, or any other networked Video-on-Demand (VoD) type of application is about 160 ms.
  • Typical human reaction time is of the order of 80 ms, and a system
  • encoded image generation apparatus comprising a graphics image generator operable to generate pixel images and an encoder coupled thereto and arranged to encode said generated images as a series of macroblocks of pixel data, each macroblock being m pixels wide by n pixels high, where m and n are integers; characterised in that the graphics image generator is configured to generate images as successive areas of pixels, to pass the pixel data for said areas to the encoder and to send a control signal to the encoder as soon as pixel data for n lines of an image have been passed, and the encoder is configured to receive and buffer said pixel data and, on receipt of the control signal, to begin to encode a macroblock line of the buffered pixel data.
  • the image may be rendered as successive blocks or areas of pixel data, with the encoder being triggered as soon as it has been sent enough data to cover the n lines of the macroblock row.
  • the graphics image generator may be configured to generate images as successive lines of pixel data, to pass said lines to the encoder and to send a control signal to the encoder every n lines, with the encoder being configured to receive and buffer said lines and, on receipt of the control signal, to encode the buffered lines of pixel data as a macroblock line.
  • the graphics image generator implements a scan line algorithm in the generation of lines of pixel data, with the generator including a memory for image primitives and means for determining, and storing a list of, those primitives active for a given pixel line.
  • Scan line algorithms are a known technique described, for example, in "Computer Graphics: Principles and Practice” at pp 680-686 by James D Foley et al, second edition, pub. Addison- Wesley Publishing Co Ltd, 1990, ISBN 0-201-12110-7. Their particular application to the present invention is described in greater detail and with reference to exemplary embodiments hereinafter.
  • the encoder is arranged to code the macroblocks of pixel data in terms of a motion vector, performing a comparison search with respect to a current macroblock and the preceding image frame to determine a motion vector for the current macroblock.
  • the graphics image generator may specify to the encoder a global motion vector for an image, with the encoder beginning the comparison search for each macroblock of that image from the global motion vector.
  • the graphics image generator may specify to the encoder a block motion vector for one or a group of macroblocks of an image, with the encoder beginning the comparison search for an identified macroblock, or each of the group of macroblocks, from the block motion vector.
  • an input is preferably provided for user control signals, in response to which signals the graphics image generator changes at least a part of an image in a first frame in one or more successive further image frames.
  • a video signal distribution system comprised of a video server (including image generation and coding apparatus with a user input as described above) together with one or more user terminals connected thereto by a data transmission network.
  • the or each of the user terminals comprises at least a decoder and display device operable to display the generated pixel images, and 5 user operable means to generate and send said user control signals to the server user input.
  • Figure 1 represents a pixel image divided into macroblocks
  • Figure 2 is a block schematic diagram of a server and remote terminal connected over a data network
  • Figure 3 shows the network server of Figure 1 in greater detail; and s Figure 4 is a flow chart representing the application of a scan line algorithm in the server of Figure 3.
  • MPEG 1 coding scheme as defined in ISO 11172, although it will be recognised o that it is applicable to other coding schemes, particularly hybrid-DCT (discrete cosine transformation) based coding schemes, such as the H.261 standard for video-telephony.
  • MPEG and H.261 are digital coding systems conventionally used for storing and compressing natural picture sequences, with the decoders being designed to interpret such encoded data and reproduce the same original 5 sequence of images.
  • the MPEG system defines a grid of macroblocks as shown in Figure 1 where the image 8 represents a rectangular block 9 formed of image primitives L, M and N against a background Q.
  • primitives L, M and N may each be made up of two or more primitives; for example each of the o parallelograms may be formed from a pair of similar-triangle primitives.
  • Each macroblock A1 ,A2,B1 ,B2 etc of the image 8 consists of 16 pixels by 16 lines: these macroblocks are the basic unit of MPEG coding.
  • Three main coded picture types are defined in MPEG, namely intra-pictures, predicted pictures and interpolated pictures; these are generally referred to as I-, P- and B-frames respectively.
  • a macroblock may be coded in one of a number of different ways: two basic macroblock coding modes from the MPEG standard are the "intra mode” and the "motion compensated, not coded modes".
  • intra mode the address of the macroblock describes which macroblock is represented, followed by the macroblock type, and the DCT coefficient data for 0 the pixel values.
  • motion compensated mode the information following the address and type information is a motion vector for the macroblock. The motion vector provides a displacement from the address of the current macroblock in the current picture to a pixel and line address on the previous picture from where it is copied to the current picture.
  • Intra coded macroblocks 5 may be used in predicted pictures to introduce image blocks that are not present in the previous frame.
  • a typical system arrangement for networked video applications consists of a server 10 and one or more remote users 12 communicating with the server over a network 14, as shown in Figure 2.
  • Graphic images are generated by o graphics engine 16 at the server site and these images are passed to an MPEG encoder 18 which codes the images for transmission over the network.
  • the remote user site 12 the user is provided with at least a display 20 (with associated MPEG decoder 22) and an interactive control 24 by use of which control signals affecting the displayed image may be sent to the server 10 via 5 a back channel 26 of the network.
  • a possible arrangement for the remote user for example in a home environment and depending on the networked application, would be a personal computer interfaced to the network and with an MPEG capability.
  • FIG. 3 shows the server in greater detail.
  • the graphics engine 16 o includes a processor 30 which assembles images one line at a time by use of a scan line algorithm (to be described below) from surface data stored in a surface table (ST) memory 32 and an active surface table memory (AST) 38 coupled to the processor via data bus 40.
  • the surface data comprises geometrical shapes (primitives) from which an image is assembled in known manner together with information as to surface detail (texture, colouring, reflectance etc).
  • the surface table memory 32 holds surface data for all primitives of an image; the active surface table memory 38 holds data for only those primitives appearing in a given line of the image.
  • the commands from the remote user on back channel 26 are passed to the processor 30 via a suitable interface 36, which commands require changes to the displayed image. Such changes may include movement of a particular displayed object (sprite) about the screen, or scrolling of a larger image.
  • the processor acts on these commands by updating the surface data memory 32 from a mass storage memory 34 or external source of possible shapes such that memory 32 contains data for each of the surfaces to be assembled to form the complete image.
  • step 101 ADD S.T
  • step 102 INIT A.S.T
  • step 103 UPDATE A.S.T
  • step 104 TOP PIXEL
  • the per-pixel contribution from two or more overlapping surfaces is determined.
  • the image is of a three-dimensional object mapped onto two-dimensional screen space
  • conventional depth buffering and/or hidden surface removal techniques may be employed to determine the extent to which each surface contributes to the pixel.
  • this step may determine a ratio in which each surface contributes to the pixel.
  • step 105 SHADE
  • the pixel colour and/or texture is calculated
  • Step 106 determines whether the pixel is a the end of a line and, if not, the next pixel is selected at step 107 (NEXT P) and the procedure reverts to step 104.
  • step 106 determines that the last pixel of a line has been reached, the o procedure moves to step 108 (LAST ML?) where a line count check is made to see if the number of lines generated is an integer multiple of n, where n is the number of lines per macroblock (16 lines in the case of MPEG). If this step indicates that the current line is not the last of a macroblock, the procedure moves to step 109 (NEXT L) where the next line is selected and then back to 5 step 103 (UPDATE A.S.T) where the active surface table is updated to contain the surface information for those surfaces which contribute to the next line.
  • step 108 determines that the current line is the last of a macroblock
  • step 110 a control signal (FLAG; Figure 3) is sent to the encoder 18 indicating that the encoder now has a complete macroblock line o which may be encoded.
  • the procedure then moves to step 111 (LAST SL?) where a check is made as to whether the current line is the last of an image. If so, the procedure ends or reverts to step 101 for the start of a new 5 image frame: if not, the next line is selected at step 109 and the procedure reverts to step 103.
  • the scan line pixel data from the processor 30 is passed to the encoder stage 18 where it is accumulated in a buffer (not shown).
  • a counter 42 (shown separate from the processor 30 although it may o be a part thereof) counts the number of scan lines sent, and at every nth scan line (where n is the number of lines in a macroblock) it sends the FLAG signal (the macroblock line command) to the encoder.
  • the encoder On receipt of this FLAG signal the encoder reads the accumulated scan lines from its buffer and begins to encode them as a line of macroblocks. With reference to Figure 1 , this means that the encoder will start coding macroblock line A1 , /&, ... A22 as soon as it has received the 16th line of pixel data.
  • the encoder stage 18 can generate only I- and P-frames: B-frames, coded in MPEG with respect to both the preceding and following frames, cannot be produced.
  • the MPEG standard does permit signal coding with I- and P-frames only, this can lead to a reduction in picture quality.
  • the global motion vector of the image to be encoded (corresponding to panning of a "virtual camera") is therefore known in the graphics engine 16 where it is used as an input to the graphical calculations. As shown in Figure 3, this global motion vector is provided as an additional input (GMV) to the encoder stage 18 where it provides assistance to the motion vector search pattern during motion compensation of macroblocks, by specifying a "most likely" direction for starting the search pattern.
  • GMV additional input

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

In an image source for multimedia applications such as networked computer games, a graphics engine (16) generates pixel images one line at a time using a scan-line algorithm and supplies the generated scan lines to an encoder (18) where they are buffered. The encoder codes the buffered pixel data as macroblocks of, for example 16x16 pixels according to MPEG or similar standards. When the graphics engine has sent sufficient scan lines to the encoder, it sends a signal (FLAG) on receipt of which the encoder begins coding the pixel lines as a macroblock line whilst continuing to receive scan lines from the graphics engine (16). To increase encoder efficiency, the graphics engine specifies to the encoder global (GMV) and macroblock (BMV) motion vectors for substantially all or selected ones of the macroblocks of an image respectively. Completed coded macroblock lines may be sent out to a remote user over a data network, with user commands affecting the composition of subsequent frames being received on a back-channel (26) of the network.

Description

DESCRIPTION
GRAPHICS IMAGE MANIPULATION
The present invention relates to the manipulation and coding of video images and particularly, but not exclusively, to the interactive manipulation and predictive coding of computer graphics images for transmission over a network to one or more remote users.
A particular problem with real time image manipulation in multimedia applications such as a network system, where user commands are sent over the network from a terminal to a remote server and images modified in response to those commands are sent back over the network to the user, is the round loop delay. In current MPEG systems the minimum delay in a decoder is around 80 ms (40 ms minimum buffer delay and around 40 ms decoding time) and operation at this extreme produces low quality pictures, assuming a network interface transfer rate of around 1.5 Mb/s. With higher speeds the quality may be restored or the latency may be reduced, by reducing the buffer delay, but the higher speed capability would produce penalties in terms of the cost and complexity of the equipment required. The commensurate figure for the encoder is around 60 ms, of which around 20 ms is the time to grab one field and 40 ms is the encoding time. Again, at this extreme, the quality would be poor. The delay attributable to the network back channel, passing control messages from user to server, is about 15 ms, typically. Given standard MPEG operation, with 1.5 Mb/s digital video transmission rate and accepting minimal quality presentation, the minimum round trip delay in a remote game, or any other networked Video-on-Demand (VoD) type of application, is about 160 ms. Typical human reaction time is of the order of 80 ms, and a system
(whether communicating over a network or directly connected) that responds more slowly than this will appear sluggish to a user. Whilst, in some applications, the remote user can become accustomed to the delay due to system latency and compensate for it, applications requiring rapid user reaction will inevitably suffer. As an example in terms of directly connected systems, the US Federal Aviation Authority has a limit of 150 ms permitted latency in commercial flight simulators.
It is, therefore an object of the present invention to reduce latency in image manipulation systems, particularly where these are incorporated in interactive networked applications.
It is a further object to reduce picture degradation which might otherwise arise due to latency minimisation.
In accordance with the present invention there is provided encoded image generation apparatus comprising a graphics image generator operable to generate pixel images and an encoder coupled thereto and arranged to encode said generated images as a series of macroblocks of pixel data, each macroblock being m pixels wide by n pixels high, where m and n are integers; characterised in that the graphics image generator is configured to generate images as successive areas of pixels, to pass the pixel data for said areas to the encoder and to send a control signal to the encoder as soon as pixel data for n lines of an image have been passed, and the encoder is configured to receive and buffer said pixel data and, on receipt of the control signal, to begin to encode a macroblock line of the buffered pixel data. By beginning the encoding process whilst the graphics image generator (graphics engine) is still assembling the frame, the generator/encoder contribution to the system latency is greatly reduced.
The image may be rendered as successive blocks or areas of pixel data, with the encoder being triggered as soon as it has been sent enough data to cover the n lines of the macroblock row. Alternately, the graphics image generator may be configured to generate images as successive lines of pixel data, to pass said lines to the encoder and to send a control signal to the encoder every n lines, with the encoder being configured to receive and buffer said lines and, on receipt of the control signal, to encode the buffered lines of pixel data as a macroblock line. Suitably, the graphics image generator implements a scan line algorithm in the generation of lines of pixel data, with the generator including a memory for image primitives and means for determining, and storing a list of, those primitives active for a given pixel line. Scan line algorithms are a known technique described, for example, in "Computer Graphics: Principles and Practice" at pp 680-686 by James D Foley et al, second edition, pub. Addison- Wesley Publishing Co Ltd, 1990, ISBN 0-201-12110-7. Their particular application to the present invention is described in greater detail and with reference to exemplary embodiments hereinafter.
Preferably, the encoder is arranged to code the macroblocks of pixel data in terms of a motion vector, performing a comparison search with respect to a current macroblock and the preceding image frame to determine a motion vector for the current macroblock. To enhance the efficiency of the encoder operation and improve image quality, the graphics image generator may specify to the encoder a global motion vector for an image, with the encoder beginning the comparison search for each macroblock of that image from the global motion vector. Furthermore, taking advantage of the relatively simple nature of the graphical images, the graphics image generator may specify to the encoder a block motion vector for one or a group of macroblocks of an image, with the encoder beginning the comparison search for an identified macroblock, or each of the group of macroblocks, from the block motion vector.
To allow for interactive image manipulation, an input is preferably provided for user control signals, in response to which signals the graphics image generator changes at least a part of an image in a first frame in one or more successive further image frames. Also in accordance with present invention there is provided a video signal distribution system comprised of a video server (including image generation and coding apparatus with a user input as described above) together with one or more user terminals connected thereto by a data transmission network. The or each of the user terminals comprises at least a decoder and display device operable to display the generated pixel images, and 5 user operable means to generate and send said user control signals to the server user input.
Preferred embodiments of the present invention will now be described in terms of the MPEG system by way of example only and with reference to the 0 accompanying drawings in which:
Figure 1 represents a pixel image divided into macroblocks; Figure 2 is a block schematic diagram of a server and remote terminal connected over a data network;
Figure 3 shows the network server of Figure 1 in greater detail; and s Figure 4 is a flow chart representing the application of a scan line algorithm in the server of Figure 3.
In the following, the invention will be described principally in terms of the MPEG 1 coding scheme as defined in ISO 11172, although it will be recognised o that it is applicable to other coding schemes, particularly hybrid-DCT (discrete cosine transformation) based coding schemes, such as the H.261 standard for video-telephony. MPEG and H.261 are digital coding systems conventionally used for storing and compressing natural picture sequences, with the decoders being designed to interpret such encoded data and reproduce the same original 5 sequence of images.
The MPEG system defines a grid of macroblocks as shown in Figure 1 where the image 8 represents a rectangular block 9 formed of image primitives L, M and N against a background Q. As will be understood, primitives L, M and N may each be made up of two or more primitives; for example each of the o parallelograms may be formed from a pair of similar-triangle primitives. Each macroblock A1 ,A2,B1 ,B2 etc of the image 8 consists of 16 pixels by 16 lines: these macroblocks are the basic unit of MPEG coding. Three main coded picture types are defined in MPEG, namely intra-pictures, predicted pictures and interpolated pictures; these are generally referred to as I-, P- and B-frames respectively. 5 Within a picture type, a macroblock may be coded in one of a number of different ways: two basic macroblock coding modes from the MPEG standard are the "intra mode" and the "motion compensated, not coded modes". In the intra mode, the address of the macroblock describes which macroblock is represented, followed by the macroblock type, and the DCT coefficient data for 0 the pixel values. In the motion compensated mode, the information following the address and type information is a motion vector for the macroblock. The motion vector provides a displacement from the address of the current macroblock in the current picture to a pixel and line address on the previous picture from where it is copied to the current picture. Intra coded macroblocks 5 may be used in predicted pictures to introduce image blocks that are not present in the previous frame.
A typical system arrangement for networked video applications consists of a server 10 and one or more remote users 12 communicating with the server over a network 14, as shown in Figure 2. Graphic images are generated by o graphics engine 16 at the server site and these images are passed to an MPEG encoder 18 which codes the images for transmission over the network. At the remote user site 12, the user is provided with at least a display 20 (with associated MPEG decoder 22) and an interactive control 24 by use of which control signals affecting the displayed image may be sent to the server 10 via 5 a back channel 26 of the network. A possible arrangement for the remote user, for example in a home environment and depending on the networked application, would be a personal computer interfaced to the network and with an MPEG capability.
Figure 3 shows the server in greater detail. The graphics engine 16 o includes a processor 30 which assembles images one line at a time by use of a scan line algorithm (to be described below) from surface data stored in a surface table (ST) memory 32 and an active surface table memory (AST) 38 coupled to the processor via data bus 40. The surface data comprises geometrical shapes (primitives) from which an image is assembled in known manner together with information as to surface detail (texture, colouring, reflectance etc). The surface table memory 32 holds surface data for all primitives of an image; the active surface table memory 38 holds data for only those primitives appearing in a given line of the image.
The commands from the remote user on back channel 26 are passed to the processor 30 via a suitable interface 36, which commands require changes to the displayed image. Such changes may include movement of a particular displayed object (sprite) about the screen, or scrolling of a larger image. The processor acts on these commands by updating the surface data memory 32 from a mass storage memory 34 or external source of possible shapes such that memory 32 contains data for each of the surfaces to be assembled to form the complete image.
As will be appreciated, where more than one user is connected to the server, some arrangement will be required at the interface 36 for controlling access and/or priority setting of user commands.
The implementation of the scan line algorithm by the graphics engine 16 to generate an image frame is shown in flow chart of Figure 4. The process begins at step 101 (ADD S.T) with the surfaces (image primitives) needed to form the image frame being loaded from external input or storage to the surface table memory 32. Following this, the next step 102 (INIT A.S.T) is to initialise the active surface table memory 38, loading it from surface table memory 32 with those surfaces which contribute to the first pixel line of the image. For the first line, the following step 103 (UPDATE A.S.T) is skipped, and the process enters a per-pixel loop 104-107.
At step 104 (TOP PIXEL), the per-pixel contribution from two or more overlapping surfaces is determined. Where, for example, the image is of a three-dimensional object mapped onto two-dimensional screen space, conventional depth buffering and/or hidden surface removal techniques may be employed to determine the extent to which each surface contributes to the pixel. Where surface effects such as transparency or partial opacity are provided for, this step may determine a ratio in which each surface contributes to the pixel. At step 105 (SHADE), the pixel colour and/or texture is calculated
5 with reference to the active surface table 38 and the surface(s) contributing at step 104. Step 106 (END P?) determines whether the pixel is a the end of a line and, if not, the next pixel is selected at step 107 (NEXT P) and the procedure reverts to step 104.
If step 106 determines that the last pixel of a line has been reached, the o procedure moves to step 108 (LAST ML?) where a line count check is made to see if the number of lines generated is an integer multiple of n, where n is the number of lines per macroblock (16 lines in the case of MPEG). If this step indicates that the current line is not the last of a macroblock, the procedure moves to step 109 (NEXT L) where the next line is selected and then back to 5 step 103 (UPDATE A.S.T) where the active surface table is updated to contain the surface information for those surfaces which contribute to the next line.
If step 108 determines that the current line is the last of a macroblock, then at step 110 (FLAG ENC) a control signal (FLAG; Figure 3) is sent to the encoder 18 indicating that the encoder now has a complete macroblock line o which may be encoded. Assuming that the image is an integer number of macroblocks high, such that the last screen pixel line will also be the last line contributing to a macroblock line, the procedure then moves to step 111 (LAST SL?) where a check is made as to whether the current line is the last of an image. If so, the procedure ends or reverts to step 101 for the start of a new 5 image frame: if not, the next line is selected at step 109 and the procedure reverts to step 103.
Returning to Figure 3, the scan line pixel data from the processor 30 is passed to the encoder stage 18 where it is accumulated in a buffer (not shown). A counter 42 (shown separate from the processor 30 although it may o be a part thereof) counts the number of scan lines sent, and at every nth scan line (where n is the number of lines in a macroblock) it sends the FLAG signal (the macroblock line command) to the encoder. On receipt of this FLAG signal the encoder reads the accumulated scan lines from its buffer and begins to encode them as a line of macroblocks. With reference to Figure 1 , this means that the encoder will start coding macroblock line A1 , /&, ... A22 as soon as it has received the 16th line of pixel data.
With the data being encoded as the frame is assembled, the normal delay is reduced, although the complete MPEG standard cannot be implemented due to the absence of the following frame. In other words, the encoder stage 18 can generate only I- and P-frames: B-frames, coded in MPEG with respect to both the preceding and following frames, cannot be produced. Although the MPEG standard does permit signal coding with I- and P-frames only, this can lead to a reduction in picture quality. In order to minimise such a reduction in quality, use is made of the relatively simple nature of motion vectors in graphics applications to provide assistance to the encoder such as to improve its efficiency of operation.
In graphics under the control of a user, the whole or a large part of the image may be required to move together. The global motion vector of the image to be encoded (corresponding to panning of a "virtual camera") is therefore known in the graphics engine 16 where it is used as an input to the graphical calculations. As shown in Figure 3, this global motion vector is provided as an additional input (GMV) to the encoder stage 18 where it provides assistance to the motion vector search pattern during motion compensation of macroblocks, by specifying a "most likely" direction for starting the search pattern. Due to the relatively simple nature of the graphics image, in comparison with, for example, a still from a video camera, it is possible to analyse the graphics image at an object level, that is to say in terms of features formed from one or more surfaces whose appearance does not change from one frame to the next but whose location within the image may change. In the processor such objects or portions of objects may be identified with one or more macroblocks of a current and a previous frame. Motion vectors for such macroblocks are passed to the encoder stage 18 as a further input (BMV) for use by the encoder as a guide to the motion vector search applied when coding those macroblocks. The result of this is an improvement in the image quality. To further enhance the functionality of the system as a whole, provision is made for switching to conventional MPEG encoder operation, through B- frame on/off toggle 50; under user or application control via the processor 30. This enables conventional encoding, with I-, P- and B-frames, where real time responsiveness to user input is not required, for example when playing back a sequence of frames without enabling user interaction.
From reading the present disclosure, other modifications will be apparent to persons skilled in the art. Such modifications may involve other features which are already known in the design, manufacture and use of interactive image coding equipment and component parts thereof and which may be used instead of or in addition to features already described herein. Although claims have been formulated in this application to particular combinations of features, it should be understood that the scope of the disclosure of the present invention also includes any novel feature or any novel combination of features disclosed herein either explicitly or implicitly or any generalisation thereof, whether or not it relates to the same invention as presently claimed in any claim and whether or not it mitigates any or all of the same technical problems as does the present invention. The applicants hereby give notice that new claims may be formulated to such features and/or combinations of features during the prosecution of the present application or of any further application derived therefrom.

Claims

1. Encoded image generation apparatus comprising a graphics image generator operable to generate pixel images and an encoder coupled
5 thereto and arranged to encode said generated images as a series of macroblocks of pixel data, each macroblock being m pixels wide by n pixels high, where m and n are integers; characterised in that the graphics image generator is configured to generate images as successive areas of pixels, to pass the pixel data for said areas to the encoder and to send a control signal o to the encoder as soon as pixel data for n lines of an image have been passed, and the encoder is configured to receive and buffer said pixel data and, on receipt of the control signal, to begin to encode a macroblock line of the buffered pixel data.
5 2. Apparatus as claimed in Claim 1 , wherein the graphics image generator is configured to generate images as successive lines of pixel data, to pass said lines to the encoder and to send the control signal to the encoder every n lines, and the encoder is configured to receive and buffer said lines and, on receipt of the control signal, to encode the buffered lines of pixel data o as a macroblock line.
3. Apparatus as claimed in Claim 2, wherein the graphics image generator implements a scan line algorithm in the generation of the lines of pixel data, the generator including a memory for image primitives and means 5 for determining, and storing a list of, those primitives active for a given pixel line.
4. Apparatus as claimed in Claim 1 , wherein the encoder is arranged to code the macroblocks of pixel data in terms of a motion vector, performing 0 a comparison search with respect to a current macroblock and the preceding image frame to determine a motion vector for the current macroblock.
5. Apparatus as claimed in Claim 4, wherein the graphics image generator specifies to the encoder a global motion vector for an image and the encoder begins the comparison search for each macroblock of that image from the global motion vector.
6. Apparatus as claimed in Claim 4, wherein the graphics image generator specifies to the encoder a block motion vector for an identified one or a group of macroblocks of an image, and the encoder begins the comparison search for that identified macroblock, or each of the group of macroblocks, from the block motion vector.
7. Apparatus as claimed in any of Claims 1 to 6, further comprising an input for user control signals, in response to which signals the graphics image generator changes at least a part of an image in a first frame in one or more successive further image frames.
8. A video signal distribution system comprised of a video server including apparatus as claimed in Claim 7, together with one or more user terminals connected thereto by a data transmission network, wherein the or each of the user terminals comprises a decoder and display device operable to display the generated pixel images, and user input means operable to generate and send said user control signals.
9. Apparatus as claimed in any one of Claims 1 to 7, wherein the encoder is operable to encode said generated images as l-frames or P-frames according to MPEG standards.
PCT/IB1996/000966 1995-09-29 1996-09-19 Graphics image manipulation WO1997013220A1 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
EP96929474A EP0800689B1 (en) 1995-09-29 1996-09-19 Graphics image manipulation
DE69619220T DE69619220T2 (en) 1995-09-29 1996-09-19 MANIPULATON OF GRAPHIC IMAGES
JP51409397A JP3884076B2 (en) 1995-09-29 1996-09-19 Graphic image processing

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
GB9519921.2 1995-09-29
GBGB9519921.2A GB9519921D0 (en) 1995-09-29 1995-09-29 Graphics image manipulation

Publications (1)

Publication Number Publication Date
WO1997013220A1 true WO1997013220A1 (en) 1997-04-10

Family

ID=10781508

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/IB1996/000966 WO1997013220A1 (en) 1995-09-29 1996-09-19 Graphics image manipulation

Country Status (6)

Country Link
US (1) US5991443A (en)
EP (1) EP0800689B1 (en)
JP (1) JP3884076B2 (en)
DE (1) DE69619220T2 (en)
GB (1) GB9519921D0 (en)
WO (1) WO1997013220A1 (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO1999060498A1 (en) * 1998-05-18 1999-11-25 Aristocrat Leisure Industries Pty. Ltd. Intelligent input/output control system
FR2940703A1 (en) * 2008-12-31 2010-07-02 Cy Play Display modeling method for application on server, involves forming image based on pixels of traces, and transmitting image and encoding information conforming to assembly of modification data to encoder by transmitting unit
FR2940689A1 (en) * 2008-12-31 2010-07-02 Cy Play Method for navigating user of mobile terminal e.g. portable computer, on application, involves detecting movement of positioning device, and modulating movement data in audio format for transmitting movement data to server
FR2940690A1 (en) * 2008-12-31 2010-07-02 Cy Play Mobile terminal i.e. mobile telephone, user navigation method, involves establishing contents to be displayed on terminal for permitting navigation on user interface having size larger than size of screen of terminal
WO2010076436A3 (en) * 2008-12-31 2010-11-25 Cy Play Method for macroblock modeling of the display of a remote terminal by means of layers characterized by a movement vector and transparency data
US9860551B2 (en) * 2011-02-09 2018-01-02 Lg Electronics Inc. Method for encoding and decoding image and device using same

Families Citing this family (37)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7310811B1 (en) * 1997-07-15 2007-12-18 At&T Corp. Interaction modalities for multimedia delivery and presentation
US20090083653A1 (en) * 2000-06-23 2009-03-26 Makoff Jeffrey T System for facilitating remote digital image manipulation services
WO2002037638A2 (en) * 2000-11-01 2002-05-10 Capstone Turbine Corporation Distributed energy network control system and method
US6678329B2 (en) * 2001-01-16 2004-01-13 Hewlett-Packard Development Company, L.P. Client-assisted motion estimation for client-server video communication
US20060033745A1 (en) * 2002-05-10 2006-02-16 Metod Koselj Graphics engine with edge draw unit, and electrical device and memopry incorporating the graphics engine
GB2388506B (en) * 2002-05-10 2004-04-21 Nec Electronics Display driver IC, display module and electrical device incorporating a graphics engine
US7027056B2 (en) * 2002-05-10 2006-04-11 Nec Electronics (Europe) Gmbh Graphics engine, and display driver IC and display module incorporating the graphics engine
US20060230428A1 (en) * 2005-04-11 2006-10-12 Rob Craig Multi-player video game system
US8284842B2 (en) 2005-07-08 2012-10-09 Activevideo Networks, Inc. Video game system using pre-encoded macro-blocks and a reference grid
US9061206B2 (en) 2005-07-08 2015-06-23 Activevideo Networks, Inc. Video game system using pre-generated motion vectors
US8074248B2 (en) 2005-07-26 2011-12-06 Activevideo Networks, Inc. System and method for providing video content associated with a source image to a television in a communication network
US20080043643A1 (en) * 2006-07-25 2008-02-21 Thielman Jeffrey L Video encoder adjustment based on latency
US8888592B1 (en) 2009-06-01 2014-11-18 Sony Computer Entertainment America Llc Voice overlay
KR101366092B1 (en) 2006-10-13 2014-02-21 삼성전자주식회사 Method and apparatus for encoding and decoding multi-view image
KR100823287B1 (en) * 2007-01-03 2008-04-21 삼성전자주식회사 Method and apparatus for encoding and decoding multi-view image based on global disparity vector
WO2008088772A2 (en) 2007-01-12 2008-07-24 Ictv, Inc. Mpeg objects and systems and methods for using mpeg objects
US9826197B2 (en) 2007-01-12 2017-11-21 Activevideo Networks, Inc. Providing television broadcasts over a managed network and interactive content over an unmanaged network to a client device
US8548261B2 (en) 2007-04-11 2013-10-01 Samsung Electronics Co., Ltd. Method and apparatus for encoding and decoding multi-view image
US8305914B2 (en) * 2007-04-30 2012-11-06 Hewlett-Packard Development Company, L.P. Method for signal adjustment through latency control
US8147339B1 (en) 2007-12-15 2012-04-03 Gaikai Inc. Systems and methods of serving game video
US8968087B1 (en) 2009-06-01 2015-03-03 Sony Computer Entertainment America Llc Video game overlay
US8613673B2 (en) 2008-12-15 2013-12-24 Sony Computer Entertainment America Llc Intelligent game loading
US8447065B2 (en) 2008-09-16 2013-05-21 Cyberlink Corp. Method of facial image reproduction and related device
US8926435B2 (en) 2008-12-15 2015-01-06 Sony Computer Entertainment America Llc Dual-mode program execution
US8840476B2 (en) 2008-12-15 2014-09-23 Sony Computer Entertainment America Llc Dual-mode program execution
US8506402B2 (en) 2009-06-01 2013-08-13 Sony Computer Entertainment America Llc Game execution environments
US8771064B2 (en) 2010-05-26 2014-07-08 Aristocrat Technologies Australia Pty Limited Gaming system and a method of gaming
US8676591B1 (en) 2010-08-02 2014-03-18 Sony Computer Entertainment America Llc Audio deceleration
US10039978B2 (en) 2010-09-13 2018-08-07 Sony Interactive Entertainment America Llc Add-on management systems
WO2012051528A2 (en) 2010-10-14 2012-04-19 Activevideo Networks, Inc. Streaming digital video between video devices using a cable television system
EP2695388B1 (en) 2011-04-07 2017-06-07 ActiveVideo Networks, Inc. Reduction of latency in video distribution networks using adaptive bit rates
US8793393B2 (en) * 2011-11-23 2014-07-29 Bluespace Corporation Video processing device, video server, client device, and video client-server system with low latency thereof
US10409445B2 (en) 2012-01-09 2019-09-10 Activevideo Networks, Inc. Rendering of an interactive lean-backward user interface on a television
US9123084B2 (en) 2012-04-12 2015-09-01 Activevideo Networks, Inc. Graphical application integration with MPEG objects
KR20150095679A (en) * 2012-11-13 2015-08-21 엘지전자 주식회사 Method and apparatus for processing video signals
US10275128B2 (en) 2013-03-15 2019-04-30 Activevideo Networks, Inc. Multiple-mode system and method for providing user selectable video content
US9219922B2 (en) 2013-06-06 2015-12-22 Activevideo Networks, Inc. System and method for exploiting scene graph information in construction of an encoded video sequence

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0405776A2 (en) * 1989-06-09 1991-01-02 Interactive Network, Inc. Remote gaming system playable by several participants
EP0625760A1 (en) * 1993-05-19 1994-11-23 Julian Dr. Menashe Interactive, computerised gaming system with remote terminals
WO1995008793A1 (en) * 1993-09-23 1995-03-30 Virtual Universe Corporation Virtual reality network

Family Cites Families (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2514115B2 (en) * 1991-02-15 1996-07-10 株式会社グラフィックス・コミュニケーション・テクノロジーズ Image signal encoder
US5265180A (en) * 1991-06-13 1993-11-23 Intel Corporation Method of encoding a sequence of images of a digital motion video signal
JPH0595540A (en) * 1991-09-30 1993-04-16 Sony Corp Dynamic picture encoder
JPH05167998A (en) * 1991-12-16 1993-07-02 Nippon Telegr & Teleph Corp <Ntt> Image-encoding controlling method
US5644660A (en) * 1992-04-09 1997-07-01 Picturetel Corporation Method and apparatus for efficiently transmitting forced updates in a moving picture codec
WO1993023816A1 (en) * 1992-05-18 1993-11-25 Silicon Engines Inc. System and method for cross correlation with application to video motion vector estimation
WO1994030013A1 (en) * 1993-06-08 1994-12-22 Sony Corporation Encoder and encoding method
US5703966A (en) * 1995-06-27 1997-12-30 Intel Corporation Block selection using motion estimation error

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0405776A2 (en) * 1989-06-09 1991-01-02 Interactive Network, Inc. Remote gaming system playable by several participants
EP0625760A1 (en) * 1993-05-19 1994-11-23 Julian Dr. Menashe Interactive, computerised gaming system with remote terminals
WO1995008793A1 (en) * 1993-09-23 1995-03-30 Virtual Universe Corporation Virtual reality network

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO1999060498A1 (en) * 1998-05-18 1999-11-25 Aristocrat Leisure Industries Pty. Ltd. Intelligent input/output control system
FR2940703A1 (en) * 2008-12-31 2010-07-02 Cy Play Display modeling method for application on server, involves forming image based on pixels of traces, and transmitting image and encoding information conforming to assembly of modification data to encoder by transmitting unit
FR2940689A1 (en) * 2008-12-31 2010-07-02 Cy Play Method for navigating user of mobile terminal e.g. portable computer, on application, involves detecting movement of positioning device, and modulating movement data in audio format for transmitting movement data to server
FR2940690A1 (en) * 2008-12-31 2010-07-02 Cy Play Mobile terminal i.e. mobile telephone, user navigation method, involves establishing contents to be displayed on terminal for permitting navigation on user interface having size larger than size of screen of terminal
WO2010076436A3 (en) * 2008-12-31 2010-11-25 Cy Play Method for macroblock modeling of the display of a remote terminal by means of layers characterized by a movement vector and transparency data
US9185159B2 (en) 2008-12-31 2015-11-10 Cy-Play Communication between a server and a terminal
US9860551B2 (en) * 2011-02-09 2018-01-02 Lg Electronics Inc. Method for encoding and decoding image and device using same
US9866861B2 (en) 2011-02-09 2018-01-09 Lg Electronics Inc. Method for encoding and decoding image and device using same
US10516895B2 (en) 2011-02-09 2019-12-24 Lg Electronics Inc. Method for encoding and decoding image and device using same
US11032564B2 (en) 2011-02-09 2021-06-08 Lg Electronics Inc. Method for encoding and decoding image and device using same
US11463722B2 (en) 2011-02-09 2022-10-04 Lg Electronics Inc. Method for encoding and decoding image and device using same
US11871027B2 (en) 2011-02-09 2024-01-09 Lg Electronics Inc. Method for encoding image and non-transitory computer readable storage medium storing a bitstream generated by a method

Also Published As

Publication number Publication date
JPH10510131A (en) 1998-09-29
DE69619220D1 (en) 2002-03-21
EP0800689A1 (en) 1997-10-15
DE69619220T2 (en) 2002-09-26
US5991443A (en) 1999-11-23
JP3884076B2 (en) 2007-02-21
GB9519921D0 (en) 1995-11-29
EP0800689B1 (en) 2002-02-13

Similar Documents

Publication Publication Date Title
US5991443A (en) Graphics image manipulation
US5742289A (en) System and method of generating compressed video graphics images
US8553772B2 (en) Moving picture prediction system
KR100350335B1 (en) Image information processing method and apparatus
EP1096800B1 (en) Image coding apparatus and image decoding apparatus
US5815604A (en) Interactive image manipulation
CN1166207C (en) Encoding a video signal
US5737023A (en) Hierarchical motion estimation for interlaced video
CA2557534A1 (en) Method and system for digital decoding 3d stereoscopic video images
US5739862A (en) Reverse playback of MPEG video
WO1996033575A1 (en) Video decoder using block oriented data structures
JP3331351B2 (en) Image data encoding method and apparatus
US5457481A (en) Memory system for use in a moving image decoding processor employing motion compensation technique
EP1147671B1 (en) Method and apparatus for performing motion compensation in a texture mapping engine
US6539058B1 (en) Methods and apparatus for reducing drift due to averaging in reduced resolution video decoders
EP0732671B1 (en) Video decoder system
CN1126411A (en) Apparatus for parallel encoding/decoding of digital video signals
KR100472564B1 (en) Image processor
US5990959A (en) Method, system and product for direct rendering of video images to a video data stream
CN1126410A (en) Apparatus for parallel decoding of digital video signals
KR19990050063A (en) Coding Mode Switching Control Device of Image Coding System
Zhang Morphing for intelligent motion picture and video interpolation: bounding distortion and autonomous implementation
KR19990041941A (en) Coding Mode Switching Control Device of Image Coding System

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): JP

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): AT BE CH DE DK ES FI FR GB GR IE IT LU MC NL PT SE

WWE Wipo information: entry into national phase

Ref document number: 1996929474

Country of ref document: EP

121 Ep: the epo has been informed by wipo that ep was designated in this application
WWP Wipo information: published in national office

Ref document number: 1996929474

Country of ref document: EP

WWG Wipo information: grant in national office

Ref document number: 1996929474

Country of ref document: EP