WO2014181892A1 - Information processing apparatus, control method and program - Google Patents

Information processing apparatus, control method and program Download PDF

Info

Publication number
WO2014181892A1
WO2014181892A1 PCT/JP2014/062761 JP2014062761W WO2014181892A1 WO 2014181892 A1 WO2014181892 A1 WO 2014181892A1 JP 2014062761 W JP2014062761 W JP 2014062761W WO 2014181892 A1 WO2014181892 A1 WO 2014181892A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
program
content
operation input
receiving
Prior art date
Application number
PCT/JP2014/062761
Other languages
French (fr)
Inventor
Cyril PERRIN
Original Assignee
Square Enix Holdings Co., Ltd.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Square Enix Holdings Co., Ltd. filed Critical Square Enix Holdings Co., Ltd.
Priority to EP14794717.0A priority Critical patent/EP2994830A4/en
Priority to JP2015546744A priority patent/JP6576245B2/en
Priority to US14/787,029 priority patent/US20160110903A1/en
Priority to CA2910655A priority patent/CA2910655A1/en
Publication of WO2014181892A1 publication Critical patent/WO2014181892A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/131Protocols for games, networked simulations or virtual reality
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T11/002D [Two Dimensional] image generation
    • G06T11/60Editing figures and text; Combining figures or text
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/30Interconnection arrangements between game servers and game devices; Interconnection arrangements between game devices; Interconnection arrangements between game servers
    • A63F13/33Interconnection arrangements between game servers and game devices; Interconnection arrangements between game devices; Interconnection arrangements between game servers using wide area network [WAN] connections
    • A63F13/335Interconnection arrangements between game servers and game devices; Interconnection arrangements between game devices; Interconnection arrangements between game servers using wide area network [WAN] connections using Internet
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/30Interconnection arrangements between game servers and game devices; Interconnection arrangements between game devices; Interconnection arrangements between game servers
    • A63F13/35Details of game servers
    • A63F13/355Performing operations on behalf of clients with restricted processing capabilities, e.g. servers transform changing game scene into an encoded video stream for transmitting to a mobile phone or a thin client
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/50Controlling the output signals based on the game progress
    • A63F13/53Controlling the output signals based on the game progress involving additional visual information provided to the game scene, e.g. by overlay to simulate a head-up display [HUD] or displaying a laser sight in a shooting game
    • A63F13/533Controlling the output signals based on the game progress involving additional visual information provided to the game scene, e.g. by overlay to simulate a head-up display [HUD] or displaying a laser sight in a shooting game for prompting the player, e.g. by displaying a game menu
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T11/002D [Two Dimensional] image generation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/60Network structure or processes for video distribution between server and client or between remote clients; Control signalling between clients, server and network components; Transmission of management data between server and client, e.g. sending from server to client commands for recording incoming content stream; Communication details between server and client 
    • H04N21/61Network physical structure; Signal processing
    • H04N21/6106Network physical structure; Signal processing specially adapted to the downstream path of the transmission network
    • H04N21/6125Network physical structure; Signal processing specially adapted to the downstream path of the transmission network involving transmission via Internet

Definitions

  • the present invention relates to an
  • a game screen rendered in a server is provided to a client device via a network, like so-called cloud gaming.
  • a game content that generates fine graphics requires sufficient rendering performance of a client device.
  • cloud gaming allows even a user who does not have a client device with
  • an application program and the like are preferably
  • the present invention was made in view of such problems in the conventional technique.
  • An aspect of the present invention provides a user experience with extended functions when providing an existing content without altering, the program of the content.
  • the present invention in its first aspect provides an information processing apparatus
  • receiving means for receiving operation input for a content comprising: receiving means for receiving operation input for a content; first generation means for
  • second generation means for generating a second image to be added to the first image by executing a second program different from the first program
  • output means for outputting a composite image obtained by compositing the first image and the second image
  • control means for, in a case where the receiving means receives the operation input to a region according to the second image in the composite image, controlling not to cause the first generation means to execute the first program in accordance with the operation input.
  • the present invention in its second aspect provides an information processing apparatus
  • first generation means for generating a first image corresponding to the content by executing a first program for the content in accordance with the operation input received by the receiving means; monitor means for monitoring a predetermined parameter that changes during execution of the first program; second generation means for generating a second image to be added to the first image by executing a second program different from the first program in a case where the predetermined
  • output means for outputting a composite image obtained by compositing the first image and the second image.
  • the present invention in its third aspect provides an information processing apparatus
  • receiving means for receiving operation input for a content comprising: receiving means for receiving operation input for a content; first generation means for
  • analysis means for analyzing the first image and detecting whether an execution state of the content meets a predetermined condition
  • generation means for generating a second image to be added to the first image by executing a second program different from the first program in a case where the execution state meets the predetermined condition; and output means for outputting a composite image obtained by compositing the first image and the second image.
  • the present invention in its fourth aspect provides an information processing apparatus
  • receiving means for receiving operation input for a content comprising: receiving means for receiving operation input for a content; first generation means for
  • analysis means for analyzing the first image and deciding a position where a display item is to be arranged; second generation means for generating a second image to be added to the first image, in which the display item is arranged at the position decided by the analysis means, by executing a second program different from the first program; and output means for outputting a composite image obtained by compositing the first image and the second image.
  • Fig. 1 is a block diagram of a cloud-based video game system architecture, according to a non- limiting embodiment of the present invention.
  • Fig. 2A is a block diagram showing various physical components of the architecture of Fig. 1, according to a non-limiting embodiment of the present invention .
  • Fig. 2B is a variant of Fig. 2A.
  • Fig. 2C is a block diagram showing various functional modules of the architecture of Fig. 1, which can be implemented by the physical components of Figs. 2A or 2B.
  • FIGs. 3A to 3C are flowcharts showing
  • FIGs. 4A and 4B are flowcharts showing
  • FIGs. 5A and 5B are block diagram showing a functional arrangement on a server side according to an embodiment or a modification of the present invention.
  • Fig. 6 is a flowchart exemplary showing menu extension processing performed on the server side according to the embodiment of the present invention.
  • Figs. 7A, 7B and 7C are views exemplary showing the structures of images before and after composite processing in the menu extension processing according to the embodiment of the present invention.
  • Figs. 8A and 8B are flowcharts exemplary showing display update processing performed on the server side according to the embodiment of the present invention .
  • Figs. 9A and 9B are views exemplary showing the structures of superimposed images generated by the display update processing.
  • Figs. 10A, 10B and IOC are views exemplary showing the structures of images before and after composite processing according to a modification of the present invention.
  • FIGs. 11A and 11B are views showing the structures of composite images according to the third modification of the present invention.
  • Fig. 1 schematically shows a cloud-based video game system architecture according to a non-limiting embodiment of the present invention.
  • the architecture includes client devices 120, 120A connected to a cloud gaming server system 100 over a data network such as the Internet 130.
  • Each of the client devices 120, 120A may connect to the Internet 130 in any suitable manner, including over a respective local access network (not shown) .
  • the cloud gaming server system 100 may also connect to the Internet 130 over a local access network (not shown) , although the server system 100 may connect directly to the Internet 130 without the intermediary of a local access network. Connections between the cloud gaming server system 100 and one or more of the client devices 120, 120A may comprise one or more channels.
  • channels can be made up of physical and/or logical links, and may travel over a variety of physical media, including radio frequency, fiber optic, free-space optical, coaxial and twisted pair.
  • the channels may abide by a protocol such as UDP or TCP/IP.
  • one or more of the channels may be supported a virtual private network (VPN) .
  • VPN virtual private network
  • one or more of the connections may be session-based.
  • the cloud gaming server system 100 enables users of the client devices 120, 120A to play video games, either individually (i.e., a single-player video game) or in groups (i.e., a multiplayer video game) .
  • video games may include games that are played for leisure, education and/or sport.
  • a video game may but need not offer participants the possibility of monetary gain.
  • client devices 120, 120A are shown, it should be appreciated that the number of client devices in the cloud-based video game system architecture is not particularly limited .
  • a user of one of the client devices 120, 120A may register with the cloud gaming server system 100 as a participant in a video game.
  • the user may register as a "player", and will have the opportunity to control a character, avatar, race car, cockpit, etc. within a virtual world maintained by the video game.
  • the virtual world is shared by two or more players, and one player's
  • gameplay may affect that of another.
  • a user of one of the client devices 120, 120A may register as a non-player "spectator", whereby such users will observe players' gameplay but otherwise do not control active characters in the game. Unless otherwise indicated, where the term “participant” is used, it is meant to apply equally to players and spectators.
  • the server system 100 may include one or more computing resources, including one or more game servers and one or more account servers.
  • the game servers and the account servers may be embodied in the same
  • a game server interacts with players in the course of a game, while an account server interacts with players outside the game environment.
  • the account server may be configured for logging a prospective player into a game portal,
  • the account server may host or have access to a participant
  • the participant database 10 that stores account information about various participants and client devices 120, 120A, such as identification data, financial data, location data, demographic data, connection data and the like.
  • the participant database 10 which can be part of the cloud gaming server system 100 or situated remotely therefrom.
  • the game server 110 may also have access to the
  • participant database 10 that stores the above mentioned account information, as this information may be used, for influencing the way in which the video game
  • any given one of the client devices 120, 120A is not particularly limited.
  • one or more of the client devices 120, 120A may be, for example, a personal computer (PC), a home game machine (console such as XBOXTM, PS3TM, iiTM, etc.), a portable game machine, a smart television, a set-top box (STB) , etc.
  • one or more of the client devices 120, 120A may be a PC or more personal computer (PC), a home game machine (console such as XBOXTM, PS3TM, iiTM, etc.), a portable game machine, a smart television, a set-top box (STB) , etc.
  • one or more of the client devices 120, 120A may be a personal computer (PC), a home game machine (console such as XBOXTM, PS3TM, iiTM, etc.), a portable game machine, a smart television, a set-top box (STB) , etc.
  • Any given one of the client devices 120, 120A may be equipped with one or more input devices (such as a touch screen, a keyboard, a game controller, a
  • the user may produce body motion or may wave an external object; these movements are detected by a camera or other sensor (e.g., KinectTM) , while software operating within the given client device attempts to correctly guess whether the user intended to provide input to the given client device and, if so, the nature of such input.
  • the given client device translates the received user inputs and detected user movements into "client device input", which is sent to the cloud gaming server system 100 over the Internet 130.
  • client device 120 produces client device input 140
  • client device 120A produces client device input 140A.
  • the cloud gaming server system 100 processes the client device input 140, 140A received from the various client devices 120, 120A and generates "media output" for the various client devices 120, 120A.
  • the media output may include a stream of encoded video data (representing images when displayed on a screen) and audio data (representing sound when played via a
  • the media output is sent over the
  • Each of the client devices 120, 120A may include circuitry for buffering and processing the media output in the packets received from the cloud gaming server system 100, as well as a display for displaying images and a transducer (e.g., a loudspeaker) for outputting audio. Additional output devices may also be provided, such as an electromechanical system to induce motion.
  • a stream of video data can be divided into “frames".
  • the term "frame” as used herein does not require the existence of a one-to-one correspondence between frames of video data and images represented by the video data. That is to say, while it is possible for a frame of video data to contain data representing a respective displayed image in its entirety, it is also possible for a frame of video data to contain data representing only part of an image, and for the image to in fact require two or more frames in order to be properly reconstructed and displayed.
  • a frame of video data may contain data representing more than one complete image, such that N images may be represented using M frames of video data, where M ⁇ N.
  • FIG. 2A shows one possible non-limiting physical arrangement of components for the cloud gaming server system 100.
  • individual servers within the cloud gaming server system 100 are
  • a compute server 200C may be primarily
  • a rendering server 200R may be primarily responsible for rendering graphics (video data) .
  • both client device 120 and client device 120A are assumed to be participating in the video game, either as players or spectators. However, it should be understood that in some cases there may be a single player and no spectator, while in other cases there may be multiple players and a single spectator, in still other cases there may be a single player and multiple spectators and in yet other cases there may be multiple players and multiple spectators.
  • the following description refers to a single compute server 200C connected to a single rendering server 200R. However, it should be appreciated that there may be more than one rendering server 200R connected to the same compute server 200C, or more than one compute server 200C connected to the same rendering server 200R. In the case where there are multiple rendering servers 200R, these may be distributed over any suitable geographic area .
  • the compute server 200C comprises one or more central processing units (CPUs) 220C, 222C and a random access memory (RAM) 230C.
  • the CPUs 220C, 222C can have access to the RAM 230C over a communication bus architecture, for example. While only two CPUs 220C, 222C are shown, it should be appreciated that a greater number of CPUs, or only a single CPU, may be provided in some example implementations of the compute server 200C.
  • the compute server 200C also comprises a network interface component (NIC) 210C2, where client device input is received over the Internet 130 from each of the client devices participating in the video game.
  • NIC network interface component
  • the compute server 200C further comprises another network interface component (NIC) 210C1, which outputs a sets of rendering commands 204.
  • the sets of rendering commands 204 output from the compute server 200C via the NIC 210C1 can be sent to the rendering server 200R.
  • the compute server 200C can be connected directly to the rendering server 200R.
  • the compute server 200C can be connected to the rendering server 200R over a network 260, which can be the Internet 130 or another network.
  • a virtual private network (VPN) may be
  • the sets of rendering commands 204 sent by the compute server 200C are received at a network interface component (NIC) 210R1 and are directed to one or more CPUs 220R, 222R.
  • the CPUs 220R, 222R are connected to graphics
  • GPUs processing units
  • GPU 240R may include a set of GPU cores 242R and a video random access memory (VRAM) 246R.
  • GPU 250R may include a set of GPU cores 252R and a video random access memory (VRAM) 256R.
  • Each of the CPUs 220R, 222R may be connected to each of the GPUs 240R, 250R or to a subset of the GPUs 240R, 250R. Communication between the CPUs 220R, 222R and the GPUs 240R, 250R can be established using, for example, a communications bus architecture.
  • the CPUs 220R, 222R cooperate with the GPUs 240R, 250R to convert the sets of rendering commands 204 into a graphics output streams, one for each of the participating client devices.
  • the CPUs 220R, 222R cooperate with the GPUs 240R, 250R to convert the sets of rendering commands 204 into a graphics output streams, one for each of the participating client devices.
  • the rendering server 200R comprises a further network interface component (NIC) 210R2, through which the graphics output streams 206, 206A are sent to the client devices 120, 120A, respectively.
  • NIC network interface component
  • Fig. 2B shows a second possible non-limiting physical arrangement of components for the cloud gaming server system 100.
  • a hybrid server 200H is responsible both for tracking state changes in a video game based on user input, and for rendering graphics (video data) .
  • the hybrid server 200H comprises one or more central processing units
  • CPUs central processing unit
  • RAM random access memory
  • the CPUs 220H, 222H can have access to the RAM 230H over a communication bus architecture, for example.
  • CPUs 220H, 222H While only two CPUs 220H, 222H are shown, it should be appreciated that a greater number of CPUs, or only a single CPU, may be provided in some example
  • the hybrid server 200H also comprises a network interface
  • NIC component 210H
  • client device input is received over the Internet 130 from each of the client devices participating in the video game.
  • client device 120 and client device 120A are assumed to be participating in the video game, and therefore the received client device input may include client device input 140 and client device input 140A.
  • GPUs 220H, 222H are connected to a graphics processing units (GPUs) 240H, 250H.
  • GPU 240H may include a set of GPU cores 242H and a video random access memory
  • VRAM video random access memory
  • GPU 250H may include a set of GPU cores 252H and a video random access memory (VRAM) 256H.
  • Each of the CPUs 220H, 222H may be connected to each of the GPUs 240H, 250H or to a subset of the GPUs 240H, 250H.
  • Communication between the CPUs 220H, 222H and the GPUs 240H, 250H can be established using, for example, a communications bus architecture. Although only two CPUs and two GPUs are shown, there may be more than two CPUs and GPUs,, or even just a single CPU or GPU, in a specific example of implementation of the hybrid server 200H.
  • the CPUs 220H, 222H cooperate with the GPUs 240H, 250H to convert the sets of rendering commands 204 into graphics output streams, one for each of the participating client devices.
  • the graphics output streams 206, 206A are sent to the client devices 120, 120A, respectively, via the NIC 210H.
  • the above-described physical components of the compute server 200C and the rendering server 200R (in Fig. 2A) and/or of the hybrid server 200H (in Fig. 2B) implement a set of functional modules, including a video game functional module 270, a rendering functional module 280 and a video encoder 285.
  • the video game functional module 270 the above-described physical components of the compute server 200C and the rendering server 200R (in Fig. 2A) and/or of the hybrid server 200H (in Fig. 2B) implement a set of functional modules, including a video game functional module 270, a rendering functional module 280 and a video encoder 285.
  • the video game functional module 270 implements a set of functional modules, including a video game functional module 270, a rendering functional module 280 and a video encoder 285.
  • the functional module 270 is implemented by the compute server 200C, while the rendering functional module 280 and the video encoder 285 are implemented by the rendering server 200R.
  • the hybrid server 200H is implemented by the compute server 200C
  • the present example embodiment discusses a single video game functional module 270 for simplicity of illustration. However, it should be noted that in an actual implementation of the cloud gaming server system 100, many video game functional modules similar to the video game functional module 270 would be executed in parallel. Thus, the cloud gaming server system 100 could support multiple independent
  • video games can be single-player video games or multi-player games of any type.
  • the video game functional module 270 may be implemented by certain physical components of the compute server 200C (in Fig. 2A) or of the hybrid server 200H (in Fig. 2B) .
  • the video game functional module 270 can be encoded as computer- readable instructions that are executable by a CPU (such as the CPUs 220C, 222C in the compute server 200C or the CPUs 220H, 222H in the hybrid server 200H) .
  • the instructions can be tangibly stored in the RAM 230C (in the compute server 200C) of the RAM 230H (in the hybrid server 200H) or in another memory area, together with constants, variables and/or other data used by the video game functional module 270.
  • the video game functional module 270 may be executed within the environment of a virtual machine that may be supported by an operating system that is also being executed by a CPU (such as the CPUs 220C, 222C in the compute server 200C or the CPUs 220H, 222H in the hybrid server 200H) .
  • a CPU such as the CPUs 220C, 222C in the compute server 200C or the CPUs 220H, 222H in the hybrid server 200H.
  • the rendering functional module 280 may be implemented by certain physical components of the rendering server 200R (in Fig. 2A) or of the hybrid server 200H (in Fig. 2B) .
  • the rendering functional module 280 may take up one or more GPUs (240R, 250R in Fig. 2A, 240H, 250H in Fig. 2B) and may or may not utilize CPU resources.
  • the video encoder 285 may be implemented by certain physical components of the rendering server 200R (in Fig. 2A) or of the hybrid server 200H (in Fig. 2B) .
  • the video encoder 285 may be implemented by the CPUs 220R, 222R and/or by the GPUs 240R, 250R.
  • the video encoder 285 may be implemented by the CPUs 220H, 222H and/or by the GPUs 240H, 250H.
  • the video encoder 285 may be implemented by a separate encoder chip (not shown) .
  • the video game functional module 270 produces the sets of rendering commands 2 ⁇ 4, based on received client device input.
  • the received client device input may carry data (e.g., an address)
  • participant participants in the video game i.e., players or
  • the received client device input includes the client device input 140, 140A received from the client devices 120, 120A.
  • Rendering commands refer to commands which can be used to instruct a specialized graphics processing unit (GPU) to produce a frame of video data or a
  • the images represented by these frames change as a function of responses to the client device input 140, 140A that are programmed into the video game functional module 270.
  • the video game functional module 270 may be programmed in such a way as to respond to certain specific stimuli to provide the user with an experience of progression
  • the instructions for the video game functional module 270 may be fixed in the form of a binary
  • the client device input 140, 140A is unknown until the moment of interaction with a player who uses the corresponding client device 120, 120A.
  • the client device input 140, 140A is unknown until the moment of interaction with a player who uses the corresponding client device 120, 120A.
  • players/spectators and the video game functional module 270 via the client devices 120, 120A can be referred to as “gameplay” or “playing a video game”.
  • the rendering functional module 280 processes the sets of rendering commands 204 to create multiple video data streams 205. Generally, there will be one video data stream per participant (or, equivalently, per client device) .
  • data for one or more objects represented in three- dimensional space (e.g., physical objects) or two- dimensional space (e.g., text) may be loaded into a cache memory (not shown) of a particular GPU 240R, 250R, 240H, 250H.
  • This data may be transformed by the GPU 240R, 250R, 240H, 250H into data representative of a two-dimensional image, which may be stored in the appropriate VRAM 246R, 256R, 246H, 256H.
  • the VRAM 246R, 256R, 246H, 256H may provide temporary storage of picture element (pixel) values for a game screen .
  • the video encoder 285 compresses and encodes the video data in each of the video data streams 205 into a corresponding stream of compressed/encoded video data.
  • the resultant streams of compressed/encoded video data referred to as graphics output streams, are produced on a per-client-device basis.
  • the video encoder 285 produces graphics output stream 206 for client device 120 and graphics output stream 206A for client device 120A. Additional functional modules may be provided for formatting the video data into packets so that they can be transmitted over the Internet 130.
  • compressed/encoded video data within a given graphics output stream may be divided into frames.
  • execution of the video game functional module 270 involves several processes, including a main game process 300A and one or more graphics control processes 300B, which are described herein below in greater detail.
  • a first process referred to as the main game process, is described with reference to Fig. 3A.
  • the main game process 300A executes continually.
  • an action 310A during which client device input may be received.
  • client device input e.g., client device input 140
  • client device 120 e.g., client device 120
  • the client device input e.g., the client device input 140 and 140A
  • client devices e.g., the client devices 120 and 120A
  • the input from a given client device may convey that the user of the given client device wishes to cause a character under his or her control to move, jump, kick, turn, swing, pull, grab, etc.
  • the input from the given client device may convey a menu selection made by the user of the given client device in order to change one or more audio, video or gameplay settings, to load/save a game or to create or join a network session.
  • the input from the given client device may convey that the user of the given client device wishes to select a particular camera view (e.g., first-person or third- person) or reposition his or her viewpoint within the virtual world.
  • the game state may be updated based at least in part on the client device input received at action 310A and other parameters. Updating the game state may involve the following actions:
  • updating the game state may involve updating certain properties of the participants (player or spectator) associated with the client devices from which the client device input may have been received. These properties may be stored in the participant database 10. Examples of participant properties that may be maintained in the participant database 10 and updated at action 320A can include a camera view selection (e.g., 1st person, 3rd person), a mode of play, a selected audio or video setting, a skill level, a customer grade (e.g., guest, premium, etc.).
  • a camera view selection e.g., 1st person, 3rd person
  • a mode of play e.g., a selected audio or video setting
  • a skill level e.g., guest, premium, etc.
  • updating the game state may involve updating the attributes of certain objects in the virtual world based on an interpretation of the client device input.
  • the objects whose attributes are to be updated may in some cases be represented by two- or three-dimensional models and may include playing characters, non-playing characters and other objects.
  • attributes that can be updated may include the object's position, strength, weapons/armor, lifetime left, special powers,
  • attributes that can be updated may include the object's position, velocity, animation, damage/health, visual effects, textual content, etc.
  • timers such as elapsed time, time since a particular event, virtual time of day, total number of players, a participant's geographic location, etc.
  • the main game process 300A returns to action 310A, whereupon new client device input received since the last pass through the main game process is gathered and processed.
  • the graphics control process 300B may execute continually, and there may be multiple separate
  • graphics control processes 300B each of which results in a respective one of the sets of rendering commands 204.
  • the graphics control process 300B may execute as an extension of the main game process 300A described above.
  • multiple distinct sets of rendering commands need to be generated for the multiple players, and therefore multiple graphics control processes 300B may execute in parallel.
  • there may again be only a single set of rendering commands 204 and therefore a single graphics control process 300B may execute in the video game functional module 270, but the resulting video data stream may be
  • the video game functional module 270 determines the objects to be rendered for the given participant. This action can include identifying the following types of objects:
  • this action can include identifying those objects from the virtual world that are in the "game screen rendering range" (also known as a "scene") for the given participant.
  • the game screen rendering range includes the portion of the virtual world that would be “visible” from the perspective of the given participant's camera. This depends on the position and orientation of that camera relative to the objects in the virtual world.
  • a frustum can be applied to the virtual world, and the objects within that frustum are retained or marked.
  • the frustum has an apex which is situated at the location of the given participant's camera and has a directionality also defined by the directionality of that camera.
  • this action can include identifying additional objects that do not appear in the virtual world, but which nevertheless are to be rendered for the given participant.
  • these additional objects may include textual messages, graphical images, etc.
  • warnings and dashboard indicators to name a few non- limiting possibilities.
  • the video game functional module 270 generates a set of commands for transforming rendering into graphics (video data) the objects that were identified at action 310B.
  • Rendering may refer to the transformation of 3-D or 2-D coordinates of an object or group of objects into data representative of a displayable image, in accordance with the viewing perspective and prevailing lighting conditions. This can be achieved using any number of different
  • the generated at action 320B are output to the rendering functional module 280. This may involve packetizing the generated rendering commands into a set of
  • rendering commands 204 that is sent to the rendering functional module 280.
  • the rendering functional module 280 interprets the sets of rendering commands 204 and produces
  • multiple video data streams 205 one for each
  • Rendering may be achieved by the GPUs 240R, 250R, 240H, 250H under control of the CPUs 220R, 222R (in Fig. 2A) or 220H, 222H (in Fig. 2B) .
  • the rate at which frames of video data are produced for a participating client device may be referred to as the frame rate.
  • N there may be N sets of rendering commands 204 (one for each participant) and also N video data streams 205 (one for each participant). In that case, rendering functionality is not shared among the
  • the N video data streams 205 may also be created from sets of rendering commands 204 (where M ⁇ N) , such that fewer sets of rendering commands need to be processed by the rendering
  • the rendering functional unit 280 may perform sharing or duplication in order to generate a larger number of video data streams 205 from a smaller number of sets of rendering commands 204. Such sharing or duplication may be prevalent when multiple participants (e.g., spectators) desire to view the same camera perspective. Thus, the rendering functional module 280 may perform functions such as duplicating a created video data stream for one or more spectators.
  • the video data in each of the video data streams 205 are encoded by the video encoder 285, resulting in a sequence of encoded video data
  • graphics output stream 206 the sequence of encoded video data destined for client device 120 is referred to as graphics output stream 206, while the sequence of encoded video data destined for client device 120A is referred to as graphics output stream 206A.
  • the video encoder 285 can be a device (or set of computer-readable instructions) that enables or - 3.1 - carries out or defines a video compression or
  • Video compression transforms an original stream of digital image data (expressed in terms of pixel locations, color values, etc.) into an output stream of digital image data that conveys substantially the same
  • the encoding process used to encode a particular frame of video data may or may not involve cryptographic encryption.
  • the graphics output streams 206, 206A created in the above manner are sent over the Internet 130 to the respective client devices.
  • the graphics output streams may be segmented and formatted into packets, each having a header and a payload.
  • the header of a packet containing video data for a given participant may include a network address of the client device associated with the given
  • the payload may include the video data, in whole or in part.
  • compression algorithm used to encode certain video data may be encoded in the content of one or more packets that convey that video data.
  • Fig. 4A shows operation of the client device associated with a given participant, which may be client device 120 or client device 120A, by way of non-limiting example.
  • a graphics output stream (e.g., 206, 206A) is received over the Internet 130 from the rendering server 200R (Fig. 2A) or from the hybrid server 200H (Fig. 2B) , depending on the embodiment.
  • the received graphics output stream comprises
  • the compressed/encoded frames of video data are decoded/decompressed in accordance with the decompression algorithm that is complementary to the encoding/compression algorithm used in the encoding/compression process.
  • encoding/compression algorithm used to encode/compress the video data may be known in advance. In other embodiments, the identity or version of the
  • encoding/compression algorithm used to encode the video data may accompany the video data itself.
  • the (decoded/decompressed) frames of video data are processed. This can include placing the decoded/decompressed frames of video data in a buffer, performing error correction, reordering and/or combining the data in multiple successive frames, alpha blending, interpolating portions of missing data, and so on.
  • the result can be video data representative of a final image to be presented to the user on a per- frame basis.
  • the final image is output via the output mechanism of the client device.
  • a composite video frame can be displayed on the display of the client device.
  • a third process referred to as the audio
  • the audio generation process may execute independently of the graphics control process 300B. In another embodiment, execution of the audio generation process and the graphics
  • control process may be coordinated.
  • the video game functional module 270 determines the sounds to be produced.
  • this action can include identifying those sounds associated with objects in the virtual world that dominate the acoustic landscape, due to their volume (loudness) and/or proximity to the participant within the virtual world.
  • the video game functional module 270 generates an audio segment.
  • the duration of the audio segment may span the duration of a video frame, although in some embodiments, audio segments may be generated less frequently than video frames, while in other embodiments, audio segments may be generated more frequently than video frames.
  • the audio segment is encoded, e.g., by an audio encoder, resulting in an encoded audio segment.
  • the audio encoder can be a device (or set of instructions) that enables or carries out or defines an audio compression or decompression algorithm. Audio compression transforms an original stream of digital audio (expressed as a sound wave changing in amplitude and phase over time) into an output stream of digital audio data that conveys substantially the same information but using fewer bits. Any suitable
  • the encoding process used to encode a particular audio segment may or may not apply
  • the audio segments may be generated by specialized hardware (e.g., a sound card) in either the compute server 200C (Fig. 2A) or the hybrid server 200H (Fig. 2B) .
  • the audio segment may be parametrized into speech parameters (e.g., LPC parameters) by the video game functional module 270, and the speech parameters can be redistributed to the destination client device (e.g., client device 120 or client device 120A) by the
  • the encoded audio created in the above manner is sent over the Internet 130.
  • the encoded audio input may be broken down and " formatted into packets, each having a header and a payload.
  • the header may carry an address of a client device associated with the participant for whom the audio generation process is being executed, while the payload may include the encoded audio.
  • the identity and/or version of the compression algorithm used to encode a given audio segment may be encoded in the content of one or more packets that convey the given segment. Other methods of transmitting the encoded audio will occur to those of skill in the art. .
  • Fig. 4B shows operation of the client device associated with a given participant, which may be client device 120 or client device 120A, by way of non-limiting example.
  • an encoded audio segment is received from the compute server 200C, the rendering server 200R or the hybrid server 200H (depending on the embodiment) .
  • the encoded audio is decoded in accordance with the decompression algorithm that is complementary to the compression algorithm used in the encoding process.
  • the identity or version of the compression algorithm used to encode the audio segment may be specified in the content of one or more packets that convey the audio segment.
  • the (decoded) audio segments are processed. This can include placing the decoded audio segments in a buffer, performing error correction, combining multiple successive waveforms, and so on.
  • the result can be a final sound to be presented to the user on a per-frame basis.
  • the final generated sound is output via the output mechanism of the client device.
  • the sound is played through a sound card or loudspeaker of the client device.
  • Menu extension processing is processing of adding a menu item of a new function (extended
  • the provided content is a game content, as described above.
  • the main process is a process of performing a series of processes for the game content by changing the game status in accordance with input of the client device input 140 received from the client device 120 and rendering and outputting a game screen
  • the main process is a process executed by the video game functional module 270 and the rendering functional module 280 described above.
  • Fig. 5A is a block diagram showing a module arrangement for execution of menu extension processing on the server side according to the embodiment of the present invention in accordance with the flow of processing and data.
  • the client device input 140 received via the Internet 130 is first checked by an input monitoring module 500.
  • the input monitoring module 500 first checks whether the input done is input for menu display, and determines whether the situation requires to execute processing of displaying a menu item for an extended function.
  • the input monitoring module 500 for example, always monitors the received client device input 140, and directly outputs the received client device input 140 to the video game functional module 270 until input for menu display is done.
  • the input monitoring module 500 monitors whether input for selecting the added menu item is done during the time display concerning the menu is included in the screen.
  • an extension processing module 510 Upon receiving an instruction to execute processing of displaying a menu item for an extended function from the input monitoring module 500, an extension processing module 510 executes various processes for adding the menu item for the extended function to the game screen generated by the main process. More specifically, the extension processing module 510 performs processing of causing a superimposed image generation module 520 to render the display item of the menu item or the like that is to be superimposed on the game screen.
  • extension processing module 510 executes the corresponding processing (extension processing) .
  • the superimposed image generation module 520 renders the display item of the menu item for the extended function and other display items, and
  • the superimposed image generation module 520 may
  • the mask data may represents the transparency level of each pixel of the superimposed image when superimposing the superimposed image on the game screen.
  • the data of the display item of the menu item and information such as a display position may be recorded in a recording device (not shown) in advance.
  • a composite module 530 performs composite processing of superimposing the superimposed image generated by the superimposed image generation module 520 on the game screen generated by processing in the main process, thereby generating a new game screen (composite screen) .
  • the composite module 530 outputs the generated composite screen to the video encoder 285. Even when no superimposed image needs to be generated, the composite module 530 receives input of the game screen generated by the main process. In this case, the game screen is directly output to the video encoder 285.
  • Processing corresponding to this flowchart can be implemented when the CPU 222 reads out a corresponding processing program stored in, for example, a recording device (not shown), loads the program to the RAM 230, and executes it independently of the main process.
  • step S601 the input monitoring module 500 determines whether the client device input 140 received from the client device 120 is input for menu display (normal menu display) . Upon determining that the client device input 140 received from the client device 120 is input for menu display (normal menu display) . Upon determining that the client device input 140 received from the client device 120 is input for menu display (normal menu display) . Upon determining that the client device input 140 received from the client device 120 is input for menu display (normal menu display) . Upon determining that the
  • client device input 140 is input for normal menu
  • the input monitoring module 500 transmits, to the extension processing module 510, an instruction to execute processing of displaying a menu item for an extended function, transfers the operation input of the client device input 140 to the main process, and
  • step S603. Upon determining that the client device input 140 is not input for normal menu display, the input monitoring module 500 transfers the operation input of the client device input 140 directly to the main process in step S602, and returns the process to step S601.
  • step S603 the extension processing module 510 causes the superimposed image generation module 520 to render a superimposed image including the menu item for the extended function and used to add the selected item to the normal menu display arranged on the game screen by processing of the main process.
  • the superimposed image can be as in Fig. 7B.
  • the image data of the menu item for the extended function data that combines the design and menu items arranged on the game screen in normal menu display is prepared in advance and recorded in a
  • the superimposed image generation module 520 upon receiving a superimposed image generation instruction, acquires the image data of the menu item and information of its arrangement position, and generates a superimposed image. Note that in this embodiment, a description will be made assuming that the superimposed image generation module 520 generates a superimposed image upon receiving an instruction from the extension processing module 510. However, the practice of the present invention is not limited to this. It should be appreciated that since the menu item for the extended function to be arranged in the superimposed image does not dynamically change, for example, the superimposed image itself may be recorded in the recording device in advance.
  • step S604 the extension processing module 510 causes the composite module 530 to composite the superimposed image generated by the superimposed image generation module 520 with the game screen generated by the rendering functional module 280 in correspondence with operation input received in the same frame and generate a composite image.
  • the composite module 530 superimposes, for example, pixels having colors other than those defined not to be superimposed out of the superimposed image at the same pixel positions of the game screen, thereby generating the composite image.
  • hatched pixels in Fig. 7B are handled as pixels to be made transparent at the time of
  • a new game screen including menu display with the additional menu item for the extended function as shown in Fig. 7C is generated.
  • step S605 the composite module 530 outputs the generated composite image to the video encoder 285 as the game screen, and terminates the menu extension processing .
  • the client device input 140 is received on a per-frame basis.
  • step S801 the input monitoring module 500 determines whether a menu item for an extended function or a display item displayed by execution of an extended function is included in a screen to be provided to the client device 120. That is, the input monitoring module 500 determines whether a superimposed image is being superimposed on the game screen currently
  • the input monitoring module 500 advances the process to step S802. Upon determining that no
  • the input monitoring module 500 advances the process to step S814.
  • step S802 the input monitoring module 500 determines whether the pointed position indicated by the received client device input 140 is included in a region where the menu item for the extended function is arranged. Upon determining that the pointed position is included in the region where the menu item for the extended function is arranged, the input monitoring module 500 advances the process to step S803. Upon determining that the pointed position is not included, the input monitoring module 500 advances the process to step S810.
  • step S803 the input monitoring module 500 determines whether the client device input 140 includes information of operation input to execute the function of the menu item. That is, the input monitoring module 500 determines whether the pointed position exists in the region where the menu item for the extended
  • operation input e.g., mouse click
  • the input monitoring module 500 Upon determining that the information of operation input to execute the function is included, the input monitoring module 500 advances the process to step S804. Upon determining that the information is not included, the input monitoring module 500 advances the process to step S807.
  • step S804 the input monitoring module 500 transfers the received client device input 140 not to the video game functional module 270 but to the
  • extension processing module 510 advances the process to step S805. Operation input to execute the function corresponding to the menu item for the
  • the extended function does not correspond to operation input to perform any processing in the main process.
  • the operation input may be determined as one to perform processing different from the extended function, and transition to an undesirable situation may occur so that, for example, the screen may transit to another screen, or a game that has paused may restart.
  • the input monitoring module 500 transfers the operation input not to the main process but only to the extension processing module 510. That is, a situation is apparently created in which a sub- process other than the main process temporarily seize authority to process operation input to execute the extended function.
  • step S805 the extension processing module 510 executes processing of the extended function corresponding to the menu item arranged at the pointed position.
  • the processing of the extended function can include above-described screen transition or a change in the display form of the menu item or another item caused by selection of the menu item.
  • the extension processing module 510 may control the video game functional module 270 to stop execution of processing in the main process.
  • the processing of the extended function may include processing for settings for an additional content element that is not implemented in the game content of the main process or settings of key
  • step S806 the superimposed image
  • the extension generation module 520 generates a superimposed image for the selected extended function under the control of the extension processing module 510.
  • the superimposed image for the selected extended function can be a setting window or a screen for a specific application, as shown in Fig. 9A.
  • the image for the extended function is an image to be superimposed on the game screen rendered by the rendering functional module 280 in the main process.
  • the image is not limited to this in the practice of the present invention.
  • the image for the extended function may be, for example, an image constructing the entire screen, and may be output as a screen to be provided to the client device 120 without being superimposed on the game screen, as will be described later.
  • step S803 upon determining in step S803 that information of operation input to execute the function is not included, the input monitoring module 500 transfers the client device input 140 to the video game functional module 270 and the extension processing module 510 in step S807, and advances the process to step S808.
  • step S808 a description will be made assuming that simple mouse-over on the menu item of the extended function leads to processing different from the extended function in the main process, and processing in the main process is not executed.
  • the input monitoring module 500 may transfer the client device input 140 only to the extension processing module 510 in this step .
  • step S808 the superimposed image
  • step S809 the composite module 530
  • step S802 upon determining in step S802 that the pointed position indicated by the client device input 140 is not included in the region where the menu item for the extended function is arranged, the input monitoring module 500 determines in step S810 whether the client device input 140 includes
  • the input monitoring module 500 determines ' whether the pointed position exists in the region where a normal menu item is arranged, and operation input to select the menu item has been performed.
  • the information of the position of each normal menu item may be recorded in a recording device in advance, or obtained by, for example, analyzing the image of the game screen before and after the display of the normal menu item and specifying the position where a display item of a predetermined shape is
  • the input monitoring module 500 advances the process to step S811. Upon determining that the information is not included, the input
  • monitoring module 500 advances the process to step S812.
  • step S811 the input monitoring module 500 transfers the received client device input 140 only to the video game functional module 270 and advances the process to step S813.
  • the input monitoring module 500 transfers the received client device input 140 only to the video game functional module 270 and advances the process to step S813.
  • monitoring module 500 notifies the extension processing module 510 that operation input to execute the function of the normal menu item has been done.
  • step S812 the input monitoring module 500 determines whether the client device input 140 includes information of operation input to end normal menu display. Upon determining that the information of operation input to end normal menu display is included, the input monitoring module 500 advances the process to step S813. Upon determining that the information is not included, the input monitoring module 500 advances the process to step S814.
  • step S813 the extension processing module 510 causes the superimposed image generation module 520 to stop generating and outputting a superimposed image.
  • the extension processing module 510 also causes the composite module 530 to stop executing composite processing. That is, in this step, the extension processing module 510 performs processing of
  • step S814 the input monitoring module 500 determines whether the client device input 140 includes operation input to end execution of the function of the normal menu item. That is, the input monitoring module 500 determines whether to superimpose the superimposed image (the menu item for the extended function) , whose superimposition has been stopped by execution of the function of the normal menu item, again in accordance with the end of display of the item displayed by executing the function. Upon determining that
  • the extension processing module 510 causes the superimposed image generation module 520 to render a superimposed image in which the menu item for the extended function is arranged, and advances the process to step S809.
  • a screen for a medium output 150 finally provided to the client device 120 can present display for function extension of a provided content to the player without a sense of incongruity.
  • the method of providing a user experience with extended functions when providing an existing content without altering the program of the content is not limited to this.
  • a method will be described in which a change in the main process caused by operation input is detected, and extended display is superimposed, thereby providing a user experience with extended functions, instead of allowing a sub-process to seize authority to process operation input.
  • Some game contents performs text display to present information of a result of processing based on done operation input, as shown in Fig. 10A.
  • Such text display is configured to display only predetermined information set at the time of development of a content. When different information other than the predetermined information is presented, the user experience can be extended.
  • a game screen shown in Fig. 10A is that of a so-called fighting game.
  • This screen includes the life gauge of each character. For example, assume a case where the total damage amount (life decrement) within a so-called combo period where a character is
  • Pieces of information of, for example, the combo period and a parameter representing life in the life gauge are managed by processing of the video game functional module 270 and stored in a predetermined storage area such as the RAM 230.
  • the input monitoring module 500 upon determining that the client device input 140 includes operation input to attack, the input monitoring module 500 notifies the extension processing module 510 of it, and the extension processing module 510 measures the decrement of the life parameter by the attack during the combo period.
  • the extension processing module 510 compares the life decrement value with the maximum life decrement value of the day managed for system users. When the life decrement value is larger than the maximum life
  • the extension processing module 510 causes the superimposed image generation module 520 to generate a superimposed image in which text display representing that the user has marked the maximum life decrement value of the day is arranged at a
  • the text display arranged in the superimposed image preferably uses the font used in the game.
  • the composite module 530 composites the superimposed image with the game screen generated by the main process to generate and output a composite image as shown in Fig. 10B under the control of the extension processing module 510. Note that the number of rows and the arrangement position of text display can change in accordance with the progress of the game content and may therefore be decided by monitoring parameters or analyzing the image of the game screen.
  • extension processing module 510 monitors intermediate data output by the main process or parameters and the like managed by the main process, extended information obtained by variously evaluating these pieces of information can be included in the screen of the content and provided.
  • a method of presenting information using a change in the life parameter as an evaluation target has been described as an example.
  • the evaluation target, evaluation method and information to be presented are not limited to these.
  • the system further includes an image analysis module 540 configured to analyze a game screen generated by the rendering functional module 280 in the main process, as shown in Fig. 5B.
  • a method of causing the image analysis module 540 to analyze a game screen to detect a change in parameters will be described. That is, instead of monitoring internally-managed parameter values, the method detects the state in the game content or the execution state of processing for the game content by analyzing the game screen.
  • a game screen as shown in Fig. 10A includes a life gauge or text display of a result of operation input, as described above.
  • the image analysis module 540 can detect the state in the game by detecting the change amount of the life gauge or performing text recognition based on the difference or correlation of the game screen between continuous frames. More specifically, when these pieces of information detected by the image analysis module 540 are transmitted to the extension processing module 510, the extension processing module 510 can perform, for example, evaluation of parameter changes as described above. That is, the extension processing module 510 can grasp the combo period or a given damage amount from the text recognition result. The extension processing module 510 can also grasp occurrence of the effect of giving an abrupt decrease in life from a change in the life gauge.
  • Information presentation is not limited to text display as described in the first modification, and may be done by, for example, superimposing
  • predetermined effect display on a portion where clash of characters takes place as shown in Fig. IOC.
  • This can be done by, for example, when the extension processing module 510 has received a detection result representing an abrupt decrease in life from the image analysis module 540, causing the superimposed image generation module 520 to generate a superimposed image in which effect display is arranged at a portion where a motion vector having a scalar of a predetermined value or more is detected between preceding and
  • the extension processing module 510 decides whether to do information presentation by image analysis of the game screen.
  • the practice of the present invention is not limited to this. It may be decided whether to do information presentation based on, for example, information such as the number of command inputs or a variation in an analog value from the history of the client device input 140 received by the input monitoring module 500. Alternatively, it may be decided whether to do information presentation based on, for example, whether an audio signal output in accordance with the game screen includes a sound of an amplitude of a predetermined value or more. Otherwise, it may be decided whether to do information
  • a sub-game may be provided, which arranges display items such as an icon in a region of interest detected in a game screen by a predetermined method, and the user collects the items by selecting them during transition of the game screen.
  • the client device input 140 includes operation input to select a display item
  • the input monitoring module 500 transfers the operation input not to the video game functional module 270 but to the extension processing module 510, and the extension processing module 510 performs processing such as score calculation for the item collection.
  • the extension processing module 510 may cause the superimposed image generation module 520 to generate a superimposed image as shown in Fig.
  • a predetermined rendering effect may be provided when generating the superimposed image so as to integrate the display items to be arranged with the atmosphere (illumination, shading, reflection and camera angle) of rendered objects on the game which exist at the arrangement positions in the game screen.
  • the region of interest may be specified by causing the image analysis module 540 shown in Fig. 5B to detect, for example, a region of the game screen without a change between continuous frames or a region where the edge components exhibit an intensity equal to or more than a threshold, and the contrast ratio to the peripheral region is high.
  • the region of interest may be specified by causing the image analysis module 540 to detect a predetermined image pattern included in the game screen. Various other methods can appropriately be used to detect the region of interest.
  • the information processing apparatus and the controlling method according to the present invention are realizable by a program executing the methods on one or more computers.
  • the program is
  • providable/distributable by being stored on a computer- readable storage medium or through an electronic communication line.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Optics & Photonics (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Processing Or Creating Images (AREA)
  • Image Generation (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

An information processing apparatus receives operation input for a content, and generates a first image corresponding to the content by executing a first program for the content in accordance with the received operation input. The apparatus also generates a second image to be added to the first image by executing a second program different from the first program, and outputs a composite image obtained by compositing the first image and the second image. In a case where the operation input to a region according to the second image in the composite image is received, the apparatus controls not to execute the first program in accordance with the operation input.

Description

DESCRIPTION
TITLE OF INVENTION INFORMATION PROCESSING APPARATUS, CONTROL METHOD AND
PROGRAM
TECHNICAL FIELD
[0001] The present invention relates to an
information processing apparatus, a control method, and a program, and particularly to a technique of extending the function of a content by adding additional displays
BACKGROUND ART
[0002] Recent development of information
communication technologies using networks such as the Internet has resulted in service providing to customers via the networks in various fields. Even contents that are conventionally provided as a service by executing an application" in a client device such as a PC operated by a customer are handled by some of these services so that processing for execution is performed by a server on a network, and a corresponding screen is rendered and transmitted to a client device. That is, since most processes to be performed in the client device are alternatively executed on the server side, the client device needs to perform only processing of transmitting operation input done by the user to the server and processing of displaying a screen received from the server .
[0003] As one of the services of such providing form, a game screen rendered in a server is provided to a client device via a network, like so-called cloud gaming. In particular, a game content that generates fine graphics requires sufficient rendering performance of a client device. However, cloud gaming allows even a user who does not have a client device with
sufficient rendering performance to play the same game as in a device having sufficient rendering performance.
[0004] In such a cloud service that renders a screen of a content on the server side and provides it, an application program and the like are preferably
optimized and configured for execution on the server. However, when providing an already developed content by the cloud service, optimizing and reconfiguring the program for cloud is not realistic because of
additional cost. In addition, since sources such as a program may be absent, there is a possibility that a program for content providing needs to be developed substantially from the beginning. Furthermore, if a work of, for example, reediting a program arises to add a process (function) of little relationship with the main process to an already released content, works such as debug need to be performed accordingly, and a time is required for release of the function. SUMMARY OF INVENTION
[0005] The present invention was made in view of such problems in the conventional technique. An aspect of the present invention provides a user experience with extended functions when providing an existing content without altering, the program of the content.
[0006] The present invention in its first aspect provides an information processing apparatus
comprising: receiving means for receiving operation input for a content; first generation means for
generating a first image corresponding to the content by executing a first program for the content in
accordance with the operation input received by the receiving means; second generation means for generating a second image to be added to the first image by executing a second program different from the first program; output means for outputting a composite image obtained by compositing the first image and the second image; and control means for, in a case where the receiving means receives the operation input to a region according to the second image in the composite image, controlling not to cause the first generation means to execute the first program in accordance with the operation input.
[0007] The present invention in its second aspect provides an information processing apparatus
comprising: receiving means for receiving operation input for a content; first generation means for generating a first image corresponding to the content by executing a first program for the content in accordance with the operation input received by the receiving means; monitor means for monitoring a predetermined parameter that changes during execution of the first program; second generation means for generating a second image to be added to the first image by executing a second program different from the first program in a case where the predetermined
parameter meets a predetermined condition; and output means for outputting a composite image obtained by compositing the first image and the second image.
[0008] The present invention in its third aspect provides an information processing apparatus
comprising: receiving means for receiving operation input for a content; first generation means for
generating a first image corresponding to the content by executing a first program for the content in
accordance with the operation input received by the receiving means; analysis means for analyzing the first image and detecting whether an execution state of the content meets a predetermined condition; second
generation means for generating a second image to be added to the first image by executing a second program different from the first program in a case where the execution state meets the predetermined condition; and output means for outputting a composite image obtained by compositing the first image and the second image.
[0009] The present invention in its fourth aspect provides an information processing apparatus
comprising: receiving means for receiving operation input for a content; first generation means for
generating a first image corresponding to the content by executing a first program for the content in
accordance with the operation input received by the receiving means; analysis means for analyzing the first image and deciding a position where a display item is to be arranged; second generation means for generating a second image to be added to the first image, in which the display item is arranged at the position decided by the analysis means, by executing a second program different from the first program; and output means for outputting a composite image obtained by compositing the first image and the second image.
[0010] Further features of the present invention will become apparent from the following description of exemplary embodiments (with reference to the attached drawings) .
BRIEF DESCRIPTION OF DRAWINGS
[0011] Fig. 1 is a block diagram of a cloud-based video game system architecture, according to a non- limiting embodiment of the present invention. [0012] Fig. 2A is a block diagram showing various physical components of the architecture of Fig. 1, according to a non-limiting embodiment of the present invention .
[0013] Fig. 2B is a variant of Fig. 2A.
[0014] Fig. 2C is a block diagram showing various functional modules of the architecture of Fig. 1, which can be implemented by the physical components of Figs. 2A or 2B.
[0015] Figs. 3A to 3C are flowcharts showing
execution of a set of processes carried out during execution of a video game, in accordance with non- limiting embodiments of the present invention.
[0016] Figs. 4A and 4B are flowcharts showing
operation of a client device to process received video and audio, respectively, in accordance with non- limiting embodiments of the present invention.
[0017] Figs. 5A and 5B are block diagram showing a functional arrangement on a server side according to an embodiment or a modification of the present invention.
[0018] Fig. 6 is a flowchart exemplary showing menu extension processing performed on the server side according to the embodiment of the present invention.
[0019] Figs. 7A, 7B and 7C are views exemplary showing the structures of images before and after composite processing in the menu extension processing according to the embodiment of the present invention. [0020] Figs. 8A and 8B are flowcharts exemplary showing display update processing performed on the server side according to the embodiment of the present invention .
[0021] ■ Figs. 9A and 9B are views exemplary showing the structures of superimposed images generated by the display update processing.
[0022] Figs. 10A, 10B and IOC are views exemplary showing the structures of images before and after composite processing according to a modification of the present invention.
[0023] Figs. 11A and 11B are views showing the structures of composite images according to the third modification of the present invention.
DESCRIPTION OF EMBODIMENTS
[0024] I. Cloud Gaming Architecture
Fig. 1 schematically shows a cloud-based video game system architecture according to a non-limiting embodiment of the present invention. The architecture includes client devices 120, 120A connected to a cloud gaming server system 100 over a data network such as the Internet 130. Each of the client devices 120, 120A may connect to the Internet 130 in any suitable manner, including over a respective local access network (not shown) . The cloud gaming server system 100 may also connect to the Internet 130 over a local access network (not shown) , although the server system 100 may connect directly to the Internet 130 without the intermediary of a local access network. Connections between the cloud gaming server system 100 and one or more of the client devices 120, 120A may comprise one or more channels. These channels can be made up of physical and/or logical links, and may travel over a variety of physical media, including radio frequency, fiber optic, free-space optical, coaxial and twisted pair. The channels may abide by a protocol such as UDP or TCP/IP. Also, one or more of the channels may be supported a virtual private network (VPN) . In some embodiments, one or more of the connections may be session-based.
[0025] The cloud gaming server system 100 enables users of the client devices 120, 120A to play video games, either individually (i.e., a single-player video game) or in groups (i.e., a multiplayer video game) . Non-limiting examples of video games may include games that are played for leisure, education and/or sport. A video game may but need not offer participants the possibility of monetary gain. Although only two client devices 120, 120A are shown, it should be appreciated that the number of client devices in the cloud-based video game system architecture is not particularly limited .
[0026] A user of one of the client devices 120, 120A may register with the cloud gaming server system 100 as a participant in a video game. The user may register as a "player", and will have the opportunity to control a character, avatar, race car, cockpit, etc. within a virtual world maintained by the video game. In the case of a multi-player video game, the virtual world is shared by two or more players, and one player's
gameplay may affect that of another. In some
embodiments, a user of one of the client devices 120, 120A may register as a non-player "spectator", whereby such users will observe players' gameplay but otherwise do not control active characters in the game. Unless otherwise indicated, where the term "participant" is used, it is meant to apply equally to players and spectators.
[0027] The server system 100 may include one or more computing resources, including one or more game servers and one or more account servers. The game servers and the account servers may be embodied in the same
hardware or they may be different servers that are connected via a communication link, including possibly over the Internet 130. In the following description, they are treated as separate servers merely in the interest of simplicity.
[0028] A game server interacts with players in the course of a game, while an account server interacts with players outside the game environment. Thus, for example, the account server may be configured for logging a prospective player into a game portal,
tracking the player's connectivity over the Internet, and responding to the player's commands to launch, join, exit or terminate an instance of a game, among several non-limiting functions. To this end, the account server may host or have access to a participant
database 10 that stores account information about various participants and client devices 120, 120A, such as identification data, financial data, location data, demographic data, connection data and the like. The participant database 10 which can be part of the cloud gaming server system 100 or situated remotely therefrom. The game server 110 may also have access to the
participant database 10 that stores the above mentioned account information, as this information may be used, for influencing the way in which the video game
progresses .
[0029] The configuration of any given one of the client devices 120, 120A is not particularly limited. In some embodiments, one or more of the client devices 120, 120A may be, for example, a personal computer (PC), a home game machine (console such as XBOX™, PS3™, ii™, etc.), a portable game machine, a smart television, a set-top box (STB) , etc. In other embodiments, one or more of the client devices 120, 120A may be a
communication or computing device such as a mobile phone, a personal digital assistant (PDA), or a tablet. [0030] Any given one of the client devices 120, 120A may be equipped with one or more input devices (such as a touch screen, a keyboard, a game controller, a
joystick, etc.) to allow users of the given client device to provide input and participate in a video game. In other embodiments, the user may produce body motion or may wave an external object; these movements are detected by a camera or other sensor (e.g., Kinect™) , while software operating within the given client device attempts to correctly guess whether the user intended to provide input to the given client device and, if so, the nature of such input. The given client device translates the received user inputs and detected user movements into "client device input", which is sent to the cloud gaming server system 100 over the Internet 130. In the illustrated embodiment, client device 120 produces client device input 140, while client device 120A produces client device input 140A.
[0031] The cloud gaming server system 100 processes the client device input 140, 140A received from the various client devices 120, 120A and generates "media output" for the various client devices 120, 120A. The media output may include a stream of encoded video data (representing images when displayed on a screen) and audio data (representing sound when played via a
loudspeaker) . The media output is sent over the
Internet 130 in the form of packets. Packets destined for a particular one of the client devices 120, 120A may be addressed in such a way as to be routed to that device over the Internet 130. Each of the client devices 120, 120A may include circuitry for buffering and processing the media output in the packets received from the cloud gaming server system 100, as well as a display for displaying images and a transducer (e.g., a loudspeaker) for outputting audio. Additional output devices may also be provided, such as an electromechanical system to induce motion.
[0032] It should be appreciated that a stream of video data can be divided into "frames". The term "frame" as used herein does not require the existence of a one-to-one correspondence between frames of video data and images represented by the video data. That is to say, while it is possible for a frame of video data to contain data representing a respective displayed image in its entirety, it is also possible for a frame of video data to contain data representing only part of an image, and for the image to in fact require two or more frames in order to be properly reconstructed and displayed. By the same token, a frame of video data may contain data representing more than one complete image, such that N images may be represented using M frames of video data, where M < N.
[0033] II. Cloud Gaming Server System 100 (Distributed Architecture ) Fig. 2A shows one possible non-limiting physical arrangement of components for the cloud gaming server system 100. In this embodiment, individual servers within the cloud gaming server system 100 are
configured to carry out specialized functions. For example, a compute server 200C may be primarily
responsible for tracking state changes in a video game based on user input, while a rendering server 200R may be primarily responsible for rendering graphics (video data) .
[0034] For the purposes of the presently described example embodiment, both client device 120 and client device 120A are assumed to be participating in the video game, either as players or spectators. However, it should be understood that in some cases there may be a single player and no spectator, while in other cases there may be multiple players and a single spectator, in still other cases there may be a single player and multiple spectators and in yet other cases there may be multiple players and multiple spectators.
[0035] For the sake of simplicity, the following description refers to a single compute server 200C connected to a single rendering server 200R. However, it should be appreciated that there may be more than one rendering server 200R connected to the same compute server 200C, or more than one compute server 200C connected to the same rendering server 200R. In the case where there are multiple rendering servers 200R, these may be distributed over any suitable geographic area .
[0036] As shown in the non-limiting physical
arrangement of components in Fig. 2A, the compute server 200C comprises one or more central processing units (CPUs) 220C, 222C and a random access memory (RAM) 230C. The CPUs 220C, 222C can have access to the RAM 230C over a communication bus architecture, for example. While only two CPUs 220C, 222C are shown, it should be appreciated that a greater number of CPUs, or only a single CPU, may be provided in some example implementations of the compute server 200C. The compute server 200C also comprises a network interface component (NIC) 210C2, where client device input is received over the Internet 130 from each of the client devices participating in the video game. In the presently described example embodiment, both client device 120 and client device 120A are assumed to be participating in the video game, and therefore the received client device input may include client device input 140 and client device input 140A.
[0037] The compute server 200C further comprises another network interface component (NIC) 210C1, which outputs a sets of rendering commands 204. The sets of rendering commands 204 output from the compute server 200C via the NIC 210C1 can be sent to the rendering server 200R. In one embodiment, the compute server 200C can be connected directly to the rendering server 200R. In another embodiment, the compute server 200C can be connected to the rendering server 200R over a network 260, which can be the Internet 130 or another network. A virtual private network (VPN) may be
established between the compute server 200C and the rendering server 200R over the network 260.
[0038] At the rendering server 200R, the sets of rendering commands 204 sent by the compute server 200C are received at a network interface component (NIC) 210R1 and are directed to one or more CPUs 220R, 222R. The CPUs 220R, 222R are connected to graphics
processing units (GPUs) 240R, 250R. By way of non- limiting example, GPU 240R may include a set of GPU cores 242R and a video random access memory (VRAM) 246R. Similarly, GPU 250R may include a set of GPU cores 252R and a video random access memory (VRAM) 256R. Each of the CPUs 220R, 222R may be connected to each of the GPUs 240R, 250R or to a subset of the GPUs 240R, 250R. Communication between the CPUs 220R, 222R and the GPUs 240R, 250R can be established using, for example, a communications bus architecture. Although only two CPUs and two GPUs are shown, there may be more than two CPUs and GPUs, or even just a single CPU or GPU, in a specific example of implementation of the rendering server 200R. [0039] The CPUs 220R, 222R cooperate with the GPUs 240R, 250R to convert the sets of rendering commands 204 into a graphics output streams, one for each of the participating client devices. In the present,
embodiment, there are two graphics output streams 206, 206A for the client devices 120, 120A, respectively. This will be described in further detail later on. The rendering server 200R comprises a further network interface component (NIC) 210R2, through which the graphics output streams 206, 206A are sent to the client devices 120, 120A, respectively.
[0040] III. Cloud Gaming Server System 100 (Hybrid
Architecture)
Fig. 2B shows a second possible non-limiting physical arrangement of components for the cloud gaming server system 100. In this embodiment, a hybrid server 200H is responsible both for tracking state changes in a video game based on user input, and for rendering graphics (video data) .
[0041] As shown in the non-limiting physical
arrangement of components in Fig. 2B, the hybrid server 200H comprises one or more central processing units
(CPUs) 220H, 222H and a random access memory (RAM) 230H. The CPUs 220H, 222H can have access to the RAM 230H over a communication bus architecture, for example.
While only two CPUs 220H, 222H are shown, it should be appreciated that a greater number of CPUs, or only a single CPU, may be provided in some example
implementations of the hybrid server 200H. The hybrid server 200H also comprises a network interface
component (NIC) 210H, where client device input is received over the Internet 130 from each of the client devices participating in the video game. In the presently described example embodiment, both client device 120 and client device 120A are assumed to be participating in the video game, and therefore the received client device input may include client device input 140 and client device input 140A.
[0042] In addition, the CPUs 220H, 222H are connected to a graphics processing units (GPUs) 240H, 250H. By way of non-limiting example, GPU 240H may include a set of GPU cores 242H and a video random access memory
(VRAM) 246H. Similarly, GPU 250H may include a set of GPU cores 252H and a video random access memory (VRAM) 256H. Each of the CPUs 220H, 222H may be connected to each of the GPUs 240H, 250H or to a subset of the GPUs 240H, 250H. Communication between the CPUs 220H, 222H and the GPUs 240H, 250H can be established using, for example, a communications bus architecture. Although only two CPUs and two GPUs are shown, there may be more than two CPUs and GPUs,, or even just a single CPU or GPU, in a specific example of implementation of the hybrid server 200H.
[0043] The CPUs 220H, 222H cooperate with the GPUs 240H, 250H to convert the sets of rendering commands 204 into graphics output streams, one for each of the participating client devices. In this embodiment, there are two graphics output streams 206, 206A for the participating client devices 120, 120A, respectively. The graphics output streams 206, 206A are sent to the client devices 120, 120A, respectively, via the NIC 210H.
[0044] IV. Cloud Gaming Server System 100
(Functionality Overview)
With additional reference now to Fig. 2C, the above-described physical components of the compute server 200C and the rendering server 200R (in Fig. 2A) and/or of the hybrid server 200H (in Fig. 2B) implement a set of functional modules, including a video game functional module 270, a rendering functional module 280 and a video encoder 285. According to the non- limiting embodiment of Fig. 2A, the video game
functional module 270 is implemented by the compute server 200C, while the rendering functional module 280 and the video encoder 285 are implemented by the rendering server 200R. According to the non-limiting embodiment of Fig. 2B, the hybrid server 200H
implements the video game functional module 270, the rendering functional module 280 and the video encoder 285.
[0045] The present example embodiment discusses a single video game functional module 270 for simplicity of illustration. However, it should be noted that in an actual implementation of the cloud gaming server system 100, many video game functional modules similar to the video game functional module 270 would be executed in parallel. Thus, the cloud gaming server system 100 could support multiple independent
instantiations of the same video game, or multiple different video games, simultaneously. Also, it should be noted that the video games can be single-player video games or multi-player games of any type.
[0046] The video game functional module 270 may be implemented by certain physical components of the compute server 200C (in Fig. 2A) or of the hybrid server 200H (in Fig. 2B) . Specifically, the video game functional module 270 can be encoded as computer- readable instructions that are executable by a CPU (such as the CPUs 220C, 222C in the compute server 200C or the CPUs 220H, 222H in the hybrid server 200H) . The instructions can be tangibly stored in the RAM 230C (in the compute server 200C) of the RAM 230H (in the hybrid server 200H) or in another memory area, together with constants, variables and/or other data used by the video game functional module 270. In some embodiments, the video game functional module 270 may be executed within the environment of a virtual machine that may be supported by an operating system that is also being executed by a CPU (such as the CPUs 220C, 222C in the compute server 200C or the CPUs 220H, 222H in the hybrid server 200H) .
[0047] The rendering functional module 280 may be implemented by certain physical components of the rendering server 200R (in Fig. 2A) or of the hybrid server 200H (in Fig. 2B) . In an embodiment, the rendering functional module 280 may take up one or more GPUs (240R, 250R in Fig. 2A, 240H, 250H in Fig. 2B) and may or may not utilize CPU resources.
[0048] The video encoder 285 may be implemented by certain physical components of the rendering server 200R (in Fig. 2A) or of the hybrid server 200H (in Fig. 2B) . Those skilled in the art will appreciate that there are various ways in which to implement the video encoder 285. In the embodiment of Fig. 2A, the video encoder 285 may be implemented by the CPUs 220R, 222R and/or by the GPUs 240R, 250R. In the embodiment of Fig. 2B, the video encoder 285 may be implemented by the CPUs 220H, 222H and/or by the GPUs 240H, 250H. In yet another embodiment, the video encoder 285 may be implemented by a separate encoder chip (not shown) .
[0049] In operation, the video game functional module 270 produces the sets of rendering commands 2Ό4, based on received client device input. The received client device input may carry data (e.g., an address)
identifying the video game functional module for which it is destined, as well as data identifying the user and/or client device from which it originates. Since the users of the client devices 120, 120A are
participants in the video game (i.e., players or
spectators), the received client device input includes the client device input 140, 140A received from the client devices 120, 120A.
[0050] Rendering commands refer to commands which can be used to instruct a specialized graphics processing unit (GPU) to produce a frame of video data or a
sequence of frames of video data. Referring to Fig. 2C, the sets of rendering commands 204 result in the
production of frames of video data by the rendering functional module 280. The images represented by these frames change as a function of responses to the client device input 140, 140A that are programmed into the video game functional module 270. For example, the video game functional module 270 may be programmed in such a way as to respond to certain specific stimuli to provide the user with an experience of progression
(with future interaction being made different, more challenging or more exciting) , while the response to certain other specific stimuli will provide the user with an experience of regression or termination.
Although the instructions for the video game functional module 270 may be fixed in the form of a binary
executable file, the client device input 140, 140A is unknown until the moment of interaction with a player who uses the corresponding client device 120, 120A. As a result, there can be a wide variety of possible outcomes, depending on the specific client device input that is provided. This interaction between
players/spectators and the video game functional module 270 via the client devices 120, 120A can be referred to as "gameplay" or "playing a video game".
[0051] The rendering functional module 280 processes the sets of rendering commands 204 to create multiple video data streams 205. Generally, there will be one video data stream per participant (or, equivalently, per client device) . When performing rendering, data for one or more objects represented in three- dimensional space (e.g., physical objects) or two- dimensional space (e.g., text) may be loaded into a cache memory (not shown) of a particular GPU 240R, 250R, 240H, 250H. This data may be transformed by the GPU 240R, 250R, 240H, 250H into data representative of a two-dimensional image, which may be stored in the appropriate VRAM 246R, 256R, 246H, 256H. As such, the VRAM 246R, 256R, 246H, 256H may provide temporary storage of picture element (pixel) values for a game screen .
[0052] The video encoder 285 compresses and encodes the video data in each of the video data streams 205 into a corresponding stream of compressed/encoded video data. The resultant streams of compressed/encoded video data, referred to as graphics output streams, are produced on a per-client-device basis. In the present example embodiment, the video encoder 285 produces graphics output stream 206 for client device 120 and graphics output stream 206A for client device 120A. Additional functional modules may be provided for formatting the video data into packets so that they can be transmitted over the Internet 130. The video data in the video data streams 205 and the
compressed/encoded video data within a given graphics output stream may be divided into frames.
[0053]V. Generation of Rendering Commands
Generation of rendering commands by the video game functional module 270 is now described in greater detail with reference to Figs. 2C, 3A and 3B.
Specifically, . execution of the video game functional module 270 involves several processes, including a main game process 300A and one or more graphics control processes 300B, which are described herein below in greater detail.
[0054]Main Game Process
A first process, referred to as the main game process, is described with reference to Fig. 3A. The main game process 300A executes continually. As part of the main game process 300A, there is provided an action 310A, during which client device input may be received. If the video game is a single-player video game without the possibility of spectating, then client device input (e.g., client device input 140) from a single client device (e.g., client device 120) is received as part of action 310A. If the video game is a multi-player video game or is a single-player video game with the possibility of spectating, then the client device input (e.g., the client device input 140 and 140A) from one or more client devices (e.g., the client devices 120 and 120A) may be received as part of action 310A.
[0055] By way of non-limiting example, the input from a given client device may convey that the user of the given client device wishes to cause a character under his or her control to move, jump, kick, turn, swing, pull, grab, etc. Alternatively or in addition, the input from the given client device may convey a menu selection made by the user of the given client device in order to change one or more audio, video or gameplay settings, to load/save a game or to create or join a network session. Alternatively or in addition, the input from the given client device may convey that the user of the given client device wishes to select a particular camera view (e.g., first-person or third- person) or reposition his or her viewpoint within the virtual world.
[0056] At action 320A, the game state may be updated based at least in part on the client device input received at action 310A and other parameters. Updating the game state may involve the following actions:
Firstly, updating the game state may involve updating certain properties of the participants (player or spectator) associated with the client devices from which the client device input may have been received. These properties may be stored in the participant database 10. Examples of participant properties that may be maintained in the participant database 10 and updated at action 320A can include a camera view selection (e.g., 1st person, 3rd person), a mode of play, a selected audio or video setting, a skill level, a customer grade (e.g., guest, premium, etc.).
Secondly, updating the game state may involve updating the attributes of certain objects in the virtual world based on an interpretation of the client device input. The objects whose attributes are to be updated may in some cases be represented by two- or three-dimensional models and may include playing characters, non-playing characters and other objects. In the case of a playing character, attributes that can be updated may include the object's position, strength, weapons/armor, lifetime left, special powers,
speed/direction (velocity) , animation, visual effects, energy, ammunition, etc. In the case of other objects (such as background, vegetation, buildings, vehicles, score board, etc.), attributes that can be updated may include the object's position, velocity, animation, damage/health, visual effects, textual content, etc.
[0057] It should be appreciated that parameters other than client device input can influence the above
properties (of participants) and attributes (of virtual world objects). For example, various timers (such as elapsed time, time since a particular event, virtual time of day, total number of players, a participant's geographic location, etc.) can have an effect on
various aspects of the game state.
[0058] Once the game state has been updated further to execution of action 320A, the main game process 300A returns to action 310A, whereupon new client device input received since the last pass through the main game process is gathered and processed.
[0059] Graphics Control Process
A second process, referred to as the graphics control process, is now described with reference to Fig. 3B. The graphics control process 300B may execute continually, and there may be multiple separate
graphics control processes 300B, each of which results in a respective one of the sets of rendering commands 204. In the case of a single-player video game without the possibility of spectating, there is only one player and therefore only one resulting set of rendering commands 204, and thus the graphics control process 300B may execute as an extension of the main game process 300A described above. In the case of a multi- player video game, multiple distinct sets of rendering commands need to be generated for the multiple players, and therefore multiple graphics control processes 300B may execute in parallel. In the case of a single- player game with the possibility of spectating, there may again be only a single set of rendering commands 204, and therefore a single graphics control process 300B may execute in the video game functional module 270, but the resulting video data stream may be
duplicated for the spectators by the rendering
functional module 280. Of course, these are only examples of implementation and are not to be taken as limiting .
[0060] At action 310B of the graphics control process 300B for a given participant requiring a distinct video data stream, the video game functional module 270 determines the objects to be rendered for the given participant. This action can include identifying the following types of objects:
Firstly, this action can include identifying those objects from the virtual world that are in the "game screen rendering range" (also known as a "scene") for the given participant. The game screen rendering range includes the portion of the virtual world that would be "visible" from the perspective of the given participant's camera. This depends on the position and orientation of that camera relative to the objects in the virtual world. In a non-limiting example of implementation of action 310B, a frustum can be applied to the virtual world, and the objects within that frustum are retained or marked. The frustum has an apex which is situated at the location of the given participant's camera and has a directionality also defined by the directionality of that camera.
Secondly, this action can include identifying additional objects that do not appear in the virtual world, but which nevertheless are to be rendered for the given participant. For example, these additional objects may include textual messages, graphical
warnings and dashboard indicators, to name a few non- limiting possibilities.
[0061] At action 320B, the video game functional module 270 generates a set of commands for transforming rendering into graphics (video data) the objects that were identified at action 310B. Rendering may refer to the transformation of 3-D or 2-D coordinates of an object or group of objects into data representative of a displayable image, in accordance with the viewing perspective and prevailing lighting conditions. This can be achieved using any number of different
algorithms and techniques, for example as described in "Computer Graphics and Geometric Modelling: Implementation & Algorithms", Max K. Agoston,
Springer-Verlag London Limited, 2005, hereby
incorporated by reference herein.
[ 0062 ] At action 330B, the rendering commands
generated at action 320B are output to the rendering functional module 280. This may involve packetizing the generated rendering commands into a set of
rendering commands 204 that is sent to the rendering functional module 280.
[ 0063 ] Those skilled in the art will appreciate that multiple instantiations of the graphics control process 300B described above may be executed, resulting in multiple sets of rendering commands 204.
[0064 ] VI. Generation of Graphics Output
The rendering functional module 280 interprets the sets of rendering commands 204 and produces
multiple video data streams 205, one for each
participating client device. Rendering may be achieved by the GPUs 240R, 250R, 240H, 250H under control of the CPUs 220R, 222R (in Fig. 2A) or 220H, 222H (in Fig. 2B) . The rate at which frames of video data are produced for a participating client device may be referred to as the frame rate.
[ 0065 ] In an embodiment where there are N
participants, there may be N sets of rendering commands 204 (one for each participant) and also N video data streams 205 (one for each participant). In that case, rendering functionality is not shared among the
participants. However, the N video data streams 205 may also be created from sets of rendering commands 204 (where M < N) , such that fewer sets of rendering commands need to be processed by the rendering
functional module 280. In that case, the rendering functional unit 280 may perform sharing or duplication in order to generate a larger number of video data streams 205 from a smaller number of sets of rendering commands 204. Such sharing or duplication may be prevalent when multiple participants (e.g., spectators) desire to view the same camera perspective. Thus, the rendering functional module 280 may perform functions such as duplicating a created video data stream for one or more spectators.
[0066] Next, the video data in each of the video data streams 205 are encoded by the video encoder 285, resulting in a sequence of encoded video data
associated with each client device, referred to as a graphics output stream. In the example embodiments of Figs. 2A to 2C, the sequence of encoded video data destined for client device 120 is referred to as graphics output stream 206, while the sequence of encoded video data destined for client device 120A is referred to as graphics output stream 206A.
[0067] The video encoder 285 can be a device (or set of computer-readable instructions) that enables or - 3.1 - carries out or defines a video compression or
decompression algorithm for digital video. Video compression transforms an original stream of digital image data (expressed in terms of pixel locations, color values, etc.) into an output stream of digital image data that conveys substantially the same
information but using fewer bits. Any suitable
compression algorithm may be used. In addition to data compression, the encoding process used to encode a particular frame of video data may or may not involve cryptographic encryption.
[0068] The graphics output streams 206, 206A created in the above manner are sent over the Internet 130 to the respective client devices. By way of non-limiting example, the graphics output streams may be segmented and formatted into packets, each having a header and a payload. The header of a packet containing video data for a given participant may include a network address of the client device associated with the given
participant, while the payload may include the video data, in whole or in part. In a non-limiting
embodiment, the identity and/or- version of the
compression algorithm used to encode certain video data may be encoded in the content of one or more packets that convey that video data. Other methods of
transmitting the encoded video data will occur to those of skill in the art. [0069] While the present description focuses on the rendering of video data representative of individual 2- D images, the present invention does not exclude the possibility of rendering video data representative of multiple 2-D images per frame to create a 3-D effect.
[0070] VII. Game Screen Reproduction at Client Device
Reference is now made to Fig. 4A, which shows operation of the client device associated with a given participant, which may be client device 120 or client device 120A, by way of non-limiting example.
[0071] At action 410A, a graphics output stream (e.g., 206, 206A) is received over the Internet 130 from the rendering server 200R (Fig. 2A) or from the hybrid server 200H (Fig. 2B) , depending on the embodiment.
The received graphics output stream comprises
compressed/encoded of video data which may be divided into frames.
[0072] At action 420A, the compressed/encoded frames of video data are decoded/decompressed in accordance with the decompression algorithm that is complementary to the encoding/compression algorithm used in the encoding/compression process. In a non-limiting embodiment, the identity or version of the
encoding/compression algorithm used to encode/compress the video data may be known in advance. In other embodiments, the identity or version of the
encoding/compression algorithm used to encode the video data may accompany the video data itself.
[ 0073 ] At action 430A, the (decoded/decompressed) frames of video data are processed. This can include placing the decoded/decompressed frames of video data in a buffer, performing error correction, reordering and/or combining the data in multiple successive frames, alpha blending, interpolating portions of missing data, and so on. The result can be video data representative of a final image to be presented to the user on a per- frame basis.
[ 0074 ] . At action 440A, the final image is output via the output mechanism of the client device. For example, a composite video frame can be displayed on the display of the client device.
[0075 ] VIII. Audio Generation
A third process, referred to as the audio
generation process, is now described with reference to Fig. 3C. The audio generation process executes
continually for each participant requiring a distinct audio stream. In one embodiment, the audio generation process may execute independently of the graphics control process 300B. In another embodiment, execution of the audio generation process and the graphics
control process may be coordinated.
[ 0076] At action 310C, the video game functional module 270 determines the sounds to be produced.
Specifically, this action can include identifying those sounds associated with objects in the virtual world that dominate the acoustic landscape, due to their volume (loudness) and/or proximity to the participant within the virtual world.
[0077] At action 320C, the video game functional module 270 generates an audio segment. The duration of the audio segment may span the duration of a video frame, although in some embodiments, audio segments may be generated less frequently than video frames, while in other embodiments, audio segments may be generated more frequently than video frames.
[0078] At action 330C, the audio segment is encoded, e.g., by an audio encoder, resulting in an encoded audio segment. The audio encoder can be a device (or set of instructions) that enables or carries out or defines an audio compression or decompression algorithm. Audio compression transforms an original stream of digital audio (expressed as a sound wave changing in amplitude and phase over time) into an output stream of digital audio data that conveys substantially the same information but using fewer bits. Any suitable
compression algorithm may be used. In addition to audio compression, the encoding process used to encode a particular audio segment may or may not apply
cryptographic encryption.
[0079] It should be appreciated that in some
embodiments, the audio segments may be generated by specialized hardware (e.g., a sound card) in either the compute server 200C (Fig. 2A) or the hybrid server 200H (Fig. 2B) . In an alternative embodiment that may be applicable to the distributed arrangement of Fig. 2A, the audio segment may be parametrized into speech parameters (e.g., LPC parameters) by the video game functional module 270, and the speech parameters can be redistributed to the destination client device (e.g., client device 120 or client device 120A) by the
rendering server 200R.
[0080] The encoded audio created in the above manner is sent over the Internet 130. By way of non-limiting example, the encoded audio input may be broken down and" formatted into packets, each having a header and a payload. The header may carry an address of a client device associated with the participant for whom the audio generation process is being executed, while the payload may include the encoded audio. In a non- limiting embodiment, the identity and/or version of the compression algorithm used to encode a given audio segment may be encoded in the content of one or more packets that convey the given segment. Other methods of transmitting the encoded audio will occur to those of skill in the art. .
[0081] Reference is now made to Fig. 4B, which shows operation of the client device associated with a given participant, which may be client device 120 or client device 120A, by way of non-limiting example.
[ 0082 ] At action 410B, an encoded audio segment is received from the compute server 200C, the rendering server 200R or the hybrid server 200H (depending on the embodiment) . At action 420B, the encoded audio is decoded in accordance with the decompression algorithm that is complementary to the compression algorithm used in the encoding process. In a non-limiting embodiment, the identity or version of the compression algorithm used to encode the audio segment may be specified in the content of one or more packets that convey the audio segment.
[ 0083] At action 430B, the (decoded) audio segments are processed. This can include placing the decoded audio segments in a buffer, performing error correction, combining multiple successive waveforms, and so on.
The result can be a final sound to be presented to the user on a per-frame basis.
[ 0084 ] At action 440B, the final generated sound is output via the output mechanism of the client device.
For example, the sound is played through a sound card or loudspeaker of the client device.
[ 0085] IX. Specific Description of Non-Limiting
Embodiments
A more detailed description of certain non- limiting embodiments of the present invention is now provided. [0086] [Embodiments]
<<Menu Extension Processing>>
Details of menu extension processing on the server side (server system 100, compute server 200C and rendering server 200R or hybrid server 200H) according to an embodiment as one form of the present invention, which is executed on the server side of the system having the above arrangement, will be described with reference to the block diagram of Fig. 5A and the flowchart of Fig. 6.
[0087] Menu extension processing is processing of adding a menu item of a new function (extended
function) to existing menu items displayed when, for example, predetermined operation input is done by the main process of a content provided by the server, thereby extending the function. In this embodiment, the provided content is a game content, as described above. The main process is a process of performing a series of processes for the game content by changing the game status in accordance with input of the client device input 140 received from the client device 120 and rendering and outputting a game screen
corresponding to the status after the change. That is, the main process is a process executed by the video game functional module 270 and the rendering functional module 280 described above.
[0088] Fig. 5A is a block diagram showing a module arrangement for execution of menu extension processing on the server side according to the embodiment of the present invention in accordance with the flow of processing and data.
[0089] The client device input 140 received via the Internet 130 is first checked by an input monitoring module 500. In this embodiment, the input monitoring module 500 first checks whether the input done is input for menu display, and determines whether the situation requires to execute processing of displaying a menu item for an extended function. The input monitoring module 500, for example, always monitors the received client device input 140, and directly outputs the received client device input 140 to the video game functional module 270 until input for menu display is done. In addition, after menu display, the input monitoring module 500 monitors whether input for selecting the added menu item is done during the time display concerning the menu is included in the screen.
[0090] Upon receiving an instruction to execute processing of displaying a menu item for an extended function from the input monitoring module 500, an extension processing module 510 executes various processes for adding the menu item for the extended function to the game screen generated by the main process. More specifically, the extension processing module 510 performs processing of causing a superimposed image generation module 520 to render the display item of the menu item or the like that is to be superimposed on the game screen. When input for selecting the menu item for the extended function
(input for an instruction to execute processing
corresponding to the item) is done, the extension processing module 510 executes the corresponding processing (extension processing) .
[0091] The superimposed image generation module 520 renders the display item of the menu item for the extended function and other display items, and
generates a superimposed image to be superimposed on the game screen generated by the main process. The superimposed image generation module 520 may
additionally generate mask data to be used in composite processing for superimposition . The mask data may represents the transparency level of each pixel of the superimposed image when superimposing the superimposed image on the game screen. The data of the display item of the menu item and information such as a display position may be recorded in a recording device (not shown) in advance.
[0092] A composite module 530 performs composite processing of superimposing the superimposed image generated by the superimposed image generation module 520 on the game screen generated by processing in the main process, thereby generating a new game screen (composite screen) . The composite module 530 outputs the generated composite screen to the video encoder 285. Even when no superimposed image needs to be generated, the composite module 530 receives input of the game screen generated by the main process. In this case, the game screen is directly output to the video encoder 285.
[0093] Details of menu extension processing
implemented by such a module arrangement will be
described with reference to the flowchart of Fig. 6.
Processing corresponding to this flowchart can be implemented when the CPU 222 reads out a corresponding processing program stored in, for example, a recording device (not shown), loads the program to the RAM 230, and executes it independently of the main process.
[0094] In step S601, the input monitoring module 500 determines whether the client device input 140 received from the client device 120 is input for menu display (normal menu display) . Upon determining that the
client device input 140 is input for normal menu
display, the input monitoring module 500 transmits, to the extension processing module 510, an instruction to execute processing of displaying a menu item for an extended function, transfers the operation input of the client device input 140 to the main process, and
advances the process to step S603. Upon determining that the client device input 140 is not input for normal menu display, the input monitoring module 500 transfers the operation input of the client device input 140 directly to the main process in step S602, and returns the process to step S601.
[0095] In step S603, the extension processing module 510 causes the superimposed image generation module 520 to render a superimposed image including the menu item for the extended function and used to add the selected item to the normal menu display arranged on the game screen by processing of the main process. When the normal menu display is done as in, for example, Fig. 7A, the superimposed image can be as in Fig. 7B. In this embodiment, as the image data of the menu item for the extended function, data that combines the design and menu items arranged on the game screen in normal menu display is prepared in advance and recorded in a
recording device. Information of the arrangement position of the menu item for the extended function is also predetermined in consideration of the arrangement interval of the menu items in the normal menu display and recorded in the recording device. Hence, upon receiving a superimposed image generation instruction, the superimposed image generation module 520 acquires the image data of the menu item and information of its arrangement position, and generates a superimposed image. Note that in this embodiment, a description will be made assuming that the superimposed image generation module 520 generates a superimposed image upon receiving an instruction from the extension processing module 510. However, the practice of the present invention is not limited to this. It should be appreciated that since the menu item for the extended function to be arranged in the superimposed image does not dynamically change, for example, the superimposed image itself may be recorded in the recording device in advance.
[0096] In step S604, the extension processing module 510 causes the composite module 530 to composite the superimposed image generated by the superimposed image generation module 520 with the game screen generated by the rendering functional module 280 in correspondence with operation input received in the same frame and generate a composite image. The composite module 530 superimposes, for example, pixels having colors other than those defined not to be superimposed out of the superimposed image at the same pixel positions of the game screen, thereby generating the composite image. When a game screen as shown in Fig. 7A is generated, and a superimposed image as shown in Fig. 7B is
generated, hatched pixels in Fig. 7B are handled as pixels to be made transparent at the time of
superimposition and are therefore not superimposed.
Conversely, the remaining pixels are processed as to be superimposed, and a new game screen (composite image) including menu display with the additional menu item for the extended function as shown in Fig. 7C is generated.
[0097] In step S605, the composite module 530 outputs the generated composite image to the video encoder 285 as the game screen, and terminates the menu extension processing .
[0098] This makes it possible to generate a new game screen in which a new menu item is arranged in normal menu display displayed on the game screen by processing of the main process without the necessity of causing the main process to perform special processing.
[0099] <<Display Update Processing»
A method of changing the superimposed image and updating display in accordance with predetermined operation input to the thus generated menu display with an extended function will be described next with reference to the flowcharts of Figs. 8A and 8B. This display update processing is executed for each
processing frame of the game program of the main process during rendering processing of menu display. At this time, the client device input 140 is received on a per-frame basis.
[0100] In this embodiment, a description will be made assuming that the operation input to menu display in the client device 120 is done using a pointing device such as a mouse for the sake of descriptive simplicity. Each menu item of menu display behaves to highlight itself when an operation (mouse-over) of making an indicator position indicated by the pointing device enter a region corresponding to the item is performed. When the pointed position exists within a region corresponding to a menu item, and operation input
(mouse click) corresponding to determination is done, the menu item behaves to execute a function
corresponding to it and make a transition to new display such as a setting screen or further superimpose a setting window. When the pointed position exists within a region corresponding to a display item (item different from the menu items) configured to end the menu display, and operation input corresponding to determination is done, the display is turned off.
[0101] In step S801, the input monitoring module 500 determines whether a menu item for an extended function or a display item displayed by execution of an extended function is included in a screen to be provided to the client device 120. That is, the input monitoring module 500 determines whether a superimposed image is being superimposed on the game screen currently
generated by the main process. Upon determining that a superimposed image is being superimposed on the game screen, the input monitoring module 500 advances the process to step S802. Upon determining that no
superimposed image is being superimposed, the input monitoring module 500 advances the process to step S814.
[0102] In step S802, the input monitoring module 500 determines whether the pointed position indicated by the received client device input 140 is included in a region where the menu item for the extended function is arranged. Upon determining that the pointed position is included in the region where the menu item for the extended function is arranged, the input monitoring module 500 advances the process to step S803. Upon determining that the pointed position is not included, the input monitoring module 500 advances the process to step S810.
[0103] In step S803, the input monitoring module 500 determines whether the client device input 140 includes information of operation input to execute the function of the menu item. That is, the input monitoring module 500 determines whether the pointed position exists in the region where the menu item for the extended
function is arranged, and operation input (e.g., mouse click) to select the menu item has been performed.
Upon determining that the information of operation input to execute the function is included, the input monitoring module 500 advances the process to step S804. Upon determining that the information is not included, the input monitoring module 500 advances the process to step S807.
[0104] In step S804, the input monitoring module 500 transfers the received client device input 140 not to the video game functional module 270 but to the
extension processing module 510 and advances the process to step S805. Operation input to execute the function corresponding to the menu item for the
extended function does not correspond to operation input to perform any processing in the main process. Alternatively, the operation input may be determined as one to perform processing different from the extended function, and transition to an undesirable situation may occur so that, for example, the screen may transit to another screen, or a game that has paused may restart. In this embodiment, to prevent the main process from performing processing of the operation input, when operation input to execute the extended function is performed, the input monitoring module 500 transfers the operation input not to the main process but only to the extension processing module 510. That is, a situation is apparently created in which a sub- process other than the main process temporarily seize authority to process operation input to execute the extended function.
[0105] In step S805, the extension processing module 510 executes processing of the extended function corresponding to the menu item arranged at the pointed position. The processing of the extended function can include above-described screen transition or a change in the display form of the menu item or another item caused by selection of the menu item. At this time, the extension processing module 510 may control the video game functional module 270 to stop execution of processing in the main process.
[0106] The processing of the extended function may include processing for settings for an additional content element that is not implemented in the game content of the main process or settings of key
assignment to play the game content in the client device 120 of various kinds of hardware. It may be processing for allowing the player to change settings such as communication settings and distribution image quality settings typical to the cloud gaming system during an experience of a game. In addition, since the service is provided via the Internet, cooperation functions with other services using the Internet such as settings for upload to an arbitrary SNS (Social Networking Site) or video uploading site, access and order functions for associated goods sales sites, and an order function to pizza delivery for a player during an experience of a game may be provided as extended functions .
[0107] In step S806, the superimposed image
generation module 520 generates a superimposed image for the selected extended function under the control of the extension processing module 510. The superimposed image for the selected extended function can be a setting window or a screen for a specific application, as shown in Fig. 9A. In this embodiment, a description will be made assuming that the image for the extended function is an image to be superimposed on the game screen rendered by the rendering functional module 280 in the main process. However, the image is not limited to this in the practice of the present invention. The image for the extended function may be, for example, an image constructing the entire screen, and may be output as a screen to be provided to the client device 120 without being superimposed on the game screen, as will be described later.
[0108] On the other hand, upon determining in step S803 that information of operation input to execute the function is not included, the input monitoring module 500 transfers the client device input 140 to the video game functional module 270 and the extension processing module 510 in step S807, and advances the process to step S808. in this embodiment, a description will be made assuming that simple mouse-over on the menu item of the extended function leads to processing different from the extended function in the main process, and processing in the main process is not executed.
However, if arbitrary processing in the main process is to be executed by an operation of changing the pointed position such as mouse-over, the input monitoring module 500 may transfer the client device input 140 only to the extension processing module 510 in this step .
[0109] In step S808, the superimposed image
generation module 520 generates a superimposed image as shown in Fig 9B, which highlights the menu item for the extended function corresponding to the pointed position indicated by the client device input 140, under the control of the extension processing module 510.
[0110] In step S809, the composite module 530
composites the game screen generated by the rendering functional module 280 with the superimposed image generated by the superimposed image generation module 520 to generate a composite image under the control of the extension processing module 510.
[0111] On the other hand, upon determining in step S802 that the pointed position indicated by the client device input 140 is not included in the region where the menu item for the extended function is arranged, the input monitoring module 500 determines in step S810 whether the client device input 140 includes
information of operation input to execute the function of a menu item (normal menu item) arranged in normal menu display. That is, the input monitoring module 500 determines' whether the pointed position exists in the region where a normal menu item is arranged, and operation input to select the menu item has been performed. The information of the position of each normal menu item may be recorded in a recording device in advance, or obtained by, for example, analyzing the image of the game screen before and after the display of the normal menu item and specifying the position where a display item of a predetermined shape is
arranged. Upon determining that the information of operation input to execute the function of the normal menu item is included, the input monitoring module 500 advances the process to step S811. Upon determining that the information is not included, the input
monitoring module 500 advances the process to step S812.
[0112] In step S811, the input monitoring module 500 transfers the received client device input 140 only to the video game functional module 270 and advances the process to step S813. In addition, the input
monitoring module 500 notifies the extension processing module 510 that operation input to execute the function of the normal menu item has been done.
[0113] In step S812, the input monitoring module 500 determines whether the client device input 140 includes information of operation input to end normal menu display. Upon determining that the information of operation input to end normal menu display is included, the input monitoring module 500 advances the process to step S813. Upon determining that the information is not included, the input monitoring module 500 advances the process to step S814.
[0114] In step S813, the extension processing module 510 causes the superimposed image generation module 520 to stop generating and outputting a superimposed image. The extension processing module 510 also causes the composite module 530 to stop executing composite processing. That is, in this step, the extension processing module 510 performs processing of
controlling not to superimpose a superimposed image to avoid an obstacle to the item displayed by executing the function of the normal menu item.
[0115] In step S814, the input monitoring module 500 determines whether the client device input 140 includes operation input to end execution of the function of the normal menu item. That is, the input monitoring module 500 determines whether to superimpose the superimposed image (the menu item for the extended function) , whose superimposition has been stopped by execution of the function of the normal menu item, again in accordance with the end of display of the item displayed by executing the function. Upon determining that
operation end execution of the function of the normal menu item is included, the input monitoring module 500 advances the process to step S815. Upon determining that the operation input is not included, the input monitoring module 500 terminates the display update processing . [0116] In step S815, the extension processing module 510 causes the superimposed image generation module 520 to render a superimposed image in which the menu item for the extended function is arranged, and advances the process to step S809.
[0117] This makes it possible to execute display transition of the display item for the extended
function and execution of the extended function by a process different from the main process on the server side of the system according to this embodiment. In addition, since the superimposed image is updated by this execution, a screen for a medium output 150 finally provided to the client device 120 can present display for function extension of a provided content to the player without a sense of incongruity.
[0118] [First Modification]
In the above-described embodiment, a method has been explained in which a menu item for an extended function is superimposed, and upon receiving operation input to the item, a sub-process temporarily seizes authority to process the operation input from the main process and executes processing of the extended
function. However, the method of providing a user experience with extended functions when providing an existing content without altering the program of the content is not limited to this. In this modification, a method will be described in which a change in the main process caused by operation input is detected, and extended display is superimposed, thereby providing a user experience with extended functions, instead of allowing a sub-process to seize authority to process operation input.
[0119] Some game contents performs text display to present information of a result of processing based on done operation input, as shown in Fig. 10A. Such text display is configured to display only predetermined information set at the time of development of a content. When different information other than the predetermined information is presented, the user experience can be extended.
[0120] A game screen shown in Fig. 10A is that of a so-called fighting game. This screen includes the life gauge of each character. For example, assume a case where the total damage amount (life decrement) within a so-called combo period where a character is
continuously damaged is measured, and users who use the corresponding content in the cloud gaming system are ranked on a daily basis (for example, leaderboard) .
Pieces of information of, for example, the combo period and a parameter representing life in the life gauge are managed by processing of the video game functional module 270 and stored in a predetermined storage area such as the RAM 230. In this case, upon determining that the client device input 140 includes operation input to attack, the input monitoring module 500 notifies the extension processing module 510 of it, and the extension processing module 510 measures the decrement of the life parameter by the attack during the combo period. At the end of the combo period, the extension processing module 510 compares the life decrement value with the maximum life decrement value of the day managed for system users. When the life decrement value is larger than the maximum life
decrement value, the extension processing module 510 causes the superimposed image generation module 520 to generate a superimposed image in which text display representing that the user has marked the maximum life decrement value of the day is arranged at a
predetermined position. At this time, the text display arranged in the superimposed image preferably uses the font used in the game. The composite module 530 composites the superimposed image with the game screen generated by the main process to generate and output a composite image as shown in Fig. 10B under the control of the extension processing module 510. Note that the number of rows and the arrangement position of text display can change in accordance with the progress of the game content and may therefore be decided by monitoring parameters or analyzing the image of the game screen.
[0121] As described above, when the extension processing module 510 monitors intermediate data output by the main process or parameters and the like managed by the main process, extended information obtained by variously evaluating these pieces of information can be included in the screen of the content and provided. Note that in this modification, a method of presenting information using a change in the life parameter as an evaluation target has been described as an example. However, it should easily be appreciated that the evaluation target, evaluation method and information to be presented are not limited to these.
[0122] [Second Modification]
In the above-described first modification, a method of monitoring parameters managed by the main process and providing a game screen in which
corresponding text display is arranged at a
predetermined position has been explained. However, the method of extending a user experience by display is not limited to this in the practice of the present invention. In this modification, the system further includes an image analysis module 540 configured to analyze a game screen generated by the rendering functional module 280 in the main process, as shown in Fig. 5B. A method of causing the image analysis module 540 to analyze a game screen to detect a change in parameters will be described. That is, instead of monitoring internally-managed parameter values, the method detects the state in the game content or the execution state of processing for the game content by analyzing the game screen.
[0123] For example, a game screen as shown in Fig. 10A includes a life gauge or text display of a result of operation input, as described above. The image analysis module 540 can detect the state in the game by detecting the change amount of the life gauge or performing text recognition based on the difference or correlation of the game screen between continuous frames. More specifically, when these pieces of information detected by the image analysis module 540 are transmitted to the extension processing module 510, the extension processing module 510 can perform, for example, evaluation of parameter changes as described above. That is, the extension processing module 510 can grasp the combo period or a given damage amount from the text recognition result. The extension processing module 510 can also grasp occurrence of the effect of giving an abrupt decrease in life from a change in the life gauge.
[0124] Information presentation is not limited to text display as described in the first modification, and may be done by, for example, superimposing
predetermined effect display on a portion where clash of characters takes place, as shown in Fig. IOC. This can be done by, for example, when the extension processing module 510 has received a detection result representing an abrupt decrease in life from the image analysis module 540, causing the superimposed image generation module 520 to generate a superimposed image in which effect display is arranged at a portion where a motion vector having a scalar of a predetermined value or more is detected between preceding and
subsequent frames.
[0125] In this modification, a description has been made assuming that the extension processing module 510 decides whether to do information presentation by image analysis of the game screen. However, the practice of the present invention is not limited to this. It may be decided whether to do information presentation based on, for example, information such as the number of command inputs or a variation in an analog value from the history of the client device input 140 received by the input monitoring module 500. Alternatively, it may be decided whether to do information presentation based on, for example, whether an audio signal output in accordance with the game screen includes a sound of an amplitude of a predetermined value or more. Otherwise, it may be decided whether to do information
presentation by combining these pieces of information.
[0126] [Third Modification]
In the above-described first and second
modifications, an explanation has been made assuming that a superimposed image aiming at information
presentation or effect display is generated. However, the practice of the present invention is not limited to this .
[0127] For example, as shown in Fig. 11A, a sub-game may be provided, which arranges display items such as an icon in a region of interest detected in a game screen by a predetermined method, and the user collects the items by selecting them during transition of the game screen. In this case, when the client device input 140 includes operation input to select a display item, the input monitoring module 500 transfers the operation input not to the video game functional module 270 but to the extension processing module 510, and the extension processing module 510 performs processing such as score calculation for the item collection. To present that an item is collected, the extension processing module 510 may cause the superimposed image generation module 520 to generate a superimposed image as shown in Fig. 11B that changes the display item to display representing that the item is collected and superimpose the image on the provided game screen. In addition, a predetermined rendering effect may be provided when generating the superimposed image so as to integrate the display items to be arranged with the atmosphere (illumination, shading, reflection and camera angle) of rendered objects on the game which exist at the arrangement positions in the game screen.
[0128] Note that the region of interest may be specified by causing the image analysis module 540 shown in Fig. 5B to detect, for example, a region of the game screen without a change between continuous frames or a region where the edge components exhibit an intensity equal to or more than a threshold, and the contrast ratio to the peripheral region is high. The region of interest may be specified by causing the image analysis module 540 to detect a predetermined image pattern included in the game screen. Various other methods can appropriately be used to detect the region of interest.
[0129] Other Embodiments
While the present invention has been described with reference to exemplary embodiments, it is to be understood that the invention is not limited to the disclosed exemplary embodiments. The scope of the following claims is to be accorded the broadest
interpretation so as to encompass all such
modifications and equivalent structures and functions. Also, the information processing apparatus and the controlling method according to the present invention are realizable by a program executing the methods on one or more computers. The program is
providable/distributable by being stored on a computer- readable storage medium or through an electronic communication line.
[0130] This application claims the benefit of U.S. Provisional Patent Application No. 61/820,909 filed May 8, 2013, which is hereby incorporated by reference herein in its entirety.

Claims

1. An information processing apparatus comprising: receiving means for receiving operation input for a content;
first generation means for generating a first image corresponding to the content by executing a first program for the content in accordance with the
operation input received by said receiving means;
second generation means for generating a second image to be added to the first image by executing a second program different from the first program;
output means for outputting a composite image obtained by compositing the first image and the second image; and
control means for, in a case where said receiving means receives the operation input to a region
according to the second image in the composite image, controlling not to cause said first generation means to execute the first program in accordance with the operation input.
2. The apparatus according to claim 1, wherein in a case where said receiving means receives the operation input to the region according to the second image, said second generation means updates the second image.
3. The apparatus according to claim 1 or 2, wherein the second image includes a display item configured to instruct to execute a function that is not implemented by the first program and
in a case where said receiving means receives the operation input to the region according to the second image, said second generation means executes, out of the second program, a program of a function
corresponding to the display item arranged in the region
4. The apparatus according to claim 3, wherein the function that is not implemented by the first program comprises one of a function of changing a setting typical to the information processing apparatus or a cooperation function with a service using the Internet.
5. The apparatus according any one of claims 1 to 3, wherein said control means causes said first generation means not to execute the first program in accordance with the operation input to the region according to the second image by not transferring the operation input.
6. An information processing apparatus comprising: receiving means for receiving operation input for a content;
first generation means for generating a first image corresponding to the content by executing a first program for the content in accordance with the
operation input received by said receiving means;
monitor means for monitoring a predetermined parameter that changes during execution of the first program; second generation means for generating a second image to be added to the first image by executing a second program different from the first program in a case where the predetermined parameter meets a
predetermined condition; and
output means for outputting a composite image obtained by compositing the first image and the second image .
7. An information processing apparatus comprising: receiving means for receiving operation input for a content;
first generation means for generating a first image corresponding to the content by executing a first program for the content in accordance with the
operation input received by said receiving means;
analysis means for analyzing the first image and detecting whether an execution state of the content meets a predetermined condition;
second generation means for generating a second image to be added to the first image by executing a second program different from the first program in a case where the execution state meets the predetermined condition; and
output means for outputting a composite image obtained by compositing the first image and the second image .
8. The apparatus according to claim 6 or 7, wherein said analysis means further decides a position where a display item is to be arranged in the second image by analyzing the first image, and
said second generation means generates the second image in which the display item is arranged at the position decided by said analysis means.
9. An information processing apparatus comprising: receiving means for receiving operation input for a content;
first generation means for generating a first image corresponding to the content by executing a first program for the content in accordance with the
operation input received by said receiving means;
analysis means for analyzing the first image and deciding a position where a display item is to be arranged;
second generation means for generating a second image to be added to the first image, in which the display item is arranged at the position decided by said analysis means, by executing a second program different from the first program; and
output means for outputting a composite image obtained by compositing the first image and the second image .
10. The apparatus according to any one of claims 1 to 9, wherein said composite means generates the composite image by superimposing the second image on the first image .
11. A control method of an information processing apparatus, comprising:
a receiving step of receiving operation input for a content;
a first generation step of generating a first image corresponding to the content by executing a first program for the content in accordance with the
operation input received in the receiving step;
a second generation step of generating a second image to be added to the first image by executing a second program different from the first program;
an output step of outputting a composite image obtained by compositing the first image and the second image; and
a control step of, in a case where the operation input to a region according to the second image in the composite image is received in the receiving step, controlling not to execute the first program in
accordance with the operation input in the first generation step.
12. A control method of an information processing apparatus, comprising:
a receiving step of receiving operation input for a content;
a first generation step of generating a first image corresponding to the content by executing a first program for the content in accordance with the
operation input received in the receiving step;
a monitor step of monitoring a predetermined parameter that changes during execution of the first program;
a second generation step of generating a second image to be added to the first image by executing a second program different from the first program in a case where the predetermined parameter meets a
predetermined condition; and
an output step of outputting a composite image obtained by compositing the first image and the second image .
13. A control method of an information processing apparatus, comprising:
a receiving step of receiving operation input for a content;
a first generation step of generating a first image corresponding to the content by executing a first program for the content in accordance with the
operation input received in the receiving step;
an analysis step of analyzing the first image and detecting whether an execution state of the content meets a predetermined condition;
a second generation step of generating a second image to be added to the first image by executing a second program different from the first .program in a case where the execution state meets the predetermined condition; and
an output step of outputting a composite image obtained by compositing the first image and the second image .
14. A control method of an information processing apparatus, comprising:
a receiving step of receiving operation input for a content;
a first generation step of generating a first image corresponding to the content by executing a first program for the content in accordance with the
operation input received in the receiving step;
an analysis step of analyzing the first image and deciding a position where a display item is to be arranged;
a second generation step of generating a second image to be added to the first image, in which the display item is arranged at the position decided in the analysis step, by executing a second program different from the first program; and
an output step of outputting a composite image obtained by compositing the first image and the second image .
15. A program that causes at least one computer to function as each means of the information processing apparatus defined in any one of claims 1 to 10.
PCT/JP2014/062761 2013-05-08 2014-05-07 Information processing apparatus, control method and program WO2014181892A1 (en)

Priority Applications (4)

Application Number Priority Date Filing Date Title
EP14794717.0A EP2994830A4 (en) 2013-05-08 2014-05-07 Information processing apparatus, control method and program
JP2015546744A JP6576245B2 (en) 2013-05-08 2014-05-07 Information processing apparatus, control method, and program
US14/787,029 US20160110903A1 (en) 2013-05-08 2014-05-07 Information processing apparatus, control method and program
CA2910655A CA2910655A1 (en) 2013-05-08 2014-05-07 Information processing apparatus, control method and program

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US201361820909P 2013-05-08 2013-05-08
US61/820,909 2013-05-08

Publications (1)

Publication Number Publication Date
WO2014181892A1 true WO2014181892A1 (en) 2014-11-13

Family

ID=51867361

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2014/062761 WO2014181892A1 (en) 2013-05-08 2014-05-07 Information processing apparatus, control method and program

Country Status (5)

Country Link
US (1) US20160110903A1 (en)
EP (1) EP2994830A4 (en)
JP (1) JP6576245B2 (en)
CA (1) CA2910655A1 (en)
WO (1) WO2014181892A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2017000271A (en) * 2015-06-05 2017-01-05 株式会社コーエーテクモゲームス Game program and recording medium
JP6154516B1 (en) * 2016-05-17 2017-06-28 株式会社ドワンゴ COMMENT DISTRIBUTION DEVICE, GAME SERVER DEVICE, COMMENT DISTRIBUTION METHOD, AND PROGRAM

Families Citing this family (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8790185B1 (en) 2012-12-04 2014-07-29 Kabam, Inc. Incentivized task completion using chance-based awards
US20160127508A1 (en) * 2013-06-17 2016-05-05 Square Enix Holdings Co., Ltd. Image processing apparatus, image processing system, image processing method and storage medium
US10565818B2 (en) 2016-09-26 2020-02-18 Everi Games, Inc. Apparatus and methods for facilitating wagering on games conducted on an independent video gaming system
CA2922062A1 (en) * 2013-09-11 2015-03-19 Square Enix Holdings Co., Ltd. Rendering apparatus, rendering method thereof, program and recording medium
US10482713B1 (en) 2013-12-31 2019-11-19 Kabam, Inc. System and method for facilitating a secondary game
US10307666B2 (en) 2014-06-05 2019-06-04 Kabam, Inc. System and method for rotating drop rates in a mystery box
US9717986B1 (en) 2014-06-19 2017-08-01 Kabam, Inc. System and method for providing a quest from a probability item bundle in an online game
US9452356B1 (en) 2014-06-30 2016-09-27 Kabam, Inc. System and method for providing virtual items to users of a virtual space
US10046236B2 (en) * 2016-06-13 2018-08-14 Sony Interactive Entertainment America, LLC Browser-based cloud gaming
US11130064B2 (en) * 2017-07-17 2021-09-28 Neuromotion, Inc. Systems and methods for biofeedback gameplay
US10814230B2 (en) * 2017-10-12 2020-10-27 Microsoft Technology Licensing, Llc Interactive event broadcasting
US11224804B2 (en) * 2018-07-17 2022-01-18 Roblox Corporation Personalized remote game update capture and recording system for multi-player online games

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2001159525A (en) * 1999-11-30 2001-06-12 Mitsubishi Electric Corp Navigation device and recording medium
JP2002238036A (en) * 2001-02-07 2002-08-23 Fujitsu Ltd Broadcast method, and program for executing the broadcast method by broadcast station device
US20100304860A1 (en) 2009-06-01 2010-12-02 Andrew Buchanan Gault Game Execution Environments
US20120004041A1 (en) 2008-12-15 2012-01-05 Rui Filipe Andrade Pereira Program Mode Transition
JP2012064000A (en) * 2010-09-16 2012-03-29 Komota Kk Information processing equipment
JP2012085821A (en) * 2010-10-19 2012-05-10 Sony Computer Entertainment Inc Information processing system, information processing method, information processing program, and computer-readable recording medium with information processing program recorded thereon
JP2012168931A (en) * 2011-02-10 2012-09-06 Sony Computer Entertainment Inc Input device, information processing device and input value acquisition method
US20120231887A1 (en) * 2011-03-07 2012-09-13 Fourth Wall Studios, Inc. Augmented Reality Mission Generators

Family Cites Families (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7454713B2 (en) * 2003-12-01 2008-11-18 Sony Ericsson Mobile Communications Ab Apparatus, methods and computer program products providing menu expansion and organization functions
US7934210B1 (en) * 2004-03-26 2011-04-26 Adobe Systems Incorporated System and method for updating one or more programs and their environment
US9740794B2 (en) * 2005-12-23 2017-08-22 Yahoo Holdings, Inc. Methods and systems for enhancing internet experiences
JP4854443B2 (en) * 2006-09-21 2012-01-18 株式会社ソニー・コンピュータエンタテインメント Reproduction device, menu screen display method, menu screen display program, and computer-readable storage medium storing menu screen display program
KR101391602B1 (en) * 2007-05-29 2014-05-07 삼성전자주식회사 Method and multimedia device for interacting using user interface based on touch screen
US8968087B1 (en) * 2009-06-01 2015-03-03 Sony Computer Entertainment America Llc Video game overlay
US9355493B2 (en) * 2007-12-31 2016-05-31 Advanced Micro Devices, Inc. Device and method for compositing video planes
US8019390B2 (en) * 2009-06-17 2011-09-13 Pradeep Sindhu Statically oriented on-screen transluscent keyboard
WO2011102656A2 (en) * 2010-02-17 2011-08-25 Samsung Electronics Co., Ltd. Apparatus and method for providing user interface
JP5147087B2 (en) * 2010-07-22 2013-02-20 シャープ株式会社 Display operation device and image processing device
US9727177B2 (en) * 2011-02-23 2017-08-08 Kyocera Corporation Electronic device with a touch sensor
KR101788060B1 (en) * 2011-04-13 2017-11-15 엘지전자 주식회사 Image display device and method of managing contents using the same
JP5854637B2 (en) * 2011-05-19 2016-02-09 日本放送協会 Receiving machine
EP2754471A4 (en) * 2011-09-06 2015-10-07 Capcom Co Game system, game control method and recording medium
EP2706432B1 (en) * 2012-05-25 2017-12-06 Nintendo Co., Ltd. Operation device, information processing system, and information processing method

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2001159525A (en) * 1999-11-30 2001-06-12 Mitsubishi Electric Corp Navigation device and recording medium
JP2002238036A (en) * 2001-02-07 2002-08-23 Fujitsu Ltd Broadcast method, and program for executing the broadcast method by broadcast station device
US20120004041A1 (en) 2008-12-15 2012-01-05 Rui Filipe Andrade Pereira Program Mode Transition
US20100304860A1 (en) 2009-06-01 2010-12-02 Andrew Buchanan Gault Game Execution Environments
JP2012064000A (en) * 2010-09-16 2012-03-29 Komota Kk Information processing equipment
JP2012085821A (en) * 2010-10-19 2012-05-10 Sony Computer Entertainment Inc Information processing system, information processing method, information processing program, and computer-readable recording medium with information processing program recorded thereon
JP2012168931A (en) * 2011-02-10 2012-09-06 Sony Computer Entertainment Inc Input device, information processing device and input value acquisition method
US20120231887A1 (en) * 2011-03-07 2012-09-13 Fourth Wall Studios, Inc. Augmented Reality Mission Generators

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See also references of EP2994830A4

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2017000271A (en) * 2015-06-05 2017-01-05 株式会社コーエーテクモゲームス Game program and recording medium
JP6154516B1 (en) * 2016-05-17 2017-06-28 株式会社ドワンゴ COMMENT DISTRIBUTION DEVICE, GAME SERVER DEVICE, COMMENT DISTRIBUTION METHOD, AND PROGRAM
JP2017205189A (en) * 2016-05-17 2017-11-24 株式会社ドワンゴ Comment distribution device, game server device, comment distribution method, and program

Also Published As

Publication number Publication date
EP2994830A1 (en) 2016-03-16
CA2910655A1 (en) 2014-11-13
JP2016526929A (en) 2016-09-08
EP2994830A4 (en) 2017-04-19
US20160110903A1 (en) 2016-04-21
JP6576245B2 (en) 2019-09-18

Similar Documents

Publication Publication Date Title
US20160110903A1 (en) Information processing apparatus, control method and program
CA2872130C (en) Information processing apparatus, rendering apparatus, method and program
JP5987060B2 (en) GAME SYSTEM, GAME DEVICE, CONTROL METHOD, PROGRAM, AND RECORDING MEDIUM
US20160293134A1 (en) Rendering system, control method and storage medium
CA2886282C (en) Dynamic allocation of rendering resources in a cloud gaming system
US9873045B2 (en) Systems and methods for a unified game experience
US20160127508A1 (en) Image processing apparatus, image processing system, image processing method and storage medium
EP3000043B1 (en) Information processing apparatus, method of controlling the same and program
CA2918725C (en) Information processing apparatus, control method, program, and recording medium

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 14794717

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2015546744

Country of ref document: JP

Kind code of ref document: A

REEP Request for entry into the european phase

Ref document number: 2014794717

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 2014794717

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 14787029

Country of ref document: US

ENP Entry into the national phase

Ref document number: 2910655

Country of ref document: CA

NENP Non-entry into the national phase

Ref country code: DE