US9344753B2 - Mobile input and sensor device for a computer-controlled video entertainment system - Google Patents

Mobile input and sensor device for a computer-controlled video entertainment system Download PDF

Info

Publication number
US9344753B2
US9344753B2 US12/974,029 US97402910A US9344753B2 US 9344753 B2 US9344753 B2 US 9344753B2 US 97402910 A US97402910 A US 97402910A US 9344753 B2 US9344753 B2 US 9344753B2
Authority
US
United States
Prior art keywords
user
sensor
input
input data
mobile
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US12/974,029
Other versions
US20110159962A1 (en
Inventor
Cevat Yerli
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Crytek GmbH
Original Assignee
Crytek GmbH
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Crytek GmbH filed Critical Crytek GmbH
Priority to US12/974,029 priority Critical patent/US9344753B2/en
Assigned to CRYTEK GMBH reassignment CRYTEK GMBH ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: YERLI, CEVAT
Publication of US20110159962A1 publication Critical patent/US20110159962A1/en
Assigned to CRYTEK GMBH reassignment CRYTEK GMBH CHANGE OF ASSIGNEE ADDRESS Assignors: CRYTEK GMBH
Assigned to GFACE GMBH reassignment GFACE GMBH ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: CRYTEK GMBH
Assigned to CRYTEK GMBH reassignment CRYTEK GMBH MERGER (SEE DOCUMENT FOR DETAILS). Assignors: GFACE GMBH
Application granted granted Critical
Publication of US9344753B2 publication Critical patent/US9344753B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/422Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
    • H04N21/42204User interfaces specially adapted for controlling a client device through a remote control device; Remote control devices therefor
    • H04N21/42206User interfaces specially adapted for controlling a client device through a remote control device; Remote control devices therefor characterized by hardware details
    • H04N21/42221Transmission circuitry, e.g. infrared [IR] or radio frequency [RF]
    • A63F13/06
    • A63F13/12
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/20Input arrangements for video game devices
    • A63F13/21Input arrangements for video game devices characterised by their sensors, purposes or types
    • A63F13/213Input arrangements for video game devices characterised by their sensors, purposes or types comprising photodetecting means, e.g. cameras, photodiodes or infrared cells
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/20Input arrangements for video game devices
    • A63F13/21Input arrangements for video game devices characterised by their sensors, purposes or types
    • A63F13/214Input arrangements for video game devices characterised by their sensors, purposes or types for locating contacts on a surface, e.g. floor mats or touch pads
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/20Input arrangements for video game devices
    • A63F13/21Input arrangements for video game devices characterised by their sensors, purposes or types
    • A63F13/214Input arrangements for video game devices characterised by their sensors, purposes or types for locating contacts on a surface, e.g. floor mats or touch pads
    • A63F13/2145Input arrangements for video game devices characterised by their sensors, purposes or types for locating contacts on a surface, e.g. floor mats or touch pads the surface being also a display device, e.g. touch screens
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/20Input arrangements for video game devices
    • A63F13/21Input arrangements for video game devices characterised by their sensors, purposes or types
    • A63F13/215Input arrangements for video game devices characterised by their sensors, purposes or types comprising means for detecting acoustic signals, e.g. using a microphone
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/20Input arrangements for video game devices
    • A63F13/23Input arrangements for video game devices for interfacing with the game device, e.g. specific interfaces between game controller and console
    • A63F13/235Input arrangements for video game devices for interfacing with the game device, e.g. specific interfaces between game controller and console using a wireless connection, e.g. infrared or piconet
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/30Interconnection arrangements between game servers and game devices; Interconnection arrangements between game devices; Interconnection arrangements between game servers
    • A63F13/33Interconnection arrangements between game servers and game devices; Interconnection arrangements between game devices; Interconnection arrangements between game servers using wide area network [WAN] connections
    • A63F13/332Interconnection arrangements between game servers and game devices; Interconnection arrangements between game devices; Interconnection arrangements between game servers using wide area network [WAN] connections using wireless networks, e.g. cellular phone networks
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/30Interconnection arrangements between game servers and game devices; Interconnection arrangements between game devices; Interconnection arrangements between game servers
    • A63F13/33Interconnection arrangements between game servers and game devices; Interconnection arrangements between game devices; Interconnection arrangements between game servers using wide area network [WAN] connections
    • A63F13/335Interconnection arrangements between game servers and game devices; Interconnection arrangements between game devices; Interconnection arrangements between game servers using wide area network [WAN] connections using Internet
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/30Interconnection arrangements between game servers and game devices; Interconnection arrangements between game devices; Interconnection arrangements between game servers
    • A63F13/35Details of game servers
    • A63F13/355Performing operations on behalf of clients with restricted processing capabilities, e.g. servers transform changing game scene into an encoded video stream for transmitting to a mobile phone or a thin client
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/40Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment
    • A63F13/42Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment by mapping the input signals into game commands, e.g. mapping the displacement of a stylus on a touch screen to the steering angle of a virtual vehicle
    • A63F13/424Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment by mapping the input signals into game commands, e.g. mapping the displacement of a stylus on a touch screen to the steering angle of a virtual vehicle involving acoustic input signals, e.g. by using the results of pitch or rhythm extraction or voice recognition
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/422Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
    • H04N21/42204User interfaces specially adapted for controlling a client device through a remote control device; Remote control devices therefor
    • H04N21/42206User interfaces specially adapted for controlling a client device through a remote control device; Remote control devices therefor characterized by hardware details
    • H04N21/42222Additional components integrated in the remote control device, e.g. timer, speaker, sensors for detecting position, direction or movement of the remote control, microphone or battery charging device
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/40Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment
    • A63F13/42Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment by mapping the input signals into game commands, e.g. mapping the displacement of a stylus on a touch screen to the steering angle of a virtual vehicle
    • A63F13/428Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment by mapping the input signals into game commands, e.g. mapping the displacement of a stylus on a touch screen to the steering angle of a virtual vehicle involving motion or position input signals, e.g. signals representing the rotation of an input controller or a player's arm motions sensed by accelerometers or gyroscopes
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F2300/00Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
    • A63F2300/10Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals
    • A63F2300/1068Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals being specially adapted to detect the point of contact of the player on a surface, e.g. floor mat, touch pad
    • A63F2300/1075Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals being specially adapted to detect the point of contact of the player on a surface, e.g. floor mat, touch pad using a touch screen
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F2300/00Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
    • A63F2300/10Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals
    • A63F2300/1081Input via voice recognition
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F2300/00Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
    • A63F2300/10Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals
    • A63F2300/1087Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals comprising photodetecting means, e.g. a camera
    • A63F2300/1093Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals comprising photodetecting means, e.g. a camera using visible light
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F2300/00Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
    • A63F2300/40Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterised by details of platform network
    • A63F2300/406Transmission via wireless network, e.g. pager or GSM
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F2300/00Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
    • A63F2300/40Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterised by details of platform network
    • A63F2300/407Data transfer via internet
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F2300/00Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
    • A63F2300/50Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by details of game servers
    • A63F2300/53Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by details of game servers details of basic data processing
    • A63F2300/538Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by details of game servers details of basic data processing for performing operations on behalf of the game client, e.g. rendering
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F2300/00Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
    • A63F2300/60Methods for processing data by generating or executing the game program
    • A63F2300/6063Methods for processing data by generating or executing the game program for sound processing
    • A63F2300/6072Methods for processing data by generating or executing the game program for sound processing of an input signal, e.g. pitch and rhythm extraction, voice recognition

Definitions

  • the present invention relates to a mobile input and sensor device providing user-related sensor signals and input to a computer-controlled video entertainment system.
  • the invention relates to a mobile input and sensor device for a computer-controlled gaming system to be used by at least one user wherein the mobile input and sensor device can detect natural input such as physical gestures or speech.
  • the invention further relates to a mobile input and sensor device to be applied to a system for providing multimedia data for entertaining presentations to a plurality of user (audience) at home or outside in cinemas, theater rooms or the like.
  • the present invention particularly relates to mobile input and sensor devices for a system providing computer-based entertainment, like interactive video games or performing interactive movies or shows (so-called theater games) and interactive advertisement games.
  • Entertainment systems such as video game systems or devices are well-known and are usually implemented by personal computers (PC) or game consoles which are controlled by at least one user.
  • PC personal computers
  • game consoles which are controlled by at least one user.
  • an input control device such as a gamepad, joystick or mouse, which is linked to the PC or game console by cable.
  • Modern game consoles provide wireless control in that the users are equipped with wireless control devices.
  • the user activity is detected by a sensor device which usually comprises infrared sensing elements. These sensor devices have to be installed in the vicinity of the display device (TV set or monitor) and close to the user's position.
  • the computing such as rendering
  • the processor of the used personal computer or game console is performed by the processor of the used personal computer or game console. Consequently any update of hardware or software has to be made locally and involves high costs. Since many producers of consoles even keep the hardware and software protected, the users are forced to wait for new releases and have to buy them from the producers.
  • cinema gaming systems the hardware and software is normally implemented by a computer or server which is installed at the cinema or theater. Consequently a high invest has to be made at each cinema and any updating has to be made on site. This makes it harder and costly to frequently update or upgrade the systems and thus to stay in line with the developing technology of rendering computer graphics and movies.
  • the known systems are not prepared to provide touch-free control equipment.
  • a mobile input and sensor device for a computer-controlled video entertainment system comprising a display device for displaying video entertainment pictures to at least one user; at least one computing device executing an entertainment software application for providing output data to be transmitted to said the display device to display the video entertainment pictures;
  • the invention also provides a mobile communications and/or entertainment device comprising a mobile input and sensor device for a computer-controlled video entertainment system and comprising a display device for displaying video entertainment pictures to at least one user; the system comprising at least one computing device executing an entertainment software application for providing output data to be transmitted to said the display device to display the video entertainment pictures;
  • the invention provides a mobile input and sensor device for rich input, representing user control input, to be processed by at least on computing device, such as a console or remote server or even a server cloud which then provides entertainment output to at least one display device.
  • a computing device such as a console or remote server or even a server cloud which then provides entertainment output to at least one display device.
  • the invention also provides a computer-controlled video entertainment system, the system comprising:
  • the invention provides an easy-to-handle mobile input and sensor device which can be wirelessly connected to at least one (local or remote) computing device and which can be installed near to or even within the display device and/or said computing device to detect user activities as rich inputs (gestures, speech, sound etc.) for the processed entertainment application.
  • the mobile input and sensor device is held by the user in front of a display device and be wirelessly linked to a remote computing server or to a cloud of several servers by having direct access to an IP-based network, in particular to the Internet, for transmitting sensor signals or data derived thereof as user-related input data via the IP-based network to a remote computing device or server, preferably to a cloud of servers, which then process these input data for controlling the running entertainment program.
  • IP-based network in particular to the Internet
  • the mobile input and sensor device can be made simple, since it is mainly designed to detect user activities and to generate sensor signals which are directly transmitted as input data to the remote computer(s) or server(s). Because of the fact that the mobile input and sensor device can directly be connected to the IP-based network (e.g. Internet) it can directly communicate with any IP-based computing device (gaming server or the like) being installed remotely and having powerful hardware and software to process the received input data. Thus there is no need to have a computer or game console being installed at the location of the user(s). Moreover, the users even do not need to have more than a mobile input and sensor device and a display device (TV set or monitor) when they want to enjoy and participate in an entertainment program.
  • IP-based network e.g. Internet
  • the entertainment or gaming program as such is remotely processed by said remote computer(s) under control of said input data.
  • the computer(s) provide(s) output data (in particular video and audio) which are transmitted via the IP-based network back to the location of the user(s) and are then output by the display device (TV set, monitor) being connected to said IP-based network.
  • the display device TV set, monitor
  • there a several remote gaming computers which form a so-called server cloud.
  • the mobile input and sensor device can be integrated together with a display device into any mobile device for providing entertainment to the user.
  • the invention allows at least one user (player, consumer) to interactively play high-definition entertainment programs, such as games and television shows, using touch-free natural inputs, such as physical gestures and speech.
  • a handheld controller device for input (such as a keyboard, game controller, or remote control) or to purchase and install expensive additional hardware such as a game console or computer.
  • handheld controllers which communicate with the sensor device, if this device is equipped with appropriate hard- and software.
  • the invention is preferably realized as a server (cloud)-based entertainment system.
  • the system mainly consists of a mobile input and sensor device (which basically is sensor detecting and signal transmitting hardware) and a remote cloud computing and storage infrastructure.
  • the mobile input and sensor device comprises a sensor module with at least one sensory element for detecting user activities, in particular natural user inputs such as physical gestures and/or speech, and for producing sensor signals.
  • the mobile input and sensor device may further comprise a processing unit for pre-processing said sensor signals to generate said user-related input data representing said detected user activities.
  • the at least one sensor element can be an optical sensor for detecting physical gestures or appearance, an acoustical sensor for detecting user-produced speech, sound or noise and/or a thermal sensor for detecting physical presence or temperature.
  • processing unit provides said user-related input data to said wireless network interface for being transmitted via said IP-based network to the at least one computing device.
  • the processing unit receives from the sensor elements the sensor signals and pre-processes these signals to generate the input data, in particular by applying signal and/or data recognition to generate input data which represent recognized user-related input patterns, in particular gestures and/or speech commands.
  • the processing unit may be part of the computing device and receive sensor signals from many mobile input and sensor devices which relate to a plurality of users to process these signals and thus to generate collective input data, in particular by applying signal and/or data recognition to generate collective input data which represent recognized multi-user or user-group-related input patterns, in particular collective gestures and/or collective speech commands.
  • the mobile input and sensor device can be integrated in a mobile communications and/or entertainment device comprising a visual output or projection module, in particular a laser-based projection module, to output visual effects and/or information (as a feedback from the game server) to said at least one user.
  • the mobile input and sensor device may further comprise a wireless interface/networking module to receive further input data from external user input devices, in particular form gamepads, joysticks, keypads, keyboards.
  • the system is a home environment gaming system.
  • the system is theater gaming system and the method of providing the data also comprises the step of controlling a display device by the output data to present images, movies, advertisements and/or games to the at least one audience.
  • the method can also comprise the step of receiving input data or signals from the at least one audience before providing the content data by at least one content server, wherein the content data is selected from a database in dependence of the input data or signals.
  • the input data may come from the gamepads being installed at the seats for the audience.
  • the method may also comprise the steps of: receiving input data or signals from the at least one audience before or during the presentation of images, movies, advertisements and/or games, wherein presentation of images, movies, advertisements and/or games audience is displayed in dependence of the input data or signals.
  • FIG. 1 shows the architecture of a computer-controlled video entertainment system of the invention
  • FIG. 2 shows a first embodiment of a mobile input and sensor device of the invention being connected with the remote cloud server(s) of the system;
  • FIG. 3 shows a second embodiment of a mobile input and sensor device of the invention being integrated into a mobile communications device.
  • FIG. 1 there is shown the architecture of a computer-controlled video entertainment system of the invention, the system being implemented as a multi-user gaming system for providing entertainment to a plurality of users U, such as members of a family at home or the members of an audience in a cinema or theatre.
  • users U such as members of a family at home or the members of an audience in a cinema or theatre.
  • a display device D which can be e.g., a flat panel LCD or a video projection screen, and for each user a mobile input and sensor device MISD, both devices being linked via an access interface to an IP-based network which is constituted by the Internet.
  • the display device D can receive output data OUT (video, audio) from (a) remote computing device(s) CS to be presented to the users U.
  • the mobile input and sensor device MISD detects user activities to provide and transmit user-related input data IN via the Internet to said remote computing device(s) CS.
  • the system can provide entertainment to a large number of users (hundreds, thousands or even much more) and/or to a plurality of locations (homes, cinemas, theatres . . . ). Thus many single users and/or a variety of audiences can be provided with entertainment services.
  • the configuration as shown in FIG. 1 relates to a single audience of users U.
  • FIGS. 2 and 3 different embodiments of a mobile input and sensor device MISD for detecting user activities and for providing user-related input data IN are shown.
  • FIG. 2 shows a mobile input and sensor device MISD comprising an input and sensor module ISM having at least one sensor element, namely a camera CAM directed towards the user to detect user movements, physical gestures or the like. Further to this the mobile input and sensor device has a touch panel TPN for receiving manual inputs from the user, a microphone MIC for sensing audio and a thermal sensor for detecting the physical presence or condition of the user(s) by detecting the body temperature(s).
  • the mobile input and sensor device MISD is held by the user in the vicinity of the display device D and near the user's face to detect his/her activities.
  • the mobile input and sensor device MISD further comprises a processing unit PU which can be implemented by a low-cost microprocessor to receive the sensor signals from the sensor elements CAM and MIC and to generate user-related input data IN which shall be transmitted to the remote cloud servers CS, one of which is shown in FIG. 2 .
  • a processing unit PU which can be implemented by a low-cost microprocessor to receive the sensor signals from the sensor elements CAM and MIC and to generate user-related input data IN which shall be transmitted to the remote cloud servers CS, one of which is shown in FIG. 2 .
  • the mobile input and sensor device MISD also comprises an interface IF in form of a networking module which can be implemented by a wireless Internet access module.
  • the mobile input and sensor device SD as shown in FIG. 2 can be realized as an universal and low-cost mobile input and sensor device performing an touch-less (optical, acoustical, thermal) and a touch-sensed (touchpad or panel) detection to provide (in uplink direction) user-related input data IN to any remote computer or server CS which runs a entertainment or gaming program.
  • the at least one server CS then produces output data OUT to be transmitted (in downlink direction) back to the users' location for being output at the display device D.
  • the display device D may have its own interface for IP network access or may use the same access interface as common interface CIF to the Internet (also see FIG. 1 ).
  • FIG. 3 a second embodiment of the invention is shown wherein the mobile input and sensor device MISD is integrated into a mobile consumer device for providing entertaining output to the user(s).
  • the mobile input and sensor device MISD can also be wirelessly connected to a game console or any other computing device.
  • the mobile consumer device MCD further comprises a projection module P for presenting visual output to the users such as laser beams to project feedback from the game server(s) CS.
  • a wireless network module N for receiving further input data from external user control devices such as keypads, gamepads, joysticks or the like. These data are then also transmitted via the interface module IF to the remote server(s) CS.
  • the mobile input and sensor device SD and the display device D can be integrated in the same housing.
  • the mobile input and sensor device SD can be part of any mobile user device being equipped with an IP interface to be connected with a remote IP-based entertainment service provided by the servers CS.
  • the user(s) only need(s) to set up a single device and can instantly profit from the latest technology since all updates/upgrades are made at the remote site, i.e. on the servers CS.
  • the mobile input and sensor device SD is integrated into a mobile consumer device MCD, such as a mobile phone or personal assistant having a display D.
  • the mobile input and sensor device SD can be a separate unit from the display screen or can be wholly integrated with the display screen in one mobile consumer device MCD to form a single unit.
  • the sensor module SM can comprise multiple types of sensory elements.
  • the preferred sensory elements are optical sensors, in particular two or more cameras, and at least one acoustical sensors such as microphones.
  • Option sensory elements can also be incorporated, such as a thermal sensor.
  • the user's raw unprocessed input (natural input such as a gesture) is detected by the sensor module SM.
  • a plurality of input streams can be delivered by each user, either simultaneously or sequentially.
  • the input streams detected by the sensor module SM can be passive or active. This means that the active input streams are generated through explicit actions by the user. These actions can consist of speech and physical gestures, for example.
  • the passive input streams can consist of the user's physical appearance (e.g. height or facial features), thermal signature (e.g. body temperature), and other data.
  • All of the raw inputs are converted into user-related input data IN. These data can be pre-processed, e.g. by applying pattern recognition to detect specific gestures, speech or the like.
  • the pre-processing stage can occur locally on the user's site, i.e. in the mobile input and sensor device SD, or can occur remotely on the cloud computing infrastructure CS. During the pre-processing stage, various parameters such as position, timing, and user differentiation are resolved for each discrete input.
  • the mobile input and sensor device can also receive further input data from external devices.
  • the data input streams from such external devices would be received by the module N and could be combined and processed with the raw input streams detected by the sensor module SM.
  • the combined and processed input could then be transmitted over the high-speed IP based data network NET to the cloud infrastructure CS for further processing and interpretation by the software entertainment or gaming application.
  • a response is computed and transmitted back as output OUT to the user over the high-speed data network NET.
  • These output OUT preferably is a (downlink) data stream consisting primarily of high-definition video and audio. Additional data such as text, three-dimensional data, and stereoscopic three-dimensional data may also be transmitted.
  • This user-related input process repeats itself with a regular frequency (e.g. 30 times per second) throughout the course of the session to deliver a smooth coherent synchronized data stream to the user, typically manifested as one or more of the following: full-motion two-dimensional video, full-motion three-dimensional stereoscopic data, audio, and graphical user interface elements.
  • a regular frequency e.g. 30 times per second
  • the cloud server infrastructure CS provides shared computing and storage resources for all users of the system.
  • the system can thus support a massive number of simultaneous users who can interact with one another either in the same location on a single device, or from different locations with multiple devices.
  • the cloud preferably is a cluster of actual servers that are connected in a self organizing form (said cloud) to distribute its performance in a dynamic and optimal way.
  • This processing cloud needs not to provide the content as such.
  • the content data is then transmitted to the processing cloud CS.
  • these cloud servers CS send the output data OUT back to the client, i.e. to the display device D (at home or in the theater room).
  • the most prominent data which have to be calculated are the visuals of the highest quality.
  • the client has always the best performance guaranteed no matter what performance the local client is capable of.
  • the service is not limited to graphics only. Any demanding process can be managed by the computing cloud CS.
  • a preferred field of use is theater gaming.
  • the small increase in lag time lapse added due to the transport of the data forth and back
  • the demand for extreme performance makes it the best choice in terms of cost efficiency.
  • the present invention relates to a computer-controlled video entertainment system and to a mobile input and sensor device providing user-related input data to said system.
  • the invention provides an easy-to-handle mobile input and sensor device having direct access to an IP-based network, in particular to the Internet, for transmitting sensor signals or data derived thereof as user-related input data via the IP-based network to a remote computing device or server, preferably to a cloud of servers, which then process these input data for controlling the running entertainment program.
  • the mobile input and sensor device can be made simple, since it is mainly designed to detect user activities and to generate sensor signals which are directly transmitted as input data to the remote computer(s) or server(s). Because of the fact that the mobile input and sensor device can directly be connected to the IP-based network (e.g.
  • IP-based computing device (gaming server or the like) being installed remotely and having powerful hardware and software to process the received input data.
  • IP-based computing device such as server or the like
  • the users just need to have at their location (at home, in a cinema . . .) their mobile input and sensor devices and a display device which is connected to the Internet.
  • both devices can communicate with the remote computer(s) which run(s) the entertainment program.
  • the users need not to take care of updating and/or upgrading, but can always profit from the latest technology.
  • the invention in particular the cloud being described here, can be used in all fields of providing interactive entertainment applications or content and also of providing non-interactive content.
  • the invention can support Internet applications via any networks and/or media, including WiFi (Wireless Fidelity), BPL (broadband power line), LAN, WiMax (Worldwide Interoperability for Microwave Access) or any other standard technology.
  • WiFi Wireless Fidelity
  • BPL broadband power line
  • LAN Worldwide Interoperability for Microwave Access
  • WiMax Worldwide Interoperability for Microwave Access

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Human Computer Interaction (AREA)
  • Signal Processing (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Selective Calling Equipment (AREA)
  • Information Transfer Between Computers (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

Mobile input and sensor device for a computer-controlled video entertainment system, in particular a mobile input and sensor device providing user-related input data to said system. This provides an easy-to-handle mobile input and sensor device having direct access to an IP-based network, for transmitting sensor signals or data derived thereof as user-related input data via the IP-based network to a remote computing device or server, preferably to a cloud of servers, which then process these input data for controlling the running entertainment program. The mobile input and sensor device can be made simple, since it is designed to detect user activities and to generate sensor signals which are transmitted as input data to the remote computer(s) or server(s). The mobile input and sensor device directly communicates with any IP-based computing device (gaming server or the like) installed remotely and having powerful hardware and software to process the received input data.

Description

The present application is related to and claims benefit under 35 U.S.C. §119(e) of U.S. Provisional Application No. 61/291,032, filed Dec. 30, 2009, the entirety of which is hereby incorporated herein by reference.
FIELD OF THE INVENTION
The present invention relates to a mobile input and sensor device providing user-related sensor signals and input to a computer-controlled video entertainment system. In particular the invention relates to a mobile input and sensor device for a computer-controlled gaming system to be used by at least one user wherein the mobile input and sensor device can detect natural input such as physical gestures or speech. The invention further relates to a mobile input and sensor device to be applied to a system for providing multimedia data for entertaining presentations to a plurality of user (audience) at home or outside in cinemas, theater rooms or the like. In this respect the present invention particularly relates to mobile input and sensor devices for a system providing computer-based entertainment, like interactive video games or performing interactive movies or shows (so-called theater games) and interactive advertisement games.
BACKGROUND OF INVENTION
Entertainment systems, such as video game systems or devices are well-known and are usually implemented by personal computers (PC) or game consoles which are controlled by at least one user. For controlling these systems the user is equipped with an input control device, such as a gamepad, joystick or mouse, which is linked to the PC or game console by cable. Modern game consoles provide wireless control in that the users are equipped with wireless control devices. The user activity is detected by a sensor device which usually comprises infrared sensing elements. These sensor devices have to be installed in the vicinity of the display device (TV set or monitor) and close to the user's position.
In US 2008/0039202 A1 such a game console with wireless sensor device is disclosed. The sensor device is linked via a cable to the game console which receives the sensor signals and processes these signals as input data for the running game. This means that the input signal and data processing is solely performed by said console, in particular by the processing unit of it. Thus the processing of the sensor signals depends on the layout and design of the processing unit which can have the effect that only a limited number of user control devices can be used. Therefore the system is not designed to be controlled by a large number of users or even by an audience. Moreover, the sensor device is exclusively designed to only work with said console. Thus there are still sensor devices cannot be used to operate with different systems. Finally, each user is equipped with a control device (wireless gamepad) which still has to be touched. But it would be desirable also to provide a contactless or touch-free control as well.
Beside these drawbacks there are further deficiencies to be mentioned here:
In present systems, the computing, such as rendering, is performed by the processor of the used personal computer or game console. Consequently any update of hardware or software has to be made locally and involves high costs. Since many producers of consoles even keep the hardware and software protected, the users are forced to wait for new releases and have to buy them from the producers. In cinema gaming systems the hardware and software is normally implemented by a computer or server which is installed at the cinema or theater. Consequently a high invest has to be made at each cinema and any updating has to be made on site. This makes it harder and costly to frequently update or upgrade the systems and thus to stay in line with the developing technology of rendering computer graphics and movies. Moreover the known systems are not prepared to provide touch-free control equipment.
SUMMARY OF INVENTION
According to a first aspect of the present invention there is provided a mobile input and sensor device for a computer-controlled video entertainment system comprising a display device for displaying video entertainment pictures to at least one user; at least one computing device executing an entertainment software application for providing output data to be transmitted to said the display device to display the video entertainment pictures; wherein
    • the mobile input and sensor device is held by the at least one user in front of said display device and/or said computing device for providing user-related input data to be processed by said entertainment software application, and wherein the mobile input and sensor device comprises at least one optical sensory element being directed towards the user and at least one input element and further comprises at least one wireless interface for transmitting input data to said at least one computing device.
The invention also provides a mobile communications and/or entertainment device comprising a mobile input and sensor device for a computer-controlled video entertainment system and comprising a display device for displaying video entertainment pictures to at least one user; the system comprising at least one computing device executing an entertainment software application for providing output data to be transmitted to said the display device to display the video entertainment pictures; wherein
    • the mobile input and sensor device comprises a touch panel being integrated in said display device for providing user-related input data to be processed by said entertainment software application, and wherein the mobile input and sensor device comprises at least one optical sensory element being directed towards the user and at least one input element and further comprises at least one wireless interface for transmitting input data to said at least one computing device.
Thus the invention provides a mobile input and sensor device for rich input, representing user control input, to be processed by at least on computing device, such as a console or remote server or even a server cloud which then provides entertainment output to at least one display device.
The invention also provides a computer-controlled video entertainment system, the system comprising:
    • a display device for displaying video entertainment pictures to at least one user;
    • at least one computing device being installed remotely from said display device and executing an entertainment software application for providing output data to be transmitted via an IP-based network to said the display device to display the video entertainment pictures;
    • a mobile input and sensor device comprising at least one optical sensory element being directed towards the user and at least one input element and further comprising at least one wireless interface for transmitting input data to said at least one computing device.
Thus the invention provides an easy-to-handle mobile input and sensor device which can be wirelessly connected to at least one (local or remote) computing device and which can be installed near to or even within the display device and/or said computing device to detect user activities as rich inputs (gestures, speech, sound etc.) for the processed entertainment application.
Instead of being installed at or within a local computing device (e.g. a game console), the mobile input and sensor device is held by the user in front of a display device and be wirelessly linked to a remote computing server or to a cloud of several servers by having direct access to an IP-based network, in particular to the Internet, for transmitting sensor signals or data derived thereof as user-related input data via the IP-based network to a remote computing device or server, preferably to a cloud of servers, which then process these input data for controlling the running entertainment program. The effort for realizing said mobile input and sensor device in terms of hardware and software can be kept very low. The mobile input and sensor device can be made simple, since it is mainly designed to detect user activities and to generate sensor signals which are directly transmitted as input data to the remote computer(s) or server(s). Because of the fact that the mobile input and sensor device can directly be connected to the IP-based network (e.g. Internet) it can directly communicate with any IP-based computing device (gaming server or the like) being installed remotely and having powerful hardware and software to process the received input data. Thus there is no need to have a computer or game console being installed at the location of the user(s). Moreover, the users even do not need to have more than a mobile input and sensor device and a display device (TV set or monitor) when they want to enjoy and participate in an entertainment program. The entertainment or gaming program as such is remotely processed by said remote computer(s) under control of said input data. The computer(s) provide(s) output data (in particular video and audio) which are transmitted via the IP-based network back to the location of the user(s) and are then output by the display device (TV set, monitor) being connected to said IP-based network. This means that the users just need to have installed at their location (at home, in a cinema . . . ) a display device which is connected to the Internet and just need to have said mobile input and sensor device which is wirelessly connected to the Internet so that both devices can communicate with the remote computer(s) which run(s) the entertainment program. In a preferred embodiment there a several remote gaming computers which form a so-called server cloud. Thus many display devices and sensor devices (i.e. many homes or cinemas) can use the processing power of this server cloud. The installation costs for each user is therefore very low. Moreover, any software updating and/or hardware upgrading only needs to be made on this central server(s). The users need not to take care of updating and/or upgrading, but can always profit from the latest technology.
Moreover, the mobile input and sensor device can be integrated together with a display device into any mobile device for providing entertainment to the user.
Thus the invention allows at least one user (player, consumer) to interactively play high-definition entertainment programs, such as games and television shows, using touch-free natural inputs, such as physical gestures and speech. With this invention, there is no need to use a handheld controller device for input (such as a keyboard, game controller, or remote control) or to purchase and install expensive additional hardware such as a game console or computer. Of course there is the option to still have handheld controllers which communicate with the sensor device, if this device is equipped with appropriate hard- and software. The invention is preferably realized as a server (cloud)-based entertainment system. The system mainly consists of a mobile input and sensor device (which basically is sensor detecting and signal transmitting hardware) and a remote cloud computing and storage infrastructure.
In further embodiments of the invention the mobile input and sensor device comprises a sensor module with at least one sensory element for detecting user activities, in particular natural user inputs such as physical gestures and/or speech, and for producing sensor signals. The mobile input and sensor device may further comprise a processing unit for pre-processing said sensor signals to generate said user-related input data representing said detected user activities. The at least one sensor element can be an optical sensor for detecting physical gestures or appearance, an acoustical sensor for detecting user-produced speech, sound or noise and/or a thermal sensor for detecting physical presence or temperature.
In one embodiment the processing unit provides said user-related input data to said wireless network interface for being transmitted via said IP-based network to the at least one computing device.
In another embodiment the processing unit receives from the sensor elements the sensor signals and pre-processes these signals to generate the input data, in particular by applying signal and/or data recognition to generate input data which represent recognized user-related input patterns, in particular gestures and/or speech commands. With respect to this, the processing unit may be part of the computing device and receive sensor signals from many mobile input and sensor devices which relate to a plurality of users to process these signals and thus to generate collective input data, in particular by applying signal and/or data recognition to generate collective input data which represent recognized multi-user or user-group-related input patterns, in particular collective gestures and/or collective speech commands.
In a further embodiment the mobile input and sensor device can be integrated in a mobile communications and/or entertainment device comprising a visual output or projection module, in particular a laser-based projection module, to output visual effects and/or information (as a feedback from the game server) to said at least one user. The mobile input and sensor device may further comprise a wireless interface/networking module to receive further input data from external user input devices, in particular form gamepads, joysticks, keypads, keyboards.
In one of the embodiments of the invention the system is a home environment gaming system. In another embodiment the system is theater gaming system and the method of providing the data also comprises the step of controlling a display device by the output data to present images, movies, advertisements and/or games to the at least one audience. The method can also comprise the step of receiving input data or signals from the at least one audience before providing the content data by at least one content server, wherein the content data is selected from a database in dependence of the input data or signals. The input data may come from the gamepads being installed at the seats for the audience. The method may also comprise the steps of: receiving input data or signals from the at least one audience before or during the presentation of images, movies, advertisements and/or games, wherein presentation of images, movies, advertisements and/or games audience is displayed in dependence of the input data or signals.
BRIEF DESCRIPTION OF THE DRAWINGS
Further features and advantages of the present invention, as well as the structure and operation of various illustrative embodiments of the present invention, are described in more detail below with reference to the accompanying schematic drawings, wherein:
FIG. 1 shows the architecture of a computer-controlled video entertainment system of the invention;
FIG. 2 shows a first embodiment of a mobile input and sensor device of the invention being connected with the remote cloud server(s) of the system; and
FIG. 3 shows a second embodiment of a mobile input and sensor device of the invention being integrated into a mobile communications device.
DETAILED DESCRIPTION OF PREFERRED EMBODIMENTS
In FIG. 1 there is shown the architecture of a computer-controlled video entertainment system of the invention, the system being implemented as a multi-user gaming system for providing entertainment to a plurality of users U, such as members of a family at home or the members of an audience in a cinema or theatre.
At the users' site there is installed a display device D which can be e.g., a flat panel LCD or a video projection screen, and for each user a mobile input and sensor device MISD, both devices being linked via an access interface to an IP-based network which is constituted by the Internet. Thus the display device D can receive output data OUT (video, audio) from (a) remote computing device(s) CS to be presented to the users U. The mobile input and sensor device MISD detects user activities to provide and transmit user-related input data IN via the Internet to said remote computing device(s) CS. In the shown embodiment there are several computing devices in form of a cluster or cloud of data processing servers which run the gaming program under control of said user-related input IN.
Since the cloud of servers CS is a very powerful processing resource which can quite easily be extended, the system can provide entertainment to a large number of users (hundreds, thousands or even much more) and/or to a plurality of locations (homes, cinemas, theatres . . . ). Thus many single users and/or a variety of audiences can be provided with entertainment services. To simplify the explanation of the invention the configuration as shown in FIG. 1 relates to a single audience of users U.
In FIGS. 2 and 3 different embodiments of a mobile input and sensor device MISD for detecting user activities and for providing user-related input data IN are shown.
As a first embodiment FIG. 2 shows a mobile input and sensor device MISD comprising an input and sensor module ISM having at least one sensor element, namely a camera CAM directed towards the user to detect user movements, physical gestures or the like. Further to this the mobile input and sensor device has a touch panel TPN for receiving manual inputs from the user, a microphone MIC for sensing audio and a thermal sensor for detecting the physical presence or condition of the user(s) by detecting the body temperature(s). The mobile input and sensor device MISD is held by the user in the vicinity of the display device D and near the user's face to detect his/her activities. The mobile input and sensor device MISD further comprises a processing unit PU which can be implemented by a low-cost microprocessor to receive the sensor signals from the sensor elements CAM and MIC and to generate user-related input data IN which shall be transmitted to the remote cloud servers CS, one of which is shown in FIG. 2. In order to transmit the input data IN via an IP-based network (see FIG. 1) the mobile input and sensor device MISD also comprises an interface IF in form of a networking module which can be implemented by a wireless Internet access module.
Thus the mobile input and sensor device SD as shown in FIG. 2 can be realized as an universal and low-cost mobile input and sensor device performing an touch-less (optical, acoustical, thermal) and a touch-sensed (touchpad or panel) detection to provide (in uplink direction) user-related input data IN to any remote computer or server CS which runs a entertainment or gaming program. The at least one server CS then produces output data OUT to be transmitted (in downlink direction) back to the users' location for being output at the display device D. The display device D may have its own interface for IP network access or may use the same access interface as common interface CIF to the Internet (also see FIG. 1).
In FIG. 3 a second embodiment of the invention is shown wherein the mobile input and sensor device MISD is integrated into a mobile consumer device for providing entertaining output to the user(s). The mobile input and sensor device MISD can also be wirelessly connected to a game console or any other computing device. In FIG. 3 the mobile consumer device MCD further comprises a projection module P for presenting visual output to the users such as laser beams to project feedback from the game server(s) CS. Further to this there is a wireless network module N for receiving further input data from external user control devices such as keypads, gamepads, joysticks or the like. These data are then also transmitted via the interface module IF to the remote server(s) CS. As shown in FIG. 3 the mobile input and sensor device SD and the display device D can be integrated in the same housing. This means that e.g. the mobile input and sensor device SD can be part of any mobile user device being equipped with an IP interface to be connected with a remote IP-based entertainment service provided by the servers CS. Thus the user(s) only need(s) to set up a single device and can instantly profit from the latest technology since all updates/upgrades are made at the remote site, i.e. on the servers CS.
In particular the mobile input and sensor device SD is integrated into a mobile consumer device MCD, such as a mobile phone or personal assistant having a display D.
By comparing FIGS. 2 and 3 it is clear that the mobile input and sensor device SD can be a separate unit from the display screen or can be wholly integrated with the display screen in one mobile consumer device MCD to form a single unit. As shown in all the FIGs the sensor module SM can comprise multiple types of sensory elements. The preferred sensory elements are optical sensors, in particular two or more cameras, and at least one acoustical sensors such as microphones. Option sensory elements can also be incorporated, such as a thermal sensor.
By this design rich input data IN which relate to the detected user activities are provided to control the system and to enable the user(s) to interact with the entertainment program running on the remote server(s) CS. The user's raw unprocessed input (natural input such as a gesture) is detected by the sensor module SM. A plurality of input streams can be delivered by each user, either simultaneously or sequentially. The input streams detected by the sensor module SM can be passive or active. This means that the active input streams are generated through explicit actions by the user. These actions can consist of speech and physical gestures, for example. The passive input streams can consist of the user's physical appearance (e.g. height or facial features), thermal signature (e.g. body temperature), and other data.
All of the raw inputs are converted into user-related input data IN. These data can be pre-processed, e.g. by applying pattern recognition to detect specific gestures, speech or the like. The pre-processing stage can occur locally on the user's site, i.e. in the mobile input and sensor device SD, or can occur remotely on the cloud computing infrastructure CS. During the pre-processing stage, various parameters such as position, timing, and user differentiation are resolved for each discrete input.
As shown in FIG. 3 by network module N the mobile input and sensor device can also receive further input data from external devices. This means that the user may also use an external input device such as a keyboard. Then the data input streams from such external devices would be received by the module N and could be combined and processed with the raw input streams detected by the sensor module SM. The combined and processed input could then be transmitted over the high-speed IP based data network NET to the cloud infrastructure CS for further processing and interpretation by the software entertainment or gaming application. After the user-related input IN is processed by the application, a response is computed and transmitted back as output OUT to the user over the high-speed data network NET. These output OUT preferably is a (downlink) data stream consisting primarily of high-definition video and audio. Additional data such as text, three-dimensional data, and stereoscopic three-dimensional data may also be transmitted.
This user-related input process repeats itself with a regular frequency (e.g. 30 times per second) throughout the course of the session to deliver a smooth coherent synchronized data stream to the user, typically manifested as one or more of the following: full-motion two-dimensional video, full-motion three-dimensional stereoscopic data, audio, and graphical user interface elements.
The cloud server infrastructure CS provides shared computing and storage resources for all users of the system. The system can thus support a massive number of simultaneous users who can interact with one another either in the same location on a single device, or from different locations with multiple devices. The cloud preferably is a cluster of actual servers that are connected in a self organizing form (said cloud) to distribute its performance in a dynamic and optimal way. This processing cloud needs not to provide the content as such. This can be provided by a further server or cloud (content cloud; not shown here) which represents an Online Game Catalogue providing all data for the desired games to be played by the audiences visiting the theaters. The content data is then transmitted to the processing cloud CS. After calculation these cloud servers CS send the output data OUT back to the client, i.e. to the display device D (at home or in the theater room).
In the field of games the most prominent data which have to be calculated are the visuals of the highest quality. By establishing a cloud of gaming servers CS the client has always the best performance guaranteed no matter what performance the local client is capable of. Depending on the application the service is not limited to graphics only. Any demanding process can be managed by the computing cloud CS. A preferred field of use is theater gaming. Here the small increase in lag (time lapse added due to the transport of the data forth and back) is negligible, and the demand for extreme performance makes it the best choice in terms of cost efficiency.
In summary the present invention relates to a computer-controlled video entertainment system and to a mobile input and sensor device providing user-related input data to said system. The invention provides an easy-to-handle mobile input and sensor device having direct access to an IP-based network, in particular to the Internet, for transmitting sensor signals or data derived thereof as user-related input data via the IP-based network to a remote computing device or server, preferably to a cloud of servers, which then process these input data for controlling the running entertainment program. The mobile input and sensor device can be made simple, since it is mainly designed to detect user activities and to generate sensor signals which are directly transmitted as input data to the remote computer(s) or server(s). Because of the fact that the mobile input and sensor device can directly be connected to the IP-based network (e.g. Internet) it can directly communicate with any IP-based computing device (gaming server or the like) being installed remotely and having powerful hardware and software to process the received input data. Thus there is no need to have a computer or game console being installed at the location of the user(s). This means that the users just need to have at their location (at home, in a cinema . . .) their mobile input and sensor devices and a display device which is connected to the Internet. Thus both devices can communicate with the remote computer(s) which run(s) the entertainment program. The users need not to take care of updating and/or upgrading, but can always profit from the latest technology.
The invention, in particular the cloud being described here, can be used in all fields of providing interactive entertainment applications or content and also of providing non-interactive content. The invention can support Internet applications via any networks and/or media, including WiFi (Wireless Fidelity), BPL (broadband power line), LAN, WiMax (Worldwide Interoperability for Microwave Access) or any other standard technology. The mobile device of the present invention can operate to receive all kinds of interactive, non-interactive applications, such as movies and music, streamed from the cloud.

Claims (27)

What is claimed is:
1. A mobile input and sensor device for a computer controlled video system, the computer-controlled video system comprising a display device for displaying video pictures to at least one user and at least one remote computing device executing a software application for providing output data to be transmitted via a network to the display device to display the video pictures, wherein the mobile input and sensor device is wirelessly linked to the at least one remote computing device, the mobile input and sensor device comprising:
a sensor module including at least one optical sensory element being directed towards the user, configured to detect user activities and produce sensor signals;
at least one input element;
at least one wireless interface configured to transmit user-related input data via the network to said at least one remote computing device;
a further wireless interface configured to receive further input data from external user control devices;
a processing unit configured to combine and process the further input data with said sensor signals to generate said user-related input data representing said detected user activities; and
a visual projection module configured to output visual data to said at least one user as a feedback from the at least one remote computing device,
wherein the mobile input and sensor device is held by the at least one user in front of said display device of the computer-controlled video system and is configured to provide the user-related input data via the network to be processed by said software application executed on the at least one remote computing device.
2. The mobile input and sensor device of claim 1, wherein said at least one optical sensory elements comprises a camera being directed towards the user, in particular the user's face, when the mobile input and sensor device being held by said user.
3. The mobile input and sensor device of claim 1, wherein said at least one input element comprises a touch panel, a keyboard, an array of keys or a key pad to be touched by the user for making user-related inputs.
4. The mobile input and sensor device of claim 1, wherein said sensor module further comprises at least one acoustic sensory element, in particular a microphone, and/or at least one thermal sensory element.
5. The mobile input and sensor device of claim 1, wherein said at least one interface of said mobile input and sensor device is at least one of the following interfaces:
a wireless network interface for an IP-based network, in particular for Internet, a wireless interface module for providing a wireless connection between said mobile input and sensor device and said at least one remote computing device being a single computing devices or a plurality of computing devices, in at least one game console, server and/or server cloud.
6. The mobile input and sensor device of claim 1, wherein said sensory elements are to detect natural user inputs such as physical gestures by said at least one optical sensory elements and/or speech by said at least one or two acoustic elements.
7. The mobile input and sensor device of claim 1, wherein said at least one optical sensory element is an optical sensor for detecting physical gestures or appearance, wherein the sensor module further includes an acoustical sensor for detecting user-produced speech, sound or noise and/or a thermal sensor for detecting physical presence or temperature.
8. The mobile input and sensor device of claim 1, wherein the processing unit provides said user-related input data to said network interface for being transmitted via said IP-based network to said at least one computing device.
9. The mobile input and sensor device of claim 1, wherein said processing unit receives from said sensor module the sensor signals and generates input data which represent said sensor signals.
10. The mobile input and sensor device of claim 1, wherein said processing unit receives from said sensor module the sensor signals and preprocesses these signals to generate said input data, in particular by applying signal and/or data recognition to generate input data which represent recognized user-related input patterns, in particular gestures and/or speech commands.
11. The mobile input and sensor device of claim 1, wherein said processing unit receives sensor signals which relate to a plurality of users and processes these signals to generate collective input data, in particular by applying signal and/or data recognition to generate collective input data which represent recognized multi-user or user-group-related input patterns, in particular collective gestures and/or collective speech commands.
12. The mobile input and sensor device of claim 1, wherein said further wireless interface is configured to receive the further input data from at least one of gamepads, joysticks, keypads, and keyboards.
13. The mobile input and sensor device of claim 1, wherein said mobile input and sensor device is a built-in device being installed in a mobile communications device.
14. A mobile communications device comprising a display device configured to display video pictures to at least one user and a mobile input and sensor device for a computer-controlled video system, the computer-controlled video system comprising at least one remote computing device executing a software application for providing output data to be transmitted via a network to the display device to display the video pictures, wherein the mobile input and sensor device is wirelessly linked to the at least one remote computing device, the mobile input and sensor device comprising:
a touch panel being integrated in said display device;
a sensor module including at least one optical sensory element being directed towards the user, configured to detect user activities and produce sensor signals;
at least one input element;
at least one wireless interface configured to transmit user-related input data via the network to said at least one remote computing device;
a further wireless interface configured to receive further input data from external user control devices;
a processing configured to combine and process the further input data with said sensor signals to generate said user-related input data representing said detected user activities; and
a visual projection module configured to output visual data to said at least one user as a feedback from the at least one remote computing device,
wherein the mobile input and sensor device is configured to provide the user-related input data via the network to be processed by said software application executed on the at least one remote computing device.
15. A computer-controlled video system, the system comprising:
a display device configured to display video pictures to at least one user;
at least one remote computing device, in particular at least one server, being installed remotely from said display device and executing a software application for providing output data to be transmitted via an IP-based network to the display device to display the video pictures; and
a mobile input and sensor device being held by the at least one user in front of said display device and wirelessly linked to the at least one remote computing device and configured to provide user-related input data via the IP-based network to be processed by said software application executed on the at least one remote computing device, wherein the mobile input and sensor device comprises:
a sensor module including at least one optical sensory element being directed towards the user, configured to detect user activities and produce sensor signals,
at least one input element,
at least one wireless interface configured to transmit the user-related input data via the IP-based network to said at least one remote computing device,
a further wireless interface configured to receive further input data from external user control devices,
a processing unit configured to combine and process the further input data with said signals to generate said user-related input data representing said detected user activities, and
a visual projection module configured to output visual data to said at least one user as a feedback from the at least one remote computing device.
16. The computer-controlled video system of claim 15, wherein the at least one remote computing device is constituted by a plurality of remotely installed computing servers being linked to each other to form a cluster or cloud of computing servers to form an overall cloud.
17. The computer-controlled video system of claim 16, wherein the computing servers are linked to each other via said IP-based network.
18. The computer-controlled video system of claim 15, wherein said IP-based network is constituted by Internet.
19. The computer-controlled video system of claim 15, wherein said network interface of said mobile input and sensor device is a wireless Internet module.
20. The computer-controlled video system of claim 15, wherein said sensor module is for detecting user activities, in particular natural user inputs such as physical gestures and/or speech.
21. The computer-controlled video system of claim 15, wherein the processing unit provides said user-related input data to said network interface for being transmitted via said IP-based network to said at least one computing device.
22. The computer-controlled video system of claim 15, wherein said at least one optical sensor element is an optical sensor for detecting physical gestures or appearance, the sensor module further including an acoustical sensor for detecting user-produced speech, sound or noise and/or a thermal sensor for detecting physical condition by temperature.
23. The computer-controlled video system of claim 15, wherein said processing unit receives from said sensor module the sensor signals and generates raw input data which represent said sensor signals.
24. The computer-controlled video system of claim 15, wherein said processing unit receives from said sensor module the sensor signals and pre-processes these signals to generate said input data, in particular by applying signal and/or data recognition to generate input data which represent recognized user-related input patterns, in particular gestures and/or speech commands.
25. The computer-controlled video system of claim 15, wherein said processing unit receives sensor signals which relate to a plurality of users and processes these signals to generate collective input data, in particular by applying signal and/or data recognition to generate collective input data which represent recognized multi-user or user-group-related input patterns, in particular collective gestures and/or collective speech commands.
26. The computer-controlled video system of claim 15, wherein the visual projection module of the mobile input and sensor device is configured to output a laser-based projection allowing user activities.
27. The computer-controlled video system of claim 15, wherein said further wireless interface is configured to receive the further input data from at least one of gamepads, joysticks, keypads, and keyboards.
US12/974,029 2009-12-30 2010-12-21 Mobile input and sensor device for a computer-controlled video entertainment system Active 2032-01-19 US9344753B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US12/974,029 US9344753B2 (en) 2009-12-30 2010-12-21 Mobile input and sensor device for a computer-controlled video entertainment system

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US29103209P 2009-12-30 2009-12-30
US12/974,029 US9344753B2 (en) 2009-12-30 2010-12-21 Mobile input and sensor device for a computer-controlled video entertainment system

Publications (2)

Publication Number Publication Date
US20110159962A1 US20110159962A1 (en) 2011-06-30
US9344753B2 true US9344753B2 (en) 2016-05-17

Family

ID=43638429

Family Applications (1)

Application Number Title Priority Date Filing Date
US12/974,029 Active 2032-01-19 US9344753B2 (en) 2009-12-30 2010-12-21 Mobile input and sensor device for a computer-controlled video entertainment system

Country Status (2)

Country Link
US (1) US9344753B2 (en)
EP (1) EP2347810A3 (en)

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170039165A1 (en) * 2015-08-04 2017-02-09 Zhejiang Dafeng Industry Co., Ltd. Multidimensional stage adaptive transmission control system
US10760803B2 (en) 2017-11-21 2020-09-01 Emerson Climate Technologies, Inc. Humidifier control systems and methods
US11226128B2 (en) 2018-04-20 2022-01-18 Emerson Climate Technologies, Inc. Indoor air quality and occupant monitoring systems and methods
US11371726B2 (en) 2018-04-20 2022-06-28 Emerson Climate Technologies, Inc. Particulate-matter-size-based fan control system
US11421901B2 (en) 2018-04-20 2022-08-23 Emerson Climate Technologies, Inc. Coordinated control of standalone and building indoor air quality devices and systems
US11486593B2 (en) 2018-04-20 2022-11-01 Emerson Climate Technologies, Inc. Systems and methods with variable mitigation thresholds
US11609004B2 (en) 2018-04-20 2023-03-21 Emerson Climate Technologies, Inc. Systems and methods with variable mitigation thresholds
US11994313B2 (en) 2018-04-20 2024-05-28 Copeland Lp Indoor air quality sensor calibration systems and methods
US12018852B2 (en) 2018-04-20 2024-06-25 Copeland Comfort Control Lp HVAC filter usage analysis system
US12078373B2 (en) 2018-04-20 2024-09-03 Copeland Lp Systems and methods for adjusting mitigation thresholds

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7938727B1 (en) * 2007-07-19 2011-05-10 Tim Konkle System and method for providing interactive content for multiple networked users in a shared venue
US9067150B2 (en) 2008-01-19 2015-06-30 Lamplight Games System and method for providing interactive content for multiple networked users in a shared venue using short messaging service communication
US8776132B2 (en) * 2011-03-17 2014-07-08 Ebay Inc. Single digital wallet across multiple payment platforms
US10075630B2 (en) 2013-07-03 2018-09-11 HJ Laboratories, LLC Providing real-time, personal services by accessing components on a mobile device
WO2015088212A1 (en) 2013-12-09 2015-06-18 Samsung Electronics Co., Ltd. Digital photographing apparatus capable of reconfiguring image signal processor and method of controlling the same

Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5801754A (en) 1995-11-16 1998-09-01 United Artists Theatre Circuit, Inc. Interactive theater network system
WO2001024473A1 (en) 1999-09-27 2001-04-05 Koninklijke Philips Electronics N.V. Pda has wireless modem for remote control via the internet
US20060230073A1 (en) * 2004-08-31 2006-10-12 Gopalakrishnan Kumar C Information Services for Real World Augmentation
US20080039202A1 (en) 2006-08-02 2008-02-14 Nintendo Co., Ltd. Game apparatus having general-purpose remote control function
WO2008069519A1 (en) 2006-12-04 2008-06-12 Electronics And Telecommunications Research Institute Gesture/speech integrated recognition system and method
US20080273755A1 (en) 2007-05-04 2008-11-06 Gesturetek, Inc. Camera-based user input for compact devices
US20090227232A1 (en) * 2008-03-04 2009-09-10 Apple Inc. Access Management
US20090327888A1 (en) * 2005-05-31 2009-12-31 Ipifini, Inc. Computer program for indentifying and automating repetitive user inputs
US7647560B2 (en) * 2004-05-11 2010-01-12 Microsoft Corporation User interface for multi-sensory emoticons in a communication system
US20100053591A1 (en) * 2007-12-05 2010-03-04 Microvision, Inc. Scanned Proximity Detection Method and Apparatus for a Scanned Image Projection System
US20110084795A1 (en) * 2009-10-14 2011-04-14 Masahiro Fukuyori Systems and Methods for Dynamically Changing Alerts of Portable Devices Using Brainwave Signals
US20110154266A1 (en) * 2009-12-17 2011-06-23 Microsoft Corporation Camera navigation for presentations

Patent Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5801754A (en) 1995-11-16 1998-09-01 United Artists Theatre Circuit, Inc. Interactive theater network system
WO2001024473A1 (en) 1999-09-27 2001-04-05 Koninklijke Philips Electronics N.V. Pda has wireless modem for remote control via the internet
US7647560B2 (en) * 2004-05-11 2010-01-12 Microsoft Corporation User interface for multi-sensory emoticons in a communication system
US20060230073A1 (en) * 2004-08-31 2006-10-12 Gopalakrishnan Kumar C Information Services for Real World Augmentation
US20090327888A1 (en) * 2005-05-31 2009-12-31 Ipifini, Inc. Computer program for indentifying and automating repetitive user inputs
US20080039202A1 (en) 2006-08-02 2008-02-14 Nintendo Co., Ltd. Game apparatus having general-purpose remote control function
WO2008069519A1 (en) 2006-12-04 2008-06-12 Electronics And Telecommunications Research Institute Gesture/speech integrated recognition system and method
US20080273755A1 (en) 2007-05-04 2008-11-06 Gesturetek, Inc. Camera-based user input for compact devices
US20100053591A1 (en) * 2007-12-05 2010-03-04 Microvision, Inc. Scanned Proximity Detection Method and Apparatus for a Scanned Image Projection System
US20090227232A1 (en) * 2008-03-04 2009-09-10 Apple Inc. Access Management
US20110084795A1 (en) * 2009-10-14 2011-04-14 Masahiro Fukuyori Systems and Methods for Dynamically Changing Alerts of Portable Devices Using Brainwave Signals
US20110154266A1 (en) * 2009-12-17 2011-06-23 Microsoft Corporation Camera navigation for presentations

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
European Search Report for corresponding Application No. 10 196 839.4 dated Apr. 4, 2011.
Nintendo DSi; Dec. 27, 2009, XP002631195, Retrieved from the Internet: URL:http://en.wikipedia.org/w/index.php?title=Nintendo DSi&oldid=334174919 [European Search Report copy retrieved on Apr. 4, 2011. Submitted copy retrieved Aug. 30, 2011].

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170039165A1 (en) * 2015-08-04 2017-02-09 Zhejiang Dafeng Industry Co., Ltd. Multidimensional stage adaptive transmission control system
US10127189B2 (en) * 2015-08-04 2018-11-13 Zhejiang Dafeng Industry Co., Ltd. Multidimensional stage adaptive transmission control system
US10760803B2 (en) 2017-11-21 2020-09-01 Emerson Climate Technologies, Inc. Humidifier control systems and methods
US10760804B2 (en) 2017-11-21 2020-09-01 Emerson Climate Technologies, Inc. Humidifier control systems and methods
US10767878B2 (en) 2017-11-21 2020-09-08 Emerson Climate Technologies, Inc. Humidifier control systems and methods
US11226128B2 (en) 2018-04-20 2022-01-18 Emerson Climate Technologies, Inc. Indoor air quality and occupant monitoring systems and methods
US11371726B2 (en) 2018-04-20 2022-06-28 Emerson Climate Technologies, Inc. Particulate-matter-size-based fan control system
US11421901B2 (en) 2018-04-20 2022-08-23 Emerson Climate Technologies, Inc. Coordinated control of standalone and building indoor air quality devices and systems
US11486593B2 (en) 2018-04-20 2022-11-01 Emerson Climate Technologies, Inc. Systems and methods with variable mitigation thresholds
US11609004B2 (en) 2018-04-20 2023-03-21 Emerson Climate Technologies, Inc. Systems and methods with variable mitigation thresholds
US11994313B2 (en) 2018-04-20 2024-05-28 Copeland Lp Indoor air quality sensor calibration systems and methods
US12018852B2 (en) 2018-04-20 2024-06-25 Copeland Comfort Control Lp HVAC filter usage analysis system
US12078373B2 (en) 2018-04-20 2024-09-03 Copeland Lp Systems and methods for adjusting mitigation thresholds

Also Published As

Publication number Publication date
US20110159962A1 (en) 2011-06-30
EP2347810A2 (en) 2011-07-27
EP2347810A3 (en) 2011-09-14

Similar Documents

Publication Publication Date Title
US9344753B2 (en) Mobile input and sensor device for a computer-controlled video entertainment system
EP2340878B1 (en) Sensor device for a computer-controlled video entertainment system
US9486701B2 (en) Computer-controlled video entertainment system
US11065551B2 (en) Virtual reality presentation of real world space
CN115068932B (en) View rendering method at view position in virtual reality environment
US20210344991A1 (en) Systems, methods, apparatus for the integration of mobile applications and an interactive content layer on a display
US11538213B2 (en) Creating and distributing interactive addressable virtual content
EP3681609B1 (en) Cross-platform interactive streaming
US20180316948A1 (en) Video processing systems, methods and a user profile for describing the combination and display of heterogeneous sources
US20200404219A1 (en) Immersive interactive remote participation in live entertainment
US20180316939A1 (en) Systems and methods for video processing, combination and display of heterogeneous sources
US20180316947A1 (en) Video processing systems and methods for the combination, blending and display of heterogeneous sources
US20180316942A1 (en) Systems and methods and interfaces for video processing, combination and display of heterogeneous sources
US11284137B2 (en) Video processing systems and methods for display, selection and navigation of a combination of heterogeneous sources
US9066144B2 (en) Interactive remote participation in live entertainment
US20180316943A1 (en) Fpga systems and methods for video processing, combination and display of heterogeneous sources
US20180316944A1 (en) Systems and methods for video processing, combination and display of heterogeneous sources
KR20230017925A (en) Methods, devices, and systems for interactive cloud gaming
US20180316946A1 (en) Video processing systems and methods for display, selection and navigation of a combination of heterogeneous sources
CN105430455A (en) Information presentation method and system
US8439751B2 (en) Multi-user computer-controlled input system and a method of compressing and processing multiple input data
WO2018071781A2 (en) Systems and methods for video processing and display
US20180316940A1 (en) Systems and methods for video processing and display with synchronization and blending of heterogeneous sources
US20180316941A1 (en) Systems and methods for video processing and display of a combination of heterogeneous sources and advertising content
KR20100136415A (en) Computer method and apparatus providing interactive control and remote identity through in-world proxy

Legal Events

Date Code Title Description
AS Assignment

Owner name: CRYTEK GMBH, GERMANY

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:YERLI, CEVAT;REEL/FRAME:025886/0221

Effective date: 20110228

AS Assignment

Owner name: CRYTEK GMBH, GERMANY

Free format text: CHANGE OF ASSIGNEE ADDRESS;ASSIGNOR:CRYTEK GMBH;REEL/FRAME:029397/0016

Effective date: 20121204

AS Assignment

Owner name: GFACE GMBH, GERMANY

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:CRYTEK GMBH;REEL/FRAME:029639/0018

Effective date: 20130110

AS Assignment

Owner name: CRYTEK GMBH, GERMANY

Free format text: MERGER;ASSIGNOR:GFACE GMBH;REEL/FRAME:032485/0763

Effective date: 20140129

STCF Information on status: patent grant

Free format text: PATENTED CASE

FEPP Fee payment procedure

Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

CC Certificate of correction
FEPP Fee payment procedure

Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

FEPP Fee payment procedure

Free format text: SURCHARGE FOR LATE PAYMENT, LARGE ENTITY (ORIGINAL EVENT CODE: M1554); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 4

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 8