WO2007078639A1 - Methods and systems for enabling direction detection when interfacing with a computer program - Google Patents
Methods and systems for enabling direction detection when interfacing with a computer program Download PDFInfo
- Publication number
- WO2007078639A1 WO2007078639A1 PCT/US2006/047111 US2006047111W WO2007078639A1 WO 2007078639 A1 WO2007078639 A1 WO 2007078639A1 US 2006047111 W US2006047111 W US 2006047111W WO 2007078639 A1 WO2007078639 A1 WO 2007078639A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- computer program
- location
- recited
- display screen
- image
- Prior art date
Links
Classifications
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/40—Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment
- A63F13/42—Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment by mapping the input signals into game commands, e.g. mapping the displacement of a stylus on a touch screen to the steering angle of a virtual vehicle
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/20—Input arrangements for video game devices
- A63F13/21—Input arrangements for video game devices characterised by their sensors, purposes or types
- A63F13/219—Input arrangements for video game devices characterised by their sensors, purposes or types for aiming at specific areas on the display, e.g. light-guns
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/017—Gesture based interaction, e.g. based on a set of recognized hand gestures
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F2300/00—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
- A63F2300/10—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals
- A63F2300/1012—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals involving biosensors worn by the player, e.g. for measuring heart beat, limb activity
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F2300/00—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
- A63F2300/10—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals
- A63F2300/1087—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals comprising photodetecting means, e.g. a camera
- A63F2300/1093—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals comprising photodetecting means, e.g. a camera using visible light
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F2300/00—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
- A63F2300/60—Methods for processing data by generating or executing the game program
- A63F2300/6045—Methods for processing data by generating or executing the game program for mapping control signals received from the input arrangement into game commands
Definitions
- Example gaming platforms may he the Sony Playstation or Sony Playstation2 (PS2), each of which is sold in the form of a game console.
- the game console is designed to connect to a monitor (usually a television) and enable user interaction through handheld controllers.
- the game console is designed with specialized processing hardware, including a CPU, a graphics synthesizer for processing intensive graphics operations, a vector unit for performing geometry transformations, and other glue hardware, firmware, and software.
- the game console is further designed with an optical disc tray for receiving game compact discs for local play through the game console. Online gaming is also possible, where a user can interactively play against or with other users over the Internet.
- the present invention fills these needs by providing an apparatus and method that facilitates interactivity with a computer program.
- the computer program is a game program, but without limitation, the apparatus and method can find applicability in any consumer electronic device that will require a user to interact therewith.
- the present invention simplifies user interaction experience through machine recognizable gestures based on pointing to the interface and discriminating commands based on factors including trigger cues and position determination of a hand or object under user control.
- a method for detecting direction when interfacing with a computer program includes capturing an image presented in front of an image capture device.
- the image capture device has a capture location in a coordinate space.
- the method includes identifying a human head in the image and assigning the human head a head location in the coordinate space.
- the method also includes identifying an object held by the person in the image and assigning the object an object location in coordinate space.
- the method further includes identifying a relative position in coordinate space between the head location and the object location when viewed from the capture location. The relative position defines a pointing direction of the object when viewed by the image capture device.
- a method for detecting pointing direction of an object directed toward a display screen that can render graphics of a computer program includes capturing an image presented in front of an image capture device.
- the image capture device has a capture location in a coordinate space that is proximate to the display screen.
- the method includes identifying a first body part of the person in the image and assigning the first body part a first location in the coordinate space.
- the method includes identifying a second body part of the person in the image and assigning the second body part a second location in coordinate space.
- an apparatus for capturing image and sound during interactivity with a computer program includes an image capture unit that is configured to capture one or more image frames. These image frames are analyzed to identify a person's head and a person's hand. Based on the relative positioning of the hand and head, the apparatus can ascertain a pointing direction, which may be translated as the pointing location on a display screen.
- Figure IA and IB illustrate the environment of one embodiment of present invention, which takes advantage of the pointing direction determination described herein.
- Figures 2-4 illustrate the analysis of the relative positions between a person's head and a person's hand (or object held by the person), when determining pointing direction.
- Figures 5A-5C illustrate tracking and triggering embodiments, which can be identified by the computer program by the analysis of the captured image data, in accordance with one embodiment of the present invention.
- Figures 6A-6F illustrate alternate embodiments of detecting an object and detecting changes in the object based on relative orientation of the object itself, in accordance with one embodiment of the present invention.
- J Figures 7 and 8A-8B illustrate examples of hand positions, when the hand positions and shapes are analyzed to determine a desired trigger or tracking response by the computer program, in accordance with one embodiment of the present invention.
- FIGS. 9 and 10 provide exemplary flow charts of the operations that can be executed in determining a pointing direction, in accordance with one embodiment of the present invention.
- Figures 11 and 12 provide exemplary embodiments of hardware that may be used in processing the computer code necessary to executed the claimed operations, in accordance with one embodiment of the present invention.
- Figures 13A-13C illustrate exemplary embodiments for when depth information is considered when identifying objects to track, when interfacing with a system that needs to ascertain pointing direction, in accordance with one embodiment of the present invention.
- FIG. IA illustrates an interactive game setup 100, in accordance with one embodiment of the present invention.
- the interactive game setup 100 includes a computer 102 that is coupled to a display screen 110.
- An image capture device 105 is placed on top of the display screen 110 and is coupled to the computer 102.
- Computer 102 is, in one embodiment, a gaming system console which allows users to play video games and interface with the video games through controllers 108.
- the image capture device 105 is shown placed on top of the display screen 110, but it should be understood that the image capture device 105 can be placed in any other proximate location that will allow it to capture images that are located about in front of the display screen 110.
- the game console can be a one manufactured by Sony Computer Entertainment Inc., Nintendo, Microsoft, or any other manufacturer.
- the image capture device 105 can be as simple as a standard web cam or can include more advanced technology. In one embodiment, the image capture device should be capable of capturing images, digitizing the images, and communicating the image data back to the computer 102.
- the image capture device will have logic integrated therein for performing the digitizing and another embodiment the image capture device 105 will simply transmit the captured data back to the computer 102 for digitizing. In either case, the image capture device 105 is capable of capturing either color or black and white images of any object located in front of the image capture device 105.
- Figure IB illustrates an exemplary embodiment of the present invention in which the computer 102 processes image data provided by the image capture device 105 to ascertain a pointing direction of an object placed in front of the image capture device 105. As shown, the computer 102 is connected to the image capture device 105. The image capture device 105 is designed to focus onto a capture region 105a. In this example, a person 112 is intending to interact with a computer program being executed by the computer 102. The computer program, in this example, is a video game which is rendered and displayed by the display screen 110.
- the video game is a target shooting game in which the person 112 wishes to aim at a target and earn points commensurate with his or her performance.
- an image 112' of the person 112 may also be placed on the display screen 110 during game play.
- the person's image 112' may be omitted from the display screen, depending on the particular device under control or game being played.
- the user experience may be enhanced by illustrating an image 112' of the person 112 during the target shooting exercise to present more reality during game play.
- a feature of the target shooting game is the ability for person 112 to point or direct an object 124 at particular interactive graphics on the display screen 110.
- the person 112 will hold the object 124 with his or her hand 122.
- the hand 122 will be directionally pointed toward the display screen 110.
- the image capture device 105 will at this point, analyze the digital image capture of the person 112 to determine the location of the person's 112 head 120, and the location of the person's 112 hand 122.
- the person's 112 hand is extended in front of his body and the image capture device will identify the object 124 when examining the captured digital image.
- the captured digital image will also be examined by code executed at the computer 102 to ascertain the location of the person's 112 head 120.
- head tracking is completed with a combination of a template matching (for speed performance), coupled to a face detection code.
- the face detection code will essentially identify the location of the user's face by locating the user's eyes and other facial features.
- the object 124 will, in one embodiment, have an identifier which may be color or lights (e.g., light emitting diodes "LEDs") coupled to the object so that the program analyzing the captured digital image will easily identify the location of the object 124.
- an identifier which may be color or lights (e.g., light emitting diodes "LEDs" coupled to the object so that the program analyzing the captured digital image will easily identify the location of the object 124.
- the computer program Once the computer program has identified the location of the person's head 120 (H) and the location of the person's hand 122 (h), the computer program will perform computations to determine a relative angle from the image capture device position, and between the detected object 124, and the head 120.
- the relative position of the object 124 and the head 120 will be calculated relative to the image capture device 105. This will produce two angle parameters (theta ⁇ and phi ⁇ ).
- the azimuth angle ⁇ will define the horizontal positioning between the head 120 and the object 124 along an X axis.
- the phi angle ⁇ will produce the altitude angle which is the relative angle between the height of the head 120 and the height of the hand 122.
- an initial calibration operation may be performed before a gaming operation begins to calibrate the object's pointing location on the display screen 110. For instance, the user may be prompted to calibrate the pointing algorithm by having the user point the object 124 at a specific location on the display screen 110.
- the computer 102 will be able to calculate the azimuth angle and the altitude angle (theta and phi) which define the relative positions of the person's head 120 and the person's hand 122, for each successive frame being captured by the image capture device 105.
- the relative positioning between the head and the hand may be calculated for each captured frame or may be captured every other frame, or after a number of frames are captured, depending on the accuracy required for the pointing operation. For example, if the game is a shooting gallery game, it would be important for the relative positioning of the head 120 and the hand 122 to be computed for each frame so that the person 112 will have accurate aiming and triggering capabilities when attempting to secure a good performing score in the video game contest.
- FIG. 2 illustrates a top view of the person 112 who is positioned in front of the display screen 110, in accordance with one embodiment of the present invention.
- an initial azimuth angle (thetai) is shown being determined as the relative angle between the position of the hand 122 (which is holding object 124), and the head 120.
- the person 112 during interactive play with the computer program, will be facing the display screen 1 10 and most likely, will maintain his body substantially parallel to the display screen 110.
- movement of the hand 122 in direction 123 will cause the azimuth angle to be recomputed and produce a new azimuth angle theta 2 .
- the person 112 is holding the object 124 out in front of his body at about arm's length.
- This distance is shown to be the approximate arm's length detect (ALD) that defines the location of the approximate place where the image capture device 105 will attempt to capture the position of the hand 122 and associated object 124.
- the approximate arm's length detect (ALD) can vary, depending upon the user's arm length, but generally should be provided to allow a distance relationship between the location of the head 120 and the hand 122. For instance, there should be at least a minor projection of the hand 122 in front of the person's body to point to different locations of the display screen 110.
- Figure 3 illustrates a front view of the person 112 where the head 120 is identified and the hand 122 is identified, from the captured digital image.
- a dashed line represents the display screen 110 and the image capture device 105 that is directionally pointed at the person 112.
- the image capture device 105 will be illustrated to be at a coordinate space of (0, 0, 0), representing the reference point of the image capture device 105, and its position in coordinate space.
- the approximate location of the head 120 will also have an associated coordinate space (x H, y H, z H).
- Figure 4 illustrates the person 112 pointing the object 124 at the display screen 110.
- the coordinate space of the head 120 is identified in the digital image captured by the capture device 105.
- the location in coordinate space of the hand 122 is also captured in the digital image captured by the image capture device 105 when the person 112 is pointing at the display screen 110.
- the image capture device 105 is the reference point, which is at the coordinate space (0, 0, 0).
- the altitude angle phi is therefore calculated between the position of the head 120 relative to the position of the hand 122. In one example, the angle is calculated as:
- Altitude angle arctan ((yh-yH)/(zh-zH)) [0031]
- Figure 5A illustrates an embodiment where the object 124 is a gun.
- the gun may be a simplified gun object that is configured to be pointed in the direction of the display screen 110 to hit particular objects or interact with objects on the display screen 110.
- the gun 124 will include a detection region 124a.
- Detection region 124a is the region which is directionally pointed toward the display screen 110.
- the detection region 124a is also the region that is captured by the image capture device 105 for analysis of the digital image by the computer 102.
- the detection region 124a is configured to include a pair of lighting objects that will assist in the interactivity with the computer program being executed by the computer 102 and displayed on the display screen 110.
- a tracking indicator 130 is provided as a light or color object that is present on the detection region 124a. Based on the tracking indicator 130, the image capture device 105 will produce a digital image that will be analyzed by the computer 102 to identify the position in coordinate space of the object 124. In this example, by providing the tracking indicator 130, the computer program being executed on the computer 102 is able to quickly identify the location of the object 124 and in relation to the head 120 of the person interacting with the computer program.
- the tracking indicator 130 may be provided by way of a number of implementations.
- One implementation might be a light indicator that can be tracked by the computer program that analyzes the captured digital images, or may be in the form of a color indicator that the computer can identify quickly from the captured digital images.
- the hand itself may be the tracking indicator 130.
- the tracing indicator 130 may be provided as a reflective tape that will have different coloring or intensity depending on the angle that it might be displayed when shown to the image capture device 105.
- the object 104 is tracked as the user moves his hand 122 to different regions pointed to on the display screen 110.
- the tracking indicator 130 will allow the computer program to provide a visual indicator on the display screen 110.
- This visual indicator on the display screen 110 will allow the user to understand where the object is currently pointing to on the display screen 110.
- the detection region 124a will also include a trigger indicator 132.
- the trigger indicator 132 may be in the form of a light that is triggered ON and OFF when the user pulls the trigger of the object 124.
- the detection region 124a is shown in Figure 5B after the trigger has been pulled and the trigger indicator 132 is lit.
- the computer program executing on the computer 102 will provide an indictor on the display screen 110 so that the user can identify whether his or her pointing has accurately hit an object of the computer game.
- the trigger indicator 132 is shown to be in the OFF position to signify that the object 124 will still remain actively tracked, but the shooting which can be continuous or intermittent, can be discontinued when the user removes his finger from the trigger of the object 124.
- the trigger indicator 132 may be an any frequency range, including audio, ultrasonic, visible light wave, infrared and radio. Passive trigger indicator 132 may be achieved through a mechanical sound generated upon actuating a trigger and receiving and decoding the audio input to the device and determining whether the trigger was actuated.
- Figure 6 A illustrates another embodiment of the present invention where tracking and trigger indicators 130a and 132a are provided.
- the track/trigger indicators 130a and 132a are provided so that determinations can be made of the relative distances between the two indicators as shown by distance (di).
- the object when pointed at the image capture device 105 may respond by having the computer program that is executed on the computer 102 to ascertain the distance di and perform an interactive action on the display screen.
- a second distance (d 2 ) is computed. This distance is the new distance between the track/trigger indicators 130a and 132a.
- the distance continues to shrink as shown by distance (d 3 ).
- distance (d 3 ) the distance between the track and trigger indicators 130a and 132a is brought to approximately zero.
- the program may read that the user intends for a trigger action to occur, or any other action that can be triggered when the detected distance from d ⁇ to d 4 has been detected.
- the response by the computer program may be gradually changed, depending on the angle at which the detection region 124b is tilted. For instance, the user may immediately begin to shoot the gun (or trigger the shooting of the gun) when the tilting begins and is executed between the tilt of Figure 6B to the tilt of Figure 6D. When the user tilts the gun back to the original position, the gun may discontinue the shooting activity. Consequently, the trigger activity caused by analyzing the patterns or colors of the tracking and trigger indicators of 130a and 132b can cause the computer program to react in different interactive ways.
- Commands and trigger states are not limited to an ON and OFF parameters, but can be incrementally changed depending on the position of the relative state and angles of the trigger and track indicators.
- the state of the trigger may be determined in a linear or variable state as opposed to ON or OFF. Any technique can be used to determine the relative trigger position including a resistive types used to control acceleration in remote control race tracks.
- the device, or gun in this example can communicate the state of its trigger by encoding and transmitting its value in any of a multitude of ways known in the art.
- a variety of commands and gestures may be formulated based on the state of the trigger and the position of the device, including those based on all machine recognizable gestures.
- the track and trigger indicators 130b and 132b are square or rectangular in dimension and can be defined by colored tapes, bar codes, light indicators, LEDs, or the like.
- FIG. 7 illustrates another embodiment of the present invention in which a hand 122 is used to control the interactivity on the display screen.
- the user may place the hand 122 in front of the image capture device 105 and it is tracked relative to the head of the person 112.
- the user's hand is extended having of an index finger 140 pointing at the display screen 110.
- the user's thumb 142 may be pointing upright to indicate to the computer program executing on the computer 102 that the trigger device has not been activated.
- the computer program executing on a computer 102 may detect from the captured digital image that the user intends to shoot or trigger or interactively point to a specific region on the display screen 1 10.
- the user's hand being placed in a different position can trigger an event or cause the interactivity of a command with a computer program being executed and shown on the display screen 110.
- the user may be able to shoot by different hand gestures, may be able to reload the gun with different gestures, and the different positions or orientations of the user's hand may cause different graphical renderings of the user or gun on the display screen when the user is interacting with a particular game program.
- the state of the trigger may be determined ON or OFF as described above, or, the trigger may be determined in a variable state.
- the relative position of the user's thumb may range from, for example, the position in which the thumb is substantially perpendicular to the pointing finger and the position where the thumb is substantially parallel to the pointing finger and where the system performs image analysis of the hand to yield a relative state of the thumb.
- This state may be mapped to various control schemes, including those relating to a scroll-wheel on mouse.
- Figures 8A and 8B illustrate another example where the user's hand 122 may be used to identify or trigger different activity on the display screen.
- Figure 8A shows the user pointing directly at the display screen 110 and the computer program being executed to identify this hand orientation on the digital image that was captured by the capture device.
- the image of Figure 8B might be captured and analyzed by the computer program.
- the user's hand is shown tilted from where the thumb 142 is pointing upright to where the thumb 142 is pointing sideways. In this orientation, the user may be able to trigger the shooting of a gun on the display screen, or the selection of an icon on a graphical display. [0044] Consequently, the detection of the user's hand and the orientation of the user's hand can be used to provide the interactivity necessary when analyzing the position of the user's hand relative to the user's head, and the pointing activity on a display screen.
- the pointing activity will allow the user to select icons, control the device under operation, shoot at graphical objects, select or scroll graphical objects, de-select graphical objects, turn ON and OFF graphical objects, disperse graphical objects, or simply interface with the graphics icons and features of a computer program being displayed on a display screen 110.
- the system employing the present invention may operate with minimal or no icons on a display screen. Instead, the system may simply just recognize the gestures of the user and provide a control input to the device under operation.
- a television or related peripheral configured or integrated with the present invention may be controlled by the present invention. Changing a channel, for example, may not necessarily involve interacting with an icon as opposed to recognizing a gesture command according to the scheme presented in the present invention.
- Figure 9 illustrates a flowchart diagram 200 where a process for identifying a pointing direction is described, in accordance with one embodiment of the present invention.
- the method begins by providing a capture device for capturing image frames.
- the frame capture rate will depend on the particular implementation, and will not limit the invention described herein.
- the capture device will, in one embodiment, include a capture location in coordinate space. As shown in Figure 4, the coordinate space of the image capture device is (0, 0, 0).
- a display screen is provided for rendering and displaying the interactive graphics of a computer program.
- the interactive graphics of the computer program may be associate with a computer game, but may also be associated with any other program or interface that may require interactivity by a user.
- the interactivity may include the selection or de-selection of objects, opening files, changing channels, recording TV shows, closing files, accessing files over a network, or interactively communicating with users by way of the internet, electronic mail or by electronic video mail, selecting a consumer electronic device, turning a device ON or OFF.
- the operation 206 will include the capturing of an image that is presented substantially in front of the image capture device.
- the image will include a person that is part of the captured space, and who is in front of the display screen and the capture device.
- operation 208 will include the identification of a human head of the person that is found in the captured image.
- the identified human head will therefore be analyzed to ascertain the head location in the coordinate space relative to the capture location.
- the method then moves to operation 210 where an object held by the person in the image is identified.
- the object's location is identified such that the coordinate space of the object is identified relative to the coordinate space of the capture location.
- the computer program can identify a relative position in coordinate space between the head location and the object location when viewed from the capture location reference point (e.g., coordinate (0,0,0)).
- an azimuth angle and an altitude angle can be computed for the relative locations of the head and the hand relative to the image capture device. This relative position in coordinate space is calculated for the captured frame.
- FIG. 10 illustrates a more detailed process diagram 250 that can be implemented when determining the pointing direction of an object that is directionally pointed at a display screen during an interactivity with a computer program.
- the method begins at operation 252 where the capture device for capturing image frames is provided.
- the capture device will have a capture location in coordinate space.
- the coordinate space of the capture location will be the reference point for performing operations to determine relative locations in the process of identifying directionality pointing.
- the method moves to operation 254 where a display .screen is provided for rendering interactive graphics of the computer program.
- the interactive graphics may be a computer game or may be any other program as defined above.
- operation 256 an image is captured in front of the image capture device and a person is captured in the image.
- the captured image may be that of a digital frame of video.
- the digital frame of video may be a JPEG frame or may be part of a compressed video frame (e.g., MPEG or the like).
- the operation moves to the identification of a human head of the person in the captured image in operation 258.
- the human head is analyzed on the captured image to determine a head location and the coordinate space relative to the image capture device.
- the method moves to the identification of an object held by the person in the image and determining an object location in the coordinate space.
- a relative position is identified in the coordinate space between the head location and the object location when viewed from the capture location of the capture device.
- the relative position will include a calculation of an azimuth angle and an altitude angle relative to the image capture device.
- operation 264 during execution of the computer program, the operations identified as A, B, C and D corresponding to operations 256, 258, 260, and 262 will be performed iteratively and continuously, depending on a rate desired for the performance of a computer program. For instance, the execution of operations A through D will occur at the rate of one time for each frame that is captured or only after a certain number of frames are captured. The rate at which operations A through D are performed will therefore depend on the specific environment and the need for accurate detection of the pointer location and the select ability of interactive objects on the display screen. If the display screen is processing a video game that has objects that are moving at rapid rates, the tracking operation may require that operations A through D be performed for each frame that is displayed on the video display screen.
- the method indicates a continual update of the point and direction of the object using the relative position.
- the pointing direction is displayed on the display screen to enable interaction with the interactive graphics of the computer program. It should again be understood that the pointing direction can be to enable a user to select icons, de-select icons, move icons, open objects, open files, save files, move files, and interact with files that may be part of a file database, or part of a graphical user interface on a computer desktop or the like.
- Figure 1 1 is a block diagram of an exemplary user input system for interaction with an object on a graphical display that can be used to implement embodiments of the present invention.
- the user input system is comprised of a video capture device 300, an input image processor 302, an output image processor 304, and a video display device 306.
- Video capture device 300 may be any device capable of capturing sequences of video images, and, in one embodiment, is a digital video camera (such as a "web-cam"), or similar image capturing device. As mentioned above, the video capture device may be configured to provide depth image.
- Input image processor 302 translates the captured video images of the control object into signals that are delivered to an output image processor.
- input image processor 302 is programmed to isolate the control object from the background in the captured video image through the depth information and generate an output signal responsive to the position and/or movement of the control object.
- the output image processor 304 is programmed to effect translational and/or rotational movement of an object on the video display device 306 in response to signals received from the input image processor 302.
- processors which execute software instructions.
- a single processor executes both input image processing and output image processing.
- the processing operations are shown as being divided between an input image processor 302 and an output image processor 304. It should be noted that the invention is in no way to be interpreted as limited to any special processor configuration, such as more than one processor.
- the multiple processing blocks shown in Figure 11 are shown only for convenience of description.
- FIG 12 is a simplified block diagram of a computer processing system configured to implement the embodiments of the invention described herein.
- the processing system may represent a computer-based entertainment system embodiment that includes central processing unit (“CPU”) 424 coupled to main memory 420 and graphical processing unit (“GPU”) 426.
- CPU 424 is also coupled to Input/Output Processor ("IOP") Bus 428.
- GPU 426 includes an internal buffer for fast processing of pixel based graphical data.
- GPU 426 can include an output processing portion or functionality to convert the image data processed into standard television signals, for example NTSC or PAL, for transmission to display device 427 connected external to the entertainment system or elements thereof.
- data output signals can be provided to a display device other than a television monitor, such as a computer monitor, LCD (Liquid Crystal Display) device, or other type of display device.
- IOP bus 428 couples CPU 424 to various input/output devices and other busses or device. IOP bus 428 is connected to input/output processor memory 430, controller 432, memory card 434, Universal Serial Bus (USB) port 436, IEEE1394 (also known as a Firewire interface) port 438, and bus 450. Bus 450 couples several other system components to CPU 424, including operating system (“OS”) ROM 440, flash memory 442, sound processing unit (“SPU”) 444, optical disc controlling 4, and hard disk drive (“HDD”) 448.
- OS operating system
- SPU sound processing unit
- HDD hard disk drive
- the video capture device can be directly connected to IOP bus 428 for transmission therethrough to CPU 424; where, data from the video capture device can be used to change or update the values used to generate the graphics images in GPU 426.
- embodiments of the present invention can use a variety of image processing configurations and techniques, such as those described in U.S. Patent Application Serial No. 10/365,120 filed February 11, 2003, and entitled METHOD AND APPARATUS FOR REAL TIME MOTION CAPTURE.
- the computer processing system may run on a CELLTM processor.
- Programs or computer instructions embodying aspects of the present invention can be provided by several different methods.
- the user input method for interaction with graphical images can be provided in the form of a program stored in HDD 448, flash memory 442, OS ROM 240, or on memory card 432.
- the program can be downloaded to the processing unit through one or more input ports coupled to CPU 424.
- the program modules defining the input method can be provided with the game or application program that is executed by CPU 424 and displayed on display device 427 or they may be provided separately from the application program, such as for execution from local main memory 420.
- the program may be executed partially on a server connected to the internet and partially on the local computer (e.g., game console, desktop, laptop, or wireless hand held device). Still further, the execution can be entirely on a remote server or processing machine, which provides the execution results to the local display screen.
- the local display or system should have minimal processing capabilities to receive the data over the network (e.g., like the Internet) and render the graphical data on the screen. The user's input, by way of the capture device can be communicated back to the server and then the response represented on the screen.
- Figures 13A-13C illustrate embodiments where depth data is taken into consideration in order to better identify the object used to perform the directional pointing.
- the object can be something the person is holding or can also be the person's hand.
- the terms "depth camera” and "three-dimensional camera” refer to any camera that is capable of obtaining distance or depth information as well as two-dimensional pixel information.
- a depth camera can utilize controlled infrared lighting to obtain distance information.
- Another exemplary depth camera can be a stereo camera pair, which triangulates distance information using two standard cameras.
- the term “depth sensing device” refers to any type of device that is capable of obtaining distance information as well as two-dimensional pixel information.
- embodiments of the present invention provide real-time interactive gaming experiences for users. For example, users can interact with various computer-generated objects in real-time. Furthermore, video scenes can be altered in real-time to enhance the user's game experience. For example, computer generated costumes can be inserted over the user's clothing, and computer generated light sources can be utilized to project virtual shadows within a video scene. Hence, using the embodiments of the present invention and a depth camera, users can experience an interactive game environment within their own living room.
- Figure 13 A is a block diagram of an exemplary system 500 for providing a real-time three-dimensional interactive environment, in accordance with an embodiment of the present invention.
- the system 500 includes a depth camera 502, an input image processor 504, an output image processor 506, and a video display device 508.
- the depth camera 502 provides the ability to capture and map the third-dimension in addition to normal two-dimensional video imagery.
- Figures 13B and 13C illustrated the images generated by a typical depth camera 502.
- Figure 13B is an illustration showing two-dimensional data 520 captured using a typical depth camera. Similar to normal cameras, a depth camera captures two-dimensional data for a plurality of pixels that comprise the video image. These values are color values for the pixels, generally red, green, and blue (RGB) values for each pixel. In this manner, objects captured by the camera appear as two-dimension objects on a monitor.
- the exemplary scene includes a cylinder object 522 and a sphere object 524 disposed on a table 526, which may be situated among hills 528.
- FIG. 13C is an illustration showing depth data 550 captured using a typical depth camera. As illustrated in Figure 13B, the depth camera captures the x and y components of a scene using RGB values for each pixel in the scene. However, as shown in Figure 13 C, the depth camera also captures the z-components of the scene, which represent the depth values for the scene. Since the depth values correspond to the z-axis, the depth values are often referred to as z-values. [0066] In operation, a z-value is captured for each pixel of the scene. Each z- value represents a distance from the camera to a particular object in the scene corresponding to the related pixel.
- z-values are illustrated for the cylinder object 552, the sphere object 554, and part of the table 556.
- a maximum detection range is defined beyond which depth values will not be detected.
- the maximum depth range 558 appears as vertical plane wherein all pixels are given the same depth value.
- this maximum range plane can be utilized by the embodiments of the present invention to provide user defined object tracking.
- each object can be tracked in three dimensions.
- a computer system of the embodiments of the present invention can utilize the z-values, along with the two- dimensional pixel data, to create an enhanced three-dimensional interactive environment for the user.
- Embodiments of the present invention also contemplate distributed image processing configurations.
- the invention is not limited to the captured image and display image processing taking place in one or even two locations, such as in the CPU or in the CPU and one other element.
- the input image processing can just as readily take place in an associated CPU, processor or device that can perform processing; essentially all of image processing can be distributed throughout the interconnected system.
- the present invention is not limited to any specific image processing hardware circuitry and/or software.
- the above described invention may be practiced with other computer system configurations including hand-held devices, microprocessor systems, microprocessor-based or programmable consumer electronics, minicomputers, mainframe computers and the like.
- the invention may also be practiced in distributing computing environments where tasks are performed by remote processing devices that are linked through a communications network.
- the invention can also be embodied as computer readable code on a computer readable medium.
- the computer readable medium is any data storage device that can store data which can be thereafter read by a computer system, including an electromagnetic wave carrier. Examples of the computer readable medium include hard drives, network attached storage (NAS), read-only memory, random-access memory, CD-ROMs, CD-Rs, CD-RWs, magnetic tapes, and other optical and non-optical data storage devices.
- the computer readable medium can also be distributed over a network coupled computer system so that the computer readable code is stored and executed in a distributed fashion.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Human Computer Interaction (AREA)
- General Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Position Input By Displaying (AREA)
- User Interface Of Digital Computer (AREA)
- Image Processing (AREA)
Abstract
A method for detecting direction when interfacing with a computer program is provided. The method includes capturing an image presented in front of an image capture device. The image capture device has a capture location in a coordinate space. When a person is captured in the' image, the method includes identifying a human head in the image and assigning the human head a head location in the coordinate space. The method also includes identifying an object held by the person in the image and assigning the object an object location in coordinate space. The method further includes identifying a relative position in coordinate space between the head location and the object location when viewed from the capture location. The relative position defines a pointing direction of the object when viewed by the image capture device. The method may be practiced on a computer system, such as one used in the gaming field.
Description
METHODS AND SYSTEMS FOR ENABLING DIRECTION DETECTION WHEN INTERFACING WITH A COMPUTER PROGRAM by Inventors
Richard L. Marks Hrishikesh R. Deshpande
BACKGROUND
Description of the Related Art [0001] The video game industry has seen many changes over the years. As computing power has expanded, developers of video games have likewise created game software that takes advantage of these increases in computing power. To this end, video game developers have been coding games that incorporate sophisticated operations and mathematics to produce a very realistic game experience. [0002] Example gaming platforms, may he the Sony Playstation or Sony Playstation2 (PS2), each of which is sold in the form of a game console. As is well known, the game console is designed to connect to a monitor (usually a television) and enable user interaction through handheld controllers. The game console is designed with specialized processing hardware, including a CPU, a graphics synthesizer for processing intensive graphics operations, a vector unit for performing geometry transformations, and other glue hardware, firmware, and software. The game console is further designed with an optical disc tray for receiving game compact discs for local play through the game console. Online gaming is also possible, where a user can interactively play against or with other users over the Internet.
[0003] As game complexity continues to intrigue players, game and hardware manufacturers have continued to innovate to enable additional interactivity. In reality, however, the way in which users interact with a game has not changed dramatically over the years. Commonly, users still play computer games using hand held controllers or interact with programs using mouse pointing devices.
[0004] In view of the foregoing, there is a need for methods and systems that enable more advanced user interactivity with game play.
SUMMARY OF THE INVENTION
[0005] Broadly speaking, the present invention fills these needs by providing an apparatus and method that facilitates interactivity with a computer program. In one embodiment, the
computer program is a game program, but without limitation, the apparatus and method can find applicability in any consumer electronic device that will require a user to interact therewith. The present invention simplifies user interaction experience through machine recognizable gestures based on pointing to the interface and discriminating commands based on factors including trigger cues and position determination of a hand or object under user control.
[0006] In one embodiment, a method for detecting direction when interfacing with a computer program is described. The method includes capturing an image presented in front of an image capture device. The image capture device has a capture location in a coordinate space. When a person is captured in the image, the method includes identifying a human head in the image and assigning the human head a head location in the coordinate space. The method also includes identifying an object held by the person in the image and assigning the object an object location in coordinate space. The method further includes identifying a relative position in coordinate space between the head location and the object location when viewed from the capture location. The relative position defines a pointing direction of the object when viewed by the image capture device.
[0007] In another embodiment, a method for detecting pointing direction of an object directed toward a display screen that can render graphics of a computer program is provided. The method includes capturing an image presented in front of an image capture device. The image capture device has a capture location in a coordinate space that is proximate to the display screen. When a person is captured in the image, the method includes identifying a first body part of the person in the image and assigning the first body part a first location in the coordinate space. Then the method includes identifying a second body part of the person in the image and assigning the second body part a second location in coordinate space. Once the first and second body parts are identified, the method moves to identifying a relative position in coordinate space between the first location and the second location when viewed from the capture location. The relative position defines a pointing direction of the second body part when viewed by the image capture device at the capture. location that is proximate to the display screen. [0008] In one embodiment, an apparatus for capturing image and sound during interactivity with a computer program is provided. The apparatus includes an image capture unit that is configured to capture one or more image frames. These image frames are analyzed to identify a person's head and a person's hand. Based on the relative positioning of the hand and head,
the apparatus can ascertain a pointing direction, which may be translated as the pointing location on a display screen.
[0009] Other aspects and advantages of the invention will become apparent from the following detailed description, taken in conjunction with the accompanying drawings, illustrating by way of example the principles of the invention.
BRIEF DESCRIPTION OF THE DRAWINGS
[0010] The invention, together with further advantages thereof, may best be understood by reference to the following description taken in conjunction with the accompanying drawings. [0011] Figure IA and IB illustrate the environment of one embodiment of present invention, which takes advantage of the pointing direction determination described herein.
[0012] Figures 2-4 illustrate the analysis of the relative positions between a person's head and a person's hand (or object held by the person), when determining pointing direction.
[0013] Figures 5A-5C illustrate tracking and triggering embodiments, which can be identified by the computer program by the analysis of the captured image data, in accordance with one embodiment of the present invention.
[0014] Figures 6A-6F illustrate alternate embodiments of detecting an object and detecting changes in the object based on relative orientation of the object itself, in accordance with one embodiment of the present invention. [0015] JFigures 7 and 8A-8B illustrate examples of hand positions, when the hand positions and shapes are analyzed to determine a desired trigger or tracking response by the computer program, in accordance with one embodiment of the present invention.
[0016] Figures. 9 and 10 provide exemplary flow charts of the operations that can be executed in determining a pointing direction, in accordance with one embodiment of the present invention.
[0017] Figures 11 and 12 provide exemplary embodiments of hardware that may be used in processing the computer code necessary to executed the claimed operations, in accordance with one embodiment of the present invention.
[0018] Figures 13A-13C illustrate exemplary embodiments for when depth information is considered when identifying objects to track, when interfacing with a system that needs to ascertain pointing direction, in accordance with one embodiment of the present invention.
DETAILED DESCRIPTION
[0019] In the following description, numerous specific details are set forth in order to provide a thorough understanding of the present invention. It will be apparent, however, to one skilled in the art that the present invention may be practiced without some or all of these specific details. In other instances, well known process steps have not been described in detail in order not to obscure the present invention.
[0020] Figure IA illustrates an interactive game setup 100, in accordance with one embodiment of the present invention. The interactive game setup 100 includes a computer 102 that is coupled to a display screen 110. An image capture device 105 is placed on top of the display screen 110 and is coupled to the computer 102. Computer 102 is, in one embodiment, a gaming system console which allows users to play video games and interface with the video games through controllers 108. The image capture device 105 is shown placed on top of the display screen 110, but it should be understood that the image capture device 105 can be placed in any other proximate location that will allow it to capture images that are located about in front of the display screen 110. Techniques for capturing these movements and interactions can vary, but exemplary techniques are described in United Kingdom Applications GB 0304024.3 (PCT/GB2004/000693) and GB 0304022.7 (PCT/GB2004/000703), each filed on February 21, 2003. [0021] In a specific example, but not limited to any brand, the game console can be a one manufactured by Sony Computer Entertainment Inc., Nintendo, Microsoft, or any other manufacturer. The image capture device 105 can be as simple as a standard web cam or can include more advanced technology. In one embodiment, the image capture device should be capable of capturing images, digitizing the images, and communicating the image data back to the computer 102. In some embodiments, the image capture device will have logic integrated therein for performing the digitizing and another embodiment the image capture device 105 will simply transmit the captured data back to the computer 102 for digitizing. In either case, the image capture device 105 is capable of capturing either color or black and white images of any object located in front of the image capture device 105. [0022] Figure IB illustrates an exemplary embodiment of the present invention in which the computer 102 processes image data provided by the image capture device 105 to ascertain a pointing direction of an object placed in front of the image capture device 105. As shown, the computer 102 is connected to the image capture device 105. The image capture device 105 is designed to focus onto a capture region 105a. In this example, a person 112 is intending to
interact with a computer program being executed by the computer 102. The computer program, in this example, is a video game which is rendered and displayed by the display screen 110.
[0023] For example purposes only, the video game is a target shooting game in which the person 112 wishes to aim at a target and earn points commensurate with his or her performance. As illustrated on the display screen 110, an image 112' of the person 112 may also be placed on the display screen 110 during game play. Alternatively, the person's image 112' may be omitted from the display screen, depending on the particular device under control or game being played. In this example, the user experience may be enhanced by illustrating an image 112' of the person 112 during the target shooting exercise to present more reality during game play. A feature of the target shooting game is the ability for person 112 to point or direct an object 124 at particular interactive graphics on the display screen 110.
[0024] To achieve accurate pointing direction of the object 124, which in this case and for example purposes is a gun, the person 112 will hold the object 124 with his or her hand 122. The hand 122 will be directionally pointed toward the display screen 110. The image capture device 105 will at this point, analyze the digital image capture of the person 112 to determine the location of the person's 112 head 120, and the location of the person's 112 hand 122. As shown, the person's 112 hand is extended in front of his body and the image capture device will identify the object 124 when examining the captured digital image. The captured digital image will also be examined by code executed at the computer 102 to ascertain the location of the person's 112 head 120. In one embodiment, head tracking is completed with a combination of a template matching (for speed performance), coupled to a face detection code. The face detection code will essentially identify the location of the user's face by locating the user's eyes and other facial features. For additional information on head and face detection, reference may be made to co-pending US Patent Application No. 10/663,236, entitled "METHOD AND APPARATUS FOR ADJUSTING A VIEW OF A SCENE BEING DISPLAYED ACCORDING TO TRACKED HEAD MOTION", filed on September 15, 2003.
[0025] The object 124 will, in one embodiment, have an identifier which may be color or lights (e.g., light emitting diodes "LEDs") coupled to the object so that the program analyzing the captured digital image will easily identify the location of the object 124. Once the computer program has identified the location of the person's head 120 (H) and the location of the person's hand 122 (h), the computer program will perform computations to determine a
relative angle from the image capture device position, and between the detected object 124, and the head 120.
[0026] As illustrated in Figure IB, the relative position of the object 124 and the head 120 will be calculated relative to the image capture device 105. This will produce two angle parameters (theta θ and phi Φ). The azimuth angle θ will define the horizontal positioning between the head 120 and the object 124 along an X axis. The phi angle Φ will produce the altitude angle which is the relative angle between the height of the head 120 and the height of the hand 122. In one embodiment, an initial calibration operation may be performed before a gaming operation begins to calibrate the object's pointing location on the display screen 110. For instance, the user may be prompted to calibrate the pointing algorithm by having the user point the object 124 at a specific location on the display screen 110. Once the calibration has been completed, the computer 102 will be able to calculate the azimuth angle and the altitude angle (theta and phi) which define the relative positions of the person's head 120 and the person's hand 122, for each successive frame being captured by the image capture device 105. The relative positioning between the head and the hand may be calculated for each captured frame or may be captured every other frame, or after a number of frames are captured, depending on the accuracy required for the pointing operation. For example, if the game is a shooting gallery game, it would be important for the relative positioning of the head 120 and the hand 122 to be computed for each frame so that the person 112 will have accurate aiming and triggering capabilities when attempting to secure a good performing score in the video game contest.
[0027] Figure 2 illustrates a top view of the person 112 who is positioned in front of the display screen 110, in accordance with one embodiment of the present invention. In this example, an initial azimuth angle (thetai) is shown being determined as the relative angle between the position of the hand 122 (which is holding object 124), and the head 120. The person 112, during interactive play with the computer program, will be facing the display screen 1 10 and most likely, will maintain his body substantially parallel to the display screen 110. When the person 112 maintains his body substantially parallel to the display screen 110, movement of the hand 122 in direction 123 will cause the azimuth angle to be recomputed and produce a new azimuth angle theta2.
[0028] In this example, the person 112 is holding the object 124 out in front of his body at about arm's length. This distance is shown to be the approximate arm's length detect (ALD) that defines the location of the approximate place where the image capture device 105 will
attempt to capture the position of the hand 122 and associated object 124. The approximate arm's length detect (ALD) can vary, depending upon the user's arm length, but generally should be provided to allow a distance relationship between the location of the head 120 and the hand 122. For instance, there should be at least a minor projection of the hand 122 in front of the person's body to point to different locations of the display screen 110.
[0029] Figure 3 illustrates a front view of the person 112 where the head 120 is identified and the hand 122 is identified, from the captured digital image. In this example, a dashed line represents the display screen 110 and the image capture device 105 that is directionally pointed at the person 112. In this example, the image capture device 105 will be illustrated to be at a coordinate space of (0, 0, 0), representing the reference point of the image capture device 105, and its position in coordinate space. The approximate location of the head 120 will also have an associated coordinate space (x H, y H, z H). Likewise, the hand 122 and the associated object 124 that is held by the hand 122 will have a coordinate space (x h, y h, z h) that is relative to the image capture device 105. [0030] Figure 4 illustrates the person 112 pointing the object 124 at the display screen 110. In this example, the coordinate space of the head 120 is identified in the digital image captured by the capture device 105. The location in coordinate space of the hand 122 is also captured in the digital image captured by the image capture device 105 when the person 112 is pointing at the display screen 110. The image capture device 105 is the reference point, which is at the coordinate space (0, 0, 0). The altitude angle phi is therefore calculated between the position of the head 120 relative to the position of the hand 122. In one example, the angle is calculated as:
Altitude angle= arctan ((yh-yH)/(zh-zH)) [0031] In a similar manner, the azimuth angle theta of Figure 2 is calculated as: Azimuth angle=arctan ((xh-xH)/(zh-zH))
[0032] When the user moves his hand down (e.g., as captured in a later frame) as illustrated in Figure 4, a new angle phi2 will be produced to define the new relative position between the hand 122 and the head 120 of the person 112. Based on this new relative positioning of the head and the hand, the computer 102 will re-position the pointing direction on the display screen.
[0033] Figure 5A illustrates an embodiment where the object 124 is a gun. The gun may be a simplified gun object that is configured to be pointed in the direction of the display screen 110 to hit particular objects or interact with objects on the display screen 110. In this example, the
gun 124 will include a detection region 124a. Detection region 124a is the region which is directionally pointed toward the display screen 110. The detection region 124a is also the region that is captured by the image capture device 105 for analysis of the digital image by the computer 102. In one embodiment, the detection region 124a is configured to include a pair of lighting objects that will assist in the interactivity with the computer program being executed by the computer 102 and displayed on the display screen 110. In this example, a tracking indicator 130 is provided as a light or color object that is present on the detection region 124a. Based on the tracking indicator 130, the image capture device 105 will produce a digital image that will be analyzed by the computer 102 to identify the position in coordinate space of the object 124. In this example, by providing the tracking indicator 130, the computer program being executed on the computer 102 is able to quickly identify the location of the object 124 and in relation to the head 120 of the person interacting with the computer program.
[0034] The tracking indicator 130 may be provided by way of a number of implementations. One implementation might be a light indicator that can be tracked by the computer program that analyzes the captured digital images, or may be in the form of a color indicator that the computer can identify quickly from the captured digital images. The hand itself may be the tracking indicator 130. In still another embodiment, the tracing indicator 130 may be provided as a reflective tape that will have different coloring or intensity depending on the angle that it might be displayed when shown to the image capture device 105. In this example, the object 104 is tracked as the user moves his hand 122 to different regions pointed to on the display screen 110.
[0035] In one embodiment, while the user moves his hand 122 relative to the head 120, the tracking indicator 130 will allow the computer program to provide a visual indicator on the display screen 110. This visual indicator on the display screen 110 will allow the user to understand where the object is currently pointing to on the display screen 110.
[0036] In another embodiment, the detection region 124a will also include a trigger indicator 132. The trigger indicator 132 may be in the form of a light that is triggered ON and OFF when the user pulls the trigger of the object 124. For instance, the detection region 124a is shown in Figure 5B after the trigger has been pulled and the trigger indicator 132 is lit. When the trigger indicator 132 is lit as shown in Figure 5B3 the computer program executing on the computer 102 will provide an indictor on the display screen 110 so that the user can identify whether his or her pointing has accurately hit an object of the computer game. In Figure 5C, the trigger indicator 132 is shown to be in the OFF position to signify that the object 124 will
still remain actively tracked, but the shooting which can be continuous or intermittent, can be discontinued when the user removes his finger from the trigger of the object 124. The trigger indicator 132 may be an any frequency range, including audio, ultrasonic, visible light wave, infrared and radio. Passive trigger indicator 132 may be achieved through a mechanical sound generated upon actuating a trigger and receiving and decoding the audio input to the device and determining whether the trigger was actuated.
[0037] Figure 6 A illustrates another embodiment of the present invention where tracking and trigger indicators 130a and 132a are provided. In this example, the track/trigger indicators 130a and 132a are provided so that determinations can be made of the relative distances between the two indicators as shown by distance (di). In one example, the object, when pointed at the image capture device 105 may respond by having the computer program that is executed on the computer 102 to ascertain the distance di and perform an interactive action on the display screen. When the object 124b is tilted relative to the starting position of Figure 6A, a second distance (d2) is computed. This distance is the new distance between the track/trigger indicators 130a and 132a. As the user continues to tilt the object 124b as shown in Figure 6C, the distance continues to shrink as shown by distance (d3). Once the object 124b has been placed in the horizontal position relative to the vertical position of Figure 6 A, the distance between the track and trigger indicators 130a and 132a is brought to approximately zero. At this point, the program may read that the user intends for a trigger action to occur, or any other action that can be triggered when the detected distance from d\ to d4 has been detected.
[0038] In another embodiment, the response by the computer program may be gradually changed, depending on the angle at which the detection region 124b is tilted. For instance, the user may immediately begin to shoot the gun (or trigger the shooting of the gun) when the tilting begins and is executed between the tilt of Figure 6B to the tilt of Figure 6D. When the user tilts the gun back to the original position, the gun may discontinue the shooting activity. Consequently, the trigger activity caused by analyzing the patterns or colors of the tracking and trigger indicators of 130a and 132b can cause the computer program to react in different interactive ways. [0039] An example of this interactivity may be to trigger a reloading operation to occur for a gun that is being used in a video game, or a change of gun type being used on the video game program. Once these changes are processed, the video display screen 110 will produce a different graphical animation for the user, depending upon the control being provided and detected by the image capture device.
[0040] Commands and trigger states are not limited to an ON and OFF parameters, but can be incrementally changed depending on the position of the relative state and angles of the trigger and track indicators. For example, the state of the trigger may be determined in a linear or variable state as opposed to ON or OFF. Any technique can be used to determine the relative trigger position including a resistive types used to control acceleration in remote control race tracks. The device, or gun in this example, can communicate the state of its trigger by encoding and transmitting its value in any of a multitude of ways known in the art. A variety of commands and gestures may be formulated based on the state of the trigger and the position of the device, including those based on all machine recognizable gestures. [0041 J Figures 6E and 6F provide yet another embodiment where different colors may be used to track tilt or relative tilt between positions of the detection regions 124c. In this example, the track and trigger indicators 130b and 132b are square or rectangular in dimension and can be defined by colored tapes, bar codes, light indicators, LEDs, or the like. As a user flips or tilts the detection region 124c from the position of Figure 6E to Figure 6F, for example, the reaction by the computer game as displayed on the computer display screen will change.
[0042J Figure 7 illustrates another embodiment of the present invention in which a hand 122 is used to control the interactivity on the display screen. The user may place the hand 122 in front of the image capture device 105 and it is tracked relative to the head of the person 112. In this example, the user's hand is extended having of an index finger 140 pointing at the display screen 110. The user's thumb 142 may be pointing upright to indicate to the computer program executing on the computer 102 that the trigger device has not been activated. When the user's thumb 142 is moved down toward the index finger 140 in the direction 122a, the computer program executing on a computer 102 may detect from the captured digital image that the user intends to shoot or trigger or interactively point to a specific region on the display screen 1 10. Thus, the user's hand being placed in a different position can trigger an event or cause the interactivity of a command with a computer program being executed and shown on the display screen 110. For example, the user may be able to shoot by different hand gestures, may be able to reload the gun with different gestures, and the different positions or orientations of the user's hand may cause different graphical renderings of the user or gun on the display screen when the user is interacting with a particular game program. In this embodiment, the state of the trigger may be determined ON or OFF as described above, or, the trigger may be determined in a variable state. In the latter, the relative position of the user's thumb may range from, for example, the position in which the thumb is substantially perpendicular to the
pointing finger and the position where the thumb is substantially parallel to the pointing finger and where the system performs image analysis of the hand to yield a relative state of the thumb. This state may be mapped to various control schemes, including those relating to a scroll-wheel on mouse. [0043] Figures 8A and 8B illustrate another example where the user's hand 122 may be used to identify or trigger different activity on the display screen. In this example, Figure 8A shows the user pointing directly at the display screen 110 and the computer program being executed to identify this hand orientation on the digital image that was captured by the capture device. When a subsequent capture period occurs at a different point in time, the image of Figure 8B might be captured and analyzed by the computer program. In this example, the user's hand is shown tilted from where the thumb 142 is pointing upright to where the thumb 142 is pointing sideways. In this orientation, the user may be able to trigger the shooting of a gun on the display screen, or the selection of an icon on a graphical display. [0044] Consequently, the detection of the user's hand and the orientation of the user's hand can be used to provide the interactivity necessary when analyzing the position of the user's hand relative to the user's head, and the pointing activity on a display screen. The pointing activity will allow the user to select icons, control the device under operation, shoot at graphical objects, select or scroll graphical objects, de-select graphical objects, turn ON and OFF graphical objects, disperse graphical objects, or simply interface with the graphics icons and features of a computer program being displayed on a display screen 110. However, in certain configurations, it may be desirable, that the system employing the present invention may operate with minimal or no icons on a display screen. Instead, the system may simply just recognize the gestures of the user and provide a control input to the device under operation. For example, a television or related peripheral configured or integrated with the present invention may be controlled by the present invention. Changing a channel, for example, may not necessarily involve interacting with an icon as opposed to recognizing a gesture command according to the scheme presented in the present invention.
[0045] Figure 9 illustrates a flowchart diagram 200 where a process for identifying a pointing direction is described, in accordance with one embodiment of the present invention. In operation 202, the method begins by providing a capture device for capturing image frames. The frame capture rate will depend on the particular implementation, and will not limit the invention described herein. The capture device will, in one embodiment, include a capture location in coordinate space. As shown in Figure 4, the coordinate space of the image capture
device is (0, 0, 0). In operation 204, a display screen is provided for rendering and displaying the interactive graphics of a computer program.
[00461 The interactive graphics of the computer program may be associate with a computer game, but may also be associated with any other program or interface that may require interactivity by a user. For instance, the interactivity may include the selection or de-selection of objects, opening files, changing channels, recording TV shows, closing files, accessing files over a network, or interactively communicating with users by way of the internet, electronic mail or by electronic video mail, selecting a consumer electronic device, turning a device ON or OFF. Next, the operation 206 will include the capturing of an image that is presented substantially in front of the image capture device. In one embodiment, the image will include a person that is part of the captured space, and who is in front of the display screen and the capture device.
[0047] Once the image has been captured in operation 206, operation 208 will include the identification of a human head of the person that is found in the captured image. The identified human head will therefore be analyzed to ascertain the head location in the coordinate space relative to the capture location. The method then moves to operation 210 where an object held by the person in the image is identified. The object's location is identified such that the coordinate space of the object is identified relative to the coordinate space of the capture location. Having the identified head location and the identified object location in memory, at operation 212, the computer program can identify a relative position in coordinate space between the head location and the object location when viewed from the capture location reference point (e.g., coordinate (0,0,0)). As mentioned above, an azimuth angle and an altitude angle can be computed for the relative locations of the head and the hand relative to the image capture device. This relative position in coordinate space is calculated for the captured frame.
[0048] In operation 214, a pointing direction is established for the object using the relative position identified between the object location and the head location. The pointing direction is displayed on the display screen to enable interaction with the interactive graphics provided by the computer program and displayed on the display screen. [0049] Figure 10 illustrates a more detailed process diagram 250 that can be implemented when determining the pointing direction of an object that is directionally pointed at a display screen during an interactivity with a computer program. The method begins at operation 252 where the capture device for capturing image frames is provided. The capture device will have
a capture location in coordinate space. The coordinate space of the capture location will be the reference point for performing operations to determine relative locations in the process of identifying directionality pointing.
[0050] The method moves to operation 254 where a display .screen is provided for rendering interactive graphics of the computer program. The interactive graphics may be a computer game or may be any other program as defined above. In operation 256, an image is captured in front of the image capture device and a person is captured in the image. The captured image may be that of a digital frame of video. In one embodiment, the digital frame of video may be a JPEG frame or may be part of a compressed video frame (e.g., MPEG or the like). [0051] Next, the operation moves to the identification of a human head of the person in the captured image in operation 258. The human head is analyzed on the captured image to determine a head location and the coordinate space relative to the image capture device. In operation 260, the method moves to the identification of an object held by the person in the image and determining an object location in the coordinate space. In operation 262, a relative position is identified in the coordinate space between the head location and the object location when viewed from the capture location of the capture device. The relative position will include a calculation of an azimuth angle and an altitude angle relative to the image capture device.
[0052] In operation 264, during execution of the computer program, the operations identified as A, B, C and D corresponding to operations 256, 258, 260, and 262 will be performed iteratively and continuously, depending on a rate desired for the performance of a computer program. For instance, the execution of operations A through D will occur at the rate of one time for each frame that is captured or only after a certain number of frames are captured. The rate at which operations A through D are performed will therefore depend on the specific environment and the need for accurate detection of the pointer location and the select ability of interactive objects on the display screen. If the display screen is processing a video game that has objects that are moving at rapid rates, the tracking operation may require that operations A through D be performed for each frame that is displayed on the video display screen.
[0053] In operation 266, the method indicates a continual update of the point and direction of the object using the relative position. The pointing direction is displayed on the display screen to enable interaction with the interactive graphics of the computer program. It should again be understood that the pointing direction can be to enable a user to select icons, de-select icons, move icons, open objects, open files, save files, move files, and interact with files that may be part of a file database, or part of a graphical user interface on a computer desktop or the like.
[0054] Figure 1 1 is a block diagram of an exemplary user input system for interaction with an object on a graphical display that can be used to implement embodiments of the present invention. As shown in Figure 11, the user input system is comprised of a video capture device 300, an input image processor 302, an output image processor 304, and a video display device 306. Video capture device 300 may be any device capable of capturing sequences of video images, and, in one embodiment, is a digital video camera (such as a "web-cam"), or similar image capturing device. As mentioned above, the video capture device may be configured to provide depth image. Input image processor 302 translates the captured video images of the control object into signals that are delivered to an output image processor. In one embodiment, input image processor 302 is programmed to isolate the control object from the background in the captured video image through the depth information and generate an output signal responsive to the position and/or movement of the control object. The output image processor 304 is programmed to effect translational and/or rotational movement of an object on the video display device 306 in response to signals received from the input image processor 302. [0055] These and additional aspects of the present invention may be implemented by one or more processors which execute software instructions. According to one embodiment of the present invention, a single processor executes both input image processing and output image processing. However, as shown in the figures and for ease of description, the processing operations are shown as being divided between an input image processor 302 and an output image processor 304. It should be noted that the invention is in no way to be interpreted as limited to any special processor configuration, such as more than one processor. The multiple processing blocks shown in Figure 11 are shown only for convenience of description.
[0056] Figure 12 is a simplified block diagram of a computer processing system configured to implement the embodiments of the invention described herein. The processing system may represent a computer-based entertainment system embodiment that includes central processing unit ("CPU") 424 coupled to main memory 420 and graphical processing unit ("GPU") 426. CPU 424 is also coupled to Input/Output Processor ("IOP") Bus 428. In one embodiment, GPU 426 includes an internal buffer for fast processing of pixel based graphical data. Additionally, GPU 426 can include an output processing portion or functionality to convert the image data processed into standard television signals, for example NTSC or PAL, for transmission to display device 427 connected external to the entertainment system or elements thereof. Alternatively, data output signals can be provided to a display device other than a
television monitor, such as a computer monitor, LCD (Liquid Crystal Display) device, or other type of display device.
[0057] IOP bus 428 couples CPU 424 to various input/output devices and other busses or device. IOP bus 428 is connected to input/output processor memory 430, controller 432, memory card 434, Universal Serial Bus (USB) port 436, IEEE1394 (also known as a Firewire interface) port 438, and bus 450. Bus 450 couples several other system components to CPU 424, including operating system ("OS") ROM 440, flash memory 442, sound processing unit ("SPU") 444, optical disc controlling 4, and hard disk drive ("HDD") 448. In one aspect of this embodiment, the video capture device can be directly connected to IOP bus 428 for transmission therethrough to CPU 424; where, data from the video capture device can be used to change or update the values used to generate the graphics images in GPU 426. Moreover, embodiments of the present invention can use a variety of image processing configurations and techniques, such as those described in U.S. Patent Application Serial No. 10/365,120 filed February 11, 2003, and entitled METHOD AND APPARATUS FOR REAL TIME MOTION CAPTURE. The computer processing system may run on a CELL™ processor.
[0058] Programs or computer instructions embodying aspects of the present invention can be provided by several different methods. For example, the user input method for interaction with graphical images can be provided in the form of a program stored in HDD 448, flash memory 442, OS ROM 240, or on memory card 432. Alternatively, the program can be downloaded to the processing unit through one or more input ports coupled to CPU 424. The program modules defining the input method can be provided with the game or application program that is executed by CPU 424 and displayed on display device 427 or they may be provided separately from the application program, such as for execution from local main memory 420. [0059] In still another embodiment, the program may be executed partially on a server connected to the internet and partially on the local computer (e.g., game console, desktop, laptop, or wireless hand held device). Still further, the execution can be entirely on a remote server or processing machine, which provides the execution results to the local display screen. In this case, the local display or system should have minimal processing capabilities to receive the data over the network (e.g., like the Internet) and render the graphical data on the screen. The user's input, by way of the capture device can be communicated back to the server and then the response represented on the screen.
[0060] Figures 13A-13C illustrate embodiments where depth data is taken into consideration in order to better identify the object used to perform the directional pointing. The object can be
something the person is holding or can also be the person's hand. In the this description, the terms "depth camera" and "three-dimensional camera" refer to any camera that is capable of obtaining distance or depth information as well as two-dimensional pixel information. For example, a depth camera can utilize controlled infrared lighting to obtain distance information. Another exemplary depth camera can be a stereo camera pair, which triangulates distance information using two standard cameras. Similarly, the term "depth sensing device" refers to any type of device that is capable of obtaining distance information as well as two-dimensional pixel information. [0061] Recent advances in three-dimensional imagery have opened the door for increased possibilities in real-time interactive computer animation. In particular, new "depth cameras" provide the ability to capture and map the third-dimension in addition to normal two- dimensional video imagery. With the new depth data, embodiments of the present invention allow the placement of computer-generated objects in various positions within a video scene in real-time, including behind other objects. [0062] Moreover, embodiments of the present invention provide real-time interactive gaming experiences for users. For example, users can interact with various computer-generated objects in real-time. Furthermore, video scenes can be altered in real-time to enhance the user's game experience. For example, computer generated costumes can be inserted over the user's clothing, and computer generated light sources can be utilized to project virtual shadows within a video scene. Hence, using the embodiments of the present invention and a depth camera, users can experience an interactive game environment within their own living room.
[0063] Figure 13 A is a block diagram of an exemplary system 500 for providing a real-time three-dimensional interactive environment, in accordance with an embodiment of the present invention. As shown in Figure 13 A, the system 500 includes a depth camera 502, an input image processor 504, an output image processor 506, and a video display device 508.
[0064] As mentioned above, the depth camera 502 provides the ability to capture and map the third-dimension in addition to normal two-dimensional video imagery. Figures 13B and 13C illustrated the images generated by a typical depth camera 502. In particular, Figure 13B is an illustration showing two-dimensional data 520 captured using a typical depth camera. Similar to normal cameras, a depth camera captures two-dimensional data for a plurality of pixels that comprise the video image. These values are color values for the pixels, generally red, green, and blue (RGB) values for each pixel. In this manner, objects captured by the camera appear as two-dimension objects on a monitor. For example, in Figure 13B, the exemplary scene
includes a cylinder object 522 and a sphere object 524 disposed on a table 526, which may be situated among hills 528.
[0065] However, unlike a conventional camera, a depth camera also captures depth values for the scene. Figure 13C is an illustration showing depth data 550 captured using a typical depth camera. As illustrated in Figure 13B, the depth camera captures the x and y components of a scene using RGB values for each pixel in the scene. However, as shown in Figure 13 C, the depth camera also captures the z-components of the scene, which represent the depth values for the scene. Since the depth values correspond to the z-axis, the depth values are often referred to as z-values. [0066] In operation, a z-value is captured for each pixel of the scene. Each z- value represents a distance from the camera to a particular object in the scene corresponding to the related pixel. For example, in Figure 13C, z-values are illustrated for the cylinder object 552, the sphere object 554, and part of the table 556. In addition, a maximum detection range is defined beyond which depth values will not be detected. For example, in Figure 13C the maximum depth range 558 appears as vertical plane wherein all pixels are given the same depth value. As will be described in greater detail below, this maximum range plane can be utilized by the embodiments of the present invention to provide user defined object tracking. Thus, using a depth camera, each object can be tracked in three dimensions. As a result, a computer system of the embodiments of the present invention can utilize the z-values, along with the two- dimensional pixel data, to create an enhanced three-dimensional interactive environment for the user. For more information on depth analysis, reference may be made to US Patent Application No. 10/448,614, entitled System and Method for Providing a Real-time three dimensional interactive environment, having a filing date of May 29, 2003. [0067] Embodiments of the present invention also contemplate distributed image processing configurations. For example, the invention is not limited to the captured image and display image processing taking place in one or even two locations, such as in the CPU or in the CPU and one other element. For example, the input image processing can just as readily take place in an associated CPU, processor or device that can perform processing; essentially all of image processing can be distributed throughout the interconnected system. Thus, the present invention is not limited to any specific image processing hardware circuitry and/or software. The embodiments described herein are also not limited to any specific combination of general hardware circuitry and/or software, nor to any particular source for the instructions executed by processing components.
[0068] With the above embodiments in mind, it should be understood that the invention may employ various computer-implemented operations involving data stored in computer systems. These operations include operations requiring physical manipulation of physical quantities. Usually, though not necessarily, these quantities take the form of electrical or magnetic signals capable of being stored, transferred, combined, compared, and otherwise manipulated. Further, the manipulations performed are often referred to in terms, such as producing, identifying, determining, or comparing.
[0069] The above described invention may be practiced with other computer system configurations including hand-held devices, microprocessor systems, microprocessor-based or programmable consumer electronics, minicomputers, mainframe computers and the like. The invention may also be practiced in distributing computing environments where tasks are performed by remote processing devices that are linked through a communications network.
[0070] The invention can also be embodied as computer readable code on a computer readable medium. The computer readable medium is any data storage device that can store data which can be thereafter read by a computer system, including an electromagnetic wave carrier. Examples of the computer readable medium include hard drives, network attached storage (NAS), read-only memory, random-access memory, CD-ROMs, CD-Rs, CD-RWs, magnetic tapes, and other optical and non-optical data storage devices. The computer readable medium can also be distributed over a network coupled computer system so that the computer readable code is stored and executed in a distributed fashion.
[0071] Although the foregoing invention has been described in some detail for purposes of clarity of understanding, it will be apparent that certain changes and modifications may be practiced within the scope of the appended claims. Accordingly, the present embodiments are to be considered as illustrative and not restrictive, and the invention is not to be limited to the details given herein, but may be modified within the scope and equivalents of the appended claims.
What is claimed is:
Claims
1. A method for detecting direction when interfacing with a computer program, comprising:
(a) capturing an image presented in front of an image capture device, the image capture device having a capture location in a coordinate space; when a person is captured in the image,
(b) identifying a human head in the image and assigning the human head a head location in the coordinate space;
(c) identifying an object held by the person in the image and assigning the object an object location in coordinate space; and
(d) identifying a relative position in coordinate space between the head location and the object location when viewed from the capture location, wherein the relative position defines a pointing direction of the object when viewed by the image capture device.
2. A method for detecting direction when interfacing with a computer program as recited in claim 1, wherein the relative position is identified by computing an azimuth angle and an altitude angle between the head location and the object location in relation to the capture location.
3. A method for detecting direction when interfacing with a computer program as recited in claim 1, wherein the capture location is at a proximate location of a display screen and the display screen is capable of rendering interactive graphics.
4. A method for detecting direction when interfacing with a computer program as recited in claim 3, wherein the pointing direction is toward the display screen.
5. A method for detecting direction when interfacing with a computer program as recited in claim 3, further comprising: repeating (a) - (d) continually to update the pointing direction; and displaying the continually updated pointing direction on the display screen.
6. A method for detecting direction when interfacing with a computer program as recited in claim 5, further comprising: ' enabling selection of particular interactive graphics using the displayed pointing direction.
7. A method for detecting direction when interfacing with a computer program as recited in claim 6, wherein the selection is in response to a detected trigger event.
8. A method for detecting direction when interfacing with a computer program as recited in claim 7, wherein the detected trigger event is identified in the image, the identification comprising, identifying a first characteristic of the object held by the person at a first point in time; and identifying a second characteristic of the object held by the person at a second point in time, wherein the trigger event is activated when a degree of difference is determined to have existed between first characteristic and the second characteristic when examined between the first point in time and the second point in time.
9. A method for detecting direction when interfacing with a computer program as recited in claim 8, wherein the trigger even being activated is indicative of interactivity with the interactive graphics.
10. A method for detecting direction when interfacing with a computer program as recited in claim 9, wherein the interactivity can include one or more of selection of a graphic, shooting of a graphic, touching a graphic, moving of a graphic, activation of a graphic, triggering of a graphic and acting upon or with a graphic.
1 1. A method for detecting direction when interfacing with a computer program as recited in claim 1, wherein identifying the human head is processed using template matching in combination with face detection code.
12. A method for detecting direction when interfacing with a computer program as recited in claim 1, wherein identifying the object held by the person is facilitated by color tracking of a portion of the object.
13. A method for detecting direction when interfacing with a computer program as recited in claim 12, wherein color tracking includes one or a combination of identifying differences in colors and identifying on/off states of colors.
14. A method for detecting direction when interfacing with a computer program as recited in claim 5, wherein identifying the object held by the person is facilitated by identification of changes in positions of the object when repeating (a) - (d).
15. A method for detecting direction when interfacing with a computer program as recited in claim, wherein the computer program is a video game.
16. A method for detecting pointing direction of an object directed toward a display screen that can render graphics of a computer program, comprising:
(a) capturing an image presented in front of an image capture device, the image capture device having a capture location in a coordinate space that is proximate to the display screen; when a person is captured in the image,
(b) identifying a first body part of the person in the image and assigning the first body part a first location in the coordinate space; (c) identifying a second body part of the person in the image and assigning the second body part a second location in coordinate space; and
(d) identifying a relative position in coordinate space between the first location and the second location when viewed from the capture location, wherein the relative position defines a pointing direction of the second body part when viewed by the image capture device at the capture location that is proximate to the display screen.
17. A method for detecting pointing direction of an object directed toward a display screen that can render graphics of a computer program as recited in claim 16, wherein the first body part is a human head and the second body part is a human hand.
18. A method for detecting pointing direction of an object directed toward a display screen that can render graphics of a computer program as recited in claim 16, wherein (a) - (d) is repeated continually during execution of the computer program, and examining a shape of the human hand during the repeating of (a) - (d) to determine particular shape changes.
19. A method for detecting pointing direction of an object directed toward a display screen that can render graphics of a computer program as recited in claim 16, wherein particular shape changes trigger interactivity with interactive graphics of the computer program.
20. A method for detecting pointing direction of an object directed toward a display screen that can render graphics of a computer program as recited in claim 19, wherein the interactivity can include one or more of selection of a graphic, shooting of a graphic, touching a graphic, moving of a graphic, activation of a graphic, triggering of a graphic and acting upon or with a graphic.
21. A method for detecting pointing direction of an object directed toward a display screen that can render graphics of a computer program as recited in claim 16, wherein the second body part is identified by way of an object held by the human hand.
22. A method for detecting pointing direction of an object directed toward a display screen that can render graphics of a computer program as recited in claim 16, wherein the object includes color.
23. A method for detecting pointing direction of an object directed toward a display screen that can render graphics of a computer program as recited in claim 22, wherein the color is capable of switching between states to trigger interactivity with interactive graphics of the computer program.
24. A method for detecting pointing direction of an object directed toward a display screen that can render graphics of a computer program as recited in claim 23, wherein additional colors are present on the object, the colors capable of being switched to trigger interactivity with interactive graphics of the computer program.
25. A method for detecting pointing direction of an object directed toward a display screen that can render graphics of a computer program as recited in claim 22, wherein the color can switch from on/off states to trigger interactivity with interactive graphics of the computer program
26. A method for detecting pointing direction of an object directed toward a display screen that can render graphics of a computer program as recited in claim 16, wherein the computer program is a video game.
27. A method for detecting pointing direction of an object directed toward a display screen that can render graphics of a computer program as recited in claim 16, wherein the relative position is identified by computing an azimuth angle and an altitude angle between the head location and the object location in relation to the capture location.
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP06839276A EP1960069A1 (en) | 2005-12-12 | 2006-12-08 | Methods and systems for enabling direction detection when interfacing with a computer program |
JP2008545704A JP5084052B2 (en) | 2005-12-12 | 2006-12-08 | Method and system for enabling direction detection when interfacing with a computer program |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/301,673 US7646372B2 (en) | 2003-09-15 | 2005-12-12 | Methods and systems for enabling direction detection when interfacing with a computer program |
US11/301,673 | 2005-12-12 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2007078639A1 true WO2007078639A1 (en) | 2007-07-12 |
Family
ID=37834200
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/US2006/047111 WO2007078639A1 (en) | 2005-12-12 | 2006-12-08 | Methods and systems for enabling direction detection when interfacing with a computer program |
Country Status (4)
Country | Link |
---|---|
US (2) | US7646372B2 (en) |
EP (1) | EP1960069A1 (en) |
JP (1) | JP5084052B2 (en) |
WO (1) | WO2007078639A1 (en) |
Cited By (45)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GB2451461A (en) * | 2007-07-28 | 2009-02-04 | Naveen Chawla | Camera based 3D user and wand tracking human-computer interaction system |
JP2009075685A (en) * | 2007-09-19 | 2009-04-09 | Sony Corp | Image processor, image processing method, and program |
WO2009078993A1 (en) | 2007-12-17 | 2009-06-25 | Sony Computer Entertainment America Inc. | Dynamic three-dimensional object mapping for user-defined control device |
JP2009165826A (en) * | 2008-01-11 | 2009-07-30 | Sony Computer Entertainment America Inc | Gesture cataloguing and recognition |
US8166421B2 (en) | 2008-01-14 | 2012-04-24 | Primesense Ltd. | Three-dimensional user interface |
US8249334B2 (en) | 2006-05-11 | 2012-08-21 | Primesense Ltd. | Modeling of humanoid forms from depth maps |
EP2583155A2 (en) * | 2010-06-21 | 2013-04-24 | Microsoft Corporation | Natural user input for driving interactive stories |
US8439733B2 (en) | 2007-06-14 | 2013-05-14 | Harmonix Music Systems, Inc. | Systems and methods for reinstating a player within a rhythm-action game |
US8444464B2 (en) | 2010-06-11 | 2013-05-21 | Harmonix Music Systems, Inc. | Prompting a player of a dance game |
US8449360B2 (en) | 2009-05-29 | 2013-05-28 | Harmonix Music Systems, Inc. | Displaying song lyrics and vocal cues |
US8465366B2 (en) | 2009-05-29 | 2013-06-18 | Harmonix Music Systems, Inc. | Biasing a musical performance input to a part |
WO2013096844A3 (en) * | 2011-12-21 | 2013-08-22 | Sony Computer Entertainment Inc. | Directional input for a video game |
US8550908B2 (en) | 2010-03-16 | 2013-10-08 | Harmonix Music Systems, Inc. | Simulating musical instruments |
US8565479B2 (en) | 2009-08-13 | 2013-10-22 | Primesense Ltd. | Extraction of skeletons from 3D maps |
US8582867B2 (en) | 2010-09-16 | 2013-11-12 | Primesense Ltd | Learning-based pose estimation from depth maps |
US8594425B2 (en) | 2010-05-31 | 2013-11-26 | Primesense Ltd. | Analysis of three-dimensional scenes |
US8663013B2 (en) | 2008-07-08 | 2014-03-04 | Harmonix Music Systems, Inc. | Systems and methods for simulating a rock band experience |
US8678896B2 (en) | 2007-06-14 | 2014-03-25 | Harmonix Music Systems, Inc. | Systems and methods for asynchronous band interaction in a rhythm action game |
US8702485B2 (en) | 2010-06-11 | 2014-04-22 | Harmonix Music Systems, Inc. | Dance game and tutorial |
US8787663B2 (en) | 2010-03-01 | 2014-07-22 | Primesense Ltd. | Tracking body parts by combined color image and depth processing |
US8872762B2 (en) | 2010-12-08 | 2014-10-28 | Primesense Ltd. | Three dimensional user interface cursor control |
US8881051B2 (en) | 2011-07-05 | 2014-11-04 | Primesense Ltd | Zoom-based gesture user interface |
US8933876B2 (en) | 2010-12-13 | 2015-01-13 | Apple Inc. | Three dimensional user interface session control |
US8959013B2 (en) | 2010-09-27 | 2015-02-17 | Apple Inc. | Virtual keyboard for a non-tactile three dimensional user interface |
US9002099B2 (en) | 2011-09-11 | 2015-04-07 | Apple Inc. | Learning-based estimation of hand and finger pose |
US9019267B2 (en) | 2012-10-30 | 2015-04-28 | Apple Inc. | Depth mapping with enhanced resolution |
US9024166B2 (en) | 2010-09-09 | 2015-05-05 | Harmonix Music Systems, Inc. | Preventing subtractive track separation |
US9030498B2 (en) | 2011-08-15 | 2015-05-12 | Apple Inc. | Combining explicit select gestures and timeclick in a non-tactile three dimensional user interface |
US9035876B2 (en) | 2008-01-14 | 2015-05-19 | Apple Inc. | Three-dimensional user interface session control |
US9047507B2 (en) | 2012-05-02 | 2015-06-02 | Apple Inc. | Upper-body skeleton extraction from depth maps |
US9122311B2 (en) | 2011-08-24 | 2015-09-01 | Apple Inc. | Visual feedback for tactile and non-tactile user interfaces |
US9158375B2 (en) | 2010-07-20 | 2015-10-13 | Apple Inc. | Interactive reality augmentation for natural interaction |
US9201501B2 (en) | 2010-07-20 | 2015-12-01 | Apple Inc. | Adaptive projector |
US9218063B2 (en) | 2011-08-24 | 2015-12-22 | Apple Inc. | Sessionless pointing user interface |
US9229534B2 (en) | 2012-02-28 | 2016-01-05 | Apple Inc. | Asymmetric mapping for tactile and non-tactile user interfaces |
US9285874B2 (en) | 2011-02-09 | 2016-03-15 | Apple Inc. | Gaze detection in a 3D mapping environment |
US9358456B1 (en) | 2010-06-11 | 2016-06-07 | Harmonix Music Systems, Inc. | Dance competition game |
US9377865B2 (en) | 2011-07-05 | 2016-06-28 | Apple Inc. | Zoom-based gesture user interface |
US9377863B2 (en) | 2012-03-26 | 2016-06-28 | Apple Inc. | Gaze-enhanced virtual touchscreen |
US9459758B2 (en) | 2011-07-05 | 2016-10-04 | Apple Inc. | Gesture-based interface with enhanced features |
US9981193B2 (en) | 2009-10-27 | 2018-05-29 | Harmonix Music Systems, Inc. | Movement based recognition and evaluation |
US10043279B1 (en) | 2015-12-07 | 2018-08-07 | Apple Inc. | Robust detection and classification of body parts in a depth map |
US10357714B2 (en) | 2009-10-27 | 2019-07-23 | Harmonix Music Systems, Inc. | Gesture-based user interface for navigating a menu |
US10362028B2 (en) | 2013-11-07 | 2019-07-23 | Sony Interactive Entertainment Inc. | Information processing apparatus |
US10366278B2 (en) | 2016-09-20 | 2019-07-30 | Apple Inc. | Curvature-based face detector |
Families Citing this family (218)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7161579B2 (en) * | 2002-07-18 | 2007-01-09 | Sony Computer Entertainment Inc. | Hand-held computer interactive device |
US8947347B2 (en) | 2003-08-27 | 2015-02-03 | Sony Computer Entertainment Inc. | Controlling actions in a video game unit |
US7809145B2 (en) * | 2006-05-04 | 2010-10-05 | Sony Computer Entertainment Inc. | Ultra small microphone array |
US8797260B2 (en) | 2002-07-27 | 2014-08-05 | Sony Computer Entertainment Inc. | Inertially trackable hand-held controller |
US7783061B2 (en) | 2003-08-27 | 2010-08-24 | Sony Computer Entertainment Inc. | Methods and apparatus for the targeted sound detection |
US8073157B2 (en) * | 2003-08-27 | 2011-12-06 | Sony Computer Entertainment Inc. | Methods and apparatus for targeted sound detection and characterization |
US7760248B2 (en) | 2002-07-27 | 2010-07-20 | Sony Computer Entertainment Inc. | Selective sound source listening in conjunction with computer interactive processing |
US8686939B2 (en) | 2002-07-27 | 2014-04-01 | Sony Computer Entertainment Inc. | System, method, and apparatus for three-dimensional input control |
US8570378B2 (en) | 2002-07-27 | 2013-10-29 | Sony Computer Entertainment Inc. | Method and apparatus for tracking three-dimensional movements of an object using a depth sensing camera |
US8139793B2 (en) | 2003-08-27 | 2012-03-20 | Sony Computer Entertainment Inc. | Methods and apparatus for capturing audio signals based on a visual image |
US8313380B2 (en) | 2002-07-27 | 2012-11-20 | Sony Computer Entertainment America Llc | Scheme for translating movements of a hand-held controller into inputs for a system |
US9174119B2 (en) | 2002-07-27 | 2015-11-03 | Sony Computer Entertainement America, LLC | Controller for providing inputs to control execution of a program when inputs are combined |
US8233642B2 (en) | 2003-08-27 | 2012-07-31 | Sony Computer Entertainment Inc. | Methods and apparatuses for capturing an audio signal based on a location of the signal |
US9393487B2 (en) | 2002-07-27 | 2016-07-19 | Sony Interactive Entertainment Inc. | Method for mapping movements of a hand-held controller to game commands |
US8160269B2 (en) * | 2003-08-27 | 2012-04-17 | Sony Computer Entertainment Inc. | Methods and apparatuses for adjusting a listening area for capturing sounds |
US7803050B2 (en) | 2002-07-27 | 2010-09-28 | Sony Computer Entertainment Inc. | Tracking device with sound emitter for use in obtaining information for controlling game program execution |
US9474968B2 (en) | 2002-07-27 | 2016-10-25 | Sony Interactive Entertainment America Llc | Method and system for applying gearing effects to visual tracking |
US7854655B2 (en) | 2002-07-27 | 2010-12-21 | Sony Computer Entertainment America Inc. | Obtaining input for controlling execution of a game program |
US9682319B2 (en) * | 2002-07-31 | 2017-06-20 | Sony Interactive Entertainment Inc. | Combiner method for altering game gearing |
US9177387B2 (en) * | 2003-02-11 | 2015-11-03 | Sony Computer Entertainment Inc. | Method and apparatus for real time motion capture |
US20040268393A1 (en) * | 2003-05-08 | 2004-12-30 | Hunleth Frank A. | Control framework with a zoomable graphical user interface for organizing, selecting and launching media items |
US8072470B2 (en) * | 2003-05-29 | 2011-12-06 | Sony Computer Entertainment Inc. | System and method for providing a real-time three-dimensional interactive environment |
US7874917B2 (en) | 2003-09-15 | 2011-01-25 | Sony Computer Entertainment Inc. | Methods and systems for enabling depth and direction detection when interfacing with a computer program |
US7746321B2 (en) | 2004-05-28 | 2010-06-29 | Erik Jan Banning | Easily deployable interactive direct-pointing system and presentation control system and calibration method therefor |
US8560972B2 (en) * | 2004-08-10 | 2013-10-15 | Microsoft Corporation | Surface UI for gesture-based interaction |
US8547401B2 (en) | 2004-08-19 | 2013-10-01 | Sony Computer Entertainment Inc. | Portable augmented reality device and method |
US7852317B2 (en) | 2005-01-12 | 2010-12-14 | Thinkoptics, Inc. | Handheld device for handheld vision based absolute pointing system |
JP4441879B2 (en) * | 2005-06-28 | 2010-03-31 | ソニー株式会社 | Signal processing apparatus and method, program, and recording medium |
US9285897B2 (en) | 2005-07-13 | 2016-03-15 | Ultimate Pointer, L.L.C. | Easily deployable interactive direct-pointing system and calibration method therefor |
JP4773170B2 (en) * | 2005-09-14 | 2011-09-14 | 任天堂株式会社 | Game program and game system |
AU2007219710B2 (en) * | 2006-03-01 | 2011-10-13 | Brennan Broadcast Group Pty Ltd | Detecting contacts during sporting activities etc |
US20070265075A1 (en) * | 2006-05-10 | 2007-11-15 | Sony Computer Entertainment America Inc. | Attachable structure for use with hand-held controller having tracking ability |
US8384665B1 (en) * | 2006-07-14 | 2013-02-26 | Ailive, Inc. | Method and system for making a selection in 3D virtual environment |
US8913003B2 (en) | 2006-07-17 | 2014-12-16 | Thinkoptics, Inc. | Free-space multi-dimensional absolute pointer using a projection marker system |
US8096880B2 (en) * | 2006-08-15 | 2012-01-17 | Nintendo Co., Ltd. | Systems and methods for reducing jitter associated with a control device |
US8781151B2 (en) | 2006-09-28 | 2014-07-15 | Sony Computer Entertainment Inc. | Object detection using video input combined with tilt angle information |
USRE48417E1 (en) | 2006-09-28 | 2021-02-02 | Sony Interactive Entertainment Inc. | Object direction using video input combined with tilt angle information |
US8310656B2 (en) | 2006-09-28 | 2012-11-13 | Sony Computer Entertainment America Llc | Mapping movements of a hand-held controller to the two-dimensional image plane of a display screen |
US20080120115A1 (en) * | 2006-11-16 | 2008-05-22 | Xiao Dong Mao | Methods and apparatuses for dynamically adjusting an audio signal based on a parameter |
US8771071B2 (en) * | 2006-11-22 | 2014-07-08 | Sony Computer Entertainment America Llc | System and method of rendering controller information |
US20090011832A1 (en) * | 2007-01-31 | 2009-01-08 | Broadcom Corporation | Mobile communication device with game application for display on a remote monitor and methods for use therewith |
US20090017910A1 (en) * | 2007-06-22 | 2009-01-15 | Broadcom Corporation | Position and motion tracking of an object |
US8254319B2 (en) * | 2007-01-31 | 2012-08-28 | Broadcom Corporation | Wireless programmable logic device |
US20080320293A1 (en) * | 2007-01-31 | 2008-12-25 | Broadcom Corporation | Configurable processing core |
US8289944B2 (en) * | 2007-01-31 | 2012-10-16 | Broadcom Corporation | Apparatus for configuration of wireless operation |
US9486703B2 (en) | 2007-01-31 | 2016-11-08 | Broadcom Corporation | Mobile communication device with game application for use in conjunction with a remote mobile communication device and methods for use therewith |
US8121541B2 (en) * | 2007-01-31 | 2012-02-21 | Broadcom Corporation | Integrated circuit with intra-chip and extra-chip RF communication |
US20090197641A1 (en) * | 2008-02-06 | 2009-08-06 | Broadcom Corporation | Computing device with handheld and extended computing units |
US8200156B2 (en) * | 2007-01-31 | 2012-06-12 | Broadcom Corporation | Apparatus for allocation of wireless resources |
US8116294B2 (en) * | 2007-01-31 | 2012-02-14 | Broadcom Corporation | RF bus controller |
US8239650B2 (en) * | 2007-01-31 | 2012-08-07 | Broadcom Corporation | Wirelessly configurable memory device addressing |
US8438322B2 (en) * | 2007-01-31 | 2013-05-07 | Broadcom Corporation | Processing module with millimeter wave transceiver interconnection |
US8125950B2 (en) * | 2007-01-31 | 2012-02-28 | Broadcom Corporation | Apparatus for wirelessly managing resources |
US8204075B2 (en) * | 2007-01-31 | 2012-06-19 | Broadcom Corporation | Inter-device wireless communication for intra-device communications |
US8280303B2 (en) * | 2007-01-31 | 2012-10-02 | Broadcom Corporation | Distributed digital signal processor |
US8238275B2 (en) * | 2007-01-31 | 2012-08-07 | Broadcom Corporation | IC with MMW transceiver communications |
US8223736B2 (en) * | 2007-01-31 | 2012-07-17 | Broadcom Corporation | Apparatus for managing frequency use |
US20090046056A1 (en) * | 2007-03-14 | 2009-02-19 | Raydon Corporation | Human motion tracking device |
RU2472206C2 (en) * | 2007-03-30 | 2013-01-10 | Конинклейке Филипс Электроникс Н.В. | Space mouse-communication device |
US9176598B2 (en) | 2007-05-08 | 2015-11-03 | Thinkoptics, Inc. | Free-space multi-dimensional absolute pointer with improved performance |
US20090062943A1 (en) * | 2007-08-27 | 2009-03-05 | Sony Computer Entertainment Inc. | Methods and apparatus for automatically controlling the sound level based on the content |
JP5307014B2 (en) * | 2007-09-07 | 2013-10-02 | 株式会社コナミデジタルエンタテインメント | Image processing apparatus, game apparatus, and computer program |
WO2009038149A1 (en) * | 2007-09-20 | 2009-03-26 | Nec Corporation | Video image providing system and video image providing method |
TWI372645B (en) * | 2007-10-17 | 2012-09-21 | Cywee Group Ltd | An electronic game controller with motion-sensing capability |
US9066047B2 (en) * | 2007-12-19 | 2015-06-23 | Echostar Technologies L.L.C. | Apparatus, systems, and methods for accessing an interactive program |
US8149210B2 (en) * | 2007-12-31 | 2012-04-03 | Microsoft International Holdings B.V. | Pointing device and method |
US20090184981A1 (en) * | 2008-01-23 | 2009-07-23 | De Matos Lucio D Orazio Pedro | system, method and computer program product for displaying images according to user position |
US8117370B2 (en) * | 2008-02-06 | 2012-02-14 | Broadcom Corporation | IC for handheld computing unit of a computing device |
US8064952B2 (en) * | 2008-02-06 | 2011-11-22 | Broadcom Corporation | A/V control for a computing device with handheld and extended computing units |
US8195928B2 (en) * | 2008-02-06 | 2012-06-05 | Broadcom Corporation | Handheld computing unit with merged mode |
US8717974B2 (en) * | 2008-02-06 | 2014-05-06 | Broadcom Corporation | Handheld computing unit coordination of femtocell AP functions |
US20090198798A1 (en) * | 2008-02-06 | 2009-08-06 | Broadcom Corporation | Handheld computing unit back-up system |
US8175646B2 (en) * | 2008-02-06 | 2012-05-08 | Broadcom Corporation | Networking of multiple mode handheld computing unit |
US8368753B2 (en) * | 2008-03-17 | 2013-02-05 | Sony Computer Entertainment America Llc | Controller with an integrated depth camera |
US20090280901A1 (en) * | 2008-05-09 | 2009-11-12 | Dell Products, Lp | Game controller device and methods thereof |
US8430750B2 (en) * | 2008-05-22 | 2013-04-30 | Broadcom Corporation | Video gaming device with image identification |
TWM347881U (en) * | 2008-06-19 | 2009-01-01 | Chia-Ching Lee | Structure of elastic cloth cushion body |
US8133119B2 (en) * | 2008-10-01 | 2012-03-13 | Microsoft Corporation | Adaptation for alternate gaming input devices |
US8221229B2 (en) * | 2008-10-27 | 2012-07-17 | Sony Computer Entertainment Inc. | Spherical ended controller with configurable modes |
US20100120531A1 (en) * | 2008-11-13 | 2010-05-13 | Microsoft Corporation | Audio content management for video game systems |
JP2010122879A (en) * | 2008-11-19 | 2010-06-03 | Sony Ericsson Mobile Communications Ab | Terminal device, display control method, and display control program |
US8659590B1 (en) * | 2008-12-17 | 2014-02-25 | Nvidia Corporation | System, method, and computer program product for modifying signals of a three-dimensional graphics application program based on a tracking algorithm |
US8682028B2 (en) | 2009-01-30 | 2014-03-25 | Microsoft Corporation | Visual target tracking |
US8267781B2 (en) * | 2009-01-30 | 2012-09-18 | Microsoft Corporation | Visual target tracking |
US8295546B2 (en) | 2009-01-30 | 2012-10-23 | Microsoft Corporation | Pose tracking pipeline |
US8588465B2 (en) | 2009-01-30 | 2013-11-19 | Microsoft Corporation | Visual target tracking |
US8565477B2 (en) * | 2009-01-30 | 2013-10-22 | Microsoft Corporation | Visual target tracking |
US8866821B2 (en) * | 2009-01-30 | 2014-10-21 | Microsoft Corporation | Depth map movement tracking via optical flow and velocity prediction |
US9652030B2 (en) | 2009-01-30 | 2017-05-16 | Microsoft Technology Licensing, Llc | Navigation of a virtual plane using a zone of restriction for canceling noise |
US8577085B2 (en) | 2009-01-30 | 2013-11-05 | Microsoft Corporation | Visual target tracking |
US8577084B2 (en) * | 2009-01-30 | 2013-11-05 | Microsoft Corporation | Visual target tracking |
US8294767B2 (en) | 2009-01-30 | 2012-10-23 | Microsoft Corporation | Body scan |
US8565476B2 (en) * | 2009-01-30 | 2013-10-22 | Microsoft Corporation | Visual target tracking |
US10180746B1 (en) | 2009-02-26 | 2019-01-15 | Amazon Technologies, Inc. | Hardware enabled interpolating sensor and display |
US9740341B1 (en) | 2009-02-26 | 2017-08-22 | Amazon Technologies, Inc. | Capacitive sensing with interpolating force-sensitive resistor array |
US8773355B2 (en) * | 2009-03-16 | 2014-07-08 | Microsoft Corporation | Adaptive cursor sizing |
US9256282B2 (en) * | 2009-03-20 | 2016-02-09 | Microsoft Technology Licensing, Llc | Virtual object manipulation |
US8988437B2 (en) * | 2009-03-20 | 2015-03-24 | Microsoft Technology Licensing, Llc | Chaining animations |
KR101666995B1 (en) * | 2009-03-23 | 2016-10-17 | 삼성전자주식회사 | Multi-telepointer, virtual object display device, and virtual object control method |
US9498718B2 (en) * | 2009-05-01 | 2016-11-22 | Microsoft Technology Licensing, Llc | Altering a view perspective within a display environment |
US8181123B2 (en) | 2009-05-01 | 2012-05-15 | Microsoft Corporation | Managing virtual port associations to users in a gesture-based computing environment |
US8340432B2 (en) | 2009-05-01 | 2012-12-25 | Microsoft Corporation | Systems and methods for detecting a tilt angle from a depth image |
US9898675B2 (en) | 2009-05-01 | 2018-02-20 | Microsoft Technology Licensing, Llc | User movement tracking feedback to improve tracking |
US8253746B2 (en) | 2009-05-01 | 2012-08-28 | Microsoft Corporation | Determine intended motions |
US8638985B2 (en) | 2009-05-01 | 2014-01-28 | Microsoft Corporation | Human body pose estimation |
US8942428B2 (en) | 2009-05-01 | 2015-01-27 | Microsoft Corporation | Isolate extraneous motions |
US9015638B2 (en) * | 2009-05-01 | 2015-04-21 | Microsoft Technology Licensing, Llc | Binding users to a gesture based system and providing feedback to the users |
US9377857B2 (en) | 2009-05-01 | 2016-06-28 | Microsoft Technology Licensing, Llc | Show body position |
US8649554B2 (en) | 2009-05-01 | 2014-02-11 | Microsoft Corporation | Method to control perspective for a camera-controlled computer |
US8503720B2 (en) | 2009-05-01 | 2013-08-06 | Microsoft Corporation | Human body pose estimation |
US20100277470A1 (en) * | 2009-05-01 | 2010-11-04 | Microsoft Corporation | Systems And Methods For Applying Model Tracking To Motion Capture |
US8953029B2 (en) * | 2009-05-08 | 2015-02-10 | Sony Computer Entertainment America Llc | Portable device interaction via motion sensitive controller |
US20100295771A1 (en) * | 2009-05-20 | 2010-11-25 | Microsoft Corporation | Control of display objects |
US9417700B2 (en) * | 2009-05-21 | 2016-08-16 | Edge3 Technologies | Gesture recognition systems and related methods |
US8320619B2 (en) * | 2009-05-29 | 2012-11-27 | Microsoft Corporation | Systems and methods for tracking a model |
US9383823B2 (en) | 2009-05-29 | 2016-07-05 | Microsoft Technology Licensing, Llc | Combining gestures beyond skeletal |
US8176442B2 (en) * | 2009-05-29 | 2012-05-08 | Microsoft Corporation | Living cursor control mechanics |
US9182814B2 (en) * | 2009-05-29 | 2015-11-10 | Microsoft Technology Licensing, Llc | Systems and methods for estimating a non-visible or occluded body part |
US8379101B2 (en) | 2009-05-29 | 2013-02-19 | Microsoft Corporation | Environment and/or target segmentation |
US8856691B2 (en) * | 2009-05-29 | 2014-10-07 | Microsoft Corporation | Gesture tool |
US8542252B2 (en) | 2009-05-29 | 2013-09-24 | Microsoft Corporation | Target digitization, extraction, and tracking |
US8145594B2 (en) * | 2009-05-29 | 2012-03-27 | Microsoft Corporation | Localized gesture aggregation |
US8803889B2 (en) * | 2009-05-29 | 2014-08-12 | Microsoft Corporation | Systems and methods for applying animations or motions to a character |
US8509479B2 (en) | 2009-05-29 | 2013-08-13 | Microsoft Corporation | Virtual object |
US20100306716A1 (en) * | 2009-05-29 | 2010-12-02 | Microsoft Corporation | Extending standard gestures |
US20100302138A1 (en) * | 2009-05-29 | 2010-12-02 | Microsoft Corporation | Methods and systems for defining or modifying a visual representation |
US8625837B2 (en) * | 2009-05-29 | 2014-01-07 | Microsoft Corporation | Protocol and format for communicating an image from a camera to a computing environment |
US8418085B2 (en) * | 2009-05-29 | 2013-04-09 | Microsoft Corporation | Gesture coach |
US8744121B2 (en) * | 2009-05-29 | 2014-06-03 | Microsoft Corporation | Device for identifying and tracking multiple humans over time |
US9400559B2 (en) | 2009-05-29 | 2016-07-26 | Microsoft Technology Licensing, Llc | Gesture shortcuts |
US7914344B2 (en) * | 2009-06-03 | 2011-03-29 | Microsoft Corporation | Dual-barrel, connector jack and plug assemblies |
US20100311512A1 (en) * | 2009-06-04 | 2010-12-09 | Timothy James Lock | Simulator with enhanced depth perception |
US8390680B2 (en) | 2009-07-09 | 2013-03-05 | Microsoft Corporation | Visual representation expression based on player expression |
US9159151B2 (en) * | 2009-07-13 | 2015-10-13 | Microsoft Technology Licensing, Llc | Bringing a visual representation to life via learned input from the user |
US20110025689A1 (en) * | 2009-07-29 | 2011-02-03 | Microsoft Corporation | Auto-Generating A Visual Representation |
US9740340B1 (en) | 2009-07-31 | 2017-08-22 | Amazon Technologies, Inc. | Visually consistent arrays including conductive mesh |
US9785272B1 (en) | 2009-07-31 | 2017-10-10 | Amazon Technologies, Inc. | Touch distinction |
US9141193B2 (en) * | 2009-08-31 | 2015-09-22 | Microsoft Technology Licensing, Llc | Techniques for using human gestures to control gesture unaware programs |
US8438503B2 (en) * | 2009-09-02 | 2013-05-07 | Universal Electronics Inc. | System and method for enhanced command input |
US8508919B2 (en) * | 2009-09-14 | 2013-08-13 | Microsoft Corporation | Separation of electrical and optical components |
US8963829B2 (en) | 2009-10-07 | 2015-02-24 | Microsoft Corporation | Methods and systems for determining and tracking extremities of a target |
US8305188B2 (en) * | 2009-10-07 | 2012-11-06 | Samsung Electronics Co., Ltd. | System and method for logging in multiple users to a consumer electronics device by detecting gestures with a sensory device |
US8867820B2 (en) | 2009-10-07 | 2014-10-21 | Microsoft Corporation | Systems and methods for removing a background of an image |
US7961910B2 (en) | 2009-10-07 | 2011-06-14 | Microsoft Corporation | Systems and methods for tracking a model |
US8564534B2 (en) | 2009-10-07 | 2013-10-22 | Microsoft Corporation | Human tracking system |
GB2483168B (en) | 2009-10-13 | 2013-06-12 | Pointgrab Ltd | Computer vision gesture based control of a device |
US20110109617A1 (en) * | 2009-11-12 | 2011-05-12 | Microsoft Corporation | Visualizing Depth |
JP5320332B2 (en) * | 2010-03-19 | 2013-10-23 | 株式会社コナミデジタルエンタテインメント | GAME DEVICE, GAME DEVICE CONTROL METHOD, AND PROGRAM |
WO2011136783A1 (en) * | 2010-04-29 | 2011-11-03 | Hewlett-Packard Development Company L. P. | System and method for providing object information |
US9180365B2 (en) * | 2010-05-10 | 2015-11-10 | Sony Computer Entertainment America Llc | Polymorphic firearm controller |
US8396252B2 (en) | 2010-05-20 | 2013-03-12 | Edge 3 Technologies | Systems and related methods for three dimensional gesture recognition in vehicles |
US8655093B2 (en) | 2010-09-02 | 2014-02-18 | Edge 3 Technologies, Inc. | Method and apparatus for performing segmentation of an image |
US8666144B2 (en) | 2010-09-02 | 2014-03-04 | Edge 3 Technologies, Inc. | Method and apparatus for determining disparity of texture |
US8467599B2 (en) | 2010-09-02 | 2013-06-18 | Edge 3 Technologies, Inc. | Method and apparatus for confusion learning |
US8582866B2 (en) | 2011-02-10 | 2013-11-12 | Edge 3 Technologies, Inc. | Method and apparatus for disparity computation in stereo images |
US9354718B2 (en) | 2010-12-22 | 2016-05-31 | Zspace, Inc. | Tightly coupled interactive stereo display |
US8570320B2 (en) * | 2011-01-31 | 2013-10-29 | Microsoft Corporation | Using a three-dimensional environment model in gameplay |
US8970589B2 (en) | 2011-02-10 | 2015-03-03 | Edge 3 Technologies, Inc. | Near-touch interaction with a stereo camera grid structured tessellations |
US8942917B2 (en) | 2011-02-14 | 2015-01-27 | Microsoft Corporation | Change invariant scene recognition by an agent |
US20120249468A1 (en) * | 2011-04-04 | 2012-10-04 | Microsoft Corporation | Virtual Touchpad Using a Depth Camera |
US8620113B2 (en) | 2011-04-25 | 2013-12-31 | Microsoft Corporation | Laser diode modes |
JP2012242948A (en) * | 2011-05-17 | 2012-12-10 | Sony Corp | Display control device, method, and program |
US8760395B2 (en) | 2011-05-31 | 2014-06-24 | Microsoft Corporation | Gesture recognition techniques |
KR101302638B1 (en) | 2011-07-08 | 2013-09-05 | 더디엔에이 주식회사 | Method, terminal, and computer readable recording medium for controlling content by detecting gesture of head and gesture of hand |
US20130100015A1 (en) * | 2011-10-25 | 2013-04-25 | Kenneth Edward Salsman | Optical input devices |
US9672609B1 (en) | 2011-11-11 | 2017-06-06 | Edge 3 Technologies, Inc. | Method and apparatus for improved depth-map estimation |
US8635637B2 (en) | 2011-12-02 | 2014-01-21 | Microsoft Corporation | User interface presenting an animated avatar performing a media reaction |
US9100685B2 (en) | 2011-12-09 | 2015-08-04 | Microsoft Technology Licensing, Llc | Determining audience state or interest using passive sensor data |
US9070019B2 (en) | 2012-01-17 | 2015-06-30 | Leap Motion, Inc. | Systems and methods for capturing motion in three-dimensional space |
US10691219B2 (en) | 2012-01-17 | 2020-06-23 | Ultrahaptics IP Two Limited | Systems and methods for machine control |
US11493998B2 (en) | 2012-01-17 | 2022-11-08 | Ultrahaptics IP Two Limited | Systems and methods for machine control |
US9501152B2 (en) | 2013-01-15 | 2016-11-22 | Leap Motion, Inc. | Free-space user interface and control using virtual constructs |
US8693731B2 (en) | 2012-01-17 | 2014-04-08 | Leap Motion, Inc. | Enhanced contrast for object detection and characterization by optical imaging |
US8638989B2 (en) | 2012-01-17 | 2014-01-28 | Leap Motion, Inc. | Systems and methods for capturing motion in three-dimensional space |
US9679215B2 (en) | 2012-01-17 | 2017-06-13 | Leap Motion, Inc. | Systems and methods for machine control |
US20150253428A1 (en) | 2013-03-15 | 2015-09-10 | Leap Motion, Inc. | Determining positional information for an object in space |
US8933912B2 (en) * | 2012-04-02 | 2015-01-13 | Microsoft Corporation | Touch sensitive user interface with three dimensional input sensor |
US8898687B2 (en) | 2012-04-04 | 2014-11-25 | Microsoft Corporation | Controlling a media program based on a media reaction |
TWI444851B (en) * | 2012-04-18 | 2014-07-11 | Au Optronics Corp | Three-dimensional interactive system and method of three-dimensional interactive |
CA2775700C (en) | 2012-05-04 | 2013-07-23 | Microsoft Corporation | Determining a future portion of a currently presented media program |
US8938124B2 (en) | 2012-05-10 | 2015-01-20 | Pointgrab Ltd. | Computer vision based tracking of a hand |
US9170667B2 (en) | 2012-06-01 | 2015-10-27 | Microsoft Technology Licensing, Llc | Contextual user interface |
US8777226B1 (en) * | 2012-06-21 | 2014-07-15 | Robert Hubert Decker, Jr. | Proxy target system |
US8779908B2 (en) * | 2012-07-16 | 2014-07-15 | Shmuel Ur | System and method for social dancing |
US20140018169A1 (en) * | 2012-07-16 | 2014-01-16 | Zhong Yuan Ran | Self as Avatar Gaming with Video Projecting Device |
KR102035134B1 (en) * | 2012-09-24 | 2019-10-22 | 엘지전자 주식회사 | Image display apparatus and method for operating the same |
US9285893B2 (en) | 2012-11-08 | 2016-03-15 | Leap Motion, Inc. | Object detection and tracking with variable-field illumination devices |
US9081413B2 (en) * | 2012-11-20 | 2015-07-14 | 3M Innovative Properties Company | Human interaction system based upon real-time intention detection |
US9857470B2 (en) | 2012-12-28 | 2018-01-02 | Microsoft Technology Licensing, Llc | Using photometric stereo for 3D environment modeling |
US10609285B2 (en) | 2013-01-07 | 2020-03-31 | Ultrahaptics IP Two Limited | Power consumption in motion-capture systems |
US9626015B2 (en) | 2013-01-08 | 2017-04-18 | Leap Motion, Inc. | Power consumption in motion-capture systems with audio and optical signals |
US9459697B2 (en) | 2013-01-15 | 2016-10-04 | Leap Motion, Inc. | Dynamic, free-space user interactions for machine control |
US9632658B2 (en) | 2013-01-15 | 2017-04-25 | Leap Motion, Inc. | Dynamic user interactions for display control and scaling responsiveness of display objects |
US9940553B2 (en) | 2013-02-22 | 2018-04-10 | Microsoft Technology Licensing, Llc | Camera/object pose from predicted coordinates |
US9393695B2 (en) | 2013-02-27 | 2016-07-19 | Rockwell Automation Technologies, Inc. | Recognition-based industrial automation control with person and object discrimination |
US9804576B2 (en) | 2013-02-27 | 2017-10-31 | Rockwell Automation Technologies, Inc. | Recognition-based industrial automation control with position and derivative decision reference |
US9498885B2 (en) | 2013-02-27 | 2016-11-22 | Rockwell Automation Technologies, Inc. | Recognition-based industrial automation control with confidence-based decision support |
US9798302B2 (en) | 2013-02-27 | 2017-10-24 | Rockwell Automation Technologies, Inc. | Recognition-based industrial automation control with redundant system input support |
US10721448B2 (en) | 2013-03-15 | 2020-07-21 | Edge 3 Technologies, Inc. | Method and apparatus for adaptive exposure bracketing, segmentation and scene organization |
US9625995B2 (en) * | 2013-03-15 | 2017-04-18 | Leap Motion, Inc. | Identifying an object in a field of view |
US10620709B2 (en) | 2013-04-05 | 2020-04-14 | Ultrahaptics IP Two Limited | Customized gesture interpretation |
US9916009B2 (en) | 2013-04-26 | 2018-03-13 | Leap Motion, Inc. | Non-tactile interface systems and methods |
US9747696B2 (en) | 2013-05-17 | 2017-08-29 | Leap Motion, Inc. | Systems and methods for providing normalized parameters of motions of objects in three-dimensional space |
US10281987B1 (en) | 2013-08-09 | 2019-05-07 | Leap Motion, Inc. | Systems and methods of free-space gestural interaction |
US10846942B1 (en) | 2013-08-29 | 2020-11-24 | Ultrahaptics IP Two Limited | Predictive information for free space gesture control and communication |
US9632572B2 (en) | 2013-10-03 | 2017-04-25 | Leap Motion, Inc. | Enhanced field of view to augment three-dimensional (3D) sensory space for free-space gesture interpretation |
US9996638B1 (en) | 2013-10-31 | 2018-06-12 | Leap Motion, Inc. | Predictive information for free space gesture control and communication |
US9613262B2 (en) | 2014-01-15 | 2017-04-04 | Leap Motion, Inc. | Object detection and tracking for providing a virtual device experience |
US9547412B1 (en) * | 2014-03-31 | 2017-01-17 | Amazon Technologies, Inc. | User interface configuration to avoid undesired movement effects |
JP2016038889A (en) | 2014-08-08 | 2016-03-22 | リープ モーション, インコーポレーテッドLeap Motion, Inc. | Extended reality followed by motion sensing |
US9696795B2 (en) | 2015-02-13 | 2017-07-04 | Leap Motion, Inc. | Systems and methods of creating a realistic grab experience in virtual reality/augmented reality environments |
JP2018528551A (en) * | 2015-06-10 | 2018-09-27 | ブイタッチ・コーポレーション・リミテッド | Gesture detection method and apparatus on user reference space coordinate system |
US20170168592A1 (en) * | 2015-12-14 | 2017-06-15 | Volodymyr Mishyn | System and method for optical tracking |
US10102423B2 (en) * | 2016-06-30 | 2018-10-16 | Snap Inc. | Object modeling and replacement in a video stream |
US10798360B2 (en) * | 2017-03-23 | 2020-10-06 | Sony Interactive Entertainment Inc. | Information processing system, method for controlling same, and program |
JP6755843B2 (en) | 2017-09-14 | 2020-09-16 | 株式会社東芝 | Sound processing device, voice recognition device, sound processing method, voice recognition method, sound processing program and voice recognition program |
US11875012B2 (en) | 2018-05-25 | 2024-01-16 | Ultrahaptics IP Two Limited | Throwable interface for augmented reality and virtual reality environments |
CN110047487B (en) * | 2019-06-05 | 2022-03-18 | 广州小鹏汽车科技有限公司 | Wake-up method and device for vehicle-mounted voice equipment, vehicle and machine-readable medium |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5818424A (en) * | 1995-10-19 | 1998-10-06 | International Business Machines Corporation | Rod shaped device and data acquisition apparatus for determining the position and orientation of an object in space |
US20040046736A1 (en) | 1997-08-22 | 2004-03-11 | Pryor Timothy R. | Novel man machine interfaces and applications |
WO2004041379A2 (en) | 2002-10-30 | 2004-05-21 | Nike Inc. | Clothes with tracking marks for computer games |
GB2398691A (en) | 2003-02-21 | 2004-08-25 | Sony Comp Entertainment Europe | Control of data processing in dependence on detection of motion in an image region associated with a processor control function |
US20050059488A1 (en) | 2003-09-15 | 2005-03-17 | Sony Computer Entertainment Inc. | Method and apparatus for adjusting a view of a scene being displayed according to tracked head motion |
Family Cites Families (285)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US3943277A (en) | 1969-02-20 | 1976-03-09 | The United States Of America As Represented By The Secretary Of The Navy | Digital memory area correlation tracker |
US4313227A (en) | 1979-01-29 | 1982-01-26 | Texas Instruments Incorporated | Light energy information transmission system |
US4263504A (en) | 1979-08-01 | 1981-04-21 | Ncr Corporation | High density matrix code |
US6772057B2 (en) | 1995-06-07 | 2004-08-03 | Automotive Technologies International, Inc. | Vehicular monitoring systems using image processing |
US4565999A (en) | 1983-04-01 | 1986-01-21 | Prime Computer, Inc. | Light pencil |
US4558864A (en) | 1984-06-13 | 1985-12-17 | Medwedeff Marion C | Handgrip exercising, computer game controller |
US4843568A (en) | 1986-04-11 | 1989-06-27 | Krueger Myron W | Real time perception of and response to the actions of an unencumbered participant/user |
JPS6347616A (en) | 1986-08-15 | 1988-02-29 | Ricoh Co Ltd | Measurement of moving quantity |
EP0348430A4 (en) | 1987-02-04 | 1992-08-19 | Mayo Foundation For Medical Education And Research | Joystick apparatus having six degrees freedom of motion |
US4802227A (en) | 1987-04-03 | 1989-01-31 | American Telephone And Telegraph Company | Noise reduction processing arrangement for microphone arrays |
GB2206716A (en) | 1987-07-06 | 1989-01-11 | Gen Electric Plc | Apparatus for monitoring the presence or movement of an object |
IT1219405B (en) | 1988-06-27 | 1990-05-11 | Fiat Ricerche | PROCEDURE AND DEVICE FOR INSTRUMENTAL VISION IN POOR CONDITIONS VISIBILITY IN PARTICULAR FOR DRIVING IN THE MIST |
NZ231883A (en) | 1988-12-20 | 1993-04-28 | Australian Meat & Live Stock | Tag with machine-readable optical code |
US5034986A (en) | 1989-03-01 | 1991-07-23 | Siemens Aktiengesellschaft | Method for detecting and tracking moving objects in a digital image sequence having a stationary background |
US5055840A (en) | 1990-01-16 | 1991-10-08 | Carroll Touch Incorporated | Infrared touch input device and light emitted activation circuit |
WO1991017522A1 (en) | 1990-05-01 | 1991-11-14 | Wang Laboratories, Inc. | Hands-free hardware keyboard |
US5111401A (en) | 1990-05-19 | 1992-05-05 | The United States Of America As Represented By The Secretary Of The Navy | Navigational control system for an autonomous vehicle |
US5662111A (en) | 1991-01-28 | 1997-09-02 | Cosman; Eric R. | Process of stereotactic optical navigation |
US5485273A (en) | 1991-04-22 | 1996-01-16 | Litton Systems, Inc. | Ring laser gyroscope enhanced resolution system |
US5534917A (en) | 1991-05-09 | 1996-07-09 | Very Vivid, Inc. | Video image based control system |
US5144594A (en) | 1991-05-29 | 1992-09-01 | Cyber Scientific | Acoustic mouse system |
US5455685A (en) | 1991-09-04 | 1995-10-03 | Fuji Photo Film Co., Ltd. | Video camera exposure control apparatus for controlling iris diaphragm and automatic gain control operating speed |
US5889670A (en) | 1991-10-24 | 1999-03-30 | Immersion Corporation | Method and apparatus for tactilely responsive user interface |
US5444462A (en) | 1991-12-16 | 1995-08-22 | Wambach; Mark L. | Computer mouse glove with remote communication |
US5790834A (en) | 1992-08-31 | 1998-08-04 | Intel Corporation | Apparatus and method using an ID instruction to identify a computer microprocessor |
JP3244798B2 (en) | 1992-09-08 | 2002-01-07 | 株式会社東芝 | Moving image processing device |
US7098891B1 (en) * | 1992-09-18 | 2006-08-29 | Pryor Timothy R | Method for providing human input to a computer |
US5394168A (en) | 1993-01-06 | 1995-02-28 | Smith Engineering | Dual-mode hand-held game controller |
US5335011A (en) | 1993-01-12 | 1994-08-02 | Bell Communications Research, Inc. | Sound localization system for teleconferencing using self-steering microphone arrays |
EP0613294B1 (en) | 1993-02-24 | 1998-10-28 | Matsushita Electric Industrial Co., Ltd. | Gradation correction device and image sensing device therewith |
JP3679426B2 (en) | 1993-03-15 | 2005-08-03 | マサチューセッツ・インスティチュート・オブ・テクノロジー | A system that encodes image data into multiple layers, each representing a coherent region of motion, and motion parameters associated with the layers. |
US5677710A (en) | 1993-05-10 | 1997-10-14 | Apple Computer, Inc. | Recognition keypad |
US5297061A (en) | 1993-05-19 | 1994-03-22 | University Of Maryland | Three dimensional pointing device monitored by computer vision |
US5959596A (en) | 1993-06-24 | 1999-09-28 | Nintendo Co., Ltd. | Airline-based video game and communications system |
US5473701A (en) | 1993-11-05 | 1995-12-05 | At&T Corp. | Adaptive microphone array |
JP2552427B2 (en) | 1993-12-28 | 1996-11-13 | コナミ株式会社 | Tv play system |
FR2714502A1 (en) | 1993-12-29 | 1995-06-30 | Philips Laboratoire Electroniq | An image processing method and apparatus for constructing from a source image a target image with perspective change. |
US5611000A (en) | 1994-02-22 | 1997-03-11 | Digital Equipment Corporation | Spline-based image registration |
JPH086708A (en) | 1994-04-22 | 1996-01-12 | Canon Inc | Display device |
US5543818A (en) | 1994-05-13 | 1996-08-06 | Sony Corporation | Method and apparatus for entering text using an input device having a small number of keys |
US5846086A (en) * | 1994-07-01 | 1998-12-08 | Massachusetts Institute Of Technology | System for human trajectory learning in virtual environments |
US5563988A (en) | 1994-08-01 | 1996-10-08 | Massachusetts Institute Of Technology | Method and system for facilitating wireless, full-body, real-time user interaction with a digitally represented visual environment |
WO1996017324A1 (en) | 1994-12-01 | 1996-06-06 | Namco Ltd. | Apparatus and method for image synthesizing |
JP3270643B2 (en) | 1994-12-22 | 2002-04-02 | キヤノン株式会社 | Pointed position detection method and device |
US5929444A (en) | 1995-01-31 | 1999-07-27 | Hewlett-Packard Company | Aiming device using radiated energy |
US5568928A (en) | 1995-02-01 | 1996-10-29 | Exertron, Inc. | Video game controller for use with an exercise apparatus |
US5638228A (en) | 1995-02-14 | 1997-06-10 | Iomega Corporation | Retroreflective marker for data storage cartridge |
US5930741A (en) | 1995-02-28 | 1999-07-27 | Virtual Technologies, Inc. | Accurate, rapid, reliable position sensing using multiple sensing technologies |
US5583478A (en) | 1995-03-01 | 1996-12-10 | Renzi; Ronald | Virtual environment tactile system |
US5900863A (en) | 1995-03-16 | 1999-05-04 | Kabushiki Kaisha Toshiba | Method and apparatus for controlling computer without touching input device |
US5706364A (en) | 1995-04-28 | 1998-01-06 | Xerox Corporation | Method of producing character templates using unsegmented samples |
EP0823683B1 (en) | 1995-04-28 | 2005-07-06 | Matsushita Electric Industrial Co., Ltd. | Interface device |
US5913727A (en) | 1995-06-02 | 1999-06-22 | Ahdoot; Ned | Interactive movement and contact simulation game |
US5649021A (en) | 1995-06-07 | 1997-07-15 | David Sarnoff Research Center, Inc. | Method and system for object detection for instrument control |
US6057909A (en) | 1995-06-22 | 2000-05-02 | 3Dv Systems Ltd. | Optical ranging camera |
IL114278A (en) | 1995-06-22 | 2010-06-16 | Microsoft Internat Holdings B | Camera and method |
DE69607472T2 (en) | 1995-07-26 | 2000-08-24 | Tegic Communications Inc | SYSTEM FOR SUPPRESSING AMBIANCE IN A REDUCED KEYBOARD |
US6311214B1 (en) | 1995-07-27 | 2001-10-30 | Digimarc Corporation | Linking of computers based on optical sensing of digital data |
US5611731A (en) | 1995-09-08 | 1997-03-18 | Thrustmaster, Inc. | Video pinball machine controller having an optical accelerometer for detecting slide and tilt |
US5768415A (en) | 1995-09-08 | 1998-06-16 | Lucent Technologies Inc. | Apparatus and methods for performing electronic scene analysis and enhancement |
US5850222A (en) | 1995-09-13 | 1998-12-15 | Pixel Dust, Inc. | Method and system for displaying a graphic image of a person modeling a garment |
US6281930B1 (en) | 1995-10-20 | 2001-08-28 | Parkervision, Inc. | System and method for controlling the field of view of a camera |
US5963250A (en) | 1995-10-20 | 1999-10-05 | Parkervision, Inc. | System and method for controlling the field of view of a camera |
US5719561A (en) | 1995-10-25 | 1998-02-17 | Gilbert R. Gonzales | Tactile communication device and method |
US5870100A (en) | 1995-11-22 | 1999-02-09 | Compaq Computer Corporation | Filling of graphical regions |
JP2000501033A (en) | 1995-11-30 | 2000-02-02 | ヴァーチャル テクノロジーズ インコーポレイテッド | Human / machine interface with tactile feedback |
JP4079463B2 (en) | 1996-01-26 | 2008-04-23 | ソニー株式会社 | Subject detection apparatus and subject detection method |
US6049619A (en) | 1996-02-12 | 2000-04-11 | Sarnoff Corporation | Method and apparatus for detecting moving objects in two- and three-dimensional scenes |
JP2891159B2 (en) | 1996-02-14 | 1999-05-17 | 日本電気株式会社 | Object detection method from multi-view images |
RU2069885C1 (en) | 1996-03-01 | 1996-11-27 | Йелстаун Корпорейшн Н.В. | Method and device for observing objects at low illumination intensity |
JPH09244793A (en) | 1996-03-04 | 1997-09-19 | Alps Electric Co Ltd | Input device |
TW419388B (en) | 1996-03-05 | 2001-01-21 | Sega Enterprises Kk | Controller and the expanding unit used by the controller |
WO1997039416A2 (en) | 1996-04-02 | 1997-10-23 | Cognex Corporation | Image formation apparatus for viewing indicia on a planar specular substrate |
US5937081A (en) | 1996-04-10 | 1999-08-10 | O'brill; Michael R. | Image composition system and method of using same |
US5923318A (en) | 1996-04-12 | 1999-07-13 | Zhai; Shumin | Finger manipulatable 6 degree-of-freedom input device |
US5917493A (en) | 1996-04-17 | 1999-06-29 | Hewlett-Packard Company | Method and apparatus for randomly generating information for subsequent correlating |
US6151009A (en) | 1996-08-21 | 2000-11-21 | Carnegie Mellon University | Method and apparatus for merging real and synthetic images |
US6400374B2 (en) | 1996-09-18 | 2002-06-04 | Eyematic Interfaces, Inc. | Video superposition system and method |
US5930383A (en) | 1996-09-24 | 1999-07-27 | Netzer; Yishay | Depth sensing camera systems and methods |
US5832931A (en) | 1996-10-30 | 1998-11-10 | Photogen, Inc. | Method for improved selectivity in photo-activation and detection of molecular diagnostic agents |
NL1004648C2 (en) | 1996-11-11 | 1998-05-14 | Johan Michiel Schaaij | Computer game system. |
US5914723A (en) | 1996-12-30 | 1999-06-22 | Sun Microsystems, Inc. | Method and system for converting images in computer systems |
US6243491B1 (en) | 1996-12-31 | 2001-06-05 | Lucent Technologies Inc. | Methods and apparatus for controlling a video system with visually recognized props |
US5850473A (en) | 1997-01-07 | 1998-12-15 | Lucent Technologies Inc. | Method and apparatus for compensating for color variation in a video system |
US6021219A (en) | 1997-01-07 | 2000-02-01 | Lucent Technologies Inc. | Methods and apparatus for distinguishing among several visual patterns |
US5796354A (en) | 1997-02-07 | 1998-08-18 | Reality Quest Corp. | Hand-attachable controller with direction sensing |
US5993314A (en) | 1997-02-10 | 1999-11-30 | Stadium Games, Ltd. | Method and apparatus for interactive audience participation by audio command |
US6009210A (en) | 1997-03-05 | 1999-12-28 | Digital Equipment Corporation | Hands-free interface to a virtual reality environment using head tracking |
WO1998039790A1 (en) | 1997-03-07 | 1998-09-11 | 3Dv Systems Ltd. | Optical shutter |
US6061055A (en) | 1997-03-21 | 2000-05-09 | Autodesk, Inc. | Method of tracking objects with an imaging device |
US6144367A (en) | 1997-03-26 | 2000-11-07 | International Business Machines Corporation | Method and system for simultaneous operation of multiple handheld control devices in a data processing system |
US6587573B1 (en) | 2000-03-20 | 2003-07-01 | Gentex Corporation | System for controlling exterior vehicle lights |
US8120652B2 (en) | 1997-04-02 | 2012-02-21 | Gentex Corporation | System for controlling vehicle equipment |
JP3009633B2 (en) | 1997-04-03 | 2000-02-14 | コナミ株式会社 | Image apparatus, image display method, and recording medium |
US6215898B1 (en) | 1997-04-15 | 2001-04-10 | Interval Research Corporation | Data processing system and method |
US6809776B1 (en) | 1997-04-23 | 2004-10-26 | Thomson Licensing S.A. | Control of video level by region and content of information displayed |
JP4732553B2 (en) | 1997-04-23 | 2011-07-27 | トムソン コンシユーマ エレクトロニクス インコーポレイテツド | Video processing system and method for providing an image having one or more regions |
US6428411B1 (en) | 1997-05-02 | 2002-08-06 | Konami Co., Ltd. | Volleyball video game system |
JP3183632B2 (en) | 1997-06-13 | 2001-07-09 | 株式会社ナムコ | Information storage medium and image generation device |
US6075895A (en) | 1997-06-20 | 2000-06-13 | Holoplex | Methods and apparatus for gesture recognition based on templates |
WO1999006956A1 (en) | 1997-07-29 | 1999-02-11 | Koninklijke Philips Electronics N.V. | Method of reconstruction of tridimensional scenes and corresponding reconstruction device and decoding system |
US6044181A (en) | 1997-08-01 | 2000-03-28 | Microsoft Corporation | Focal length estimation method and apparatus for construction of panoramic mosaic images |
US20020036617A1 (en) | 1998-08-21 | 2002-03-28 | Timothy R. Pryor | Novel man machine interfaces and applications |
US6297838B1 (en) | 1997-08-29 | 2001-10-02 | Xerox Corporation | Spinning as a morpheme for a physical manipulatory grammar |
US6243074B1 (en) | 1997-08-29 | 2001-06-05 | Xerox Corporation | Handedness detection for a physical manipulatory grammar |
US6101289A (en) | 1997-10-15 | 2000-08-08 | Electric Planet, Inc. | Method and apparatus for unencumbered capture of an object |
US6031934A (en) | 1997-10-15 | 2000-02-29 | Electric Planet, Inc. | Computer vision system for subject characterization |
US6072494A (en) | 1997-10-15 | 2000-06-06 | Electric Planet, Inc. | Method and apparatus for real-time gesture recognition |
WO1999019828A1 (en) | 1997-10-15 | 1999-04-22 | Electric Planet, Inc. | Method and apparatus for performing a clean background subtraction |
JPH11154240A (en) | 1997-11-20 | 1999-06-08 | Nintendo Co Ltd | Image producing device to produce image by using fetched image |
US6677987B1 (en) | 1997-12-03 | 2004-01-13 | 8×8, Inc. | Wireless user-interface arrangement and method |
US6195104B1 (en) | 1997-12-23 | 2001-02-27 | Philips Electronics North America Corp. | System and method for permitting three-dimensional navigation through a virtual reality environment using camera-based gesture inputs |
US6690268B2 (en) | 2000-03-02 | 2004-02-10 | Donnelly Corporation | Video mirror systems incorporating an accessory module |
US6160540A (en) | 1998-01-12 | 2000-12-12 | Xerox Company | Zoomorphic computer user interface |
US6134346A (en) | 1998-01-16 | 2000-10-17 | Ultimatte Corp | Method for removing from an image the background surrounding a selected object |
JP4195189B2 (en) | 1998-02-08 | 2008-12-10 | スリーディーヴィー システムズ リミテッド | Large aperture light image shutter |
US6115052A (en) | 1998-02-12 | 2000-09-05 | Mitsubishi Electric Information Technology Center America, Inc. (Ita) | System for reconstructing the 3-dimensional motions of a human figure from a monocularly-viewed image sequence |
US6037942A (en) | 1998-03-10 | 2000-03-14 | Magellan Dis, Inc. | Navigation system character input device |
US6411392B1 (en) | 1998-04-15 | 2002-06-25 | Massachusetts Institute Of Technology | Method and apparatus for data hiding in printed images |
US6393142B1 (en) | 1998-04-22 | 2002-05-21 | At&T Corp. | Method and apparatus for adaptive stripe based patch matching for depth estimation |
US6173059B1 (en) | 1998-04-24 | 2001-01-09 | Gentner Communications Corporation | Teleconferencing system with visual feedback |
US6593956B1 (en) | 1998-05-15 | 2003-07-15 | Polycom, Inc. | Locating an audio source |
US6473516B1 (en) | 1998-05-22 | 2002-10-29 | Asa Systems, Inc. | Large capacity steganography |
JP3646969B2 (en) | 1998-05-25 | 2005-05-11 | 富士通株式会社 | 3D image display device |
JP3841132B2 (en) | 1998-06-01 | 2006-11-01 | 株式会社ソニー・コンピュータエンタテインメント | Input position detection device and entertainment system |
FR2780176B1 (en) | 1998-06-17 | 2001-01-26 | Gabriel Guary | SHOOTING GUN FOR VIDEO GAME |
US6513160B2 (en) * | 1998-06-17 | 2003-01-28 | Opentv, Inc. | System and method for promoting viewer interaction in a television system |
IL125221A0 (en) | 1998-07-06 | 1999-03-12 | Toy Control Ltd | Motion activation using passive sound source |
US6256398B1 (en) | 1998-08-22 | 2001-07-03 | Kenneth H. P. Chang | Encoding and decoding a message within an image |
US6970183B1 (en) | 2000-06-14 | 2005-11-29 | E-Watch, Inc. | Multimedia surveillance and monitoring system including network configuration |
US6621938B1 (en) | 1998-09-18 | 2003-09-16 | Fuji Photo Film Co., Ltd. | Image capture apparatus and method |
US6184863B1 (en) | 1998-10-13 | 2001-02-06 | The George Washington University | Direct pointing apparatus and method therefor |
DE19849515C1 (en) | 1998-10-19 | 2000-12-28 | Gerhard Wergen | Method for transferring characters in particular to a computer and input device using this method |
US6307568B1 (en) | 1998-10-28 | 2001-10-23 | Imaginarix Ltd. | Virtual dressing over the internet |
US6147678A (en) * | 1998-12-09 | 2000-11-14 | Lucent Technologies Inc. | Video hand image-three-dimensional computer interface with multiple degrees of freedom |
JP2000172431A (en) | 1998-12-09 | 2000-06-23 | Sony Corp | Information input device and game device |
US6533420B1 (en) | 1999-01-22 | 2003-03-18 | Dimension Technologies, Inc. | Apparatus and method for generating and projecting autostereoscopic images |
JP2000261752A (en) | 1999-03-05 | 2000-09-22 | Hewlett Packard Co <Hp> | Device for recording and reproducing image having sound |
US7139767B1 (en) | 1999-03-05 | 2006-11-21 | Canon Kabushiki Kaisha | Image processing apparatus and database |
US6323942B1 (en) | 1999-04-30 | 2001-11-27 | Canesta, Inc. | CMOS-compatible three-dimensional image sensor IC |
US7164413B2 (en) | 1999-05-19 | 2007-01-16 | Digimarc Corporation | Enhanced input peripheral |
US6791531B1 (en) | 1999-06-07 | 2004-09-14 | Dot On, Inc. | Device and method for cursor motion control calibration and object selection |
JP2000350865A (en) | 1999-06-11 | 2000-12-19 | Mr System Kenkyusho:Kk | Game device for composite real space, image processing method therefor and program storage medium |
JP2001009154A (en) * | 1999-06-30 | 2001-01-16 | Namco Ltd | Shooting game device |
US6952198B2 (en) | 1999-07-06 | 2005-10-04 | Hansen Karl C | System and method for communication with enhanced optical pointer |
US6819318B1 (en) | 1999-07-23 | 2004-11-16 | Z. Jason Geng | Method and apparatus for modeling via a three-dimensional image mosaic system |
US6545706B1 (en) | 1999-07-30 | 2003-04-08 | Electric Planet, Inc. | System, method and article of manufacture for tracking a head of a camera-generated image of a person |
US6417836B1 (en) | 1999-08-02 | 2002-07-09 | Lucent Technologies Inc. | Computer input device having six degrees of freedom for controlling movement of a three-dimensional object |
JP2001056742A (en) | 1999-08-19 | 2001-02-27 | Alps Electric Co Ltd | Input device |
US6556704B1 (en) | 1999-08-25 | 2003-04-29 | Eastman Kodak Company | Method for forming a depth image from digital image data |
WO2001018563A1 (en) | 1999-09-08 | 2001-03-15 | 3Dv Systems, Ltd. | 3d imaging system |
EP1220547A4 (en) | 1999-09-17 | 2005-09-21 | Nature Technology Co Ltd | Image pickup system, image processor, and camera |
US6441825B1 (en) | 1999-10-04 | 2002-08-27 | Intel Corporation | Video token tracking system for animation |
JP3847058B2 (en) | 1999-10-04 | 2006-11-15 | 任天堂株式会社 | GAME SYSTEM AND GAME INFORMATION STORAGE MEDIUM USED FOR THE SAME |
US6753849B1 (en) | 1999-10-27 | 2004-06-22 | Ken Curran & Associates | Universal remote TV mouse |
US20050037844A1 (en) | 2002-10-30 | 2005-02-17 | Nike, Inc. | Sigils for use with apparel |
JP4403474B2 (en) | 1999-12-09 | 2010-01-27 | ソニー株式会社 | Tactile sense presentation mechanism and force-tactile sense presentation device using the same |
US6785329B1 (en) | 1999-12-21 | 2004-08-31 | Microsoft Corporation | Automatic video object extraction |
KR100384406B1 (en) | 2000-01-24 | 2003-05-22 | (주) 베스트소프트 | A program drive divice for computer |
WO2001061519A1 (en) | 2000-02-15 | 2001-08-23 | Sorceron, Inc. | Method and system for distributing captured motion data over a network |
JP2001265275A (en) | 2000-03-16 | 2001-09-28 | Olympus Optical Co Ltd | Picture display device |
US6924787B2 (en) | 2000-04-17 | 2005-08-02 | Immersion Corporation | Interface for controlling a graphical image |
TW527518B (en) | 2000-07-14 | 2003-04-11 | Massachusetts Inst Technology | Method and system for high resolution, ultra fast, 3-D imaging |
US6795068B1 (en) | 2000-07-21 | 2004-09-21 | Sony Computer Entertainment Inc. | Prop input device and method for mapping an object from a two-dimensional camera image to a three-dimensional space for controlling action in a game program |
AUPQ896000A0 (en) | 2000-07-24 | 2000-08-17 | Seeing Machines Pty Ltd | Facial image processing system |
US7227526B2 (en) | 2000-07-24 | 2007-06-05 | Gesturetek, Inc. | Video-based image control system |
AU2001287948A1 (en) | 2000-07-24 | 2002-02-05 | Herman Ehrenburg | Computer-compatible, color-coded manual input system |
JP3561463B2 (en) | 2000-08-11 | 2004-09-02 | コナミ株式会社 | Virtual camera viewpoint movement control method and 3D video game apparatus in 3D video game |
JP4815661B2 (en) | 2000-08-24 | 2011-11-16 | ソニー株式会社 | Signal processing apparatus and signal processing method |
EP1211640A3 (en) * | 2000-09-15 | 2003-10-15 | Canon Kabushiki Kaisha | Image processing methods and apparatus for detecting human eyes, human face and other objects in an image |
FR2814965A1 (en) | 2000-10-11 | 2002-04-12 | Janick Simeray | Remote communicating game personal computer having game transducers sound/ultrasonic signals transmitting and computer with microphone processing detections/identifying signals/computer programme interacting. |
US7061507B1 (en) | 2000-11-12 | 2006-06-13 | Bitboys, Inc. | Antialiasing method and apparatus for video applications |
KR100385563B1 (en) | 2000-12-01 | 2003-05-27 | 한국과학기술원 | Spectrophotometer With Driving Means And Intensity Of Light Measurement Method |
US6751338B1 (en) | 2000-12-15 | 2004-06-15 | Cognex Corporation | System and method of using range image data with machine vision tools |
US20020085097A1 (en) | 2000-12-22 | 2002-07-04 | Colmenarez Antonio J. | Computer vision-based wireless pointing system |
AU2002303082A1 (en) | 2001-01-26 | 2002-09-12 | Zaxel Systems, Inc. | Real-time virtual viewpoint in simulated reality environment |
DE10103922A1 (en) | 2001-01-30 | 2002-08-01 | Physoptics Opto Electronic Gmb | Interactive data viewing and operating system |
US6741741B2 (en) | 2001-02-01 | 2004-05-25 | Xerox Corporation | System and method for automatically detecting edges of scanned documents |
US6789967B1 (en) | 2001-02-02 | 2004-09-14 | George Forester | Distal chording keyboard |
US20020134151A1 (en) | 2001-02-05 | 2002-09-26 | Matsushita Electric Industrial Co., Ltd. | Apparatus and method for measuring distances |
US6746124B2 (en) | 2001-02-06 | 2004-06-08 | Robert E. Fischer | Flashlight producing uniform high brightness |
US7116330B2 (en) | 2001-02-28 | 2006-10-03 | Intel Corporation | Approximating motion using a three-dimensional model |
GB2376397A (en) | 2001-06-04 | 2002-12-11 | Hewlett Packard Co | Virtual or augmented reality |
US7259747B2 (en) * | 2001-06-05 | 2007-08-21 | Reactrix Systems, Inc. | Interactive video display system |
JP2003018604A (en) | 2001-07-04 | 2003-01-17 | Matsushita Electric Ind Co Ltd | Image signal encoding method, device thereof and recording medium |
US6931125B2 (en) | 2001-07-10 | 2005-08-16 | Gregory N. Smallwood | Telephone equipment compatible, twelve button alphanumeric keypad |
JP3611807B2 (en) * | 2001-07-19 | 2005-01-19 | コナミ株式会社 | Video game apparatus, pseudo camera viewpoint movement control method and program in video game |
US7039253B2 (en) | 2001-07-24 | 2006-05-02 | Casio Computer Co., Ltd. | Image display device, image display method, program, and projection system |
KR20030009919A (en) | 2001-07-24 | 2003-02-05 | 삼성전자주식회사 | Inputting device for computer game having inertial sense |
US7148922B2 (en) | 2001-08-02 | 2006-12-12 | Olympus Optical Co., Ltd. | Electronic camera which detects flash influence on an image and controls white balance in accordance with the flash influence |
JP3442754B2 (en) | 2001-08-10 | 2003-09-02 | 株式会社コナミコンピュータエンタテインメント東京 | Gun shooting game apparatus, computer control method and program |
US6709108B2 (en) | 2001-08-31 | 2004-03-23 | Adaptive Optics Associates, Inc. | Ophthalmic instrument with adaptive optic subsystem that measures aberrations (including higher order aberrations) of a human eye and that provides a view of compensation of such aberrations to the human eye |
KR100846761B1 (en) | 2001-09-11 | 2008-07-16 | 삼성전자주식회사 | Pointer control method, pointing apparatus and host apparatus therefor |
JP2003114640A (en) | 2001-10-04 | 2003-04-18 | Nec Corp | Plasma display panel and its driving method |
US20030093591A1 (en) | 2001-11-09 | 2003-05-15 | David Hohl | System and method for fast data transfer to display driver |
FR2832892B1 (en) | 2001-11-27 | 2004-04-02 | Thomson Licensing Sa | SPECIAL EFFECTS VIDEO CAMERA |
US20030100363A1 (en) | 2001-11-28 | 2003-05-29 | Ali Guiseppe C. | Method and apparatus for inputting appearance of computer operator into a computer program |
US20040070565A1 (en) | 2001-12-05 | 2004-04-15 | Nayar Shree K | Method and apparatus for displaying images |
KR20030048570A (en) | 2001-12-12 | 2003-06-25 | 한국전자통신연구원 | A keypad assembly with the supplementary buttons and its operating method |
US7106366B2 (en) | 2001-12-19 | 2006-09-12 | Eastman Kodak Company | Image capture system incorporating metadata to facilitate transcoding |
US7305114B2 (en) | 2001-12-26 | 2007-12-04 | Cognex Technology And Investment Corporation | Human/machine interface for a machine vision sensor and method for installing and operating the same |
US7436887B2 (en) | 2002-02-06 | 2008-10-14 | Playtex Products, Inc. | Method and apparatus for video frame sequence-based object tracking |
US6990639B2 (en) | 2002-02-07 | 2006-01-24 | Microsoft Corporation | System and process for controlling electronic components in a ubiquitous computing environment using multimodal integration |
US20030160862A1 (en) | 2002-02-27 | 2003-08-28 | Charlier Michael L. | Apparatus having cooperating wide-angle digital camera system and microphone array |
GB2388418A (en) | 2002-03-28 | 2003-11-12 | Marcus James Eales | Input or pointing device with a camera |
US6847311B2 (en) | 2002-03-28 | 2005-01-25 | Motorola Inc. | Method and apparatus for character entry in a wireless communication device |
US20030199324A1 (en) | 2002-04-23 | 2003-10-23 | Xiaoling Wang | Apparatus and a method for more realistic shooting video games on computers or similar devices using visible or invisible light |
JP2004021345A (en) | 2002-06-12 | 2004-01-22 | Toshiba Corp | Image processing device and its method |
US20040001082A1 (en) | 2002-06-26 | 2004-01-01 | Amir Said | System and method of interaction with a computer controlled image display system using a projected light source |
CA2390072C (en) | 2002-06-28 | 2018-02-27 | Adrian Gh Podoleanu | Optical mapping apparatus with adjustable depth resolution and multiple functionality |
US7227976B1 (en) | 2002-07-08 | 2007-06-05 | Videomining Corporation | Method and system for real-time facial image enhancement |
US8073157B2 (en) | 2003-08-27 | 2011-12-06 | Sony Computer Entertainment Inc. | Methods and apparatus for targeted sound detection and characterization |
US7102615B2 (en) | 2002-07-27 | 2006-09-05 | Sony Computer Entertainment Inc. | Man-machine interface using a deformable device |
US7613310B2 (en) | 2003-08-27 | 2009-11-03 | Sony Computer Entertainment Inc. | Audio input system |
US8797260B2 (en) | 2002-07-27 | 2014-08-05 | Sony Computer Entertainment Inc. | Inertially trackable hand-held controller |
US7783061B2 (en) | 2003-08-27 | 2010-08-24 | Sony Computer Entertainment Inc. | Methods and apparatus for the targeted sound detection |
US7697700B2 (en) | 2006-05-04 | 2010-04-13 | Sony Computer Entertainment Inc. | Noise removal for electronic device with far field microphone on console |
US7809145B2 (en) | 2006-05-04 | 2010-10-05 | Sony Computer Entertainment Inc. | Ultra small microphone array |
US7623115B2 (en) | 2002-07-27 | 2009-11-24 | Sony Computer Entertainment Inc. | Method and apparatus for light input device |
US7545926B2 (en) | 2006-05-04 | 2009-06-09 | Sony Computer Entertainment Inc. | Echo and noise cancellation |
US7970147B2 (en) | 2004-04-07 | 2011-06-28 | Sony Computer Entertainment Inc. | Video game controller with noise canceling logic |
US20040017355A1 (en) | 2002-07-24 | 2004-01-29 | Youngtack Shim | Cursor control systems and methods |
US7391409B2 (en) | 2002-07-27 | 2008-06-24 | Sony Computer Entertainment America Inc. | Method and system for applying gearing effects to multi-channel mixed input |
US20070015559A1 (en) | 2002-07-27 | 2007-01-18 | Sony Computer Entertainment America Inc. | Method and apparatus for use in determining lack of user activity in relation to a system |
US8139793B2 (en) | 2003-08-27 | 2012-03-20 | Sony Computer Entertainment Inc. | Methods and apparatus for capturing audio signals based on a visual image |
US7918733B2 (en) | 2002-07-27 | 2011-04-05 | Sony Computer Entertainment America Inc. | Multi-input game control mixer |
US9474968B2 (en) | 2002-07-27 | 2016-10-25 | Sony Interactive Entertainment America Llc | Method and system for applying gearing effects to visual tracking |
US8160269B2 (en) | 2003-08-27 | 2012-04-17 | Sony Computer Entertainment Inc. | Methods and apparatuses for adjusting a listening area for capturing sounds |
US10086282B2 (en) | 2002-07-27 | 2018-10-02 | Sony Interactive Entertainment Inc. | Tracking device for use in obtaining information for controlling game program execution |
US8313380B2 (en) | 2002-07-27 | 2012-11-20 | Sony Computer Entertainment America Llc | Scheme for translating movements of a hand-held controller into inputs for a system |
US20070260517A1 (en) | 2006-05-08 | 2007-11-08 | Gary Zalewski | Profile detection |
US7352359B2 (en) | 2002-07-27 | 2008-04-01 | Sony Computer Entertainment America Inc. | Method and system for applying gearing effects to inertial tracking |
US8686939B2 (en) | 2002-07-27 | 2014-04-01 | Sony Computer Entertainment Inc. | System, method, and apparatus for three-dimensional input control |
US7760248B2 (en) | 2002-07-27 | 2010-07-20 | Sony Computer Entertainment Inc. | Selective sound source listening in conjunction with computer interactive processing |
US20060264260A1 (en) | 2002-07-27 | 2006-11-23 | Sony Computer Entertainment Inc. | Detectable and trackable hand-held controller |
US7803050B2 (en) | 2002-07-27 | 2010-09-28 | Sony Computer Entertainment Inc. | Tracking device with sound emitter for use in obtaining information for controlling game program execution |
US20060256081A1 (en) | 2002-07-27 | 2006-11-16 | Sony Computer Entertainment America Inc. | Scheme for detecting and tracking user manipulation of a game controller body |
US7854655B2 (en) | 2002-07-27 | 2010-12-21 | Sony Computer Entertainment America Inc. | Obtaining input for controlling execution of a game program |
US7627139B2 (en) | 2002-07-27 | 2009-12-01 | Sony Computer Entertainment Inc. | Computer image and audio processing of intensity and input devices for interfacing with a computer program |
US20070061413A1 (en) | 2005-09-15 | 2007-03-15 | Larsen Eric J | System and method for obtaining user information from voices |
US20070261077A1 (en) | 2006-05-08 | 2007-11-08 | Gary Zalewski | Using audio/visual environment to select ads on game platform |
US7850526B2 (en) | 2002-07-27 | 2010-12-14 | Sony Computer Entertainment America Inc. | System for tracking user manipulations within an environment |
US8233642B2 (en) | 2003-08-27 | 2012-07-31 | Sony Computer Entertainment Inc. | Methods and apparatuses for capturing an audio signal based on a location of the signal |
US20060282873A1 (en) | 2002-07-27 | 2006-12-14 | Sony Computer Entertainment Inc. | Hand-held controller having detectable elements for tracking purposes |
US9393487B2 (en) | 2002-07-27 | 2016-07-19 | Sony Interactive Entertainment Inc. | Method for mapping movements of a hand-held controller to game commands |
JP3996015B2 (en) * | 2002-08-09 | 2007-10-24 | 本田技研工業株式会社 | Posture recognition device and autonomous robot |
US7039199B2 (en) | 2002-08-26 | 2006-05-02 | Microsoft Corporation | System and process for locating a speaker using 360 degree sound source localization |
US6917688B2 (en) | 2002-09-11 | 2005-07-12 | Nanyang Technological University | Adaptive noise cancelling microphone system |
US20040063480A1 (en) | 2002-09-30 | 2004-04-01 | Xiaoling Wang | Apparatus and a method for more realistic interactive video games on computers or similar devices |
EP1411461A1 (en) | 2002-10-14 | 2004-04-21 | STMicroelectronics S.r.l. | User controlled device for sending control signals to an electric appliance, in particular user controlled pointing device such as mouse or joystick, with 3D-motion detection |
US6995666B1 (en) | 2002-10-16 | 2006-02-07 | Luttrell Clyde K | Cellemetry-operated railroad switch heater |
JP2004145448A (en) | 2002-10-22 | 2004-05-20 | Toshiba Corp | Terminal device, server device, and image processing method |
US8206219B2 (en) * | 2002-10-30 | 2012-06-26 | Nike, Inc. | Interactive gaming apparel for interactive gaming |
US20040095327A1 (en) | 2002-11-14 | 2004-05-20 | Lo Fook Loong | Alphanumeric data input system and method |
US7576727B2 (en) * | 2002-12-13 | 2009-08-18 | Matthew Bell | Interactive directed light/sound system |
US8012025B2 (en) | 2002-12-13 | 2011-09-06 | Applied Minds, Llc | Video game controller hub with control input reduction and combination schemes |
US7212308B2 (en) | 2002-12-18 | 2007-05-01 | Morgan Carol L | Interactive photo kiosk |
US20040140955A1 (en) | 2003-01-21 | 2004-07-22 | Metz Kristofer Erik | Input device for a computer and method of operation |
US9177387B2 (en) | 2003-02-11 | 2015-11-03 | Sony Computer Entertainment Inc. | Method and apparatus for real time motion capture |
JP2004261236A (en) | 2003-02-20 | 2004-09-24 | Konami Co Ltd | Game system |
GB2398690B (en) | 2003-02-21 | 2006-05-10 | Sony Comp Entertainment Europe | Control of data processing |
US7161634B2 (en) | 2003-03-06 | 2007-01-09 | Huaya Microelectronics, Ltd. | Encoding system for error diffusion dithering |
US7519186B2 (en) | 2003-04-25 | 2009-04-14 | Microsoft Corporation | Noise reduction systems and methods for voice applications |
US7379559B2 (en) | 2003-05-28 | 2008-05-27 | Trw Automotive U.S. Llc | Method and apparatus for determining an occupant's head location in an actuatable occupant restraining system |
US8072470B2 (en) | 2003-05-29 | 2011-12-06 | Sony Computer Entertainment Inc. | System and method for providing a real-time three-dimensional interactive environment |
US6881147B2 (en) | 2003-06-06 | 2005-04-19 | Nyko Technologies, Inc. | Video game controller with integrated microphone and speaker |
TW571812U (en) | 2003-06-11 | 2004-01-11 | Vision Electronics Co Ltd | Audio device for TV game machine |
DE60308342T2 (en) | 2003-06-17 | 2007-09-06 | Sony Ericsson Mobile Communications Ab | Method and apparatus for voice activity detection |
AU2004287478A1 (en) | 2003-10-28 | 2005-05-19 | Welch Allyn, Inc. | Digital documenting ophthalmoscope |
US20050105777A1 (en) | 2003-11-18 | 2005-05-19 | Kozlowski William J.Jr. | Personalized multimedia summary |
US7296007B1 (en) | 2004-07-06 | 2007-11-13 | Ailive, Inc. | Real time context learning by software agents |
US7263462B2 (en) | 2004-07-30 | 2007-08-28 | Ailive, Inc. | Non-disruptive embedding of specialized elements |
US7613610B1 (en) | 2005-03-14 | 2009-11-03 | Escription, Inc. | Transcription data extraction |
KR100630118B1 (en) * | 2005-04-21 | 2006-09-27 | 삼성전자주식회사 | Internetwork optical fiber sharing system |
US7927216B2 (en) | 2005-09-15 | 2011-04-19 | Nintendo Co., Ltd. | Video game system with wireless modular handheld controller |
US7620316B2 (en) | 2005-11-28 | 2009-11-17 | Navisense | Method and device for touchless control of a camera |
US7834850B2 (en) | 2005-11-29 | 2010-11-16 | Navisense | Method and system for object control |
US7636645B1 (en) | 2007-06-18 | 2009-12-22 | Ailive Inc. | Self-contained inertial navigation system for interactive control using movable controllers |
JP4481280B2 (en) | 2006-08-30 | 2010-06-16 | 富士フイルム株式会社 | Image processing apparatus and image processing method |
US8277316B2 (en) | 2006-09-14 | 2012-10-02 | Nintendo Co., Ltd. | Method and apparatus for using a common pointing input to control 3D viewpoint and object targeting |
GB0622451D0 (en) | 2006-11-10 | 2006-12-20 | Intelligent Earth Ltd | Object position and orientation detection device |
US7937243B2 (en) | 2007-08-03 | 2011-05-03 | Ailive, Inc. | Method and apparatus for non-disruptive embedding of specialized elements |
US8419545B2 (en) | 2007-11-28 | 2013-04-16 | Ailive, Inc. | Method and system for controlling movements of objects in a videogame |
US20090221368A1 (en) | 2007-11-28 | 2009-09-03 | Ailive Inc., | Method and system for creating a shared game space for a networked game |
US8655622B2 (en) | 2008-07-05 | 2014-02-18 | Ailive, Inc. | Method and apparatus for interpreting orientation invariant motion |
-
2005
- 2005-12-12 US US11/301,673 patent/US7646372B2/en active Active
-
2006
- 2006-12-08 JP JP2008545704A patent/JP5084052B2/en active Active
- 2006-12-08 EP EP06839276A patent/EP1960069A1/en not_active Ceased
- 2006-12-08 WO PCT/US2006/047111 patent/WO2007078639A1/en active Application Filing
-
2009
- 2009-11-10 US US12/616,117 patent/US8568230B2/en not_active Expired - Lifetime
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5818424A (en) * | 1995-10-19 | 1998-10-06 | International Business Machines Corporation | Rod shaped device and data acquisition apparatus for determining the position and orientation of an object in space |
US20040046736A1 (en) | 1997-08-22 | 2004-03-11 | Pryor Timothy R. | Novel man machine interfaces and applications |
WO2004041379A2 (en) | 2002-10-30 | 2004-05-21 | Nike Inc. | Clothes with tracking marks for computer games |
GB2398691A (en) | 2003-02-21 | 2004-08-25 | Sony Comp Entertainment Europe | Control of data processing in dependence on detection of motion in an image region associated with a processor control function |
US20050059488A1 (en) | 2003-09-15 | 2005-03-17 | Sony Computer Entertainment Inc. | Method and apparatus for adjusting a view of a scene being displayed according to tracked head motion |
Non-Patent Citations (1)
Title |
---|
See also references of EP1960069A1 |
Cited By (71)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8888593B2 (en) | 2005-10-26 | 2014-11-18 | Sony Computer Entertainment Inc. | Directional input for a video game |
US8249334B2 (en) | 2006-05-11 | 2012-08-21 | Primesense Ltd. | Modeling of humanoid forms from depth maps |
US8690670B2 (en) | 2007-06-14 | 2014-04-08 | Harmonix Music Systems, Inc. | Systems and methods for simulating a rock band experience |
US8444486B2 (en) | 2007-06-14 | 2013-05-21 | Harmonix Music Systems, Inc. | Systems and methods for indicating input actions in a rhythm-action game |
US8678895B2 (en) | 2007-06-14 | 2014-03-25 | Harmonix Music Systems, Inc. | Systems and methods for online band matching in a rhythm action game |
US8439733B2 (en) | 2007-06-14 | 2013-05-14 | Harmonix Music Systems, Inc. | Systems and methods for reinstating a player within a rhythm-action game |
US8678896B2 (en) | 2007-06-14 | 2014-03-25 | Harmonix Music Systems, Inc. | Systems and methods for asynchronous band interaction in a rhythm action game |
GB2451461A (en) * | 2007-07-28 | 2009-02-04 | Naveen Chawla | Camera based 3D user and wand tracking human-computer interaction system |
US8643598B2 (en) | 2007-09-19 | 2014-02-04 | Sony Corporation | Image processing apparatus and method, and program therefor |
US8896535B2 (en) | 2007-09-19 | 2014-11-25 | Sony Corporation | Image processing apparatus and method, and program therefor |
JP2009075685A (en) * | 2007-09-19 | 2009-04-09 | Sony Corp | Image processor, image processing method, and program |
EP2227786A4 (en) * | 2007-12-17 | 2014-01-08 | Sony Comp Entertainment Us | Dynamic three-dimensional object mapping for user-defined control device |
EP2227786A1 (en) * | 2007-12-17 | 2010-09-15 | Sony Computer Entertainment America, Inc. | Dynamic three-dimensional object mapping for user-defined control device |
WO2009078993A1 (en) | 2007-12-17 | 2009-06-25 | Sony Computer Entertainment America Inc. | Dynamic three-dimensional object mapping for user-defined control device |
US9009747B2 (en) | 2008-01-11 | 2015-04-14 | Sony Computer Entertainment America, LLC | Gesture cataloging and recognition |
JP2012160199A (en) * | 2008-01-11 | 2012-08-23 | Sony Computer Entertainment America Llc | Gesture catalog generation and recognition |
US8839279B2 (en) | 2008-01-11 | 2014-09-16 | Sony Computer Entertainment America, LLC | Gesture cataloging and recognition |
JP2009165826A (en) * | 2008-01-11 | 2009-07-30 | Sony Computer Entertainment America Inc | Gesture cataloguing and recognition |
US8225343B2 (en) | 2008-01-11 | 2012-07-17 | Sony Computer Entertainment America Llc | Gesture cataloging and recognition |
US8166421B2 (en) | 2008-01-14 | 2012-04-24 | Primesense Ltd. | Three-dimensional user interface |
US9035876B2 (en) | 2008-01-14 | 2015-05-19 | Apple Inc. | Three-dimensional user interface session control |
US8663013B2 (en) | 2008-07-08 | 2014-03-04 | Harmonix Music Systems, Inc. | Systems and methods for simulating a rock band experience |
US8449360B2 (en) | 2009-05-29 | 2013-05-28 | Harmonix Music Systems, Inc. | Displaying song lyrics and vocal cues |
US8465366B2 (en) | 2009-05-29 | 2013-06-18 | Harmonix Music Systems, Inc. | Biasing a musical performance input to a part |
US8565479B2 (en) | 2009-08-13 | 2013-10-22 | Primesense Ltd. | Extraction of skeletons from 3D maps |
US10421013B2 (en) | 2009-10-27 | 2019-09-24 | Harmonix Music Systems, Inc. | Gesture-based user interface |
US10357714B2 (en) | 2009-10-27 | 2019-07-23 | Harmonix Music Systems, Inc. | Gesture-based user interface for navigating a menu |
US9981193B2 (en) | 2009-10-27 | 2018-05-29 | Harmonix Music Systems, Inc. | Movement based recognition and evaluation |
US8787663B2 (en) | 2010-03-01 | 2014-07-22 | Primesense Ltd. | Tracking body parts by combined color image and depth processing |
US8550908B2 (en) | 2010-03-16 | 2013-10-08 | Harmonix Music Systems, Inc. | Simulating musical instruments |
US8636572B2 (en) | 2010-03-16 | 2014-01-28 | Harmonix Music Systems, Inc. | Simulating musical instruments |
US8874243B2 (en) | 2010-03-16 | 2014-10-28 | Harmonix Music Systems, Inc. | Simulating musical instruments |
US8568234B2 (en) | 2010-03-16 | 2013-10-29 | Harmonix Music Systems, Inc. | Simulating musical instruments |
US9278286B2 (en) | 2010-03-16 | 2016-03-08 | Harmonix Music Systems, Inc. | Simulating musical instruments |
US8824737B2 (en) | 2010-05-31 | 2014-09-02 | Primesense Ltd. | Identifying components of a humanoid form in three-dimensional scenes |
US8781217B2 (en) | 2010-05-31 | 2014-07-15 | Primesense Ltd. | Analysis of three-dimensional scenes with a surface model |
US8594425B2 (en) | 2010-05-31 | 2013-11-26 | Primesense Ltd. | Analysis of three-dimensional scenes |
US8444464B2 (en) | 2010-06-11 | 2013-05-21 | Harmonix Music Systems, Inc. | Prompting a player of a dance game |
US8562403B2 (en) | 2010-06-11 | 2013-10-22 | Harmonix Music Systems, Inc. | Prompting a player of a dance game |
US8702485B2 (en) | 2010-06-11 | 2014-04-22 | Harmonix Music Systems, Inc. | Dance game and tutorial |
US9358456B1 (en) | 2010-06-11 | 2016-06-07 | Harmonix Music Systems, Inc. | Dance competition game |
EP2583155A4 (en) * | 2010-06-21 | 2013-10-16 | Microsoft Corp | Natural user input for driving interactive stories |
EP2583155A2 (en) * | 2010-06-21 | 2013-04-24 | Microsoft Corporation | Natural user input for driving interactive stories |
US9274747B2 (en) | 2010-06-21 | 2016-03-01 | Microsoft Technology Licensing, Llc | Natural user input for driving interactive stories |
US9158375B2 (en) | 2010-07-20 | 2015-10-13 | Apple Inc. | Interactive reality augmentation for natural interaction |
US9201501B2 (en) | 2010-07-20 | 2015-12-01 | Apple Inc. | Adaptive projector |
US9024166B2 (en) | 2010-09-09 | 2015-05-05 | Harmonix Music Systems, Inc. | Preventing subtractive track separation |
US8582867B2 (en) | 2010-09-16 | 2013-11-12 | Primesense Ltd | Learning-based pose estimation from depth maps |
US8959013B2 (en) | 2010-09-27 | 2015-02-17 | Apple Inc. | Virtual keyboard for a non-tactile three dimensional user interface |
US8872762B2 (en) | 2010-12-08 | 2014-10-28 | Primesense Ltd. | Three dimensional user interface cursor control |
US8933876B2 (en) | 2010-12-13 | 2015-01-13 | Apple Inc. | Three dimensional user interface session control |
US9454225B2 (en) | 2011-02-09 | 2016-09-27 | Apple Inc. | Gaze-based display control |
US9285874B2 (en) | 2011-02-09 | 2016-03-15 | Apple Inc. | Gaze detection in a 3D mapping environment |
US9342146B2 (en) | 2011-02-09 | 2016-05-17 | Apple Inc. | Pointing-based display interaction |
US8881051B2 (en) | 2011-07-05 | 2014-11-04 | Primesense Ltd | Zoom-based gesture user interface |
US9459758B2 (en) | 2011-07-05 | 2016-10-04 | Apple Inc. | Gesture-based interface with enhanced features |
US9377865B2 (en) | 2011-07-05 | 2016-06-28 | Apple Inc. | Zoom-based gesture user interface |
US9030498B2 (en) | 2011-08-15 | 2015-05-12 | Apple Inc. | Combining explicit select gestures and timeclick in a non-tactile three dimensional user interface |
US9218063B2 (en) | 2011-08-24 | 2015-12-22 | Apple Inc. | Sessionless pointing user interface |
US9122311B2 (en) | 2011-08-24 | 2015-09-01 | Apple Inc. | Visual feedback for tactile and non-tactile user interfaces |
US9002099B2 (en) | 2011-09-11 | 2015-04-07 | Apple Inc. | Learning-based estimation of hand and finger pose |
WO2013096844A3 (en) * | 2011-12-21 | 2013-08-22 | Sony Computer Entertainment Inc. | Directional input for a video game |
JP2015506739A (en) * | 2011-12-21 | 2015-03-05 | 株式会社ソニー・コンピュータエンタテインメント | Directional input for video games |
US9229534B2 (en) | 2012-02-28 | 2016-01-05 | Apple Inc. | Asymmetric mapping for tactile and non-tactile user interfaces |
US9377863B2 (en) | 2012-03-26 | 2016-06-28 | Apple Inc. | Gaze-enhanced virtual touchscreen |
US11169611B2 (en) | 2012-03-26 | 2021-11-09 | Apple Inc. | Enhanced virtual touchpad |
US9047507B2 (en) | 2012-05-02 | 2015-06-02 | Apple Inc. | Upper-body skeleton extraction from depth maps |
US9019267B2 (en) | 2012-10-30 | 2015-04-28 | Apple Inc. | Depth mapping with enhanced resolution |
US10362028B2 (en) | 2013-11-07 | 2019-07-23 | Sony Interactive Entertainment Inc. | Information processing apparatus |
US10043279B1 (en) | 2015-12-07 | 2018-08-07 | Apple Inc. | Robust detection and classification of body parts in a depth map |
US10366278B2 (en) | 2016-09-20 | 2019-07-30 | Apple Inc. | Curvature-based face detector |
Also Published As
Publication number | Publication date |
---|---|
JP2009519105A (en) | 2009-05-14 |
US8568230B2 (en) | 2013-10-29 |
US20070298882A1 (en) | 2007-12-27 |
US20100056277A1 (en) | 2010-03-04 |
JP5084052B2 (en) | 2012-11-28 |
US7646372B2 (en) | 2010-01-12 |
EP1960069A1 (en) | 2008-08-27 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US7646372B2 (en) | Methods and systems for enabling direction detection when interfacing with a computer program | |
US7874917B2 (en) | Methods and systems for enabling depth and direction detection when interfacing with a computer program | |
US10445884B2 (en) | Control device for communicating visual information | |
EP1636762B1 (en) | System and method for providing a real-time three-dimensional interactive environment | |
JP4425274B2 (en) | Method and apparatus for adjusting the view of a scene being displayed according to the motion of the head being tracked | |
US7102615B2 (en) | Man-machine interface using a deformable device | |
EP2293172A2 (en) | System and method for interfacing and computer program | |
KR20110023227A (en) | Robot game system relating virtual space to real space | |
EP3600578B1 (en) | Zoom apparatus and associated methods |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
WWE | Wipo information: entry into national phase |
Ref document number: 2006839276 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2008545704 Country of ref document: JP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |