WO2012032687A1 - 画像認識装置および操作判定方法並びにプログラム - Google Patents
画像認識装置および操作判定方法並びにプログラム Download PDFInfo
- Publication number
- WO2012032687A1 WO2012032687A1 PCT/JP2011/002136 JP2011002136W WO2012032687A1 WO 2012032687 A1 WO2012032687 A1 WO 2012032687A1 JP 2011002136 W JP2011002136 W JP 2011002136W WO 2012032687 A1 WO2012032687 A1 WO 2012032687A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- operator
- operation surface
- virtual operation
- image
- virtual
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/017—Gesture based interaction, e.g. based on a set of recognized hand gestures
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/20—Input arrangements for video game devices
- A63F13/21—Input arrangements for video game devices characterised by their sensors, purposes or types
- A63F13/213—Input arrangements for video game devices characterised by their sensors, purposes or types comprising photodetecting means, e.g. cameras, photodiodes or infrared cells
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/50—Controlling the output signals based on the game progress
- A63F13/53—Controlling the output signals based on the game progress involving additional visual information provided to the game scene, e.g. by overlay to simulate a head-up display [HUD] or displaying a laser sight in a shooting game
- A63F13/533—Controlling the output signals based on the game progress involving additional visual information provided to the game scene, e.g. by overlay to simulate a head-up display [HUD] or displaying a laser sight in a shooting game for prompting the player, e.g. by displaying a game menu
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/0304—Detection arrangements using opto-electronic means
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/033—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
- G06F3/0346—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of the device orientation or free movement in a 3D space, e.g. 3D mice, 6-DOF [six degrees of freedom] pointers using gyroscopes, accelerometers or tilt-sensors
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/107—Static hand or arm
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/20—Movements or behaviour, e.g. gesture recognition
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F2300/00—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
- A63F2300/10—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals
- A63F2300/1087—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals comprising photodetecting means, e.g. a camera
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F2300/00—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
- A63F2300/30—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by output arrangements for receiving control signals generated by the game device
- A63F2300/308—Details of the user interface
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
- H04N13/361—Reproducing mixed stereoscopic images; Reproducing mixed monoscopic and stereoscopic images, e.g. a stereoscopic image overlay window on a monoscopic image background
Definitions
- the present invention relates to an image recognition apparatus and an operation determination method, and more particularly, to an image recognition apparatus and an operation determination method for determining an operation of a measurement target from an image captured by a video camera or the like.
- Patent Document 1 includes a host computer that recognizes the shape and movement of an object in an image captured by a CCD camera, and a display that displays the shape and movement of the object recognized by the host computer. If you give an instruction by hand gesture or the like, the given hand gesture will be displayed on the display screen of the display, you can select the virtual switch etc. displayed on the display screen with the arrow cursor icon by hand gesture, and input devices such as a mouse There has been proposed a technique that enables operation of a very simple device without being required.
- an operation input is performed by recognizing as a kind of gesture from an image obtained by capturing the movement and shape of a finger.
- an operator who faces a large screen usually points to a camera installed at the bottom of the screen.
- the contents are displayed on the large screen.
- the operator's shape and movement are extracted from the image thus captured by a method known in the art, and compared with, for example, a predetermined pattern stored in the database, the operator's shape, The meaning of the movement, that is, what kind of operation is intended is determined and used for controlling the device.
- the operator's image reading technology can be used to reproduce a stereoscopic image by photographing the operator with a three-dimensional or stereoscopic camera. Used for applications.
- reproducing the stereoscopic image it is possible to grasp the movement of the operator in a three-dimensional manner. For example, since the movement of the operator, particularly the movement of the hand, can be recognized back and forth, the gesture can be compared with that using a two-dimensional image. Variety increases. Also, even if multiple operators are extracted as images, if it is a large number of people, it is a three-dimensional image, so the context can be known, and only the movement of the first operator can be extracted and used for operation input .
- the conventional gesture detection device is difficult to capture the clear intention of the operator.
- the present invention has been made in view of such problems, and by causing the operator to recognize that the user is performing an operation that leads to some input, the device can be operated accurately.
- An object is to provide an image recognition device and an operation determination method that enable determination. As a result, it is possible to determine as an operation that accurately indicates the will of the operator by operating the whole body or a part without learning the operation or learning special gestures.
- the invention described in claim 1 is an image recognition apparatus comprising: a three-dimensional imaging unit that reads an operator's image and generates stereoscopic image data; and a three-dimensional imaging unit.
- Three-dimensional imaging of operation surface forming means for forming a virtual operation surface at a shape and position based on the read operator's image and position, and movement of at least a part of the operator's image relative to the formed virtual operation surface Means for determining whether or not the movement is an operation based on the positional relationship between the part of the operator and the virtual operation surface, and when the movement is determined to be an operation, And a signal output means for outputting a signal.
- the operation candidate that performs a predetermined specific motion is selected as the operation. It is characterized by determining as a person.
- the operation candidate determined as the operator indicated by the images of the plurality of operation candidates read by the three-dimensional imaging means is currently And an operator display means for displaying the positional relationship between the operator and other operation candidates.
- the operation position on the virtual operation surface is determined by being arranged at a predetermined position on the operator side of the virtual operation surface.
- a position determination surface for performing the operation, and the operation determination unit is configured to select one of the operators on the position determination surface when crossing the virtual operation surface in a series of operations in which a part of the operator moves from the position determination surface to the virtual operation surface.
- the operation of the operator is determined on the assumption that a part of the operator has crossed the position on the virtual operation surface corresponding to the position crossed by the unit.
- the shape and position of the virtual operation surface are obtained by extracting the positions of the arm and the face from the read operator image. The determination is made based on the extracted arm and face positions.
- a sixth aspect of the present invention is the image recognition apparatus according to any one of the first to fifth aspects, wherein the position of the virtual operation surface is between an operator and the three-dimensional imaging means. .
- the operation surface forming means includes an operator image read by the three-dimensional imaging means and an operator's predetermined value.
- the virtual operation surface is formed with a shape and a position determined based on the specified specific movement.
- the operation surface forming means includes an operator image read by the three-dimensional imaging means and an operator's predetermined value.
- the virtual operation surface is formed at an angle determined based on the specified specific movement.
- the operation surface forming means is determined based on the image and position of the operator read by the three-dimensional imaging means.
- a virtual operation surface is formed at an angle.
- a tenth aspect of the present invention is the image recognition device according to any one of the first to ninth aspects, further comprising an operation surface stereoscopic display means for displaying a stereoscopic image indicating the formed virtual operation surface.
- the stereoscopic image is based on left and right eye parallax.
- the operator calculates a distance from a positional relationship between a part of the operator and the virtual operation surface, and moves the operator to a position corresponding to the distance.
- a predetermined sign indicating a positional relationship with respect to a part of the virtual operation surface is stereoscopically displayed.
- the invention according to claim 13 is an image recognition apparatus, which is a three-dimensional imaging unit that reads an operator's image and generates stereoscopic image data, and an operator's predetermined read by the three-dimensional imaging unit.
- An operation surface forming unit that forms a virtual operation surface with a shape and position based on a specific movement, and a motion of at least a part of an image of the operator with respect to the formed virtual operation surface is read by a three-dimensional imaging unit and operated.
- Operation determining means for determining whether or not the motion is an operation based on a positional relationship between a part of the person and the virtual operation surface, and a signal output for outputting a predetermined signal when the motion is determined to be an operation Means.
- the invention according to claim 14 is an operation determination method, comprising: a three-dimensional imaging step of reading an operator's image to generate stereoscopic image data; and an operator's image and position read in the three-dimensional imaging step.
- An operation surface forming step for forming a virtual operation surface with a shape and a position based on it, and a movement of at least a part of the image of the operator with respect to the formed virtual operation surface is read by a three-dimensional imaging means, and a part of the operator
- an operation determining step for determining whether or not the motion is an operation based on a positional relationship between the virtual operation surface and a signal output step for outputting a predetermined signal when the motion is determined to be an operation. It is characterized by that.
- a part of the operator is arranged at a predetermined position on the operator side of the virtual operation surface.
- Virtual operation surface corresponding to the position of a part of the operator on the position determination surface when crossing the virtual operation surface in a series of movements from the position determination surface to the virtual operation surface for determining the operation position at The operation of the operator is determined as crossing the upper position.
- the invention described in claim 16 is the operation determination method according to claim 14 or 15, further comprising an operation surface stereoscopic display step for displaying a stereoscopic image showing the formed virtual operation surface.
- the invention according to claim 17 is a program for causing an image recognition apparatus to execute an operation determination method for recognizing an operator's image and determining the operation content.
- the operation determination method reads an image of the operator.
- 3D imaging step for generating stereoscopic image data
- an operation surface forming step for forming a virtual operation surface with a shape and position based on the image and position of the operator read in the 3D imaging step
- the formed virtual operation An operation for determining whether or not the movement is an operation based on the positional relationship between the part of the operator and the virtual operation surface by reading the movement of at least a part of the image of the operator with respect to the surface with a three-dimensional imaging unit.
- a signal output step of outputting a predetermined signal when it is determined that the movement is an operation.
- a part of the operator is arranged at a predetermined position on the operator side of the virtual operation surface, and the operation determination step is performed on the virtual operation surface.
- the invention according to claim 19 is the program according to claim 17 or 18, further comprising an operation surface stereoscopic display step for displaying a stereoscopic image showing the formed virtual operation surface.
- the three-dimensional imaging unit that reads the operator's image and generates stereoscopic image data, and the operator's image and position read by the three-dimensional imaging unit.
- An operation surface forming unit that forms a virtual operation surface with a shape and a position based on the image and a movement of at least a part of the image of the operator with respect to the formed virtual operation surface is read by a three-dimensional imaging unit, and a part of the operator Operation determining means for determining whether or not the movement is an operation based on a positional relationship between the virtual operation surface and the virtual operation surface; and a signal output means for outputting a predetermined signal when the movement is determined to be an operation. Therefore, it is possible to determine as an operation that accurately indicates the will of the operator by operating the whole body or a part without learning the operation or learning a special gesture.
- FIG. 1 is a diagram illustrating an example of an operation input system according to the present embodiment.
- FIG. 2 is a block diagram schematically showing a relationship between the operation input system of this embodiment and a computer.
- FIG. 3 is a block diagram showing an example of a functional module of a program processed in the CPU of the computer according to the present embodiment.
- FIG. 4 is a flowchart of the processing of this embodiment.
- FIG. 5 is a diagram for explaining a virtual operation surface that is stereoscopically displayed according to an embodiment of the present invention.
- FIG. 6 is a diagram illustrating a virtual operation surface according to the embodiment of the present invention.
- FIG. 7 is a diagram showing a state of a virtual operation surface in a television display device which is a usage example of one embodiment of the present invention.
- FIG. 1 is a diagram illustrating an example of an operation input system according to the present embodiment.
- FIG. 2 is a block diagram schematically showing a relationship between the operation input system of this embodiment and a computer.
- FIG. 8 is a diagram showing the position of a virtual operation surface in a television display device that is an example of use of an embodiment of the present invention.
- FIG. 9 is a diagram illustrating a virtual operation surface formed based on the operation surface formation reference according to the embodiment of the present invention.
- FIG. 10 is a diagram illustrating an example of an image when a plurality of operators' images are captured using a conventional 3D camera.
- FIG. 11 is a diagram showing an example of operation area setting for operation input support according to an embodiment of the present invention.
- FIG. 12 is a diagram showing an example of adjustment of the operation area depending on the screen or the position of the camera according to the embodiment of the present invention.
- FIG. 13 is a diagram showing another example of adjustment of the operation area depending on the screen or the position of the camera according to the embodiment of the present invention.
- FIG. 14 is a diagram showing another example of the adjustment of the operation area according to the screen or the position of the camera according to the embodiment of the present invention.
- FIG. 15 is a diagram for explaining a method of adjusting the operation area according to the screen or the position of the camera according to the embodiment of the present invention.
- FIG. 16 is a diagram showing an example of adjustment of the operation area depending on the screen or the position of the camera according to the embodiment of the present invention.
- FIG. 17 is a diagram illustrating a method for capturing an image of an operator using a conventional 3D camera.
- FIG. 18 is a diagram for explaining an operation input using the stereoscopic display icon according to the embodiment of the present invention.
- FIG. 19 is a diagram showing an example of a specific display of operation input support according to an embodiment of the present invention.
- FIG. 20 is a diagram illustrating a virtual operation surface and an operation region according to the embodiment of the present invention.
- FIG. 21 is a diagram showing a relationship between an operator's movement and icons displayed on the screen according to the embodiment of the present invention.
- FIG. 22 is a diagram showing an example of a specific display of the operation input screen according to the embodiment of the present invention.
- FIG. 23 is a diagram showing examples of various icons that can be used on the operation input screen according to the embodiment of the present invention.
- FIG. 19 is a diagram for explaining an operation input using the stereoscopic display icon according to the embodiment of the present invention.
- FIG. 19 is a diagram showing an example of a specific display of operation input support according to an embodiment of the present invention.
- FIG. 20 is a
- FIG. 24 is a diagram showing a relationship between an operator's movement and an icon displayed on the screen according to the embodiment of the present invention.
- FIG. 25 is a diagram showing an example display screen for inputting an instruction to move a graphic displayed on the screen according to the present embodiment.
- FIG. 26 is a diagram showing a relationship between an operator's movement and a menu displayed on the screen according to the embodiment of the present invention.
- FIG. 27 is a diagram showing a relationship between an operator's movement and a menu displayed on the screen according to another embodiment of the present invention.
- FIG. 28 is a diagram showing a relationship between an operator's movement and a menu displayed on the screen according to still another embodiment of the present invention.
- FIG. 29 is a diagram illustrating a virtual operation surface and an operation surface formation reference according to an embodiment of the present invention.
- FIG. 30 is a diagram showing an example of adjustment of the operation area depending on the screen or the position of the camera by the projector according to the embodiment of the present invention.
- FIG. 31 is a diagram showing a relationship between an operator's movement and a menu displayed on the screen according to the embodiment of the present invention.
- FIG. 32 is a view showing an image photographed for determining an area for setting the position of the operation surface by the operator's gesture according to the embodiment of the present invention.
- FIG. 33 is a view for explaining extraction of the position of the designated operation surface according to the embodiment of the present invention.
- FIG. 34 is a diagram showing a distance data image acquired from the camera for extracting the position of the designated operation surface according to the embodiment of the present invention.
- FIG. 35 is a flowchart showing a process for setting the position of the operation surface by the operator's gesture according to the embodiment of the present invention.
- FIG. 36 is a flowchart showing processing for extracting the position of the designated operation surface according to the embodiment of the present invention.
- FIG. 37A is a diagram showing variations of the operation surface setting gesture of the operator according to the embodiment of the present invention.
- FIG. 37B is a diagram showing variations of the operation surface setting gesture of the operator according to the embodiment of the present invention.
- FIG. 37C is a diagram showing variations of the operation surface setting gesture of the operator according to the embodiment of the present invention.
- FIG. 37A is a diagram showing variations of the operation surface setting gesture of the operator according to the embodiment of the present invention.
- FIG. 37B is a diagram showing variations of the operation surface setting gesture of the operator according to the embodiment of the present
- FIG. 37D is a diagram showing variations of the operation surface setting gesture of the operator according to the embodiment of the present invention.
- FIG. 37E is a diagram showing variations of the operation surface setting gesture of the operator according to the embodiment of the present invention.
- FIG. 37F is a diagram showing variations of the operation surface setting gesture of the operator according to the embodiment of the present invention.
- FIG. 38 is a diagram showing an example of a target screen used for the grid suction processing according to the embodiment of the present invention.
- FIG. 39 is a diagram illustrating a setting example of a grid used for the grid suction processing according to the embodiment of the present invention.
- FIG. 40 is a view for explaining grid suction processing according to an embodiment of the present invention.
- FIG. 41 is a diagram showing an example of displaying an icon at the bottom of the three-dimensional display in order to specify the current operator of this embodiment.
- FIG. 42 shows the relationship between FIG. 42A and FIG. 42B.
- FIG. 42A is a diagram showing another example in which an icon is displayed at the bottom of the three-dimensional display in order to specify the current operator of this embodiment.
- FIG. 42B is a diagram showing another example in which an icon is displayed at the bottom of the three-dimensional display in order to specify the current operator of the present embodiment.
- FIG. 43A is a diagram showing examples of various variations for displaying the operator of the present embodiment.
- FIG. 43B is a diagram showing examples of various variations for displaying the operator of this embodiment.
- FIG. 43C is a diagram showing examples of various variations for displaying the operator of this embodiment.
- FIG. 1 is a diagram illustrating an example of an operation input system according to the present embodiment.
- the stereoscopic display three-dimensional display 111 of the present embodiment is disposed in front of the operator 102, and the operator 102 displays the virtual operation surface stereoscopically displayed at a fixed position between the operator 102 and the three-dimensional display 111. It is possible to perform an operation while being aware that the shape of a finger or the like is a target for operation determination.
- the three-dimensional display 111 displays various images for various applications targeted by the system.
- the three-dimensional display 111 supports operation input as described later, that is, for example, a part of the target operator 102. Can be displayed toward the corner of the screen to allow the operator 102 to recognize an action that can be determined as an operation at the present time.
- a 3D display is a display device that can present stereoscopic images with a sense of depth or popping out to viewers. There are various types of devices, but basically both eyes. On the other hand, by presenting separate images, the operator can make a stereoscopic view. In general, in order to perceive an object as if it is in the same space as you, an image with parallax between the left and right eyes is used.
- motion parallax ⁇ ⁇ ⁇ motion parallax It is necessary to change the appearance of the image according to the operator's head movement (called motion parallax ⁇ ⁇ ⁇ motion parallax) at the same time as presenting the binocular parallax binocular parallax.
- the presence of the operator's three-dimensional eye can be supplemented and correction can be performed in real time to further improve the sense of reality.
- Three-dimensional displays can be broadly divided into two types: those that allow viewers to wear glasses with special optical characteristics and display images with parallax in both eyes, and those that do not use glasses. Those that do not use glasses are particularly called autostereoscopic displays.
- the anaglyph type, the polarized glasses type, and the liquid crystal shutter are generally used as the three-dimensional display type that wears the glasses.
- the anaglyph method is to project the left and right images with red and blue light respectively and separate them with glasses with red and blue color filters. Although it was the simplest and lowest cost technically, it was previously limited to monochrome video, but now it is possible to produce video with color information remaining. However, the color balance is inevitably lost because you must pass the red and blue color filters when viewing.
- the polarized glasses type is designed to superimpose and project linearly polarized light orthogonal to the left and right images and separate them with a polarizing filter to provide different images to both eyes.
- a silver screen or the like is used to preserve the polarization state.
- the liquid crystal shutter glasses enable stereoscopic viewing by delivering different images to the left and right eyes using glasses driven by the liquid crystal shutter so that the left and right images are alternately shielded. Since images with parallax on the left and right are projected alternately at a rate twice the frame rate of the source, the color reproducibility is good, but the cost of the glasses increases and equipment for wirelessly transmitting signals to the glasses is also required become.
- the frame rate depends on the response frequency of the liquid crystal shutter.
- the movement of the operator 102 is photographed by the video camera 201, and the photographed image is processed by the computer 110, and is optimal depending on the position, height and arm length of the operator 102, or body dimension information such as height and shoulder width.
- the position and size of the virtual operation surface and the operation area including the virtual operation surface are set, and it is determined what operation is meant by the gesture of the portion coming out from the virtual operation surface to the 3D display 111 side. That is, the computer 110 creates a stereoscopic image of the operator 102 from the data obtained from the video camera 201, calculates the position of the virtual operation surface optimal for the operator from the created stereoscopic image, and will be described later.
- the operation content is determined using that portion as an operation target.
- the three-dimensional display 111 displays the virtual operation surface so that the virtual operation surface appears to exist at the set or adjusted position when viewed from the operator 102.
- the operator 102 can recognize the virtual operation surface as a 3D image as if it existed, and execute an operation on the recognized 3D image using a hand or a finger. This point will be described with reference to FIGS.
- FIG. 5 is a view for explaining a virtual operation surface displayed in three dimensions according to the present embodiment.
- the operator 102 views the image 502 and the virtual operation surface image 501 displayed on the three-dimensional display 111 with the right eye 504 and the left eye 503, and the virtual operation surface image 501 includes, for example, parallax in order to display the image in three dimensions.
- a virtual operation surface that is displayed in a three-dimensional manner as if it is in front of the operator is formed from the three-dimensional display 111, as shown in FIG. can do.
- the position where the virtual operation surface image 501 is displayed in three dimensions is determined from the image and position of the operator photographed by the video camera 201 described above, and is adjusted and displayed on the three-dimensional display 111 so that the convergence angle ⁇ 1 is obtained.
- the image 502 is displayed from the display normal two-dimensional at the position of the display 111, but is adjusted so that a congestion angle theta 2, the depth of be determined congestion angle theta 2 in accordance with the usage
- a certain stereoscopic image can be provided.
- the position of the virtual operation surface image 501 shown in FIG. 5 is a position for stereoscopic display of the virtual operation surface, but is displayed at a position slightly different from the position where the virtual operation surface is actually formed as an operational feeling of the operator. In some cases, the formation position of the virtual operation surface and the stereoscopic display position of the virtual operation surface image do not necessarily match.
- the virtual operation surface image 501 is stereoscopically displayed to be adjusted to the convergence angle ⁇ 1.
- the icon 1801 is further adjusted to be the convergence angle ⁇ 3 to be stereoscopically displayed.
- the three-dimensional display is performed such that the icon 1801 comes to the position of the hand or the like, thereby making it easy to grasp the sense of distance in the z direction to the virtual operation surface and the positional relationship in the xy direction.
- icons can be three-dimensionally displayed in various forms and shapes. For example, in addition to displaying the icon at the same position as the fingertip, the icon can be displayed with a certain distance in the depth direction.
- the positional relationship between the virtual surface and the icon may be easier to visually recognize than the method of looking at the fingertip, and if the raised distance to the front of the virtual surface is insufficient, this can be compensated.
- a stereoscopic display icon can be displayed in any shape or form of the focus icon described in detail in the second embodiment. For example, as a 3D display icon, while displaying an icon near the fingertip, the size of the icon can be changed or the color can be changed so that the operator can grasp the sense of distance to the virtual operation surface. it can.
- FIG. 6 is a diagram showing a state of a virtual operation surface according to the present embodiment, and a virtual operation surface image 501 including operation icons is displayed. Is displayed so that it can be touched if the hand on the front side of the 3D display is extended. Therefore, the operator can operate the virtual operation surface as if the touch panel or the operation panel exists there.
- the virtual operation surface image 501 once displayed in a three-dimensional manner can be changed in position and shape so that the operation can be performed more naturally in accordance with the movement of the arm and body of the operator 102.
- FIGS. 7 and 8 are views showing a virtual operation surface in the television display device which is an example of use of the present embodiment.
- An operation surface can be generated and operated individually, or various settings can be made, such as receiving instructions from any viewer preferentially.
- one viewer is determined by some method, and operation input is permitted only to the viewers.
- a plurality of viewers as shown in FIG. 7 or 8 are candidates to be operators who have authority to operate the apparatus, they are called operation candidates, and there are a plurality of operation candidates below. The determination of the operator and the processing after the operator determination will be described. Therefore, the operation candidates include a candidate as shown in FIG.
- an operator candidate who is a subject of the present invention for example, a plurality of players of a game machine, a presenter of a presentation, and the like.
- the person who is the target of the operator is limited to those that can be recognized as the target by the system, so that the operation candidates are limited to those that exist in the imaging range of the video camera 201, for example.
- the setting as to which of the plurality of viewers is set as the operator, the setting of the virtual operation surface, and the like are performed using the method of the second embodiment described later, that is, the virtual operation surface of the second embodiment.
- the setting it is possible to first determine who will be the operator from the operation candidates and then the determined operator can set the virtual operation surface, but any other known in this technical field It can also be done by the method. For example, in the example of FIG. 7, immediately after starting the television viewing, it is determined which operator (the operator is the viewer 702 in FIGS. 7 and 8) by some gesture, or a certain gesture or The setting mode can be set by operating the remote control. If the operator decides, the virtual operation surface image 501 is stereoscopically displayed according to the operator 702.
- the virtual operation surface is displayed in 3D as shown in the figures for viewers other than the operator 702 as shown in FIGS.
- the position where the virtual operation surface is formed is actually in front of the operator 702, so that it is difficult for other viewers to operate.
- a normal stereoscopic display system when there is a possibility that a plurality of persons can operate, a single operator is first determined, a virtual operation surface suitable for the operator is formed, and a stereoscopic display is performed. There are many cases that need to be done.
- FIG. 41 is a diagram showing an example in which icons 4202 to 4204 are displayed at the bottom of the three-dimensional display 111 in order to specify the current operator. Referring to FIG. 41, if three people watching TV 111 become operation candidates and one of them is an operator 702, in operation screen example 4201, three operation candidates are displayed at the bottom of the screen.
- Three boxes 4202 to 4204 are displayed, and the middle box 4203 is highlighted by coloring or the like so as to indicate the operator 702 who is the priority user.
- Such a display can be performed by extracting the position and order of the operation candidates from the read image since the operation candidates can be read by the video camera 201 as described above.
- FIGS. 42A and 42B are diagrams showing another example in which icons 4302 to 4304 are displayed at the bottom of the three-dimensional display 111 in order to specify the current operator.
- person icons 4302 to 4304 are used and displayed in an area 4301 at the bottom of the monitor screen 111, and the sense of distance is also shown. Since it is expressed by its size, it is possible to more intuitively recognize who is the operator.
- icons representing operation candidates can be displayed in the icon display area in consideration of the position and arrangement order as shown in FIG. 43A. .
- each icon is displayed with a depth based on the actual position of the corresponding operation candidate, and the operator who is the current priority user is more intuitive. Can be recognized.
- the display of the operator described with reference to FIGS. 41 to 42 is as described in the second embodiment described later without using a system that uses a normal two-dimensional display, that is, a three-dimensional icon. Needless to say, this is effective even when the virtual operation surface is set in a form.
- the video camera 201 is attached to the upper part of the three-dimensional display 111 in order to acquire an image.
- Any imaging means known in this technical field such as a camera can be used, and the installation location can be selected anywhere near the monitor.
- a three-dimensional (or 3D) camera as the video camera 201, a stereoscopic image including the operator can be created.
- a voice output device such as a speaker (not shown) is attached to the system of the present embodiment, and information regarding display contents and operations can be transmitted to the operator by voice.
- a voice output device such as a speaker (not shown) is attached to the system of the present embodiment, and information regarding display contents and operations can be transmitted to the operator by voice.
- the virtual operation surface 701 is set based on the height and arm length of the operator 102 or body dimension information such as height and shoulder width, and the user 102 virtually displays the operation surface 701 as a stereoscopic image.
- the hand 601 can be projected forward with the operation surface 701 as a reference to show a gesture.
- the action is determined by pushing (determining) the operation to the front of the virtual operation surface, or the operation is determined and then pushed out. Therefore, it is easy for the user to recognize and the operability is close to that of a conventional touch panel operation.
- operation variations are overwhelmingly larger than conventional touch panels (two-handed operation, behavior, multiple fingers, etc.).
- a virtual operation surface 701 as shown in FIGS. 9 and 10 is formed in real time when the camera 201 captures an image of the operator 102, but until the operator starts the operation. Since the operator's standing position is not constant, the virtual operation surface is not determined and the operation determination is not easy. Therefore, in the present embodiment, the setting process of the virtual operation surface can be started at the timing when the operator's body is stationary for a certain time within the imaging range of the three-dimensional camera.
- the virtual operation surface of the present embodiment can be formed in real time, but even in this case, the operation position can be limited by limiting the operator's standing position to a certain range optimum for the system.
- the determination can be made more accurate.
- a footprint indicating the standing position is drawn on the floor, or the operator can recognize the presence of a certain limited range by placing a monitor or system, or make a screen to operate within a certain range. It can also be made to do.
- the position and size of the virtual operation surface that the operator can recognize naturally are greatly affected by the positional relationship between the operator and the display, and the position of the display, camera, operator, etc. are assumed in advance throughout the system. Sometimes it is better.
- FIG. 2 is a block diagram schematically showing the structure of the computer 110 of the image recognition apparatus of this embodiment.
- the computer 110 is connected to a monitor 701 and connected to a video camera 201 that captures the operator 102 and the like, and the captured image is captured by the computer 110.
- the CPU 210 performs image extraction, position calculation, and the like, which are features of the present embodiment, in the CPU 210, and whether a part of the body comes out from the operation surface to the video camera based on the calculated position. To decide.
- the computer 110 generally includes a CPU 210, executes a program stored in the ROM 211 or the like on the RAM 212, and outputs a processing result based on an image input from the image recognition apparatus to the three-dimensional display 111 or the like.
- the three-dimensional display 111 can output various images provided by various applications that the operator wants to experience in addition to the above-described virtual operation surface, and operation input as described later. Also displays information that can help.
- FIG. 3 is a block diagram showing an example of a functional module of a program processed in the CPU 210 of the computer 110 of this embodiment.
- the processing in this system is executed by an image reading unit 301, an image extraction unit 302, an image position calculation unit 303, an operation determination unit 304, and a stereoscopic image display unit 305.
- processing from receiving an image from the video camera 201 to outputting data is executed by these five modules.
- the present invention is not limited to this, and other modules are used. It can be processed with few modules.
- the virtual operation surface is formed as a stereoscopic image based on the image of the operator 102 photographed by the video camera 201, and a hand that is a part of the operator 102 that is also photographed.
- a process for determining the position of the finger and calculating the positional relationship between the virtual operation surface 701 and the finger 601 of the operator 102 is performed.
- initial settings known in the art for example, when the image recognition apparatus of the present embodiment is newly installed, a video camera used as a preliminary preparation It is necessary to input information such as the distortion of the used lens 201 and the distance between the three-dimensional display 111 and the lens to the apparatus. Further, the threshold setting or the like is adjusted in advance. When the initial setting of the system is completed, the processing of this embodiment is performed. This processing will be described below with reference to FIG.
- FIG. 4 is a flowchart of processing according to this embodiment.
- data captured by the video camera 201 is read by the image reading unit 301 (S401), and an image of the operator is extracted from the data by the image extracting unit 302 (S402).
- the virtual operation surface and the operation area are determined based on the extracted image of the operator 102 (S403).
- the shape of the operation surface is a rectangle standing vertically from the floor surface with reference to FIG. 11 and the like, but is not limited thereto, and various shapes (sizes, sizes with respect to the display) depending on the operation mode of the operator.
- An operation surface (including an angle) can be formed. Specifically, the operator's body shape and position are photographed as a stereoscopic image with a three-dimensional camera, as in the setting of the virtual operation surface of the second embodiment to be described later, and the position and size of each part of the operator. To determine the optimal shape and position of the virtual operation surface.
- the formation position of the virtual operation surface can be set to a position where the stroke at the time of operation is optimal based on the arm length from the position of the eye of the operator, but the length and size of other parts
- the distance may be determined on the basis of, and the starting point may be not the eye but the shoulder. If it is difficult to extract the position of the eye, the position of the head can be used as the starting point.
- a virtual operation can be performed by attaching a marker on the front of the glasses or the like that can easily extract the position from the image of the three-dimensional camera. The formation position of the surface can be easily determined.
- the operation area includes a virtual operation surface that is a feature of the present embodiment, and is an area in which a hand, a finger, or the like that is a main operation of the operator is mainly moved.
- a certain area beyond the virtual operation surface from the operator's trunk is used for the operation recognition of the present invention.
- an adult operator 810 can be formed like an operation region 811 in consideration of the height (the position of the line of sight) and the length of the arm, and the child operator 820 In this case, the height becomes lower and the arm becomes shorter, so that the operation area 821 can be set accordingly.
- a virtual operation surface is set in such an operation area, the operator can move the hand or finger naturally so that the operation intended by the operator can be determined by the movement of the hand or finger. become.
- the depth is up to the fingertip where the operator has extended his hand
- the horizontal width is the length of the left and right wrists when the operator has extended his hand
- the height is the height of the operator.
- the range can be from the head position to the waist position.
- the target person of the system of the present embodiment is an elementary school elementary school student to an adult
- the height width is approximately 100 cm to 195 cm
- the correction width of the vertical position of the operation area or the virtual operation surface is as the height difference. , About 100 cm is required.
- the virtual operation surface and the operation area can be executed each time, can be executed under a certain condition, or the setting timing can be selected in advance or each time.
- the shape and position of the virtual operation surface are determined, the shape (size, angle with respect to the display) determined at the position determined by the operator is determined by any method known in the art.
- a stereoscopic image forming a virtual operation surface is generated by the stereoscopic image display unit 305 and displayed on the three-dimensional display 111 (S410). Accordingly, the operator can accurately touch the virtual operation surface displayed in three dimensions with a finger or a hand, and thus can perform an operation such as touching an icon displayed on the virtual operation surface.
- a series of flows until the virtual operation surface is displayed in three dimensions has been described.
- it is basically difficult to operate on the initially set virtual operation surface because the operator moves as a whole.
- the optimal shape and position are further calculated and the three-dimensional display is performed again, but the same processing is repeatedly executed depending on the content displayed as the virtual operation surface.
- the virtual operation surface may be displayed again with an optimal position and shape according to the movement of the finger or hand in order to make the operator's icon designation process easier. it can.
- the operation determination unit 304 uses the relative relationship between the three-dimensional virtual operation surface of the operation input system and the operator 102 (S404), and a part of the operator 102 is in front of the operation camera when viewed from the video camera 201 on the operation surface.
- S404 the operation has been started
- S406 any operation whose shape and movement are assumed in advance from the shape and movement of each part (open hand or two fingers etc.)
- S406 what kind of shape and movement corresponds to what kind of operation can be determined independently by the system, can be determined by adopting any method known in this technical field, or It can also be determined that the position has been touched by the operator, for example, that an icon has been selected or a button has been pressed.
- the determination result is executed by the computer 110 assuming that such an operation has been input (S407). If the hand has not come out from the virtual operation surface, it is determined that the operation has not been performed. The process ends (S408).
- the determination of the operation content is not limited to the method described here, and any method known in the present embodiment can be used. Although a specific determination method has also been omitted, generally, the shape and movement of the operator's body, such as a predetermined gesture, and the operation content meaning the same are stored in a database or the like, and after image extraction Then, this database is accessed to determine the operation content. Also in this case, of course, it is possible to improve the determination accuracy by using an image recognition technique, artificial intelligence, or the like by a method known in the technical field.
- the virtual operation surface is formed at which position and with a certain size depending on whether the operator is a child or an adult.
- a three-dimensional camera can measure the distance to an object in parallel or concentric with a CCD or lens surface. If the monitor is installed at the height of the operator's line of sight, the cameras are close to each other, and each is installed vertically on the floor surface, if the operator is also in a standing position, the mutual operation will be It can be said that there is no need for adjustment or correction, such as the positional relationship.
- various situations are assumed for the camera installation position and the positional relationship with the monitor and the operator.
- the virtual operation surface is always arranged perpendicular to the straight line connecting the operator's line of sight and the operation target screen. If the operation area along that line is not generated, the angle of the push stroke of the operator in the Z direction will not match, and even if the push operation to the point aimed by the operator is performed, any angle will be It may flow along the line and normal operation may not be possible. Therefore, when forming the virtual operation surface, it is preferable to adjust the angle, size, or position depending on the positions of the monitor, the camera and the operator, and the arrangement mode.
- the operation area 821 and the virtual operation surface 601 are determined according to the operator 820 as shown in FIG. 11, but the camera 201 is placed on the upper part of the three-dimensional display 111 as in the example shown in FIG.
- the virtual operation surface 601 is perpendicular to the viewing direction of the camera 201 because the operator 820 cannot obtain a good feeling of operation with respect to the virtual operation surface unless the operator 820 is perpendicular to the direction 910 in which the operator 820 extends his arm. It is preferable not to use a flat surface.
- the virtual operation surface 701 can be operated by the operator 820 looking up at the 3D display 111. It becomes a plane perpendicular to the direction 1010 having an upward inclination. Also in this case, since the field of view 1011 of the camera 201 is tilted at a constant angle with the line-of-sight direction 1010 as in the example shown in FIG. 12, the information read by the camera 201 is corrected to match the tilted virtual operation surface 701. Is desirable. Furthermore, referring to FIG. 14, the camera 201 is placed near the floor away from the 3D display 111, and the line of sight 1110 of the operator 820 and the field of view of the camera 201 have a larger angle. It is desirable to correct.
- FIG. 15 is a diagram for explaining an example of defining the virtual operation surface 701 and the operation area 821.
- the arm stroke may be extracted from the image of the operator, or may be determined from the information on the average arm length for each height separately from the obtained height information.
- the virtual operation surface and the operation area of the present embodiment are determined so that natural operation and easier operation determination can be performed based on the positions, arrangement modes, and the like of the camera, the monitor, and the operator. It is determined which operation is being performed by detecting the movement of the operator.
- specific processing that is not described here, such as how to specify the position and shape from the image of the three-dimensional camera, whether or not part of the operator has passed through the virtual operation surface, etc. Processing necessary for implementation of the embodiment can be achieved using any method known in the art.
- the virtual operation surface is stereoscopically displayed by the 3D video camera, so that the operator can recognize as if there is an operation surface such as a touch panel in the space.
- operation input such as displaying an operator's image of the virtual operation surface on the three-dimensional display 111 is also possible.
- the system of this embodiment can be utilized more easily.
- FIG. 19 is a diagram showing an example in which guidance for assisting such an operation input is displayed on the display 111.
- the operator points the desired location by projecting a finger against the image that is itself a virtual operation surface.
- the operator can execute the next operation while recognizing and confirming the currently performed operation.
- the pointer 901 is displayed on the screen when a finger is protruded from the operation surface, and disappears when the finger is retracted, or is displayed with shading, and is displayed on the display 111 with the movement of the hand.
- an operation screen 902 representing the state of the operator himself in the form as shown in FIGS. 9 and 10 is displayed small in the upper right corner of the three-dimensional display 111, and what kind of movement is currently performed in the system.
- a line graph 903 that graphs the movement of the hand so that the operator is aware of how the hand moves back and forth. More accurate operation can be expected.
- gestures that can be used in the system can be displayed in the guidance, and the operator can be urged to input an operation following the gesture.
- the operator operates the input operation device as if there is an input device such as a touch panel on the basis of the virtual operation surface virtually displayed in three dimensions in the space.
- the contents of the operation thus determined are to be determined with certainty, but the body such as the virtual operation surface and the operator's hand in the direction away from the operator, which is the back side direction from the virtual operation surface. It is determined by the positional relationship of a part of or a worn object.
- the operation area is set as a virtual operation hierarchy in two or three layers in the z-axis direction, which is the direction away from the operator, and the type of operation is determined according to which layer the operator's hand is in, The operation content is determined by the movement of the hand in the layer.
- the position of the hand, the type of operation, and the like are displayed on the display screen visually recognized by the operator, the operator can more easily recognize the operation.
- the distance in the z direction between a part of the operator and the plane dividing each layer can be obtained by a method of calculating the distance between the formed virtual operation surface and a part of the operator.
- the trigger surface 701 shown in FIG. 27 is a virtual operation surface of the present embodiment.
- the operation area ahead of the trigger plane 701 is divided into three levels A to C by planes 4501 and 4502, and different types of operations are assigned to each.
- an object rotation operation is assigned to the hierarchy A
- an enlargement / reduction operation is assigned to the hierarchy B
- an object movement operation is assigned to the hierarchy C.
- An operation determined by moving the finger 601 in each layer is executed.
- an icon indicating the finger 601 for example, an object designated around the position indicated by the rotation icon 4503 rotates in accordance with the movement of the finger 601.
- an enlargement / reduction icon 4504 is displayed on the three-dimensional display 111, and when the finger 601 is moved in the z direction, the object is enlarged, and when the finger 601 is moved in the opposite direction, the object can be reduced.
- the movement icon 4505 can be displayed at the position of the finger 601 on the designated object displayed on the three-dimensional display 111, and can be moved in accordance with the movement of the finger 601.
- the planes 4501 and 4502 separating the layers can be arranged so that each layer has the same thickness, or can be arranged so as to have different thicknesses according to the operation type assigned to the layer. For example, in the example of FIG. 27, an operation of enlargement / reduction is assigned to the hierarchy B. However, since the enlargement / reduction must be expressed by a forward / backward movement, it is usually in the z direction compared to the hierarchy A and hierarchy C. Since the movement becomes large, the operation can be facilitated by making the layer B thicker.
- the hierarchy as described above can be used with a natural feel of the operator by determining the interval by a standard stroke.
- the surfaces 4501 and 4502 that divide each layer are used as virtual operation surfaces.
- the boundary of the hierarchy can be surely recognized by the operator by performing a stereoscopic display at an appropriately set position.
- any or all of hierarchies A, B, and C can be stereoscopically displayed with gradation at the depth of the hierarchies to allow the operator to recognize the existence and depth of the hierarchies. It can be expressed more effectively by any display method known in the technical field.
- FIG. 28 is a diagram showing an example of another icon of the present embodiment.
- the operation for specifying the operation position on the three-dimensional display 111 is performed for the hierarchy A
- the operation for “grabbing” the object at the specified position is performed for the hierarchy B
- the object grasped for the hierarchy C is assigned to throw or move.
- the above icons can also be used in two-dimensional display, and can be displayed stereoscopically by adjusting the icon for stereoscopic display.
- a multi-sensing state in which a hand opposite to the one to be operated is inserted into the operation area is set to no operation (or vice versa).
- it is determined whether or not to operate each layer by putting in and out the hand opposite to the one to be operated in this example, depending on the two-hand operation, but on the XY plane)
- Various methods such as providing a slip-through area are conceivable).
- the operator can operate the system by the movement without memorizing or arranging the gesture in advance, and the posture of the operator and each part, for example, the movement of the hand
- so-called mixed reality (MR) can be realized in a game using the whole body.
- the present embodiment is different in that the virtual operation surface stereoscopically displayed in the first embodiment described above is set by causing the operator to specify in advance, but the operator is different from the system configuration and the virtual operation surface of the first embodiment. Processing such as how to recognize the operation performed by the operation is the same. That is, in this embodiment, when an operator performs an operation as will be described later, first, an instruction is given as to where to set the virtual operation surface by a certain operation, and a virtual operation surface is formed according to the instruction.
- the operator can recognize in advance where and what kind of operation surface exists without having to display the virtual operation surface in a three-dimensional manner using the three-dimensional display as in the system of the first embodiment, It is not always necessary to display the surface in three dimensions, and a normal two-dimensional display can also be used. However, in order to make the recognition of the virtual operation surface more reliable, a three-dimensional display can also be used for stereoscopic display.
- the process after the virtual operation surface is formed in accordance with the operator's instruction is basically the same as that in the first embodiment, and therefore the virtual operation surface setting process will be mainly described below.
- FIG. 32 is a diagram illustrating an image photographed to determine a region for setting the position of the operation surface by the operator's gesture according to the present embodiment.
- the shape (including size and angle with respect to the display) of the operation surface is determined by the hands 3202 and 3203 of the operator image 3210 captured by the three-dimensional camera 201, and the distance in the z direction.
- the position of the virtual operation surface 3201 is determined by the distance from the display of the data image 3211 to the hands 3202 and 3203 and the distance from the display to the trunk 3204.
- the depth can be grasped from the image data of the operator photographed by the three-dimensional camera 201.
- FIG. 33 is a diagram for explaining the extraction of the position of the designated operation surface according to the present embodiment, and shows the image of the operator in the depth direction (z direction). Referring to FIG. 33, it can be understood that the operator may form the operation surface 3201 at the position indicated by the hand 3202.
- FIG. 35 is a flowchart illustrating processing for setting the position of the operation surface by the operator's gesture according to the present embodiment.
- the person who sets the virtual operation surface can be determined as the operator, and the virtual set at this time is set.
- the operation surface is for the determined operator.
- operation candidates including the determined operator are displayed by displaying the operator's icons and boxes as shown in FIGS. 41 to 42 on the screen. , Which one is the operator can be recognized to prevent erroneous operation.
- a system program is started by turning on the power (S3501), and device management such as various settings of devices used in the system is executed (S3502).
- the virtual operation surface frame acquisition process is started (S3503), the instruction input from the operator is awaited (S3504), image data is acquired when the instruction is input (S3505), and operation surface extraction described later is performed.
- the device post-process is executed to finish the setting of the virtual operation surface (S3509).
- FIG. 34 is a diagram showing a distance data image acquired from the camera for extracting the position of the designated operation surface according to this embodiment
- FIG. 36 is a diagram of the position of the designated operation surface according to this embodiment. It is a flowchart which shows the process to extract.
- data that is, the operator's stereoscopic image data is input (S3601)
- the hand shape and head are recognized (S3602, S3603), and it is determined whether or not a virtual operation surface is to be formed (S3602). S3604).
- the hand when the hand has maintained a certain predetermined gesture for a certain period of time, it is determined that the setting process of the virtual operation surface is performed, and the position and size indicated by the operator by the hand are included.
- a virtual operation surface having a shape is extracted from a distance data image as shown in FIG. 34 to form a virtual operation surface (S3605), and measured values of the operator's body, for example, arm length, torso and head positions
- the operation area is set based on (S3606) and the data is output and the process ends (S3609).
- the movement of the hand or the like is determined to be determined as a virtual operation surface setting instruction by standing still for a certain period of time.
- the present invention is not limited to this.
- the instruction can be determined by any method known in the technical field.
- the virtual operation surface is basically formed in the shape and position formed by the hand.
- the virtual operation surface can be adjusted according to the physical characteristics of the operator, the position of the display, the usage situation, and the like.
- the virtual operation surface setting process is performed based on the operator's instruction.
- the position of the virtual operation surface is indicated to the operator by some other method. It can be recognized without it.
- various gestures for instructing the shape and position of the virtual operation surface can be considered with ordinary knowledge in this technical field, and any of them can be used in this embodiment. Examples of such variations are shown in FIGS.
- the operator operates the input operation device as if there is an input device such as a touch panel on the basis of the virtual operation surface virtually displayed in three dimensions in the space.
- the operation support principle is based on the operator's part on the virtual operation surface, for example, what kind of operation the operator wants to perform on the three-dimensional display 111 according to the movement of the position of the hand or finger. By visually displaying whether or not, the operator is guided to enable accurate operation input.
- the operator when the operator can operate at a certain standing position in advance, it is suitable to operate the virtual operation surface at the standing position.
- the virtual operation surface 701 is three-dimensionally displayed at an appropriate position according to the position or the operator's standing position.
- an appropriate operation area 821 for the operator 820 is set.
- the operation that is about to be performed on the three-dimensional display 111 is shown in various forms so that the operator can recognize his / her own operation.
- FIG. 22 shows how the icon changes on the screen 2501 of the three-dimensional display 111 as a result of the above operation.
- a television program guide is stereoscopically displayed on a screen 2501 of the three-dimensional display 111, and an operation regarding a certain program is about to be performed.
- the operator tries to select by pushing the finger 601 toward the three-dimensional display 111 as described above.
- an icon 2503 is displayed on the screen 2501. Since the finger is still far from this icon, a relatively large icon on the right side of the icon shown in FIG.
- the icon approaches the target selection item “setting change”, becomes smaller and becomes a special icon when the icon 2502 has a certain size, and the finger crosses the virtual operation surface. Then, it is determined that the item at the indicated position has been selected.
- the TV program guide is stereoscopically displayed at the position of the virtual operation surface for the operator, but the icon does not necessarily have to be displayed at the same position and can be displayed in two dimensions, and can be displayed at an appropriate position by 3D display. Can be arranged.
- the icons displayed on the screen are circular and change in size according to the operation of the operator.
- the present invention is not limited to this, and various forms of icons can be used as shown in FIG. it can. That is, referring to FIG. 23, (1) is an icon in the form of a finger, which is made smaller as it approaches the virtual operation surface as in the example of FIG. (2) shows a circular shape that gradually becomes smaller, but changes to a special shape when input or selection is confirmed, and is confirmed. In the case of this icon and other icons, the color of the icon can be changed instead of or in combination with the change in shape (including size and angle with respect to the display).
- (3) is a shape like X, and when it is far away, it is not only large but also blurred, and as it gets closer, the size of the icon becomes smaller and the blur disappears and becomes a sharp shape.
- the size of the entire icon is not changed, and the figure drawn therein is recognized to change its shape and become focused. In this case, the color of the figure can also be changed.
- (5) shown in FIG. 23 also changes the shape. In FIG. 23, the shape and color of the icon change according to the movement of the finger.
- the shape and color are changed or flashed as shown in a column 2601 at that moment. It is also possible to make the operator recognize that it has been determined as an operation. In addition, although not shown in the figure, as other icon changes, a change that is transparent at first and becomes opaque as the finger approaches the virtual operation surface is also effective. Although not shown, it is possible to create an image that makes the icon appear to pop out of the screen when focusing using the stereoscopic display function of the three-dimensional display, and to inform the operator of the focus state.
- an icon is displayed at the position of the virtual operation surface to confirm the operation determination status, and the color and shape are changed according to the operation of the operator.
- the position is fixed, it is determined according to which item button of the three-dimensional menu is closest to the position indicated by the finger, even if the icon is not displayed.
- the position of the virtual operation surface can be more easily understood and the operation input can be facilitated by changing the fill color or the fill density of the item button indicated in accordance with the distance from.
- FIG. 26 A similar menu selection example is shown in FIG. 26, and will be described.
- the menu 4301 is displayed on the screen.
- a large icon 2610 is displayed on, for example, the item 4302 of the menu shown in FIG.
- the selection of the item 4302 is confirmed and a small icon 2611 is displayed to notify this.
- the selected item in the menu moves, and when the desired item is stopped for a certain period of time, processing corresponding to the selected item can be performed. Further, if the finger 601 is moved backward from a certain area in front of the virtual operation surface 701 before selection is performed, the menu may be deleted.
- the menu is stereoscopically displayed at the position of the virtual operation surface, but here is an example of video image control.
- the menu can be operated by the large icon 3110 and the small icon 3111 as in the example shown in FIG.
- FIG. 25 is a diagram illustrating an example display screen for inputting an instruction to move a graphic displayed on the screen according to the present embodiment.
- the display screen is three-dimensionally displayed at the position of the virtual operation surface, and an instruction is given by touching and moving the operator's hand or finger against the virtual operation surface.
- the icon is reduced from the icon 4201 on the screen 4211 to the icon 4202 on the screen 4212 to indicate that the user is approaching the virtual operation surface.
- the icon 4203 on the screen 4213 is changed and left to stand, and when the finger is moved upward in that state, the rubber band 4204 or the like on the screen 4214 is displayed and the moving direction is displayed. By indicating, the operator can confirm his / her own operation. Further, when the finger is moved in the right direction, the rubber band 4205 of the screen 4215 can be displayed.
- a rubber band (arrow in the above figure) that expands and contracts according to the drag distance of the up / down / left / right after the finger or the like arrives on the virtual operation surface (the position of the icon 4203 is fixed until it comes off the virtual operation surface),
- the moving direction in the 3D space can be changed according to the moving speed and the extending angle according to the expansion / contraction distance (the tip of the arrow follows the movement of the arm tip or fingertip).
- the principle of this embodiment has been described in the case where the operator and the monitor are almost at the same height as shown in FIG. 20, that is, the virtual operation surface is formed substantially perpendicular to the front surface in the horizontal direction of the operator.
- this principle is not affected by the positional relationship and shape between the operator and the display, and various arrangements and configurations are possible.
- the present invention can be applied to a system arrangement as shown in FIGS.
- the three-dimensional camera 201 is also tilted together with the three-dimensional display 111, there is basically no significant difference from the case where the three-dimensional camera 201 is arranged at the horizontal position described above.
- By performing position correction or the like by any method known in the technical field, it is possible to determine the operation by calculating the positional relationship between the operator's part and the virtual operation surface.
- the virtual operation surface itself is set at the operator's will before the operation input processing based on the virtual operation surface described in the first embodiment, so only the setting processing is performed.
- the operation input process can be executed in combination with any function of the first embodiment.
- This embodiment can basically be used in common with the first and second embodiments described above, and therefore the processing such as the system configuration and the setting of the virtual operation surface is the same, but the operation determination processing Is an improvement. That is, in the present embodiment, the processing is basically the same as in the first and second embodiments, but the direction in which the finger is pushed in is particularly small for icons such as when selecting an icon using a virtual operation surface. If you make a mistake, you may not be able to specify the icon properly, or you may specify another icon, so you can set a grid around the icon and make it easy to specify the desired icon It is.
- the operator designates an icon at a slightly wrong position and direction, for example, in the case of the example shown in FIG. An icon can be selected.
- FIG. 38 is a diagram illustrating an example of a target screen used for the grid suction process according to the present embodiment
- FIG. 39 is a diagram illustrating a setting example of a grid used for the grid suction process according to the present embodiment.
- FIG. 39 for example, by setting a grid 4001 to be larger than each icon and surrounding the icon 3902, even when the operator cannot accurately press the icon 3902 by trying to point the icon 3902 with a finger. If the finger enters the range of the grid 4001 that surrounds the icon 3902, the grid can be adsorbed and it can be determined that the icon 3902 has been pointed. Furthermore, it is possible to determine that the icon closest to the position pointed to when pointing between the icons is pointed by adjusting the division position of the grid.
- the focus is attracted from the periphery of a certain area toward the center of the target (icon or link) by means of a dedicated browser or content creation tool, etc., with respect to the center position of the icon or link on the operation target display. .
- the selection operation can be facilitated, and even if the virtual operation surface is inclined, there is no problem in the plane pointing operation of the operator.
- the suction area (area) arbitrarily defined on the XY plane is expanded / enlarged with respect to the Z-axis depth direction ( Grid adsorption can be made more effective by making the area corresponding to the angle of the “flow” due to the difference in camera and display mounting angle.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Multimedia (AREA)
- Human Computer Interaction (AREA)
- General Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Social Psychology (AREA)
- Health & Medical Sciences (AREA)
- Computer Vision & Pattern Recognition (AREA)
- General Health & Medical Sciences (AREA)
- Psychiatry (AREA)
- Optics & Photonics (AREA)
- Signal Processing (AREA)
- User Interface Of Digital Computer (AREA)
- Position Input By Displaying (AREA)
- Processing Or Creating Images (AREA)
Abstract
Description
図1は、本実施形態の操作入力システムの一例を示す図である。本実施形態の立体表示三次元ディスプレイ111は、操作者102の前面に配置され、操作者102は、三次元ディスプレイ111との間の一定の位置に立体表示された仮想的な操作面に対し、指などの形状が操作判定の対象になることを意識して操作を行うことができる。三次元ディスプレイ111には、本システムが目的とする様々なアプリケーション用の種々の映像が表示されるが、これに加え後述するように操作入力を支援、すなわち例えば対象となる操作者102の部位などを画面の隅の方に表示して、操作者102に、現時点で操作として判定されうる動作を認識させることができる。
また、図7及び8は、本実施形態の使用例であるテレビ表示装置における仮想的な操作面の様子を示す図であるが、図7及び8に示すように複数の視聴者に対し各々仮想操作面を生成して、個別に操作することもできるし、いずれかの視聴者の指示を優先的に受付けるようにするなど、様々な設定を行うことができる。しかし、通常は複数の視聴者の各々異なる操作による混乱を回避するため、何らかの方法で1人の視聴者を決定してその視聴者だけに操作入力を認めるようにする。ここで、図7または8に示すような複数の視聴者は、装置に対する操作の権限を有する操作者となる候補者であることから、操作候補者と呼び、以下複数の操作候補者がいる場合の操作者の決定、操作者決定後の処理について説明する。したがって、操作候補者には、図7または8に示すような視聴者を含め、本発明の対象である操作者の候補者、例えばゲーム機の複数のプレイヤや、プレゼンテーションの発表者などが含まれる。なお、本実施形態のシステムでは、操作者の対象となる者はシステムが対象として認識できるものに限られるので、操作候補者は、例えばビデオカメラ201の撮像範囲に存在するものに限られる。
本実施形態では、図10に示すように、ビデオカメラ201で撮影された操作者102の像に基づいて仮想操作面を立体画像として形成し、同じく撮影した操作者102の一部である手や指の位置を定めるとともに、仮想的な操作面701と操作者102の手指601との位置関係を算出する処理を行う。本実施形態では、このような処理を行う前提として、本技術分野で知られた初期設定、例えば本実施形態の画像認識装置が新たに設置された場合を想定すると、事前準備として利用するビデオカメラ201の利用レンズの歪み、三次元ディスプレイ111とレンズとの距離等の情報を装置に入力しておく必要がある。さらに閾値設定等を予め調整しておく。システムの初期設定が終了すると、本実施形態の処理を行うこととなるが、この処理については図4を参照して以下に説明する。
以上説明したように、三次元ビデオカメラにより仮想操作面を立体表示することにより、操作者は空間上にあたかもタッチパネルのような操作面が存在するように認識することができ、この操作面に対し、種々の操作を行うことにより、身体の全部または一部を用いた操作入力が可能となるが、さらに仮想的な操作面に対する操作者の映像を三次元ディスプレイ111に表示する等、操作入力を支援することにより、より容易に本実施形態のシステムを活用することができる。
本実施形態では、操作者が空間上に仮想的に立体表示された仮想操作面を基準に、そこにあたかもタッチパネルのような入力機器が存在するかのように操作することにより、その操作内容を確実に判定しようとするものであるが、このようにして判定された操作の内容を仮想操作面から奥側方向である、操作者から離れる方向への仮想操作面と操作者の手等の身体の一部あるいは身につけた物体の位置関係で決定する。例えば、操作者から離れる方向であるz軸方向に2層あるいは3層に操作領域を仮想操作階層として設定し、操作者の手がどの層に入っているかにより操作の種別を決定して、その層内での手の動きにより操作内容を決定する。この際、操作者が視認している表示画面上に手の位置や操作の種別などを表示させれば、より操作者は操作の認識を容易にすることができる。なお、操作者の一部と各階層を分割する面とのz方向の距離は、上述した形成された仮想操作面と操作者の一部との距離を算出する手法によって取得することができる。
本実施形態は、上述の第1実施形態で立体表示した仮想操作面を予め操作者に指定させることにより設定する点で異なるが、第1実施形態のシステム構成および仮想操作面に対して操作者が行った動作をどのように操作として認識するか等の処理は同様である。すなわち、本実施形態では、後述するように操作者が操作を行う場合は先ず、仮想操作面をどこに設定するか一定の操作で指示させ、その指示に従って仮想操作面を形成する。したがって、第1実施形態のシステムのように三次元ディスプレイにより仮想操作面を立体表示させなくても操作者はどこにどのような操作面が存在しているかを予め認識することができるので、仮想操作面を必ずしも立体表示する必要はなく、通常の二次元ディスプレイを使用することもできる。ただし、仮想操作面の認識をより確実にするため、三次元ディスプレイを用いて立体表示させることもできる。以上の通り、本実施形態は、仮想操作面を操作者の指示に応じて形成した後の処理は基本的に第1実施形態と同様なので、以下おもに仮想操作面の設定処理について説明する。
図32は、本実施形態にかかる操作者のジェスチャにより操作面の位置を設定する領域を判定するため撮影した画像を示す図である。図32に示す例のように、三次元カメラ201で撮影された操作者の画像3210の手3202及び3203により操作面の形状(大きさ、ディスプレイに対する角度を含む)が決定され、z方向の距離データ画像3211のディスプレイから手3202及び3203までの距離、ディスプレイから胴3204までの距離により仮想操作面3201の位置を決定する。具体的には、図33に示すように三次元カメラ201により撮影された操作者の画像データから奥行きを把握することができる。図33は、本実施形態にかかる指定された操作面の位置の抽出を説明するための図であり、操作者の像を奥行き方向(z方向)で示したものである。図33を参照すると、操作者が手3202で示した位置に操作面3201を形成すれば良いことが理解される。
本実施形態では、操作者が空間上に仮想的に立体表示された仮想操作面を基準に、そこにあたかもタッチパネルのような入力機器が存在するかのように操作することにより、その操作内容を確実に判定しようとするものであるが、特に上述の第1実施形態のように仮想操作面を立体表示しない場合、操作者の一部である手または指などが仮想操作面に至るまで、つまり操作者が何らかの操作を実行しようと手または指を動かし始めてから、仮想操作面を押下するまでの間も操作支援することにより、さらに操作入力を容易に、より高精度に行うようにすることができる。
本実施形態は、基本的に上述の第1および第2実施形態で共通に用いることができるものであり、従ってシステム構成や仮想操作面の設定などの処理は同様であるが、操作の判定処理を改良したものである。すなわち、本実施形態では、第1および2実施形態とその処理は基本的に同じであるが、例えば仮想操作面を使用してアイコンを選択するときなど、特に小さいアイコンの場合指の押し込み方向を誤ると、適切にアイコンを指定できなかったり、別のアイコンを指定してしまったりするおそれがあるため、アイコンの周辺にグリッドを設定してグリッド吸着させ、所望のアイコンを指定しやすくするというものである。本発明のように仮想的に操作面を設定して操作の基準とする場合、第1及び2実施形態のように操作面の認識を確実にする様々な処理を盛り込んだとしても、操作者は常に正確な位置指定ができるわけではない。本実施形態では、操作者が多少誤った位置や方向でアイコンを指定、例えば図15に示す例のような場合に、操作面の傾きや位置などの調整がされていないときでも的確に所望のアイコンを選択できるようにすることができる。
Claims (19)
- 操作者の像を読取って立体画像データを生成する三次元撮像手段と、
前記三次元撮像手段により読取られた操作者の像及び位置に基づいた形状及び位置で仮想操作面を形成する操作面形成手段と、
前記形成された仮想操作面に対する操作者の少なくとも一部の像の動きを、前記三次元撮像手段で読取って、前記操作者の一部と前記仮想操作面との位置関係に基づいて該動きが操作であるか否かを判定する操作判定手段と、
前記動きが操作であると判定されると、所定の信号を出力する信号出力手段と
を備えたことを特徴とする画像認識装置。 - 前記三次元撮像手段が複数の操作候補者を読み取った場合、予め定められた特定の動きをする操作候補者を前記操作者と判定することを特徴とする請求項1に記載の画像認識装置。
- 前記三次元撮像手段が読み取った複数の操作候補者の画像の示す、前記操作者と判定された操作候補者が現在の操作者と、他の操作候補者との位置関係を表示する操作者表示手段をさらに備えたことを特徴とする請求項1または2に記載の画像認識装置。
- 前記仮想操作面の操作者側の所定の位置に配置され、該仮想操作面上での操作位置を決定するための位置決定面をさらに備え、
前記操作判定手段は前記操作者の一部が前記位置決定面から前記仮想操作面へ移動する一連の動作において前記仮想操作面を横切る際、前記位置決定面上の前記操作者の一部が横切った位置に対応する前記仮想操作面上の位置を、該操作者の一部が横切ったものとして前記操作者の操作を判定することを特徴とする請求項1ないし3のいずれかに記載の画像認識装置。 - 前記仮想操作面の形状および位置は、前記読取られた操作者の像から腕および顔の位置を抽出し、当該抽出した腕および顔の位置に基づいて決定することを特徴とする請求項1ないし4のいずれかに記載の画像認識装置。
- 前記仮想操作面の位置は、前記操作者と前記三次元撮像手段との間にあることを特徴とする請求項1ないし5のいずれかに記載の画像認識装置。
- 前記操作面形成手段は、前記三次元撮像手段により読取られた操作者の像、及び前記操作者の予め定められた特定の動きに基づいて定められる形状および位置で、仮想操作面を形成することを特徴とする請求項1ないし6のいずれかに記載の画像認識装置。
- 前記操作面形成手段は、前記三次元撮像手段により読取られた操作者の像、及び前記操作者の予め定められた特定の動きに基づいて定められる角度で、仮想操作面を形成することを特徴とする請求項1ないし7のいずれかに記載の画像認識装置。
- 前記操作面形成手段は、前記三次元撮像手段により読取られた操作者の像及び位置に基づいた定められる角度で、仮想操作面を形成することを特徴とする請求項1ないし7のいずれかに記載の画像認識装置。
- 前記形成された仮想操作面を示す立体画像を表示する操作面立体表示手段をさらに備えたことを特徴とする請求項1ないし9のいずれかに記載の画像認識装置。
- 前記立体画像は、左右眼視差によることを特徴とする請求項10に記載の画像認識装置。
- 前記操作者の一部と前記仮想操作面との位置関係から当該距離を算出して該距離に応じた位置に、前記操作者の一部の前記仮想操作面に対する位置関係を示す所定の標示を立体表示することを特徴とする請求項10または11に記載の画像認識装置。
- 操作者の像を読取って立体画像データを生成する三次元撮像手段と、
前記三次元撮像手段により読取られた操作者の予め定められた特定の動きに基づいた形状及び位置で仮想操作面を形成する操作面形成手段と、
前記形成された仮想操作面に対する操作者の少なくとも一部の像の動きを、前記三次元撮像手段で読取って、前記操作者の一部と前記仮想操作面との位置関係に基づいて該動きが操作であるか否かを判定する操作判定手段と、
前記動きが操作であると判定されると、所定の信号を出力する信号出力手段と
を備えたことを特徴とする画像認識装置。 - 操作者の像を読取って立体画像データを生成する三次元撮像ステップと、
前記三次元撮像ステップにおいて読取られた操作者の像及び位置に基づいた形状及び位置で仮想操作面を形成する操作面形成ステップと、
前記形成された仮想操作面に対する操作者の少なくとも一部の像の動きを、前記三次元撮像手段で読取って、前記操作者の一部と前記仮想操作面との位置関係に基づいて該動きが操作であるか否かを判定する操作判定ステップと、
前記動きが操作であると判定されると、所定の信号を出力する信号出力ステップと
を備えたことを特徴とする操作判定方法。 - 前記操作判定ステップは、前記操作者の一部が前記仮想操作面の操作者側の所定の位置に配置され、該仮想操作面上での操作位置を決定するための位置決定面から前記仮想操作面へ移動する一連の動作において前記仮想操作面を横切る際、前記位置決定面上の前記操作者の一部が横切った位置に対応する前記仮想操作面上の位置を横切ったものとして前記操作者の操作を判定することを特徴とする請求項14に記載の操作判定方法。
- 前記形成された仮想操作面を示す立体画像を表示する操作面立体表示ステップをさらに備えたことを特徴とする請求項14または15に記載の操作判定方法。
- 画像認識装置に、操作者の画像を認識して操作内容を判定する操作判定方法を実行させるプログラムであって、該操作判定方法は、
操作者の像を読取って立体画像データを生成する三次元撮像ステップと、
前記三次元撮像ステップにおいて読取られた操作者の像及び位置に基づいた形状及び位置で仮想操作面を形成する操作面形成ステップと、
前記形成された仮想操作面に対する操作者の少なくとも一部の像の動きを、前記三次元撮像手段で読取って、前記操作者の一部と前記仮想操作面との位置関係に基づいて該動きが操作であるか否かを判定する操作判定ステップと、
前記動きが操作であると判定されると、所定の信号を出力する信号出力ステップと
を備えたことを特徴とするプログラム。 - 前記操作判定ステップは、前記操作者の一部が前記仮想操作面の操作者側の所定の位置に配置され、該仮想操作面上での操作位置を決定するための位置決定面から前記仮想操作面へ移動する一連の動作において前記仮想操作面を横切る際、前記位置決定面上の前記操作者の一部が横切った位置に対応する前記仮想操作面上の位置を横切ったものとして前記操作者の操作を判定することを特徴とする請求項17に記載のプログラム。
- 前記形成された仮想操作面を示す立体画像を表示する操作面立体表示ステップをさらに備えたことを特徴とする請求項17または18に記載のプログラム。
Priority Applications (7)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201180042578.8A CN103124945B (zh) | 2010-09-06 | 2011-04-11 | 图像识别装置和操作判断方法以及程序 |
KR1020177009386A KR20170040389A (ko) | 2010-09-06 | 2011-04-11 | 화상인식장치 및 조작판정방법, 그리고 프로그램 |
US13/145,030 US20120056989A1 (en) | 2010-09-06 | 2011-04-11 | Image recognition apparatus, operation determining method and program |
EP11823172.9A EP2615523A4 (en) | 2010-09-06 | 2011-04-11 | PICTURE DETECTION DEVICE, OPERATING ASSESSMENT METHOD THEREFOR AND PROGRAM |
KR1020137005743A KR101541803B1 (ko) | 2010-09-06 | 2011-04-11 | 화상인식장치 및 조작판정방법, 그리고 프로그램 |
CA2810307A CA2810307A1 (en) | 2010-09-06 | 2011-04-11 | Image recognition apparatus, operation determining method and computer readable medium |
KR1020147030384A KR20140140109A (ko) | 2010-09-06 | 2011-04-11 | 화상인식장치 및 조작판정방법, 그리고 프로그램 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2010-199368 | 2010-09-06 | ||
JP2010199368A JP4900741B2 (ja) | 2010-01-29 | 2010-09-06 | 画像認識装置および操作判定方法並びにプログラム |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2012032687A1 true WO2012032687A1 (ja) | 2012-03-15 |
Family
ID=45811215
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2011/002136 WO2012032687A1 (ja) | 2010-09-06 | 2011-04-11 | 画像認識装置および操作判定方法並びにプログラム |
Country Status (6)
Country | Link |
---|---|
EP (1) | EP2615523A4 (ja) |
JP (1) | JP4900741B2 (ja) |
KR (3) | KR101541803B1 (ja) |
CN (1) | CN103124945B (ja) |
CA (1) | CA2810307A1 (ja) |
WO (1) | WO2012032687A1 (ja) |
Cited By (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP5174979B1 (ja) * | 2012-04-26 | 2013-04-03 | 株式会社三菱東京Ufj銀行 | 情報処理装置、電子機器およびプログラム |
JP2013200784A (ja) * | 2012-03-26 | 2013-10-03 | Yahoo Japan Corp | 操作入力装置、操作入力方法およびプログラム |
JP2013200815A (ja) * | 2012-03-26 | 2013-10-03 | Yahoo Japan Corp | 操作入力装置、操作入力方法、及び、プログラム |
EP2662705A1 (en) * | 2012-05-07 | 2013-11-13 | Hexagon Technology Center GmbH | Surveying apparatus having a range camera |
JP2016118947A (ja) * | 2014-12-22 | 2016-06-30 | 日本ユニシス株式会社 | 角度調整可能な仮想平面を用いた空間手書き入力システム |
US20190286280A1 (en) * | 2014-12-26 | 2019-09-19 | Nikon Corporation | Detection device, electronic apparatus, detection method and program |
US10928919B2 (en) | 2016-03-29 | 2021-02-23 | Sony Corporation | Information processing device and information processing method for virtual objects operability |
WO2023188774A1 (ja) * | 2022-03-31 | 2023-10-05 | パナソニックIpマネジメント株式会社 | 情報処理装置、情報処理システム、および情報処理方法 |
Families Citing this family (46)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8181123B2 (en) | 2009-05-01 | 2012-05-15 | Microsoft Corporation | Managing virtual port associations to users in a gesture-based computing environment |
DE102011112620B3 (de) * | 2011-09-08 | 2013-02-21 | Eads Deutschland Gmbh | Abgewinkeltes Display zur dreidimensionalen Darstellung eines Szenarios |
DE102011112619A1 (de) | 2011-09-08 | 2013-03-14 | Eads Deutschland Gmbh | Auswahl von Objekten in einem dreidimensionalen virtuellen Szenario |
JPWO2013046596A1 (ja) * | 2011-09-26 | 2015-03-26 | Necカシオモバイルコミュニケーションズ株式会社 | 携帯型情報処理端末 |
RU2658790C2 (ru) | 2011-11-21 | 2018-06-22 | Никон Корпорейшн | Устройство отображения и программа управления отображением |
CN102662577B (zh) * | 2012-03-29 | 2016-08-10 | 华为终端有限公司 | 一种基于三维显示的光标操作方法及移动终端 |
JP5735453B2 (ja) * | 2012-05-16 | 2015-06-17 | ヤフー株式会社 | 表示制御装置、表示制御方法、情報表示システム、およびプログラム。 |
JP2013250882A (ja) | 2012-06-01 | 2013-12-12 | Sharp Corp | 注目位置検出装置、注目位置検出方法、及び注目位置検出プログラム |
JP2014056462A (ja) * | 2012-09-13 | 2014-03-27 | Toshiba Alpine Automotive Technology Corp | 操作装置 |
WO2014073384A1 (ja) * | 2012-11-06 | 2014-05-15 | 株式会社ソニー・コンピュータエンタテインメント | 情報処理装置 |
EP2951811A4 (en) | 2013-01-03 | 2016-08-17 | Meta Co | DIGITAL EYES FOR EXTRAMISSIVE SPATIAL IMAGING, PROCESSES AND SYSTEMS FOR VIRTUAL OR ENLARGED VISIBILITY, MANIPULATION, PRODUCTION OR INTERACTION WITH OBJECTS, MATERIALS, OR OTHER UNITS |
US9979946B2 (en) * | 2013-02-19 | 2018-05-22 | Mirama Service Inc | I/O device, I/O program, and I/O method |
WO2014128748A1 (ja) * | 2013-02-19 | 2014-08-28 | 株式会社ブリリアントサービス | キャリブレーション装置、キャリブレーションプログラム、およびキャリブレーション方法 |
US10171800B2 (en) * | 2013-02-19 | 2019-01-01 | Mirama Service Inc. | Input/output device, input/output program, and input/output method that provide visual recognition of object to add a sense of distance |
US20140282275A1 (en) * | 2013-03-15 | 2014-09-18 | Qualcomm Incorporated | Detection of a zooming gesture |
JP6266229B2 (ja) * | 2013-05-14 | 2018-01-24 | 東芝メディカルシステムズ株式会社 | 画像処理装置、方法、及びプログラム |
US9063578B2 (en) * | 2013-07-31 | 2015-06-23 | Microsoft Technology Licensing, Llc | Ergonomic physical interaction zone cursor mapping |
JP6524589B2 (ja) * | 2013-08-30 | 2019-06-05 | 国立大学法人山梨大学 | クリック動作検出装置,方法およびプログラム |
TW201528049A (zh) * | 2014-01-03 | 2015-07-16 | Utechzone Co Ltd | 基於圖樣的校正方法及電子裝置 |
WO2015159768A1 (ja) * | 2014-04-15 | 2015-10-22 | シャープ株式会社 | 入力装置 |
US20170168584A1 (en) * | 2014-05-08 | 2017-06-15 | Nec Solution Innovators, Ltd. | Operation screen display device, operation screen display method, and non-temporary recording medium |
US9841844B2 (en) * | 2014-06-20 | 2017-12-12 | Funai Electric Co., Ltd. | Image display device |
CN104065949B (zh) * | 2014-06-26 | 2016-08-10 | 深圳奥比中光科技有限公司 | 一种电视虚拟触控方法及系统 |
JP2016110177A (ja) * | 2014-12-02 | 2016-06-20 | 株式会社トプコン | 3次元入力装置及び入力システム |
KR101844883B1 (ko) | 2014-12-23 | 2018-04-03 | 메타 컴퍼니 | 관심 객체의 임의의 깊이에 있는 동일한 평면으로 시각적 조절과 시각적 수렴을 결합시키는 장치, 방법 및 시스템 |
WO2016103521A1 (ja) | 2014-12-26 | 2016-06-30 | 株式会社ニコン | 検出装置およびプログラム |
KR20170105404A (ko) | 2014-12-26 | 2017-09-19 | 가부시키가이샤 니콘 | 제어 장치, 전자 기기, 제어 방법 및 프로그램 |
US9836130B2 (en) | 2014-12-26 | 2017-12-05 | Nextedge Technology K.K. | Operation input device, operation input method, and program |
JP2018032055A (ja) * | 2015-01-13 | 2018-03-01 | 三菱電機株式会社 | ジェスチャー認識装置及び方法、並びにプログラム及び記録媒体 |
JP6354653B2 (ja) * | 2015-04-25 | 2018-07-11 | 京セラドキュメントソリューションズ株式会社 | 拡張現実操作システムおよび拡張現実操作プログラム |
KR101708455B1 (ko) | 2015-09-08 | 2017-02-21 | 엠더블유엔테크 주식회사 | 핸드 플로트 입체 메뉴 시스템 |
US10665020B2 (en) | 2016-02-15 | 2020-05-26 | Meta View, Inc. | Apparatuses, methods and systems for tethering 3-D virtual elements to digital content |
JP6514416B2 (ja) * | 2016-06-15 | 2019-05-15 | ナーブ株式会社 | 画像表示装置、画像表示方法、及び画像表示プログラム |
US11635827B2 (en) | 2016-06-28 | 2023-04-25 | Nikon Corporation | Control device, display device, program, and detection method |
JP6822472B2 (ja) * | 2016-06-28 | 2021-01-27 | 株式会社ニコン | 表示装置、プログラム、表示方法および制御装置 |
JP6444345B2 (ja) * | 2016-08-23 | 2018-12-26 | 株式会社コロプラ | 仮想空間における入力を支援するための方法および装置ならびに当該方法をコンピュータに実行させるプログラム |
CN107977071B (zh) * | 2016-10-24 | 2020-02-28 | 中国移动通信有限公司研究院 | 一种适用于空间系统的操作方法及装置 |
CN106774938A (zh) * | 2017-01-16 | 2017-05-31 | 广州弥德科技有限公司 | 基于体感设备的人机交互集成装置 |
WO2018158813A1 (ja) * | 2017-02-28 | 2018-09-07 | 三菱電機株式会社 | エレベーターの呼び登録システム |
JP6728111B2 (ja) * | 2017-08-01 | 2020-07-22 | 株式会社コロプラ | 仮想空間を提供する方法、仮想体験を提供する方法、プログラム、および記録媒体 |
WO2019054037A1 (ja) * | 2017-09-12 | 2019-03-21 | ソニー株式会社 | 情報処理装置、情報処理方法、およびプログラム |
JP7119398B2 (ja) * | 2018-02-06 | 2022-08-17 | 富士フイルムビジネスイノベーション株式会社 | 情報処理装置及びプログラム |
JP6658809B2 (ja) * | 2018-07-04 | 2020-03-04 | 株式会社ニコン | 検出装置、電子機器、検出方法およびプログラム |
US12017791B2 (en) | 2020-09-03 | 2024-06-25 | Rockwell Collins, Inc. | System and method for interpreting gestures and providing control signals |
KR102437979B1 (ko) * | 2022-02-22 | 2022-08-30 | 주식회사 마인드포지 | 제스처 기반의 객체 지향적 인터페이싱 방법 및 장치 |
US11531448B1 (en) | 2022-06-01 | 2022-12-20 | VR-EDU, Inc. | Hand control interfaces and methods in virtual reality environments |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH0612177A (ja) * | 1992-06-29 | 1994-01-21 | Canon Inc | 情報入力方法及びその装置 |
JP2010015553A (ja) * | 2008-06-03 | 2010-01-21 | Shimane Pref Gov | 画像認識装置および操作判定方法並びにプログラム |
JP2010134629A (ja) * | 2008-12-03 | 2010-06-17 | Sony Corp | 情報処理装置および情報処理方法 |
Family Cites Families (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH10207620A (ja) * | 1997-01-28 | 1998-08-07 | Atr Chinou Eizo Tsushin Kenkyusho:Kk | 立体インタラクション装置および立体インタラクション方法 |
JP2004303251A (ja) * | 1997-11-27 | 2004-10-28 | Matsushita Electric Ind Co Ltd | 制御方法 |
US7227526B2 (en) * | 2000-07-24 | 2007-06-05 | Gesturetek, Inc. | Video-based image control system |
JP4286556B2 (ja) * | 2003-02-24 | 2009-07-01 | 株式会社東芝 | 画像表示装置 |
JP2004334590A (ja) * | 2003-05-08 | 2004-11-25 | Denso Corp | 操作入力装置 |
JP3847753B2 (ja) * | 2004-01-30 | 2006-11-22 | 株式会社ソニー・コンピュータエンタテインメント | 画像処理装置、画像処理方法、記録媒体、コンピュータプログラム、半導体デバイス |
JP2006235771A (ja) * | 2005-02-23 | 2006-09-07 | Victor Co Of Japan Ltd | 遠隔操作装置 |
US9910497B2 (en) * | 2006-02-08 | 2018-03-06 | Oblong Industries, Inc. | Gestural control of autonomous and semi-autonomous systems |
-
2010
- 2010-09-06 JP JP2010199368A patent/JP4900741B2/ja not_active Expired - Fee Related
-
2011
- 2011-04-11 KR KR1020137005743A patent/KR101541803B1/ko active IP Right Grant
- 2011-04-11 KR KR1020147030384A patent/KR20140140109A/ko active Application Filing
- 2011-04-11 WO PCT/JP2011/002136 patent/WO2012032687A1/ja active Application Filing
- 2011-04-11 EP EP11823172.9A patent/EP2615523A4/en not_active Withdrawn
- 2011-04-11 CA CA2810307A patent/CA2810307A1/en not_active Abandoned
- 2011-04-11 KR KR1020177009386A patent/KR20170040389A/ko active Search and Examination
- 2011-04-11 CN CN201180042578.8A patent/CN103124945B/zh not_active Expired - Fee Related
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH0612177A (ja) * | 1992-06-29 | 1994-01-21 | Canon Inc | 情報入力方法及びその装置 |
JP2010015553A (ja) * | 2008-06-03 | 2010-01-21 | Shimane Pref Gov | 画像認識装置および操作判定方法並びにプログラム |
JP2010134629A (ja) * | 2008-12-03 | 2010-06-17 | Sony Corp | 情報処理装置および情報処理方法 |
Non-Patent Citations (1)
Title |
---|
See also references of EP2615523A4 * |
Cited By (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2013200784A (ja) * | 2012-03-26 | 2013-10-03 | Yahoo Japan Corp | 操作入力装置、操作入力方法およびプログラム |
JP2013200815A (ja) * | 2012-03-26 | 2013-10-03 | Yahoo Japan Corp | 操作入力装置、操作入力方法、及び、プログラム |
JP2013242845A (ja) * | 2012-04-26 | 2013-12-05 | Bank Of Tokyo-Mitsubishi Ufj Ltd | 情報処理装置、電子機器およびプログラム |
JP5174979B1 (ja) * | 2012-04-26 | 2013-04-03 | 株式会社三菱東京Ufj銀行 | 情報処理装置、電子機器およびプログラム |
CN104204848A (zh) * | 2012-05-07 | 2014-12-10 | 赫克斯冈技术中心 | 具有测距相机的勘测设备 |
WO2013167472A1 (en) * | 2012-05-07 | 2013-11-14 | Hexagon Technology Center Gmbh | Surveying apparatus having a range camera |
EP2662705A1 (en) * | 2012-05-07 | 2013-11-13 | Hexagon Technology Center GmbH | Surveying apparatus having a range camera |
US9683833B2 (en) | 2012-05-07 | 2017-06-20 | Hexagon Technology Center Gmbh | Surveying apparatus having a range camera |
JP2016118947A (ja) * | 2014-12-22 | 2016-06-30 | 日本ユニシス株式会社 | 角度調整可能な仮想平面を用いた空間手書き入力システム |
US20190286280A1 (en) * | 2014-12-26 | 2019-09-19 | Nikon Corporation | Detection device, electronic apparatus, detection method and program |
US10860139B2 (en) * | 2014-12-26 | 2020-12-08 | Nikon Corporation | Detection device, electronic apparatus, detection method and program |
US10928919B2 (en) | 2016-03-29 | 2021-02-23 | Sony Corporation | Information processing device and information processing method for virtual objects operability |
WO2023188774A1 (ja) * | 2022-03-31 | 2023-10-05 | パナソニックIpマネジメント株式会社 | 情報処理装置、情報処理システム、および情報処理方法 |
Also Published As
Publication number | Publication date |
---|---|
JP4900741B2 (ja) | 2012-03-21 |
EP2615523A1 (en) | 2013-07-17 |
KR101541803B1 (ko) | 2015-08-04 |
KR20170040389A (ko) | 2017-04-12 |
EP2615523A4 (en) | 2014-05-14 |
KR20140140109A (ko) | 2014-12-08 |
CA2810307A1 (en) | 2012-03-15 |
CN103124945B (zh) | 2016-04-13 |
CN103124945A (zh) | 2013-05-29 |
KR20130049199A (ko) | 2013-05-13 |
JP2011175617A (ja) | 2011-09-08 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP4900741B2 (ja) | 画像認識装置および操作判定方法並びにプログラム | |
JP5114795B2 (ja) | 画像認識装置および操作判定方法並びにプログラム | |
US20120056989A1 (en) | Image recognition apparatus, operation determining method and program | |
JP7283506B2 (ja) | 情報処理装置、情報処理方法、及び情報処理プログラム | |
JP4701424B2 (ja) | 画像認識装置および操作判定方法並びにプログラム | |
JP4725595B2 (ja) | 映像処理装置、映像処理方法、プログラム及び記録媒体 | |
US9651782B2 (en) | Wearable tracking device | |
US8894486B2 (en) | Handheld information processing apparatus and handheld game apparatus | |
KR101822471B1 (ko) | 혼합현실을 이용한 가상현실 시스템 및 그 구현방법 | |
WO2009148064A1 (ja) | 画像認識装置および操作判定方法並びにプログラム | |
KR20130108643A (ko) | 응시 및 제스처 인터페이스를 위한 시스템들 및 방법들 | |
JP6576536B2 (ja) | 情報処理装置 | |
US20160249043A1 (en) | Three dimensional (3d) glasses, 3d display system and 3d display method | |
JP2012256110A (ja) | 情報処理装置、情報処理方法およびプログラム | |
WO2018074045A1 (ja) | 情報処理装置、情報処理方法、及びプログラム | |
JP4678428B2 (ja) | 仮想空間内位置指示装置 | |
JP2009212582A (ja) | バーチャルスタジオ用フィードバックシステム | |
WO2021220407A1 (ja) | ヘッドマウントディスプレイ装置および表示制御方法 | |
JP5520772B2 (ja) | 立体画像の表示システム及び表示方法 | |
JP2005050120A (ja) | 仮想空間内位置指示装置 | |
CN113170077A (zh) | 显示装置 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 201180042578.8 Country of ref document: CN |
|
WWE | Wipo information: entry into national phase |
Ref document number: 13145030 Country of ref document: US |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 11823172 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2810307 Country of ref document: CA |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2011823172 Country of ref document: EP |
|
ENP | Entry into the national phase |
Ref document number: 20137005743 Country of ref document: KR Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |