US5801704A - Three-dimensional input device with displayed legend and shape-changing cursor - Google Patents

Three-dimensional input device with displayed legend and shape-changing cursor Download PDF

Info

Publication number
US5801704A
US5801704A US08515450 US51545095A US5801704A US 5801704 A US5801704 A US 5801704A US 08515450 US08515450 US 08515450 US 51545095 A US51545095 A US 51545095A US 5801704 A US5801704 A US 5801704A
Authority
US
Grant status
Grant
Patent type
Prior art keywords
image
cursor
means
operating
function
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
US08515450
Inventor
Shunichi Oohara
Masakazu Ejiri
Yasuhiro Nemoto
Naoki Sasaki
Hidefumi Ohtsuka
Shogo Matsumoto
Ryoko Sato
Kazushi Yoshida
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hitachi Ltd
Original Assignee
Hitachi Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Grant date

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/016Input arrangements with force or tactile feedback as computer generated output to the user
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • G06F3/014Hand-worn input/output arrangements, e.g. data gloves
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • G06F3/04812Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance interaction techniques based on cursor appearance or behaviour being affected by the presence of displayed objects, e.g. visual feedback during interaction with elements of a graphical user interface through change in cursor appearance, constraint movement or attraction/repulsion with respect to a displayed object
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2203/00Indexing scheme relating to G06F3/00 - G06F3/048
    • G06F2203/01Indexing scheme relating to G06F3/01
    • G06F2203/014Force feedback applied to GUI

Abstract

In an image processing apparatus for performing image processing of a processed object by displaying the processed object and an operating cursor on an image display device, and by operating the operating cursor according to three-dimensional actions of the hand and the fingers of an operator, a processing function and a shape of the operating cursor are changed by judging the meaning of the three-dimensional actions of the hand and the fingers to support understanding of the processing function.

Description

"THREE-DIMENSIONAL INPUT DEVICE WITH DISPLAYED LEGEND AND SHAPE-CHANGING CURSOR" BACKGROUND OF THE INVENTION

1. Field of the Invention

The present invention relates to an image processing apparatus and a method of image processing which performs image processing and simulation on a processed object produced and displayed through data processing using a computer, and more particularly relates to an image processing apparatus and a method of image processing for processing a processed object by detecting movement of hands and fingers of an operator.

2. Description of the Related Art

Processing such as displacement, modification or deformation of a processed object produced and displayed through data processing using a computer is performed by operation of controlling a cursor using a keyboard or a mouse.

An image processing apparatus and a method of image processing for processing a processed object by detecting movement of hands and fingers of an operator are disclosed in Japanese Patent Application Laid-Open No.62-44827 (1987), where a glove-shaped input device is used to detect the motion of the hand and fingers of an operator, and the processing operation of a processed object is performed by moving a handshaped cursor corresponding to the detected results.

An operation of a three-dimensional image using a glove-shaped input device is described in a paper titled "Network Virtual Reality System", K.Shinohara; PIXEL, No.132, pp 26 to 28 (1993).

In the image processing apparatus in which control of a cursor for operating image processing of a processed object displayed as an image is performed using a keyboard, it is difficult to freely control the image processing since the moving directions of the cursor are limited. In the image processing apparatus in which control of a cursor for operating image processing of a processed object displayed as an image is performed using a mouse, there is a problem of lack of reality in the operation since the operating direction of the mouse is on a horizontal plane, but the moving direction of the cursor is on a vertical plane of the image display. Further, both of the input operations of the keyboard and the mouse are two-dimensional operation, and accordingly it is difficult to operate and control a three-dimensional image processing.

On the other hand, although, in the glove-shaped input device, the input operation can be performed by three-dimensional operation, there is a disadvantage in that it is difficult to clearly display the function of cursor when processing operation is performed to a processed object and an acting point of a cursor to the processed object since the shape of the cursor is a projection of a hand.

SUMMARY OF THE INVENTION

An object of the present invention is to provide an image processing apparatus and a method of image processing capable of accurately performing control operation processing of an image using a cursor.

Another object of the present invention is to provide an image processing apparatus and a method of image processing capable of speedy and easily inputting operating condition for acting the processed object in a computer simulation.

One of the features of the present invention is characterized by an image processing apparatus comprising an image displaying means for displaying a processed object and an operating cursor, a cursor generating means for operating the operating cursor in corresponding to three-dimensional actions of the hand and the fingers of an operator, a processed object generating means for generating image data of the processed object processed according to the action of the operating cursor, which comprises a control changing means for changing the processing function and the shape of the operating cursor by judging the meaning of the three-dimensional actions of the hand and the fingers.

Another feature of the present invention is characterized by an image processing apparatus comprising an image displaying means for displaying an image, a three-dimensional action detecting means for detecting action and the three-dimensional positions of the hand and the fingers of an operator, a cursor generating means for generating operating cursor image data based on the detected result of the three-dimensional action detecting means, processed object generating means for generating image data of a processed object to be operated by an operating cursor displayed based on the operating cursor image data, and a display image generating means for synthetically displaying the operating cursor image data and the processed object image data on the image display means, which comprises a meaning judging means for judging the meaning of an instruction from the form of action of the hand and the fingers detected by the three-dimensional action detecting means, and a control changing means for changing the processing function and the shape of the operating cursor corresponding to the judged result of the meaning judging means.

A further feature of the present invention is characterized by an image processing apparatus comprising an image displaying means for displaying an image, a three-dimensional action detecting means for detecting action and the three-dimensional positions of the hand and the fingers of an operator, a cursor generating means for generating operating cursor image data based on the detected result of the three-dimensional action detecting means, a processed object generating means for generating image data of a processed object to be operated by an operating cursor displayed based on the operating cursor image data, and a display image generating means for synthetically displaying the operating cursor image data and the processed object image data on the image display means, which comprises a meaning judging means for judging the meaning of an instruction from the form of action of the hand and the fingers detected by the three-dimensional action detecting means, a legend displaying means for displaying the processing function and plural kinds of operating cursors corresponding to the result of the judgement of the meaning judging means, a condition setting means for setting a processing operating condition corresponding to each of the operating cursor, and the processed object generating means generating image data of the processed object corresponding to the processing operating condition.

A still further feature of the present invention is characterized by a method of image processing in which image processing of a processed object is performed by displaying the processed object and an operating cursor on an image display means and by operating the operating cursor corresponding to three-dimensional actions of the hand and the fingers of an operator, wherein the operating function and the shape of the operating cursor are changed by judging the meaning of the three-dimensional actions of the hand and the fingers.

According to the present invention, image processing of a processed object is performed by an image processing function and selection and moving operation of an operating cursor corresponding to the processing function through a three-dimensional actions of a hand and fingers.

The shape of the cursor expresses a processing function by symbolic shape of the processing function to support understanding of the processing function.

Further, a simulation image is expressed in a visually displayed state of the condition by setting operating conditions to a plurality of operating cursors and by acting the operating cursors on a processed object.

BRIEF DESCRIPTION OF THE DRAWINGS

FIG.1 is a perspective view of the outer appearance of an image processing apparatus in accordance with the present invention.

FIG.2 is an example of a displayed image on an image display device in an image processing apparatus in accordance with the present invention.

FIG.3 is a view showing a function selecting legend in an image processing apparatus in accordance with the present invention.

FIG.4 is a view showing a cursor operating legend in an image processing apparatus in accordance with the present invention.

FIG.5 is a block diagram showing the function of an image processing apparatus in accordance with the present invention.

FIG.6 is a flow chart of information processing executed by a computer in an image processing apparatus in accordance with the present invention.

FIG.7 is a view showing a simulation displayed image executed by an image processing apparatus in accordance with the present invention.

FIG.8 is a flow chart describing the details of a part of the flow chart shown in FIG.6.

FIG.9 is a perspective view showing the outer appearance of a modified embodiment of a position indicating means in an image processing apparatus in accordance with the present invention.

FIGS. 10A-B are perspective views showing the outer appearance of a state of use of the position indicating means shown in FIG.9.

FIG. 11 is a perspective view showing the outer appearance of another modified embodiment of a position indicating means in an image processing apparatus in accordance with the present invention.

FIG.12 is a perspective view showing the outer appearance of a state of use of the position indicating means shown in FIG.11.

DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS

FIG.1 is a perspective view of the outer appearance of an image processing apparatus in accordance with the present invention. The image processing apparatus performs image processing to a processed object (image) 12 such as deformation, displacement, cutting, rotation and the like by operating the processed object displayed on a displaying screen 11 of an image display device 10 using an operating cursor 13. The image processing apparatus is characterized by that when one of the processing functions such as deformation, displacement, cutting, rotation and so on is selected to a processed object, operation of the processing in the processing function is made easy by displaying a shape of the operating cursor 13 suitable for the selected processing function to give recognition to an operator.

A three-dimensional position detecting means for detecting the instructing action form and motion of a hand 101 and fingers of an operator 100 has a position indicating means 21 for transmitting position indicating signals mounted on the hand 101 of the operator 100 and three receivers 22a to 22c for receiving the position indicating signals placed in the image display device 10.

The position indicating means 21 has a plurality of ring-shaped finger position transmitters 21b receiving electric power and signals through signal lines 21a mounted on the individual fingers of the hand 101, and a hand position transmitter 21d mounted and positioned on the back of the hand 101 with a band 21c. The finger position transmitters 21b and the hand position transmitter 21d and the receivers 22a to 22c are used such that each of the three-dimensional positions of the transmitters 21b, 21d, that is, each of the positions of the three fingers and the hand 101, is measured (detected) by measuring a time lag from the time each of the transmitters 21b, 21d transmits an ultrasonic signal to the time the three receivers 22a to 22c receive the ultrasonic signal at three positions. Each of the transmitters 22b, 22d is controlled by a sync signal so that the ultrasonic signal is transmitted with time-sharing, and the generating source of the ultrasonic signal received by the each of the receivers 22a to 22c is specified, and the position detecting processing for each of the hand 101 and the fingers is executed with time-sharing.

Advantageously, the three-dimensional position detecting means may be modified by replacing the receiver 22b with a video camera, and then processing image data taken from the hand of the operator using a computer to detect the position. By doing so, the operator is not required to mount each of the transmitters for detecting the positions of the hand and the fingers on his hand, which leads to an advantage to eliminate work to mount them.

A computer 30 generates image data for displaying a processed object and cursor image data for displaying cursor and transfers them to the image display device 10. The computer 30 processes the cursor image data and the processed object data using the received signals from the receivers 22a to 22c, the instruction signals from a keyboard 41 and the mouse 42, and the input signal from a microphone 43 as a voice inputting means. The computer 30 also controls speakers 44a, 44b as an imitation sound generating means, and transmits a sync signal for switching control of a pair of stereo-glasses 46 for stereoscopically seeing a three-dimensional image by giving image switching sync signals to an image sync signal generator 45.

The stereo-glasses 46 have liquid-crystal shutters which are opened and shut corresponding to an image for left side eye or an image for right side eye displayed on the image display device 20 so that the operator 100 can see the images with his left eye or right eye, and the sync signal receiver 47 receives the switching control sync signal from the transmitter 45 to control opening and shutting operation of the liquid-crystal shutter.

FIG.2 is a displayed image displayed on the display screen 11 of the image display device 10 in accordance with the present invention. A processed object 12, a cursor 13, a reference cursor 14, a legend display portion 15 and so on are displayed on the display screen 11.

In the legend display portion 15, plural kinds (plural shapes) of operating cursors 152, the corresponding processing function display characters 153 and the corresponding function selecting instruction form FIGS. 154 using form of fingers are displayed as a function selection legend.

Selection of function (operation cursor 152) is performed by forming the finger form in the corresponding selecting instruction action form, or by inputting number for the function selecting number from the keyboard 41, or by moving the reference cursor 14 to the frame of a desired function using the keyboard 41 or the mouse 42 to instruct. The function selecting instruction form FIGS. 154 showing examples of selecting instruction using the action forms of fingers provide instructing action forms for the function selection apt to make a mistake due to resemblance in the acting form of fingers as reference information to support the selecting operation by an inexperienced operator.

FIG.3 shows the relationship among the image processing functions, the operating cursor shapes and the selecting instruction action forms defined in order to display the function selecting legend in the embodiment. In the figure, operating cursors displaying together regular hexahedral processed objects and selecting instruction action forms of fingers are shown with pairing as examples.

Instruction (point) function is expressed by an arrow-shaped operating cursor, and the selection instructing action form for selecting the function is the form of straightening forefinger from the fist. The instructing point is the tip of the arrow.

Deforming function for deforming a processed object to the pushing direction is expressed by a cone-shaped operating cursor, and the selection instructing action form for selecting the function is the form of straightening forefinger and middle finger from the fist.

Deforming function for deforming a processed object to the pulling direction is expressed by a bent-tip cone-shaped operating cursor, and the selection instructing action form for selecting the function is the form of straightening forefinger and middle finger with bending the tips from the fist.

Displacement function for displacing a processed object is expressed by a grasping-hand-shaped operating cursor, and the selection instructing action form for selecting the function is the form of hand grasping an object.

Parallel displacement function for displacing a processed object in a parallel direction is expressed by a spatula-shaped operating cursor, and the selection instructing action form for selecting the function is the form of open-palm state.

Plane forming function for forming a surface of a processed object is expressed by a brush-shaped operating cursor, and the selection instructing action form for selecting the function is the same form of open-palm state as in the parallel displacement function.

Cutting function for cutting a processed object is expressed by a kitchen knife- or knife-shaped operating cursor, and the selection instructing action form for selecting the function is the same form of open-palm state as in the parallel displacement function.

In a case where selection instructing action forms are the same or similar such as the parallel displacement function, the plane forming function and the cutting function, the functions are displayed in the legend portion and the selection is performed by the selecting instruction by fingers together with the other instructing means such as voice or the mouse. Such a combined selecting instruction decreases number of kinds of the selection instructing action forms and consequently selection of many functions can be performed without complicated selecting instruction due to a slightly difference.

Rotation function for rotating a processed object is expressed by a top-shaped operating cursor, and the selection instructing action form for selecting the function is the form of rubbing forefinger with the inside of the thumb as to rotate a top.

By expressing the shape of the operating cursor with such a shape symbolizing the processing function to the processed object, the operator can easily recognize the contents of processing. By expressing the selection instructing action form of fingers for selecting a processing function with the form associating the processing function, an operator knowing the apparatus not so well can easily perform the processing operation.

FIG.4 is an example showing the relationship among the image processing function, the content of image processing corresponding to the motion of the operating cursor to a processed object expressed by a regular hexahedron and the instructing operation to the operating cursor in this embodiment. This image is displayed in the legend portion 15 as an operation legend after the processing function is selected.

Scroll processing function is such an image processing function that when an operating cursor is performed moving operation to be hidden in the back of a processed object or the operating cursor is operated to arrive at an arbitrarily set position near the processed object and further to be moved, the processed object is rotated in the opposite direction to the moving direction. That is, the scroll processing function is the image processing function that when the cursor is operated to move so as to instruct the shade portion of the processed object, the processed object is automatically rotated so as to bring the shaded portion to the front. There is an advantage in that the instruction input operation to the shaded portion of the processed object to bring the front becomes easy.

Displacement-on-surface processing function is such a function that even when a cursor operation to push a cursor in the surface of a processed object is performed, the operating cursor is moved in keeping the cursor on the surface. In a case where there is a component of moving operation parallel to the surface of the processed object, the operating cursor is moved on the surface corresponding to the component. This processing function makes the operation to instruct a point on the surface of the processed object easy, since the operating cursor does not enter into the inside of the processing object even if the position of finger to instruct moving of the operating cursor fluctuates.

Directing-to-proceeding-direction function is such a function that an operating cursor is always directed to the proceeding direction regardless of the positional relationship of fingers of an operating hand. For example, it is difficult to make an action to instruct the left side surface of a processing object using right hand. In such a case, this function directs an operating cursor in a proceeding direction (to the right) by moving a finger toward the proceeding direction (from left to right) to realizes the instruction to the left side surface of the processed object using the operating cursor.

Upright function is such a function that by contacting an operating cursor to a surface of a processed object, the processed object is raised at an arbitrary angle to the surface. This function is used when it is necessary to clarify an instructing position of an operating cursor to a processed object. In a case of deformation processing operation, the direction of deformation is clearly specified and therefore deformation toward the accurate direction can be realized.

Step displacement function is such a function that the amount of displacement of an operating cursor per one instructing operation and the amount of processing (displacement) of a processed object using the operating cursor are specified. This makes the image processing accurate or speedy.

Plural cursor display function is such a function that when a processed object is operated with one operating cursor, another operating cursor is displayed and another operation is performed to the processing object with the operating cursor. For instance, in a case of applying plural loads or plural temperatures to a processed object, the loads or temperatures can be applied to instructed arbitrary positions using plural operating cursors. Since such image processing display screens can visually display the simulation conditions, it becomes easy to understand the selecting conditions and the simulation results.

The selection of processing function and the operation of cursor are performed on the legend portion by using the instruction by finger forms together with the other instructing means such as voice or the mouse.

FIG.5 is a functional block diagram of an information processing program for realizing the image processing functions described above executed by the computer 30 in an image processing apparatus in accordance with the present invention.

The computer 30 comprises a three-dimensional action detecting means 301, a meaning judging means 302, voice recognition means 303, a command signal input means 304, legend image generating means 305, a cursor generating means 306, an image memory 307, a processed object generating means 308, a three-dimensional image generating means 309, a position judging means 310 and an imitation sound generating means 311.

The three-dimensional action detecting means 301 obtains three-dimensional positions of the hand 101 and the individual fingers of an operator 100 based on the individual signals from individual transmitters 21b, 21d received by the receivers 22a to 22c, and obtains the action forms of the fingers by sequentially monitoring the positions, and then transmits each of the obtained data to the meaning judging means 302 in the next stage.

The voice recognition means 303 analyzes voice signals input from a microphone 43 and converts to command signals. The command signals support the judgement in the meaning judging means 302 in the next stage and also the instruction input such as the operation or support the function selection displayed in a legend display portion 15. If the instruction input from the keyboard or the mouse during operating a cursor is eliminated by the support of the voice input, the cursor operating efficiency is improved and the operating cursor 13 can be operated with the both hands and consequently a more complex and real operation can be realized. Even in a case of instruction input with one hand, it is possible to use many kinds of processing functions since number of kinds of commands can be increased.

The meaning judging means 302 judges the function and the action selected and instructed based on the each of the signals given from the three-dimensional action detecting means 301 and the command signal given from the voice recognition means 303, and generates a control signal corresponding to the judged result. The displayed position of the operating cursor 13 is determined by moving on the display screen 11 corresponding to the action form of the hand and the fingers of the operator.

The command signal input means 304 generates a command signal corresponding to the instruction input signals from the keyboard 41 and the mouse 42. This command signal performs display control and selecting instruction of legend images generated in the legend image generating means 305 to be described later.

The legend image generating means 305 generates a legend image displayed on the legend display portion 15 for selecting and operating processing of the plural image processing functions. The legend image supplies the reference information for function selection and processing operation by the instructing action of the hand and fingers. The legend image makes it possible to perform selecting instruction input with other operating means and command means when other processing operation is selected during selecting or operating a function which is apt to make a mistake due to resemblance in the acting form of fingers.

The cursor generating means 306 corresponds to the judged result of the meaning judging means 302 and the input signal from the command signal input means 304, specifies an operating cursor 13 and a reference cursor corresponding to the selected and instructed function, refers the image data in the image memory 307, and generates cursor image data so as to display or displace it in the position specified by the display position data.

The processed object generating means 308 generates a processed object image data to display a processed object 12 to be processed and operated by the operating cursor 13 based on the data of the image memory 307. When the operating cursor 13 contacts to the processed object 12, image processing such as deformation, displacement, cutting, rotation or the like of the processed object 12 is performed corresponding to the function of the operating cursor 13, the operating content of the cursor after then and the property if the processed object 12.

The three-dimensional image generating means 309 generates the three-dimensional image display data to be displayed on the display screen 11 of the image display device 10 and the switching control sync signal. The three-dimensional image displayed on the display screen 11 contains the processed object 12, the operating cursor 13, the reference cursor 14 and the legend display portion 15. The image data is generated by synthesizing image data generated by the legend image generating means 305, the cursor generating means 306 and the processed object generating means 308. In order to obtain three-dimensional feeling from an image displayed on the plane display screen 11, image data for left eye and image data for right eye are alternatively generated so that the displayed images corresponding to the azimuth difference of the left eye and the right eye of the operator 100 are alternatively displayed, and the switching control sync signals to switch the liquid crystal shutters of the stereo-glasses 46 are generated so that the operator 100 can see the two kinds of the displayed images with the corresponding eyes.

The image display means 10 provide the images based on the image data for left eye and the image data for right eye generated by the three-dimensional image generating means 309 to the operator 100 by displaying the images alternatively. The switching control sync signal generator 45 transmits the switching control sync signals to open and shut the liquid crystal shutters so that the stereo-glasses 46 supplies the left eye with the image for left eye and the right eye with the image for right eye.

The position judging means 310 judges presence and absence of contact between the processed object 12 and the operating cursor 13 by comparing mutual relationship between the position of the processed object 12 and the position of the operating cursor 13.

The imitation sound generating means 311 generates a contact imitating sound signal corresponding to the function of the operating cursor 13 and the properties of the processed object 12 when the position judging means 310 judges that the processed object 12 and the operating cursor 13 are contact to each other. The generated contact imitating sound signal is given to the speakers 44a, 44b to convert to a contact imitating sound.

FIG.6 is a flow chart of information processing executed by the computer 30 to realize each of the function means as described above.

Step 601; An initial image as shown in FIG.2 is displayed. Although the function selecting legend as described in referring to FIG.3 is displayed in the legend display portion 15, "instruction" function is set as the initial state in this embodiment, and therefore the operating cursor 13 is displayed as a form for the "instruction". The processed object 12 is displayed with a pre-generated form, or not displayed.

Step 602; When the position instructing means 21 is brought to operation and a receiving signal of receiving signals from the individual transmitters 21b, 21d is generated, the positions and movements of the fingers are detected and the instructing action form is detected.

Step 603; The voice input signals of the operator 100 are analyzed with the voice recognition means, and the function selecting instruction or the cursor operating instruction is obtained.

Step 604; the instruction action form and the voice input signal are logically processed to perform meaning judgement.

Step 605; Based on the result of the meaning judgement, functional arranging order of the function selecting legend image displayed in the legend display portion 15 is updated. The selected and instructed processing function is placed in the top and the others are re-arranged in order of similar function, or in order of similar function in selecting instruction action form, or in order of high frequency of use. With arranging the processing functions in order of similar function in selecting instruction action form, there is an advantage in that it is possible to refer the other functions when selection or operating instruction is performed. And with arranging the processing functions in order of high frequency of use, there is an advantage in that the function selection is performed speedy.

The above is the same when a desired function frame on the legend display portion 15 is instructed by moving the reference cursor 14 using the mouse 42, or when a numerical value of the function selecting number is input by the keyboard 41.

Step 606; A processing function which the operator 100 requires is selected (selective input). Instruction input for selection is performed by a command such as the instruction action form of the fingers, the voice input, the keyboard or mouse input. Then an operating cursor 13 corresponding to a new processing function selected according to the meaning judgement is specified.

The processing functions are, as described in referring to FIG. 3, the instructing function to instruct a certain point, displacement function to displace a processed object, the deformation function to deform a processed object, the cutting function to cut a displayed object and so on.

Step 607; Cursor image data to display the shape of an operating cursor 13 symbolizing the processing function is generated corresponding to the selected processing function.

Step 608; Processed object image data to display the processed object 12 on the display screen 11 is generated. The processed object 12 is displayed by reading out pre-generated image data from the image memory, or is generated according to data input from the keyboard 41 or the mouse 42.

Step 609; The image position data of the processed object 12 and the image position data of the operating cursor 13 are detected.

Step 610; By comparing the image position data of the processed object 12 and the image position data of the operating cursor 13, presence or absence of the contact between the processed object 12 and the operating cursor 13 is judged.

Step 611; When the contact (closely approach) between the processed object 12 and the operating cursor 13 is detected, a processing operation legend to be executed by the operating cursor after then is displayed in the legend display portion 15.

Although a part of the processing operating legend is described in referring to FIG.4, as to the instruction functions there are the pass function for passing a moving operating cursor 13 through a processed object without causing anything, the stop function for stopping an operating cursor 13 on the surface of a processed object 12 even if the operating cursor 13 is operated to enter the inside of the processed object 12, the upright function for raising a processed object 12 instructed by an operating cursor 13 so as to have an arbitrary angle to a surface, the on-surface displacement function for moving an operating cursor 13 on the surface of a processed object 12 by the component in the direction along the surface of the processed object when the cursor 13 is operated to enter the inside of the processed object 13. As to the deformation function,there is the simulation function for deforming a processed object 12 by giving conditions using plural cursors 13.

Step 612; Selection of cursor operation suitable for a desired image processing is performed. The selecting instruction is performed by the selecting instruction action form of fingers, the voice input, the keyboard input or the mouse input. Since the selection of processing function and the selection of processing operation can be performed by the action of fingers, it is possible to perform the selecting instruction input while the operating cursor 13 is being operating. If the keyboard and the mouse are not used in the selection, the selection of processing operation can be performed speedy since motion of hand and motion of eye-line for inputting the selecting instruction are not required.

Step 613; After selecting the processing operation, cursor image data to display an operating cursor corresponding to the processing operation is generated. Since as to the stop, upright, on-surface displacement functions the three-dimensional action of hand and the operation of the operating cursor 13 on the display screen 11 do not correspond one-to-one, function cursor image data is re-generated.

Step 614; After the processed object 12 is processed by operating the operating cursor 13, image data of the processed object 12 is generated.

Step 615; Necessity of generating the imitation sound generated when the processed object 12 is processed by operating the operating cursor 13 is judged from the processing function, the processed object 12 and the operating cursor 13.

Step 616; If generation of the imitation sound is required, imitation sound data is generated corresponding to the properties of the processing function, the cursor operation and the processed object.

Step 617; The speakers are controlled so as to generate an imitation sound based on the imitation sound data. Since the imitation sound supplies auditory sense information together with visual sense information of the screen display to the operator, it makes easy to operate the operating cursor 13 as well as it makes accurate to operate the operating cursor 13.

Step 618; New display image data re-constructing the relationship between operation of the operating cursor 13 and the processed object 12 processed based on the operation is generated.

Step 619; The new display image data is displayed on the image display device 10 to provide it to the operator 100.

Step 620; Whether another operating cursor display is required is confirmed. If a processing operation requiring plural operating cursors is selected, the processing is branched to the block of new operating cursor setting, operating and image processing.

Step 621; A new operating cursor 13 required for processing operation is set. The old operating cursor 13 is stopped at the present position, or cooperates with new operating cursor. Then necessary image processing is performed and displayed.

Step 622; If completion of the processing is indicated, the image processing is ended. If not, the processing returns to Step 602 to repeat the same processing.

In a case where selective instruction of processing function and cursor operating instruction input are performed using a plurality of three-dimensional action detection means, instruction input from each of the three-dimensional action detecting means is discriminated, and function selection and cursor operation obtained by combining the independent function selection and cursor operation corresponding to each of the detecting means are performed.

FIG.7 shows a display screen when an image processing (operation) is performed using a plurality of operating cursors 13a, 13b, 13c in an image processing apparatus in accordance with the present invention. This image processing function is suitable for a simulation such as deformation of a processed object 13, and each of the operating cursors 13a to 12c is used to give the simulation condition.

Various kinds of the boundary conditions set to the operating cursors 13a to 13c can be set by inputting from the keyboard 41, and the set boundary conditions 155 are displayed in the legend display portion 15. The operating cursors shown as examples are an operating cursor 13a for concentrated load, an operating cursor 13b for distributed load and an operating cursor 13c for temperature.

Each of the operating cursors 13a to 13c is selected by the instruction action form of fingers or by instructing the function of the legend display portion 15 using a numerical value or the cursor, and each of the cursors is positioned in each of the arbitrary positions. The operating cursor 13b for temperature expresses that the acting heat source is moving toward the arrow direction.

Since the setting state of the boundary conditions in such an image processing display is visually expressed by the operating cursors 13a to 13c, there is an advantage in that the simulation state can be easily understood.

FIG.8 shows a flow chart of an information processing contained in Step 621 described above in order to execute the simulation processing shown in FIG.7. This information processing comprises the processing steps of setting of each of operating cursors 13a to 13c and display processing 621a, judging processing for completion of setting 621b, simulation calculation processing 621c such as deformation of the processed object 12 executed according to the boundary conditions set by the operating cursors 13a to 13c after completion of setting, display image data generating processing 621d according to the result of the calculation, and image display processing 621e according to the display image data.

FIG.9 shows a modified embodiment of a position indicating means 21 in an information processing apparatus in accordance with the present invention. Constructing parts common to the embodiment shown in FIG.1 are identified by the same reference characters, and detailed description is omitted here.

This modified embodiment is characterized by providing a auxiliary switch means 21e in the middle finger side of a transmitter 21b mounted on the thumb. The auxiliary switch means 21e is used for generating an auxiliary command such as command to instruct displacement and stopping of the operating cursor 13.

The ON/OFF operation of the auxiliary switch means 21e is performed by pushing the auxiliary switch means 21e to the middle finger side or detaching the auxiliary switch means from the middle finger side. If the command using the auxiliary switch means is used so effectively that other command means is unnecessary to use, the selection and the operability of the operating cursor 13 are improved since motion of eye line and hands in operating other command means are not required .

It may be possible to detect the three-dimensional position by employing light generating means as the transmitters 21b, 21d and optical two-dimensional sensors as the receivers 23a to 23c.

FIG.11 shows another modified embodiment of a position indicating means 21 in an information processing apparatus in accordance with the present invention. And FIG.12 shows a state of mounting the position indicating means shown in FIG.11 on a hand (fingers).

This modified position indicating means 21 comprises a main body 50 which can be gripped by hand, a band 51 for support the main body 50 in the palm of hand, fitting parts 52a to 52e for fit to fingers, a take-up member 53 arranged corresponding to each of the fitting parts, a wire member 56 wound in each of the take-up members 53 and having the extending outer end connected to the fitting part 52, a detecting member 54 for detecting rotating angle of the take-up member 53, a hand position detecting means, not shown, for specifying the three-dimensional position of the main body 50.

The wire member 56 is wound up with the take-up member 53. When fingers are stretched, wire members 56 connected to the fitting parts 52a to 52e fit to the tips of fingers are extracted, and accordingly the take-up parts 53 are rotated. When the fingers are returned to the original positions, the take-up members 56 are rotated so as to wind up the wire members 56.

The detecting means 54 detects the amount of stretching of each of the fingers by measuring the amount of rotation of the take-up member 53, and the three-dimensional position of the hand and the instructing action form of each finger are specified from the detected values of the hand position detecting means and the detecting means 54.

In the optical or ultrasonic three-dimensional position detection, it sometime becomes difficult to detect the position when there is an obstacle between the transmitting side and the receiving side. According to this embodiment, if a magnetic sensor is employed to the hand position detecting means provided in the main body 50, the sensor is hardly affected by the obstacle since magnetic field is transmitted through the obstacle unless it is magnetic material. The positional relationship among the fingers are detected by the detecting member 54 mechanically linked with each of the fingers. Therefore, there is an advantage in that the positional detection of fingers is hardly affected by an external obstacle.

According to the present invention, processing operation of an image (processed object) can be speedy and easily performed by operating a cursor. Since the operating cursor having a shape symbolizing the image processing function is displayed, the processing operation can be performed accurately.

Further, in a simulation using a computer, the condition to act on a processed object can be speedy and easily set.

Claims (8)

What is claimed is:
1. An image processing apparatus comprising:
image displaying means for displaying an image;
three-dimensional action detecting means for detecting actions and three-dimensional positions of a hand and fingers of an operator;
cursor generating means for generating operating cursor image data based on a detected result of said three-dimensional action detecting means;
processed object generating means for generating image data of a processed object to be operated upon by an operating cursor displayed based on said operating cursor image data;
display image generating means for synthetically displaying said operating cursor image data and said processed object image data on said image display means;
an image memory for storing image data indicative of a plurality of shaves of the operating cursor corresponding to a plurality of processing functions of said image processing apparatus, each of said Plurality of shapes visually symbolizing a respective one of said plurality of functions;
meaning judging means for judging a meaning of an instruction from a form of action of the hand and fingers detected by said three-dimensional action detecting means;
control changing means for changing a processing function of the image processing apparatus corresponding to a judged result of said meaning judging means;
cursor changing means for changing a shape of the operating cursor corresponding to the processing function of the image processing apparatus as changed by said control changing means, said cursor changing means changing said shape by referring to said image data in said image memory; and
legend image generating means for generating function selecting legend image data which displays image processing function and an operating cursor corresponding to said image processing function.
2. An image processing apparatus according to claim 1, wherein said function selecting legend image data includes an instruction action form for selecting a specified image processing function.
3. An image processing apparatus according to claim 2, wherein said legend image generating means generates function selecting legend image data of the processing function approximated by the instruction action form.
4. An image processing apparatus according to claim 1, which comprises legend image generating means for generating operating legend image data which displays a cursor operation and the image processing corresponding to said cursor operation.
5. An image processing apparatus according to claim 4, which further comprises position judging means for judging presence or absence of contact between the processed object and the operating cursor, and said legend image generating means generates operating legend image data based on the result of the judgement of contact.
6. An image processing apparatus comprising image displaying means for displaying an image, three-dimensional action detecting means for detecting actions and three-dimensional positions of the hand and the fingers of an operator, cursor generating means for generating operating cursor image data based on the detected result of said three-dimensional action detecting means, processed object generating means for generating image data of a processed object to be operated by an operating cursor displayed based on said operating cursor image data, and display image generating means for synthetically displaying said operating cursor image data and said processed object image data on said image display means, which comprises:
meaning judging means for judging the meaning of an instruction from the form of action of the hand and the fingers detected by said three-dimensional action detecting means;
an image memory for storing image data of shape which is shape of operating cursor corresponding to each of processing functions of said image processing means and visually symbolizes each of said functions;
control changing means for changing the processing function of the image processing apparatus corresponding to the judged result of said meaning judging means;
cursor changing means for changing the shape of the operating cursor corresponding to the processing function of said image processing means referring to said image data in said image memory;
legend displaying means for displaying the processing function and plural kinds of operating cursors corresponding to the result of the judgement of said meaning judging means;
condition setting means for setting a processing operating condition corresponding to each of said operating cursor; and
said processed object generating means generating image data of the processed object corresponding to said processing operating condition.
7. A method of image processing in which image processing of a processed object is performed by displaying said processed object and an operating cursor on image display means and by operating said operating cursor corresponding to a three-dimensional action of a hand and fingers of an operator. said method comprising steps of:
storing image data indicative of a plurality of shapes of the operating cursor corresponding to a plurality of processing functions, each of said plurality of shapes visually symbolizing a respective one of said plurality of processing functions;
judging a meaning of an instruction from a form of action of the hand and fingers detected by a three-dimensional action detecting means;
changing a processing function of the image processing apparatus corresponding to a judged result of said judging step; and
changing a shape of the operating cursor corresponding to the processing function as changed by said processing function changing step, said shape being changed by referring to said image data; and
displaying a function selecting legend image expressing an image processing function and an operating cursor corresponding to said image processing function.
8. A method of image processing according to claim 7, further comprising a step of:
displaying an operating legend image expressing a cursor operation and an image processing function corresponding to said cursor operation.
US08515450 1994-08-22 1995-08-15 Three-dimensional input device with displayed legend and shape-changing cursor Expired - Fee Related US5801704A (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
JP19663294A JPH0863326A (en) 1994-08-22 1994-08-22 Image processing device/method
JP6-196632 1994-08-22

Publications (1)

Publication Number Publication Date
US5801704A true US5801704A (en) 1998-09-01

Family

ID=16360996

Family Applications (1)

Application Number Title Priority Date Filing Date
US08515450 Expired - Fee Related US5801704A (en) 1994-08-22 1995-08-15 Three-dimensional input device with displayed legend and shape-changing cursor

Country Status (2)

Country Link
US (1) US5801704A (en)
JP (1) JPH0863326A (en)

Cited By (117)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5880733A (en) * 1996-04-30 1999-03-09 Microsoft Corporation Display system and method for displaying windows of an operating system to provide a three-dimensional workspace for a computer system
US5936610A (en) * 1993-07-27 1999-08-10 Canon Kabushiki Kaisha Control device for image input apparatus
WO1999042946A2 (en) * 1998-02-18 1999-08-26 Geovector Corporation Apparatus and methods for presentation of information relating to objects being addressed
US6118888A (en) * 1997-02-28 2000-09-12 Kabushiki Kaisha Toshiba Multi-modal interface apparatus and method
US6128003A (en) * 1996-12-20 2000-10-03 Hitachi, Ltd. Hand gesture recognition system and method
US6144366A (en) * 1996-10-18 2000-11-07 Kabushiki Kaisha Toshiba Method and apparatus for generating information input using reflected light image of target object
US6144384A (en) * 1996-02-20 2000-11-07 Yugen Kashia Aloalo International Voxel data processing using attributes thereof
US6154558A (en) * 1998-04-22 2000-11-28 Hsieh; Kuan-Hong Intention identification method
US6166718A (en) * 1996-06-18 2000-12-26 Konami Co., Ltd. Video game system with vertical array of cursor images
US6266061B1 (en) * 1997-01-22 2001-07-24 Kabushiki Kaisha Toshiba User interface apparatus and operation range presenting method
US6363160B1 (en) 1999-01-22 2002-03-26 Intel Corporation Interface using pattern recognition and tracking
US6396476B1 (en) * 1998-12-01 2002-05-28 Intel Corporation Synthesizing computer input events
US6476834B1 (en) * 1999-05-28 2002-11-05 International Business Machines Corporation Dynamic creation of selectable items on surfaces
US20030026170A1 (en) * 2001-07-31 2003-02-06 Yang Jong H. Same-hand control of a multi-function device
US6522292B1 (en) 2000-02-23 2003-02-18 Geovector Corp. Information systems having position measuring capacity
US6538649B2 (en) 1998-12-01 2003-03-25 Intel Corporation Computer vision control variable transformation
US6563512B1 (en) * 1996-06-25 2003-05-13 Fujitsu Limited Object editing method, object editing system and computer memory product
US20030122839A1 (en) * 2001-12-26 2003-07-03 Eastman Kodak Company Image format including affective information
US20030132909A1 (en) * 2002-01-11 2003-07-17 International Business Machines Corporation Pointing device attribute variance based on distance/time ratio
US20030187532A1 (en) * 1999-06-10 2003-10-02 Dassault Systemes Knowledge-based polymorph undockable toolbar
US6647131B1 (en) 1999-08-27 2003-11-11 Intel Corporation Motion detection using normal optical flow
US6654483B1 (en) 1999-12-22 2003-11-25 Intel Corporation Motion detection using normal optical flow
EP1369822A2 (en) * 2002-05-31 2003-12-10 CAD Center Corporation Apparatus and method for controlling the shift of the viewpoint in a virtual space
US20030234782A1 (en) * 2002-06-21 2003-12-25 Igor Terentyev System and method for adaptively labeling multi-dimensional images
US20040001112A1 (en) * 2002-01-25 2004-01-01 Silicon Graphics, Inc. Volume management system for volumetric displays
US6681031B2 (en) 1998-08-10 2004-01-20 Cybernet Systems Corporation Gesture-controlled interfaces for self-service machines and other applications
US6690393B2 (en) * 1999-12-24 2004-02-10 Koninklijke Philips Electronics N.V. 3D environment labelling
US6690370B2 (en) 1995-06-07 2004-02-10 Geovector Corp. Vision system computer modeling apparatus including interaction with real scenes with respect to perspective and spatial relationship as measured in real-time
US20040125076A1 (en) * 2001-06-08 2004-07-01 David Green Method and apparatus for human interface with a computer
US20040136062A1 (en) * 2002-05-13 2004-07-15 Avner Farkash Portable periscope
US20040161132A1 (en) * 1998-08-10 2004-08-19 Cohen Charles J. Gesture-controlled interfaces for self-service machines and other applications
US20040193413A1 (en) * 2003-03-25 2004-09-30 Wilson Andrew D. Architecture for controlling a computer using hand gestures
US20040189720A1 (en) * 2003-03-25 2004-09-30 Wilson Andrew D. Architecture for controlling a computer using hand gestures
US20040219961A1 (en) * 2003-04-08 2004-11-04 Ellenby Thomas William Computer games having variable execution dependence with respect to spatial properties of a mobile unit.
US20040246234A1 (en) * 2001-03-01 2004-12-09 Luis Serra Display apparatus
US20050088410A1 (en) * 2003-10-23 2005-04-28 Apple Computer, Inc. Dynamically changing cursor for user interface
US6938218B1 (en) 2000-04-28 2005-08-30 James Nolen Method and apparatus for three dimensional internet and computer file interface
US20050226469A1 (en) * 2004-04-07 2005-10-13 Jong-Hwan Ho Method for displaying finger images in wireless telecommunication terminal
US20060031874A1 (en) * 2004-08-07 2006-02-09 Samsung Electronics Co., Ltd. Three-dimensional motion graphic user interface and method and apparatus for providing the same
US20060067576A1 (en) * 2004-03-17 2006-03-30 James Marggraff Providing a user interface having interactive elements on a writable surface
US7031875B2 (en) 2001-01-24 2006-04-18 Geo Vector Corporation Pointing systems for addressing objects
US20060190812A1 (en) * 2005-02-22 2006-08-24 Geovector Corporation Imaging systems including hyperlink associations
US20060202974A1 (en) * 2005-03-10 2006-09-14 Jeffrey Thielman Surface
US20060209016A1 (en) * 2005-03-17 2006-09-21 Microsoft Corporation Computer interaction based upon a currently active input device
US20060238490A1 (en) * 2003-05-15 2006-10-26 Qinetiq Limited Non contact human-computer interface
US7134095B1 (en) 1999-10-20 2006-11-07 Gateway, Inc. Simulated three-dimensional navigational menu system
US20070256032A1 (en) * 2006-04-28 2007-11-01 Petri John E Presenting identifiers and states of processes in a stacked cursor
US20080024484A1 (en) * 2006-06-26 2008-01-31 University Of Southern California Seamless Image Integration Into 3D Models
US7342574B1 (en) * 1999-10-29 2008-03-11 Ricoh Company, Ltd. Method and apparatus for inputting information including coordinate data
US20080192007A1 (en) * 2002-02-07 2008-08-14 Microsoft Corporation Determining a position of a pointing device
US20090116692A1 (en) * 1998-08-10 2009-05-07 Paul George V Realtime object tracking system
US20090245691A1 (en) * 2008-03-31 2009-10-01 University Of Southern California Estimating pose of photographic images in 3d earth model using human assistance
US20090286653A1 (en) * 2006-06-21 2009-11-19 Wiber Laurent Remote control device for an electronic apparatus in particular for performing a physical exercise
US7705830B2 (en) * 2001-02-10 2010-04-27 Apple Inc. System and method for packing multitouch gestures onto a hand
US20110083112A1 (en) * 2009-10-05 2011-04-07 Takashi Matsubara Input apparatus
US20110083106A1 (en) * 2009-10-05 2011-04-07 Seiko Epson Corporation Image input system
US20110179368A1 (en) * 2010-01-19 2011-07-21 King Nicholas V 3D View Of File Structure
US8006183B1 (en) * 2006-12-08 2011-08-23 Trading Technologies International Inc. System and method for using a curser to convey information
US20110221669A1 (en) * 2010-02-28 2011-09-15 Osterhout Group, Inc. Gesture control in an augmented reality eyepiece
USRE42794E1 (en) 1999-12-27 2011-10-04 Smart Technologies Ulc Information-inputting device inputting contact point of object on recording surfaces as information
US8055022B2 (en) 2000-07-05 2011-11-08 Smart Technologies Ulc Passive touch system and method of detecting user input
US8089462B2 (en) 2004-01-02 2012-01-03 Smart Technologies Ulc Pointer tracking across multiple overlapping coordinate input sub-regions defining a generally contiguous input region
US8094137B2 (en) 2007-07-23 2012-01-10 Smart Technologies Ulc System and method of detecting contact on a display
US8115753B2 (en) 2007-04-11 2012-02-14 Next Holdings Limited Touch screen system with hover and click input methods
US8120596B2 (en) 2004-05-21 2012-02-21 Smart Technologies Ulc Tiled touch system
US8149221B2 (en) 2004-05-07 2012-04-03 Next Holdings Limited Touch panel display system with illumination and detection provided from a single edge
US8228304B2 (en) 2002-11-15 2012-07-24 Smart Technologies Ulc Size/scale orientation determination of a pointer in a camera-based touch system
US20120235946A1 (en) * 2006-01-30 2012-09-20 Microsoft Corporation Controlling application windows in an operating system
US8274496B2 (en) 2004-04-29 2012-09-25 Smart Technologies Ulc Dual mode touch systems
US8289299B2 (en) 2003-02-14 2012-10-16 Next Holdings Limited Touch screen signal processing
US20120272180A1 (en) * 2011-04-20 2012-10-25 Nokia Corporation Method and apparatus for providing content flipping based on a scrolling operation
US8339378B2 (en) 2008-11-05 2012-12-25 Smart Technologies Ulc Interactive input system with multi-angle reflector
WO2012177322A1 (en) * 2011-06-21 2012-12-27 Qualcomm Incorporated Gesture-controlled technique to expand interaction radius in computer vision applications
US8384693B2 (en) 2007-08-30 2013-02-26 Next Holdings Limited Low profile touch panel systems
US8405637B2 (en) 2008-01-07 2013-03-26 Next Holdings Limited Optical position sensing system and optical position sensor assembly with convex imaging window
US8432377B2 (en) 2007-08-30 2013-04-30 Next Holdings Limited Optical touchscreen with improved illumination
US8456447B2 (en) 2003-02-14 2013-06-04 Next Holdings Limited Touch screen signal processing
US8456418B2 (en) 2003-10-09 2013-06-04 Smart Technologies Ulc Apparatus for determining the location of a pointer within a region of interest
US8456451B2 (en) 2003-03-11 2013-06-04 Smart Technologies Ulc System and method for differentiating between pointers used to contact touch surface
US8467133B2 (en) 2010-02-28 2013-06-18 Osterhout Group, Inc. See-through display with an optical assembly including a wedge-shaped illumination system
US8472120B2 (en) 2010-02-28 2013-06-25 Osterhout Group, Inc. See-through near-eye display glasses with a small scale image source
US8477425B2 (en) 2010-02-28 2013-07-02 Osterhout Group, Inc. See-through near-eye display glasses including a partially reflective, partially transmitting optical element
US8482859B2 (en) 2010-02-28 2013-07-09 Osterhout Group, Inc. See-through near-eye display glasses wherein image light is transmitted to and reflected from an optically flat film
US8488246B2 (en) 2010-02-28 2013-07-16 Osterhout Group, Inc. See-through near-eye display glasses including a curved polarizing film in the image source, a partially reflective, partially transmitting optical element and an optically flat film
US8508508B2 (en) 2003-02-14 2013-08-13 Next Holdings Limited Touch screen signal processing with single-point calibration
US20130311952A1 (en) * 2011-03-09 2013-11-21 Maiko Nakagawa Image processing apparatus and method, and program
US8620113B2 (en) 2011-04-25 2013-12-31 Microsoft Corporation Laser diode modes
US8635637B2 (en) 2011-12-02 2014-01-21 Microsoft Corporation User interface presenting an animated avatar performing a media reaction
US20140026103A1 (en) * 2012-07-18 2014-01-23 DS Zodiac, Inc. Multi-dimensional file system
US8692768B2 (en) 2009-07-10 2014-04-08 Smart Technologies Ulc Interactive input system
US8760395B2 (en) 2011-05-31 2014-06-24 Microsoft Corporation Gesture recognition techniques
US8799821B1 (en) * 2008-04-24 2014-08-05 Pixar Method and apparatus for user inputs for three-dimensional animation
US8898687B2 (en) 2012-04-04 2014-11-25 Microsoft Corporation Controlling a media program based on a media reaction
US8902193B2 (en) 2008-05-09 2014-12-02 Smart Technologies Ulc Interactive input system and bezel therefor
US8959541B2 (en) 2012-05-04 2015-02-17 Microsoft Technology Licensing, Llc Determining a future portion of a currently presented media program
US9091851B2 (en) 2010-02-28 2015-07-28 Microsoft Technology Licensing, Llc Light control in head mounted displays
US9097890B2 (en) 2010-02-28 2015-08-04 Microsoft Technology Licensing, Llc Grating in a light transmissive illumination system for see-through near-eye display glasses
US9100685B2 (en) 2011-12-09 2015-08-04 Microsoft Technology Licensing, Llc Determining audience state or interest using passive sensor data
US9097891B2 (en) 2010-02-28 2015-08-04 Microsoft Technology Licensing, Llc See-through near-eye display glasses including an auto-brightness control for the display brightness based on the brightness in the environment
US9128281B2 (en) 2010-09-14 2015-09-08 Microsoft Technology Licensing, Llc Eyepiece with uniformly illuminated reflective display
US9129295B2 (en) 2010-02-28 2015-09-08 Microsoft Technology Licensing, Llc See-through near-eye display glasses with a fast response photochromic film system for quick transition from dark to clear
US9134534B2 (en) 2010-02-28 2015-09-15 Microsoft Technology Licensing, Llc See-through near-eye display glasses including a modular image source
USD741361S1 (en) * 2011-06-07 2015-10-20 Cornell D. M. Judge Cornish Portion of display screen with an icon
US9182596B2 (en) 2010-02-28 2015-11-10 Microsoft Technology Licensing, Llc See-through near-eye display glasses with the optical assembly including absorptive polarizers or anti-reflective coatings to reduce stray light
DE102014106960A1 (en) * 2014-05-16 2015-11-19 Faindu Gmbh A method for displaying a virtual interaction on at least one screen and input device, system and method for a virtual application by means of a computing unit
DE102014107220A1 (en) * 2014-05-22 2015-11-26 Atlas Elektronik Gmbh Input device, computer or the operating system as well as vehicle
US9223134B2 (en) 2010-02-28 2015-12-29 Microsoft Technology Licensing, Llc Optical imperfections in a light transmissive illumination system for see-through near-eye display glasses
US9223405B2 (en) 2011-03-16 2015-12-29 Electronics And Telecommunications Research Institute Apparatus and method for inputting information based on events
US9229227B2 (en) 2010-02-28 2016-01-05 Microsoft Technology Licensing, Llc See-through near-eye display glasses with a light transmissive wedge shaped illumination system
US9285589B2 (en) 2010-02-28 2016-03-15 Microsoft Technology Licensing, Llc AR glasses with event and sensor triggered control of AR eyepiece applications
US9304593B2 (en) 1998-08-10 2016-04-05 Cybernet Systems Corporation Behavior recognition system
US9341843B2 (en) 2010-02-28 2016-05-17 Microsoft Technology Licensing, Llc See-through near-eye display glasses with a small scale image source
US9366862B2 (en) 2010-02-28 2016-06-14 Microsoft Technology Licensing, Llc System and method for delivering content to a group of see-through near eye display eyepieces
US9442607B2 (en) 2006-12-04 2016-09-13 Smart Technologies Inc. Interactive input system and method
US9596643B2 (en) 2011-12-16 2017-03-14 Microsoft Technology Licensing, Llc Providing a user interface experience based on inferred vehicle state
US9759917B2 (en) 2010-02-28 2017-09-12 Microsoft Technology Licensing, Llc AR glasses with event and sensor triggered AR eyepiece interface to external devices
USD822036S1 (en) * 2014-04-16 2018-07-03 Regents Of The University Of Minnesota Display screen or portion thereof with graphical user interface

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2338114B1 (en) * 2008-09-03 2017-03-15 Oblong Industries, Inc. Control system for navigating a principal dimension of a data space
CN106896952A (en) * 2011-03-31 2017-06-27 富士胶片株式会社 Stereoscopic display device, and method for accepting instruction
JP6107271B2 (en) * 2013-03-21 2017-04-05 カシオ計算機株式会社 The information processing apparatus, an information processing system and program
JP6315122B2 (en) * 2017-03-08 2018-04-25 カシオ計算機株式会社 Display control apparatus, a display control method, and program
JP6244069B1 (en) * 2017-04-17 2017-12-06 株式会社オプティム Remote work support system, remote work support method, and program

Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPS6244827A (en) * 1985-08-20 1987-02-26 Zetsudo Rania Jiyaron Apparatus and method for generating control signal accordingto action and position of hand
US4988981A (en) * 1987-03-17 1991-01-29 Vpl Research, Inc. Computer data entry and manipulation apparatus and method
US5252951A (en) * 1989-04-28 1993-10-12 International Business Machines Corporation Graphical user interface with gesture recognition in a multiapplication environment
US5347295A (en) * 1990-10-31 1994-09-13 Go Corporation Control of a computer through a position-sensed stylus
US5381158A (en) * 1991-07-12 1995-01-10 Kabushiki Kaisha Toshiba Information retrieval apparatus
US5423554A (en) * 1993-09-24 1995-06-13 Metamedia Ventures, Inc. Virtual reality game method and apparatus
US5454043A (en) * 1993-07-30 1995-09-26 Mitsubishi Electric Research Laboratories, Inc. Dynamic and static hand gesture recognition through low-level image analysis
US5481454A (en) * 1992-10-29 1996-01-02 Hitachi, Ltd. Sign language/word translation system
US5512919A (en) * 1992-03-31 1996-04-30 Pioneer Electronic Corporation Three-dimensional coordinates input apparatus
US5524195A (en) * 1993-05-24 1996-06-04 Sun Microsystems, Inc. Graphical user interface for interactive television with an animated agent
US5524187A (en) * 1991-03-25 1996-06-04 The Trustees Of Columbia University Worlds-within-worlds nested display and interaction system and method
US5558619A (en) * 1991-04-23 1996-09-24 Olympus Optical Co., Ltd. Endoscope system with automatic control according to movement of an operator

Patent Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPS6244827A (en) * 1985-08-20 1987-02-26 Zetsudo Rania Jiyaron Apparatus and method for generating control signal accordingto action and position of hand
US4988981A (en) * 1987-03-17 1991-01-29 Vpl Research, Inc. Computer data entry and manipulation apparatus and method
US4988981B1 (en) * 1987-03-17 1999-05-18 Vpl Newco Inc Computer data entry and manipulation apparatus and method
US5252951A (en) * 1989-04-28 1993-10-12 International Business Machines Corporation Graphical user interface with gesture recognition in a multiapplication environment
US5347295A (en) * 1990-10-31 1994-09-13 Go Corporation Control of a computer through a position-sensed stylus
US5524187A (en) * 1991-03-25 1996-06-04 The Trustees Of Columbia University Worlds-within-worlds nested display and interaction system and method
US5558619A (en) * 1991-04-23 1996-09-24 Olympus Optical Co., Ltd. Endoscope system with automatic control according to movement of an operator
US5381158A (en) * 1991-07-12 1995-01-10 Kabushiki Kaisha Toshiba Information retrieval apparatus
US5512919A (en) * 1992-03-31 1996-04-30 Pioneer Electronic Corporation Three-dimensional coordinates input apparatus
US5481454A (en) * 1992-10-29 1996-01-02 Hitachi, Ltd. Sign language/word translation system
US5524195A (en) * 1993-05-24 1996-06-04 Sun Microsystems, Inc. Graphical user interface for interactive television with an animated agent
US5454043A (en) * 1993-07-30 1995-09-26 Mitsubishi Electric Research Laboratories, Inc. Dynamic and static hand gesture recognition through low-level image analysis
US5423554A (en) * 1993-09-24 1995-06-13 Metamedia Ventures, Inc. Virtual reality game method and apparatus

Non-Patent Citations (12)

* Cited by examiner, † Cited by third party
Title
Ishibuchi et al., "Real Time Hand Gesture Recognition Using 3D Prediction Model", IEEE, pp. 324-328, Oct. 1993.
Ishibuchi et al., Real Time Hand Gesture Recognition Using 3D Prediction Model , IEEE, pp. 324 328, Oct. 1993. *
Kahaner, "Japanese Activities in Virtual Reality", IEEE Computer Graphics and Applications, Jan. 1994.
Kahaner, Japanese Activities in Virtual Reality , IEEE Computer Graphics and Applications, Jan. 1994. *
Morita et al., "A Computer Music System That Follows a Human Conductor", IEEE Computer, pp. 44-53, Jul. 1991.
Morita et al., A Computer Music System That Follows a Human Conductor , IEEE Computer, pp. 44 53, Jul. 1991. *
Shinohara, "Network Virtual Reality System", Pixel, No. 32, non-english, pp. 26-28, 1993.
Shinohara, Network Virtual Reality System , Pixel, No. 32, non english, pp. 26 28, 1993. *
Su et al, "The Virtual Panel Architecture: A 3D Gesture Framework", IEEE, pp. 387-393, 1993.
Su et al, The Virtual Panel Architecture: A 3D Gesture Framework , IEEE, pp. 387 393, 1993. *
Zimmerman et al., "A Hand Guesture Interface Device", CHI+GI 1987, pp. 189-192, Apr. 9, 1987.
Zimmerman et al., A Hand Guesture Interface Device , CHI GI 1987, pp. 189 192, Apr. 9, 1987. *

Cited By (183)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5936610A (en) * 1993-07-27 1999-08-10 Canon Kabushiki Kaisha Control device for image input apparatus
US6690370B2 (en) 1995-06-07 2004-02-10 Geovector Corp. Vision system computer modeling apparatus including interaction with real scenes with respect to perspective and spatial relationship as measured in real-time
US20040095345A1 (en) * 1995-06-07 2004-05-20 John Ellenby Vision system computer modeling apparatus
US6144384A (en) * 1996-02-20 2000-11-07 Yugen Kashia Aloalo International Voxel data processing using attributes thereof
US6016145A (en) * 1996-04-30 2000-01-18 Microsoft Corporation Method and system for transforming the geometrical shape of a display window for a computer system
US5880733A (en) * 1996-04-30 1999-03-09 Microsoft Corporation Display system and method for displaying windows of an operating system to provide a three-dimensional workspace for a computer system
US6166718A (en) * 1996-06-18 2000-12-26 Konami Co., Ltd. Video game system with vertical array of cursor images
US6563512B1 (en) * 1996-06-25 2003-05-13 Fujitsu Limited Object editing method, object editing system and computer memory product
US6144366A (en) * 1996-10-18 2000-11-07 Kabushiki Kaisha Toshiba Method and apparatus for generating information input using reflected light image of target object
US6128003A (en) * 1996-12-20 2000-10-03 Hitachi, Ltd. Hand gesture recognition system and method
US6266061B1 (en) * 1997-01-22 2001-07-24 Kabushiki Kaisha Toshiba User interface apparatus and operation range presenting method
US6118888A (en) * 1997-02-28 2000-09-12 Kabushiki Kaisha Toshiba Multi-modal interface apparatus and method
US6345111B1 (en) 1997-02-28 2002-02-05 Kabushiki Kaisha Toshiba Multi-modal interface apparatus and method
WO1999042946A3 (en) * 1998-02-18 1999-10-28 Geovector Corp Apparatus and methods for presentation of information relating to objects being addressed
WO1999042946A2 (en) * 1998-02-18 1999-08-26 Geovector Corporation Apparatus and methods for presentation of information relating to objects being addressed
US6154558A (en) * 1998-04-22 2000-11-28 Hsieh; Kuan-Hong Intention identification method
US20060013440A1 (en) * 1998-08-10 2006-01-19 Cohen Charles J Gesture-controlled interfaces for self-service machines and other applications
US7668340B2 (en) 1998-08-10 2010-02-23 Cybernet Systems Corporation Gesture-controlled interfaces for self-service machines and other applications
US20040161132A1 (en) * 1998-08-10 2004-08-19 Cohen Charles J. Gesture-controlled interfaces for self-service machines and other applications
US7684592B2 (en) 1998-08-10 2010-03-23 Cybernet Systems Corporation Realtime object tracking system
US6681031B2 (en) 1998-08-10 2004-01-20 Cybernet Systems Corporation Gesture-controlled interfaces for self-service machines and other applications
US7460690B2 (en) 1998-08-10 2008-12-02 Cybernet Systems Corporation Gesture-controlled interfaces for self-service machines and other applications
US6950534B2 (en) 1998-08-10 2005-09-27 Cybernet Systems Corporation Gesture-controlled interfaces for self-service machines and other applications
US20090116692A1 (en) * 1998-08-10 2009-05-07 Paul George V Realtime object tracking system
US9304593B2 (en) 1998-08-10 2016-04-05 Cybernet Systems Corporation Behavior recognition system
US6396476B1 (en) * 1998-12-01 2002-05-28 Intel Corporation Synthesizing computer input events
US6538649B2 (en) 1998-12-01 2003-03-25 Intel Corporation Computer vision control variable transformation
US6363160B1 (en) 1999-01-22 2002-03-26 Intel Corporation Interface using pattern recognition and tracking
US6476834B1 (en) * 1999-05-28 2002-11-05 International Business Machines Corporation Dynamic creation of selectable items on surfaces
EP2328081A1 (en) * 1999-06-10 2011-06-01 Dassault Systèmes Knowledge-based polymorph undockable toolbar
US20110016434A1 (en) * 1999-06-10 2011-01-20 Dassault Systemes Knowledge-based polymorph undockable toolbar
EP2328079A1 (en) * 1999-06-10 2011-06-01 Dassault Systèmes Knowledge-based polymorph undockable toolbar
US20030187532A1 (en) * 1999-06-10 2003-10-02 Dassault Systemes Knowledge-based polymorph undockable toolbar
US9454287B2 (en) 1999-06-10 2016-09-27 Dassault Systemes Knowledge-based polymorph undockable toolbar
US7823085B2 (en) 1999-06-10 2010-10-26 Dassault Systemes Knowledge-based polymorph undockable toolbar
EP1059581A3 (en) * 1999-06-10 2004-04-28 Dassault Systèmes Knowledge-based polymorph undockable toolbar
US6647131B1 (en) 1999-08-27 2003-11-11 Intel Corporation Motion detection using normal optical flow
US7134095B1 (en) 1999-10-20 2006-11-07 Gateway, Inc. Simulated three-dimensional navigational menu system
USRE43084E1 (en) 1999-10-29 2012-01-10 Smart Technologies Ulc Method and apparatus for inputting information including coordinate data
US7342574B1 (en) * 1999-10-29 2008-03-11 Ricoh Company, Ltd. Method and apparatus for inputting information including coordinate data
US6654483B1 (en) 1999-12-22 2003-11-25 Intel Corporation Motion detection using normal optical flow
US6690393B2 (en) * 1999-12-24 2004-02-10 Koninklijke Philips Electronics N.V. 3D environment labelling
USRE42794E1 (en) 1999-12-27 2011-10-04 Smart Technologies Ulc Information-inputting device inputting contact point of object on recording surfaces as information
US6522292B1 (en) 2000-02-23 2003-02-18 Geovector Corp. Information systems having position measuring capacity
US6938218B1 (en) 2000-04-28 2005-08-30 James Nolen Method and apparatus for three dimensional internet and computer file interface
US8203535B2 (en) 2000-07-05 2012-06-19 Smart Technologies Ulc Passive touch system and method of detecting user input
US8055022B2 (en) 2000-07-05 2011-11-08 Smart Technologies Ulc Passive touch system and method of detecting user input
US8378986B2 (en) 2000-07-05 2013-02-19 Smart Technologies Ulc Passive touch system and method of detecting user input
US7031875B2 (en) 2001-01-24 2006-04-18 Geo Vector Corporation Pointing systems for addressing objects
US7705830B2 (en) * 2001-02-10 2010-04-27 Apple Inc. System and method for packing multitouch gestures onto a hand
US20040246234A1 (en) * 2001-03-01 2004-12-09 Luis Serra Display apparatus
US20040125076A1 (en) * 2001-06-08 2004-07-01 David Green Method and apparatus for human interface with a computer
US20030026170A1 (en) * 2001-07-31 2003-02-06 Yang Jong H. Same-hand control of a multi-function device
WO2003012615A2 (en) * 2001-07-31 2003-02-13 The Timberland Company Hand worn input device
WO2003012615A3 (en) * 2001-07-31 2003-12-04 Timberland Co Hand worn input device
US6991364B2 (en) 2001-07-31 2006-01-31 The Timberland Company Same-hand control of a multi-function device
US20030122839A1 (en) * 2001-12-26 2003-07-03 Eastman Kodak Company Image format including affective information
US20030132909A1 (en) * 2002-01-11 2003-07-17 International Business Machines Corporation Pointing device attribute variance based on distance/time ratio
US20040001112A1 (en) * 2002-01-25 2004-01-01 Silicon Graphics, Inc. Volume management system for volumetric displays
US7839400B2 (en) 2002-01-25 2010-11-23 Autodesk, Inc. Volume management system for volumetric displays
US8707216B2 (en) 2002-02-07 2014-04-22 Microsoft Corporation Controlling objects via gesturing
US20110004329A1 (en) * 2002-02-07 2011-01-06 Microsoft Corporation Controlling electronic components in a computing environment
US9454244B2 (en) 2002-02-07 2016-09-27 Microsoft Technology Licensing, Llc Recognizing a movement of a pointing device
US20080192007A1 (en) * 2002-02-07 2008-08-14 Microsoft Corporation Determining a position of a pointing device
US8456419B2 (en) 2002-02-07 2013-06-04 Microsoft Corporation Determining a position of a pointing device
US20040136062A1 (en) * 2002-05-13 2004-07-15 Avner Farkash Portable periscope
US7477243B2 (en) 2002-05-31 2009-01-13 Eit Co., Ltd. Apparatus for controlling the shift of virtual space and method and program for controlling same
EP1369822A2 (en) * 2002-05-31 2003-12-10 CAD Center Corporation Apparatus and method for controlling the shift of the viewpoint in a virtual space
US20040051709A1 (en) * 2002-05-31 2004-03-18 Eit Co., Ltd. Apparatus for controlling the shift of virtual space and method and program for controlling same
EP1369822A3 (en) * 2002-05-31 2004-03-31 CAD Center Corporation Apparatus and method for controlling the shift of the viewpoint in a virtual space
US6917360B2 (en) * 2002-06-21 2005-07-12 Schlumberger Technology Corporation System and method for adaptively labeling multi-dimensional images
US20030234782A1 (en) * 2002-06-21 2003-12-25 Igor Terentyev System and method for adaptively labeling multi-dimensional images
US20060125822A1 (en) * 2002-06-28 2006-06-15 Alias Systems Corp. Volume management system for volumetric displays
US7986318B2 (en) * 2002-06-28 2011-07-26 Autodesk, Inc. Volume management system for volumetric displays
US8228304B2 (en) 2002-11-15 2012-07-24 Smart Technologies Ulc Size/scale orientation determination of a pointer in a camera-based touch system
US8466885B2 (en) 2003-02-14 2013-06-18 Next Holdings Limited Touch screen signal processing
US8456447B2 (en) 2003-02-14 2013-06-04 Next Holdings Limited Touch screen signal processing
US8289299B2 (en) 2003-02-14 2012-10-16 Next Holdings Limited Touch screen signal processing
US8508508B2 (en) 2003-02-14 2013-08-13 Next Holdings Limited Touch screen signal processing with single-point calibration
US8456451B2 (en) 2003-03-11 2013-06-04 Smart Technologies Ulc System and method for differentiating between pointers used to contact touch surface
US7665041B2 (en) 2003-03-25 2010-02-16 Microsoft Corporation Architecture for controlling a computer using hand gestures
US9652042B2 (en) 2003-03-25 2017-05-16 Microsoft Technology Licensing, Llc Architecture for controlling a computer using hand gestures
US20100146455A1 (en) * 2003-03-25 2010-06-10 Microsoft Corporation Architecture For Controlling A Computer Using Hand Gestures
US20040189720A1 (en) * 2003-03-25 2004-09-30 Wilson Andrew D. Architecture for controlling a computer using hand gestures
US8745541B2 (en) * 2003-03-25 2014-06-03 Microsoft Corporation Architecture for controlling a computer using hand gestures
US20040193413A1 (en) * 2003-03-25 2004-09-30 Wilson Andrew D. Architecture for controlling a computer using hand gestures
US20040219961A1 (en) * 2003-04-08 2004-11-04 Ellenby Thomas William Computer games having variable execution dependence with respect to spatial properties of a mobile unit.
US20060238490A1 (en) * 2003-05-15 2006-10-26 Qinetiq Limited Non contact human-computer interface
US8456418B2 (en) 2003-10-09 2013-06-04 Smart Technologies Ulc Apparatus for determining the location of a pointer within a region of interest
US8230366B2 (en) * 2003-10-23 2012-07-24 Apple Inc. Dynamically changing cursor for user interface
US20050088410A1 (en) * 2003-10-23 2005-04-28 Apple Computer, Inc. Dynamically changing cursor for user interface
US8089462B2 (en) 2004-01-02 2012-01-03 Smart Technologies Ulc Pointer tracking across multiple overlapping coordinate input sub-regions defining a generally contiguous input region
US20060067576A1 (en) * 2004-03-17 2006-03-30 James Marggraff Providing a user interface having interactive elements on a writable surface
US20050226469A1 (en) * 2004-04-07 2005-10-13 Jong-Hwan Ho Method for displaying finger images in wireless telecommunication terminal
US8274496B2 (en) 2004-04-29 2012-09-25 Smart Technologies Ulc Dual mode touch systems
US8149221B2 (en) 2004-05-07 2012-04-03 Next Holdings Limited Touch panel display system with illumination and detection provided from a single edge
US8120596B2 (en) 2004-05-21 2012-02-21 Smart Technologies Ulc Tiled touch system
US7917868B2 (en) * 2004-08-07 2011-03-29 Samsung Electronics Co., Ltd. Three-dimensional motion graphic user interface and method and apparatus for providing the same
US20060031874A1 (en) * 2004-08-07 2006-02-09 Samsung Electronics Co., Ltd. Three-dimensional motion graphic user interface and method and apparatus for providing the same
US20060190812A1 (en) * 2005-02-22 2006-08-24 Geovector Corporation Imaging systems including hyperlink associations
US20060202974A1 (en) * 2005-03-10 2006-09-14 Jeffrey Thielman Surface
US20060209016A1 (en) * 2005-03-17 2006-09-21 Microsoft Corporation Computer interaction based upon a currently active input device
US7802202B2 (en) * 2005-03-17 2010-09-21 Microsoft Corporation Computer interaction based upon a currently active input device
US9354771B2 (en) * 2006-01-30 2016-05-31 Microsoft Technology Licensing, Llc Controlling application windows in an operating system
US20150089445A1 (en) * 2006-01-30 2015-03-26 Microsoft Corporation Controlling Application Windows In An Operating System
US8910066B2 (en) * 2006-01-30 2014-12-09 Microsoft Corporation Controlling application windows in an operating system
US20120235946A1 (en) * 2006-01-30 2012-09-20 Microsoft Corporation Controlling application windows in an operating system
US7549132B2 (en) * 2006-04-28 2009-06-16 International Business Machines Corporation Presenting identifiers and states of processes in a stacked cursor
US20070256032A1 (en) * 2006-04-28 2007-11-01 Petri John E Presenting identifiers and states of processes in a stacked cursor
US20090286653A1 (en) * 2006-06-21 2009-11-19 Wiber Laurent Remote control device for an electronic apparatus in particular for performing a physical exercise
US8264504B2 (en) 2006-06-26 2012-09-11 University Of Southern California Seamlessly overlaying 2D images in 3D model
US20080024484A1 (en) * 2006-06-26 2008-01-31 University Of Southern California Seamless Image Integration Into 3D Models
US8026929B2 (en) * 2006-06-26 2011-09-27 University Of Southern California Seamlessly overlaying 2D images in 3D model
US9442607B2 (en) 2006-12-04 2016-09-13 Smart Technologies Inc. Interactive input system and method
US8006183B1 (en) * 2006-12-08 2011-08-23 Trading Technologies International Inc. System and method for using a curser to convey information
US8732578B2 (en) 2006-12-08 2014-05-20 Trading Technologies International, Inc. System and method for using a cursor to convey information
US20110239165A1 (en) * 2006-12-08 2011-09-29 Trading Technologies International Inc. System and Method for Using a Cursor to Convey Information
US8115753B2 (en) 2007-04-11 2012-02-14 Next Holdings Limited Touch screen system with hover and click input methods
US8094137B2 (en) 2007-07-23 2012-01-10 Smart Technologies Ulc System and method of detecting contact on a display
US8432377B2 (en) 2007-08-30 2013-04-30 Next Holdings Limited Optical touchscreen with improved illumination
US8384693B2 (en) 2007-08-30 2013-02-26 Next Holdings Limited Low profile touch panel systems
US8405637B2 (en) 2008-01-07 2013-03-26 Next Holdings Limited Optical position sensing system and optical position sensor assembly with convex imaging window
US8405636B2 (en) 2008-01-07 2013-03-26 Next Holdings Limited Optical position sensing system and optical position sensor assembly
US20090245691A1 (en) * 2008-03-31 2009-10-01 University Of Southern California Estimating pose of photographic images in 3d earth model using human assistance
US9619106B2 (en) 2008-04-24 2017-04-11 Pixar Methods and apparatus for simultaneous user inputs for three-dimensional animation
US8836646B1 (en) 2008-04-24 2014-09-16 Pixar Methods and apparatus for simultaneous user inputs for three-dimensional animation
US8799821B1 (en) * 2008-04-24 2014-08-05 Pixar Method and apparatus for user inputs for three-dimensional animation
US8902193B2 (en) 2008-05-09 2014-12-02 Smart Technologies Ulc Interactive input system and bezel therefor
US8339378B2 (en) 2008-11-05 2012-12-25 Smart Technologies Ulc Interactive input system with multi-angle reflector
US8692768B2 (en) 2009-07-10 2014-04-08 Smart Technologies Ulc Interactive input system
US20110083106A1 (en) * 2009-10-05 2011-04-07 Seiko Epson Corporation Image input system
US20110083112A1 (en) * 2009-10-05 2011-04-07 Takashi Matsubara Input apparatus
US10007393B2 (en) * 2010-01-19 2018-06-26 Apple Inc. 3D view of file structure
US20110179368A1 (en) * 2010-01-19 2011-07-21 King Nicholas V 3D View Of File Structure
US20110225536A1 (en) * 2010-02-28 2011-09-15 Osterhout Group, Inc. Sliding keyboard input control in an augmented reality eyepiece
US9223134B2 (en) 2010-02-28 2015-12-29 Microsoft Technology Licensing, Llc Optical imperfections in a light transmissive illumination system for see-through near-eye display glasses
US9366862B2 (en) 2010-02-28 2016-06-14 Microsoft Technology Licensing, Llc System and method for delivering content to a group of see-through near eye display eyepieces
US9759917B2 (en) 2010-02-28 2017-09-12 Microsoft Technology Licensing, Llc AR glasses with event and sensor triggered AR eyepiece interface to external devices
US9875406B2 (en) 2010-02-28 2018-01-23 Microsoft Technology Licensing, Llc Adjustable extension for temple arm
US9229227B2 (en) 2010-02-28 2016-01-05 Microsoft Technology Licensing, Llc See-through near-eye display glasses with a light transmissive wedge shaped illumination system
US8814691B2 (en) 2010-02-28 2014-08-26 Microsoft Corporation System and method for social networking gaming with an augmented reality
US8467133B2 (en) 2010-02-28 2013-06-18 Osterhout Group, Inc. See-through display with an optical assembly including a wedge-shaped illumination system
US8488246B2 (en) 2010-02-28 2013-07-16 Osterhout Group, Inc. See-through near-eye display glasses including a curved polarizing film in the image source, a partially reflective, partially transmitting optical element and an optically flat film
US8482859B2 (en) 2010-02-28 2013-07-09 Osterhout Group, Inc. See-through near-eye display glasses wherein image light is transmitted to and reflected from an optically flat film
US9329689B2 (en) 2010-02-28 2016-05-03 Microsoft Technology Licensing, Llc Method and apparatus for biometric data capture
US9341843B2 (en) 2010-02-28 2016-05-17 Microsoft Technology Licensing, Llc See-through near-eye display glasses with a small scale image source
US20110227812A1 (en) * 2010-02-28 2011-09-22 Osterhout Group, Inc. Head nod detection and control in an augmented reality eyepiece
US9091851B2 (en) 2010-02-28 2015-07-28 Microsoft Technology Licensing, Llc Light control in head mounted displays
US9097890B2 (en) 2010-02-28 2015-08-04 Microsoft Technology Licensing, Llc Grating in a light transmissive illumination system for see-through near-eye display glasses
US8477425B2 (en) 2010-02-28 2013-07-02 Osterhout Group, Inc. See-through near-eye display glasses including a partially reflective, partially transmitting optical element
US8472120B2 (en) 2010-02-28 2013-06-25 Osterhout Group, Inc. See-through near-eye display glasses with a small scale image source
US9097891B2 (en) 2010-02-28 2015-08-04 Microsoft Technology Licensing, Llc See-through near-eye display glasses including an auto-brightness control for the display brightness based on the brightness in the environment
US20110231757A1 (en) * 2010-02-28 2011-09-22 Osterhout Group, Inc. Tactile control in an augmented reality eyepiece
US9129295B2 (en) 2010-02-28 2015-09-08 Microsoft Technology Licensing, Llc See-through near-eye display glasses with a fast response photochromic film system for quick transition from dark to clear
US9134534B2 (en) 2010-02-28 2015-09-15 Microsoft Technology Licensing, Llc See-through near-eye display glasses including a modular image source
US9182596B2 (en) 2010-02-28 2015-11-10 Microsoft Technology Licensing, Llc See-through near-eye display glasses with the optical assembly including absorptive polarizers or anti-reflective coatings to reduce stray light
US20110221669A1 (en) * 2010-02-28 2011-09-15 Osterhout Group, Inc. Gesture control in an augmented reality eyepiece
US9285589B2 (en) 2010-02-28 2016-03-15 Microsoft Technology Licensing, Llc AR glasses with event and sensor triggered control of AR eyepiece applications
US9128281B2 (en) 2010-09-14 2015-09-08 Microsoft Technology Licensing, Llc Eyepiece with uniformly illuminated reflective display
US20130311952A1 (en) * 2011-03-09 2013-11-21 Maiko Nakagawa Image processing apparatus and method, and program
US9223405B2 (en) 2011-03-16 2015-12-29 Electronics And Telecommunications Research Institute Apparatus and method for inputting information based on events
US20120272180A1 (en) * 2011-04-20 2012-10-25 Nokia Corporation Method and apparatus for providing content flipping based on a scrolling operation
US8620113B2 (en) 2011-04-25 2013-12-31 Microsoft Corporation Laser diode modes
US9372544B2 (en) 2011-05-31 2016-06-21 Microsoft Technology Licensing, Llc Gesture recognition techniques
US8760395B2 (en) 2011-05-31 2014-06-24 Microsoft Corporation Gesture recognition techniques
USD741361S1 (en) * 2011-06-07 2015-10-20 Cornell D. M. Judge Cornish Portion of display screen with an icon
KR20140040246A (en) * 2011-06-21 2014-04-02 퀄컴 인코포레이티드 Gesture-controlled technique to expand interaction radius in computer vision applications
CN103620526B (en) * 2011-06-21 2017-07-21 高通股份有限公司 Extended interaction radius gesture control technologies in the computer vision applications
WO2012177322A1 (en) * 2011-06-21 2012-12-27 Qualcomm Incorporated Gesture-controlled technique to expand interaction radius in computer vision applications
CN103620526A (en) * 2011-06-21 2014-03-05 高通股份有限公司 Gesture-controlled technique to expand interaction radius in computer vision applications
US8635637B2 (en) 2011-12-02 2014-01-21 Microsoft Corporation User interface presenting an animated avatar performing a media reaction
US9154837B2 (en) 2011-12-02 2015-10-06 Microsoft Technology Licensing, Llc User interface presenting an animated avatar performing a media reaction
US9100685B2 (en) 2011-12-09 2015-08-04 Microsoft Technology Licensing, Llc Determining audience state or interest using passive sensor data
US9628844B2 (en) 2011-12-09 2017-04-18 Microsoft Technology Licensing, Llc Determining audience state or interest using passive sensor data
US9596643B2 (en) 2011-12-16 2017-03-14 Microsoft Technology Licensing, Llc Providing a user interface experience based on inferred vehicle state
US8898687B2 (en) 2012-04-04 2014-11-25 Microsoft Corporation Controlling a media program based on a media reaction
US9788032B2 (en) 2012-05-04 2017-10-10 Microsoft Technology Licensing, Llc Determining a future portion of a currently presented media program
US8959541B2 (en) 2012-05-04 2015-02-17 Microsoft Technology Licensing, Llc Determining a future portion of a currently presented media program
US9098516B2 (en) * 2012-07-18 2015-08-04 DS Zodiac, Inc. Multi-dimensional file system
US20140026103A1 (en) * 2012-07-18 2014-01-23 DS Zodiac, Inc. Multi-dimensional file system
USD822036S1 (en) * 2014-04-16 2018-07-03 Regents Of The University Of Minnesota Display screen or portion thereof with graphical user interface
DE102014106960A1 (en) * 2014-05-16 2015-11-19 Faindu Gmbh A method for displaying a virtual interaction on at least one screen and input device, system and method for a virtual application by means of a computing unit
DE102014107220A1 (en) * 2014-05-22 2015-11-26 Atlas Elektronik Gmbh Input device, computer or the operating system as well as vehicle

Also Published As

Publication number Publication date Type
JPH0863326A (en) 1996-03-08 application

Similar Documents

Publication Publication Date Title
US9104271B1 (en) Gloved human-machine interface
US5594469A (en) Hand gesture machine control system
US6320559B1 (en) Head-mounted image display device and data processing apparatus including the same
US5668573A (en) Management method for a man-machine interaction system
US20090172606A1 (en) Method and apparatus for two-handed computer user interface with gesture recognition
US6714213B1 (en) System and method for providing interactive haptic collision detection
US20040113885A1 (en) New input devices for augmented reality applications
US20050162384A1 (en) Pointing device, method for displaying point image, and program therefor
US5598187A (en) Spatial motion pattern input system and input method
US5914720A (en) Method of using multiple perceptual channels to increase user absorption of an N-dimensional presentation environment
Chen et al. A study in interactive 3-D rotation using 2-D control devices
US5680524A (en) Synthetic environment employing a craft for providing user perspective reference
US5982353A (en) Virtual body modeling apparatus having dual-mode motion processing
US6437794B1 (en) Interactive image generation method and apparatus utilizing a determination of the visual point position of an operator
Brooks Jr et al. Research directions in virtual environments
US20030023347A1 (en) Authoring system and authoring method, and storage medium
US6853881B2 (en) Robot information processing system
Ong et al. Augmented reality applications in manufacturing: a survey
US20120304059A1 (en) Interactive Build Instructions
Gobbetti et al. Virtual reality: Past, present, and future
US20050156904A1 (en) Input control apparatus and method for responding to input
US20120122062A1 (en) Reconfigurable platform management apparatus for virtual reality-based training simulator
GB2281838A (en) Input for a virtual reality system
US20140201666A1 (en) Dynamic, free-space user interactions for machine control
CN102117117A (en) System and method for control through identifying user posture by image extraction device

Legal Events

Date Code Title Description
AS Assignment

Owner name: HITACHI, LTD., JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:OOHARA, SHUNICHI;EJIRI, MASAKAZU;NEMOTO, YASUHIRO;AND OTHERS;REEL/FRAME:009283/0351

Effective date: 19950720

FPAY Fee payment

Year of fee payment: 4

REMI Maintenance fee reminder mailed
LAPS Lapse for failure to pay maintenance fees
FP Expired due to failure to pay maintenance fee

Effective date: 20060901