US20120062603A1 - Information Processing Apparatus, Information Processing Method, and Program Therefor - Google Patents
Information Processing Apparatus, Information Processing Method, and Program Therefor Download PDFInfo
- Publication number
- US20120062603A1 US20120062603A1 US13/297,777 US201113297777A US2012062603A1 US 20120062603 A1 US20120062603 A1 US 20120062603A1 US 201113297777 A US201113297777 A US 201113297777A US 2012062603 A1 US2012062603 A1 US 2012062603A1
- Authority
- US
- United States
- Prior art keywords
- indicator
- display device
- pointer
- input
- display
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/041—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
- G06F3/0416—Control or interface arrangements specially adapted for digitisers
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/041—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
- G06F3/0412—Digitisers structurally integrated in a display
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
- G06F3/04883—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
- G06F3/04886—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures by partitioning the display area of the touch-screen or the surface of the digitising tablet into independently controllable areas, e.g. virtual keyboards or menus
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2203/00—Indexing scheme relating to G06F3/00 - G06F3/048
- G06F2203/041—Indexing scheme relating to G06F3/041 - G06F3/045
- G06F2203/04104—Multi-touch detection in digitiser, i.e. details about the simultaneous detection of a plurality of touching locations, e.g. multiple fingers or pen and finger
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2203/00—Indexing scheme relating to G06F3/00 - G06F3/048
- G06F2203/041—Indexing scheme relating to G06F3/041 - G06F3/045
- G06F2203/04105—Pressure sensors for measuring the pressure or force exerted on the touch surface without providing the touch position
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2203/00—Indexing scheme relating to G06F3/00 - G06F3/048
- G06F2203/041—Indexing scheme relating to G06F3/041 - G06F3/045
- G06F2203/04108—Touchless 2D- digitiser, i.e. digitiser detecting the X/Y position of the input means, finger or stylus, also when it does not touch, but is proximate to the digitiser's interaction surface without distance measurement in the Z direction
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2203/00—Indexing scheme relating to G06F3/00 - G06F3/048
- G06F2203/048—Indexing scheme relating to G06F3/048
- G06F2203/04803—Split screen, i.e. subdividing the display area or the window area into separate subareas
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G5/00—Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
- G09G5/14—Display of multiple viewports
Definitions
- the present disclosure relates to an information processing apparatus and an information processing method, which enable a multi-touch operation to be made, and a program therefor.
- Patent Document 1 discloses a portable electronic apparatus that enables, through a multi-touch operation by a user, an enlargement operation and the like with respect to an object displayed on a display screen to be made (see paragraph [0018], FIG. 1 and the like in Patent Document 1).
- the present disclosure is directed towards an apparatus comprising a memory storing instructions and a control unit executing the instructions to send signals to display first and second indicators on a display device.
- the control unit further executes the instructions to receive first and second user inputs and, in response to the received inputs, to send signals to change a display state of the first indicator according to the first input.
- the control unit further executes the instructions to send signals to change a display state of the second indicator according to the second input and initiate an operation to be performed based on a combination of the first and second inputs.
- the present disclosure is directed toward a method comprising sending signals to display first and second indicators on a display device.
- the method further comprises receiving first and second user inputs and, in response to the received inputs, sending signals to change a display state of the first indicator according to the first input.
- the method further comprises sending signals to change a display state of the second indicator according to the second input.
- the method further comprises initiating an operation to be performed based on a combination of the first and second inputs.
- the present disclosure is directed toward a tangibly embodied non-transitory computer-readable medium storing instructions which, when executed by a processor, perform a method comprising sending signals to display first and second indicators on a display device.
- the method further comprises receiving first and second user inputs.
- the method further comprises, in response to the received inputs, sending signals to change a display state of the first indicator according to the first input and sending signals to change a display state of the second indicator according to the second input.
- the method further comprises initiating an operation to be performed based on a combination of the first and second inputs.
- FIG. 1 is a view schematically showing an outer appearance of an information processing apparatus according to an embodiment of the present disclosure
- FIG. 2 is a block diagram showing a configuration example of the information processing apparatus shown in FIG. 1 ;
- FIG. 3 is a view schematically showing the information processing apparatus shown in FIG. 1 and a display apparatus connected thereto;
- FIG. 4 is a flowchart showing an operation of the information processing apparatus shown in FIG. 1 ;
- FIG. 5 is a view for describing an example of a method of determining a correspondence relation between an operator and a cursor pointer shown in FIG. 3 ;
- FIGS. 6A and 6B are views for describing touch operations for controlling a left pointer and a right pointer and an action assigned to the touch operations;
- FIGS. 7A and 7B are views for describing touch operations for controlling the left pointer and the right pointer and an action assigned to the touch operations;
- FIG. 8 is a view for describing touch operations for controlling the left pointer and the right pointer and an action assigned to the touch operations;
- FIG. 9 is a view for describing touch operations for controlling the left pointer and the right pointer and an action assigned to the touch operations
- FIG. 10 is a view for describing touch operations for controlling the left pointer and the right pointer and an action assigned to the touch operations;
- FIG. 11 is a view for describing touch operations for controlling the left pointer and the right pointer and an action assigned to the touch operations;
- FIGS. 12A and 12B are views for describing touch operations for controlling the left pointer and the right pointer and an action assigned to the touch operations;
- FIG. 13 is a view for describing touch operations for controlling the left pointer and the right pointer and an action assigned to the touch operations;
- FIGS. 14A and 14B are schematic views showing a modification of the information processing apparatus shown in FIG. 1 ;
- FIGS. 15A and 15B are schematic views showing a modification of the information processing apparatus shown in FIG. 1 ;
- FIG. 16 is a view schematically showing a modification of variations of the present disclosure.
- FIG. 17 is a schematic view showing another example of the information processing apparatus shown in FIG. 1 .
- FIG. 1 is a view schematically showing an outer appearance of an information processing apparatus according to an embodiment of the present disclosure.
- FIG. 2 is a block diagram showing a configuration example of the information processing apparatus shown in FIG. 1 .
- the information processing apparatus 100 includes a casing 1 that is held by a user 30 and a touch pad 2 provided in front of the casing 1 to serve as an input unit.
- the touch pad 2 according to this embodiment is capable of receiving various types of multi-touch operations with one or more operators 3 .
- the casing 1 is formed of, for example, metal such as aluminum and stainless steel, plastic, or other various materials.
- the casing 1 has such a size such that user 30 is capable of carrying the apparatus 100 .
- the size of the casing 1 is about 15 cm in a longitudinal direction, about 10 cm in a lateral direction, and about 1 cm in thickness.
- the size of the casing 1 can be appropriately set.
- the information processing apparatus 100 including the casing 1 having a size of 30 cm or more in the longitudinal and lateral directions may be used while placed on a desk.
- thumbs of the user 30 are used as the plurality of operators 3 serving to perform touch operations with respect to the touch pad 2 .
- other fingers may also be used (see FIG. 17 ).
- a stylus, pen, joystick, or other suitable apparatus may be used as the operators 3 .
- the touch pad 2 includes a position detector 101 and a pressure sensor 102 .
- the position detector 101 includes a sensor unit and an arithmetic circuit unit (not shown).
- the sensor unit When the operator 3 comes close to or in contact with the touch pad 2 , the sensor unit outputs a proximity or contact signal to the arithmetic circuit unit.
- the arithmetic circuit unit generates, based on the output proximity or contact signal, information on a proximity or contact position of the operator 3 , for example. Alternatively, information on a movement direction, movement amount, movement speed, acceleration speed, or the like of the operator 3 may be generated. In addition, information on a contact area, contact angle, or the like of the operator 3 may be generated.
- touch operation refers to various touch operations performed when the operator 3 comes close to or in contact with the touch pad 2 . It should be understood that, in certain variations, the operator 3 need not touch or make contact with touch pad 2 in order to perform a touch operation. Further, the proximity and contact positions are collectively referred to as touch position of the operator 3 .
- the touch pad 2 is a touch pad of a so-called multi-touch type.
- the user 30 can perform a pinch operation with two fingers.
- the pinch operation is, for example, an operation of bringing the two fingers into contact with the touch pad 2 at the same time and opening or closing the two fingers or the like while keeping the in-contact state.
- the operation of opening the two fingers or the like held in contact with the touch pad 2 is referred to as pinch-out operation.
- the operation of closing the two fingers or the like held in contact with the touch pad 2 is referred to as pinch-in operation.
- the pressure sensor 102 serves to detect whether or not the touch pad 2 is pressed by the operator 3 . For example, when the pressure sensor 102 detects the pressing force above a threshold level, it is determined that the touch pad 2 is pressed. A method of determining existence, absence of pressure, change in pressure, or amount of pressure by the pressure sensor 102 , the structure of the pressure sensor 102 , a method of detecting the pressing force, and the like can be appropriately set.
- the touch pad 2 of a capacitive type is used.
- a touch pad of another type such as, for example, a resistive touch pad, a surface acoustic wave touch pad, and an infrared touch pad may be used instead.
- a “touch operation” may be detected by any suitable method and/or by using any suitable apparatus.
- various optical detection apparatuses e.g., cameras, electric eyes or other light-sensing devices
- detection apparatuses that detect infrared radiation or other radiation outside of the visible spectrum may be used to sense a “touch operation” or any of the other operations discussed herein.
- These devices may be used in conjunction with other equipment that may, for example, aid in sensing the location of the user's fingers, stylus or other user control apparatus (e.g., gloves may be used to enhance the visual signature of the users fingers, for example).
- the information processing apparatus 100 includes a control unit such as a CPU (Central Processing Unit) 103 , a RAM (Random Access Memory) 104 , a storage 105 , and a display interface 106 .
- a control unit such as a CPU (Central Processing Unit) 103 , a RAM (Random Access Memory) 104 , a storage 105 , and a display interface 106 .
- CPU Central Processing Unit
- RAM Random Access Memory
- the storage 105 is a non-volatile storage device.
- the storage 105 may be an HDD (Hard Disk Drive), a flash memory, or other solid-state memories.
- a part or all of the RAM 104 and the storage 105 corresponds to a storage means according to this embodiment.
- the information processing apparatus 100 may include a ROM (Read Only Memory) (not shown), and thus the storage means may include the ROM.
- the display interface 106 serves to connect the information processing apparatus 100 to a display apparatus 40 (see FIG. 3 ) including a display screen 41 . Via the display interface 106 , image data of still images, moving images, and the like, control signals for controlling the operation of the display apparatus 40 , and the like are output. The image data, the control signals, and the like are output to the display apparatus 40 in a wireless or wired manner.
- the information processing apparatus 100 may be provided with a drive unit and a communication unit (not shown).
- the drive unit may be, for example, a device capable of driving a removable recording medium such as an optical recording medium, a floppy (registered trademark) disc, a magnetic recording tape, and a flash memory.
- the communication unit may be, for example, a modem, a router, and other communication apparatuses for communicating with other devices that are capable of connecting to a LAN (Local Area Network), a WAN (Wide Area Network), and the like.
- the communication unit may communicate in a wired or wireless manner.
- Data processing by the information processing apparatus 100 is realized by software stored in the storage 105 or the like in cooperation with hardware resources of the information processing apparatus 100 .
- the CPU 103 loading a program of instructions stored in the storage 105 or the like into the RAM 104 and executing it, various types of data processing are realized.
- the CPU 103 functions as a first display processor, a second display processor, and an action executing unit.
- the RAM 104 may include a video RAM (VRAM) (not shown) for displaying images.
- VRAM video RAM
- FIG. 3 is a view schematically showing the information processing apparatus 100 and the display apparatus 40 connected thereto.
- a plurality of cursor pointers 43 are displayed on the display screen 41 .
- the plurality of cursor pointers 43 are displayed as the CPU 103 serving as the second display processor outputs display signals of the cursor pointers 43 to the display apparatus 40 .
- a cursor pointer 43 a and a cursor pointer 43 b may be displayed on the display screen 41 .
- the cursor pointer 43 a is, for example, operated mainly by the left operator 3 a being a left finger or the like.
- the cursor pointer 43 b is, for example, operated mainly by the right operator 3 b being a right finger or the like.
- each cursor pointer which respectively control those two cursor pointers 43 a and 43 b , may be input to the touch pad 2 by the left operator 3 a and the right operator 3 b .
- the cursor pointer 43 a is referred to as the left pointer 43 a
- the cursor pointer 43 b is referred to as the right pointer 43 b.
- the left pointer 43 a and the right pointer 43 b may be displayed in different colors, or exhibit different shading as shown in FIG. 3 , for example, so that the pointers can be discriminated from each other. With this, the pointers 43 a and 43 b can be operated with high operability. In order to discriminate the pointers 43 a and 43 b from each other, the shape, size, or the like of the pointers 43 a and 43 b may be appropriately set. However, the pointers 43 a and 43 b may be displayed in an indiscriminable manner.
- a coordinate system may be set with a center O of the touch pad 2 being a reference. Coordinate information including information of positional coordinates of the operators 3 a and 3 b in that coordinate system is output to the display apparatus 40 .
- a coordinate system may be set with a center O′ of the display screen 41 being a reference.
- the pointers 43 a and 43 b are displayed at positions corresponding to touch positions 5 a and 5 b of the operators 3 , respectively.
- the coordinate systems may be appropriately set in the touch pad 2 and the display screen 41 .
- FIG. 4 is a flowchart showing the operation. It may be determined whether or not a contact or proximity of the operator 3 with respect to the touch pad 2 is detected as an initial input operation (Step 101 ). When the contact or the like with the touch pad 2 is detected (“Yes” in Step 101 ), a correspondence relation between the operator 3 and the cursor pointer 43 displayed on the display screen 41 is determined.
- FIG. 5 is a view for describing an example of a method of determining the correspondence relation between the operator 3 and the cursor pointer 43 .
- the correspondence relation is determined.
- the touch pad 2 may be divided into the left area 2 a and the right area 2 b .
- the operator 3 and the left pointer 43 a may be caused to correspond to each other.
- the operator 3 functions as the left operator 3 a .
- the left pointer 43 a is displayed at a position on the display screen 41 , which corresponds to the touch position 5 of the left operator 3 a .
- the operator 3 and the right pointer 43 b may be caused to correspond to each other.
- the operator 3 functions as the right operator 3 b .
- the right pointer 43 b is displayed at a position on the display screen 41 , which corresponds to the touch position 5 of the right operator 3 b.
- the touch pad 2 is divided into the left area 2 a and the right area 2 b .
- the operator is caused to correspond to the left pointer 43 a .
- the operator is caused to correspond to the right pointer 43 b .
- This may allow the user 30 to perform intuitive operations. It should be noted that the correspondence relations between the touch positions 5 of the operators 3 and the pointers 43 a and 43 b are not limited to that described above and can be appropriately set. In addition, a method of setting areas for dividing the touch pad 2 can be set.
- the correspondence relations are determined in the following manner, for example.
- one operator 3 inputs the initial input operation on the left area 2 a
- the other operator 3 inputs the initial input operation on the right area 2 b .
- the operator 3 on the left area 2 a is caused to correspond to the left pointer 43 a
- the operator 3 on the right area 2 b is caused to correspond to the right pointer 43 b .
- the two operators 3 input the initial input operations almost at the same time. In this case, for example, the operator 3 closer to the left side of the touch pad 2 is caused to correspond to the left pointer 43 a .
- the correspondence relations between the operators 3 and the pointers may be determined according to the order in which the initial input operations are input. For example, the operator 3 that has input the initial input operation to the touch pad 2 first is caused to correspond to the left pointer 43 a . A method of determining the correspondence relations between the operators 3 and the pointers can be appropriately determined.
- Step 103 It is determined whether or not touch operations for respectively controlling the left pointer 43 a and the right pointer 43 b , which are displayed on the display screen 41 , have been received.
- Step 104 an action assigned to a combination of the touch operations is executed (Step 104 ).
- FIGS. 6A to 13 are views for describing examples of the touch operations for respectively controlling the left pointer 43 a and the right pointer 43 b and actions assigned to combinations of the touch operations.
- an icon 47 as a GUI object is displayed on the display screen 41 .
- the icon 47 is displayed as the CPU 103 serving as the first display processor outputs a display signal of the icon 47 to the display apparatus 40 . Further, on the display screen 41 , a window 44 a of a folder A storing the icon 47 and a window 44 b of another folder B are displayed.
- the icon 47 is designated by the left pointer 43 a serving as a first pointer. At this time, in order to express that the icon 47 has been designated by the left pointer 43 a , a frame, a color, or the like may be applied to the icon 47 . As shown in FIG. 6A , the right pointer 43 b serving as a second pointer is located within the window 44 b of the folder B.
- an indication indicating the movement processing and an indication indicating the copy processing may be displayed on the display screen 41 . Then, by moving the operators 3 a and 3 b away from the touch pad 2 , the processing displayed at that time may be executed as the action.
- the left pointer 43 a designates the icon 47 .
- the right pointer 43 b determines the destination position. In this manner, the icon 47 can be moved with high operability. Further, the right pointer 43 b determines the display position for the copy 47 ′, and hence the icon 47 can be copied with high operability.
- the right pointer 43 b may designate the icon 47 .
- the left pointer 43 a may determine the destination position for the icon 47 or the display position for the copy 47 ′.
- the icon 47 is stored in the folder A, and the right pointer 43 b is located within the window 44 b of the folder B.
- the icon 47 may be stored on a desk top.
- the right pointer 43 b may be located on the desk top. In either case, the movement processing or copy processing may be executed as described above.
- the icon 47 is displayed on the display screen 41 .
- the icon 47 is designated by the left pointer 43 a serving as a first cursor pointer and the right pointer 43 b serving as a second cursor pointer.
- the right operator 3 b is moved with the left operator 3 a and the right operator 3 b being pressed against the screen, so that a relative distance between the operators 3 a and 3 b may be increased.
- the right pointer 43 b moves on the display screen 41 so that the left pointer 43 a and the right pointer 43 b move away from each other.
- copy processing of the icon 47 is executed.
- the created copy 47 ′ moves together with the right pointer 43 b .
- the movement of the copy 47 ′ is terminated.
- the left pointer 43 a and the right pointer 43 b designate a plurality of icons 47 a to 47 c at the same time.
- copy processing may be executed with respect to the plurality of icons 47 a to 47 c at the same time.
- FIGS. 7A and 7B when the right pointer 43 b is moved, the copy processing may be executed. However, the copy processing may be executed when the left pointer 43 a is moved.
- the left pointer 43 a and the right pointer 43 b designate the icon 47 .
- the operators 3 a and 3 b are moved with the operators 3 a and 3 b being pressed against the screen, so that a relative distance between the operators 3 a and 3 b is increased.
- the left pointer 43 a and the right pointer 43 b move away from each other on the display screen 41 .
- deletion processing of the icon 47 is executed as the assigned action.
- images 48 obtained by separating the icon 47 into two are displayed on the display screen 41 .
- the moved operators 3 a and 3 b are moved back to their original positions with the operators 3 a and 3 b being pressed against the screen. Specifically, when the pointers 43 a and 43 b are moved back to their original positions, execution of the deletion processing may be cancelled.
- the left pointer 43 a and the right pointer 43 b designate the plurality of icons 47 a to 47 c at the same time.
- the pointers 43 a and 43 b are moved.
- the deletion processing may be executed on the plurality of icons 47 a to 47 c at the same time.
- the pointers 43 a and 43 b are moved away from each other in the left- and right-hand directions on the display screen 41 .
- the copy processing shown in FIGS. 7A and 7B may be executed.
- the pointers 43 a and 43 b are moved away from each other in upper and lower directions on the display screen 41 .
- the deletion processing shown in FIG. 8 may be executed.
- the icon 47 is designated by the left pointer 43 a and the right pointer 43 b .
- the copy processing of the icon 47 is executed. With this, the icon 47 can be copied with relative ease and high operability. Further, when the pointers 43 a and 43 b are moved away from each other, the deletion processing of the icon 47 is executed. With this, the icon 47 can be deleted with high operability.
- the icon 47 of an image file is displayed on the display screen 41 .
- the icon 47 is designated by the left pointer 43 a serving as the first cursor pointer. Further, the icon 47 is designated by the right pointer 43 b serving as the second cursor pointer.
- the operators 3 a and 3 b are moved relative to each other with the left operator 3 a and the right operator 3 b being pressed against the screen so that a relative distance between the operators 3 a and 3 b is decreased. At this time, only one or both of the operators 3 a and 3 b may be moved.
- the pointers 43 a and 43 b are moved relative to each other on the display screen 41 so that the left pointer 43 a and the right pointer 43 b approach each other.
- compression processing of the image file as data related to the icon 47 is executed as the action.
- Compression rate for the image file to be compressed may depend on an amount of change of the relative distance between the pointers 43 a and 43 b .
- the compression rate may be set such that it becomes higher as the pointers 43 a and 43 b approach relative to each other.
- the image file may be compressed at a compression rate.
- the data to be compressed is not limited to the image file; and a moving image file, an audio file, or the like may be compressed. Further, a compression method and the like can be appropriately set. As shown in FIG. 14A , an icon 47 ′′ of the compressed image file is displayed in a size smaller than that of the icon 47 of the image file before the compression. With this, the user 30 is allowed to visually recognize that the compression processing has been executed, with the result that intuitive operations become possible.
- the icon 47 is designated by the left pointer 43 a and the right pointer 43 b .
- the compression processing of the data related to the icon 47 is executed. With this, the data related to the icon 47 can be compressed with high operability.
- the icon 47 of the image file is displayed on the display screen 41 .
- the icon 47 is designated by the left pointer 43 a serving as the first cursor pointer.
- the left operator 3 a performs a click operation.
- reproduction processing of the image file is executed.
- an image 45 is displayed on the display screen 41 .
- the icon 47 is designated by the right pointer 43 b serving as the second cursor pointer.
- the right operator 3 b performs a click operation.
- a context menu 46 related to the icon 47 is displayed.
- various processing items that will be executed with respect to the image file are displayed. For example, in the context menu 46 , items of the reproduction processing, print processing, copy processing, cut processing, deletion processing, and the like are displayed.
- FIG. 10 shows a state in which the icon 47 is designated by the left pointer 43 a and the right pointer 43 b . However, either the left pointer 43 a or the right pointer 43 b may designate the icon 47 . Then, the operator 3 a or 3 b may perform the click operation.
- the operation using only the left operator 3 a or the right operator 3 b alone may be classified as a combination with a non-operation by the other operator.
- a window A is displayed on the display screen 41 .
- the window A is, for example, a window of a folder, a window of an application for creating mails, documents, or the like, a window of a web browser, or the like.
- an icon of the folder, the application, or the like is designated by the left pointer 43 a or the right pointer 43 b .
- the operator 3 a or 3 b performs an operation such as a click operation.
- display processing of the window A is executed as the action assigned to this click operation.
- both of the operators 3 a and 3 b are operated so that a relative distance between the left pointer 43 a and the right pointer 43 b is decreased in the left- and right-hand directions on the display screen 41 .
- the pointers 43 a and 43 b are moved in the same direction. That is, the operators 3 a and 3 b are operated so that the left pointer 43 a and the right pointer 43 b intersect in the left- and right-hand directions on the display screen 41 .
- window switch processing of switching the window to be in a focused state is executed as the assigned action. With this, as shown in FIG. 11 , the window in the focused state is switched from the window A to a window B.
- the window switch processing may be executed with respect to an operation of changing the relative distance between the left pointer 43 a and the right pointer 43 b .
- the operation of causing the left pointer 43 a and the right pointer 43 b to intersect is set assuming a physical operation of turning sheets of paper or the like. That is, the user can execute the window switch processing by intuitive operations.
- the window switch processing may be executed with respect to the operation of changing the relative distance between the pointers 43 a and 43 b in the upper and lower directions on the display screen 41 .
- a plurality of icons 47 a to 47 c are displayed on the display screen 41 .
- operations of moving the left pointer 43 a and the right pointer 43 b are input.
- designation processing of designating icons 47 a and 47 b out of the plurality of icons 47 a to 47 c at the same time is executed as an action assigned to combination of the above-mentioned touch operations.
- the positions of a start point 49 a and an end point 50 a of the left pointer 43 a and the positions of a start point 49 b and an end point 50 b of the right pointer 43 b may be calculated. Then, a designation area 52 surrounded by straight lines 51 linking the four points is determined. The icons 47 a and 47 b located within this designation area 52 are designated. With this, for example, as compared to the case of setting the designation area by one pointer, it is possible to easily set a larger area as the designation area 52 with high operability.
- positional information representing movement tracks of the pointers 43 a and 43 b is calculated.
- a straight line 51 linking the start points 49 a and 49 b a straight line 51 linking the end points 50 a and 50 b , and a curved line 53 representing tracks of the pointers 43 a and 43 b
- a designation area 52 ′ is set.
- the icons 47 a and 47 b located within this designation area 52 ′ are designated. With this, according to the positions of the plurality of icons 47 a to 47 c displayed on the display screen 41 , icons 47 a and 47 b can be designated with high operability.
- an icon designated from the icons 47 a and 47 b located within the designation area 52 or 52 ′ may be additionally selected. For example, before or after the designation area 52 or 52 ′ is determined, attribute information of the designated icon may be set. Then, only an icon including such attribute information may be selected as the designated icon. Before or after the designation area 52 or 52 ′ is determined, an icon is selected by the left pointer 43 a or the right pointer 43 b . Then, only an icon including the same attribute information as the selected icon may be designated.
- the icons 47 a and 47 b located within the designation area 52 ( 52 ′) surrounded by the start points and the end points of the two pointers 43 a and 43 b may be designated.
- the icons 47 a and 47 b can be designated with high operability.
- the left pointer 43 a and the right pointer 43 b determine the designation area 52 , and the icon 47 of the image file is designated. Then, from the positions of the end points 50 a and 50 b of the pointers 43 a and 43 b , the pointers 43 a and 43 b are brought close to each other. At this time, the compression processing of the image file may be executed as the assigned action. In the case where the compression rate depends on the amount of change of the relative distance between the pointers 43 a and 43 b , this amount of change can be easily adjusted.
- the touch pad 2 receives the operations of the left pointer 43 a and the right pointer 43 b . Then, the action previously assigned to the combination of the operations of the pointers 43 a and 43 b is executed with respect to the icon 47 . With this, by appropriately setting the action assigned to the combination of the operations, multi-touch operations with high operability become possible.
- the processing shown in FIGS. 6 to 11 can be executed. That is, when at least one pointer designates the icon 47 and the operations of the pointers 43 a and 43 b in this state are received, the action previously assigned to the combination of those operations is executed with respect to the designated icon 47 . With this, the action with respect to the icon 47 can be executed with high operability.
- Embodiments according to the present disclosure are not limited to the above-mentioned embodiments and various modifications can be made.
- FIGS. 14A-15B are schematic views showing a modified example of the information processing apparatus 100 shown in FIG. 1 .
- an information processing apparatus 300 includes a touch pad 302 , a D-pad 310 , and operation buttons 320 .
- the information processing apparatus 300 may be provided with the operation buttons 320 and the like.
- an information processing apparatus 400 includes a touch pad 402 a and a touch pad 402 b .
- the touch pad 402 a is used for a touch operation by a left operator.
- the touch pad 402 b is used for a touch operation by a right operator.
- the correspondence relations between the operators and the pointers can be easily determined.
- the operations for respectively controlling the pointers can be intuitively performed with high operability.
- the operability of each pointer can be kept high.
- each pointer is operated based on relative coordinate information being information of the movement direction and movement amount of each pointer.
- an information processing apparatus 500 or 600 As shown in FIGS. 15A and 15B , as an information processing apparatus 500 or 600 according to a modified example of the present disclosure, an information processing apparatus including a display screen 541 or 641 may be used.
- the information processing apparatuses 500 or 600 is typically manufactured in a size that the user is capable of carrying.
- the touch pad serving as the input unit, the CPU, and the like are integrally provided.
- a main body including the CPU and an input controller including the touch pad may be provided separately. That is, for example, a remote controller having an outer appearance that is substantially the same as the outer appearance of the information processing apparatus 100 shown in FIG. 1 and a main body including the CPU provided separately from the controller may constitute an information processing apparatus according to an embodiment of the present disclosure.
- an information processing apparatus without the touch pad may be used.
- the touch pad serving as an external input apparatus is connected in a wired or wireless manner. Then, a contact signal or the like output from this touch pad serving as the external input apparatus is received. That is, as the information processing apparatus according to this embodiment, an information processing apparatus having a structure for receiving the operation of each pointer through the external input apparatus may be used. In this case, an interface or the like provided for connection to the external input apparatus functions as the input unit.
- the touch pad serving as the input unit includes the pressure sensor.
- the touch pad does not need to include the pressure sensor and detect the pressing operations of the operators. In this case, it is only sufficient that various types of processing are executed according to operations of contact and separation of the operators.
- embodiments herein need not include a pressure sensor. “Touch operations” may be detected by, in addition to or alternative to pressure sensing, any other suitable methods such as by using optical or infrared detectors.
- the operators control the cursor pointers.
- a relative coordinate system may be used.
- the cursor pointer may be hidden.
- the correspondence relations between the operators and the cursor pointers are determined.
- the correspondence relations may be determined based on contact angles obtained when the operators come in contact with the touch pad. For example, the contact areas of the operators are detected, and the contact angles are calculated from the shape of each area, a change of the shape, or the like. A method of calculating the contact angles can be appropriately set.
- FIG. 16 is a view schematically showing an embodiment of the present disclosure.
- An information processing apparatus 700 shown in FIG. 16 includes a touch panel 750 of a multi-touch type as the input unit capable of receiving touch operations by one or more operators. Thus, no pointer needs to be displayed on a display screen 741 .
- the user 730 inputs touch operations to the display screen 741 by use of a plurality of operators 703 a and 703 b . Then, with respect to an icon 747 displayed on the display screen 741 , an action assigned to those touch operations is executed. With this, multi-touch operations with high operability become possible.
Landscapes
- Engineering & Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- User Interface Of Digital Computer (AREA)
- Position Input By Displaying (AREA)
Abstract
An apparatus comprising a memory storing instructions and a control unit executing the instructions is provided. The control unit is configured to send signals to display first and second indicators on a display device. The control unit is further configured to receive first and second user inputs and, in response to the received inputs, to send signals to change a display state of the first indicator according to the first input. The control unit is further configured to send signals to change a display state of the second indicator according to the second input and initiate an operation to be performed based on a combination of the first and second inputs.
Description
- The present disclosure relates to an information processing apparatus and an information processing method, which enable a multi-touch operation to be made, and a program therefor.
- In recent years, there have been known various terminal apparatuses that enable multi-touch operations to be made. For example, Japanese Unexamined Patent Application Publication No. 2009-522669 (hereinafter, referred to as Patent Document 1) discloses a portable electronic apparatus that enables, through a multi-touch operation by a user, an enlargement operation and the like with respect to an object displayed on a display screen to be made (see paragraph [0018], FIG. 1 and the like in Patent Document 1).
- The use of such terminal apparatuses that enable multi-touch operations to be made will likely increase. Thus, it is desired to improve operability for users who perform the multi-touch operations.
- In view of the above-mentioned circumstances, there is a need for providing an information processing apparatus and an information processing method, which enable a multi-touch operation to be made with high operability, and a program therefore. In one exemplary embodiment, the present disclosure is directed towards an apparatus comprising a memory storing instructions and a control unit executing the instructions to send signals to display first and second indicators on a display device. The control unit further executes the instructions to receive first and second user inputs and, in response to the received inputs, to send signals to change a display state of the first indicator according to the first input. The control unit further executes the instructions to send signals to change a display state of the second indicator according to the second input and initiate an operation to be performed based on a combination of the first and second inputs.
- In another exemplary embodiment, the present disclosure is directed toward a method comprising sending signals to display first and second indicators on a display device. The method further comprises receiving first and second user inputs and, in response to the received inputs, sending signals to change a display state of the first indicator according to the first input. The method further comprises sending signals to change a display state of the second indicator according to the second input. The method further comprises initiating an operation to be performed based on a combination of the first and second inputs.
- In another exemplary embodiment, the present disclosure is directed toward a tangibly embodied non-transitory computer-readable medium storing instructions which, when executed by a processor, perform a method comprising sending signals to display first and second indicators on a display device. The method further comprises receiving first and second user inputs. The method further comprises, in response to the received inputs, sending signals to change a display state of the first indicator according to the first input and sending signals to change a display state of the second indicator according to the second input. The method further comprises initiating an operation to be performed based on a combination of the first and second inputs.
- As described above, according to the embodiments of the present disclosure, multi-touch operations with high operability become possible.
- These and other objects, features and advantages of the present disclosure will become more apparent in light of the following detailed description of best mode embodiments thereof, as illustrated in the accompanying drawings.
-
FIG. 1 is a view schematically showing an outer appearance of an information processing apparatus according to an embodiment of the present disclosure; -
FIG. 2 is a block diagram showing a configuration example of the information processing apparatus shown inFIG. 1 ; -
FIG. 3 is a view schematically showing the information processing apparatus shown inFIG. 1 and a display apparatus connected thereto; -
FIG. 4 is a flowchart showing an operation of the information processing apparatus shown inFIG. 1 ; -
FIG. 5 is a view for describing an example of a method of determining a correspondence relation between an operator and a cursor pointer shown inFIG. 3 ; -
FIGS. 6A and 6B are views for describing touch operations for controlling a left pointer and a right pointer and an action assigned to the touch operations; -
FIGS. 7A and 7B are views for describing touch operations for controlling the left pointer and the right pointer and an action assigned to the touch operations; -
FIG. 8 is a view for describing touch operations for controlling the left pointer and the right pointer and an action assigned to the touch operations; -
FIG. 9 is a view for describing touch operations for controlling the left pointer and the right pointer and an action assigned to the touch operations; -
FIG. 10 is a view for describing touch operations for controlling the left pointer and the right pointer and an action assigned to the touch operations; -
FIG. 11 is a view for describing touch operations for controlling the left pointer and the right pointer and an action assigned to the touch operations; -
FIGS. 12A and 12B are views for describing touch operations for controlling the left pointer and the right pointer and an action assigned to the touch operations; -
FIG. 13 is a view for describing touch operations for controlling the left pointer and the right pointer and an action assigned to the touch operations; -
FIGS. 14A and 14B are schematic views showing a modification of the information processing apparatus shown inFIG. 1 ; -
FIGS. 15A and 15B are schematic views showing a modification of the information processing apparatus shown inFIG. 1 ; -
FIG. 16 is a view schematically showing a modification of variations of the present disclosure; and -
FIG. 17 is a schematic view showing another example of the information processing apparatus shown inFIG. 1 . - Hereinafter, embodiments of the present disclosure will be described with reference to the drawings.
-
FIG. 1 is a view schematically showing an outer appearance of an information processing apparatus according to an embodiment of the present disclosure.FIG. 2 is a block diagram showing a configuration example of the information processing apparatus shown inFIG. 1 . - The
information processing apparatus 100 includes acasing 1 that is held by auser 30 and atouch pad 2 provided in front of thecasing 1 to serve as an input unit. Thetouch pad 2 according to this embodiment is capable of receiving various types of multi-touch operations with one ormore operators 3. - The
casing 1 is formed of, for example, metal such as aluminum and stainless steel, plastic, or other various materials. Thecasing 1 has such a size such thatuser 30 is capable of carrying theapparatus 100. For example, when seen from theuser 30 inFIG. 1 , the size of thecasing 1 is about 15 cm in a longitudinal direction, about 10 cm in a lateral direction, and about 1 cm in thickness. However, the size of thecasing 1 can be appropriately set. For example, as shown inFIG. 17 , theinformation processing apparatus 100 including thecasing 1 having a size of 30 cm or more in the longitudinal and lateral directions may be used while placed on a desk. - As shown in
FIG. 1 , in this embodiment, as the plurality ofoperators 3 serving to perform touch operations with respect to thetouch pad 2, thumbs of theuser 30 are used. However, other fingers may also be used (seeFIG. 17 ). Alternatively, as theoperators 3, a stylus, pen, joystick, or other suitable apparatus may be used. - As shown in
FIG. 2 , thetouch pad 2 according to this embodiment includes aposition detector 101 and apressure sensor 102. Theposition detector 101 includes a sensor unit and an arithmetic circuit unit (not shown). When theoperator 3 comes close to or in contact with thetouch pad 2, the sensor unit outputs a proximity or contact signal to the arithmetic circuit unit. The arithmetic circuit unit generates, based on the output proximity or contact signal, information on a proximity or contact position of theoperator 3, for example. Alternatively, information on a movement direction, movement amount, movement speed, acceleration speed, or the like of theoperator 3 may be generated. In addition, information on a contact area, contact angle, or the like of theoperator 3 may be generated. - As used herein, the term “touch operation” refers to various touch operations performed when the
operator 3 comes close to or in contact with thetouch pad 2. It should be understood that, in certain variations, theoperator 3 need not touch or make contact withtouch pad 2 in order to perform a touch operation. Further, the proximity and contact positions are collectively referred to as touch position of theoperator 3. - As mentioned above, the
touch pad 2 according to this embodiment is a touch pad of a so-called multi-touch type. Thus, it is possible to detect touch operations by theoperators 3 with respect to thetouch pad 2. For example, theuser 30 can perform a pinch operation with two fingers. The pinch operation is, for example, an operation of bringing the two fingers into contact with thetouch pad 2 at the same time and opening or closing the two fingers or the like while keeping the in-contact state. Here, the operation of opening the two fingers or the like held in contact with thetouch pad 2 is referred to as pinch-out operation. Meanwhile, the operation of closing the two fingers or the like held in contact with thetouch pad 2 is referred to as pinch-in operation. - The
pressure sensor 102 serves to detect whether or not thetouch pad 2 is pressed by theoperator 3. For example, when thepressure sensor 102 detects the pressing force above a threshold level, it is determined that thetouch pad 2 is pressed. A method of determining existence, absence of pressure, change in pressure, or amount of pressure by thepressure sensor 102, the structure of thepressure sensor 102, a method of detecting the pressing force, and the like can be appropriately set. In this embodiment, thetouch pad 2 of a capacitive type is used. However, a touch pad of another type, such as, for example, a resistive touch pad, a surface acoustic wave touch pad, and an infrared touch pad may be used instead. - Alternatively, or in addition to
touch pad 2, a “touch operation” may be detected by any suitable method and/or by using any suitable apparatus. For example, various optical detection apparatuses (e.g., cameras, electric eyes or other light-sensing devices) may be used to sense a “touch operation” or any of the other operations discussed herein. Also, detection apparatuses that detect infrared radiation or other radiation outside of the visible spectrum may be used to sense a “touch operation” or any of the other operations discussed herein. These devices may be used in conjunction with other equipment that may, for example, aid in sensing the location of the user's fingers, stylus or other user control apparatus (e.g., gloves may be used to enhance the visual signature of the users fingers, for example). - Further, as shown in
FIG. 2 , theinformation processing apparatus 100 includes a control unit such as a CPU (Central Processing Unit) 103, a RAM (Random Access Memory) 104, astorage 105, and adisplay interface 106. - The
storage 105 is a non-volatile storage device. For example, thestorage 105 may be an HDD (Hard Disk Drive), a flash memory, or other solid-state memories. A part or all of theRAM 104 and thestorage 105 corresponds to a storage means according to this embodiment. Alternatively, theinformation processing apparatus 100 may include a ROM (Read Only Memory) (not shown), and thus the storage means may include the ROM. - The
display interface 106 serves to connect theinformation processing apparatus 100 to a display apparatus 40 (seeFIG. 3 ) including adisplay screen 41. Via thedisplay interface 106, image data of still images, moving images, and the like, control signals for controlling the operation of thedisplay apparatus 40, and the like are output. The image data, the control signals, and the like are output to thedisplay apparatus 40 in a wireless or wired manner. - The
information processing apparatus 100 may be provided with a drive unit and a communication unit (not shown). The drive unit may be, for example, a device capable of driving a removable recording medium such as an optical recording medium, a floppy (registered trademark) disc, a magnetic recording tape, and a flash memory. - The communication unit may be, for example, a modem, a router, and other communication apparatuses for communicating with other devices that are capable of connecting to a LAN (Local Area Network), a WAN (Wide Area Network), and the like. The communication unit may communicate in a wired or wireless manner.
- Data processing by the
information processing apparatus 100 is realized by software stored in thestorage 105 or the like in cooperation with hardware resources of theinformation processing apparatus 100. For example, by theCPU 103 loading a program of instructions stored in thestorage 105 or the like into theRAM 104 and executing it, various types of data processing are realized. In this embodiment, theCPU 103 functions as a first display processor, a second display processor, and an action executing unit. It should be noted that theRAM 104 may include a video RAM (VRAM) (not shown) for displaying images. -
FIG. 3 is a view schematically showing theinformation processing apparatus 100 and thedisplay apparatus 40 connected thereto. As shown inFIG. 3 , a plurality ofcursor pointers 43 are displayed on thedisplay screen 41. The plurality ofcursor pointers 43 are displayed as theCPU 103 serving as the second display processor outputs display signals of thecursor pointers 43 to thedisplay apparatus 40. In this embodiment, acursor pointer 43 a and acursor pointer 43 b may be displayed on thedisplay screen 41. Thecursor pointer 43 a is, for example, operated mainly by theleft operator 3 a being a left finger or the like. Further, thecursor pointer 43 b is, for example, operated mainly by theright operator 3 b being a right finger or the like. That is, operations for each cursor pointer, which respectively control those twocursor pointers touch pad 2 by theleft operator 3 a and theright operator 3 b. Hereinafter, thecursor pointer 43 a is referred to as theleft pointer 43 a and thecursor pointer 43 b is referred to as theright pointer 43 b. - The
left pointer 43 a and theright pointer 43 b may be displayed in different colors, or exhibit different shading as shown inFIG. 3 , for example, so that the pointers can be discriminated from each other. With this, thepointers pointers pointers pointers - In this embodiment, a coordinate system may be set with a center O of the
touch pad 2 being a reference. Coordinate information including information of positional coordinates of theoperators display apparatus 40. In thedisplay apparatus 40, a coordinate system may be set with a center O′ of thedisplay screen 41 being a reference. In the coordinate system, based on the coordinate information received from theinformation processing apparatus 100, thepointers positions operators 3, respectively. The coordinate systems may be appropriately set in thetouch pad 2 and thedisplay screen 41. - An operation of the
information processing apparatus 100 according to this embodiment will be described.FIG. 4 is a flowchart showing the operation. It may be determined whether or not a contact or proximity of theoperator 3 with respect to thetouch pad 2 is detected as an initial input operation (Step 101). When the contact or the like with thetouch pad 2 is detected (“Yes” in Step 101), a correspondence relation between theoperator 3 and thecursor pointer 43 displayed on thedisplay screen 41 is determined. -
FIG. 5 is a view for describing an example of a method of determining the correspondence relation between theoperator 3 and thecursor pointer 43. In this embodiment, based on atouch position 5 of theoperator 3 with respect to thetouch pad 2, the correspondence relation is determined. - For example, as shown in
FIG. 5 , thetouch pad 2 may be divided into theleft area 2 a and theright area 2 b. When thetouch position 5 of theoperator 3 is located within theleft area 2 a, theoperator 3 and theleft pointer 43 a may be caused to correspond to each other. Thus, theoperator 3 functions as theleft operator 3 a. Theleft pointer 43 a is displayed at a position on thedisplay screen 41, which corresponds to thetouch position 5 of theleft operator 3 a. When thetouch position 5 of theoperator 3 is located within theright area 2 b, theoperator 3 and theright pointer 43 b may be caused to correspond to each other. - After the
left operator 3 a and theleft pointer 43 a are caused to correspond to each other, a contact or the like of theother operator 3 with thetouch pad 2 may be detected. In this case, irrespective of atouch position 5 of theother operator 3, theoperator 3 and theright pointer 43 b may be caused to correspond to each other. Thus, theoperator 3 functions as theright operator 3 b. Theright pointer 43 b is displayed at a position on thedisplay screen 41, which corresponds to thetouch position 5 of theright operator 3 b. - Thus, in this embodiment, the
touch pad 2 is divided into theleft area 2 a and theright area 2 b. With respect to the initial input operation on theleft area 2 a, the operator is caused to correspond to theleft pointer 43 a. On the other hand, with respect to the initial input operation on theright area 2 b, the operator is caused to correspond to theright pointer 43 b. This may allow theuser 30 to perform intuitive operations. It should be noted that the correspondence relations between thetouch positions 5 of theoperators 3 and thepointers touch pad 2 can be set. - When the two
operators 3 input the initial input operations almost at the same time, the correspondence relations are determined in the following manner, for example. For example, oneoperator 3 inputs the initial input operation on theleft area 2 a, and theother operator 3 inputs the initial input operation on theright area 2 b. In this case, theoperator 3 on theleft area 2 a is caused to correspond to theleft pointer 43 a, and theoperator 3 on theright area 2 b is caused to correspond to theright pointer 43 b. Within either theleft area 2 a or theright area 2 b, the twooperators 3 input the initial input operations almost at the same time. In this case, for example, theoperator 3 closer to the left side of thetouch pad 2 is caused to correspond to theleft pointer 43 a. Then, theother operator 3 is caused to correspond to theright pointer 43 b. The correspondence relations between theoperators 3 and the pointers may be determined according to the order in which the initial input operations are input. For example, theoperator 3 that has input the initial input operation to thetouch pad 2 first is caused to correspond to theleft pointer 43 a. A method of determining the correspondence relations between theoperators 3 and the pointers can be appropriately determined. - It is determined whether or not touch operations for respectively controlling the
left pointer 43 a and theright pointer 43 b, which are displayed on thedisplay screen 41, have been received (Step 103). When it is determined that the touch operations for each of thepointers FIGS. 6A to 13 are views for describing examples of the touch operations for respectively controlling theleft pointer 43 a and theright pointer 43 b and actions assigned to combinations of the touch operations. - As shown in
FIG. 6A , anicon 47 as a GUI object is displayed on thedisplay screen 41. Theicon 47 is displayed as theCPU 103 serving as the first display processor outputs a display signal of theicon 47 to thedisplay apparatus 40. Further, on thedisplay screen 41, awindow 44 a of a folder A storing theicon 47 and awindow 44 b of another folder B are displayed. - The
icon 47 is designated by theleft pointer 43 a serving as a first pointer. At this time, in order to express that theicon 47 has been designated by theleft pointer 43 a, a frame, a color, or the like may be applied to theicon 47. As shown inFIG. 6A , theright pointer 43 b serving as a second pointer is located within thewindow 44 b of the folder B. - In this state, the
left operator 3 a and theright operator 3 b perform click operations almost at the same time. Consequently, as an action assigned to a combination of the click operations, movement processing of theicon 47 to a position of theright pointer 43 b is executed. With this, an instant drag and drop operation is realized. It should be noted that instead of the click operations, other touch operations (e.g., double click operations, operations including any number of clicks, tap, double tap, operations including any number of taps, drag operations, etc. may be input in order to execute the movement processing of theicon 47. - As shown in
FIG. 6B , as the action assigned to the click operations by theleft operator 3 a and theright operator 3 b, copy processing of theicon 47 may be realized. At the position of theright pointer 43 b, that is, within the folder B, acopy 47′ of theicon 47 is created. Alternatively, the click operations by theleft operator 3 a and theright operator 3 b may be assigned with the movement processing shown inFIG. 6A . Moreover, double click operations by theleft operator 3 a and theright operator 3 b may be assigned with the copy processing shown inFIG. 6B . - Every time the
left operator 3 a and theright operator 3 b perform click operations, an indication indicating the movement processing and an indication indicating the copy processing may be displayed on thedisplay screen 41. Then, by moving theoperators touch pad 2, the processing displayed at that time may be executed as the action. - As shown in
FIGS. 6A and 6B , theleft pointer 43 a designates theicon 47. Then, theright pointer 43 b determines the destination position. In this manner, theicon 47 can be moved with high operability. Further, theright pointer 43 b determines the display position for thecopy 47′, and hence theicon 47 can be copied with high operability. - It should be noted that the
right pointer 43 b may designate theicon 47. Further, theleft pointer 43 a may determine the destination position for theicon 47 or the display position for thecopy 47′. Further, inFIGS. 6A and 6B , theicon 47 is stored in the folder A, and theright pointer 43 b is located within thewindow 44 b of the folder B. However, theicon 47 may be stored on a desk top. Alternatively, theright pointer 43 b may be located on the desk top. In either case, the movement processing or copy processing may be executed as described above. - As shown in
FIG. 7A , theicon 47 is displayed on thedisplay screen 41. Theicon 47 is designated by theleft pointer 43 a serving as a first cursor pointer and theright pointer 43 b serving as a second cursor pointer. - In this state, the
right operator 3 b is moved with theleft operator 3 a and theright operator 3 b being pressed against the screen, so that a relative distance between theoperators right pointer 43 b moves on thedisplay screen 41 so that theleft pointer 43 a and theright pointer 43 b move away from each other. At this time, as the action assigned to a combination of the touch operations, copy processing of theicon 47 is executed. - As shown in
FIG. 7A , the createdcopy 47′ moves together with theright pointer 43 b. When the pressing of theoperators copy 47′ is terminated. - As shown in
FIG. 7B , theleft pointer 43 a and theright pointer 43 b designate a plurality oficons 47 a to 47 c at the same time. By moving theright pointer 43 b, copy processing may be executed with respect to the plurality oficons 47 a to 47 c at the same time. It should be noted that inFIGS. 7A and 7B , when theright pointer 43 b is moved, the copy processing may be executed. However, the copy processing may be executed when theleft pointer 43 a is moved. - As shown in
FIG. 8 , theleft pointer 43 a and theright pointer 43 b designate theicon 47. In this state, theoperators operators operators left pointer 43 a and theright pointer 43 b move away from each other on thedisplay screen 41. At this time, deletion processing of theicon 47 is executed as the assigned action. - In this embodiment,
images 48 obtained by separating theicon 47 into two are displayed on thedisplay screen 41. With this, it is possible to visually recognize that the deletion processing has been executed. It should be noted that the movedoperators operators pointers - As shown in
FIG. 7B , theleft pointer 43 a and theright pointer 43 b designate the plurality oficons 47 a to 47 c at the same time. In this state, thepointers icons 47 a to 47 c at the same time. - The
pointers display screen 41. At this time, the copy processing shown inFIGS. 7A and 7B may be executed. Then, thepointers display screen 41. At this time, the deletion processing shown inFIG. 8 may be executed. - As shown in
FIGS. 7A-8 , theicon 47 is designated by theleft pointer 43 a and theright pointer 43 b. When theright pointer 43 b is moved so that thepointers icon 47 is executed. With this, theicon 47 can be copied with relative ease and high operability. Further, when thepointers icon 47 is executed. With this, theicon 47 can be deleted with high operability. - As shown in
FIG. 9 , theicon 47 of an image file is displayed on thedisplay screen 41. Theicon 47 is designated by theleft pointer 43 a serving as the first cursor pointer. Further, theicon 47 is designated by theright pointer 43 b serving as the second cursor pointer. - In this state, the
operators left operator 3 a and theright operator 3 b being pressed against the screen so that a relative distance between theoperators operators pointers display screen 41 so that theleft pointer 43 a and theright pointer 43 b approach each other. At this time, compression processing of the image file as data related to theicon 47 is executed as the action. - Compression rate for the image file to be compressed may depend on an amount of change of the relative distance between the
pointers pointers pointers - The data to be compressed is not limited to the image file; and a moving image file, an audio file, or the like may be compressed. Further, a compression method and the like can be appropriately set. As shown in
FIG. 14A , anicon 47″ of the compressed image file is displayed in a size smaller than that of theicon 47 of the image file before the compression. With this, theuser 30 is allowed to visually recognize that the compression processing has been executed, with the result that intuitive operations become possible. - As shown in
FIG. 9 , theicon 47 is designated by theleft pointer 43 a and theright pointer 43 b. When thepointers pointers icon 47 is executed. With this, the data related to theicon 47 can be compressed with high operability. - As shown in
FIG. 10 , theicon 47 of the image file is displayed on thedisplay screen 41. Theicon 47 is designated by theleft pointer 43 a serving as the first cursor pointer. In this state, theleft operator 3 a performs a click operation. Then, as a first action assigned to this click operation, reproduction processing of the image file is executed. As shown inFIG. 10 , animage 45 is displayed on thedisplay screen 41. As shown inFIG. 10 , theicon 47 is designated by theright pointer 43 b serving as the second cursor pointer. In this state, theright operator 3 b performs a click operation. Then, as a second action assigned to this click operation, acontext menu 46 related to theicon 47 is displayed. In thecontext menu 46, various processing items that will be executed with respect to the image file are displayed. For example, in thecontext menu 46, items of the reproduction processing, print processing, copy processing, cut processing, deletion processing, and the like are displayed. -
FIG. 10 shows a state in which theicon 47 is designated by theleft pointer 43 a and theright pointer 43 b. However, either theleft pointer 43 a or theright pointer 43 b may designate theicon 47. Then, theoperator - As shown in
FIG. 10 , when the click operation is input as an operation with respect to theleft pointer 43 a designating theicon 47, reproduction processing of the image file is executed. Then, when the same click operation is input with respect to theright pointer 43 b designating theicon 47, display processing of thecontext menu 46, which is different from the reproduction processing, is executed. That is, for each of thepointers left pointer 43 a or theright pointer 43 b, there are exemplified various touch operations such as a click operation, a double click operation, a tap operation, and a drag operation. As the action to be assigned to the operation, there are exemplified various types of processing with respect to various GUIs displayed on thedisplay screen 41, such as designation, movement, enlargement, reduction in size, rotation, and deletion of the object displayed on thedisplay screen 41, and reproduction of the data related to the object. - The operation using only the
left operator 3 a or theright operator 3 b alone may be classified as a combination with a non-operation by the other operator. - As shown in
FIG. 11 , a window A is displayed on thedisplay screen 41. The window A is, for example, a window of a folder, a window of an application for creating mails, documents, or the like, a window of a web browser, or the like. - In this embodiment, an icon of the folder, the application, or the like is designated by the
left pointer 43 a or theright pointer 43 b. Then, theoperator FIG. 10 , display processing of the window A is executed as the action assigned to this click operation. - In this state, both of the
operators left pointer 43 a and theright pointer 43 b is decreased in the left- and right-hand directions on thedisplay screen 41. Then, also after thepointers display screen 41, thepointers operators left pointer 43 a and theright pointer 43 b intersect in the left- and right-hand directions on thedisplay screen 41. - As shown in
FIG. 11 , until the relative distance between thepointers pointers operators FIG. 11 , the window in the focused state is switched from the window A to a window B. - As mentioned above, the window switch processing may be executed with respect to an operation of changing the relative distance between the
left pointer 43 a and theright pointer 43 b. The operation of causing theleft pointer 43 a and theright pointer 43 b to intersect is set assuming a physical operation of turning sheets of paper or the like. That is, the user can execute the window switch processing by intuitive operations. It should be noted that the window switch processing may be executed with respect to the operation of changing the relative distance between thepointers display screen 41. - As shown in
FIGS. 12A-12B , a plurality oficons 47 a to 47 c are displayed on thedisplay screen 41. In this state, operations of moving theleft pointer 43 a and theright pointer 43 b are input. Then, designation processing of designatingicons icons 47 a to 47 c at the same time is executed as an action assigned to combination of the above-mentioned touch operations. - In
FIG. 12A , the positions of astart point 49 a and anend point 50 a of theleft pointer 43 a and the positions of astart point 49 b and anend point 50 b of theright pointer 43 b may be calculated. Then, adesignation area 52 surrounded bystraight lines 51 linking the four points is determined. Theicons designation area 52 are designated. With this, for example, as compared to the case of setting the designation area by one pointer, it is possible to easily set a larger area as thedesignation area 52 with high operability. - In
FIG. 12B , positional information representing movement tracks of thepointers straight line 51 linking the start points 49 a and 49 b, astraight line 51 linking the end points 50 a and 50 b, and acurved line 53 representing tracks of thepointers designation area 52′ is set. Theicons designation area 52′ are designated. With this, according to the positions of the plurality oficons 47 a to 47 c displayed on thedisplay screen 41,icons - It should be noted that an icon designated from the
icons designation area designation area designation area left pointer 43 a or theright pointer 43 b. Then, only an icon including the same attribute information as the selected icon may be designated. - As shown in
FIGS. 12A-12B , by moving theleft pointer 43 a and theright pointer 43 b, theicons pointers icons - As shown in
FIG. 13 , theleft pointer 43 a and theright pointer 43 b determine thedesignation area 52, and theicon 47 of the image file is designated. Then, from the positions of the end points 50 a and 50 b of thepointers pointers pointers - As mentioned above, in the
information processing apparatus 100 according to this embodiment, thetouch pad 2 receives the operations of theleft pointer 43 a and theright pointer 43 b. Then, the action previously assigned to the combination of the operations of thepointers icon 47. With this, by appropriately setting the action assigned to the combination of the operations, multi-touch operations with high operability become possible. - Further, in the
information processing apparatus 100 according to this embodiment, the processing shown inFIGS. 6 to 11 can be executed. That is, when at least one pointer designates theicon 47 and the operations of thepointers icon 47. With this, the action with respect to theicon 47 can be executed with high operability. - As the touch operations of the
pointers FIGS. 6 to 13 may be input. Further, as the action assigned to the combination of the touch operations of the pointers, an action other than the actions described above may be executed. - Embodiments according to the present disclosure are not limited to the above-mentioned embodiments and various modifications can be made.
-
FIGS. 14A-15B are schematic views showing a modified example of theinformation processing apparatus 100 shown inFIG. 1 . As shown inFIG. 14A , aninformation processing apparatus 300 includes atouch pad 302, a D-pad 310, andoperation buttons 320. As shown in the figure, theinformation processing apparatus 300 may be provided with theoperation buttons 320 and the like. - As shown in
FIG. 14B , aninformation processing apparatus 400 includes atouch pad 402 a and atouch pad 402 b. Thetouch pad 402 a is used for a touch operation by a left operator. Thetouch pad 402 b is used for a touch operation by a right operator. As shown in the figure, due to the provision of the plurality of touch pads for each of the operators, the correspondence relations between the operators and the pointers can be easily determined. Thus, the operations for respectively controlling the pointers can be intuitively performed with high operability. Further, in the case of using a so-called relative coordinate system, the operability of each pointer can be kept high. In the relative coordinate system, each pointer is operated based on relative coordinate information being information of the movement direction and movement amount of each pointer. - As shown in
FIGS. 15A and 15B , as aninformation processing apparatus display screen information processing apparatuses - In each of the above-mentioned embodiments, the touch pad serving as the input unit, the CPU, and the like are integrally provided. However, a main body including the CPU and an input controller including the touch pad may be provided separately. That is, for example, a remote controller having an outer appearance that is substantially the same as the outer appearance of the
information processing apparatus 100 shown inFIG. 1 and a main body including the CPU provided separately from the controller may constitute an information processing apparatus according to an embodiment of the present disclosure. - Further, as the information processing apparatus according to this embodiment, an information processing apparatus without the touch pad may be used. In this case, to this information processing apparatus, the touch pad serving as an external input apparatus is connected in a wired or wireless manner. Then, a contact signal or the like output from this touch pad serving as the external input apparatus is received. That is, as the information processing apparatus according to this embodiment, an information processing apparatus having a structure for receiving the operation of each pointer through the external input apparatus may be used. In this case, an interface or the like provided for connection to the external input apparatus functions as the input unit.
- In the above description, the touch pad serving as the input unit includes the pressure sensor. However, the touch pad does not need to include the pressure sensor and detect the pressing operations of the operators. In this case, it is only sufficient that various types of processing are executed according to operations of contact and separation of the operators.
- Moreover, as mentioned above, embodiments herein need not include a pressure sensor. “Touch operations” may be detected by, in addition to or alternative to pressure sensing, any other suitable methods such as by using optical or infrared detectors.
- In the above description, two operators of the left operator and the right operator are illustrated. However, two or more operators may input the touch operations to the touch pad.
- In the above description, in the absolute coordinate system, the operators control the cursor pointers. However, a relative coordinate system may be used. Further, when a contact or proximity with respect to the
touch pad 2 is not detected for a period of time, the cursor pointer may be hidden. - In the above description, based on the touch positions of the operator or the order in which the touch operations are input, the correspondence relations between the operators and the cursor pointers are determined. However, the correspondence relations may be determined based on contact angles obtained when the operators come in contact with the touch pad. For example, the contact areas of the operators are detected, and the contact angles are calculated from the shape of each area, a change of the shape, or the like. A method of calculating the contact angles can be appropriately set.
- In the
touch pad 2 shown inFIG. 1 , an operator comes in contact with thetouch pad 2 in order to move in a direction from the left-hand side to the upper right-hand side. In this case, this operator is caused to correspond to the left pointer. An operator comes in contact with thetouch pad 2 in order to move in a direction from the right-hand side to the upper left-hand side of thetouch pad 2. In this case, this operator is caused to correspond to the right pointer. With this, generally, the left pointer is caused to correspond to operations by the left hand and the right pointer is caused to correspond to operations by the right hand. As a result, intuitive operations become possible. -
FIG. 16 is a view schematically showing an embodiment of the present disclosure. Aninformation processing apparatus 700 shown inFIG. 16 includes atouch panel 750 of a multi-touch type as the input unit capable of receiving touch operations by one or more operators. Thus, no pointer needs to be displayed on adisplay screen 741. Theuser 730 inputs touch operations to thedisplay screen 741 by use of a plurality ofoperators icon 747 displayed on thedisplay screen 741, an action assigned to those touch operations is executed. With this, multi-touch operations with high operability become possible. - The present disclosure contains subject matter related to that disclosed in Japanese Priority Patent Application JP 2010-268479 filed in the Japan Patent Office on Dec. 1, 2010, the entire content of which is hereby incorporated by reference.
- It should be understood by those skilled in the art that various modifications, combinations, sub-combinations and alterations may occur depending on design requirements and other factors insofar as they are within the scope of the appended claims or the equivalents thereof.
Claims (20)
1. An apparatus comprising:
a memory storing instructions; and
a control unit executing the instructions to:
send signals to display first and second indicators on a display device;
receive first and second user inputs;
in response to the received inputs:
send signals to change a display state of the first indicator according to the first input;
send signals to change a display state of the second indicator according to the second input; and
initiate an operation to be performed based on a combination of the first and second inputs.
2. The apparatus of claim 1 , wherein:
the control unit is further configured to select a first object displayed on the screen based on at least one of the first and second inputs; and
initiating an operation further comprises initiating the operation for performing on the selected first object.
3. The apparatus of claim 2 , wherein:
selecting the first object further comprises selecting a second object based on an association between the first object and the second object; and
initiating an operation further comprises initiating the operation for performing on the second selected object.
4. The apparatus of claim 2 , wherein:
the first display state represents a first location on the display device;
the second display state represents a second location on the display device; and
the operation comprises moving the selected first object from the first location to the second location.
5. The apparatus of claim 2 , wherein the operation comprises at least one of copying, deleting, compressing, or decompressing the selected first object.
6. The apparatus of claim 1 , wherein:
the first input represents a location of the first indicator on the display device;
the second input represents a location of the second indicator on the display device; and
initiating an operation further comprises initiating the operation when the combination results in one of an increase or a decrease in a distance between the locations of the first and second indicators on the display device.
7. The apparatus of claim 6 , wherein the location of the first indicator on the display device and the location of the second indicator on the display device both correspond to an object on the display device and the operation is performed on the object.
8. The apparatus of claim 7 , wherein when the combination results in an increase in the distance between the locations of the first and second indicators and the location of one of the first or second indicators does not change, the operation comprises copying a file associated with the object.
9. The apparatus of claim 7 , wherein when the combination results in an increase in the distance between the locations of the first and second indicators and the location of both of the first and second indicators changes, the operation comprises deleting a file associated with the object.
10. The apparatus of claim 7 , wherein when the combination results in a decrease in the distance between the locations of the first and second indicators, the operation comprises compressing a file associated with the object.
11. The apparatus of claim 1 , wherein:
the first input represents a location of the first indicator on the display device;
the second input represents a location of the second indicator on the display device; and
the operation is initiated when the combination comprises a user click operation associated with at least one of the first or second indicators.
12. The apparatus of claim 11 , wherein the operation is different when the user click is associated with the first indicator than when the user click is associated with the second indicator.
13. The apparatus of claim 1 , wherein:
the first input represents a series of locations of the first indicator on the display device;
the second input represents a series of locations of the second indicator on the display device; and
the operation comprises selecting an object on the display device located between the series of locations of the first indicator and the series of locations of the second indicator.
14. The apparatus of claim 13 , wherein the operation further comprises at least one of copying, deleting, compressing, or decompressing the selected object.
15. The apparatus of claim 1 , further comprising a touch pad and wherein the touch pad detects the first and second user inputs.
16. The apparatus of claim 15 , wherein the touch pad detects a pressure of a user pressure input and the control unit receives a signal associated with the detected pressure.
17. The apparatus of claim 1 , further comprising an optical detection device which detects the first and second user inputs.
18. The apparatus of claim 1 , further comprising an infrared radiation detection device that detects the first and second user inputs.
19. A method comprising:
sending signals to display first and second indicators on a display device;
receiving first and second user inputs; and
in response to the received inputs:
sending signals to change a display state of the first indicator according to the first input;
sending signals to change a display state of the second indicator according to the second input; and
initiating an operation to be performed based on a combination of the first and second inputs.
20. A tangibly embodied non-transitory computer-readable medium storing instructions which, when executed by a processor, perform a method comprising:
sending signals to display first and second indicators on a display device;
receiving first and second user inputs;
in response to the received inputs:
sending signals to change a display state of the first indicator according to the first input;
sending signals to change a display state of the second indicator according to the second input; and
initiating an operation to be performed based on a combination of the first and second inputs.
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US15/260,204 US11281324B2 (en) | 2010-12-01 | 2016-09-08 | Information processing apparatus, information processing method, and program inputs to a graphical user interface |
US17/668,880 US20220164059A1 (en) | 2010-12-01 | 2022-02-10 | Information processing apparatus, information processing method, and program therefor |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JPP2010-268479 | 2010-01-12 | ||
JP2010268479A JP5750875B2 (en) | 2010-12-01 | 2010-12-01 | Information processing apparatus, information processing method, and program |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/260,204 Continuation US11281324B2 (en) | 2010-12-01 | 2016-09-08 | Information processing apparatus, information processing method, and program inputs to a graphical user interface |
Publications (1)
Publication Number | Publication Date |
---|---|
US20120062603A1 true US20120062603A1 (en) | 2012-03-15 |
Family
ID=45806267
Family Applications (3)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/297,777 Abandoned US20120062603A1 (en) | 2010-01-12 | 2011-11-16 | Information Processing Apparatus, Information Processing Method, and Program Therefor |
US15/260,204 Active US11281324B2 (en) | 2010-12-01 | 2016-09-08 | Information processing apparatus, information processing method, and program inputs to a graphical user interface |
US17/668,880 Pending US20220164059A1 (en) | 2010-12-01 | 2022-02-10 | Information processing apparatus, information processing method, and program therefor |
Family Applications After (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/260,204 Active US11281324B2 (en) | 2010-12-01 | 2016-09-08 | Information processing apparatus, information processing method, and program inputs to a graphical user interface |
US17/668,880 Pending US20220164059A1 (en) | 2010-12-01 | 2022-02-10 | Information processing apparatus, information processing method, and program therefor |
Country Status (3)
Country | Link |
---|---|
US (3) | US20120062603A1 (en) |
JP (1) | JP5750875B2 (en) |
CN (1) | CN102591564B (en) |
Cited By (25)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130127754A1 (en) * | 2011-11-17 | 2013-05-23 | Samsung Electronics Co., Ltd. | Display apparatus and control method thereof |
US8587422B2 (en) | 2010-03-31 | 2013-11-19 | Tk Holdings, Inc. | Occupant sensing system |
US20140043535A1 (en) * | 2011-04-04 | 2014-02-13 | Sharp Kabushiki Kaisha | Display apparatus, information processing system and recording medium |
US8725230B2 (en) | 2010-04-02 | 2014-05-13 | Tk Holdings Inc. | Steering wheel with hand sensors |
CN104220978A (en) * | 2012-04-19 | 2014-12-17 | 索尼公司 | Information processing apparatus, information processing method, program, and information processing system |
US9007190B2 (en) | 2010-03-31 | 2015-04-14 | Tk Holdings Inc. | Steering wheel sensors |
US9032818B2 (en) | 2012-07-05 | 2015-05-19 | Nextinput, Inc. | Microelectromechanical load sensor and methods of manufacturing the same |
US20160117081A1 (en) * | 2014-10-27 | 2016-04-28 | Thales Avionics, Inc. | Controlling entertainment system using combination of inputs from proximity sensor and touch sensor of remote controller |
US9487388B2 (en) | 2012-06-21 | 2016-11-08 | Nextinput, Inc. | Ruggedized MEMS force die |
WO2017032079A1 (en) * | 2015-08-27 | 2017-03-02 | 广东欧珀移动通信有限公司 | Information browsing method and mobile terminal |
US9696223B2 (en) | 2012-09-17 | 2017-07-04 | Tk Holdings Inc. | Single layer force sensor |
US9727031B2 (en) | 2012-04-13 | 2017-08-08 | Tk Holdings Inc. | Pressure sensor including a pressure sensitive material for use with control systems and methods of using the same |
US9902611B2 (en) | 2014-01-13 | 2018-02-27 | Nextinput, Inc. | Miniaturized and ruggedized wafer level MEMs force sensors |
US10466119B2 (en) | 2015-06-10 | 2019-11-05 | Nextinput, Inc. | Ruggedized wafer level MEMS force sensor with a tolerance trench |
US10599252B2 (en) | 2014-06-25 | 2020-03-24 | Huawei Technologies Co., Ltd. | Intelligent terminal control method utilizing touch contact location and pressure |
US10962427B2 (en) | 2019-01-10 | 2021-03-30 | Nextinput, Inc. | Slotted MEMS force sensor |
US11194471B1 (en) | 2021-01-28 | 2021-12-07 | Honda Motor Co., Ltd. | Apparatus and method for display control based on touch interface |
US11221263B2 (en) | 2017-07-19 | 2022-01-11 | Nextinput, Inc. | Microelectromechanical force sensor having a strain transfer layer arranged on the sensor die |
US11243126B2 (en) | 2017-07-27 | 2022-02-08 | Nextinput, Inc. | Wafer bonded piezoresistive and piezoelectric force sensor and related methods of manufacture |
US11243125B2 (en) | 2017-02-09 | 2022-02-08 | Nextinput, Inc. | Integrated piezoresistive and piezoelectric fusion force sensor |
US11255737B2 (en) | 2017-02-09 | 2022-02-22 | Nextinput, Inc. | Integrated digital force sensors and related methods of manufacture |
US11385108B2 (en) | 2017-11-02 | 2022-07-12 | Nextinput, Inc. | Sealed force sensor with etch stop layer |
US11423686B2 (en) | 2017-07-25 | 2022-08-23 | Qorvo Us, Inc. | Integrated fingerprint and force sensor |
US11579028B2 (en) | 2017-10-17 | 2023-02-14 | Nextinput, Inc. | Temperature coefficient of offset compensation for force sensor and strain gauge |
US11874185B2 (en) | 2017-11-16 | 2024-01-16 | Nextinput, Inc. | Force attenuator for force sensor |
Families Citing this family (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP5750875B2 (en) * | 2010-12-01 | 2015-07-22 | ソニー株式会社 | Information processing apparatus, information processing method, and program |
JP6251957B2 (en) * | 2013-01-23 | 2017-12-27 | セイコーエプソン株式会社 | Display device, head-mounted display device, and display device control method |
JP2015028690A (en) * | 2013-07-30 | 2015-02-12 | ソニー株式会社 | Information processing device, information processing method, and program |
KR101422060B1 (en) | 2013-10-30 | 2014-07-28 | 전자부품연구원 | Information display apparatus and method for vehicle using touch-pad, and information input module thereof |
JP5711409B1 (en) * | 2014-06-26 | 2015-04-30 | ガンホー・オンライン・エンターテイメント株式会社 | Terminal device |
KR102248741B1 (en) * | 2015-01-29 | 2021-05-07 | 삼성전자주식회사 | Display appaeatus and control method thereof |
KR102507884B1 (en) | 2018-01-05 | 2023-03-09 | 삼성디스플레이 주식회사 | Apparatus for separating window and method for separating window using the same |
Citations (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20050162402A1 (en) * | 2004-01-27 | 2005-07-28 | Watanachote Susornpol J. | Methods of interacting with a computer using a finger(s) touch sensing input device with visual feedback |
US20070152984A1 (en) * | 2005-12-30 | 2007-07-05 | Bas Ording | Portable electronic device with multi-touch input |
US20070177804A1 (en) * | 2006-01-30 | 2007-08-02 | Apple Computer, Inc. | Multi-touch gesture dictionary |
US20080005703A1 (en) * | 2006-06-28 | 2008-01-03 | Nokia Corporation | Apparatus, Methods and computer program products providing finger-based and hand-based gesture commands for portable electronic device applications |
US20080036743A1 (en) * | 1998-01-26 | 2008-02-14 | Apple Computer, Inc. | Gesturing with a multipoint sensing device |
US20080168403A1 (en) * | 2007-01-06 | 2008-07-10 | Appl Inc. | Detecting and interpreting real-world and security gestures on touch and hover sensitive devices |
US20090021475A1 (en) * | 2007-07-20 | 2009-01-22 | Wolfgang Steinle | Method for displaying and/or processing image data of medical origin using gesture recognition |
US20090051671A1 (en) * | 2007-08-22 | 2009-02-26 | Jason Antony Konstas | Recognizing the motion of two or more touches on a touch-sensing surface |
US20090073194A1 (en) * | 2007-01-07 | 2009-03-19 | Bas Ording | Device, Method, and Graphical User Interface for List Scrolling on a Touch-Screen Display |
US20110154268A1 (en) * | 2009-12-18 | 2011-06-23 | Synaptics Incorporated | Method and apparatus for operating in pointing and enhanced gesturing modes |
US20110169750A1 (en) * | 2010-01-14 | 2011-07-14 | Continental Automotive Systems, Inc. | Multi-touchpad multi-touch user interface |
US8054296B2 (en) * | 2007-01-03 | 2011-11-08 | Apple Inc. | Storing baseline information in EEPROM |
Family Cites Families (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH06282368A (en) * | 1993-02-01 | 1994-10-07 | Wacom Co Ltd | Position information input system for information processor |
EP0622722B1 (en) * | 1993-04-30 | 2002-07-17 | Xerox Corporation | Interactive copying system |
US7046230B2 (en) * | 2001-10-22 | 2006-05-16 | Apple Computer, Inc. | Touch pad handheld device |
US8373660B2 (en) * | 2003-07-14 | 2013-02-12 | Matt Pallakoff | System and method for a portable multimedia client |
US8446373B2 (en) * | 2008-02-08 | 2013-05-21 | Synaptics Incorporated | Method and apparatus for extended adjustment based on relative positioning of multiple objects contemporaneously in a sensing region |
JP5164675B2 (en) * | 2008-06-04 | 2013-03-21 | キヤノン株式会社 | User interface control method, information processing apparatus, and program |
JP5191314B2 (en) * | 2008-08-28 | 2013-05-08 | アルパイン株式会社 | Music selection method and apparatus |
KR101586627B1 (en) * | 2008-10-06 | 2016-01-19 | 삼성전자주식회사 | A method for controlling of list with multi touch and apparatus thereof |
KR101503835B1 (en) * | 2008-10-13 | 2015-03-18 | 삼성전자주식회사 | Apparatus and method for object management using multi-touch |
US8363020B2 (en) * | 2009-08-27 | 2013-01-29 | Symbol Technologies, Inc. | Methods and apparatus for pressure-based manipulation of content on a touch screen |
US8390583B2 (en) * | 2009-08-31 | 2013-03-05 | Qualcomm Incorporated | Pressure sensitive user interface for mobile devices |
US8436821B1 (en) * | 2009-11-20 | 2013-05-07 | Adobe Systems Incorporated | System and method for developing and classifying touch gestures |
JP5750875B2 (en) * | 2010-12-01 | 2015-07-22 | ソニー株式会社 | Information processing apparatus, information processing method, and program |
US8839150B2 (en) * | 2010-02-10 | 2014-09-16 | Apple Inc. | Graphical objects that respond to touch or motion input |
JP5657269B2 (en) * | 2010-04-26 | 2015-01-21 | シャープ株式会社 | Image processing apparatus, display apparatus, image processing method, image processing program, and recording medium |
KR20110128220A (en) * | 2010-05-21 | 2011-11-29 | 옥윤선 | Calculation method interminal heving touch pad |
KR101498457B1 (en) * | 2010-10-22 | 2015-03-05 | 에스케이플래닛 주식회사 | Terminal and method for compression of files using multi touch |
-
2010
- 2010-12-01 JP JP2010268479A patent/JP5750875B2/en active Active
-
2011
- 2011-11-16 US US13/297,777 patent/US20120062603A1/en not_active Abandoned
- 2011-11-24 CN CN201110378036.XA patent/CN102591564B/en active Active
-
2016
- 2016-09-08 US US15/260,204 patent/US11281324B2/en active Active
-
2022
- 2022-02-10 US US17/668,880 patent/US20220164059A1/en active Pending
Patent Citations (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20080036743A1 (en) * | 1998-01-26 | 2008-02-14 | Apple Computer, Inc. | Gesturing with a multipoint sensing device |
US20050162402A1 (en) * | 2004-01-27 | 2005-07-28 | Watanachote Susornpol J. | Methods of interacting with a computer using a finger(s) touch sensing input device with visual feedback |
US20070152984A1 (en) * | 2005-12-30 | 2007-07-05 | Bas Ording | Portable electronic device with multi-touch input |
US20070177804A1 (en) * | 2006-01-30 | 2007-08-02 | Apple Computer, Inc. | Multi-touch gesture dictionary |
US20080005703A1 (en) * | 2006-06-28 | 2008-01-03 | Nokia Corporation | Apparatus, Methods and computer program products providing finger-based and hand-based gesture commands for portable electronic device applications |
US8054296B2 (en) * | 2007-01-03 | 2011-11-08 | Apple Inc. | Storing baseline information in EEPROM |
US20080168403A1 (en) * | 2007-01-06 | 2008-07-10 | Appl Inc. | Detecting and interpreting real-world and security gestures on touch and hover sensitive devices |
US20090073194A1 (en) * | 2007-01-07 | 2009-03-19 | Bas Ording | Device, Method, and Graphical User Interface for List Scrolling on a Touch-Screen Display |
US20090021475A1 (en) * | 2007-07-20 | 2009-01-22 | Wolfgang Steinle | Method for displaying and/or processing image data of medical origin using gesture recognition |
US20090051671A1 (en) * | 2007-08-22 | 2009-02-26 | Jason Antony Konstas | Recognizing the motion of two or more touches on a touch-sensing surface |
US20110154268A1 (en) * | 2009-12-18 | 2011-06-23 | Synaptics Incorporated | Method and apparatus for operating in pointing and enhanced gesturing modes |
US20110169750A1 (en) * | 2010-01-14 | 2011-07-14 | Continental Automotive Systems, Inc. | Multi-touchpad multi-touch user interface |
Cited By (39)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9007190B2 (en) | 2010-03-31 | 2015-04-14 | Tk Holdings Inc. | Steering wheel sensors |
US8587422B2 (en) | 2010-03-31 | 2013-11-19 | Tk Holdings, Inc. | Occupant sensing system |
US8725230B2 (en) | 2010-04-02 | 2014-05-13 | Tk Holdings Inc. | Steering wheel with hand sensors |
US20140043535A1 (en) * | 2011-04-04 | 2014-02-13 | Sharp Kabushiki Kaisha | Display apparatus, information processing system and recording medium |
US20130127754A1 (en) * | 2011-11-17 | 2013-05-23 | Samsung Electronics Co., Ltd. | Display apparatus and control method thereof |
US9727031B2 (en) | 2012-04-13 | 2017-08-08 | Tk Holdings Inc. | Pressure sensor including a pressure sensitive material for use with control systems and methods of using the same |
CN108829333A (en) * | 2012-04-19 | 2018-11-16 | 索尼公司 | Information processing unit |
CN104220978A (en) * | 2012-04-19 | 2014-12-17 | 索尼公司 | Information processing apparatus, information processing method, program, and information processing system |
US20150040070A1 (en) * | 2012-04-19 | 2015-02-05 | Sony Corporation | Information processing apparatus, information processing method, program, and information processing system |
US10942620B2 (en) | 2012-04-19 | 2021-03-09 | Sony Corporation | Information processing apparatus, information processing method, program, and information processing system |
US10162480B2 (en) | 2012-04-19 | 2018-12-25 | Sony Corporation | Information processing apparatus, information processing method, program, and information processing system |
US9772744B2 (en) * | 2012-04-19 | 2017-09-26 | Sony Corporation | Information processing apparatus, information processing method, program, and information processing system |
US9487388B2 (en) | 2012-06-21 | 2016-11-08 | Nextinput, Inc. | Ruggedized MEMS force die |
US9493342B2 (en) | 2012-06-21 | 2016-11-15 | Nextinput, Inc. | Wafer level MEMS force dies |
US9032818B2 (en) | 2012-07-05 | 2015-05-19 | Nextinput, Inc. | Microelectromechanical load sensor and methods of manufacturing the same |
US9696223B2 (en) | 2012-09-17 | 2017-07-04 | Tk Holdings Inc. | Single layer force sensor |
US9902611B2 (en) | 2014-01-13 | 2018-02-27 | Nextinput, Inc. | Miniaturized and ruggedized wafer level MEMs force sensors |
US10599252B2 (en) | 2014-06-25 | 2020-03-24 | Huawei Technologies Co., Ltd. | Intelligent terminal control method utilizing touch contact location and pressure |
US20160117081A1 (en) * | 2014-10-27 | 2016-04-28 | Thales Avionics, Inc. | Controlling entertainment system using combination of inputs from proximity sensor and touch sensor of remote controller |
US10466119B2 (en) | 2015-06-10 | 2019-11-05 | Nextinput, Inc. | Ruggedized wafer level MEMS force sensor with a tolerance trench |
WO2017032079A1 (en) * | 2015-08-27 | 2017-03-02 | 广东欧珀移动通信有限公司 | Information browsing method and mobile terminal |
US11946817B2 (en) | 2017-02-09 | 2024-04-02 | DecaWave, Ltd. | Integrated digital force sensors and related methods of manufacture |
US11808644B2 (en) | 2017-02-09 | 2023-11-07 | Qorvo Us, Inc. | Integrated piezoresistive and piezoelectric fusion force sensor |
US11243125B2 (en) | 2017-02-09 | 2022-02-08 | Nextinput, Inc. | Integrated piezoresistive and piezoelectric fusion force sensor |
US11255737B2 (en) | 2017-02-09 | 2022-02-22 | Nextinput, Inc. | Integrated digital force sensors and related methods of manufacture |
US11604104B2 (en) | 2017-02-09 | 2023-03-14 | Qorvo Us, Inc. | Integrated piezoresistive and piezoelectric fusion force sensor |
US11221263B2 (en) | 2017-07-19 | 2022-01-11 | Nextinput, Inc. | Microelectromechanical force sensor having a strain transfer layer arranged on the sensor die |
US11423686B2 (en) | 2017-07-25 | 2022-08-23 | Qorvo Us, Inc. | Integrated fingerprint and force sensor |
US11243126B2 (en) | 2017-07-27 | 2022-02-08 | Nextinput, Inc. | Wafer bonded piezoresistive and piezoelectric force sensor and related methods of manufacture |
US11609131B2 (en) | 2017-07-27 | 2023-03-21 | Qorvo Us, Inc. | Wafer bonded piezoresistive and piezoelectric force sensor and related methods of manufacture |
US11946816B2 (en) | 2017-07-27 | 2024-04-02 | Nextinput, Inc. | Wafer bonded piezoresistive and piezoelectric force sensor and related methods of manufacture |
US11579028B2 (en) | 2017-10-17 | 2023-02-14 | Nextinput, Inc. | Temperature coefficient of offset compensation for force sensor and strain gauge |
US11898918B2 (en) | 2017-10-17 | 2024-02-13 | Nextinput, Inc. | Temperature coefficient of offset compensation for force sensor and strain gauge |
US11385108B2 (en) | 2017-11-02 | 2022-07-12 | Nextinput, Inc. | Sealed force sensor with etch stop layer |
US11965787B2 (en) | 2017-11-02 | 2024-04-23 | Nextinput, Inc. | Sealed force sensor with etch stop layer |
US11874185B2 (en) | 2017-11-16 | 2024-01-16 | Nextinput, Inc. | Force attenuator for force sensor |
US11698310B2 (en) | 2019-01-10 | 2023-07-11 | Nextinput, Inc. | Slotted MEMS force sensor |
US10962427B2 (en) | 2019-01-10 | 2021-03-30 | Nextinput, Inc. | Slotted MEMS force sensor |
US11194471B1 (en) | 2021-01-28 | 2021-12-07 | Honda Motor Co., Ltd. | Apparatus and method for display control based on touch interface |
Also Published As
Publication number | Publication date |
---|---|
CN102591564A (en) | 2012-07-18 |
JP5750875B2 (en) | 2015-07-22 |
CN102591564B (en) | 2017-05-10 |
JP2012118796A (en) | 2012-06-21 |
US20220164059A1 (en) | 2022-05-26 |
US11281324B2 (en) | 2022-03-22 |
US20160378248A1 (en) | 2016-12-29 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20220164059A1 (en) | Information processing apparatus, information processing method, and program therefor | |
US20210089201A1 (en) | Apparatus and method of copying and pasting content in a computing device | |
US9389722B2 (en) | User interface device that zooms image in response to operation that presses screen, image zoom method, and program | |
EP2752755B1 (en) | Information processing apparatus, information processing method, and computer program | |
US8610678B2 (en) | Information processing apparatus and method for moving a displayed object between multiple displays | |
US10282081B2 (en) | Input and output method in touch screen terminal and apparatus therefor | |
JP5523090B2 (en) | INPUT DEVICE, CONTROL METHOD FOR INPUT DEVICE, PROGRAM, AND STORAGE MEDIUM | |
US9690475B2 (en) | Information processing apparatus, information processing method, and program | |
US9448714B2 (en) | Touch and non touch based interaction of a user with a device | |
US20140362016A1 (en) | Electronic book display device that performs page turning in response to user operation pressing screen, page turning method, and program | |
US20230342025A1 (en) | Simultaneous gesture and touch control on a display | |
WO2015085919A1 (en) | Clicked object magnifying method and apparatus based on floating touch | |
JP6053332B2 (en) | Information processing apparatus, information processing apparatus control method, and program | |
JP5861638B2 (en) | Display device, display method, program thereof, and terminal device | |
KR102323440B1 (en) | Method for operating page and electronic device thereof | |
JP6034281B2 (en) | Object selection method, apparatus, and computer program | |
US10963137B2 (en) | Information display apparatus and non-transitory recording medium storing program for controlling information display apparatus | |
JP2017045298A (en) | User interface of electronic device, input processing method, and electronic device | |
CN113064532B (en) | Content transmitting method, device, electronic equipment and storage medium | |
KR101436805B1 (en) | Method and apparatus for selecting multiple objects on a touch-screen display | |
JP6971573B2 (en) | Electronic devices, their control methods and programs | |
JP2015102946A (en) | Information processing apparatus, control method of information processing apparatus, and program | |
JP6210664B2 (en) | Information processing apparatus, control method therefor, program, and storage medium | |
JP2018116605A (en) | Display control device and display control method | |
CN117193574A (en) | Split screen control method, device, equipment and storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: SONY CORPORATION, JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:MIZUNUMA, HIROYUKI;YAMAMOTO, KAZUYUKI;SATOH, NARIAKI;AND OTHERS;SIGNING DATES FROM 20111028 TO 20111105;REEL/FRAME:027239/0859 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |