US10768740B2 - Input axis rotations - Google Patents
Input axis rotations Download PDFInfo
- Publication number
- US10768740B2 US10768740B2 US15/772,470 US201615772470A US10768740B2 US 10768740 B2 US10768740 B2 US 10768740B2 US 201615772470 A US201615772470 A US 201615772470A US 10768740 B2 US10768740 B2 US 10768740B2
- Authority
- US
- United States
- Prior art keywords
- touch
- axis
- input
- sensitive
- orientation
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 230000004044 response Effects 0.000 claims abstract description 15
- 238000005259 measurement Methods 0.000 claims 1
- 238000000034 method Methods 0.000 description 8
- 230000001960 triggered effect Effects 0.000 description 4
- 230000003287 optical effect Effects 0.000 description 2
- 230000003247 decreasing effect Effects 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 230000001902 propagating effect Effects 0.000 description 1
- 238000010079 rubber tapping Methods 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/041—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
- G06F3/0416—Control or interface arrangements specially adapted for digitisers
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/041—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
- G06F3/042—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by opto-electronic means
- G06F3/0421—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by opto-electronic means by interrupting or reflecting a light beam, e.g. optical touch-screen
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
- G06F3/04883—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2203/00—Indexing scheme relating to G06F3/00 - G06F3/048
- G06F2203/033—Indexing scheme relating to G06F3/033
- G06F2203/0339—Touch strips, e.g. orthogonal touch strips to control cursor movement or scrolling; single touch strip to adjust parameter or to implement a row of soft keys
Definitions
- An electronic device may be attached to an input device to receive input from a user of the computing device.
- an input device may be a keyboard or a computer mouse.
- Some electronic devices, such as a notebook computer, may have an integrated input device, such as an integrated keyboard.
- FIG. 1 illustrates an electronic device with a touch-sensitive input device that rotates an input axis based on a re-orientation input, according to an example
- FIGS. 2A-2B illustrate a process of the touch-sensitive input device of FIG. 1 rotating an input axis based on a re-orientation input, according to an example
- FIG. 3 illustrates an electronic device with a touch-sensitive input device that rotates an input axis based on a triggering of a sensor, according to an example
- FIG. 4 illustrates an electronic device with a touch-sensitive input device that rotates an input axis based on a triggering of a sensor, according to another example
- FIGS. 5A-5B illustrate a process of a touch-sensitive input device rotating an input axis based on a triggering of a sensor, according to an example
- FIG. 5C illustrates a process of a touch-sensitive input device rotating an input axis based on a triggering of a sensor, according to another example
- FIG. 5D illustrates a process of a touch-sensitive input device rotating an input axis based on a triggering of a sensor, according to another example.
- FIG. 6 illustrates a touch-sensitive input device that rotates an input axis based on a triggering of a sensor, according to another example.
- a touch-sensitive input device may translate the movements of a writing tool (e.g., a user's finger or a stylus) to a relative position in a graphical user interface (GUI).
- GUI graphical user interface
- An example of a touch-sensitive input device may be a touchpad.
- Another touch-sensitive input device may be a touchscreen.
- a touch-sensitive input device may include an input axis to translate a movement of a writing tool to a relative position in a GUI.
- the input axis may be a fixed input axis designed to face a sole, individual user.
- the touch-sensitive input device may not translate the movements from the user's finger to the correct position in a GUI.
- the convenience of using the touch-sensitive input device may be decreased.
- Examples described herein provide a touch-sensitive input device that rotates an input axis to align the input axis to a user of the touch-sensitive input device.
- a non-transitory computer readable storage medium comprising instructions that when executed cause a controller of an electronic device to receive, via a touch-sensitive input device of the electronic device, a touch input.
- the instructions also cause the controller to determine whether the touch input corresponds to a re-orientation input.
- the instructions further cause the controller to, in response to a determination that the touch input corresponds to the re-orientation input, rotate an input axis associated with the touch sensitive input device from a first orientation to a second orientation based on the re-orientation input.
- a non-transitory computer-readable storage medium comprising instructions that when executed cause a controller of an electronic device to detect, via a light-sensitive sensor of the electronic device, a direction of a user gesture relative to a side of a touch-sensitive input device of the electronic device.
- the instructions also cause the controller to rotate an input axis associated with the touch-sensitive input device from a first orientation to a second orientation based on the detected direction, where a vertical axis of the input axis is aligned with the side in the second orientation.
- FIG. 1 illustrates an electronic device 100 with a touch-sensitive input device that rotates an input axis based on a re-orientation input, according to an example.
- Electronic device 100 may be, for example, a desktop computer, an all-in-one computer, a tablet computing device, a mobile phone, a display, or any other electronic device suitable to receive touch inputs.
- Electronic device 100 may include a controller 102 , a computer-readable storage medium 104 , and a touch-sensitive input device 106 .
- Controller 102 may be a central processing unit (CPU), a semiconductor-based microprocessor, and/or other hardware devices suitable for retrieval and execution of instructions stored in computer-readable storage medium 104 . Controller 102 may fetch, decode, and execute instructions 108 , 110 , and 112 to control a process of rotating an input axis of touch-sensitive input device 106 . As an alternative or in addition to retrieving and executing instructions, controller 102 may include at least one electronic circuit that includes electronic components for performing the functionality of instructions 108 , 110 , 112 , or a combination thereof.
- Computer-readable storage medium 104 may be any electronic, magnetic, optical, or other physical storage device that contains or stores executable instructions.
- computer-readable storage medium 104 may be, for example, Random Access Memory (RAM), an Electrically Erasable Programmable Read-Only Memory (EEPROM), a storage device, an optical disc, etc.
- RAM Random Access Memory
- EEPROM Electrically Erasable Programmable Read-Only Memory
- computer-readable storage medium 104 may be a non-transitory storage medium, where the term “non-transitory” does not encompass transitory propagating signals.
- computer-readable storage medium 104 may be encoded with a series of processor executable instructions 108 - 112 .
- Touch-sensitive input device 106 may be a touchpad, a touchscreen, or any other electronic device suitable to translate the movements of a writing tool (e.g., a user's finger or a stylus) to a relative position in a GUI based on an input axis.
- a writing tool e.g., a user's finger or a stylus
- An example of an input axis is described in more detail in FIGS. 2A-2B .
- Touch input reception instructions 108 may receive a touch input 114 via touch-sensitive input device 106 .
- Re-orientation input determination instructions 110 may determine whether touch input 114 corresponds to a re-orientation input.
- a re-orientation input may have a distinct characteristic.
- a re-orientation input may correspond to a tapping gesture on touch-sensitive input device 106 .
- a re-orientation input may correspond to a two finger swipe from one side of touch-sensitive input device 106 to another side of touch-sensitive input device 106 .
- re-orientation input determination instructions 110 may determine that a re-orientation input has been received.
- input axis rotation instructions 112 may rotate the input axis of touch-sensitive input device 106 from a first orientation to a second orientation based on the re-orientation input.
- input axis rotation instructions 112 may identify a location of the re-orientation input in touch-sensitive input device 106 .
- Input axis rotation instructions 112 may determine an anchor point based on the location and rotate the input axis based on the anchor point from a first orientation to a second orientation. A vertical axis of the input axis may be aligned with the anchor point in the second orientation.
- FIGS. 2A-2B illustrate a process of touch-sensitive input device 106 rotating an input axis based on a re-orientation input, according to an example.
- touch-sensitive input device 106 may include an input axis 202 .
- Input axis 202 may be virtual and may be implemented using instructions executable by a controller, such as controller 102 of FIG. 1 .
- input axis 202 may not be visible to a user of touch-sensitive input device 106 .
- Input axis 202 may be a two dimensional axis.
- Input axis 202 may include a horizontal axis 204 and a vertical axis 206 .
- Input axis 202 may determine how movements detected on touch-sensitive input device 106 is translated to a relative position in a GUI 208 .
- GUI 208 may be shown in a display attached to electronic device 100 .
- input axis 202 may be in a first orientation as shown in FIG. 2A . In the first orientation, input axis 202 may be aligned with a first user 210 .
- a pointer 214 may move across GUI 208 from a first position to a second position (as shown by an arrow 216 ).
- second user 218 may re-orient input axis 202 to better suit second user's 218 position relative to touch-sensitive input device 106 .
- second user 218 may physically face a corner of touch-sensitive input device 106 and may not be aligned with input axis 202 in the first orientation.
- Second user 218 may provide a re-orientation input 220 to touch-sensitive input device 106 to re-align input axis 202 with second user 218 .
- Touch-sensitive input device 106 may identify that re-orientation input 220 is received at a location 222 in touch-sensitive input device 106 .
- Location 222 may correspond to a region of touch-sensitive input device 106 where re-orientation input 220 initially makes physical contact with touch-sensitive input device 106 .
- Touch-sensitive input device 106 may determine an anchor point 224 based on location 222 .
- Anchor point 224 may be a location in touch-sensitive input device 106 that serves as a reference point to re-orient input axis 202 .
- anchor point 224 may be a center of location 222 .
- touch-sensitive input device 106 may rotate input axis 202 to a second orientation based on anchor point 224 while touch-sensitive input device 106 and electronic device 100 are to remain stationary.
- vertical axis 206 of input axis 202 may align with anchor point 224 such that a proximal side 226 of vertical axis 206 is to intersect anchor point 224 .
- Proximal side 226 may be a side of vertical axis 206 that is below horizontal axis 204 .
- input axis 202 is aligned with second user 218 .
- second user 218 may navigate GUI 208 without having to physically re-orient touch-sensitive input device 106 .
- second user 218 may move a finger across touch-sensitive input device 106 (as shown by an arrow 230 ).
- FIG. 3 illustrates an electronic device 300 with a touch-sensitive input device that rotates an input axis based on a triggering of a sensor, according to an example.
- Electronic device 300 may be similar to electronic device 100 of FIG. 1 .
- Electronic device 300 may include controller 102 , touch-sensitive input device 106 , a computer-readable storage medium 302 , and a set of sensors 304 .
- Computer-readable storage medium 302 may be similar to computer-readable storage medium 104 .
- a set of sensors 304 may include a plurality of sensors. Examples of set of sensors 304 are described in more detail in FIGS. 4 and 6 .
- Set of sensors 304 may be light-sensitive sensors that surround touch-sensitive input device 106 .
- Computer-readable storage medium 302 may be encoded with a series of processor executable instructions 306 - 308 .
- User gesture direction detection instructions 306 may detect a direction of a user gesture relative to touch-sensitive input device 106 based on a triggering of a sensor in set of sensors 304 .
- Input axis rotation instructions 308 may determine an anchor point based on the detected direction.
- Input axis rotation instructions 308 may also rotate an input axis of touch-sensitive input device 106 based on the anchor point.
- FIG. 4 illustrates an electronic device 400 with a touch-sensitive input device that rotates an input axis based on a triggering of a sensor, according to another example.
- Electronic device 400 may implement electronic device 300 of FIG. 3 .
- Electronic device 400 may include controller 102 , a touch-sensitive input device 402 , and a plurality of sensors 404 - 410 that surrounds touch-sensitive input device 402 .
- Touch-sensitive input device 402 may be similar to touch-sensitive input device 106 .
- Each of sensors 404 - 410 may be a light-sensitive sensor that is triggered (e.g., generates a voltage) in response to a change in light intensity detected at the light-sensitive sensor.
- each of sensors 404 - 410 may be implemented using a photo transistor that detects shadow.
- Touch-sensitive input device 402 may have a square sharp or a rectangular shape. Touch-sensitive input device 402 may have a first side 412 , a second side 414 , a third side 416 , and a fourth side 418 . Each of sides 412 - 418 may be aligned with a distinct sensor from the plurality of sensors 404 - 410 . For example, a first sensor 404 is aligned with first side 412 , a second sensor 406 is aligned with second side 414 , a third sensor 408 is aligned with third side 416 , and a fourth sensor 410 is aligned with fourth side 418 .
- sensors 404 - 410 may detect a direction of a user gesture relative to touch-sensitive input device 402 so that touch-sensitive input device 402 may rotate to align with the direction of the user gesture, as described in more detail in FIGS. 5A-5D .
- FIGS. 5A-5B illustrate a process of touch-sensitive input device 402 rotating an input axis based on a triggering of a sensor, according to an example.
- touch-sensitive input device 402 may include input axis 202 .
- Input axis 202 may be in a first orientation.
- a user 502 of touch-sensitive input device 402 facing second side 414 of touch-sensitive input device 402 may reach a hand or a finger towards touch-sensitive input device 402 .
- second sensor 406 may detect a shadow caused by the hand or finger.
- second sensor 406 may be triggered.
- the triggering of second sensor 406 may indicate that the direction of the hand or finger is coming from second side 414 .
- touch-sensitive input device 402 may rotate input axis 202 , as described in more detail in FIG. 5B .
- Touch-sensitive input device 402 may determine an anchor point 504 based on the detected direction. For example, since second sensor 406 is triggered, touch-sensitive input device 402 may determine that the direction of user's 502 hand or finger is coming from second side 414 . In some examples, anchor point 504 may be located at a mid-point of second side 414 . In some examples, anchor point 504 may be located anywhere in a region of touch-sensitive input device 402 that is aligned with second sensor 406 . Touch-sensitive input device 402 may rotate input axis 202 from the first orientation to a second orientation to align vertical axis 206 with anchor point 504 . In the second orientation, proximal side 226 may intersect with anchor point 504 .
- FIG. 5C another example of touch-sensitive input device 402 rotating input axis 202 based on a triggering of a sensor is described.
- user 502 may face fourth side 418 of touch-sensitive input device 402 .
- fourth sensor 410 may be triggered.
- touch-sensitive input device 402 may determine that the direction of user's 502 hand or finger is coming from fourth side 418 .
- touch-sensitive input device 402 may determine an anchor point 506 based on the detected direction.
- Touch-sensitive input device 402 may rotate input axis 202 from the first orientation to a third orientation to align vertical axis 206 with anchor point 506 . In the third orientation, proximal side 226 may intersect with anchor point 506 .
- Electronic device 400 may further include a set of indicators 508 , 510 , 512 , and 514 .
- Each of indicators 508 - 514 may be aligned with a corresponding side of touch-sensitive input device 402 .
- a first indicator 508 may be aligned with first side 412
- a second indicator 510 may be aligned with second side 414
- a third indicator 512 may be aligned with third side 416
- a fourth indicator 514 may be aligned with fourth side 418 .
- Indicators 508 - 514 may be implemented using light emitting diodes. Indicators 508 - 514 may provide a visual indication of which side of touch-sensitive input device 402 proximal side 226 is aligned with. Thus, a user of touch-sensitive input device 402 may determine the current orientation of touch-sensitive input device 402 using indicators 508 - 514 . For example, proximal side 226 may be aligned with third side 416 initially. Third indicator 512 may light up. Input axis 202 may be rotated as described in FIGS. 5A-5C and proximal side 226 may be aligned with fourth side 418 after the rotation. Thus, third indicator 512 may turn dark and fourth indicator 514 may light up.
- FIG. 6 illustrates a touch-sensitive input device 600 that rotates an input axis based on a triggering of a sensor, according to another example.
- Touch-sensitive input device 600 may have a circular shape.
- a band of sensors 602 may surround touch-sensitive input device 600 .
- Band of sensors 602 may include a plurality of sensors, such as sensors 604 , 606 , and 608 .
- Band of sensors may be light-sensitive sensors.
- Operations of touch-sensitive input device 600 and band of sensors 602 may be similar to touch-sensitive input device 402 and set of sensors 404 - 410 of FIG. 4 . However, compared to set of sensors 404 - 410 , band of sensors 602 may include more sensors. Thus, direction of gesture from a user may be more precisely detected and an anchor point to align an input axis of touch-sensitive input device 600 may be more precisely placed in touch-sensitive input device 600 .
- touch-sensitive input device 600 may determine an anchor in different manners based on the number of the contiguous sensors.
- the middle sensor among the contiguous sensors may define an anchor point.
- a halfway point between the middle two sensors may define the anchor point.
Landscapes
- Engineering & Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
Description
Claims (20)
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/US2016/020693 WO2017151136A1 (en) | 2016-03-03 | 2016-03-03 | Input axis rotations |
Publications (2)
Publication Number | Publication Date |
---|---|
US20180329564A1 US20180329564A1 (en) | 2018-11-15 |
US10768740B2 true US10768740B2 (en) | 2020-09-08 |
Family
ID=59744250
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/772,470 Active US10768740B2 (en) | 2016-03-03 | 2016-03-03 | Input axis rotations |
Country Status (4)
Country | Link |
---|---|
US (1) | US10768740B2 (en) |
EP (1) | EP3362884A4 (en) |
CN (1) | CN108292192A (en) |
WO (1) | WO2017151136A1 (en) |
Citations (38)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6567101B1 (en) | 1999-10-13 | 2003-05-20 | Gateway, Inc. | System and method utilizing motion input for manipulating a display of data |
US20070063987A1 (en) | 2005-09-21 | 2007-03-22 | Alps Electric Co., Ltd. | Input device |
US20070300182A1 (en) * | 2006-06-22 | 2007-12-27 | Microsoft Corporation | Interface orientation using shadows |
US20090100343A1 (en) * | 2007-10-10 | 2009-04-16 | Samsung Electronics Co. Ltd. | Method and system for managing objects in a display environment |
US20090184936A1 (en) | 2008-01-22 | 2009-07-23 | Mathematical Inventing - Slicon Valley | 3D touchpad |
US20100053109A1 (en) * | 2008-08-29 | 2010-03-04 | Tomoya Narita | Information Processing Apparatus and Information Processing Method |
US20100177931A1 (en) * | 2009-01-15 | 2010-07-15 | Microsoft Corporation | Virtual object adjustment via physical object detection |
US20110001628A1 (en) * | 2009-07-03 | 2011-01-06 | Sony Corporation | Map information display device, map information display method and program |
US20110016749A1 (en) | 2009-07-21 | 2011-01-27 | Reebok International Ltd. | Article Of Footwear And Methods Of Making Same |
US20110102333A1 (en) * | 2009-10-30 | 2011-05-05 | Wayne Carl Westerman | Detection of Gesture Orientation on Repositionable Touch Surface |
US20110134047A1 (en) * | 2009-12-04 | 2011-06-09 | Microsoft Corporation | Multi-modal interaction on multi-touch display |
US20110169749A1 (en) * | 2010-01-13 | 2011-07-14 | Lenovo (Singapore) Pte, Ltd. | Virtual touchpad for a touch device |
US20110279384A1 (en) | 2010-05-14 | 2011-11-17 | Google Inc. | Automatic Derivation of Analogous Touch Gestures From A User-Defined Gesture |
US20120001858A1 (en) * | 2010-06-30 | 2012-01-05 | Kabushiki Kaisha Toshiba | Information processor, information processing method, and computer program product |
US20120038546A1 (en) * | 2010-08-10 | 2012-02-16 | Daryl Cromer | Gesture control |
EP2437147A1 (en) | 2009-05-26 | 2012-04-04 | Sony Corporation | Information processing device, information processing method, and program |
US8402391B1 (en) * | 2008-09-25 | 2013-03-19 | Apple, Inc. | Collaboration system |
US20130083074A1 (en) * | 2011-10-03 | 2013-04-04 | Nokia Corporation | Methods, apparatuses and computer program products utilizing hovering, in part, to determine user interface orientation |
US20130125045A1 (en) | 2011-11-16 | 2013-05-16 | Samsung Electronics Co. Ltd. | Apparatus including a touch screen under a multiapplication environment and controlling method thereof |
US20130127750A1 (en) * | 2011-11-18 | 2013-05-23 | International Business Machines Corporation | Facilitating operation of controls displayed in a display surface independently of the size of the display surface |
US8508475B2 (en) * | 2008-10-24 | 2013-08-13 | Microsoft Corporation | User interface elements positioned for display |
US20130222275A1 (en) | 2012-02-29 | 2013-08-29 | Research In Motion Limited | Two-factor rotation input on a touchscreen device |
US20130229361A1 (en) * | 2012-03-05 | 2013-09-05 | International Business Machines Corporation | Computer action detection |
US8553001B2 (en) * | 2011-03-22 | 2013-10-08 | Adobe Systems Incorporated | Methods and apparatus for determining local coordinate frames for a human hand |
US20130275907A1 (en) * | 2010-10-14 | 2013-10-17 | University of Technology ,Sydney | Virtual keyboard |
US20140015792A1 (en) * | 2012-07-11 | 2014-01-16 | Wistron Corp. | Frame with sensing function and touch control method |
US8749493B2 (en) | 2003-08-18 | 2014-06-10 | Apple Inc. | Movable touch pad with added functionality |
US20140168118A1 (en) * | 2012-12-17 | 2014-06-19 | Hon Hai Precision Industry Co., Ltd. | Touch display device and control method thereof |
US20140218290A1 (en) | 2013-02-07 | 2014-08-07 | Universal Electronics Inc. | System and methods for providing orientation compensation in pointing devices |
US20140354695A1 (en) * | 2012-01-13 | 2014-12-04 | Sony Corporation | Information processing apparatus and information processing method, and computer program |
WO2015042444A1 (en) | 2013-09-20 | 2015-03-26 | Handscape Inc. | Method for controlling a control region of a computerized device from a touchpad |
US9128552B2 (en) * | 2013-07-17 | 2015-09-08 | Lenovo (Singapore) Pte. Ltd. | Organizing display data on a multiuser display |
US20150293616A1 (en) | 2014-04-09 | 2015-10-15 | Wei-Chih Cheng | Operating system with shortcut touch panel having shortcut function |
US20150370414A1 (en) * | 2014-06-20 | 2015-12-24 | International Business Machines Corporation | Touch panel information processing |
US9223340B2 (en) * | 2013-08-14 | 2015-12-29 | Lenovo (Singapore) Pte. Ltd. | Organizing display data on a multiuser display |
US20160109861A1 (en) * | 2014-09-11 | 2016-04-21 | Samsung Electronics Co., Ltd. | Wearable Device |
US9417733B2 (en) * | 2011-12-21 | 2016-08-16 | Wistron Corporation | Touch method and touch system |
US20170118402A1 (en) * | 2015-10-22 | 2017-04-27 | Samsung Electronics Co., Ltd | Electronic device and camera control method therefor |
Family Cites Families (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8060840B2 (en) * | 2005-12-29 | 2011-11-15 | Microsoft Corporation | Orientation free user interface |
US8872854B1 (en) * | 2011-03-24 | 2014-10-28 | David A. Levitt | Methods for real-time navigation and display of virtual worlds |
-
2016
- 2016-03-03 EP EP16892883.6A patent/EP3362884A4/en not_active Ceased
- 2016-03-03 CN CN201680069439.7A patent/CN108292192A/en active Pending
- 2016-03-03 WO PCT/US2016/020693 patent/WO2017151136A1/en active Application Filing
- 2016-03-03 US US15/772,470 patent/US10768740B2/en active Active
Patent Citations (39)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6567101B1 (en) | 1999-10-13 | 2003-05-20 | Gateway, Inc. | System and method utilizing motion input for manipulating a display of data |
US8749493B2 (en) | 2003-08-18 | 2014-06-10 | Apple Inc. | Movable touch pad with added functionality |
US20070063987A1 (en) | 2005-09-21 | 2007-03-22 | Alps Electric Co., Ltd. | Input device |
US20070300182A1 (en) * | 2006-06-22 | 2007-12-27 | Microsoft Corporation | Interface orientation using shadows |
US20090100343A1 (en) * | 2007-10-10 | 2009-04-16 | Samsung Electronics Co. Ltd. | Method and system for managing objects in a display environment |
US20090184936A1 (en) | 2008-01-22 | 2009-07-23 | Mathematical Inventing - Slicon Valley | 3D touchpad |
US20100053109A1 (en) * | 2008-08-29 | 2010-03-04 | Tomoya Narita | Information Processing Apparatus and Information Processing Method |
US8402391B1 (en) * | 2008-09-25 | 2013-03-19 | Apple, Inc. | Collaboration system |
US8508475B2 (en) * | 2008-10-24 | 2013-08-13 | Microsoft Corporation | User interface elements positioned for display |
US20100177931A1 (en) * | 2009-01-15 | 2010-07-15 | Microsoft Corporation | Virtual object adjustment via physical object detection |
EP2437147A1 (en) | 2009-05-26 | 2012-04-04 | Sony Corporation | Information processing device, information processing method, and program |
US20120092283A1 (en) * | 2009-05-26 | 2012-04-19 | Reiko Miyazaki | Information processing apparatus, information processing method, and program |
US20110001628A1 (en) * | 2009-07-03 | 2011-01-06 | Sony Corporation | Map information display device, map information display method and program |
US20110016749A1 (en) | 2009-07-21 | 2011-01-27 | Reebok International Ltd. | Article Of Footwear And Methods Of Making Same |
US20110102333A1 (en) * | 2009-10-30 | 2011-05-05 | Wayne Carl Westerman | Detection of Gesture Orientation on Repositionable Touch Surface |
US20110134047A1 (en) * | 2009-12-04 | 2011-06-09 | Microsoft Corporation | Multi-modal interaction on multi-touch display |
US20110169749A1 (en) * | 2010-01-13 | 2011-07-14 | Lenovo (Singapore) Pte, Ltd. | Virtual touchpad for a touch device |
US20110279384A1 (en) | 2010-05-14 | 2011-11-17 | Google Inc. | Automatic Derivation of Analogous Touch Gestures From A User-Defined Gesture |
US20120001858A1 (en) * | 2010-06-30 | 2012-01-05 | Kabushiki Kaisha Toshiba | Information processor, information processing method, and computer program product |
US20120038546A1 (en) * | 2010-08-10 | 2012-02-16 | Daryl Cromer | Gesture control |
US20130275907A1 (en) * | 2010-10-14 | 2013-10-17 | University of Technology ,Sydney | Virtual keyboard |
US8553001B2 (en) * | 2011-03-22 | 2013-10-08 | Adobe Systems Incorporated | Methods and apparatus for determining local coordinate frames for a human hand |
US20130083074A1 (en) * | 2011-10-03 | 2013-04-04 | Nokia Corporation | Methods, apparatuses and computer program products utilizing hovering, in part, to determine user interface orientation |
US20130125045A1 (en) | 2011-11-16 | 2013-05-16 | Samsung Electronics Co. Ltd. | Apparatus including a touch screen under a multiapplication environment and controlling method thereof |
US20130127750A1 (en) * | 2011-11-18 | 2013-05-23 | International Business Machines Corporation | Facilitating operation of controls displayed in a display surface independently of the size of the display surface |
US9417733B2 (en) * | 2011-12-21 | 2016-08-16 | Wistron Corporation | Touch method and touch system |
US20140354695A1 (en) * | 2012-01-13 | 2014-12-04 | Sony Corporation | Information processing apparatus and information processing method, and computer program |
US20130222275A1 (en) | 2012-02-29 | 2013-08-29 | Research In Motion Limited | Two-factor rotation input on a touchscreen device |
US20130229361A1 (en) * | 2012-03-05 | 2013-09-05 | International Business Machines Corporation | Computer action detection |
US20140015792A1 (en) * | 2012-07-11 | 2014-01-16 | Wistron Corp. | Frame with sensing function and touch control method |
US20140168118A1 (en) * | 2012-12-17 | 2014-06-19 | Hon Hai Precision Industry Co., Ltd. | Touch display device and control method thereof |
US20140218290A1 (en) | 2013-02-07 | 2014-08-07 | Universal Electronics Inc. | System and methods for providing orientation compensation in pointing devices |
US9128552B2 (en) * | 2013-07-17 | 2015-09-08 | Lenovo (Singapore) Pte. Ltd. | Organizing display data on a multiuser display |
US9223340B2 (en) * | 2013-08-14 | 2015-12-29 | Lenovo (Singapore) Pte. Ltd. | Organizing display data on a multiuser display |
WO2015042444A1 (en) | 2013-09-20 | 2015-03-26 | Handscape Inc. | Method for controlling a control region of a computerized device from a touchpad |
US20150293616A1 (en) | 2014-04-09 | 2015-10-15 | Wei-Chih Cheng | Operating system with shortcut touch panel having shortcut function |
US20150370414A1 (en) * | 2014-06-20 | 2015-12-24 | International Business Machines Corporation | Touch panel information processing |
US20160109861A1 (en) * | 2014-09-11 | 2016-04-21 | Samsung Electronics Co., Ltd. | Wearable Device |
US20170118402A1 (en) * | 2015-10-22 | 2017-04-27 | Samsung Electronics Co., Ltd | Electronic device and camera control method therefor |
Non-Patent Citations (1)
Title |
---|
Heo, S, et al. "Designing Rich Touch Interaction Through Proximity and 2.5D Force Sensing Touchpad." Ozchi 13, Nov. 25-29, 2013. Adelaide, Australia. |
Also Published As
Publication number | Publication date |
---|---|
WO2017151136A1 (en) | 2017-09-08 |
EP3362884A1 (en) | 2018-08-22 |
US20180329564A1 (en) | 2018-11-15 |
EP3362884A4 (en) | 2019-06-26 |
CN108292192A (en) | 2018-07-17 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8217909B2 (en) | Multi-finger sub-gesture reporting for a user interface device | |
US20150268789A1 (en) | Method for preventing accidentally triggering edge swipe gesture and gesture triggering | |
US8466934B2 (en) | Touchscreen interface | |
US9423877B2 (en) | Navigation approaches for multi-dimensional input | |
US9348466B2 (en) | Touch discrimination using fisheye lens | |
US20140118268A1 (en) | Touch screen operation using additional inputs | |
US9965095B2 (en) | Information processing apparatus, information processing method and program | |
WO2019223461A1 (en) | Touch detection method and computer-readable storage medium | |
US9330249B2 (en) | Information terminal | |
EP3100151B1 (en) | Virtual mouse for a touch screen device | |
KR20140031254A (en) | Method for selecting an element of a user interface and device implementing such a method | |
JP6500041B2 (en) | Stochastic touch sensing | |
US20140082559A1 (en) | Control area for facilitating user input | |
US10073586B2 (en) | Method and system for mouse pointer to automatically follow cursor | |
US20130088427A1 (en) | Multiple input areas for pen-based computing | |
WO2018177156A1 (en) | Method for operating mobile terminal, and mobile terminal | |
AU2014242161B2 (en) | Off-center sensor target region | |
US9836082B2 (en) | Wearable electronic apparatus | |
US10895934B2 (en) | Information input device | |
US10768740B2 (en) | Input axis rotations | |
KR20140033726A (en) | Method and apparatus for distinguishing five fingers in electronic device including touch screen | |
US10303299B2 (en) | Use of groove analysis in a touch screen device to determine occurrence of an elongated touch by a single finger | |
US20150116281A1 (en) | Portable electronic device and control method | |
TWI668604B (en) | Electronic device and method for preventing unintentional touch | |
US20160266773A1 (en) | Information processing method and electronic device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
AS | Assignment |
Owner name: HEWLETT-PACKARD DEVELOPMENT COMPANY, L.P., TEXAS Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:CLARK, ALEXANDER W.;BIGGS, KENT E.;REEL/FRAME:045898/0275 Effective date: 20160303 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |