WO2013048495A1 - Mécanisme pour employer et faciliter un pavé capteur de pouce de panneau tactile dans un dispositif informatique - Google Patents

Mécanisme pour employer et faciliter un pavé capteur de pouce de panneau tactile dans un dispositif informatique Download PDF

Info

Publication number
WO2013048495A1
WO2013048495A1 PCT/US2011/054427 US2011054427W WO2013048495A1 WO 2013048495 A1 WO2013048495 A1 WO 2013048495A1 US 2011054427 W US2011054427 W US 2011054427W WO 2013048495 A1 WO2013048495 A1 WO 2013048495A1
Authority
WO
WIPO (PCT)
Prior art keywords
sensor pad
touch panel
computing device
user
pad
Prior art date
Application number
PCT/US2011/054427
Other languages
English (en)
Inventor
David L. Graumann
Original Assignee
Intel Corporation
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Intel Corporation filed Critical Intel Corporation
Priority to US13/977,395 priority Critical patent/US20130271415A1/en
Priority to CN201180075092.4A priority patent/CN103946774B/zh
Priority to EP11873407.8A priority patent/EP2761412A4/fr
Priority to PCT/US2011/054427 priority patent/WO2013048495A1/fr
Publication of WO2013048495A1 publication Critical patent/WO2013048495A1/fr

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/041Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
    • G06F3/0416Control or interface arrangements specially adapted for digitisers
    • G06F3/04166Details of scanning methods, e.g. sampling time, grouping of sub areas or time sharing with display driving
    • G06F3/041661Details of scanning methods, e.g. sampling time, grouping of sub areas or time sharing with display driving using detection at multiple resolutions, e.g. coarse and fine scanning; using detection within a limited area, e.g. object tracking window
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/033Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
    • G06F3/039Accessories therefor, e.g. mouse pads
    • G06F3/0393Accessories for touch pads or touch screens, e.g. mechanical guides added to touch screens for drawing straight lines, hard keys overlaying touch screens or touch pads
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/041Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • G06F3/04886Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures by partitioning the display area of the touch-screen or the surface of the digitising tablet into independently controllable areas, e.g. virtual keyboards or menus
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2203/00Indexing scheme relating to G06F3/00 - G06F3/048
    • G06F2203/048Indexing scheme relating to G06F3/048
    • G06F2203/04804Transparency, e.g. transparent or translucent windows
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2203/00Indexing scheme relating to G06F3/00 - G06F3/048
    • G06F2203/048Indexing scheme relating to G06F3/048
    • G06F2203/04809Textured surface identifying touch areas, e.g. overlay structure for a virtual keyboard
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/041Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
    • G06F3/044Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by capacitive means

Definitions

  • the field relates generally to computing devices and, more particularly, to employing a mechanism for employing and facilitating a touch panel thumb sensor pad at a computing device.
  • touch panel sensitivity or functionality is to add a number of extra components to the device (e.g., a microcontroller, Central Processing Unit (CPU) drivers, cables or connectors, etc.) which results in increased cost, size, complexity, etc.
  • a microcontroller Central Processing Unit (CPU) drivers, cables or connectors, etc.
  • Figure 1 illustrates a computing device employing and facilitating a touch panel thumb sensor pad mechanism according to one embodiment of the invention
  • Figure 2 illustrates a touch panel thumb pad mechanism employed at a computing device according to one embodiment of the invention
  • Figures 3A-3C illustrate a computing device employing and facilitating a thumb pad using a touch panel thumb pad mechanism according to one embodiment of the invention
  • Figures 4 illustrates a method for employing and facilitating a thumb pad using a touch panel thumb pad mechanism according to one embodiment of the invention.
  • Figure 5 illustrates a computing system according to one
  • Embodiments of the invention provide a mechanism for employing and facilitating a sensor pad transparently placed at a touch panel of a computing device.
  • a method of embodiments of the invention includes sensing a use of a sensor pad transparently placed over and within a dedicated section of a touch panel of a computing device.
  • the sensor pad and its relevant sensor pad interaction may be employed using hardware of the touch panel, the use including touching of the sensor pad by a user, while sensing may include detecting a change at one or more intersecting points of a plurality of intersecting points.
  • the method may further include facilitating an action in response to the use of the sensor pad.
  • a computing device e.g., a mobile computing device, such as a smartphone
  • a thumb pad provided at a corner (e.g., a top right-hand corner) of a touch panel.
  • the computing device may be controlled without placing a finger or thumb in front of the displayed content (that is being viewed by the user using the display screen of the computing device).
  • Embodiments of the invention provide a novel approach of using an edge or corner of the touch panel to keep the various factors (e.g., cost, size,
  • FIG. 1 illustrates a computing device employing and facilitating a touch panel thumb sensor pad mechanism according to one embodiment of the invention.
  • a computing device 100 is illustrated as having a touch panel (TP) thumb sensor pad mechanism 108 ("TP thumb pad mechanism") to employ and facilitate a TP thumb pad.
  • TP thumb pad mechanism touch panel
  • Computing device 100 includes a mobile computing device, such as a smartphone (e.g., iPhone®, BlackBerry®, etc.), a handheld computing device, a personal digital assistant (PDA), a tablet computer (e.g., iPad®, Samsung® Galaxy Tab®, etc.), a laptop computer (e.g., notebooks, netbooks, etc.), and other similar mobile computing devices, etc., having a touchscreen or touch panel having a virtual keyboard, etc., serving as a source of input.
  • Computing device 100 further includes an operating system 106 serving as an interface between any hardware or physical resources of the computer device 100 and a user.
  • Computing device 100 may further include a processor 102, memory devices 104, network devices, drivers, or the like. It is to be noted that terms like “machine”, “device”, “computing device”, “computer”, “computing system”, and the like, are used interchangeably and synonymously throughout this document.
  • the computing device 100 further includes a hardware extension 110 that is used to provide the necessary hardware and/or circuitry to employ and adopt a thumb pad; for example, in one embodiment, a portion (e.g., a triangular shaped space at the top right corner of the touch panel) of the touch panel of the computing device 100 may be reduced to provide a thumb pad using the hardware extension 110 as will be further described with reference to the subsequent figures.
  • a hardware extension 110 that is used to provide the necessary hardware and/or circuitry to employ and adopt a thumb pad; for example, in one embodiment, a portion (e.g., a triangular shaped space at the top right corner of the touch panel) of the touch panel of the computing device 100 may be reduced to provide a thumb pad using the hardware extension 110 as will be further described with reference to the subsequent figures.
  • FIG. 2 illustrates a TP thumb pad mechanism employed at a computing device according to one embodiment of the invention.
  • TP thumb pad mechanism 108 includes various components 202-208 to facilitate the use of a thumb sensor as described throughout this document.
  • the TP thumb pad mechanism 108 includes a thumb pad extension module 202 ("extension module”) to work with hardware extension and
  • a TP thumb pad of any shape is placed in a corner of the computing device's display or touch pad/panel corner that does not typically require touch interaction by the user, such as the top right-hand corner of a touch pad that is not typically touched or used by the user.
  • various hardware components may be removed or rearranged, such as various radio strength indicators, application status indicators, and the like, so that the particular corner or area can be freed up to employ and accommodate the TP thumb pad.
  • a number of not-so-important and/or not- so-frequently-used components may be placed in the corner or area that is occupied by the TP thumb pad.
  • any existing icons or hyperlinks, etc. may be moved from under the potential thumb pad to another area of the touch panel or screen so that they can be seen visually but prevented from being activated by accidental touches to the thumb pad, while other readable content and ambient status icons may continue to remain under the thumb pad.
  • the extension module 202 may be further used to reduce the amount of hardware needed to employ the thumb pad on one side (e.g., top left side) of the touch panel of the computing device.
  • the extension module 202 may provide overlaying hardware (e.g., virtual scroll and flick area) over a corner of (and/or along the right and/or left edge) the touch panel such that the overlaying hardware is in communication with the rest of the touch panel through common conductive lines and other hardware components.
  • overlaying hardware e.g., virtual scroll and flick area
  • the triangular (or another shape) region at corner (e.g., top right corner) of the touch panel acts as the mouse pad or trackball equivalent.
  • the user may move up/down/left/right and flick or brush to navigate the thumb pad user interface provided by the user interface module 210.
  • the design and interactions of the thumb pad may be matched with the thumb arc and potential movement. This will be illustrated and further discussed with reference to Figures 3A-3C.
  • a sensitivity module unit 204 is employed to sense the user touch to the thumb pad and the sensitivity resulting from the touch.
  • the sensitivity unit 204 can sense the sensitivity and the signal that contained within that sensitivity and provides the sensitivity data to a signal analysis module 206 provided by the TP thumb pad mechanism 108.
  • the signal analysis module 206 analyzes the sensitivity data to interpret the signal contained within the sensitivity obtained from the user touch of the thumb pad.
  • the sensitivity data my reveal whether the user may have brushed, flicked, or pressed the thumb pad which can then be made a basis for determining the type of action intended or anticipated by the user, such as pressed the thumb pad to obtain additional information relating to an open contact or website, etc., or flicked in a certain way to scroll up or down the current page being displayed on the display screen of the computing device, and the like.
  • the signal analysis or interpretation is then sent on to an action facilitator 208 to facilitate any action requested by the user through the thumb sensor. For example, if the user wishes to scroll down or up the screen using the thumb sensor, the action facilitator 208, working with other components 202-206, 210, ensures that the screen is scrolled up or down as requested by the user.
  • the TP thumb pad mechanism 108 works with the hardware extension 110 to, for example, provide common conductive lines between the touch panel and the thumb pad such that the thumb pad can be floated and located at any place over the touch panel (e.g., top right or left corners, bottom right or left corners, etc.) as desired or necessitated by the user while maintaining the functionality and operability of the thumb pad.
  • a user interface module 210 works with the hardware extension 110 to provide the user a user interface to conveniently work the thumb pad in accordance with its intended purpose.
  • any number and type of components may be added to and removed from the TP thumb pad mechanism 108 to facilitate the workings and operability of the thumb sensor.
  • many of the default or known components of a computing device are not shown or discussed here.
  • the display of the computing device may be continuously (readjusted or (re-)align based on the changing information being communicated by user inputs via the thumb pad.
  • Figures 3A-3C illustrate a computing device employing and facilitating a thumb pad using a TP thumb pad mechanism according to one embodiment of the invention.
  • Figure 3A illustrates the computer system 100 (e.g., a mobile computing device, such as a smartphone) having a touch panel 302 and a novel thumb pad 304 transparently overlaying in the upper right corner of the touch panel 302.
  • the thumb pad 304 may be moveable and can be moved around the touch pad 302 by the user by touching and holding the thumb pad 304 and placing it anywhere (e.g., upper left corner, lower right corner, etc.) on the touch pad 302.
  • Figure 3B illustrates an embodiment of the thumb pad 304 illustrating is various movements.
  • the user may place his or her thumb or finger to move the thumb pad 304 in various directions 316, such as up, down, left, right, circular, etc.
  • Figure 3C it illustrates the computing device 100 having a thumb pad capacitive column 306 to employ and facilitate the thumb pad 304 with the touch panel 302.
  • the illustrated capacitive column 306 may refer to a side column where the thumb pad 304 is situated, such as the right most column or the column to the right of the thumb pad 304 as illustrated here (or, for example, the left column if the thumb pad 304 occupied the left side of the device, and the like), may be electrically driven with a higher and more sensitivity charge, such as to sense the thumb that is off the panel to the right side (e.g., similar to a "hover" detection, etc., such as a non-uniform capacitive charge that is used in the thumb pad 304 to detect extreme edge interactions.
  • the capacitive column 306 provided by the thumb pad capacitive lines may be lighter, but probably not less sensitive, than the normal capacitive charge provided on the edge of the touch panel by the touch panel capacitive lines.
  • the thumb pad interaction may be implemented directly using the touch panel hardware, such as without having to employ separate or additional hardware to employ the thumb pad 304, by using non-uniformed sensitivity across the touch panel 302 and increasing the column sensitivity (e.g., top corner right column sensitivity) for managing the touch-on-edge interaction (e.g., thumb-on-edge interaction, sensor-on-edge interaction, etc.).
  • any user transaction using the thumb pad 306 be analyzed and interpreted and then reflected as an action (e.g., a new set of information, such as a new website, is displayed on the screen, the screen is scrolled up and/or down, an error message is displayed, etc.) being displayed by and on a display screen (e.g., the touch panel 302) of the computing device 100.
  • an action e.g., a new set of information, such as a new website, is displayed on the screen, the screen is scrolled up and/or down, an error message is displayed, etc.
  • Figure 4 illustrates a method for employing and facilitating a thumb pad using a TP thumb pad mechanism according to one embodiment of the invention.
  • Method 400 may be performed by processing logic that may comprise hardware (e.g., circuitry, dedicated logic, programmable logic, etc.), software (such as instructions run on a processing device), or a combination thereof.
  • processing logic may comprise hardware (e.g., circuitry, dedicated logic, programmable logic, etc.), software (such as instructions run on a processing device), or a combination thereof.
  • method 400 may be performed by the TP thumb pad mechanism of Figure 1.
  • Method 400 starts at processing block 405 with sensing of a use of a thumb pad of a computing device (e.g., a mobile computing device, such as a smartphone, a tablet computer, an e-reader, etc.).
  • the thumb pad may be placed over within a specified space (e.g., top right corner) of a touch panel of the computing device.
  • a specified space e.g., top right corner
  • the thumb pad use is sensed by detecting the relevant intersecting points formed by thumb pad capacitive lines and the touch panel capacitive lines.
  • a signal analysis of the thumb sensor use is performed to interpret the thumb pad use, such as to determine the intended purpose of the use of the thumb pad, such as whether the use was accidental and, if not, whether the thumb sensor was pressed or rolled or scrolled or flicked, etc., in a given direction so that the intended purpose of the use can be accurately determined.
  • an appropriate action is determined.
  • the appropriate action is facilitated by the action facilitator.
  • the display screen e.g., touch panel
  • the computing device displays the appropriate content (e.g., a website).
  • FIG. 5 illustrates a computing system 500 employing and facilitating a thumb pad as referenced throughout this document according to one embodiment of the invention.
  • the exemplary computing system 500 may be the same as or similar to the computing device 100 of Figure 1 and include: 1) one or more processors 501 at least one of which may include features described above; 2) a memory control hub (MCH) 502; 3) a system memory 503 (of which different types exist such as double data rate RAM (DDR RAM), extended data output RAM (EDO RAM) etc.); 4) a cache 504; 5) an input/output (I/O) control hub (ICH) 505; 6) a graphics processor 506; 7) a display/screen 507 (of which different types exist such as Cathode Ray Tube (CRT), Thin Film Transistor (TFT), Light Emitting Diode (LED), Molecular Organic LED (MOLED), Liquid Crystal Display (LCD), Digital Light Projector (DLP), etc.; and 8) one or more I/O devices 508.
  • the one or more processors 501 execute instructions in order to perform whatever software routines the computing system implements.
  • the instructions frequently involve some sort of operation performed upon data.
  • Both data and instructions are stored in system memory 503 and cache 504.
  • Cache 504 is typically designed to have shorter latency times than system memory 503.
  • cache 504 might be integrated onto the same silicon chip(s) as the processor(s) and/or constructed with faster static RAM (SRAM) cells whilst system memory 503 might be constructed with slower dynamic RAM (DRAM) cells.
  • SRAM static RAM
  • DRAM dynamic RAM
  • System memory 503 is deliberately made available to other components within the computing system.
  • the data received from various interfaces to the computing system e.g., keyboard and mouse, printer port, Local Area Network (LAN) port, modem port, etc.
  • an internal storage element of the computer system e.g., hard disk drive
  • system memory 503 prior to their being operated upon by the one or more processor(s) 501 in the implementation of a software program.
  • data that a software program determines should be sent from the computing system to an outside entity through one of the computing system interfaces, or stored into an internal storage element is often temporarily queued in system memory 503 prior to its being transmitted or stored.
  • the ICH 505 is responsible for ensuring that such data is properly passed between the system memory 503 and its appropriate corresponding computing system interface (and internal storage device if the computing system is so designed).
  • the MCH 502 is responsible for managing the various contending requests for system memory 503 accesses amongst the processor(s) 501, interfaces and internal storage elements that may proximately arise in time with respect to one another.
  • the MCH 502 and ICH 505 may not be separately employed; but rather, be provided as part of a chipset that includes the MCH 502, ICH 505, other controller hubs, and the like.
  • One or more I/O devices 508 are also implemented in a typical computing system.
  • I/O devices generally are responsible for transferring data to and/or from the computing system (e.g., a networking adapter); or, for large scale non-volatile storage within the computing system (e.g., hard disk drive).
  • ICH 505 has bi-directional point-to-point links between itself and the observed I/O devices 508.
  • Portions of various embodiments of the present invention may be provided as a computer program product, which may include a computer-readable medium having stored thereon computer program instructions, which may be used to program a computer (or other electronic devices) to perform a process according to the embodiments of the present invention.
  • the machine-readable medium may include, but is not limited to, floppy diskettes, optical disks, compact disk read-only memory (CD-ROM), and magneto-optical disks, ROM, RAM, erasable
  • EPROM programmable read-only memory
  • EEPROM electrically EPROM
  • magnet or optical cards flash memory, or other type of media/machine-readable medium suitable for storing electronic instructions.
  • the techniques shown in the figures can be implemented using code and data stored and executed on one or more electronic devices (e.g., an end station, a network element).
  • electronic devices store and communicate (internally and/or with other electronic devices over a network) code and data using computer - readable media, such as non-transitory computer -readable storage media (e.g., magnetic disks; optical disks; random access memory; read only memory; flash memory devices; phase-change memory) and transitory computer -readable transmission media (e.g., electrical, optical, acoustical or other form of propagated signals - such as carrier waves, infrared signals, digital signals).
  • non-transitory computer -readable storage media e.g., magnetic disks; optical disks; random access memory; read only memory; flash memory devices; phase-change memory
  • transitory computer -readable transmission media e.g., electrical, optical, acoustical or other form of propagated signals - such as carrier waves, infrared signals,
  • such electronic devices typically include a set of one or more processors coupled to one or more other components, such as one or more storage devices (non-transitory machine-readable storage media), user input/output devices (e.g., a keyboard, a touchscreen, and/or a display), and network connections.
  • the coupling of the set of processors and other components is typically through one or more busses and bridges (also termed as bus controllers).
  • bus controllers also termed as bus controllers
  • the storage device of a given electronic device typically stores code and/or data for execution on the set of one or more processors of that electronic device.
  • one or more parts of an embodiment of the invention may be implemented using different combinations of software, firmware, and/or hardware.

Landscapes

  • Engineering & Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • User Interface Of Digital Computer (AREA)
  • Position Input By Displaying (AREA)

Abstract

L'invention décrit un mécanisme pour employer et faciliter un pavé capteur placé de façon transparente sur un panneau tactile d'un dispositif informatique. Un procédé des modes de réalisation de l'invention comprend la détection d'une utilisation d'un pavé capteur placé de façon transparente sur et à l'intérieur d'une section dédiée d'un panneau tactile d'un dispositif informatique. Le pavé capteur et son interaction de pavé capteur concerné peuvent être employés en utilisant le matériel du panneau tactile, l'utilisation comprenant le toucher du pavé capteur par un utilisateur, alors que la détection peut comprendre la détection d'un changement sur un ou plusieurs points d'intersection d'une pluralité de points d'intersection. Le procédé peut en outre comprendre la facilitation d'une action en réponse à l'utilisation du pavé capteur.
PCT/US2011/054427 2011-09-30 2011-09-30 Mécanisme pour employer et faciliter un pavé capteur de pouce de panneau tactile dans un dispositif informatique WO2013048495A1 (fr)

Priority Applications (4)

Application Number Priority Date Filing Date Title
US13/977,395 US20130271415A1 (en) 2011-09-30 2011-09-30 Mechanism for employing and facilitating a touch panel thumb sensor pad at a computing device
CN201180075092.4A CN103946774B (zh) 2011-09-30 2011-09-30 用于采用和促进计算装置处的触摸板拇指传感器垫的机构
EP11873407.8A EP2761412A4 (fr) 2011-09-30 2011-09-30 Mécanisme pour employer et faciliter un pavé capteur de pouce de panneau tactile dans un dispositif informatique
PCT/US2011/054427 WO2013048495A1 (fr) 2011-09-30 2011-09-30 Mécanisme pour employer et faciliter un pavé capteur de pouce de panneau tactile dans un dispositif informatique

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/US2011/054427 WO2013048495A1 (fr) 2011-09-30 2011-09-30 Mécanisme pour employer et faciliter un pavé capteur de pouce de panneau tactile dans un dispositif informatique

Publications (1)

Publication Number Publication Date
WO2013048495A1 true WO2013048495A1 (fr) 2013-04-04

Family

ID=47996226

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2011/054427 WO2013048495A1 (fr) 2011-09-30 2011-09-30 Mécanisme pour employer et faciliter un pavé capteur de pouce de panneau tactile dans un dispositif informatique

Country Status (4)

Country Link
US (1) US20130271415A1 (fr)
EP (1) EP2761412A4 (fr)
CN (1) CN103946774B (fr)
WO (1) WO2013048495A1 (fr)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2015060279A1 (fr) * 2013-10-25 2015-04-30 株式会社村田製作所 Dispositif d'affichage pour terminal portable
JP2017157079A (ja) * 2016-03-03 2017-09-07 富士通株式会社 情報処理装置、表示制御方法、及び表示制御プログラム
EP4012543A1 (fr) * 2020-12-14 2022-06-15 Siemens Aktiengesellschaft Dispositif électronique et procédé pour un fonctionnement d'interface amélioré

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040135824A1 (en) 2002-10-18 2004-07-15 Silicon Graphics, Inc. Tracking menus, system and method
US20080042983A1 (en) * 2006-06-27 2008-02-21 Samsung Electronics Co., Ltd. User input device and method using fingerprint recognition sensor
US20090160792A1 (en) * 2007-12-21 2009-06-25 Kabushiki Kaisha Toshiba Portable device
US20100231548A1 (en) * 2009-03-10 2010-09-16 William Henry Mangione-Smith Touch-sensitive display device and method
KR20110081736A (ko) * 2010-01-08 2011-07-14 크루셜텍 (주) 이종의 입력 장치를 이용하여 화면을 제어하는 방법 및 그 단말장치

Family Cites Families (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6037937A (en) * 1997-12-04 2000-03-14 Nortel Networks Corporation Navigation tool for graphical user interface
US6459424B1 (en) * 1999-08-10 2002-10-01 Hewlett-Packard Company Touch-sensitive input screen having regional sensitivity and resolution properties
JP4185825B2 (ja) * 2003-07-01 2008-11-26 キヤノン株式会社 座標入力装置及びその制御方法、情報処理装置、プログラム
US7489306B2 (en) * 2004-12-22 2009-02-10 Microsoft Corporation Touch screen accuracy
EP3835920B1 (fr) * 2006-03-03 2023-10-11 Apple Inc. Dispositif électronique doté d'un affichage avec périmètre sensible au toucher pour l'interface utilisateur et la commande
KR100686165B1 (ko) * 2006-04-18 2007-02-26 엘지전자 주식회사 오에스디 기능 아이콘을 갖는 휴대용 단말기 및 이를이용한 오에스디 기능 아이콘의 디스플레이 방법
US8754855B2 (en) * 2008-06-27 2014-06-17 Microsoft Corporation Virtual touchpad
CN101719032B (zh) * 2008-10-09 2014-07-02 联想(北京)有限公司 一种多点触摸系统及其方法
CN101739159B (zh) * 2008-11-07 2012-10-10 联想(北京)有限公司 一种具有指点输入功能的计算机及其指点输入实现方法
KR101007042B1 (ko) * 2009-01-07 2011-01-12 주식회사평화발레오 자동변속기용 직결 댐퍼플라이휠과 이를 구비한 차량구동장치
US9207806B2 (en) * 2009-05-28 2015-12-08 Microsoft Technology Licensing, Llc Creating a virtual mouse input device
US8970475B2 (en) * 2009-06-19 2015-03-03 Apple Inc. Motion sensitive input control
EP2517091A4 (fr) * 2009-12-23 2013-11-06 Nokia Corp Procédé et appareil pour dispositif d'affichage
JP5717270B2 (ja) * 2009-12-28 2015-05-13 任天堂株式会社 情報処理プログラム、情報処理装置および情報処理方法

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040135824A1 (en) 2002-10-18 2004-07-15 Silicon Graphics, Inc. Tracking menus, system and method
US20080042983A1 (en) * 2006-06-27 2008-02-21 Samsung Electronics Co., Ltd. User input device and method using fingerprint recognition sensor
US20090160792A1 (en) * 2007-12-21 2009-06-25 Kabushiki Kaisha Toshiba Portable device
US20100231548A1 (en) * 2009-03-10 2010-09-16 William Henry Mangione-Smith Touch-sensitive display device and method
KR20110081736A (ko) * 2010-01-08 2011-07-14 크루셜텍 (주) 이종의 입력 장치를 이용하여 화면을 제어하는 방법 및 그 단말장치

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See also references of EP2761412A4

Also Published As

Publication number Publication date
EP2761412A1 (fr) 2014-08-06
EP2761412A4 (fr) 2015-07-01
CN103946774B (zh) 2018-11-16
US20130271415A1 (en) 2013-10-17
CN103946774A (zh) 2014-07-23

Similar Documents

Publication Publication Date Title
CN106575203B (zh) 与再现的内容的基于悬停的交互
US8525776B2 (en) Techniques for controlling operation of a device with a virtual touchscreen
KR101577106B1 (ko) 인간 기계가 가전 기기와 인터페이싱하기 위한 방법, 회로, 장치 및 시스템
US20090315841A1 (en) Touchpad Module which is Capable of Interpreting Multi-Object Gestures and Operating Method thereof
US20100090983A1 (en) Techniques for Creating A Virtual Touchscreen
US20120233545A1 (en) Detection of a held touch on a touch-sensitive display
AU2014200701B2 (en) Method and electronic device for displaying virtual keypad
US11366579B2 (en) Controlling window using touch-sensitive edge
WO2013044735A1 (fr) Navigateur web et procédé de navigation de page web
US9645729B2 (en) Precise object selection in touch sensing systems
JP6349015B2 (ja) タッチ入力装置のディスプレイ方法
US10509563B2 (en) Dynamic modification of displayed elements of obstructed region
US9501182B2 (en) Mechanism for interpreting touches to a pad cover over a sensor pad at a computing device
US20130271415A1 (en) Mechanism for employing and facilitating a touch panel thumb sensor pad at a computing device
TWI497357B (zh) 多點觸控板控制方法
US10416795B2 (en) Mechanism for employing and facilitating an edge thumb sensor at a computing device
WO2022134954A1 (fr) Procédé de mise en signet de page web, dispositif électronique, et support de stockage
CN115867883A (zh) 用于接收用户输入的方法和装置
US10042440B2 (en) Apparatus, system, and method for touch input
JP6400622B2 (ja) コンピューティングデバイスにおけるエッジ親指センサを使用し且つ促進するためのメカニズム
US20160378209A1 (en) Single stylus for use with multiple inking technologies
US20130249813A1 (en) Apparatus, system, and method for touch input
JP2014241097A (ja) ディスプレイを有するポータブルマルチファンクション装置におけるリストの横列表示部の複数表示、リスト内項目アイテムを複数選択による詳細一括表示の方法

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 11873407

Country of ref document: EP

Kind code of ref document: A1

WWE Wipo information: entry into national phase

Ref document number: 13977395

Country of ref document: US

WWE Wipo information: entry into national phase

Ref document number: 2011873407

Country of ref document: EP

NENP Non-entry into the national phase

Ref country code: DE