US20100117970A1 - Methods of Operating Electronic Devices Using Touch Sensitive Interfaces with Contact and Proximity Detection and Related Devices and Computer Program Products - Google Patents

Methods of Operating Electronic Devices Using Touch Sensitive Interfaces with Contact and Proximity Detection and Related Devices and Computer Program Products Download PDF

Info

Publication number
US20100117970A1
US20100117970A1 US12/268,502 US26850208A US2010117970A1 US 20100117970 A1 US20100117970 A1 US 20100117970A1 US 26850208 A US26850208 A US 26850208A US 2010117970 A1 US2010117970 A1 US 2010117970A1
Authority
US
United States
Prior art keywords
finger
contact
touch sensitive
user interface
sensing
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US12/268,502
Inventor
David Per BURSTROM
Anders Wilhelm Ostsjo
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Sony Mobile Communications AB
Original Assignee
Sony Ericsson Mobile Communications AB
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Sony Ericsson Mobile Communications AB filed Critical Sony Ericsson Mobile Communications AB
Priority to US12/268,502 priority Critical patent/US20100117970A1/en
Assigned to SONY ERICSSON MOBILE AB reassignment SONY ERICSSON MOBILE AB ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: BURSTROM, DAVID PER, OSTSJO, ANDERS WILHELM
Assigned to SONY ERICSSON MOBILE COMMUNICATIONS AB reassignment SONY ERICSSON MOBILE COMMUNICATIONS AB CORRECTIVE ASSIGNMENT TO CORRECT THE ASSIGNEE NAME SONY ERICSSON MOBILE AB PREVIOUSLY RECORDED ON REEL 021822 FRAME 0674. ASSIGNOR(S) HEREBY CONFIRMS THE ASSIGNEE NAME IS SONY ERICSSON MOBILE COMMUNICATIONS AB. Assignors: BURSTROM, DAVID PER, OSTSJO, ANDERS WILHELM
Priority to PCT/IB2009/051941 priority patent/WO2010055424A1/en
Publication of US20100117970A1 publication Critical patent/US20100117970A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/041Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • G06F3/04883Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2203/00Indexing scheme relating to G06F3/00 - G06F3/048
    • G06F2203/041Indexing scheme relating to G06F3/041 - G06F3/045
    • G06F2203/04108Touchless 2D- digitiser, i.e. digitiser detecting the X/Y position of the input means, finger or stylus, also when it does not touch, but is proximate to the digitiser's interaction surface without distance measurement in the Z direction

Definitions

  • This invention relates to user interfaces for electronic devices, and more particularly to touch panel interfaces for electronic devices such as wireless communication terminals and/or computer keyboards.
  • a touch sensitive user interface (also referred to as a touch sensitive panel), such as a touch sensitive screen or a touch sensitive pad, may be used to provide an interface(s) on an electronic device for a user to enter commands and/or data used in the operation of the device.
  • Touch sensitive screens may be used in mobile radiotelephones, particularly cellular radiotelephones having integrated PDA (personal digital assistant) features and other phone operation related features.
  • PDA personal digital assistant
  • the touch sensitive screens are generally designed to operate and respond to a finger touch, a stylus touch, and/or finger/stylus movement on the touch screen surface.
  • a touch sensitive screen may be used in addition to, in combination with, or in place of physical keys traditionally used in a cellular phone to carry out the phone functions and features.
  • Touch sensitive pads may be provided below the spacebar of a keyboard of a computer (such as a laptop computer), and may be used to accept pointer and click inputs. In other words, a touch sensitive pad may be used to accept user input equivalent to input accepted by a computer mouse.
  • Touching a specific point on a touch sensitive screen may activate a virtual button, feature, or function found or shown at that location on the touch screen display.
  • Typical phone features which may be operated by touching the touch screen display include entering a telephone number, for example, by touching virtual keys of a virtual keyboard shown on the display, making a call or ending a call, bringing up, adding to or editing and navigating through an address book, accepting inputs for internet browsing, and/or other phone functions such as text messaging, wireless connection to the global computer network, and/or other phone functions.
  • a method of operating an electronic device using a touch sensitive user interface may include detecting contact between a first finger and the touch sensitive user interface, and detecting non-contact proximity of a second finger to the touch sensitive user interface. Responsive to detecting contact between the first finger and the touch sensitive user interface and responsive to detecting non-contact proximity of the second finger to the touch sensitive user interface, one of a plurality of operations may be selected. Responsive to selecting one of the plurality of operations, the selected operation may be performed.
  • the touch sensitive user interface may include a touch sensitive screen and/or a touch sensitive pad.
  • Detecting contact may include detecting contact between the first finger and the touch sensitive user interface using infrared (IR) contact sensing, acoustic wave contact sensing, capacitive contact sensing, and/or resistive contact sensing.
  • Detecting non-contact proximity may include detecting non-contact proximity of the second finger to the touch sensitive user interface using optical sensing.
  • detecting contact may include detecting contact using a first sensing technology, and wherein detecting non-contact proximity comprises detecting non-contact proximity using a second sensing technology different than the first sensing technology.
  • the first sensing technology may be selected from infrared sensing, acoustic sensing, capacitive sensing, and/or resistive sensing
  • the second sensing technology may be selected from acoustic sensing and/or optical sensing.
  • the first operation may include initiating a link to a website identified by detecting contact between the first finger and the touch sensitive user interface
  • the second operation may include an editing operation and/or a bookmarking operation.
  • selecting one of the plurality of operations may include selecting a first of the plurality of operations when the first finger is between the second and third fingers, and selecting a second of the plurality of operations when the second and third fingers are on a same side of the first finger.
  • an electronic device may include a touch sensitive user interface with a contact detector and a non-contact proximity detector.
  • the contact detector may be configured to detect contact between a first finger and the touch sensitive user interface
  • the non-contact proximity detector may be configured to detect a proximity of a second finger to the touch sensitive user interface.
  • a controller may be coupled to the touch sensitive user interface.
  • the controller may be configured to select one of a plurality of operations responsive to detecting contact between the first finger and the touch sensitive user interface and responsive to detecting non-contact proximity of the second finger to the touch sensitive user interface.
  • the controller may be configured to perform the selected operation responsive to selecting one of the plurality of operations.
  • the touch sensitive user interface may include a touch sensitive screen and/or a touch sensitive pad.
  • the contact detector may be configured to detect contact between the first finger and the touch sensitive user interface using infrared (IR) contact sensing, acoustic wave contact sensing, capacitive contact sensing, and/or resistive contact sensing.
  • the non-contact proximity detector may be configured to detect non-contact proximity of the second finger to the touch sensitive user interface using optical sensing.
  • the contact detector may be configured to detect contact using a first sensing technology
  • the non-contact proximity detector may be configured to detect non-contact proximity using a second sensing technology different than the first sensing technology.
  • the first sensing technology may be selected from infrared sensing, acoustic sensing, capacitive sensing, and/or resistive sensing
  • the second sensing technology is selected from acoustic sensing and/or optical sensing.
  • the non-contact proximity detector may be configured to detect non-contact proximity of the second finger to the touch sensitive user interface without contact between the second finger and the touch sensitive user interface.
  • the non-contact proximity detector may be configured to detect non-contact proximity of the second finger while detecting contact between the first finger and the touch sensitive user interface.
  • the controller may be configured to select one of the plurality of operations by determining an orientation of the second finger relative to the first finger, selecting a first of the plurality of operations when the second finger is in a first orientation relative to the first finger, and selecting a second of the plurality of operations when the second finger is in a second orientation relative to the first finger different than the first orientation.
  • the first operation may include initiating a link to a website identified by detecting contact between the first finger and the touch sensitive user interface
  • the second operation may include an editing operation and/or a bookmarking operation.
  • the non-contact proximity detector may be further configured to detect non-contact proximity of a third finger to the touch sensitive user interface, and the controller may be configured to select a first of the plurality of operations when the first finger is between the second and third fingers, and to select a second of the plurality of operations when the second and third fingers are on a same side of the first finger.
  • a computer program product may be provided to operate an electronic device using a touch sensitive user interface
  • the computer program product may include a computer readable storage medium having computer readable program code embodied therein.
  • the computer readable program code may include computer readable program code configured to detect contact between a first finger and the touch sensitive user interface, and computer readable program code configured to detect non-contact proximity of a second finger to the touch sensitive user interface.
  • the computer readable program code may further include computer readable program code configured to select one of a plurality of operations responsive to detecting contact between the first finger and the touch sensitive user interface and responsive to detecting non-contact proximity of the second finger to the touch sensitive user interface.
  • the computer readable program code may include computer readable program code configured to perform the selected operation responsive to selecting one of the plurality of operations.
  • FIG. 1 is a block diagram of an electronic device including a touch sensitive user interface according to some embodiments of the present invention.
  • FIG. 2 is a block diagram of an electronic device including a touch sensitive user interface according to some other embodiments of the present invention.
  • FIGS. 3A and 3B are schematic illustrations of a touch sensitive user interface according to some embodiments of the present invention.
  • FIG. 4 is a flow chart illustrating operations of an electronic device including a touch sensitive interface according to some embodiments of the present invention.
  • the present invention may be embodied as methods, electronic devices, and/or computer program products. Accordingly, the present invention may be embodied in hardware and/or in software (including firmware, resident software, micro-code, etc.). Furthermore, the present invention may take the form of a computer program product on a computer-usable or computer-readable storage medium having computer-usable or computer-readable program code embodied in the medium for use by or in connection with an instruction execution system.
  • a computer-usable or computer-readable medium may be any medium that can contain, store, communicate, propagate, or transport the program for use by or in connection with the instruction execution system, apparatus, or device.
  • the computer-usable or computer-readable medium may be, for example but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, device, or propagation medium. More specific examples (a nonexhaustive list) of the computer-readable medium would include the following: an electrical connection having one or more wires, a portable computer diskette, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), an optical fiber, and a compact disc read-only memory (CD-ROM).
  • RAM random access memory
  • ROM read-only memory
  • EPROM or Flash memory erasable programmable read-only memory
  • CD-ROM compact disc read-only memory
  • the computer-usable or computer-readable medium could even be paper or another suitable medium upon which the program is printed, as the program can be electronically captured, via, for instance, optical scanning of the paper or other medium, then compiled, interpreted, or otherwise processed in a suitable manner, if necessary, and then stored in a computer memory.
  • FIG. 1 is a block diagram of an electronic device 100 (such as a cellular radiotelephone) including a touch sensitive user interface 101 according to some embodiments of the present invention.
  • the electronic device 100 may be a wireless communications device (such as a cellular radiotelephone), a PDA, an audio/picture/video player/recorder, a global positioning (GPS) unit, a gaming device, or any other electronic device including a touch sensitive screen display.
  • Electronic device 100 may also include a controller 111 coupled to touch sensitive user interface 101 , a radio transceiver 115 coupled to controller 111 , and a memory 117 coupled to controller 111 .
  • a keyboard/keypad 119 may be coupled to controller 111 .
  • electronic device 100 may be a cellular radiotelephone configured to provide PDA functionality, data network connectivity (such as Internet browsing), and/or other data functionality.
  • the controller 111 may be configured to communicate through transceiver 115 and antenna 125 over a wireless air interface with one or more RF transceiver base stations and/or other wireless communication devices using one or more wireless communication protocols such as, for example, Global Standard for Mobile (GSM) communication, General Packet Radio Service (GPRS), enhanced data rates for GSM evolution (EDGE), Integrated Digital Enhancement Network (iDEN), code division multiple access (CDMA), wideband-CDMA, CDMA2000, Universal Mobile Telecommunications System (UMTS), WiMAX, and/or HIPERMAN, wireless local area network (e.g., 802.11), and/or Bluetooth. Controller 111 may be configured to carry out wireless communications functionality, such as conventional cellular phone functionality including, but not limited to, voice/video telephone calls and/or data messaging such as text/picture/video messaging.
  • GSM Global Standard for Mobile
  • GPRS General Packet Radio Service
  • EDGE enhanced data rates for GSM evolution
  • iDEN Integrated Digital Enhancement Network
  • CDMA code division multiple
  • the controller 111 may be further configured to provide various user applications which can include a music/picture/video recorder/player application, an e-mail/messaging application, a calendar/appointment application, and/or other user applications.
  • the audio/picture/video recorder/player application can be configured to record and playback audio, digital pictures, and/or video that are captured by a sensor (e.g., microphone 123 and/or a camera) within electronic device 100 , downloaded into electronic device 100 via radio transceiver 115 and controller 111 , downloaded into electronic device 100 via a wired connection (e.g., via USB), and/or installed within electronic device 100 such as through a removable memory media.
  • a sensor e.g., microphone 123 and/or a camera
  • An e-mail/messaging application may be configured to allow a user to generate e-mail/messages (e.g., short messaging services messages and/or instant messages) for transmission via controller 111 and transceiver 115 .
  • a calendar/appointment application may provide a calendar and task schedule that can be viewed and edited by a user to schedule appointments and other tasks.
  • touch sensitive user interface 101 may be a touch sensitive screen including a display 103 , a contact detector 105 , and a proximity detector 107 .
  • contact detector 105 may be configured to detect contact between a first finger and display 103
  • proximity detector 107 may be configured to detect proximity of a second finger to display 103 without contact between the second finger and touch sensitive user interface 101 .
  • contact detector 105 may be configured to detect contact between first finger and touch sensitive user interface 101 using infrared (IR) contact sensing, acoustic wave contact sensing, capacitive contact sensing, and/or resistive contact sensing.
  • IR infrared
  • Proximity detector 107 may be configured to detect proximity of the second finger to touch sensitive user interface 101 using acoustic sensing and/or optical sensing.
  • Optical sensing may be provided, for example, using a High Ambient Light Independent Optical System (HALIOS®) as discussed in the reference by Rottmann et al. in “Electronic Concept Fulfils Optical Sensor Dream” published by ELMOS Semiconductor AG at http://www.mechaless.com/images/pdf/Elektronikheft_ENG.pdf.
  • HALIOS® High Ambient Light Independent Optical System
  • HALIOS® Optical sensing is also discussed in the reference entitled “HALIOS®—Optics For Human Machine Interfaces,” ELMOS Semiconductor AC, Version 1.0, pages 1-15, Mar. 3, 2008, the disclosure of which is also incorporated herein in its entirety by reference.
  • contact detector 105 may be configured to detect contact using a first sensing technology
  • proximity detector 107 may be configured to detect non-contact proximity using a second technology different than the first technology. More particularly, proximity detector 107 may be configured to detect non-contact proximity while the contact detector 105 is detecting contact.
  • contact detector 105 may be configured to detect contact using a first sensing technology such as infrared sensing, acoustic wave sensing, capacitive sensing, and/or resistive sensing
  • proximity detector 107 may be configured to detect non-contact proximity using a second sensing technology such as acoustic sensing and/or optical sensing.
  • a same technology such as an optical sensing technology
  • controller 111 may be configured to select one of a plurality of different operations responsive to detecting contact between a first finger and touch sensitive user interface 101 and responsive to detecting non-contact proximity of a second finger to touch sensitive user interface 101 , and then perform the selected operation. As discussed in greater detail below with respect to FIGS. 3A and 3B , by detecting contact of a first finger and non-contact proximity of a second finger relative to display 103 of touch sensitive user interface 101 at the same time, controller 111 may determine which finger (e.g., pointer finger, middle finger, etc.) is in contact with display 103 . Accordingly, different operations may be performed depending on the finger making contact with display 103 .
  • finger e.g., pointer finger, middle finger, etc.
  • a web address may be shown on display 103 , and contact with the portion of display 103 where the web address is shown may select the web address.
  • one of a plurality of operations relating to the web address may be performed depending on an orientation of a proximate finger relative to the contacting finger.
  • a proximate finger relative to the contacting finger.
  • a communications link may be established with a website identified by the selected web address, and if the contacting finger is the middle finger, another operation (such as a bookmarking operation and/or an editing operation) may be performed using the selected web address.
  • a contact alias may be shown on display 103 . If pointer finger contact is made with the contact alias, a communication (e.g., a telephone call, an e-mail, a text message, etc.) with the contact may be initiated, while if middle finger contact is made with the contact alias, a property(ies) (e.g., telephone number, e-mail address, text message address, etc.) may be shown, and/or an editing operation may be initiated. While differentiation between two fingers is discussed by way of example, differentiation between three or more fingers may be provided as discussed in greater detail below.
  • a communication e.g., a telephone call, an e-mail, a text message, etc.
  • FIG. 2 is a block diagram of an electronic device 200 including a touch sensitive user interface 201 according to some embodiments of the present invention.
  • the electronic device 200 may be a computing device (such as a laptop computer) including a touch sensitive pad.
  • Device 200 may also include a controller 211 coupled to touch sensitive user interface 201 , a network interface 215 coupled to controller 211 , and a memory 217 coupled to controller 211 .
  • a display 227 , a keyboard/keypad 219 , a speaker 221 , and/or a microphone 223 may be coupled to controller 211 .
  • device 200 may be a laptop computer configured to provide data network connectivity (such as Internet browsing), and/or other data functionality.
  • touch sensitive pad 203 may be provided below a spacebar of keyboard 219 to accept user input of pointer and/or click commands similar to pointer and click commands normally accepted though a computer mouse.
  • the controller 211 may be configured to communicate through network interface 215 with one or more other remote devices over a local area network, a wide area network, and/or the Internet. Controller 211 may be further configured to provide various user applications which can include an audio/picture/video recorder/player application, an e-mail/messaging application, a calendar/appointment application, and/or other user applications.
  • various user applications can include an audio/picture/video recorder/player application, an e-mail/messaging application, a calendar/appointment application, and/or other user applications.
  • the audio/picture/video recorder/player application can be configured to record and playback audio, digital pictures, and/or video that are captured by a sensor (e.g., microphone 223 and/or a camera) within device 200 , downloaded into device 200 via network interface 215 and controller 211 , downloaded into device 200 via a wired connection (e.g., via USB), and/or installed within device 200 such as through a removable memory media.
  • An e-mail/messaging application may be configured to allow a user to generate e-mail/messages for transmission via controller 211 and network interface 215 .
  • a calendar/appointment application may provide a calendar and task schedule that can be viewed and edited by a user to schedule appointments and other tasks.
  • touch sensitive user interface 201 may include a touch sensitive pad 203 , a contact detector 205 , and a non-contact proximity detector 207 .
  • contact detector 205 may be configured to detect contact between a first finger and pad 203
  • non-contact proximity detector 207 may be configured to detect non-contact proximity of a second finger to pad 203 without contact between the second finger and the touch sensitive user interface.
  • contact detector 205 may be configured to detect contact between the first finger and pad 203 using infrared (IR) contact sensing, acoustic wave contact sensing, capacitive contact sensing, and/or resistive contact sensing.
  • IR infrared
  • Non-contact proximity detector 207 may be configured to detect non-contact proximity of the second finger to pad 203 using acoustic sensing and/or optical sensing.
  • Optical sensing may be provided, for example, using a High Ambient Light Independent Optical System (HALIOS) as discussed in the reference by Rottmann et al. in “Electronic Concept Fulfils Optical Sensor Dream” published by ELMOS Semiconductor AG at http://www.mechaless.com/images/pdf/Elektronikheft_ENG.pdf.
  • HALIOS High Ambient Light Independent Optical System
  • HALIOS® Optical sensing is also discussed in the reference entitled “HALIOS®—Optics For Human Machine Interfaces,” ELMOS Semiconductor AG, Version 1.0, pages 1-15, Mar. 3, 2008, the disclosure of which is also incorporated herein in its entirety by reference.
  • contact detector 205 may be configured to detect contact using a first sensing technology
  • non-contact proximity detector 207 may be configured to detect non-contact proximity using a second technology different than the first technology. More particularly, non-contact proximity detector 207 may be configured to detect non-contact proximity while the contact detector 205 is detecting contact.
  • contact detector 205 may be configured to detect contact using a first sensing technology such as infrared sensing, acoustic wave sensing, capacitive sensing, and/or resistive sensing
  • non-contact proximity detector 207 may be configured to detect non-contact proximity using a second sensing technology such as acoustic sensing and/or optical sensing.
  • a same technology such as an optical sensing technology
  • controller 211 may be configured to select one of a plurality of different operations responsive to detecting contact between a first finger and touch sensitive user interface 201 and responsive to detecting non-contact proximity of a second finger to touch sensitive user interface 201 , and then perform the selected operation.
  • controller 211 may determine which finger (e.g., pointer finger, middle finger, etc.) is in contact with pad 203 . Accordingly, different operations may be performed depending on the finger making contact with pad 203 .
  • touch sensitive user interface 201 may be configured to differentiate between three different fingers (e.g., pointer, middle, and ring fingers) to provide three different command types.
  • fingers e.g., pointer, middle, and ring fingers
  • there will be no proximate finger to the left of the contacting finger if the pointer finger is the contacting finger
  • there will be one non-contacting proximate finger i.e., the pointer finger
  • there will be two non-contacting proximate fingers i.e., the pointer and middle fingers
  • movement of a pointer finger in contact with pad 203 may be interpreted as a pointer command to move a pointer on display 227 ; contact of a middle finger with pad 203 may be interpreted as a left mouse click operation; and contact of a ring finger with pad 203 may be interpreted as a right mouse click operation. While differentiation between three fingers is discussed by way of example, differentiation between two or four fingers may be provided as discussed in greater detail below.
  • FIGS. 3A and 3B are schematic illustrations showing operations of a touch sensitive user interface 311 according to some embodiments of the present invention.
  • the operations shown in FIGS. 3A and 3B may be applied to touch sensitive user interface 101 (implemented with touch sensitive screen display 103 ) of FIG. 1 or to touch sensitive user interface 201 (implemented with touch sensitive pad 203 ) of FIG. 2 .
  • the touch sensitive user interface 311 may be a touch sensitive screen display or a touch sensitive pad.
  • the touch sensitive user interface 311 may be configured to differentiate between contact from a pointer finger 331 and a middle finger 332 for right hand use.
  • middle finger 332 may contact interface 331 while pointer finger 331 , ring finger 333 , and pinky finger 334 are proximate to interface 331 without contacting interface 331 .
  • pointer finger 331 i.e., pointer finger 331
  • middle finger 332 By detecting proximity of one non-contacting finger (i.e., pointer finger 331 ) to the left of the contacting finger (i.e., middle finger 332 ), a determination can be made that the contacting finger is middle finger 332 , and an appropriate operation corresponding to a middle finger contact may be initiated.
  • pointer finger 331 may contact interface 331 while middle finger 332 , ring finger 333 , and pinky finger 334 are proximate to interface 331 without contacting interface 331 .
  • middle finger 332 , ring finger 333 , and pinky finger 334 are proximate to interface 331 without contacting interface 331 .
  • pointer finger 331 By detecting a lack of proximity of any fingers to the left of the contacting finger (i.e., pointer finger 331 ), a determination can be made that the contacting finger is pointer finger 331 , and an appropriate operation corresponding to pointer finger contact may be initiated (different than the operation corresponding to middle finger contact).
  • ring finger 333 may be determined by detecting proximity of two non-contacting fingers (i.e., pointer and middle fingers 331 and 332 ) to the left of the contacting finger (i.e., ring finger 333 ), and/or by detecting proximity only one non-contacting finger (i.e., pinky finger 334 ) to the right of contacting finger (i.e., ring finger 333 ).
  • Contact by pinky finger 334 may be determined by detecting proximity of three non-contacting fingers (i.e., pointer finger 331 , middle finger 332 , and ring finger 333 ) to the right of contacting finger (i.e., pinky finger 334 ), and/or by detecting proximity of no fingers to the right of the contacting finger (i.e., pinky finger 334 ).
  • Alternate detection criteria may be used to provide redundancy in the determination and/or to accommodate a situation where the contacting finger is near an edge of interface 311 so that proximate non-contacting fingers on one side of the contacting finger are not within range of detection.
  • the examples discussed above are discussed for right hand use. Left hand use, however, may be provided by using a reversed consideration of fingers proximate to the contacting finger.
  • an electronic device 100 / 200 incorporating touch sensitive user interface 311 / 101 / 201 may provide user selection of right or left hand use.
  • a set-up routine of the electronic device 100 / 200 may prompt the user to enter a right hand or left hand preference, and the preference may be stored in memory 117 / 217 of the electronic device 100 / 200 .
  • the controller 111 / 211 of the electronic device 100 / 200 may use the stored preference to determine how to interpret finger contact with interface 311 / 101 / 201 .
  • operations may be restricted to use of two fingers (e.g., pointer and middle fingers), and determination of the contacting finger may be performed automatically without requiring prior selection/assumption regarding right or left handed use.
  • touch sensitive user interface 311 / 101 / 201 may be configured to differentiate between pointer and middle fingers to provide two different command types responsive to contact with touch sensitive user interface 311 / 101 / 201 .
  • the pointer finger is the contacting finger, there will be no non-contacting proximate fingers on one side of the contacting finger regardless of light or left handed use.
  • determination of pointer or middle finger contact may be performed regardless of right or left handedness and/or regardless of user orientation relative to touch sensitive user interface 311 / 101 / 201 .
  • determination of pointer or middle finger contact may be performed if the user is oriented normally with respect to touch sensitive user interface 311 / 101 / 201 (e.g., with the wrist/arm below the touch sensitive user interface), if the user is oriented sideways with respect to touch sensitive user interface 311 / 101 / 201 (e.g., with the wrist/arm to the side of touch sensitive user interface), or if the user is oriented upside down with respect to touch sensitive user interface 311 / 101 / 201 (e.g., with the wrist/arm above the touch sensitive user interface).
  • FIG. 4 is a flow chart illustrating operations of an electronic device including a touch sensitive interface according to some embodiments of the present invention. Operations of FIG. 4 may be performed, for example, by an electronic device including a touch sensitive screen display as discussed above with respect to FIG. 1 , or by an electronic device including a touch sensitive pad as discussed above with respect to FIG. 2 .
  • contact between a first finger and the touch sensitive user interface may be detected, for example, using infrared (IR) contact sensing, acoustic wave contact sensing, capacitive contact sensing, and/or resistive contact sensing.
  • non-contact proximity of a second finger to the touch sensitive user interface may be detected, for example, using optical sensing. More particularly, non-contact proximity of the second finger may be detected at block 403 while detecting contact of the first finger at block 401 , and/or contact of the first finger and non-contact proximity of the second finger may be detected at the same time.
  • one of a plurality of operations may be selected at block 405 .
  • the selection may be based on a determination of relative orientations of the first and second fingers as discussed above with respect to FIGS. 3A and 3B . More particularly, the selection may be based on a determination of which finger (i.e., pointer, middle, ring, or pinky) is the contacting finger, and different operations may be assigned to at least two of the fingers. Responsive to selecting one of the plurality of operations, the selected operation may be performed at block 407 .
  • Computer program code for carrying out operations of devices and/or systems discussed above may be written in a high-level programming language, such as Java, C, and/or C++, for development convenience.
  • computer program code for carrying out operations of embodiments of the present invention may also be written in other programming languages, such as, but not limited to, interpreted languages.
  • Some modules or routines may be written in assembly language or even micro-code to enhance performance and/or memory usage. It will be further appreciated that the functionality of any or all of the program modules may also be implemented using discrete hardware components, one or more application specific integrated circuits (ASICs), or a programmed digital signal processor or microcontroller.
  • ASICs application specific integrated circuits
  • These computer program instructions may be provided to a processor of a general purpose computer, a special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions specified in the flowchart and/or block diagram block or blocks.
  • These computer program instructions may also be stored in a computer usable or computer-readable memory that may direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer usable or computer-readable memory produce an article of manufacture including instructions that implement the function specified in the flowchart and/or block diagram block or blocks.
  • the computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions that execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart and/or block diagram block or blocks.

Abstract

A method of operating an electronic device using a touch sensitive user interface may include detecting contact between a first finger and the touch sensitive user interface, and detecting non-contact proximity of a second finger to the touch sensitive user interface. Responsive to detecting contact between the first finger and the touch sensitive user interface and responsive to detecting non-contact proximity of the second finger to the touch sensitive user interface, one of a plurality of operations may be selected. Responsive to selecting one of the plurality of operations, the selected operation may be performed. Related devices and computer program products are also discussed.

Description

    FIELD OF THE INVENTION
  • This invention relates to user interfaces for electronic devices, and more particularly to touch panel interfaces for electronic devices such as wireless communication terminals and/or computer keyboards.
  • BACKGROUND OF THE INVENTION
  • A touch sensitive user interface (also referred to as a touch sensitive panel), such as a touch sensitive screen or a touch sensitive pad, may be used to provide an interface(s) on an electronic device for a user to enter commands and/or data used in the operation of the device. Touch sensitive screens, for example, may be used in mobile radiotelephones, particularly cellular radiotelephones having integrated PDA (personal digital assistant) features and other phone operation related features. The touch sensitive screens are generally designed to operate and respond to a finger touch, a stylus touch, and/or finger/stylus movement on the touch screen surface. A touch sensitive screen may be used in addition to, in combination with, or in place of physical keys traditionally used in a cellular phone to carry out the phone functions and features. Touch sensitive pads may be provided below the spacebar of a keyboard of a computer (such as a laptop computer), and may be used to accept pointer and click inputs. In other words, a touch sensitive pad may be used to accept user input equivalent to input accepted by a computer mouse.
  • Touching a specific point on a touch sensitive screen may activate a virtual button, feature, or function found or shown at that location on the touch screen display. Typical phone features which may be operated by touching the touch screen display include entering a telephone number, for example, by touching virtual keys of a virtual keyboard shown on the display, making a call or ending a call, bringing up, adding to or editing and navigating through an address book, accepting inputs for internet browsing, and/or other phone functions such as text messaging, wireless connection to the global computer network, and/or other phone functions.
  • Commercial pressure to provide increased functionality is continuing to drive demand for even more versatile user interfaces.
  • SUMMARY OF THE INVENTION
  • According to some embodiments of the present invention, a method of operating an electronic device using a touch sensitive user interface may include detecting contact between a first finger and the touch sensitive user interface, and detecting non-contact proximity of a second finger to the touch sensitive user interface. Responsive to detecting contact between the first finger and the touch sensitive user interface and responsive to detecting non-contact proximity of the second finger to the touch sensitive user interface, one of a plurality of operations may be selected. Responsive to selecting one of the plurality of operations, the selected operation may be performed. For example, the touch sensitive user interface may include a touch sensitive screen and/or a touch sensitive pad.
  • Detecting contact may include detecting contact between the first finger and the touch sensitive user interface using infrared (IR) contact sensing, acoustic wave contact sensing, capacitive contact sensing, and/or resistive contact sensing. Detecting non-contact proximity may include detecting non-contact proximity of the second finger to the touch sensitive user interface using optical sensing. For example, detecting contact may include detecting contact using a first sensing technology, and wherein detecting non-contact proximity comprises detecting non-contact proximity using a second sensing technology different than the first sensing technology. More particularly, the first sensing technology may be selected from infrared sensing, acoustic sensing, capacitive sensing, and/or resistive sensing, and the second sensing technology may be selected from acoustic sensing and/or optical sensing.
  • Detecting non-contact proximity may include detecting non-contact proximity of the second finger to the touch sensitive user interface without contact between the second finger and the touch sensitive user interface. Detecting non-contact proximity of the second finger may include detecting non-contact proximity of the second finger while detecting contact between the first finger and the touch sensitive user interface. Moreover, selecting one of a plurality of operations may include determining an orientation of the second finger relative to the first finger, selecting a first of the plurality of operations when the second finger is in a first orientation relative to the first finger, and selecting a second of the plurality of operations when the second finger is in a second orientation relative to the first finger different than the first orientation. The first operation may include initiating a link to a website identified by detecting contact between the first finger and the touch sensitive user interface, and the second operation may include an editing operation and/or a bookmarking operation.
  • In addition, non-contact proximity of a third finger to the touch sensitive user interface may be detected. Accordingly, selecting one of the plurality of operations may include selecting a first of the plurality of operations when the first finger is between the second and third fingers, and selecting a second of the plurality of operations when the second and third fingers are on a same side of the first finger.
  • According to other embodiments of the present invention, an electronic device may include a touch sensitive user interface with a contact detector and a non-contact proximity detector. The contact detector may be configured to detect contact between a first finger and the touch sensitive user interface, and the non-contact proximity detector may be configured to detect a proximity of a second finger to the touch sensitive user interface. In addition, a controller may be coupled to the touch sensitive user interface. The controller may be configured to select one of a plurality of operations responsive to detecting contact between the first finger and the touch sensitive user interface and responsive to detecting non-contact proximity of the second finger to the touch sensitive user interface. In addition, the controller may be configured to perform the selected operation responsive to selecting one of the plurality of operations. For example, the touch sensitive user interface may include a touch sensitive screen and/or a touch sensitive pad.
  • The contact detector may be configured to detect contact between the first finger and the touch sensitive user interface using infrared (IR) contact sensing, acoustic wave contact sensing, capacitive contact sensing, and/or resistive contact sensing. The non-contact proximity detector may be configured to detect non-contact proximity of the second finger to the touch sensitive user interface using optical sensing. For example, the contact detector may be configured to detect contact using a first sensing technology, and the non-contact proximity detector may be configured to detect non-contact proximity using a second sensing technology different than the first sensing technology. More particularly, the first sensing technology may be selected from infrared sensing, acoustic sensing, capacitive sensing, and/or resistive sensing, and the second sensing technology is selected from acoustic sensing and/or optical sensing.
  • The non-contact proximity detector may be configured to detect non-contact proximity of the second finger to the touch sensitive user interface without contact between the second finger and the touch sensitive user interface. The non-contact proximity detector may be configured to detect non-contact proximity of the second finger while detecting contact between the first finger and the touch sensitive user interface. The controller may be configured to select one of the plurality of operations by determining an orientation of the second finger relative to the first finger, selecting a first of the plurality of operations when the second finger is in a first orientation relative to the first finger, and selecting a second of the plurality of operations when the second finger is in a second orientation relative to the first finger different than the first orientation. For example, the first operation may include initiating a link to a website identified by detecting contact between the first finger and the touch sensitive user interface, and the second operation may include an editing operation and/or a bookmarking operation.
  • The non-contact proximity detector may be further configured to detect non-contact proximity of a third finger to the touch sensitive user interface, and the controller may be configured to select a first of the plurality of operations when the first finger is between the second and third fingers, and to select a second of the plurality of operations when the second and third fingers are on a same side of the first finger.
  • According to still other embodiments of the present invention, a computer program product may be provided to operate an electronic device using a touch sensitive user interface, and the computer program product may include a computer readable storage medium having computer readable program code embodied therein. The computer readable program code may include computer readable program code configured to detect contact between a first finger and the touch sensitive user interface, and computer readable program code configured to detect non-contact proximity of a second finger to the touch sensitive user interface. The computer readable program code may further include computer readable program code configured to select one of a plurality of operations responsive to detecting contact between the first finger and the touch sensitive user interface and responsive to detecting non-contact proximity of the second finger to the touch sensitive user interface. In addition, the computer readable program code may include computer readable program code configured to perform the selected operation responsive to selecting one of the plurality of operations.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 is a block diagram of an electronic device including a touch sensitive user interface according to some embodiments of the present invention.
  • FIG. 2 is a block diagram of an electronic device including a touch sensitive user interface according to some other embodiments of the present invention.
  • FIGS. 3A and 3B are schematic illustrations of a touch sensitive user interface according to some embodiments of the present invention.
  • FIG. 4 is a flow chart illustrating operations of an electronic device including a touch sensitive interface according to some embodiments of the present invention.
  • DETAILED DESCRIPTION
  • While the invention is susceptible to various modifications and alternative forms, specific embodiments thereof are shown by way of example in the drawings and will herein be described in detail. It should be understood, however, that there is no intent to limit the invention to the particular forms disclosed, but on the contrary, the invention is to cover all modifications, equivalents, and alternatives falling within the spirit and scope of the invention as defined by the claims. Like reference numbers signify like elements throughout the description of the figures.
  • As used herein, the singular forms “a,” “an,” and “the” are intended to include the plural forms as well, unless expressly stated otherwise. It should be further understood that the terms “comprises” and/or “comprising” when used in this specification is taken to specify the presence of stated features, integers, steps, operations, elements, and/or components, but does not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof. It will be understood that when an element is referred to as being “connected” or “coupled” to another element, it can be directly connected or coupled to the other element or intervening elements may be present. Furthermore, “connected” or “coupled” as used herein may include wirelessly connected or coupled. As used herein, the term “and/or” includes any and all combinations of one or more of the associated listed items.
  • Unless otherwise defined, all terms (including technical and scientific terms) used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this invention belongs. It will be further understood that terms, such as those defined in commonly used dictionaries, should be interpreted as having a meaning that is consistent with their meaning in the context of the relevant art and will not be interpreted in an idealized or overly formal sense unless expressly so defined herein.
  • The present invention may be embodied as methods, electronic devices, and/or computer program products. Accordingly, the present invention may be embodied in hardware and/or in software (including firmware, resident software, micro-code, etc.). Furthermore, the present invention may take the form of a computer program product on a computer-usable or computer-readable storage medium having computer-usable or computer-readable program code embodied in the medium for use by or in connection with an instruction execution system. In the context of this document, a computer-usable or computer-readable medium may be any medium that can contain, store, communicate, propagate, or transport the program for use by or in connection with the instruction execution system, apparatus, or device.
  • The computer-usable or computer-readable medium may be, for example but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, device, or propagation medium. More specific examples (a nonexhaustive list) of the computer-readable medium would include the following: an electrical connection having one or more wires, a portable computer diskette, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), an optical fiber, and a compact disc read-only memory (CD-ROM). Note that the computer-usable or computer-readable medium could even be paper or another suitable medium upon which the program is printed, as the program can be electronically captured, via, for instance, optical scanning of the paper or other medium, then compiled, interpreted, or otherwise processed in a suitable manner, if necessary, and then stored in a computer memory.
  • Embodiments are described below with reference to block diagrams and operational flow charts. It is to be understood that the functions/acts noted in the blocks may occur out of the order noted in the operational illustrations. For example, two blocks shown in succession may in fact be executed substantially concurrently or the blocks may sometimes be executed in the reverse order, depending upon the functionality/acts involved. Although some of the diagrams include arrows on communication paths to show a primary direction of communication, it is to be understood that communication may occur in the opposite direction to the depicted arrows.
  • Although various embodiments of the present invention are described in the context of wireless communication terminals for purposes of illustration and explanation only, the present invention is not limited thereto. It is to be understood that the present invention can be more broadly used in any sort of electronic device to identify and respond to input on a touch sensitive user input.
  • It will be understood that, although the terms first, second, third etc. may be used herein to describe various elements, components, and/or sections, these elements, components, and/or sections should not be limited by these terms. These terms are only used to distinguish one element, component, or section from another element, component, or section. Thus, a first element, component, or section discussed below could be termed a second element, component, or section without departing from the teachings of the present invention.
  • FIG. 1 is a block diagram of an electronic device 100 (such as a cellular radiotelephone) including a touch sensitive user interface 101 according to some embodiments of the present invention. The electronic device 100, for example, may be a wireless communications device (such as a cellular radiotelephone), a PDA, an audio/picture/video player/recorder, a global positioning (GPS) unit, a gaming device, or any other electronic device including a touch sensitive screen display. Electronic device 100 may also include a controller 111 coupled to touch sensitive user interface 101, a radio transceiver 115 coupled to controller 111, and a memory 117 coupled to controller 111. In addition, a keyboard/keypad 119, a speaker 121, and/or a microphone 123 may be coupled to controller 111. As discussed herein, electronic device 100 may be a cellular radiotelephone configured to provide PDA functionality, data network connectivity (such as Internet browsing), and/or other data functionality.
  • The controller 111 may be configured to communicate through transceiver 115 and antenna 125 over a wireless air interface with one or more RF transceiver base stations and/or other wireless communication devices using one or more wireless communication protocols such as, for example, Global Standard for Mobile (GSM) communication, General Packet Radio Service (GPRS), enhanced data rates for GSM evolution (EDGE), Integrated Digital Enhancement Network (iDEN), code division multiple access (CDMA), wideband-CDMA, CDMA2000, Universal Mobile Telecommunications System (UMTS), WiMAX, and/or HIPERMAN, wireless local area network (e.g., 802.11), and/or Bluetooth. Controller 111 may be configured to carry out wireless communications functionality, such as conventional cellular phone functionality including, but not limited to, voice/video telephone calls and/or data messaging such as text/picture/video messaging.
  • The controller 111 may be further configured to provide various user applications which can include a music/picture/video recorder/player application, an e-mail/messaging application, a calendar/appointment application, and/or other user applications. The audio/picture/video recorder/player application can be configured to record and playback audio, digital pictures, and/or video that are captured by a sensor (e.g., microphone 123 and/or a camera) within electronic device 100, downloaded into electronic device 100 via radio transceiver 115 and controller 111, downloaded into electronic device 100 via a wired connection (e.g., via USB), and/or installed within electronic device 100 such as through a removable memory media. An e-mail/messaging application may be configured to allow a user to generate e-mail/messages (e.g., short messaging services messages and/or instant messages) for transmission via controller 111 and transceiver 115. A calendar/appointment application may provide a calendar and task schedule that can be viewed and edited by a user to schedule appointments and other tasks.
  • More particularly, touch sensitive user interface 101 may be a touch sensitive screen including a display 103, a contact detector 105, and a proximity detector 107. For example, contact detector 105 may be configured to detect contact between a first finger and display 103, and proximity detector 107 may be configured to detect proximity of a second finger to display 103 without contact between the second finger and touch sensitive user interface 101. More particularly, contact detector 105 may be configured to detect contact between first finger and touch sensitive user interface 101 using infrared (IR) contact sensing, acoustic wave contact sensing, capacitive contact sensing, and/or resistive contact sensing. Proximity detector 107 may be configured to detect proximity of the second finger to touch sensitive user interface 101 using acoustic sensing and/or optical sensing. Optical sensing may be provided, for example, using a High Ambient Light Independent Optical System (HALIOS®) as discussed in the reference by Rottmann et al. in “Electronic Concept Fulfils Optical Sensor Dream” published by ELMOS Semiconductor AG at http://www.mechaless.com/images/pdf/Elektronikartikel_ENG.pdf. The disclosure of the Rottmann et al. reference is hereby incorporated herein in its entirety by reference. Optical sensing is also discussed in the reference entitled “HALIOS®—Optics For Human Machine Interfaces,” ELMOS Semiconductor AC, Version 1.0, pages 1-15, Mar. 3, 2008, the disclosure of which is also incorporated herein in its entirety by reference.
  • Accordingly, contact detector 105 may be configured to detect contact using a first sensing technology, and proximity detector 107 may be configured to detect non-contact proximity using a second technology different than the first technology. More particularly, proximity detector 107 may be configured to detect non-contact proximity while the contact detector 105 is detecting contact. For example, contact detector 105 may be configured to detect contact using a first sensing technology such as infrared sensing, acoustic wave sensing, capacitive sensing, and/or resistive sensing, and proximity detector 107 may be configured to detect non-contact proximity using a second sensing technology such as acoustic sensing and/or optical sensing. According to other embodiments of the present invention, a same technology (such as an optical sensing technology) may provide both contact and non-contact proximity sensing so that contact detector 105 and proximity detector 107 may be implemented using a single detector.
  • Accordingly, controller 111 may be configured to select one of a plurality of different operations responsive to detecting contact between a first finger and touch sensitive user interface 101 and responsive to detecting non-contact proximity of a second finger to touch sensitive user interface 101, and then perform the selected operation. As discussed in greater detail below with respect to FIGS. 3A and 3B, by detecting contact of a first finger and non-contact proximity of a second finger relative to display 103 of touch sensitive user interface 101 at the same time, controller 111 may determine which finger (e.g., pointer finger, middle finger, etc.) is in contact with display 103. Accordingly, different operations may be performed depending on the finger making contact with display 103.
  • For example, a web address may be shown on display 103, and contact with the portion of display 103 where the web address is shown may select the web address. Once the web address has been selected, however, one of a plurality of operations relating to the web address may be performed depending on an orientation of a proximate finger relative to the contacting finger. With a right handed user, for example, if the pointer finger is the contacting finger, there will be no proximate finger to the left of the contacting finger, and if the middle finger is the contacting finger, there will be a proximate non-contacting finger (i.e., the pointer finger) to the left of the contacting finger. If the contacting finger is the pointer finger, for example, a communications link may be established with a website identified by the selected web address, and if the contacting finger is the middle finger, another operation (such as a bookmarking operation and/or an editing operation) may be performed using the selected web address.
  • According to other embodiments of the present invention, a contact alias may be shown on display 103. If pointer finger contact is made with the contact alias, a communication (e.g., a telephone call, an e-mail, a text message, etc.) with the contact may be initiated, while if middle finger contact is made with the contact alias, a property(ies) (e.g., telephone number, e-mail address, text message address, etc.) may be shown, and/or an editing operation may be initiated. While differentiation between two fingers is discussed by way of example, differentiation between three or more fingers may be provided as discussed in greater detail below.
  • FIG. 2 is a block diagram of an electronic device 200 including a touch sensitive user interface 201 according to some embodiments of the present invention. The electronic device 200 may be a computing device (such as a laptop computer) including a touch sensitive pad. Device 200 may also include a controller 211 coupled to touch sensitive user interface 201, a network interface 215 coupled to controller 211, and a memory 217 coupled to controller 211. In addition, a display 227, a keyboard/keypad 219, a speaker 221, and/or a microphone 223 may be coupled to controller 211. As discussed herein, device 200 may be a laptop computer configured to provide data network connectivity (such as Internet browsing), and/or other data functionality. Moreover, touch sensitive pad 203 may be provided below a spacebar of keyboard 219 to accept user input of pointer and/or click commands similar to pointer and click commands normally accepted though a computer mouse.
  • The controller 211 may be configured to communicate through network interface 215 with one or more other remote devices over a local area network, a wide area network, and/or the Internet. Controller 211 may be further configured to provide various user applications which can include an audio/picture/video recorder/player application, an e-mail/messaging application, a calendar/appointment application, and/or other user applications. The audio/picture/video recorder/player application can be configured to record and playback audio, digital pictures, and/or video that are captured by a sensor (e.g., microphone 223 and/or a camera) within device 200, downloaded into device 200 via network interface 215 and controller 211, downloaded into device 200 via a wired connection (e.g., via USB), and/or installed within device 200 such as through a removable memory media. An e-mail/messaging application may be configured to allow a user to generate e-mail/messages for transmission via controller 211 and network interface 215. A calendar/appointment application may provide a calendar and task schedule that can be viewed and edited by a user to schedule appointments and other tasks.
  • More particularly, touch sensitive user interface 201 may include a touch sensitive pad 203, a contact detector 205, and a non-contact proximity detector 207. For example, contact detector 205 may be configured to detect contact between a first finger and pad 203, and non-contact proximity detector 207 may be configured to detect non-contact proximity of a second finger to pad 203 without contact between the second finger and the touch sensitive user interface. More particularly, contact detector 205 may be configured to detect contact between the first finger and pad 203 using infrared (IR) contact sensing, acoustic wave contact sensing, capacitive contact sensing, and/or resistive contact sensing. Non-contact proximity detector 207 may be configured to detect non-contact proximity of the second finger to pad 203 using acoustic sensing and/or optical sensing. Optical sensing may be provided, for example, using a High Ambient Light Independent Optical System (HALIOS) as discussed in the reference by Rottmann et al. in “Electronic Concept Fulfils Optical Sensor Dream” published by ELMOS Semiconductor AG at http://www.mechaless.com/images/pdf/Elektronikartikel_ENG.pdf. The disclosure of the Rottmann et al. reference is hereby incorporated herein in its entirety by reference. Optical sensing is also discussed in the reference entitled “HALIOS®—Optics For Human Machine Interfaces,” ELMOS Semiconductor AG, Version 1.0, pages 1-15, Mar. 3, 2008, the disclosure of which is also incorporated herein in its entirety by reference.
  • Accordingly, contact detector 205 may be configured to detect contact using a first sensing technology, and non-contact proximity detector 207 may be configured to detect non-contact proximity using a second technology different than the first technology. More particularly, non-contact proximity detector 207 may be configured to detect non-contact proximity while the contact detector 205 is detecting contact. For example, contact detector 205 may be configured to detect contact using a first sensing technology such as infrared sensing, acoustic wave sensing, capacitive sensing, and/or resistive sensing, and non-contact proximity detector 207 may be configured to detect non-contact proximity using a second sensing technology such as acoustic sensing and/or optical sensing. According to other embodiments of the present invention, a same technology (such as an optical sensing technology) may provide both contact and non-contact proximity sensing so that contact detector 205 and non-contact proximity detector 207 may be implemented using a single detector.
  • Accordingly, controller 211 may be configured to select one of a plurality of different operations responsive to detecting contact between a first finger and touch sensitive user interface 201 and responsive to detecting non-contact proximity of a second finger to touch sensitive user interface 201, and then perform the selected operation. As discussed in greater detail below with respect to FIGS. 3A and 3B, by detecting contact of a first finger and non-contact proximity of a second finger relative to pad 203 of touch sensitive user interface 201 at the same time, controller 211 may determine which finger (e.g., pointer finger, middle finger, etc.) is in contact with pad 203. Accordingly, different operations may be performed depending on the finger making contact with pad 203.
  • For example, touch sensitive user interface 201 may be configured to differentiate between three different fingers (e.g., pointer, middle, and ring fingers) to provide three different command types. With a right handed user, for example, there will be no proximate finger to the left of the contacting finger if the pointer finger is the contacting finger, there will be one non-contacting proximate finger (i.e., the pointer finger) to the left of the contacting finger if the middle finger is the contacting finger, and there will be two non-contacting proximate fingers (i.e., the pointer and middle fingers) to the left of the contacting finger if the ring finger is the contacting finger. To emulate functionality of a computer mouse (without requiring separate click buttons), for example, movement of a pointer finger in contact with pad 203 may be interpreted as a pointer command to move a pointer on display 227; contact of a middle finger with pad 203 may be interpreted as a left mouse click operation; and contact of a ring finger with pad 203 may be interpreted as a right mouse click operation. While differentiation between three fingers is discussed by way of example, differentiation between two or four fingers may be provided as discussed in greater detail below.
  • FIGS. 3A and 3B are schematic illustrations showing operations of a touch sensitive user interface 311 according to some embodiments of the present invention. The operations shown in FIGS. 3A and 3B may be applied to touch sensitive user interface 101 (implemented with touch sensitive screen display 103) of FIG. 1 or to touch sensitive user interface 201 (implemented with touch sensitive pad 203) of FIG. 2. Accordingly, the touch sensitive user interface 311 may be a touch sensitive screen display or a touch sensitive pad. In the example of FIGS. 3A and 3B, the touch sensitive user interface 311 may be configured to differentiate between contact from a pointer finger 331 and a middle finger 332 for right hand use.
  • As shown in FIG. 3A, middle finger 332 may contact interface 331 while pointer finger 331, ring finger 333, and pinky finger 334 are proximate to interface 331 without contacting interface 331. By detecting proximity of one non-contacting finger (i.e., pointer finger 331) to the left of the contacting finger (i.e., middle finger 332), a determination can be made that the contacting finger is middle finger 332, and an appropriate operation corresponding to a middle finger contact may be initiated. In addition, or in an alternative, a determination can be made that the contacting finger is middle finger 332 by detecting proximity of two non-contacting fingers (i.e., ring and pinky fingers 333 and 334) to the right of the contacting finger (i.e., middle finger 332).
  • As shown in FIG. 3B, pointer finger 331 may contact interface 331 while middle finger 332, ring finger 333, and pinky finger 334 are proximate to interface 331 without contacting interface 331. By detecting a lack of proximity of any fingers to the left of the contacting finger (i.e., pointer finger 331), a determination can be made that the contacting finger is pointer finger 331, and an appropriate operation corresponding to pointer finger contact may be initiated (different than the operation corresponding to middle finger contact). In addition, or in an alternative, a determination can be made that the contacting finger is pointer finger 331 by detecting proximity of three non-contacting fingers (i.e., middle, ring, and pinky fingers 332, 333, and 334) to the right of the contacting finger (i.e., pointer finger 332).
  • Moreover, different operations may be assigned to each of the four fingers, and detection operations may be used to determine which of the four fingers is contacting interface 311. Contact by ring finger 333, for example, may be determined by detecting proximity of two non-contacting fingers (i.e., pointer and middle fingers 331 and 332) to the left of the contacting finger (i.e., ring finger 333), and/or by detecting proximity only one non-contacting finger (i.e., pinky finger 334) to the right of contacting finger (i.e., ring finger 333). Contact by pinky finger 334 may be determined by detecting proximity of three non-contacting fingers (i.e., pointer finger 331, middle finger 332, and ring finger 333) to the right of contacting finger (i.e., pinky finger 334), and/or by detecting proximity of no fingers to the right of the contacting finger (i.e., pinky finger 334).
  • Alternate detection criteria (e.g., considering non-contacting proximate fingers to the left and right of the contacting finger) may be used to provide redundancy in the determination and/or to accommodate a situation where the contacting finger is near an edge of interface 311 so that proximate non-contacting fingers on one side of the contacting finger are not within range of detection. Moreover, the examples discussed above are discussed for right hand use. Left hand use, however, may be provided by using a reversed consideration of fingers proximate to the contacting finger. In addition, an electronic device 100/200 incorporating touch sensitive user interface 311/101/201 may provide user selection of right or left hand use. For example, a set-up routine of the electronic device 100/200 may prompt the user to enter a right hand or left hand preference, and the preference may be stored in memory 117/217 of the electronic device 100/200. The controller 111/211 of the electronic device 100/200 may use the stored preference to determine how to interpret finger contact with interface 311/101/201.
  • According to other embodiments of the present invention, operations may be restricted to use of two fingers (e.g., pointer and middle fingers), and determination of the contacting finger may be performed automatically without requiring prior selection/assumption regarding right or left handed use. Stated in other words, touch sensitive user interface 311/101/201 may be configured to differentiate between pointer and middle fingers to provide two different command types responsive to contact with touch sensitive user interface 311/101/201. By way of example, if the pointer finger is the contacting finger, there will be no non-contacting proximate fingers on one side of the contacting finger regardless of light or left handed use. If the middle finger is the contacting finger, there will be non-contacting proximate fingers on both sides of the contacting finger regardless of right or left handed use. Accordingly, determination of pointer or middle finger contact may be performed regardless of right or left handedness and/or regardless of user orientation relative to touch sensitive user interface 311/101/201. For example, determination of pointer or middle finger contact may be performed if the user is oriented normally with respect to touch sensitive user interface 311/101/201 (e.g., with the wrist/arm below the touch sensitive user interface), if the user is oriented sideways with respect to touch sensitive user interface 311/101/201 (e.g., with the wrist/arm to the side of touch sensitive user interface), or if the user is oriented upside down with respect to touch sensitive user interface 311/101/201 (e.g., with the wrist/arm above the touch sensitive user interface).
  • FIG. 4 is a flow chart illustrating operations of an electronic device including a touch sensitive interface according to some embodiments of the present invention. Operations of FIG. 4 may be performed, for example, by an electronic device including a touch sensitive screen display as discussed above with respect to FIG. 1, or by an electronic device including a touch sensitive pad as discussed above with respect to FIG. 2. At block 401, contact between a first finger and the touch sensitive user interface may be detected, for example, using infrared (IR) contact sensing, acoustic wave contact sensing, capacitive contact sensing, and/or resistive contact sensing. At block 403, non-contact proximity of a second finger to the touch sensitive user interface may be detected, for example, using optical sensing. More particularly, non-contact proximity of the second finger may be detected at block 403 while detecting contact of the first finger at block 401, and/or contact of the first finger and non-contact proximity of the second finger may be detected at the same time.
  • Responsive to detecting contact between the first finger and the touch sensitive user interface and responsive to detecting non-contact proximity of the second finger to the touch sensitive user interface, one of a plurality of operations may be selected at block 405. For example, the selection may be based on a determination of relative orientations of the first and second fingers as discussed above with respect to FIGS. 3A and 3B. More particularly, the selection may be based on a determination of which finger (i.e., pointer, middle, ring, or pinky) is the contacting finger, and different operations may be assigned to at least two of the fingers. Responsive to selecting one of the plurality of operations, the selected operation may be performed at block 407.
  • Computer program code for carrying out operations of devices and/or systems discussed above may be written in a high-level programming language, such as Java, C, and/or C++, for development convenience. In addition, computer program code for carrying out operations of embodiments of the present invention may also be written in other programming languages, such as, but not limited to, interpreted languages. Some modules or routines may be written in assembly language or even micro-code to enhance performance and/or memory usage. It will be further appreciated that the functionality of any or all of the program modules may also be implemented using discrete hardware components, one or more application specific integrated circuits (ASICs), or a programmed digital signal processor or microcontroller.
  • Some embodiments of the present invention have been described above with reference to flowchart and/or block diagram illustrations of methods, mobile terminals, electronic devices, data processing systems, and/or computer program products. These flowchart and/or block diagrams further illustrate exemplary operations of processing user input in accordance with various embodiments of the present invention. It will be understood that each block of the flowchart and/or block diagram illustrations, and combinations of blocks in the flowchart and/or block diagram illustrations, may be implemented by computer program instructions and/or hardware operations. These computer program instructions may be provided to a processor of a general purpose computer, a special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions specified in the flowchart and/or block diagram block or blocks.
  • These computer program instructions may also be stored in a computer usable or computer-readable memory that may direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer usable or computer-readable memory produce an article of manufacture including instructions that implement the function specified in the flowchart and/or block diagram block or blocks.
  • The computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions that execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart and/or block diagram block or blocks.
  • In the drawings and specification, there have been disclosed examples of embodiments of the invention and, although specific terms are employed, they are used in a generic and descriptive sense only and not for purposes of limitation, the scope of the invention being set forth in the following claims.

Claims (20)

1. A method of operating an electronic device using a touch sensitive user interface, the method comprising:
detecting contact between a first finger and the touch sensitive user interface;
detecting non-contact proximity of a second finger to the touch sensitive user interface;
responsive to detecting contact between the first finger and the touch sensitive user interface and responsive to detecting non-contact proximity of the second finger to the touch sensitive user interface, selecting one of a plurality of operations; and
responsive to selecting one of the plurality of operations, performing the selected operation.
2. A method according to claim 1 wherein detecting contact comprises detecting contact between the first finger and the touch sensitive user interface using infrared (IR) contact sensing, acoustic wave contact sensing, capacitive contact sensing, and/or resistive contact sensing.
3. A method according to claim 1 wherein detecting non-contact proximity comprises detecting non-contact proximity of the second finger to the touch sensitive user interface using optical sensing.
4. A method according to claim 1 wherein detecting contact comprises detecting contact using a first sensing technology, and wherein detecting non-contact proximity comprises detecting non-contact proximity using a second sensing technology different than the first sensing technology.
5. A method according to claim 4 wherein the first sensing technology is selected from infrared sensing, acoustic sensing, capacitive sensing, and/or resistive sensing, and wherein the second sensing technology is selected from acoustic sensing and/or optical sensing.
6. A method according to claim 1 wherein detecting non-contact proximity of the second finger comprises detecting non-contact proximity of the second finger while detecting contact between the first finger and the touch sensitive user interface.
7. A method according to claim 1, wherein selecting one of a plurality of operations comprises,
determining an orientation of the second finger relative to the first finger,
when the second finger is in a first orientation relative to the first finger, selecting a first of the plurality of operations, and
when the second finger is in a second orientation relative to the first finger different than the first orientation, selecting a second of the plurality of operations.
8. A method according to claim 7 wherein the first operation comprises initiating a link to a website identified by detecting contact between the first finger and the touch sensitive user interface, and wherein the second operation comprises an editing operation and/or a bookmarking operation.
9. A method according to claim 1 further comprising:
detecting non-contact proximity of a third finger to the touch sensitive user interface;
wherein selecting one of the plurality of operations comprises selecting a first of the plurality of operations when the first finger is between the second and third fingers, and selecting a second of the plurality of operations when the second and third fingers are on a same side of the first finger.
10. A method according to claim 1 wherein the touch sensitive user interface comprises a touch sensitive screen and/or a touch sensitive pad.
11. An electronic device comprising:
a touch sensitive user interface including a contact detector configured to detect contact between a first finger and the touch sensitive user interface, and a proximity detector configured to detect non-contact proximity of a second finger to the touch sensitive user interface; and
a controller coupled to the touch sensitive user interface, wherein the controller is configured to select one of a plurality of operations responsive to detecting contact between the first finger and the touch sensitive user interface and responsive to detecting non-contact proximity of the second finger to the touch sensitive user interface, and to perform the selected operation responsive to selecting one of the plurality of operations.
12. An electronic device according to claim 11 wherein the contact detector is configured to detect contact between the first finger and the touch sensitive user interface using infrared (IR) contact sensing, acoustic wave contact sensing, capacitive contact sensing, and/or resistive contact sensing.
13. An electronic device according to claim 11 wherein the proximity detector is configured to detect non-contact proximity of the second finger to the touch sensitive user interface using optical sensing.
14. An electronic device according to claim 11 wherein the contact detector is configured to detect contact using a first sensing technology, and wherein the proximity detector is configured to detect non-contact proximity using a second sensing technology different than the first sensing technology.
15. An electronic device according to claim 14 wherein the first sensing technology is selected from infrared sensing, acoustic sensing, capacitive sensing, and/or resistive sensing, and wherein the second sensing technology is selected from acoustic sensing and/or optical sensing.
16. An electronic device according to claim 11 wherein detecting non-contact proximity of the second finger comprises detecting non-contact proximity of the second finger while detecting contact between the first finger and the touch sensitive user interface.
17. An electronic device according to claim 11, wherein the controller is configured to select one of the plurality of operations by determining an orientation of the second finger relative to the first finger, selecting a first of the plurality of operations when the second finger is in a first orientation relative to the first finger, and selecting a second of the plurality of operations when the second finger is in a second orientation relative to the first finger different than the first orientation.
18. An electronic device according to claim 17 wherein the first operation comprises initiating a link to a website identified by detecting contact between the first finger and the touch sensitive user interface, and wherein the second operation comprises an editing operation and/or a bookmarking operation.
19. An electronic device according to claim 11 wherein the proximity detector is further configured to detect non-contact proximity of a third finger to the touch sensitive user interface, and wherein the controller is configured to select a first of the plurality of operations when the first finger is between the second and third fingers, and to select a second of the plurality of operations when the second and third fingers are on a same side of the first finger.
20. An electronic device according to claim 11 wherein the touch sensitive user interface comprises a touch sensitive screen and/or a touch sensitive pad.
US12/268,502 2008-11-11 2008-11-11 Methods of Operating Electronic Devices Using Touch Sensitive Interfaces with Contact and Proximity Detection and Related Devices and Computer Program Products Abandoned US20100117970A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
US12/268,502 US20100117970A1 (en) 2008-11-11 2008-11-11 Methods of Operating Electronic Devices Using Touch Sensitive Interfaces with Contact and Proximity Detection and Related Devices and Computer Program Products
PCT/IB2009/051941 WO2010055424A1 (en) 2008-11-11 2009-05-12 Methods of operating electronic devices using touch sensitive interfaces with contact and proximity detection and related devices and computer program products

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US12/268,502 US20100117970A1 (en) 2008-11-11 2008-11-11 Methods of Operating Electronic Devices Using Touch Sensitive Interfaces with Contact and Proximity Detection and Related Devices and Computer Program Products

Publications (1)

Publication Number Publication Date
US20100117970A1 true US20100117970A1 (en) 2010-05-13

Family

ID=41020928

Family Applications (1)

Application Number Title Priority Date Filing Date
US12/268,502 Abandoned US20100117970A1 (en) 2008-11-11 2008-11-11 Methods of Operating Electronic Devices Using Touch Sensitive Interfaces with Contact and Proximity Detection and Related Devices and Computer Program Products

Country Status (2)

Country Link
US (1) US20100117970A1 (en)
WO (1) WO2010055424A1 (en)

Cited By (38)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100066696A1 (en) * 2008-09-12 2010-03-18 Samsung Electronics Co. Ltd. Proximity sensor based input system and method for operating the same
US20100271304A1 (en) * 2009-04-24 2010-10-28 Chia-Hsin Yang Method for determining mouse command according to trigger points
US20110175830A1 (en) * 2010-01-19 2011-07-21 Sony Corporation Display control apparatus, display control method and display control program
US20110242055A1 (en) * 2010-04-02 2011-10-06 Samsung Electronics Co., Ltd. Composite touch screen panel
WO2012001412A1 (en) * 2010-06-29 2012-01-05 Elliptic Laboratories As User control of electronic devices
US20120092261A1 (en) * 2010-10-15 2012-04-19 Sony Corporation Information processing apparatus, information processing method, and computer program
US20120268388A1 (en) * 2011-04-21 2012-10-25 Mahmoud Razzaghi Touch screen text selection
US20120313891A1 (en) * 2011-06-08 2012-12-13 Sitronix Technology Corp Distance sensing circuit and touch-control electronic apparatus
US20130076121A1 (en) * 2011-09-22 2013-03-28 Ford Global Technologies, Llc Proximity switch having lockout control for controlling movable panel
US20130088430A1 (en) * 2010-06-25 2013-04-11 Gwangju Institute Of Science And Technology Ultra thin light scanning apparatus for portable information device
WO2013083737A1 (en) * 2011-12-09 2013-06-13 Microchip Technology Germany Ii Gmbh & Co. Kg Electronic device with a user interface that has more than two degrees of freedom, the user interface comprising a touch-sensitive surface and contact-free detection means
US20130201151A1 (en) * 2012-02-08 2013-08-08 Sony Mobile Communications Japan, Inc. Method for detecting a contact
US20130278524A1 (en) * 2012-04-23 2013-10-24 Hon Hai Precision Industry Co., Ltd. Electronic device with touch screen and method and system for protecting same
US20140277825A1 (en) * 2013-03-12 2014-09-18 Audi Ag Vehicle signal lever proximity sensing for lane change intention detection with following recommendation to driver
US9136840B2 (en) 2012-05-17 2015-09-15 Ford Global Technologies, Llc Proximity switch assembly having dynamic tuned threshold
US9184745B2 (en) 2012-04-11 2015-11-10 Ford Global Technologies, Llc Proximity switch assembly and method of sensing user input based on signal rate of change
US9197206B2 (en) 2012-04-11 2015-11-24 Ford Global Technologies, Llc Proximity switch having differential contact surface
US9219472B2 (en) 2012-04-11 2015-12-22 Ford Global Technologies, Llc Proximity switch assembly and activation method using rate monitoring
US9287864B2 (en) 2012-04-11 2016-03-15 Ford Global Technologies, Llc Proximity switch assembly and calibration method therefor
US9311204B2 (en) 2013-03-13 2016-04-12 Ford Global Technologies, Llc Proximity interface development system having replicator and method
US9337832B2 (en) 2012-06-06 2016-05-10 Ford Global Technologies, Llc Proximity switch and method of adjusting sensitivity therefor
US9342214B2 (en) * 2013-04-26 2016-05-17 Spreadtrum Communications (Shanghai) Co., Ltd. Apparatus and method for setting a two hand mode to operate a touchscreen
WO2016119827A1 (en) * 2015-01-28 2016-08-04 Huawei Technologies Co., Ltd. Hand or finger detection device and a method thereof
US9447613B2 (en) 2012-09-11 2016-09-20 Ford Global Technologies, Llc Proximity switch based door latch release
US9465429B2 (en) 2013-06-03 2016-10-11 Qualcomm Incorporated In-cell multifunctional pixel and display
US9507469B2 (en) * 2010-01-19 2016-11-29 Sony Corporation Information processing device, operation input method and operation input program
US9520875B2 (en) 2012-04-11 2016-12-13 Ford Global Technologies, Llc Pliable proximity switch assembly and activation method
US9531379B2 (en) 2012-04-11 2016-12-27 Ford Global Technologies, Llc Proximity switch assembly having groove between adjacent proximity sensors
US9548733B2 (en) 2015-05-20 2017-01-17 Ford Global Technologies, Llc Proximity sensor assembly having interleaved electrode configuration
US9559688B2 (en) 2012-04-11 2017-01-31 Ford Global Technologies, Llc Proximity switch assembly having pliable surface and depression
US9568527B2 (en) 2012-04-11 2017-02-14 Ford Global Technologies, Llc Proximity switch assembly and activation method having virtual button mode
US9654103B2 (en) 2015-03-18 2017-05-16 Ford Global Technologies, Llc Proximity switch assembly having haptic feedback and method
US9660644B2 (en) 2012-04-11 2017-05-23 Ford Global Technologies, Llc Proximity switch assembly and activation method
EP2677404A4 (en) * 2011-02-16 2017-09-27 NEC Corporation Touch input device, electronic apparatus, and input method
US9831870B2 (en) 2012-04-11 2017-11-28 Ford Global Technologies, Llc Proximity switch assembly and method of tuning same
US9944237B2 (en) 2012-04-11 2018-04-17 Ford Global Technologies, Llc Proximity switch assembly with signal drift rejection and method
US10038443B2 (en) 2014-10-20 2018-07-31 Ford Global Technologies, Llc Directional proximity switch assembly
US10112556B2 (en) 2011-11-03 2018-10-30 Ford Global Technologies, Llc Proximity switch having wrong touch adaptive learning and method

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5661635A (en) * 1995-12-14 1997-08-26 Motorola, Inc. Reusable housing and memory card therefor
US20030234768A1 (en) * 2002-05-16 2003-12-25 Junichi Rekimoto Input method and input device
US20060010400A1 (en) * 2004-06-28 2006-01-12 Microsoft Corporation Recognizing gestures and using gestures for interacting with software applications
US20060026521A1 (en) * 2004-07-30 2006-02-02 Apple Computer, Inc. Gestures for touch sensitive input devices
US20070177804A1 (en) * 2006-01-30 2007-08-02 Apple Computer, Inc. Multi-touch gesture dictionary
US20080042979A1 (en) * 2007-08-19 2008-02-21 Navid Nikbin Method and apparatus for executing commands or inputting data based on finger's characteristics and Multi-Finger key

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2003017244A1 (en) * 2001-08-17 2003-02-27 Multidigit, Inc. System and method for selecting actions based on the identification of user's fingers
US10437459B2 (en) * 2007-01-07 2019-10-08 Apple Inc. Multitouch data fusion

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5661635A (en) * 1995-12-14 1997-08-26 Motorola, Inc. Reusable housing and memory card therefor
US20030234768A1 (en) * 2002-05-16 2003-12-25 Junichi Rekimoto Input method and input device
US20060010400A1 (en) * 2004-06-28 2006-01-12 Microsoft Corporation Recognizing gestures and using gestures for interacting with software applications
US20060026521A1 (en) * 2004-07-30 2006-02-02 Apple Computer, Inc. Gestures for touch sensitive input devices
US20070177804A1 (en) * 2006-01-30 2007-08-02 Apple Computer, Inc. Multi-touch gesture dictionary
US20080042979A1 (en) * 2007-08-19 2008-02-21 Navid Nikbin Method and apparatus for executing commands or inputting data based on finger's characteristics and Multi-Finger key

Cited By (61)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100066696A1 (en) * 2008-09-12 2010-03-18 Samsung Electronics Co. Ltd. Proximity sensor based input system and method for operating the same
US20100271304A1 (en) * 2009-04-24 2010-10-28 Chia-Hsin Yang Method for determining mouse command according to trigger points
US9841838B2 (en) 2010-01-19 2017-12-12 Sony Corporation Information processing device, operation input method and operation input program
US11567656B2 (en) 2010-01-19 2023-01-31 Sony Group Corporation Information processing device, operation input method and operation input program
US10386959B2 (en) 2010-01-19 2019-08-20 Sony Corporation Information processing device, operation input method and operation input program
US10013110B2 (en) 2010-01-19 2018-07-03 Sony Corporation Information processing device, operation input method and operation input program
US10606405B2 (en) 2010-01-19 2020-03-31 Sony Corporation Information processing device, operation input method and operation input program
US9507469B2 (en) * 2010-01-19 2016-11-29 Sony Corporation Information processing device, operation input method and operation input program
US8760418B2 (en) * 2010-01-19 2014-06-24 Sony Corporation Display control apparatus, display control method and display control program
US20110175830A1 (en) * 2010-01-19 2011-07-21 Sony Corporation Display control apparatus, display control method and display control program
US11169698B2 (en) 2010-01-19 2021-11-09 Sony Group Corporation Information processing device, operation input method and operation input program
US20230185447A1 (en) * 2010-01-19 2023-06-15 Sony Group Corporation Information Processing Device, Operation Input Method And Operation Input Program
US20110242055A1 (en) * 2010-04-02 2011-10-06 Samsung Electronics Co., Ltd. Composite touch screen panel
US9270854B2 (en) * 2010-06-25 2016-02-23 Gwangju Institute Of Science And Technology Ultra thin light scanning apparatus for portable information device
US20130088430A1 (en) * 2010-06-25 2013-04-11 Gwangju Institute Of Science And Technology Ultra thin light scanning apparatus for portable information device
WO2012001412A1 (en) * 2010-06-29 2012-01-05 Elliptic Laboratories As User control of electronic devices
US9024881B2 (en) * 2010-10-15 2015-05-05 Sony Corporation Information processing apparatus, information processing method, and computer program
US20120092261A1 (en) * 2010-10-15 2012-04-19 Sony Corporation Information processing apparatus, information processing method, and computer program
EP2677404A4 (en) * 2011-02-16 2017-09-27 NEC Corporation Touch input device, electronic apparatus, and input method
US20120268388A1 (en) * 2011-04-21 2012-10-25 Mahmoud Razzaghi Touch screen text selection
US20120313891A1 (en) * 2011-06-08 2012-12-13 Sitronix Technology Corp Distance sensing circuit and touch-control electronic apparatus
US20130076121A1 (en) * 2011-09-22 2013-03-28 Ford Global Technologies, Llc Proximity switch having lockout control for controlling movable panel
US9143126B2 (en) * 2011-09-22 2015-09-22 Ford Global Technologies, Llc Proximity switch having lockout control for controlling movable panel
US10501027B2 (en) 2011-11-03 2019-12-10 Ford Global Technologies, Llc Proximity switch having wrong touch adaptive learning and method
US10112556B2 (en) 2011-11-03 2018-10-30 Ford Global Technologies, Llc Proximity switch having wrong touch adaptive learning and method
JP2015505393A (en) * 2011-12-09 2015-02-19 マイクロチップ テクノロジー ジャーマニー ツー ゲーエ An electronic device with a user interface having three or more degrees of freedom, wherein the user interface includes a touch-sensitive surface and non-contact detection means
KR102008348B1 (en) 2011-12-09 2019-08-07 마이크로칩 테크놀로지 저머니 게엠베하 Eletronic device with a user interface that has more than two degrees of freedom, the user interface comprising a touch-sensitive surface and contact-free detection means
US9323379B2 (en) 2011-12-09 2016-04-26 Microchip Technology Germany Gmbh Electronic device with a user interface that has more than two degrees of freedom, the user interface comprising a touch-sensitive surface and contact-free detection means
KR20140100575A (en) * 2011-12-09 2014-08-14 마이크로칩 테크놀로지 저머니 Ⅱ 게엠베하 운트 콤파니 카게 Eletronic device with a user interface that has more than two degrees of freedom, the user interface comprising a touch-sensitive surface and contact-free detection means
WO2013083737A1 (en) * 2011-12-09 2013-06-13 Microchip Technology Germany Ii Gmbh & Co. Kg Electronic device with a user interface that has more than two degrees of freedom, the user interface comprising a touch-sensitive surface and contact-free detection means
US9182860B2 (en) * 2012-02-08 2015-11-10 Sony Corporation Method for detecting a contact
US20130201151A1 (en) * 2012-02-08 2013-08-08 Sony Mobile Communications Japan, Inc. Method for detecting a contact
US9660644B2 (en) 2012-04-11 2017-05-23 Ford Global Technologies, Llc Proximity switch assembly and activation method
US9831870B2 (en) 2012-04-11 2017-11-28 Ford Global Technologies, Llc Proximity switch assembly and method of tuning same
US9184745B2 (en) 2012-04-11 2015-11-10 Ford Global Technologies, Llc Proximity switch assembly and method of sensing user input based on signal rate of change
US9197206B2 (en) 2012-04-11 2015-11-24 Ford Global Technologies, Llc Proximity switch having differential contact surface
US9219472B2 (en) 2012-04-11 2015-12-22 Ford Global Technologies, Llc Proximity switch assembly and activation method using rate monitoring
US9520875B2 (en) 2012-04-11 2016-12-13 Ford Global Technologies, Llc Pliable proximity switch assembly and activation method
US9531379B2 (en) 2012-04-11 2016-12-27 Ford Global Technologies, Llc Proximity switch assembly having groove between adjacent proximity sensors
US9287864B2 (en) 2012-04-11 2016-03-15 Ford Global Technologies, Llc Proximity switch assembly and calibration method therefor
US9559688B2 (en) 2012-04-11 2017-01-31 Ford Global Technologies, Llc Proximity switch assembly having pliable surface and depression
US9568527B2 (en) 2012-04-11 2017-02-14 Ford Global Technologies, Llc Proximity switch assembly and activation method having virtual button mode
US9944237B2 (en) 2012-04-11 2018-04-17 Ford Global Technologies, Llc Proximity switch assembly with signal drift rejection and method
US9348454B2 (en) * 2012-04-23 2016-05-24 Fu Tai Hua Industry (Shenzhen) Co., Ltd. Electronic device with touch screen and method and system for protecting same
US20130278524A1 (en) * 2012-04-23 2013-10-24 Hon Hai Precision Industry Co., Ltd. Electronic device with touch screen and method and system for protecting same
US9136840B2 (en) 2012-05-17 2015-09-15 Ford Global Technologies, Llc Proximity switch assembly having dynamic tuned threshold
US9337832B2 (en) 2012-06-06 2016-05-10 Ford Global Technologies, Llc Proximity switch and method of adjusting sensitivity therefor
US9447613B2 (en) 2012-09-11 2016-09-20 Ford Global Technologies, Llc Proximity switch based door latch release
US20140277825A1 (en) * 2013-03-12 2014-09-18 Audi Ag Vehicle signal lever proximity sensing for lane change intention detection with following recommendation to driver
US8989916B2 (en) * 2013-03-12 2015-03-24 Volkswagen Ag Vehicle signal lever proximity sensing for lane change intention detection with following recommendation to driver
US9311204B2 (en) 2013-03-13 2016-04-12 Ford Global Technologies, Llc Proximity interface development system having replicator and method
US9342214B2 (en) * 2013-04-26 2016-05-17 Spreadtrum Communications (Shanghai) Co., Ltd. Apparatus and method for setting a two hand mode to operate a touchscreen
US10031602B2 (en) 2013-06-03 2018-07-24 Qualcomm Incorporated Multifunctional pixel and display
US9494995B2 (en) 2013-06-03 2016-11-15 Qualcomm Incorporated Devices and methods of sensing
US9465429B2 (en) 2013-06-03 2016-10-11 Qualcomm Incorporated In-cell multifunctional pixel and display
US9606606B2 (en) 2013-06-03 2017-03-28 Qualcomm Incorporated Multifunctional pixel and display
US9798372B2 (en) 2013-06-03 2017-10-24 Qualcomm Incorporated Devices and methods of sensing combined ultrasonic and infrared signal
US10038443B2 (en) 2014-10-20 2018-07-31 Ford Global Technologies, Llc Directional proximity switch assembly
WO2016119827A1 (en) * 2015-01-28 2016-08-04 Huawei Technologies Co., Ltd. Hand or finger detection device and a method thereof
US9654103B2 (en) 2015-03-18 2017-05-16 Ford Global Technologies, Llc Proximity switch assembly having haptic feedback and method
US9548733B2 (en) 2015-05-20 2017-01-17 Ford Global Technologies, Llc Proximity sensor assembly having interleaved electrode configuration

Also Published As

Publication number Publication date
WO2010055424A1 (en) 2010-05-20

Similar Documents

Publication Publication Date Title
US20100117970A1 (en) Methods of Operating Electronic Devices Using Touch Sensitive Interfaces with Contact and Proximity Detection and Related Devices and Computer Program Products
US9678659B2 (en) Text entry for a touch screen
RU2605359C2 (en) Touch control method and portable terminal supporting same
US8739053B2 (en) Electronic device capable of transferring object between two display units and controlling method thereof
US9110587B2 (en) Method for transmitting and receiving data between memo layer and application and electronic device using the same
US8421756B2 (en) Two-thumb qwerty keyboard
US8381118B2 (en) Methods and devices that resize touch selection zones while selected on a touch sensitive display
US20090160806A1 (en) Method for controlling electronic apparatus and apparatus and recording medium using the method
CN102119376B (en) Multidimensional navigation for touch-sensitive display
US20100053111A1 (en) Multi-touch control for touch sensitive display
US20120056817A1 (en) Location of a touch-sensitive control method and apparatus
US20070263014A1 (en) Multi-function key with scrolling in electronic devices
JP2011512584A (en) Identify and respond to multiple temporally overlapping touches on the touch panel
US20110177798A1 (en) Mobile communication terminal and method for controlling application program
WO2009111138A1 (en) Handwriting recognition interface on a device
KR20150007048A (en) Method for displaying in electronic device
JP5305545B2 (en) Handwritten character input device and portable terminal
KR20110133450A (en) Portable electronic device and method of controlling same
EP2677413B1 (en) Method for improving touch recognition and electronic device thereof
EP1847913A1 (en) Multifunction activation methods and related devices
US20070211039A1 (en) Multifunction activation methods and related devices
EP2685367B1 (en) Method and apparatus for operating additional function in mobile device
US20130069881A1 (en) Electronic device and method of character entry
WO2014003012A1 (en) Terminal device, display-control method, and program
EP2570892A1 (en) Electronic device and method of character entry

Legal Events

Date Code Title Description
AS Assignment

Owner name: SONY ERICSSON MOBILE AB,SWEDEN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:BURSTROM, DAVID PER;OSTSJO, ANDERS WILHELM;REEL/FRAME:021822/0674

Effective date: 20081103

AS Assignment

Owner name: SONY ERICSSON MOBILE COMMUNICATIONS AB,SWEDEN

Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE ASSIGNEE NAME SONY ERICSSON MOBILE AB PREVIOUSLY RECORDED ON REEL 021822 FRAME 0674. ASSIGNOR(S) HEREBY CONFIRMS THE ASSIGNEE NAME IS SONY ERICSSON MOBILE COMMUNICATIONS AB;ASSIGNORS:BURSTROM, DAVID PER;OSTSJO, ANDERS WILHELM;REEL/FRAME:021859/0117

Effective date: 20081103

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION