WO2013075137A1 - Detecting touch input force - Google Patents

Detecting touch input force Download PDF

Info

Publication number
WO2013075137A1
WO2013075137A1 PCT/US2012/065939 US2012065939W WO2013075137A1 WO 2013075137 A1 WO2013075137 A1 WO 2013075137A1 US 2012065939 W US2012065939 W US 2012065939W WO 2013075137 A1 WO2013075137 A1 WO 2013075137A1
Authority
WO
WIPO (PCT)
Prior art keywords
signal
force
touch input
received
location
Prior art date
Application number
PCT/US2012/065939
Other languages
French (fr)
Inventor
Lapoe E. LYNN
Samuel W. Sheng
Shih-Ming Shih
Yenyu Hsieh
Original Assignee
Sentons Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US13/451,288 external-priority patent/US9477350B2/en
Application filed by Sentons Inc. filed Critical Sentons Inc.
Priority to KR1020177016603A priority Critical patent/KR101803261B1/en
Priority to EP12849599.1A priority patent/EP2780783B1/en
Priority to KR1020177033898A priority patent/KR101850680B1/en
Priority to KR1020167021667A priority patent/KR101750300B1/en
Priority to KR1020147013247A priority patent/KR101648143B1/en
Priority to CN201280067220.5A priority patent/CN104169848B/en
Publication of WO2013075137A1 publication Critical patent/WO2013075137A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/041Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
    • G06F3/0414Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means using force sensing means to determine a position
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/041Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/041Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
    • G06F3/0416Control or interface arrangements specially adapted for digitisers
    • G06F3/0418Control or interface arrangements specially adapted for digitisers for error correction or compensation, e.g. based on parallax, calibration or alignment
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/041Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
    • G06F3/043Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means using propagating acoustic waves
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/041Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
    • G06F3/043Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means using propagating acoustic waves
    • G06F3/0433Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means using propagating acoustic waves in which the acoustic waves are either generated by a movable member and propagated within a surface layer or propagated within a surface layer and captured by a movable member
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/041Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
    • G06F3/043Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means using propagating acoustic waves
    • G06F3/0436Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means using propagating acoustic waves in which generating transducers and detecting transducers are attached to a single acoustic waves transmission substrate
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • G06F3/0482Interaction with lists of selectable items, e.g. menus
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • G06F3/04847Interaction techniques to control parameter settings, e.g. interaction with sliders or dials
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • G06F3/04886Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures by partitioning the display area of the touch-screen or the surface of the digitising tablet into independently controllable areas, e.g. virtual keyboards or menus
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • G06F3/165Management of the audio stream, e.g. setting of volume, audio stream path
    • HELECTRICITY
    • H03ELECTRONIC CIRCUITRY
    • H03GCONTROL OF AMPLIFICATION
    • H03G3/00Gain control in amplifiers or frequency changers
    • H03G3/02Manually-operated control
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2203/00Indexing scheme relating to G06F3/00 - G06F3/048
    • G06F2203/041Indexing scheme relating to G06F3/041 - G06F3/045
    • G06F2203/04105Pressure sensors for measuring the pressure or force exerted on the touch surface without providing the touch position
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2203/00Indexing scheme relating to G06F3/00 - G06F3/048
    • G06F2203/041Indexing scheme relating to G06F3/041 - G06F3/045
    • G06F2203/04106Multi-sensing digitiser, i.e. digitiser using at least two different sensing technologies simultaneously or alternatively, e.g. for detecting pen and finger, for saving power or for improving position detection

Definitions

  • APR Elo Touch Systems Acoustic Pulse Recognition
  • the APR system includes transducers attached to the edges of a touchscreen glass that pick up the sound emitted on the glass due to a touch.
  • the surface glass may pick up other external sounds and vibrations that reduce the accuracy and effectiveness of the APR system to efficiently detect a touch input.
  • SAW Surface Acoustic Wave-based technology
  • SAW Surface Acoustic Wave-based technology
  • the SAW technology sends ultrasonic waves in a guided pattern using reflectors on the touch screen to detect a touch.
  • sending the ultrasonic waves in the guided pattern increases costs and may be difficult to achieve.
  • Detecting additional types of inputs, such as multi-touch inputs, may not be possible or may be difficult using SAW or APR technology.
  • Figure 1 is a block diagram illustrating an embodiment of a system for detecting a surface disturbance.
  • Figure 2 is a block diagram illustrating an embodiment of a system for detecting a touch input.
  • Figure 3 is a flow chart illustrating an embodiment of a process for calibrating and validating touch detection.
  • Figure 4 is a flow chart illustrating an embodiment of a process for detecting a user touch input.
  • Figure 5 is a flow chart illustrating an embodiment of a process for determining a location associated with a disturbance on a surface.
  • Figure 6 is a flow chart illustrating an embodiment of a process for determining a location associated with a disturbance.
  • Figure 7 is a flowchart illustrating an embodiment of a process of determining a force associated with a user input.
  • Figure 8 is a flowchart illustrating an embodiment of a process for determining an entry of a data structure used to determine a force intensity identifier.
  • Figure 9 includes graphs illustrating examples of a relationship between a normalized amplitude value of a measured disturbance and an applied force.
  • Figure 10 is a flowchart illustrating an embodiment of a process for providing a combined force.
  • Figure 11 is a flowchart illustrating an embodiment of a process for providing a user interface interaction.
  • Figure 12 is a diagram showing an example user interface interaction using force information to drag and drop an item into a file system folder.
  • Figure 13 is a diagram showing an example user interface interaction using force information to provide a context menu.
  • Figure 14 and Figure 15 are diagrams showing examples of user interface interactions using force information to navigate a menu.
  • Figure 16 is a diagram showing an example user interface interaction using force information to interact with a virtual keyboard.
  • Figure 17 and Figure 18 are diagrams showing example user interface interactions using force information to zoom and select user interface objects.
  • Figure 19 is a graph illustrating an example of a relationship between detected touch input force and direction of change in audio volume.
  • Figure 20 is a diagram showing an example user interface interaction using force information to interact with a slider bar.
  • the invention can be implemented in numerous ways, including as a process; an apparatus; a system; a composition of matter; a computer program product embodied on a computer readable storage medium; and/or a processor, such as a processor configured to execute instructions stored on and/or provided by a memory coupled to the processor.
  • these implementations, or any other form that the invention may take, may be referred to as techniques.
  • the order of the steps of disclosed processes may be altered within the scope of the invention.
  • a component such as a processor or a memory described as being configured to perform a task may be implemented as a general component that is temporarily configured to perform the task at a given time or a specific component that is manufactured to perform the task.
  • the term 'processor' refers to one or more devices, circuits, and/or processing cores configured to process data, such as computer program instructions.
  • an acoustic transducer transmits an acoustic wave through a medium of a touch input surface.
  • the acoustic wave may be scattered by the touch input producing a scattered acoustic wave.
  • An acoustic detector that detects the scattered acoustic wave and the acoustic detector outputs a signal indicating variation of the acoustic wave that is indicative of an amount of force associated with the touch input.
  • the force of a touch input is associated with the amount of deflection or movement of a touch surface medium caused by a touch input.
  • the pressure of a touch input is the force of touch input per unit area of the touch input. For example, the total force of a touch input divided by the area of contact of the touch input equals the pressure of the touch input.
  • pressure of a touch input may be used as well.
  • the pressure of the touch input may stay substantially constant because the size of the fingertip in contact with the surface becomes larger due to the softness of the fingertip.
  • the total force of the touch input may be used instead of the pressure of the touch input.
  • a force of a touch input is used to provide user interface interaction.
  • a user touch input on the glass surface of a display screen is detected.
  • a signal such as an acoustic or ultrasonic signal is propagated freely through a propagating medium with a surface using a transmitter coupled to the medium.
  • the propagated signal is disturbed (e.g., the touch causes an interference with the propagated signal).
  • the disturbed signal is received at a sensor coupled to the propagating medium. By processing the received signal and comparing it against an expected signal without the disturbance, a location on the surface associated with the touch input is at least in part determined.
  • the disturbed signal is received at a plurality of sensors and a relative time difference between when the disturbed signal was received at different sensors is used to determine the location on the surface.
  • the touch includes a physical contact to a surface using a human finger, pen, pointer, stylus, and/or any other body parts or objects that can be used to contact or disturb the surface.
  • the touch includes an input gesture and/or a multi-touch input.
  • the disturbed signal is used to determine one or more of the following associated with a touch input: a gesture, a coordinate position, a time, a time frame, a direction, a velocity, a force magnitude, a proximity magnitude, a pressure, a size, and other measurable or derived parameters.
  • touch input detection technology can be applied to larger surface regions with less or no additional cost due to a larger surface region as compared to certain previous touch detection technologies. Additionally, the optical transparency of a touch screen may not have to be affected as compared to resistive and capacitive touch technologies.
  • the touch detection described herein can be applied to a variety of objects such as a kiosk, an ATM, a computing device, an entertainment device, a digital signage apparatus, a cell phone, a tablet computer, a point of sale terminal, a food and restaurant apparatus, a gaming device, a casino game and application, a piece of furniture, a vehicle, an industrial application, a financial application, a medical device, an appliance, and any other objects or devices having surfaces.
  • objects such as a kiosk, an ATM, a computing device, an entertainment device, a digital signage apparatus, a cell phone, a tablet computer, a point of sale terminal, a food and restaurant apparatus, a gaming device, a casino game and application, a piece of furniture, a vehicle, an industrial application, a financial application, a medical device, an appliance, and any other objects or devices having surfaces.
  • Figure 1 is a block diagram illustrating an embodiment of a system for detecting a surface disturbance.
  • the system shown in Figure 1 is included in a kiosk, an ATM, a computing device, an entertainment device, a digital signage apparatus, a cell phone, a tablet computer, a point of sale terminal, a food and restaurant apparatus, a gaming device, a casino game and application, a piece of furniture, a vehicle, an industrial application, a financial application, a medical device, an appliance, and any other objects or devices having surfaces.
  • Propagating signal medium 102 is coupled to transmitters 104, 106, 108, and 110 and sensors 112, 114, 116, and 118.
  • the propagating medium includes one or more of the following: panel, table, glass, screen, door, floor, whiteboard, glass, plastic, wood, steel, metal, semiconductor, insulator, conductor, and any medium that is able to propagate an acoustic or ultrasonic signal.
  • medium 102 is glass of a display screen.
  • a first surface of medium 102 includes a surface area where a user may touch to provide a selection input and a substantially opposite surface of medium 102 is coupled to the transmitters and sensors shown in Figure 1.
  • a surface of medium 102 is substantially flat, curved, or combinations thereof and may be configured in a variety of shapes such as rectangular, square, oval, circular, trapezoidal, annular, or any combination of these, and the like.
  • transmitters 104, 106, 108, and 110 include piezoelectric transducers, electromagnetic transducers, transmitters, sensors and/or any other transmitters and transducers capable of propagating a signal through medium 102.
  • sensors 112, 114, 116, and 118 include piezoelectric transducers, electromagnetic transducers, transmitters and/or any other sensors and transducers capable of detecting a signal on medium 102.
  • the transmitters and sensors shown in Figure 1 are coupled to medium 102 in a manner that allows a user input to be detected in a predetermined region of medium 102. Although four transmitters and four sensors are shown, any number of transmitters and any number of sensors may be used in other embodiments.
  • transmitters and three sensors may be used.
  • a single transducer acts as both a transmitter and a sensor.
  • transmitter 104 and sensor 112 represent a single piezoelectric transducer.
  • transmitter 104 may propagate a signal through medium 102.
  • Sensors 112, 114, 116, and 118 receive the propagated signal.
  • the transmitters/sensors in Figure 1 are attached to a flexible cable coupled to medium 102 via an encapsulant and/or glue material and/or fasteners.
  • Touch detector 120 is connected to the transmitters and sensors shown in Figure 1.
  • detector 120 includes one or more of the following: an integrated circuit chip, a printed circuit board, a processor, and other electrical components and connectors. Detector 120 determines and sends a signal to be propagated by transmitters 104, 106, 108, and 110.
  • Detector 120 also receives the signal detected by sensors 112, 114, 116, and 118. The received signals are processed by detector 120 to determine whether a disturbance associated with a user input has been detected at a location on a surface of medium 102 associated with the disturbance. Detector 120 is in communication with application system 122. Application system 122 uses information provided by detector 120. For example, application system 122 receives from detector 120 a coordinate associated with a user touch input that is used by application system 122 to control a software application of application system 122. In some embodiments, application system 122 includes a processor and/or memory/storage. In other embodiments, detector 120 and application system 122 are at least in part included/processed in a single processor.
  • An example of data provided by detector 120 to application system 122 includes one or more of the following associated with a user indication: a location coordinate of a surface of medium 102, a gesture, simultaneous user indications (e.g., multi-touch input), a time, a status, a direction, a velocity, a force magnitude, a proximity magnitude, a pressure, a size, and other measurable or derived information.
  • a user indication e.g., a location coordinate of a surface of medium 102, a gesture, simultaneous user indications (e.g., multi-touch input), a time, a status, a direction, a velocity, a force magnitude, a proximity magnitude, a pressure, a size, and other measurable or derived information.
  • a touch input is received at location 130 on a surface of medium 102.
  • a user touches the surface of medium 102 at location 130.
  • one or more of transmitters 104, 106, 108, and 110 transmit one or more active signals that are propagated through medium 102.
  • the touch input at location 130 disturbs (e.g., scatters) the propagated signal(s) and the disturbed signals are received at sensors 112, 114, 116, and 118. By measuring the disturbance(s) of the propagated signal(s), the location and/or a force associated with the touch input may be determined.
  • FIG. 2 is a block diagram illustrating an embodiment of a system for detecting a touch input.
  • touch detector 202 is included in touch detector 120 of Figure 1.
  • the system of Figure 2 is integrated in an integrated circuit chip.
  • Touch detector 202 includes system clock 204 that provides a synchronous system time source to one or more other components of detector 202.
  • Controller 210 controls data flow and/or commands between microprocessor 206, interface 208, DSP engine 220, and signal generator 212.
  • microprocessor 206 processes instructions and/or calculations that can be used to program software/firmware and/or process data of detector 202.
  • a memory is coupled to microprocessor 206 and is configured to provide microprocessor 206 with
  • Signal generator 212 generates a signal to be used to propagate a signal such as a signal propagated by transmitter 104 of Figure 1. For example, signal generator 212 generates a pseudorandom binary sequence signal.
  • Driver 214 receives the signal from generator 212 and drives one or more transmitters, such as transmitters 104, 106, 108, and 110 of Figure 1, to propagate a signal through a medium.
  • a signal detected from a sensor such as sensor 112 of Figure 1 is received by detector 202 and signal conditioner 216 conditions (e.g., filters) the received analog signal for further processing.
  • signal conditioner 216 receives the signal outputted by driver 214 and performs echo cancellation of the signal received by signal conditioner 216.
  • the conditioned signal is converted to a digital signal by analog-to-digital converter 218.
  • the converted signal is processed by digital signal processor engine 220.
  • DSP engine 220 correlates the converted signal against a reference signal. The result of the correlation may be used by
  • microprocessor 206 to determine a location associated with a user touch input.
  • the DSP engine determines an amplitude change associated with the converted signal and a reference signal.
  • the amplitude change may be used by microprocessor 206 to determine a force associated with a user touch input.
  • Interface 208 provides an interface for microprocessor 206 and controller 210 that allows an external component to access and/or control detector 202.
  • interface 208 allows detector 202 to communicate with application system 122 of Figure 1 and provides the application system with location information associated with a user touch input.
  • Figure 3 is a flow chart illustrating an embodiment of a process for calibrating and validating touch detection.
  • the process of Figure 3 is used at least in part to calibrate and validate the system of Figure 1 and/or the system of Figure 2.
  • locations of signal transmitters and sensors with respect to a surface are determined. For example, locations of transmitters and sensors shown in Figure 1 are determined with respect to their location on a surface of medium 102.
  • determining the locations includes receiving location information.
  • one or more of the locations may be fixed and/or variable.
  • calibrating the transmitter includes calibrating a characteristic of a signal driver and/or transmitter (e.g., strength).
  • calibrating the sensor includes calibrating a characteristic of a sensor (e.g., sensitivity).
  • the calibration of 304 is performed to optimize the coverage and improve signal-to-noise transmission/detection of a signal (e.g., acoustic or ultrasonic) to be propagated through a medium and/or a disturbance to be detected.
  • a signal e.g., acoustic or ultrasonic
  • one or more components of the system of Figure 1 and/or the system of Figure 2 are tuned to meet a signal-to-noise requirement.
  • the calibration of 304 depends on the size and type of a transmission/propagation medium and geometric configuration of the transmitters/sensors. In some embodiments, the calibration of step 304 includes detecting a failure or aging of a transmitter or sensor. In some embodiments, the calibration of step 304 includes cycling the transmitter and/or receiver. For example, to increase the stability and reliability of a piezoelectric transmitter and/or receiver, a burn-in cycle is performed using a burn-in signal. In some
  • the step of 304 includes configuring at least one sensing device within a vicinity of a predetermined spatial region to capture an indication associated with a disturbance using the sensing device.
  • the disturbance is caused in a selected portion of the input signal corresponding to a selection portion of the predetermined spatial region.
  • a test signal is propagated through a medium such as medium 102 of Figure 1 to determine an expected sensed signal when no disturbance has been applied.
  • a test signal is propagated through a medium to determine a sensed signal when one or more predetermined disturbances (e.g., predetermined touch) are applied at a predetermined location. Using the sensed signal, one or more components may be adjusted to calibrate the disturbance detection.
  • predetermined disturbances e.g., predetermined touch
  • a validation of a touch detection system is performed.
  • the system of Figure 1 and/or Figure 2 is testing using predetermined disturbance patterns to determine detection accuracy, detection resolution, multi-touch detection, and/or response time. If the validation fails, the process of Figure 3 may be at least in part repeated and/or one or more components may be adjusted before performing another validation.
  • Figure 4 is a flow chart illustrating an embodiment of a process for detecting a user touch input.
  • the process of Figure 4 is at least in part implemented on touch detector 120 of Figure 1 and/or touch detector 202 of Figure 2.
  • a signal that can be used to propagate an active signal through a surface region is sent.
  • sending the signal includes driving (e.g., using driver 214 of Figure 2) a transmitter such as a transducer (e.g., transmitter 104 of Figure 1) to propagate an active signal (e.g., acoustic or ultrasonic) through a propagating medium with the surface region.
  • a transducer e.g., transmitter 104 of Figure 1
  • the signal includes a sequence selected to optimize autocorrelation (e.g., resulting in narrow/short peak) of the signal.
  • the signal includes a Zadoff-Chu sequence.
  • the signal includes a pseudorandom binary sequence with or without modulation.
  • the propagated signal is an acoustic signal.
  • the propagated signal is an ultrasonic signal (e.g., outside the range of human hearing).
  • the propagated signal is a signal above 20 kHz (e.g., within the range between 80 kHz to 100 kHz). In other embodiments, the propagated signal may be within the range of human hearing.
  • a user input on or near the surface region can be detected by detecting disturbances in the active signal when it is received by a sensor on the propagating medium.
  • the active signal is used in addition to receiving a passive signal from a user input to determine the user input.
  • the active signal that has been disturbed by a disturbance of the surface region is received.
  • the disturbance may be associated with a user touch indication.
  • the disturbance causes the active signal that is propagating through a medium to be attenuated and/or delayed.
  • the disturbance in a selected portion of the active signal corresponds to a location on the surface that has been indicated (e.g., touched) by a user.
  • the received signal is processed to at least in part determine a location associated with the disturbance.
  • determining the location includes extracting a desired signal from the received signal at least in part by removing or reducing undesired components of the received signal such as disturbances caused by extraneous noise and vibrations not useful in detecting a touch input.
  • determining the location includes comparing the received signal to a reference signal that has not been affected by the disturbance. The result of the comparison may be used with a result of other comparisons performed using the reference signal and other signal(s) received at a plurality of sensors.
  • the location in some embodiments, is a location (e.g., a location coordinate) on the surface region where a user has provided a touch input.
  • one or more of the following information associated with the disturbance may be determined at 406: a gesture, simultaneous user indications (e.g., multi-touch input), a time, a status, a direction, a velocity, a force magnitude, a proximity magnitude, a pressure, a size, and other measurable or derived information.
  • the location is not determined at 406 if a location cannot be determined using the received signal and/or the disturbance is determined to be not associated with a user input.
  • Figure 4 shows receiving and processing an active signal that has been disturbed
  • a received signal has not been disturbed by a touch input and the received signal is processed to determine that a touch input has not been detected.
  • An indication that a touch input has not been detected may be provided/outputted.
  • Figure 5 is a flow chart illustrating an embodiment of a process for determining a location associated with a disturbance on a surface.
  • the process of Figure 5 is included in 406 of Figure 4.
  • the process of Figure 5 may be implemented in touch detector 120 of Figure 1 and/or touch detector 202 of Figure 2.
  • a received signal is conditioned.
  • the received signal is a signal including a pseudorandom binary sequence that has been freely propagated through a medium with a surface that can be used to receive a user input.
  • the received signal is the signal that has been received at 404 of Figure 4.
  • conditioning the signal includes filtering or otherwise modifying the received signal to improve signal quality (e.g., signal-to-noise ratio) for detection of a pseudorandom binary sequence included in the received signal and/or user touch input.
  • conditioning the received signal includes filtering out from the signal extraneous noise and/or vibrations not likely associated with a user touch indication.
  • an analog to digital signal conversion is performed on the signal that has been conditioned at 502.
  • any number of standard analog to digital signal converters may be used.
  • the resulting digital signal is used to perform a first correlation at 506.
  • performing the first correlation includes correlating the converted signal with a reference signal.
  • Performing the correlation includes cross-correlating or determining a convolution (e.g., interferometry) of the converted signal with a reference signal to measure the similarity of the two signals as a time-lag is applied to one of the signals. By performing the correlation, the location of a portion of the converted signal that most corresponds to the reference signal can be located.
  • a result of the correlation can be plotted as a graph of time within the received and converted signal (e.g., time-lag between the signals) vs. a measure of similarity.
  • the associated time value of the largest value of the measure of similarity corresponds to the location where the two signals most correspond.
  • a reference time value e.g., at 306 of Figure 3
  • a force associated with a touch indication may be determined.
  • the reference signal is determined based at least in part on the signal that was propagated through a medium (e.g., based on a source pseudorandom binary sequence signal that was propagated). In some embodiments, the reference signal is at least in part determined using information determined during calibration at 306 of Figure 3. The reference signal may be chosen so that calculations required to be performed during the correlation may be simplified. For example, the reference signal used in 506 is a simplified reference signal that can be used to efficiently correlate the reference signal over a relatively large time difference (e.g., lag-time) between the received and converted signal and the reference signal.
  • a relatively large time difference e.g., lag-time
  • a second correlation is performed based on a result of the first correlation.
  • Performing the second correlation includes correlating (e.g., cross-correlation or convolution similar to step 506) the converted signal in 504 with a second reference signal.
  • the second reference signal is a more complex/detailed (e.g., more computationally intensive) reference signal as compared to the first reference signal used in 506.
  • the second correlation is performed in 508 because using the second reference signal in 506 may be too computationally intensive for the time interval required to be correlated in 506.
  • Performing the second correlation based on the result of the first correlation includes using one or more time values determined as a result of the first correlation.
  • a range of likely time values e.g., time-lag
  • the second correlation is performed using the second reference signal only across the determined range of time values to fine tune and determine the time value that most corresponds to where the second reference signal (and, by association, also the first reference signal) matched the received signal.
  • the first and second correlations have been used to determine a portion within the received signal that correspond to a disturbance caused by a touch input at a location on a surface of a propagating medium.
  • the second correlation is optional. For example, only a single correlation step is performed.
  • a result of the second correlation is used to at least in part determine a location associated with a disturbance.
  • determining the location includes comparing a determined time value where the signals of the second correlation are most correlated and comparing the determined time value with a reference time value (e.g., determined at 306 of Figure 3) not associated with a touch input disturbance, to determine a time delay/offset or phase difference caused on the received signal due to the disturbance (e.g., caused by a touch input).
  • This time delay is associated with a signal received at a first sensor and other time delays due to the disturbance at other signals received at other sensors are used to calculate a location of the disturbance relative to the locations of the sensors.
  • Figure 6 is a flowchart illustrating an embodiment of a process for determining a location associated with a disturbance.
  • the process of Figure 6 is included in 510 of Figure 5.
  • a plurality of results of correlations performed on a plurality of signals disturbed by a disturbance of a surface is received.
  • a result of the correlation performed at 508 of Figure 5 is received.
  • a signal is propagated using transmitter 104 and sensors 114, 116, and 118 each receives the propagated signal that has been disturbed by a touch input on or near a surface of medium 102 of Figure 1.
  • the propagated signal may contain a predetermined signal and the predetermined signal is received at the various sensors.
  • Each of the received signals is correlated with a reference signal to determine the results received at 602.
  • the received results are associated with a same signal content (e.g., same binary sequence) that has been freely propagated on a medium at the same time.
  • the received results are associated with different signal contents that have been disturbed by the same disturbance.
  • time differences associated with the plurality of results are used to determine a location associated with the disturbance.
  • each of the time differences is associated with a time when signals used in the correlation are most correlated.
  • the time differences are associated with a determined time delay/offset or phase difference caused on the received signal due to the disturbance.
  • This time delay may be calculated by comparing a time value determined using a correlation with a reference time value that is associated with a scenario where a touch input has not been specified. The result of the comparison may be used to calculate a location of the disturbance relative to the locations of sensors that received the plurality of signals. By using the location of the sensors relative to a surface of a medium that has propagated the received signal, a location on the surface where the disturbance originated may be determined.
  • Figure 7 is a flowchart illustrating an embodiment of a process of determining a force associated with a user input.
  • the process of Figure 7 may be implemented on touch detector 120 of Figure 1 and/or touch detector 202 of Figure 2.
  • a location associated with a user input on a touch input surface is determined.
  • at least a portion of the process of Figure 4 is included in step 702.
  • the process of Figure 4 is used to determine a location associated with a user touch input.
  • a location associated with a user input at location 130 on a surface of medium 102 of Figure 1 is determined.
  • one or more received signals are selected to be evaluated.
  • selecting the signal(s) to be evaluated include selecting one or more desired signals from a plurality of received signals used to detect the location associated with the user input. For example, one or more signals received in step 404 of Figure 4 are selected. In some embodiments, the selected signal(s) are selected based at least in part on a signal-to-noise ratio associated with signals. In some embodiments, one or more signals with the highest signal-to-noise ratio are selected. For example, when an active signal that is propagated through a touch input surface medium is disturbed/scattered by a touch input, the disturbed signal is detected/received at various detectors/sensors/receivers coupled to the medium. The received disturbed signals may be subject to other undesirable disturbances such as other minor vibration sources (e.g., due to external audio vibration, device movement, etc.) that also disturb the active signal. The effects of these
  • undesirable disturbances may be larger on received signals that were received further away from the location of the touch input.
  • a variation (e.g., disturbance such as amplitude change) detected in an active signal received at a receiver/sensor may be greater at certain receivers (e.g., receivers located closest to the location of the touch input) as compared to other receivers.
  • touch input at location 130 disturbs an active signal sent by transmitter 104.
  • the disturbed active signal is received at sensors/receivers 112, 114, 116, and 118. Because sensor/receiver 114 is located closest to touch input location 130, it has received a disturbed signal with the largest amplitude variation that is proportional to the force of the touch input.
  • the selected signals may have been selected at least in part by examining the amplitude of a detected disturbance. For example, one or more signals with the highest amplitude associated with a detected touch input disturbance are selected. In some embodiments, based at least in part on a location determined in 702, one or more signals received at one or more receivers located closest to the touch input location are selected. In some
  • a plurality of active signals is used to detect a touch input location and/or touch input force intensity.
  • One or more received signals to be used to determine a force intensity may be selected for each of the active signals.
  • one or more received signals to be used to determine the force intensity may be selected across the received signals of all the active signals.
  • the one or more selected signals are normalized.
  • normalizing a selected signal includes adjusting (e.g., scaling) an amplitude of the selected signal based on a distance value associated with the selected signal. For example, although an
  • amount/intensity of force of a touch input may be detected by measuring an amplitude of a received active signal that has been disturbed by the force of the touch input, other factors such as the location of the touch input with respect to a receiver that has received the disturbed signal and/or location of the transmitter transmitting the active signal may also affect the amplitude of the received signal used to determine the intensity of the force.
  • a distance value/identifier associated with one or more of the following is used to determine a scaling factor used to scale a selected signal: a distance between a location of a touch input and a location of a receiver that has received the selected signal, a distance between a location of a touch input and a location of a transmitter that has transmitted an active signal that has been disturbed by a touch input and received as the selected signal, a distance between a location of a receiver that has received the selected signal and a location of a transmitter that has transmitted an active signal that has been disturbed by a touch input and received as the selected signal, and a combined distance of a first distance between a location of a touch input and a location of a receiver that has received the selected signal and a second distance between the location of the touch input and a location of a transmitter that has transmitted an active signal that has been disturbed by a touch input and received as the selected signal.
  • each of one or more selected signals is normalized by a different amount (e.g., different amount
  • a force intensity identifier associated with the one or more normalized signals is determined.
  • the force intensity identifier may include a numerical value and/or other identifier identifying a force intensity.
  • an associated force may be determined for each normalized signal and the determined forces may be averaged and/or weighted-averaged to determine the amount of the force. For example, in the case of weighted averaging of the force values, each determined force value is weighted based on an associated signal-to-noise ratio, an associated amplitude value, and/or an associated distance value between a receiver of the normalized signal and the location of the touch input.
  • the amount of force is determined using a measured amplitude associated with a disturbed portion of the normalized signal.
  • the normalized signal represents a received active signal that has been disturbed when a touch input was provided on a surface of a medium that was propagating the active signal.
  • a reference signal may indicate a reference amplitude of a received active signal if the active signal was not disturbed by a touch input.
  • an amplitude value associated with an amplitude change to the normalized signal caused by a force intensity of a touch input is determined.
  • the amplitude value may be a measured amplitude of a disturbance detected in a normalized signal or a difference between a reference amplitude and the measured amplitude of the disturbance detected in the normalized signal.
  • the amplitude value is used to obtain an amount/intensity of a force.
  • the use of the amplitude value includes using the amplitude value to look up in a data structure (e.g., table, database, chart, graph, lookup table, list, etc.) a corresponding associated force intensity.
  • a data structure e.g., table, database, chart, graph, lookup table, list, etc.
  • the data structure includes entries associating a signal disturbance amplitude value and a corresponding force intensity identifier.
  • the data structure may be predetermined/pre-computed. For example, for a given device, a controlled amount of force is applied and the disturbance effect on an active signal due to the controlled amount of force is measured to determine an entry for the data structure.
  • the force intensity may be varied to determine other entries of the data structure.
  • the data structure is associated with a specific receiver that received the signal included in the normalized signal.
  • the data structure includes data that has been specifically determined for characteristics of a specific receiver (e.g., for sensor/receiver 114 of Figure 1).
  • the use of the amplitude value to look up a corresponding force intensity identifier stored in a data structure includes selecting a specific data structure and/or a specific portion of a data structure
  • the data structure is associated with a plurality of receivers.
  • the data structure includes entries associated with averages of data determined for characteristics of each receiver in the plurality of receivers.
  • the same data structure may be used for a plurality of normalized signals associated with various receivers.
  • the use of the amplitude value includes using the amplitude value in a formula that can be used to simulate and/or calculate a corresponding force intensity.
  • the amplitude value is used as an input to a predetermined formula used to compute a corresponding force intensity.
  • the formula is associated with a specific receiver that received the signal of the normalized signal.
  • the formula includes one or more parameters (e.g., coefficients) that have been specifically determined for characteristics of a specific receiver (e.g., for sensor/receiver 114 of Figure 1).
  • the use of the amplitude value in a formula calculation includes selecting a specific formula corresponding to the normalized signal and/or a receiver that received the signal included in the normalized signal.
  • a single formula is associated with a plurality of receivers.
  • a formula includes averaged parameter values of parameter values that have been specifically determined for characteristics for each of the receivers in the plurality of receivers.
  • the same formula may be used for a plurality of normalized signals associated with different receivers.
  • the determined force intensity identifier is provided.
  • providing the force intensity identifier includes providing the identifier (e.g., a numerical value, an identifier within a scale, etc.) to an application such as an application of application system 122 of Figure 1.
  • the provided force intensity identifier is provided with a corresponding touch input location identifier determined in step 406 of Figure 4.
  • the provided force intensity identifier is used to provide a user interface interaction.
  • Figure 8 is a flowchart illustrating an embodiment of a process for determining an entry of a data structure used to determine a force intensity identifier.
  • the process of Figure 8 is included in step 304 of Figure 3.
  • the process of Figure 8 is used at least in part to create the data structure that may be used in step 708 of Figure 7.
  • the process of Figure 8 is used at least in part to calibrate the system of Figure 1 and/or the system of Figure 2.
  • the process of Figure 8 is used at least in part to determine a data structure that can be included in one or more devices to be manufactured to determine a force intensity identifier/value corresponding to an amplitude value of a disturbance detected in the received active signal.
  • the data structure may be determined for a plurality of similar devices to be manufactured or the data structure may be determined for a specific device taking into account the manufacturing variation of the device.
  • a controlled amount of force is applied at a selected location on a touch input surface.
  • the force is provided on a location of a surface of medium 102 of Figure 1 where a touch input may be provided.
  • a tip of a pointer e.g., stylus
  • a controlled amount of force is applied on a touch input surface while an active signal is being propagated through a medium of the touch input surface.
  • the amount of force applied in 802 may be one of a plurality of different amounts of force that will be applied on the touch input surface.
  • an effect of the applied force is measured using one or more receivers.
  • measuring the effect includes measuring an amplitude associated with a disturbed portion of an active signal that has been disturbed when the force was applied in 802 and that has been received by the one or more receivers.
  • the amplitude may be a directly measured amplitude value or a difference between a reference amplitude and a detected amplitude.
  • the signal received by the one or more receivers is normalized before the amplitude is measured.
  • normalizing a received signal includes adjusting (e.g., scaling) an amplitude of the signal based on a distance value associated with the selected signal.
  • a reference signal may indicate a reference amplitude of a received active signal that has not been disturbed by a touch input.
  • an amplitude value associated with an amplitude change caused by a disturbance of a touch input is determined.
  • the amplitude value may be a measured amplitude value of a disturbance detected in a normalized signal or a difference between a reference amplitude and the measured amplitude value of the disturbance detected in the normalized signal.
  • the amplitude value is used to obtain an identifier of a force intensity.
  • a distance value associated with one or more of the following is used to determine a scaling factor used to scale a received signal before an effect of a disturbance is measured using the received signal: a distance between a location of a touch input and a location of a receiver that has received the selected signal, a distance between a location of the force input and a location of a transmitter that has transmitted an active signal that has been disturbed by the force input and received by the receiver, a distance between a location of the receiver and a location of a transmitter that has transmitted an active signal that has been disturbed by the force input and received by the receiver, and a combined distance of a first distance between a location of a force input and a location of the receiver and a second distance between the location of the force input and a location of a transmitter that has transmitted an active signal that has been disturbed by the force input and received by the receiver.
  • each of one or more signals received by different receivers is normalized by a different amount (e.g., different amplitude scaling
  • storing the data includes storing an entry in a data structure such as the data structure that may be used in step 708 of Figure 7. For example, an entry that associates the amplitude value determined in 804 and an identifier associated with an amount of force applied in 802 is stored in the data structure.
  • storing the data includes indexing the data by an amplitude value determined in 804. For example, the stored data may be retrieved from the storage using the amplitude value.
  • the data structure is determined for a specific signal receiver.
  • a data structure is determined for a plurality of signal receivers. For example, data associated with the measured effect on signals received at each receiver of a plurality of receivers is averaged and stored.
  • storing the data includes storing the data in a format that can be used to generate a graph such as the graph of Figure 9.
  • the process of Figure 8 is repeated for different applied force intensities, different receivers, different force application locations, and/or different types of applied forces (e.g., different force application tip).
  • Data stored from the repeated execution of the steps of Figure 8 may be used to fill the data structure that may be used in step 708 of Figure 7.
  • Figure 9 includes graphs illustrating examples of a relationship between a normalized amplitude value of a measured disturbance and an applied force.
  • Graph 900 plots an applied force intensity (in grams of force) of a touch input vs. a measured amplitude of a disturbance caused by the applied force for a single receiver.
  • Graph 902 plots an applied force intensity of a touch input vs. a measured amplitude of a disturbance caused by the applied force for different receivers. The plots of the different receivers may be averaged and combined into a single plot.
  • graph 900 and/or graph 902 may be derived from data stored in the data structure that may be used in step 708 of Figure 7.
  • graph 900 and/or graph 902 may be generated using data stored in step 806 of Figure 8.
  • Graphs 900 and 902 show that there exists an increasing functional relationship between measured amplitude and applied force.
  • an associated force intensity identifier may be determined for a given amplitude value (e.g., such as in step 708 of Figure 7).
  • Figure 10 is a flowchart illustrating an embodiment of a process for providing a combined force. The process of Figure 10 may be implemented on touch detector 120 of Figure 1 and/or touch detector 202 of Figure 2.
  • a user touch input may be represented by a plurality of touch input locations (e.g., multi-touch input, touch input covering a relatively large area, etc.).
  • at least a portion of the process of Figure 7 is used to determine an associated force.
  • a force intensity identifier is determined for each input location in the plurality of touch input locations.
  • the determined forces are combined to determine a combined force.
  • the combined force represents a total amount of force applied on a touch input surface.
  • combining the forces includes adding a numerical representation of the forces together to determine the combined force.
  • a numerical representation of each determined force is weighted before being added together. For example, each numerical value of a determined force is weighted (e.g., multiplied by a scalar) based on an associated signal-to- noise ratio, an associated amplitude value, and/or an associated distance value between a receiver and a location of a touch input. In some embodiments, the weights of the forces being weighted must sum to the number of forces being combined.
  • the combined force is provided.
  • providing the combined force includes providing a force intensity identifier to an application such as an application of application system 122 of Figure 1.
  • provided combined force is used to provide a user interface interaction.
  • the determined forces for each touch input location point of a plurality of touch input location points are provided.
  • Figure 11 is a flowchart illustrating an embodiment of a process for providing a user interface interaction. The process of Figure 11 may be implemented on touch detector 120 of Figure 1 and/or touch detector 202 of Figure 2.
  • one or more indicators associated with a location and a force intensity of a user input are received.
  • the indicator(s) include data provided in step 710 of Figure 7 and/or step 1006 of Figure 10.
  • indicators associated with a sequence of locations and associated force intensities are received.
  • a user interface object associated with the location is determined.
  • the user input is a touch screen user interface input and the user interface element desired to be indicated by the user input is determined.
  • the user input is detected at a location where an icon has been displayed and it is determined that a user has desired to select the user icon by providing a touch input at a location of the icon.
  • the user interface object includes an object displayed on a touchscreen.
  • the user interface object is not an object already displayed on a screen. For example, a hidden keyboard user interface object appears when a user touches a specific area of a touch input screen.
  • a user interface interaction based at least in part on the user interface object and the force intensity is provided.
  • a user may indicate a desired user interface action by varying the amount of force applied on a touch input surface and the user interaction indicated by the received data in 1102 is provided. Examples of the possible user interface interactions are described in the following paragraphs.
  • Figure 12 is a diagram showing an example user interface interaction using force information to drag and drop an item into a file system folder.
  • a user may drag a desired item (e.g., a file, a folder, a reference, a link, an object, etc.) by touching the desired item with a relatively "light" force applied to a pointer (e.g., finger, stylus, etc.) and dragging the pointer.
  • a pointer e.g., finger, stylus, etc.
  • a user may desire to drag and drop the desired item to a folder to move or copy the item into the folder.
  • the user wants to drag and drop the desired item into a subfolder of the folder a user typically has to open the folder to reveal the desired subfolder before dragging and dropping the desired item.
  • a user may drag the desired item by touching the desired item with a relatively "light" force applied to a pointer (e.g., finger, stylus, etc.) and dragging the pointer to the displayed folder and applying a force intensity above a threshold level to descend into the subfolders of the displayed folder and releasing the pointer once a desired subfolder is found.
  • a pointer e.g., finger, stylus, etc.
  • a file may be moved by a "light" touch input to an icon representing the file and dragging the touch input to a displayed folder and applying greater force intensity to the touch input to descend into the contents of the displayed folder until a desired destination subfolder is displayed.
  • a file system hierarchy may be explored.
  • a touch input force intensity greater than a first threshold level indicates a command to navigate into a lower file system hierarchy and a touch input force less than a second threshold level (in some cases, the second threshold level may be the same as the first threshold level) indicates a command to navigate to a higher file system hierarchy.
  • the threshold levels may be preconfigured, dynamically determined, and/or may be configurable.
  • Figure 13 is a diagram showing an example user interface interaction using force information to provide a context menu.
  • traditional touch input device button e.g., mouse button
  • traditional touch input device button e.g., mouse button
  • a "left button click" input may be performed by a touch input with a force within a first intensity range
  • a "right button click” input may be performed by a touch input with a force within a second intensity range.
  • a "middle button click” input may be performed by a touch input with a force within a third intensity range.
  • a user may select an area (e.g., spreadsheet cells) or text by performing a touch and drag operation with a force intensity below a predetermined threshold.
  • Diagram 1300 shows text selected using a touch and drag operation and a context menu displayed when the force of the touch input was increased above a predetermined/dynamic/configurable threshold level.
  • Figure 14 and Figure 15 are diagrams showing examples of user interface interactions using force information to navigate a menu.
  • a user may navigate a menu by touching and dragging a touch input to desired menu items.
  • a user selects a menu item by increasing the force of the touch input above a threshold level and a user cancels the menu by releasing the touch input without ever increasing the force of the touch input above the threshold level.
  • a user can navigate a cascading menu by touching and dragging a touch input to desired cascading menu items.
  • a user selects a cascading menu item by increasing the force of the touch input above a threshold level and a user cancels the cascading menu by releasing touch input without ever increasing the force of the touch input above the threshold level.
  • the threshold levels may be preconfigured, dynamically determined, and/or configurable.
  • Figure 16 is a diagram showing an example user interface interaction using force information to interact with a virtual keyboard.
  • the virtual keyboard includes a keyboard that is displayed on a screen or projected on a surface.
  • a touch input key of a virtual keyboard is only registered as a key press if the force of the touch input is above a threshold level or within a first intensity range. For example, "lightly" resting fingers on a virtual keyboard will not register key presses on the virtual keyboard and a touch input will only be registered a key press when a greater force intensity is provided on the key of the virtual keyboard. This may reduce spurious key press events (e.g., often generated simply due to a finger lightly brushing or contacting the surface).
  • alternate key functionality may be indicated based on a force of touch input. For example, if a force of a touch input on a key is within a first range, a lower case or normal character of the key is indicated and if the force of the touch input is within a second range (e.g., greater than the first range), then a shifted/capitalized character of the key is indicated.
  • the threshold levels may be preconfigured, dynamically determined, and/or configurable.
  • a touch input gesture and a force associated with the gesture indicates that a virtual keyboard should be displayed and/or not displayed.
  • a keyboard is displayed.
  • a displayed virtual keyboard is oriented and/or located on a screen based at least in part on one or more touch inputs received.
  • a virtual keyboard is oriented and placed on a touch input display surface such that when fingers of a user are rested on the surface, the keys of the home row of the virtual keyboard are placed under the location and orientation of the placed fingers of the user to place the virtual keyboard in standard touch typing position with respect to the user's fingers.
  • the keyboard may be split to match the orientation of fingers of the user's two hands.
  • Diagram 1600 shows a virtual keyboard that has been displayed for a user that has placed fingers of the user's left hand higher and angled out as compared to fingers of the user's right hand that has been placed lower in the opposite angle.
  • a touch input to a key of the virtual keyboard of diagram 1600 is only registered as a keyboard key press if the force of the touch input is above a threshold.
  • the threshold levels may be preconfigured, dynamically determined, and/or configurable.
  • Figure 17 and Figure 18 are diagrams showing example user interface interactions using force information to zoom and select user interface objects.
  • force information is used to aid in navigating a dense array of objects on the screen (such as icons, keys, or several hyperlinks close by one another in a body of text).
  • a region around the touch contact point is shown magnified on a display, indicating in greater detail what object on the screen is underneath the contact point that is being used to provide touch input.
  • a predetermined threshold level e.g., configurable
  • a physical feedback e.g., haptic feedback
  • visual feedback e.g., audio feedback
  • audio feedback e.g., audio feedback
  • force information of touch input is used to distinguish between different gestures that otherwise might be identical or very similar.
  • a swipe touchscreen gesture of a first force intensity within a first threshold range may be interpreted as a scrolling/panning indication and a swipe touchscreen gesture of a second force intensity within a second threshold range may be interpreted as a "change to the next window/tab" indication.
  • Figure 19 is a graph illustrating an example of a relationship between detected touch input force and direction of change in audio volume.
  • Graph 1900 shows that when a touch input is within a first intensity range, volume is not changed, when a touch input is within a second intensity range, volume decrease functionality is indicated, and when a touch input is within a third intensity range, volume increase functionality is indicated.
  • force information of a touch input is used to control audio volume level of a device.
  • volume is increased if a force of a touch input is above a threshold value (e.g., predetermined, dynamically determined, and/or configurable) and the volume is decreased if the force is below the threshold value.
  • a threshold value e.g., predetermined, dynamically determined, and/or configurable
  • the touch input must be received in a specified area (e.g., displayed volume adjustment bar or a designated area of a device for adjusting volume) to control the volume.
  • the rate of change of the volume is proportional to an amount of force applied in a touch input.
  • an audio output destination is selected based at least in part on a force of a touch input. For example, the audio is outputted to an earpiece of a device with no touch input, and as a touch input is provided with an increasing force that meets a threshold level, a speakerphone function engages at a volume proportional to a detected force.
  • Figure 20 is a diagram showing an example user interface interaction using force information to interact with a slider bar.
  • a slider bar may be used to indicate an intensity level or a time location (e.g., video position during playback).
  • a user when navigating through a video sequence, a user wants the slider to move quickly to a particular time index/portion of the sequence, but then move with greater precision to focus on a particular scene or even a single frame of video.
  • Diagram 2000 shows a slider bar that can be moved by touching and dragging on the slider bar with a touch input.
  • a speed or precision of slider bar movement using touch input dragging may be proportional to the force intensity level of the touch input.
  • a slider control moves with detailed/fine precision when "light” pressure is applied but moves with coarse/faster precision when “harder” pressure is applied.
  • the slider bar may be moved with greater (e.g., fine or less granular) precision when a touch input force intensity within a first intensity range is applied and moved with less (e.g., coarse or more granular) precision when a touch input force intensity within a second intensity range is applied.
  • the threshold levels may be preconfigured, dynamically determined, and/or configurable.
  • a velocity at which an object such as a finger or stylus contacts a touch input surface is used to control a user interface.
  • video games, virtual musical instruments (drums and pianos are two common examples), and other applications may utilize velocity information to provide desired functionality.
  • measurement of contact velocity may be achieved by measuring the rate of change of the force. For example, if the touch force changes at a given point from 0 to 0.5 pounds in 20 milliseconds, it can be inferred that the finger or other object impacted the touch input screen at high velocity. On the other hand, a change in force from 0 to 0.1 pounds in 100 milliseconds could be construed as a relatively low velocity. Both the absolute measure of pressure and the rate-of-change of pressure may be useful measures of information in user interface design.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Human Computer Interaction (AREA)
  • General Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • General Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • Position Input By Displaying (AREA)
  • User Interface Of Digital Computer (AREA)
  • Electronic Switches (AREA)

Abstract

Detecting a touch input force is disclosed. A signal to be used to propagate a propagating signal through a propagating medium with a surface is sent. The propagating signal that has been disturbed by a touch input with an amount of force on the surface is received. The received signal is processed to determine an identifier associated with the amount of force.

Description

DETECTING TOUCH INPUT FORCE
CROSS REFERENCE TO OTHER APPLICATIONS
[0001] This application claims priority to U.S. Provisional Patent Application No.
61/561,697 entitled TOUCH SCREEN SYSTEM UTILIZING ADDITIONAL AXIS
INFORMATION filed November 18, 2011 which is incorporated herein by reference for all purposes.
[0002] This application claims priority to U.S. Provisional Patent Application No.
61/561,660 entitled TOUCH SCREEN SYSTEM UTILIZING ADDITIONAL AXIS
INFORMATION FOR SELECTED APPLICATIONS filed November 18, 2011 which is incorporated herein by reference for all purposes.
[0003] This application claims priority to U.S. Provisional Patent Application No.
61/673,102 entitled UTILIZING TOUCH PRESSURE INFORMATION IN GRAPHICAL USER INTERFACES filed July 18, 2012 which is incorporated herein by reference for all purposes.
[0004] This application is a continuation in part of co-pending U.S. Patent Application No.
13/451,288 entitled METHOD AND APPARATUS FOR ACTIVE ULTRASONIC TOUCH DEVICES filed April 19, 2012, which is incorporated herein by reference for all purposes, which claims priority to U.S. Provisional Patent Application No. 61/479,331 entitled METHOD AND APPARATUS FOR ACTIVE ULTRASONIC TOUCH DEVICES filed April 26, 2011 which is incorporated herein by reference for all purposes.
BACKGROUND OF THE INVENTION
[0005] Various technologies have been used to detect a touch input on a display area. The most popular technologies today include capacitive and resistive touch detection technology. Using resistive touch technology, often a glass panel is coated with multiple conductive layers that register touches when physical pressure is applied to the layers to force the layers to make physical contact. Using capacitive touch technology, often a glass panel is coated with material that can hold an electrical charge sensitive to a human finger. By detecting the change in the electrical charge due to a touch, a touch location can be detected. However, with resistive and capacitive touch detection technologies, the glass screen is required to be coated with a material that reduces the clarity of the glass screen. Additionally, because the entire glass screen is required to be coated with a material, manufacturing and component costs can become prohibitively expensive as larger screens are desired.
[0006] Another type of touch detection technology includes surface acoustic wave technology. One example includes the Elo Touch Systems Acoustic Pulse Recognition, commonly called APR, manufactured by Elo Touch Systems of 301 Constitution Drive, Menlo Park, CA 94025. The APR system includes transducers attached to the edges of a touchscreen glass that pick up the sound emitted on the glass due to a touch. However, the surface glass may pick up other external sounds and vibrations that reduce the accuracy and effectiveness of the APR system to efficiently detect a touch input. Another example includes the Surface Acoustic Wave-based technology, commonly called SAW, such as the Elo IntelliTouch Plus(TM) of Elo Touch Systems. The SAW technology sends ultrasonic waves in a guided pattern using reflectors on the touch screen to detect a touch. However, sending the ultrasonic waves in the guided pattern increases costs and may be difficult to achieve. Detecting additional types of inputs, such as multi-touch inputs, may not be possible or may be difficult using SAW or APR technology.
[0007] Additionally, current touch detection technology cannot reliably, accurately, and efficiently detect pressure or force of a touch input. Although prior attempts have been made to detect pressure of touch input by measuring the relative size of a touch input (e.g., as a finger presses harder on a screen, area of the finger contacting the screen proportionally increases), these attempts produce unreliable results when a hard stylus or different sized fingers are used. Therefore there exists a need for a better way to detect an input on a surface. Once force or pressure of a touch input can be reliably detected, user interface interaction utilizing force or pressure may be provided.
BRIEF DESCRIPTION OF THE DRAWINGS
[0008] Various embodiments of the invention are disclosed in the following detailed description and the accompanying drawings.
[0009] Figure 1 is a block diagram illustrating an embodiment of a system for detecting a surface disturbance.
[0010] Figure 2 is a block diagram illustrating an embodiment of a system for detecting a touch input.
[0011] Figure 3 is a flow chart illustrating an embodiment of a process for calibrating and validating touch detection. [0012] Figure 4 is a flow chart illustrating an embodiment of a process for detecting a user touch input.
[0013] Figure 5 is a flow chart illustrating an embodiment of a process for determining a location associated with a disturbance on a surface.
[0014] Figure 6 is a flow chart illustrating an embodiment of a process for determining a location associated with a disturbance.
[0015] Figure 7 is a flowchart illustrating an embodiment of a process of determining a force associated with a user input.
[0016] Figure 8 is a flowchart illustrating an embodiment of a process for determining an entry of a data structure used to determine a force intensity identifier.
[0017] Figure 9 includes graphs illustrating examples of a relationship between a normalized amplitude value of a measured disturbance and an applied force.
[0018] Figure 10 is a flowchart illustrating an embodiment of a process for providing a combined force.
[0019] Figure 11 is a flowchart illustrating an embodiment of a process for providing a user interface interaction.
[0020] Figure 12 is a diagram showing an example user interface interaction using force information to drag and drop an item into a file system folder.
[0021] Figure 13 is a diagram showing an example user interface interaction using force information to provide a context menu.
[0022] Figure 14 and Figure 15 are diagrams showing examples of user interface interactions using force information to navigate a menu.
[0023] Figure 16 is a diagram showing an example user interface interaction using force information to interact with a virtual keyboard.
[0024] Figure 17 and Figure 18 are diagrams showing example user interface interactions using force information to zoom and select user interface objects. [0025] Figure 19 is a graph illustrating an example of a relationship between detected touch input force and direction of change in audio volume.
[0026] Figure 20 is a diagram showing an example user interface interaction using force information to interact with a slider bar.
DETAILED DESCRIPTION
[0027] The invention can be implemented in numerous ways, including as a process; an apparatus; a system; a composition of matter; a computer program product embodied on a computer readable storage medium; and/or a processor, such as a processor configured to execute instructions stored on and/or provided by a memory coupled to the processor. In this specification, these implementations, or any other form that the invention may take, may be referred to as techniques. In general, the order of the steps of disclosed processes may be altered within the scope of the invention. Unless stated otherwise, a component such as a processor or a memory described as being configured to perform a task may be implemented as a general component that is temporarily configured to perform the task at a given time or a specific component that is manufactured to perform the task. As used herein, the term 'processor' refers to one or more devices, circuits, and/or processing cores configured to process data, such as computer program instructions.
[0028] A detailed description of one or more embodiments of the invention is provided below along with accompanying figures that illustrate the principles of the invention. The invention is described in connection with such embodiments, but the invention is not limited to any embodiment. The scope of the invention is limited only by the claims and the invention
encompasses numerous alternatives, modifications and equivalents. Numerous specific details are set forth in the following description in order to provide a thorough understanding of the invention. These details are provided for the purpose of example and the invention may be practiced according to the claims without some or all of these specific details. For the purpose of clarity, technical material that is known in the technical fields related to the invention has not been described in detail so that the invention is not unnecessarily obscured.
[0029] Detecting a force of a touch input is disclosed. In some embodiments, an acoustic transducer transmits an acoustic wave through a medium of a touch input surface. The acoustic wave may be scattered by the touch input producing a scattered acoustic wave. An acoustic detector that detects the scattered acoustic wave and the acoustic detector outputs a signal indicating variation of the acoustic wave that is indicative of an amount of force associated with the touch input. In some embodiments, the force of a touch input is associated with the amount of deflection or movement of a touch surface medium caused by a touch input. For example, as a finger or stylus touches and pushes a touch input surface harder, the amount of force detected gets functionally larger as well. The pressure of a touch input is the force of touch input per unit area of the touch input. For example, the total force of a touch input divided by the area of contact of the touch input equals the pressure of the touch input. Although force of a touch input is utilized in the
specification, pressure of a touch input may be used as well. In some cases, when a user pushes harder on a surface such as a touch screen display with a fingertip, the pressure of the touch input may stay substantially constant because the size of the fingertip in contact with the surface becomes larger due to the softness of the fingertip. In order to detect that the user is pushing harder on the surface, the total force of the touch input may be used instead of the pressure of the touch input. In some embodiments, a force of a touch input is used to provide user interface interaction.
[0030] In some embodiments, a user touch input on the glass surface of a display screen is detected. In some embodiments, a signal such as an acoustic or ultrasonic signal is propagated freely through a propagating medium with a surface using a transmitter coupled to the medium. When the surface is touched, the propagated signal is disturbed (e.g., the touch causes an interference with the propagated signal). In some embodiments, the disturbed signal is received at a sensor coupled to the propagating medium. By processing the received signal and comparing it against an expected signal without the disturbance, a location on the surface associated with the touch input is at least in part determined. For example, the disturbed signal is received at a plurality of sensors and a relative time difference between when the disturbed signal was received at different sensors is used to determine the location on the surface. In various embodiments, the touch includes a physical contact to a surface using a human finger, pen, pointer, stylus, and/or any other body parts or objects that can be used to contact or disturb the surface. In some embodiments, the touch includes an input gesture and/or a multi-touch input.
[0031] In some embodiments, the disturbed signal is used to determine one or more of the following associated with a touch input: a gesture, a coordinate position, a time, a time frame, a direction, a velocity, a force magnitude, a proximity magnitude, a pressure, a size, and other measurable or derived parameters. In some embodiments, by detecting disturbances of a freely propagated signal, touch input detection technology can be applied to larger surface regions with less or no additional cost due to a larger surface region as compared to certain previous touch detection technologies. Additionally, the optical transparency of a touch screen may not have to be affected as compared to resistive and capacitive touch technologies. Merely by way of example, the touch detection described herein can be applied to a variety of objects such as a kiosk, an ATM, a computing device, an entertainment device, a digital signage apparatus, a cell phone, a tablet computer, a point of sale terminal, a food and restaurant apparatus, a gaming device, a casino game and application, a piece of furniture, a vehicle, an industrial application, a financial application, a medical device, an appliance, and any other objects or devices having surfaces.
[0032] Figure 1 is a block diagram illustrating an embodiment of a system for detecting a surface disturbance. In some embodiments, the system shown in Figure 1 is included in a kiosk, an ATM, a computing device, an entertainment device, a digital signage apparatus, a cell phone, a tablet computer, a point of sale terminal, a food and restaurant apparatus, a gaming device, a casino game and application, a piece of furniture, a vehicle, an industrial application, a financial application, a medical device, an appliance, and any other objects or devices having surfaces.
Propagating signal medium 102 is coupled to transmitters 104, 106, 108, and 110 and sensors 112, 114, 116, and 118. In various embodiments, the propagating medium includes one or more of the following: panel, table, glass, screen, door, floor, whiteboard, glass, plastic, wood, steel, metal, semiconductor, insulator, conductor, and any medium that is able to propagate an acoustic or ultrasonic signal. For example, medium 102 is glass of a display screen. A first surface of medium 102 includes a surface area where a user may touch to provide a selection input and a substantially opposite surface of medium 102 is coupled to the transmitters and sensors shown in Figure 1. In various embodiments, a surface of medium 102 is substantially flat, curved, or combinations thereof and may be configured in a variety of shapes such as rectangular, square, oval, circular, trapezoidal, annular, or any combination of these, and the like.
[0033] Examples of transmitters 104, 106, 108, and 110 include piezoelectric transducers, electromagnetic transducers, transmitters, sensors and/or any other transmitters and transducers capable of propagating a signal through medium 102. Examples of sensors 112, 114, 116, and 118 include piezoelectric transducers, electromagnetic transducers, transmitters and/or any other sensors and transducers capable of detecting a signal on medium 102. In some embodiments, the transmitters and sensors shown in Figure 1 are coupled to medium 102 in a manner that allows a user input to be detected in a predetermined region of medium 102. Although four transmitters and four sensors are shown, any number of transmitters and any number of sensors may be used in other embodiments. For example, two transmitters and three sensors may be used. In some embodiments, a single transducer acts as both a transmitter and a sensor. For example, transmitter 104 and sensor 112 represent a single piezoelectric transducer. In the example shown, transmitter 104 may propagate a signal through medium 102. Sensors 112, 114, 116, and 118 receive the propagated signal. In another embodiment, the transmitters/sensors in Figure 1 are attached to a flexible cable coupled to medium 102 via an encapsulant and/or glue material and/or fasteners.
[0034] Touch detector 120 is connected to the transmitters and sensors shown in Figure 1.
In some embodiments, detector 120 includes one or more of the following: an integrated circuit chip, a printed circuit board, a processor, and other electrical components and connectors. Detector 120 determines and sends a signal to be propagated by transmitters 104, 106, 108, and 110.
Detector 120 also receives the signal detected by sensors 112, 114, 116, and 118. The received signals are processed by detector 120 to determine whether a disturbance associated with a user input has been detected at a location on a surface of medium 102 associated with the disturbance. Detector 120 is in communication with application system 122. Application system 122 uses information provided by detector 120. For example, application system 122 receives from detector 120 a coordinate associated with a user touch input that is used by application system 122 to control a software application of application system 122. In some embodiments, application system 122 includes a processor and/or memory/storage. In other embodiments, detector 120 and application system 122 are at least in part included/processed in a single processor. An example of data provided by detector 120 to application system 122 includes one or more of the following associated with a user indication: a location coordinate of a surface of medium 102, a gesture, simultaneous user indications (e.g., multi-touch input), a time, a status, a direction, a velocity, a force magnitude, a proximity magnitude, a pressure, a size, and other measurable or derived information.
[0035] In some embodiments, a touch input is received at location 130 on a surface of medium 102. For example, a user touches the surface of medium 102 at location 130. In some embodiments, one or more of transmitters 104, 106, 108, and 110 transmit one or more active signals that are propagated through medium 102. The touch input at location 130 disturbs (e.g., scatters) the propagated signal(s) and the disturbed signals are received at sensors 112, 114, 116, and 118. By measuring the disturbance(s) of the propagated signal(s), the location and/or a force associated with the touch input may be determined.
[0036] Figure 2 is a block diagram illustrating an embodiment of a system for detecting a touch input. In some embodiments, touch detector 202 is included in touch detector 120 of Figure 1. In some embodiments, the system of Figure 2 is integrated in an integrated circuit chip. Touch detector 202 includes system clock 204 that provides a synchronous system time source to one or more other components of detector 202. Controller 210 controls data flow and/or commands between microprocessor 206, interface 208, DSP engine 220, and signal generator 212. In some embodiments, microprocessor 206 processes instructions and/or calculations that can be used to program software/firmware and/or process data of detector 202. In some embodiments, a memory is coupled to microprocessor 206 and is configured to provide microprocessor 206 with
instructions. Signal generator 212 generates a signal to be used to propagate a signal such as a signal propagated by transmitter 104 of Figure 1. For example, signal generator 212 generates a pseudorandom binary sequence signal. Driver 214 receives the signal from generator 212 and drives one or more transmitters, such as transmitters 104, 106, 108, and 110 of Figure 1, to propagate a signal through a medium.
[0037] A signal detected from a sensor such as sensor 112 of Figure 1 is received by detector 202 and signal conditioner 216 conditions (e.g., filters) the received analog signal for further processing. For example, signal conditioner 216 receives the signal outputted by driver 214 and performs echo cancellation of the signal received by signal conditioner 216. The conditioned signal is converted to a digital signal by analog-to-digital converter 218. The converted signal is processed by digital signal processor engine 220. For example, DSP engine 220 correlates the converted signal against a reference signal. The result of the correlation may be used by
microprocessor 206 to determine a location associated with a user touch input. In some
embodiments, the DSP engine determines an amplitude change associated with the converted signal and a reference signal. The amplitude change may be used by microprocessor 206 to determine a force associated with a user touch input. Interface 208 provides an interface for microprocessor 206 and controller 210 that allows an external component to access and/or control detector 202. For example, interface 208 allows detector 202 to communicate with application system 122 of Figure 1 and provides the application system with location information associated with a user touch input.
[0038] Figure 3 is a flow chart illustrating an embodiment of a process for calibrating and validating touch detection. In some embodiments, the process of Figure 3 is used at least in part to calibrate and validate the system of Figure 1 and/or the system of Figure 2. At 302, locations of signal transmitters and sensors with respect to a surface are determined. For example, locations of transmitters and sensors shown in Figure 1 are determined with respect to their location on a surface of medium 102. In some embodiments, determining the locations includes receiving location information. In various embodiments, one or more of the locations may be fixed and/or variable.
[0039] At 304, signal transmitters and sensors are calibrated. In some embodiments, calibrating the transmitter includes calibrating a characteristic of a signal driver and/or transmitter (e.g., strength). In some embodiments, calibrating the sensor includes calibrating a characteristic of a sensor (e.g., sensitivity). In some embodiments, the calibration of 304 is performed to optimize the coverage and improve signal-to-noise transmission/detection of a signal (e.g., acoustic or ultrasonic) to be propagated through a medium and/or a disturbance to be detected. For example, one or more components of the system of Figure 1 and/or the system of Figure 2 are tuned to meet a signal-to-noise requirement. In some embodiments, the calibration of 304 depends on the size and type of a transmission/propagation medium and geometric configuration of the transmitters/sensors. In some embodiments, the calibration of step 304 includes detecting a failure or aging of a transmitter or sensor. In some embodiments, the calibration of step 304 includes cycling the transmitter and/or receiver. For example, to increase the stability and reliability of a piezoelectric transmitter and/or receiver, a burn-in cycle is performed using a burn-in signal. In some
embodiments, the step of 304 includes configuring at least one sensing device within a vicinity of a predetermined spatial region to capture an indication associated with a disturbance using the sensing device. The disturbance is caused in a selected portion of the input signal corresponding to a selection portion of the predetermined spatial region.
[0040] At 306, surface disturbance detection is calibrated. In some embodiments, a test signal is propagated through a medium such as medium 102 of Figure 1 to determine an expected sensed signal when no disturbance has been applied. In some embodiments, a test signal is propagated through a medium to determine a sensed signal when one or more predetermined disturbances (e.g., predetermined touch) are applied at a predetermined location. Using the sensed signal, one or more components may be adjusted to calibrate the disturbance detection.
[0041] At 308, a validation of a touch detection system is performed. For example, the system of Figure 1 and/or Figure 2 is testing using predetermined disturbance patterns to determine detection accuracy, detection resolution, multi-touch detection, and/or response time. If the validation fails, the process of Figure 3 may be at least in part repeated and/or one or more components may be adjusted before performing another validation.
[0042] Figure 4 is a flow chart illustrating an embodiment of a process for detecting a user touch input. In some embodiments, the process of Figure 4 is at least in part implemented on touch detector 120 of Figure 1 and/or touch detector 202 of Figure 2. At 402, a signal that can be used to propagate an active signal through a surface region is sent. In some embodiments, sending the signal includes driving (e.g., using driver 214 of Figure 2) a transmitter such as a transducer (e.g., transmitter 104 of Figure 1) to propagate an active signal (e.g., acoustic or ultrasonic) through a propagating medium with the surface region. In some embodiments, the signal includes a sequence selected to optimize autocorrelation (e.g., resulting in narrow/short peak) of the signal. For example, the signal includes a Zadoff-Chu sequence. In some embodiments, the signal includes a pseudorandom binary sequence with or without modulation. In some embodiments, the propagated signal is an acoustic signal. In some embodiments, the propagated signal is an ultrasonic signal (e.g., outside the range of human hearing). For example, the propagated signal is a signal above 20 kHz (e.g., within the range between 80 kHz to 100 kHz). In other embodiments, the propagated signal may be within the range of human hearing. In some embodiments, by using the active signal, a user input on or near the surface region can be detected by detecting disturbances in the active signal when it is received by a sensor on the propagating medium. By using an active signal rather than merely listening passively for a user touch indication on the surface, other vibrations and disturbances that are not likely associated with a user touch indication can be more easily discerned/filtered out. In some embodiments, the active signal is used in addition to receiving a passive signal from a user input to determine the user input.
[0043] At 404, the active signal that has been disturbed by a disturbance of the surface region is received. The disturbance may be associated with a user touch indication. In some embodiments, the disturbance causes the active signal that is propagating through a medium to be attenuated and/or delayed. In some embodiments, the disturbance in a selected portion of the active signal corresponds to a location on the surface that has been indicated (e.g., touched) by a user.
[0044] At 406, the received signal is processed to at least in part determine a location associated with the disturbance. In some embodiments, determining the location includes extracting a desired signal from the received signal at least in part by removing or reducing undesired components of the received signal such as disturbances caused by extraneous noise and vibrations not useful in detecting a touch input. In some embodiments, determining the location includes comparing the received signal to a reference signal that has not been affected by the disturbance. The result of the comparison may be used with a result of other comparisons performed using the reference signal and other signal(s) received at a plurality of sensors. The location, in some embodiments, is a location (e.g., a location coordinate) on the surface region where a user has provided a touch input. In addition to determining the location, one or more of the following information associated with the disturbance may be determined at 406: a gesture, simultaneous user indications (e.g., multi-touch input), a time, a status, a direction, a velocity, a force magnitude, a proximity magnitude, a pressure, a size, and other measurable or derived information. In some embodiments, the location is not determined at 406 if a location cannot be determined using the received signal and/or the disturbance is determined to be not associated with a user input.
Information determined at 406 may be provided and/or outputted. [0045] Although Figure 4 shows receiving and processing an active signal that has been disturbed, in some embodiments, a received signal has not been disturbed by a touch input and the received signal is processed to determine that a touch input has not been detected. An indication that a touch input has not been detected may be provided/outputted.
[0046] Figure 5 is a flow chart illustrating an embodiment of a process for determining a location associated with a disturbance on a surface. In some embodiments, the process of Figure 5 is included in 406 of Figure 4. The process of Figure 5 may be implemented in touch detector 120 of Figure 1 and/or touch detector 202 of Figure 2. At 502, a received signal is conditioned. In some embodiments, the received signal is a signal including a pseudorandom binary sequence that has been freely propagated through a medium with a surface that can be used to receive a user input. For example, the received signal is the signal that has been received at 404 of Figure 4. In some embodiments, conditioning the signal includes filtering or otherwise modifying the received signal to improve signal quality (e.g., signal-to-noise ratio) for detection of a pseudorandom binary sequence included in the received signal and/or user touch input. In some embodiments, conditioning the received signal includes filtering out from the signal extraneous noise and/or vibrations not likely associated with a user touch indication.
[0047] At 504, an analog to digital signal conversion is performed on the signal that has been conditioned at 502. In various embodiments, any number of standard analog to digital signal converters may be used. The resulting digital signal is used to perform a first correlation at 506. In some embodiments, performing the first correlation includes correlating the converted signal with a reference signal. Performing the correlation includes cross-correlating or determining a convolution (e.g., interferometry) of the converted signal with a reference signal to measure the similarity of the two signals as a time-lag is applied to one of the signals. By performing the correlation, the location of a portion of the converted signal that most corresponds to the reference signal can be located. For example, a result of the correlation can be plotted as a graph of time within the received and converted signal (e.g., time-lag between the signals) vs. a measure of similarity. The associated time value of the largest value of the measure of similarity corresponds to the location where the two signals most correspond. By comparing this measured time value against a reference time value (e.g., at 306 of Figure 3) not associated with a touch indication disturbance, a time delay/offset or phase difference caused on the received signal due to a disturbance caused by a touch input can be determined. In some embodiments, by measuring the amplitude/intensity difference of the received signal at the determined time vs. a reference signal, a force associated with a touch indication may be determined. In some embodiments, the reference signal is determined based at least in part on the signal that was propagated through a medium (e.g., based on a source pseudorandom binary sequence signal that was propagated). In some embodiments, the reference signal is at least in part determined using information determined during calibration at 306 of Figure 3. The reference signal may be chosen so that calculations required to be performed during the correlation may be simplified. For example, the reference signal used in 506 is a simplified reference signal that can be used to efficiently correlate the reference signal over a relatively large time difference (e.g., lag-time) between the received and converted signal and the reference signal.
[0048] At 508, a second correlation is performed based on a result of the first correlation.
Performing the second correlation includes correlating (e.g., cross-correlation or convolution similar to step 506) the converted signal in 504 with a second reference signal. The second reference signal is a more complex/detailed (e.g., more computationally intensive) reference signal as compared to the first reference signal used in 506. In some embodiments, the second correlation is performed in 508 because using the second reference signal in 506 may be too computationally intensive for the time interval required to be correlated in 506. Performing the second correlation based on the result of the first correlation includes using one or more time values determined as a result of the first correlation. For example, using a result of the first correlation, a range of likely time values (e.g., time-lag) that most correlate between the received signal and the first reference signal is determined and the second correlation is performed using the second reference signal only across the determined range of time values to fine tune and determine the time value that most corresponds to where the second reference signal (and, by association, also the first reference signal) matched the received signal. In various embodiments, the first and second correlations have been used to determine a portion within the received signal that correspond to a disturbance caused by a touch input at a location on a surface of a propagating medium. In other embodiments, the second correlation is optional. For example, only a single correlation step is performed.
[0049] At 510, a result of the second correlation is used to at least in part determine a location associated with a disturbance. In some embodiments, determining the location includes comparing a determined time value where the signals of the second correlation are most correlated and comparing the determined time value with a reference time value (e.g., determined at 306 of Figure 3) not associated with a touch input disturbance, to determine a time delay/offset or phase difference caused on the received signal due to the disturbance (e.g., caused by a touch input). This time delay is associated with a signal received at a first sensor and other time delays due to the disturbance at other signals received at other sensors are used to calculate a location of the disturbance relative to the locations of the sensors. By using the location of the sensors relative to a surface of a medium that has propagated the received signal, a location on the surface where the disturbance originated may be determined.
[0050] Figure 6 is a flowchart illustrating an embodiment of a process for determining a location associated with a disturbance. In some embodiments, the process of Figure 6 is included in 510 of Figure 5. At 602, a plurality of results of correlations performed on a plurality of signals disturbed by a disturbance of a surface is received. For example, a result of the correlation performed at 508 of Figure 5 is received. In some embodiments, a signal is propagated using transmitter 104 and sensors 114, 116, and 118 each receives the propagated signal that has been disturbed by a touch input on or near a surface of medium 102 of Figure 1. The propagated signal may contain a predetermined signal and the predetermined signal is received at the various sensors. Each of the received signals is correlated with a reference signal to determine the results received at 602. In some embodiments, the received results are associated with a same signal content (e.g., same binary sequence) that has been freely propagated on a medium at the same time. In some embodiments, the received results are associated with different signal contents that have been disturbed by the same disturbance.
[0051] At 604, time differences associated with the plurality of results are used to determine a location associated with the disturbance. In some embodiments, each of the time differences is associated with a time when signals used in the correlation are most correlated. In some
embodiments, the time differences are associated with a determined time delay/offset or phase difference caused on the received signal due to the disturbance. This time delay may be calculated by comparing a time value determined using a correlation with a reference time value that is associated with a scenario where a touch input has not been specified. The result of the comparison may be used to calculate a location of the disturbance relative to the locations of sensors that received the plurality of signals. By using the location of the sensors relative to a surface of a medium that has propagated the received signal, a location on the surface where the disturbance originated may be determined.
[0052] Figure 7 is a flowchart illustrating an embodiment of a process of determining a force associated with a user input. The process of Figure 7 may be implemented on touch detector 120 of Figure 1 and/or touch detector 202 of Figure 2.
[0053] At 702, a location associated with a user input on a touch input surface is determined. In some embodiments, at least a portion of the process of Figure 4 is included in step 702. For example, the process of Figure 4 is used to determine a location associated with a user touch input. In another example, a location associated with a user input at location 130 on a surface of medium 102 of Figure 1 is determined.
[0054] At 704, one or more received signals are selected to be evaluated. In some embodiments, selecting the signal(s) to be evaluated include selecting one or more desired signals from a plurality of received signals used to detect the location associated with the user input. For example, one or more signals received in step 404 of Figure 4 are selected. In some embodiments, the selected signal(s) are selected based at least in part on a signal-to-noise ratio associated with signals. In some embodiments, one or more signals with the highest signal-to-noise ratio are selected. For example, when an active signal that is propagated through a touch input surface medium is disturbed/scattered by a touch input, the disturbed signal is detected/received at various detectors/sensors/receivers coupled to the medium. The received disturbed signals may be subject to other undesirable disturbances such as other minor vibration sources (e.g., due to external audio vibration, device movement, etc.) that also disturb the active signal. The effects of these
undesirable disturbances may be larger on received signals that were received further away from the location of the touch input.
[0055] In some embodiments, a variation (e.g., disturbance such as amplitude change) detected in an active signal received at a receiver/sensor may be greater at certain receivers (e.g., receivers located closest to the location of the touch input) as compared to other receivers. For example, in the example of Figure 1, touch input at location 130 disturbs an active signal sent by transmitter 104. The disturbed active signal is received at sensors/receivers 112, 114, 116, and 118. Because sensor/receiver 114 is located closest to touch input location 130, it has received a disturbed signal with the largest amplitude variation that is proportional to the force of the touch input. In some embodiments, the selected signals may have been selected at least in part by examining the amplitude of a detected disturbance. For example, one or more signals with the highest amplitude associated with a detected touch input disturbance are selected. In some embodiments, based at least in part on a location determined in 702, one or more signals received at one or more receivers located closest to the touch input location are selected. In some
embodiments, a plurality of active signals is used to detect a touch input location and/or touch input force intensity. One or more received signals to be used to determine a force intensity may be selected for each of the active signals. In some embodiments, one or more received signals to be used to determine the force intensity may be selected across the received signals of all the active signals. [0056] At 706, the one or more selected signals are normalized. In some embodiments, normalizing a selected signal includes adjusting (e.g., scaling) an amplitude of the selected signal based on a distance value associated with the selected signal. For example, although an
amount/intensity of force of a touch input may be detected by measuring an amplitude of a received active signal that has been disturbed by the force of the touch input, other factors such as the location of the touch input with respect to a receiver that has received the disturbed signal and/or location of the transmitter transmitting the active signal may also affect the amplitude of the received signal used to determine the intensity of the force. In some embodiments, a distance value/identifier associated with one or more of the following is used to determine a scaling factor used to scale a selected signal: a distance between a location of a touch input and a location of a receiver that has received the selected signal, a distance between a location of a touch input and a location of a transmitter that has transmitted an active signal that has been disturbed by a touch input and received as the selected signal, a distance between a location of a receiver that has received the selected signal and a location of a transmitter that has transmitted an active signal that has been disturbed by a touch input and received as the selected signal, and a combined distance of a first distance between a location of a touch input and a location of a receiver that has received the selected signal and a second distance between the location of the touch input and a location of a transmitter that has transmitted an active signal that has been disturbed by a touch input and received as the selected signal. In some embodiments, each of one or more selected signals is normalized by a different amount (e.g., different amplitude scaling factors).
[0057] At 708, a force intensity identifier associated with the one or more normalized signals is determined. The force intensity identifier may include a numerical value and/or other identifier identifying a force intensity. In some embodiments, if a plurality of normalized signals is used, an associated force may be determined for each normalized signal and the determined forces may be averaged and/or weighted-averaged to determine the amount of the force. For example, in the case of weighted averaging of the force values, each determined force value is weighted based on an associated signal-to-noise ratio, an associated amplitude value, and/or an associated distance value between a receiver of the normalized signal and the location of the touch input.
[0058] In some embodiments, the amount of force is determined using a measured amplitude associated with a disturbed portion of the normalized signal. For example, the normalized signal represents a received active signal that has been disturbed when a touch input was provided on a surface of a medium that was propagating the active signal. A reference signal may indicate a reference amplitude of a received active signal if the active signal was not disturbed by a touch input. In some embodiments, an amplitude value associated with an amplitude change to the normalized signal caused by a force intensity of a touch input is determined. For example, the amplitude value may be a measured amplitude of a disturbance detected in a normalized signal or a difference between a reference amplitude and the measured amplitude of the disturbance detected in the normalized signal. In some embodiments, the amplitude value is used to obtain an amount/intensity of a force.
[0059] In some embodiments, the use of the amplitude value includes using the amplitude value to look up in a data structure (e.g., table, database, chart, graph, lookup table, list, etc.) a corresponding associated force intensity. For example, the data structure includes entries associating a signal disturbance amplitude value and a corresponding force intensity identifier. The data structure may be predetermined/pre-computed. For example, for a given device, a controlled amount of force is applied and the disturbance effect on an active signal due to the controlled amount of force is measured to determine an entry for the data structure. The force intensity may be varied to determine other entries of the data structure. In some embodiments, the data structure is associated with a specific receiver that received the signal included in the normalized signal. For example, the data structure includes data that has been specifically determined for characteristics of a specific receiver (e.g., for sensor/receiver 114 of Figure 1). In some embodiments, the use of the amplitude value to look up a corresponding force intensity identifier stored in a data structure includes selecting a specific data structure and/or a specific portion of a data structure
corresponding to the normalized signal and/or a receiver that received the signal included in the normalized signal. In some embodiments, the data structure is associated with a plurality of receivers. For example, the data structure includes entries associated with averages of data determined for characteristics of each receiver in the plurality of receivers. In this example, the same data structure may be used for a plurality of normalized signals associated with various receivers.
[0060] In some embodiments, the use of the amplitude value includes using the amplitude value in a formula that can be used to simulate and/or calculate a corresponding force intensity. For example, the amplitude value is used as an input to a predetermined formula used to compute a corresponding force intensity. In some embodiments, the formula is associated with a specific receiver that received the signal of the normalized signal. For example, the formula includes one or more parameters (e.g., coefficients) that have been specifically determined for characteristics of a specific receiver (e.g., for sensor/receiver 114 of Figure 1). In some embodiments, the use of the amplitude value in a formula calculation includes selecting a specific formula corresponding to the normalized signal and/or a receiver that received the signal included in the normalized signal. In some embodiments, a single formula is associated with a plurality of receivers. For example, a formula includes averaged parameter values of parameter values that have been specifically determined for characteristics for each of the receivers in the plurality of receivers. In this example, the same formula may be used for a plurality of normalized signals associated with different receivers.
[0061] At 710, the determined force intensity identifier is provided. In some embodiments, providing the force intensity identifier includes providing the identifier (e.g., a numerical value, an identifier within a scale, etc.) to an application such as an application of application system 122 of Figure 1. In some embodiments, the provided force intensity identifier is provided with a corresponding touch input location identifier determined in step 406 of Figure 4. In some embodiments, the provided force intensity identifier is used to provide a user interface interaction.
[0062] Figure 8 is a flowchart illustrating an embodiment of a process for determining an entry of a data structure used to determine a force intensity identifier. In some embodiments, the process of Figure 8 is included in step 304 of Figure 3. In some embodiments, the process of Figure 8 is used at least in part to create the data structure that may be used in step 708 of Figure 7. In some embodiments, the process of Figure 8 is used at least in part to calibrate the system of Figure 1 and/or the system of Figure 2. In some embodiments, the process of Figure 8 is used at least in part to determine a data structure that can be included in one or more devices to be manufactured to determine a force intensity identifier/value corresponding to an amplitude value of a disturbance detected in the received active signal. For example, the data structure may be determined for a plurality of similar devices to be manufactured or the data structure may be determined for a specific device taking into account the manufacturing variation of the device.
[0063] At 802, a controlled amount of force is applied at a selected location on a touch input surface. In some embodiments, the force is provided on a location of a surface of medium 102 of Figure 1 where a touch input may be provided. In some embodiments, a tip of a pointer (e.g., stylus) is pressing at the surface with a controllable amount of force. For example, a controlled amount of force is applied on a touch input surface while an active signal is being propagated through a medium of the touch input surface. The amount of force applied in 802 may be one of a plurality of different amounts of force that will be applied on the touch input surface.
[0064] At 804, an effect of the applied force is measured using one or more receivers.
Examples of the receivers include sensors 112-118 of Figure 1 and transducer transmitters used as receivers (e.g., transmitters 104-110 of Figure 1). In some embodiments, measuring the effect includes measuring an amplitude associated with a disturbed portion of an active signal that has been disturbed when the force was applied in 802 and that has been received by the one or more receivers. The amplitude may be a directly measured amplitude value or a difference between a reference amplitude and a detected amplitude. In some embodiments, the signal received by the one or more receivers is normalized before the amplitude is measured. In some embodiments, normalizing a received signal includes adjusting (e.g., scaling) an amplitude of the signal based on a distance value associated with the selected signal.
[0065] A reference signal may indicate a reference amplitude of a received active signal that has not been disturbed by a touch input. In some embodiments, an amplitude value associated with an amplitude change caused by a disturbance of a touch input is determined. For example, the amplitude value may be a measured amplitude value of a disturbance detected in a normalized signal or a difference between a reference amplitude and the measured amplitude value of the disturbance detected in the normalized signal. In some embodiments, the amplitude value is used to obtain an identifier of a force intensity.
[0066] In some embodiments, a distance value associated with one or more of the following is used to determine a scaling factor used to scale a received signal before an effect of a disturbance is measured using the received signal: a distance between a location of a touch input and a location of a receiver that has received the selected signal, a distance between a location of the force input and a location of a transmitter that has transmitted an active signal that has been disturbed by the force input and received by the receiver, a distance between a location of the receiver and a location of a transmitter that has transmitted an active signal that has been disturbed by the force input and received by the receiver, and a combined distance of a first distance between a location of a force input and a location of the receiver and a second distance between the location of the force input and a location of a transmitter that has transmitted an active signal that has been disturbed by the force input and received by the receiver. In some embodiments, each of one or more signals received by different receivers is normalized by a different amount (e.g., different amplitude scaling factors).
[0067] At 806, data associated with the measured effect is stored. In some embodiments, storing the data includes storing an entry in a data structure such as the data structure that may be used in step 708 of Figure 7. For example, an entry that associates the amplitude value determined in 804 and an identifier associated with an amount of force applied in 802 is stored in the data structure. In some embodiments, storing the data includes indexing the data by an amplitude value determined in 804. For example, the stored data may be retrieved from the storage using the amplitude value. In some embodiments, the data structure is determined for a specific signal receiver. In some embodiments, a data structure is determined for a plurality of signal receivers. For example, data associated with the measured effect on signals received at each receiver of a plurality of receivers is averaged and stored. In some embodiments, storing the data includes storing the data in a format that can be used to generate a graph such as the graph of Figure 9.
[0068] In some embodiments, the process of Figure 8 is repeated for different applied force intensities, different receivers, different force application locations, and/or different types of applied forces (e.g., different force application tip). Data stored from the repeated execution of the steps of Figure 8 may be used to fill the data structure that may be used in step 708 of Figure 7.
[0069] Figure 9 includes graphs illustrating examples of a relationship between a normalized amplitude value of a measured disturbance and an applied force. Graph 900 plots an applied force intensity (in grams of force) of a touch input vs. a measured amplitude of a disturbance caused by the applied force for a single receiver. Graph 902 plots an applied force intensity of a touch input vs. a measured amplitude of a disturbance caused by the applied force for different receivers. The plots of the different receivers may be averaged and combined into a single plot. In some embodiments, graph 900 and/or graph 902 may be derived from data stored in the data structure that may be used in step 708 of Figure 7. In some embodiments, graph 900 and/or graph 902 may be generated using data stored in step 806 of Figure 8. Graphs 900 and 902 show that there exists an increasing functional relationship between measured amplitude and applied force. Using a predetermined graph, data structure, and/or formula that model this relationship, an associated force intensity identifier may be determined for a given amplitude value (e.g., such as in step 708 of Figure 7).
[0070] Figure 10 is a flowchart illustrating an embodiment of a process for providing a combined force. The process of Figure 10 may be implemented on touch detector 120 of Figure 1 and/or touch detector 202 of Figure 2.
[0071] At 1002, forces associated with each touch input location point of a plurality of touch input location points are determined. In some embodiments, a user touch input may be represented by a plurality of touch input locations (e.g., multi-touch input, touch input covering a relatively large area, etc.). In some embodiments, for each touch input location point, at least a portion of the process of Figure 7 is used to determine an associated force. For example, a force intensity identifier is determined for each input location in the plurality of touch input locations. [0072] At 1004, the determined forces are combined to determine a combined force. For example, the combined force represents a total amount of force applied on a touch input surface. In some embodiments, combining the forces includes adding a numerical representation of the forces together to determine the combined force. In some embodiments, a numerical representation of each determined force is weighted before being added together. For example, each numerical value of a determined force is weighted (e.g., multiplied by a scalar) based on an associated signal-to- noise ratio, an associated amplitude value, and/or an associated distance value between a receiver and a location of a touch input. In some embodiments, the weights of the forces being weighted must sum to the number of forces being combined.
[0073] At 1006, the combined force is provided. In some embodiments, providing the combined force includes providing a force intensity identifier to an application such as an application of application system 122 of Figure 1. In some embodiments, provided combined force is used to provide a user interface interaction. In an alternative embodiment, rather than providing the combine force, the determined forces for each touch input location point of a plurality of touch input location points are provided.
[0074] Figure 11 is a flowchart illustrating an embodiment of a process for providing a user interface interaction. The process of Figure 11 may be implemented on touch detector 120 of Figure 1 and/or touch detector 202 of Figure 2.
[0075] At 1102, one or more indicators associated with a location and a force intensity of a user input are received. In some embodiments, the indicator(s) include data provided in step 710 of Figure 7 and/or step 1006 of Figure 10. In some embodiments, indicators associated with a sequence of locations and associated force intensities are received.
[0076] At 1104, a user interface object associated with the location is determined. In some embodiments, the user input is a touch screen user interface input and the user interface element desired to be indicated by the user input is determined. For example, the user input is detected at a location where an icon has been displayed and it is determined that a user has desired to select the user icon by providing a touch input at a location of the icon. In some embodiments, the user interface object includes an object displayed on a touchscreen. In some embodiments, the user interface object is not an object already displayed on a screen. For example, a hidden keyboard user interface object appears when a user touches a specific area of a touch input screen. [0077] At 1106, a user interface interaction based at least in part on the user interface object and the force intensity is provided. For example, a user may indicate a desired user interface action by varying the amount of force applied on a touch input surface and the user interaction indicated by the received data in 1102 is provided. Examples of the possible user interface interactions are described in the following paragraphs.
[0078] Figure 12 is a diagram showing an example user interface interaction using force information to drag and drop an item into a file system folder. In some embodiments, a user may drag a desired item (e.g., a file, a folder, a reference, a link, an object, etc.) by touching the desired item with a relatively "light" force applied to a pointer (e.g., finger, stylus, etc.) and dragging the pointer. A user may desire to drag and drop the desired item to a folder to move or copy the item into the folder. However if the user wants to drag and drop the desired item into a subfolder of the folder, a user typically has to open the folder to reveal the desired subfolder before dragging and dropping the desired item. In some embodiments, in order to move or copy an item to a subfolder of a displayed folder, a user may drag the desired item by touching the desired item with a relatively "light" force applied to a pointer (e.g., finger, stylus, etc.) and dragging the pointer to the displayed folder and applying a force intensity above a threshold level to descend into the subfolders of the displayed folder and releasing the pointer once a desired subfolder is found. As shown in diagram 1200, a file may be moved by a "light" touch input to an icon representing the file and dragging the touch input to a displayed folder and applying greater force intensity to the touch input to descend into the contents of the displayed folder until a desired destination subfolder is displayed. In some embodiments, by varying the amount of pressure of a touch input, a file system hierarchy may be explored. In some embodiments, a touch input force intensity greater than a first threshold level indicates a command to navigate into a lower file system hierarchy and a touch input force less than a second threshold level (in some cases, the second threshold level may be the same as the first threshold level) indicates a command to navigate to a higher file system hierarchy. The threshold levels may be preconfigured, dynamically determined, and/or may be configurable.
[0079] Figure 13 is a diagram showing an example user interface interaction using force information to provide a context menu. In some embodiments, traditional touch input device button (e.g., mouse button) functionality may be mapped to one or more force intensity levels. For example, a "left button click" input may be performed by a touch input with a force within a first intensity range and a "right button click" input may be performed by a touch input with a force within a second intensity range. In some embodiments, a "middle button click" input may be performed by a touch input with a force within a third intensity range. In some embodiments, a user may select an area (e.g., spreadsheet cells) or text by performing a touch and drag operation with a force intensity below a predetermined threshold. Before the touch input is released, a user may indicate that a context menu is desired (e.g., "right button click") by increasing the force of the touch input above a predetermined/dynamic/configurable threshold level. Diagram 1300 shows text selected using a touch and drag operation and a context menu displayed when the force of the touch input was increased above a predetermined/dynamic/configurable threshold level.
[0080] Figure 14 and Figure 15 are diagrams showing examples of user interface interactions using force information to navigate a menu. As shown in diagram 1400, a user may navigate a menu by touching and dragging a touch input to desired menu items. A user selects a menu item by increasing the force of the touch input above a threshold level and a user cancels the menu by releasing the touch input without ever increasing the force of the touch input above the threshold level. As shown in diagram 1500, a user can navigate a cascading menu by touching and dragging a touch input to desired cascading menu items. A user selects a cascading menu item by increasing the force of the touch input above a threshold level and a user cancels the cascading menu by releasing touch input without ever increasing the force of the touch input above the threshold level. The threshold levels may be preconfigured, dynamically determined, and/or configurable.
[0081] Figure 16 is a diagram showing an example user interface interaction using force information to interact with a virtual keyboard. In some embodiments, the virtual keyboard includes a keyboard that is displayed on a screen or projected on a surface. In some embodiments, a touch input key of a virtual keyboard is only registered as a key press if the force of the touch input is above a threshold level or within a first intensity range. For example, "lightly" resting fingers on a virtual keyboard will not register key presses on the virtual keyboard and a touch input will only be registered a key press when a greater force intensity is provided on the key of the virtual keyboard. This may reduce spurious key press events (e.g., often generated simply due to a finger lightly brushing or contacting the surface). In some embodiments, alternate key functionality may be indicated based on a force of touch input. For example, if a force of a touch input on a key is within a first range, a lower case or normal character of the key is indicated and if the force of the touch input is within a second range (e.g., greater than the first range), then a shifted/capitalized character of the key is indicated. The threshold levels may be preconfigured, dynamically determined, and/or configurable. [0082] In some embodiments, a touch input gesture and a force associated with the gesture indicates that a virtual keyboard should be displayed and/or not displayed. For example, when a predetermined number of distinct touch inputs are detected simultaneously (e.g., 4 or 5 fingers of each hand resting on a touch input surface), a keyboard is displayed. In some embodiments, a displayed virtual keyboard is oriented and/or located on a screen based at least in part on one or more touch inputs received. For example, a virtual keyboard is oriented and placed on a touch input display surface such that when fingers of a user are rested on the surface, the keys of the home row of the virtual keyboard are placed under the location and orientation of the placed fingers of the user to place the virtual keyboard in standard touch typing position with respect to the user's fingers. The keyboard may be split to match the orientation of fingers of the user's two hands. Diagram 1600 shows a virtual keyboard that has been displayed for a user that has placed fingers of the user's left hand higher and angled out as compared to fingers of the user's right hand that has been placed lower in the opposite angle. In some embodiments, a touch input to a key of the virtual keyboard of diagram 1600 is only registered as a keyboard key press if the force of the touch input is above a threshold. The threshold levels may be preconfigured, dynamically determined, and/or configurable.
[0083] Figure 17 and Figure 18 are diagrams showing example user interface interactions using force information to zoom and select user interface objects. In some embodiments, force information is used to aid in navigating a dense array of objects on the screen (such as icons, keys, or several hyperlinks close by one another in a body of text). As shown in diagram 1700 and 1800, by "lightly" touching a touch input display surface, a region around the touch contact point is shown magnified on a display, indicating in greater detail what object on the screen is underneath the contact point that is being used to provide touch input. When the object to be selected (e.g., icon in diagram 1700 and keyboard key in diagram 1800) is underneath the contact point, increasing the force of the touch input to a level greater than a predetermined threshold level (e.g., configurable) selects the object. In some embodiments, when a user selects an object and/or completes a user interface action, a physical feedback (e.g., haptic feedback), visual feedback, and/or audio feedback may be provided. In some embodiments, when a force of a touch input reaches a threshold level, a physical feedback (e.g., haptic feedback), visual feedback, and/or audio feedback may be provided.
[0084] In some embodiments, force information of touch input is used to distinguish between different gestures that otherwise might be identical or very similar. For example, a swipe touchscreen gesture of a first force intensity within a first threshold range may be interpreted as a scrolling/panning indication and a swipe touchscreen gesture of a second force intensity within a second threshold range may be interpreted as a "change to the next window/tab" indication.
[0085] Figure 19 is a graph illustrating an example of a relationship between detected touch input force and direction of change in audio volume. Graph 1900 shows that when a touch input is within a first intensity range, volume is not changed, when a touch input is within a second intensity range, volume decrease functionality is indicated, and when a touch input is within a third intensity range, volume increase functionality is indicated. In some embodiments, force information of a touch input is used to control audio volume level of a device. In some embodiments, volume is increased if a force of a touch input is above a threshold value (e.g., predetermined, dynamically determined, and/or configurable) and the volume is decreased if the force is below the threshold value. In some embodiments, the touch input must be received in a specified area (e.g., displayed volume adjustment bar or a designated area of a device for adjusting volume) to control the volume. In some embodiments, the rate of change of the volume is proportional to an amount of force applied in a touch input. In some embodiments, an audio output destination is selected based at least in part on a force of a touch input. For example, the audio is outputted to an earpiece of a device with no touch input, and as a touch input is provided with an increasing force that meets a threshold level, a speakerphone function engages at a volume proportional to a detected force.
[0086] Figure 20 is a diagram showing an example user interface interaction using force information to interact with a slider bar. In some embodiments, a slider bar may be used to indicate an intensity level or a time location (e.g., video position during playback). In some embodiments, when navigating through a video sequence, a user wants the slider to move quickly to a particular time index/portion of the sequence, but then move with greater precision to focus on a particular scene or even a single frame of video. Diagram 2000 shows a slider bar that can be moved by touching and dragging on the slider bar with a touch input.
[0087] In some embodiments, a speed or precision of slider bar movement using touch input dragging may be proportional to the force intensity level of the touch input. For example, a slider control moves with detailed/fine precision when "light" pressure is applied but moves with coarse/faster precision when "harder" pressure is applied. In some embodiments, the slider bar may be moved with greater (e.g., fine or less granular) precision when a touch input force intensity within a first intensity range is applied and moved with less (e.g., coarse or more granular) precision when a touch input force intensity within a second intensity range is applied. The threshold levels may be preconfigured, dynamically determined, and/or configurable. [0088] In some embodiments, a velocity at which an object such as a finger or stylus contacts a touch input surface is used to control a user interface. For example, video games, virtual musical instruments (drums and pianos are two common examples), and other applications may utilize velocity information to provide desired functionality. In some embodiments, measurement of contact velocity may be achieved by measuring the rate of change of the force. For example, if the touch force changes at a given point from 0 to 0.5 pounds in 20 milliseconds, it can be inferred that the finger or other object impacted the touch input screen at high velocity. On the other hand, a change in force from 0 to 0.1 pounds in 100 milliseconds could be construed as a relatively low velocity. Both the absolute measure of pressure and the rate-of-change of pressure may be useful measures of information in user interface design.
[0089] Although the foregoing embodiments have been described in some detail for purposes of clarity of understanding, the invention is not limited to the details provided. There are many alternative ways of implementing the invention. The disclosed embodiments are illustrative and not restrictive.
[0090] WHAT IS CLAIMED IS:

Claims

1. A system for detecting a touch input force, comprising:
a communication interface configured to send a signal to be used to propagate a propagating signal through a propagating medium with a surface and receive the propagating signal that has been disturbed by a touch input with an amount of force on the surface; and
a processor coupled to the communication interface and configured to process the received signal to determine an identifier associated with the amount of force.
2. The system of claim 1, wherein processing the received signal to determine the identifier includes determining a user indication associated with the disturbance.
3. The system of claim 2, wherein determining the user indication associated with the disturbance includes determining a location of the disturbance on the surface.
4. The system of claim 1, wherein the processed received signal is one of a plurality of versions of the disturbed propagating signal received by the communication interface.
5. The system of claim 4, wherein the processed received signal was selected among the plurality of versions due to a signal-to-noise ratio associated with the processed received signal.
6. The system of claim 4, wherein the processed received signal was selected among the plurality of versions due to a location associated with a detector that received the processed received signal.
7. The system of claim 1, wherein processing the received signal includes normalizing the received signal.
8. The system of claim 7, wherein normalizing the received signal includes adjusting an amplitude of the received signal.
9. The system of claim 8, wherein adjusting the amplitude includes scaling the amplitude using a value determined using a distance value associated with a location of the disturbance on the surface.
10. The system of claim 1, wherein processing the received signal includes determining a signal amplitude associated with a portion of the received signal disturbed by the touch input.
11. The system of claim 10, wherein determining the signal amplitude associated with the portion of the received signal disturbed by the touch input includes determining a signal amplitude of a normalized version of the received signal.
12. The system of claim 10, wherein a value of the signal amplitude is a function of the amount of the force.
13. The system of claim 10, wherein determining the identifier includes using the signal amplitude to locate a data structure entry, wherein the data structure entry includes the identifier associated with the amount of force.
14. The system of claim 13, wherein the data structure entry was located in a data structure identified to match a detector that received the processed received signal.
15. The system of claim 10, wherein determining the identifier includes averaging the determined signal amplitude with one or more other signal amplitudes of one or more other received signals.
16. The system of claim 1, wherein the processor is further configured to combine the identifier associated with the amount of force with one or more other identifiers associated with an amount of force of one or more other touch input location points to determine a combined force identifier.
17. The system of claim 1 , wherein the identifier is used at least in part to provide a user interface interaction.
18. The system of claim 1, wherein the identifier indicates an amount of pressure.
19. A method for detecting a touch input force, comprising:
sending a signal to be used to propagate a propagating signal through a propagating medium with a surface;
receiving the propagating signal that has been disturbed by a touch input with an amount of force on the surface; and
using a processor to process the received signal to determine an identifier associated with the amount of force.
20. A computer program product for detecting a touch input force, the computer program product being embodied in a tangible computer readable storage medium and comprising computer instructions for:
sending a signal to be used to propagate a propagating signal through a propagating medium with a surface;
receiving the propagating signal that has been disturbed by a touch input with an amount of force on the surface; and
processing the received signal to determine an identifier associated with the amount of force.
21. A system for detecting a touch input force including:
a surface configured to receive a touch input with an amount of force;
a transmitter that propagates a propagating signal through a medium of the surface, wherein the propagating signal is disturbed by the touch input that disturbs the propagating signal; and a detector that detects the disturbed propagating signal, wherein the detector outputs data indicating a disturbance of the propagating signal that is indicative of the amount of feree.
PCT/US2012/065939 2011-11-18 2012-11-19 Detecting touch input force WO2013075137A1 (en)

Priority Applications (6)

Application Number Priority Date Filing Date Title
KR1020177016603A KR101803261B1 (en) 2011-11-18 2012-11-19 Detecting touch input force
EP12849599.1A EP2780783B1 (en) 2011-11-18 2012-11-19 Detecting touch input force
KR1020177033898A KR101850680B1 (en) 2011-11-18 2012-11-19 Detecting touch input force
KR1020167021667A KR101750300B1 (en) 2011-11-18 2012-11-19 Detecting touch input force
KR1020147013247A KR101648143B1 (en) 2011-11-18 2012-11-19 Detecting touch input force
CN201280067220.5A CN104169848B (en) 2011-11-18 2012-11-19 Detect touch input force

Applications Claiming Priority (8)

Application Number Priority Date Filing Date Title
US201161561697P 2011-11-18 2011-11-18
US201161561660P 2011-11-18 2011-11-18
US61/561,660 2011-11-18
US61/561,697 2011-11-18
US13/451,288 2012-04-19
US13/451,288 US9477350B2 (en) 2011-04-26 2012-04-19 Method and apparatus for active ultrasonic touch devices
US201261673102P 2012-07-18 2012-07-18
US61/673,102 2012-07-18

Publications (1)

Publication Number Publication Date
WO2013075137A1 true WO2013075137A1 (en) 2013-05-23

Family

ID=48430256

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2012/065939 WO2013075137A1 (en) 2011-11-18 2012-11-19 Detecting touch input force

Country Status (5)

Country Link
US (10) US10248262B2 (en)
EP (1) EP2780783B1 (en)
KR (4) KR101803261B1 (en)
CN (2) CN104169848B (en)
WO (1) WO2013075137A1 (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2016074919A1 (en) * 2014-11-11 2016-05-19 Fresenius Vial Sas Method for processing an input for controlling an infusion operation
AU2013259614B2 (en) * 2012-05-09 2016-08-25 Apple Inc. Device, method, and graphical user interface for providing feedback for changing activation states of a user interface object
EP3065037A4 (en) * 2014-06-11 2017-01-11 Boe Technology Group Co. Ltd. Touch device and driving method
WO2022075976A1 (en) * 2020-10-06 2022-04-14 Google Llc Touch sensitive audio surface
US11698701B1 (en) 2022-03-23 2023-07-11 Infineon Technologies Ag Force sensing in touch sensor by directivity control of transmit transducers

Families Citing this family (154)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8487759B2 (en) 2009-09-30 2013-07-16 Apple Inc. Self adapting haptic device
US8823644B2 (en) 2009-12-08 2014-09-02 Contour Design Inc. Inner-sensor based pointing device
US10013058B2 (en) 2010-09-21 2018-07-03 Apple Inc. Touch-based user interface with haptic feedback
US10120446B2 (en) 2010-11-19 2018-11-06 Apple Inc. Haptic input device
US9477350B2 (en) 2011-04-26 2016-10-25 Sentons Inc. Method and apparatus for active ultrasonic touch devices
US9639213B2 (en) 2011-04-26 2017-05-02 Sentons Inc. Using multiple signals to detect touch input
US10198097B2 (en) 2011-04-26 2019-02-05 Sentons Inc. Detecting touch input force
US11327599B2 (en) 2011-04-26 2022-05-10 Sentons Inc. Identifying a contact type
US9189109B2 (en) 2012-07-18 2015-11-17 Sentons Inc. Detection of type of object used to provide a touch contact input
US9417754B2 (en) 2011-08-05 2016-08-16 P4tents1, LLC User interface system, method, and computer program product
CN104169848B (en) 2011-11-18 2017-10-20 森顿斯公司 Detect touch input force
US10235004B1 (en) 2011-11-18 2019-03-19 Sentons Inc. Touch input detector with an integrated antenna
US9449476B2 (en) 2011-11-18 2016-09-20 Sentons Inc. Localized haptic feedback
KR101691278B1 (en) 2012-05-03 2017-01-09 애플 인크. Moment compensated bending beam sensor for load measurement on platform supported by bending beams
WO2013169849A2 (en) 2012-05-09 2013-11-14 Industries Llc Yknots Device, method, and graphical user interface for displaying user interface objects corresponding to an application
CN108052264B (en) 2012-05-09 2021-04-27 苹果公司 Device, method and graphical user interface for moving and placing user interface objects
WO2013169845A1 (en) 2012-05-09 2013-11-14 Yknots Industries Llc Device, method, and graphical user interface for scrolling nested regions
WO2013169842A2 (en) 2012-05-09 2013-11-14 Yknots Industries Llc Device, method, and graphical user interface for selecting object within a group of objects
WO2013169875A2 (en) 2012-05-09 2013-11-14 Yknots Industries Llc Device, method, and graphical user interface for displaying content associated with a corresponding affordance
KR101683868B1 (en) 2012-05-09 2016-12-07 애플 인크. Device, method, and graphical user interface for transitioning between display states in response to gesture
WO2013169851A2 (en) 2012-05-09 2013-11-14 Yknots Industries Llc Device, method, and graphical user interface for facilitating user interaction with controls in a user interface
EP3594797B1 (en) * 2012-05-09 2024-10-02 Apple Inc. Device, method, and graphical user interface for providing tactile feedback for operations performed in a user interface
WO2013169865A2 (en) 2012-05-09 2013-11-14 Yknots Industries Llc Device, method, and graphical user interface for moving a user interface object based on an intensity of a press input
CN106201316B (en) 2012-05-09 2020-09-29 苹果公司 Apparatus, method and graphical user interface for selecting user interface objects
CN108958550B (en) 2012-05-09 2021-11-12 苹果公司 Device, method and graphical user interface for displaying additional information in response to user contact
WO2013169843A1 (en) 2012-05-09 2013-11-14 Yknots Industries Llc Device, method, and graphical user interface for manipulating framed graphical objects
US9348468B2 (en) 2013-06-07 2016-05-24 Sentons Inc. Detecting multi-touch inputs
WO2014018121A1 (en) 2012-07-26 2014-01-30 Changello Enterprise Llc Fingerprint-assisted force estimation
WO2014018116A1 (en) 2012-07-26 2014-01-30 Changello Enterprise Llc Ultrasound-based force sensing and touch sensing
WO2014018119A1 (en) * 2012-07-26 2014-01-30 Changello Enterprise Llc Ultrasound-based force and touch sensing
WO2014018115A1 (en) * 2012-07-26 2014-01-30 Changello Enterprise Llc Ultrasound-based force sensing of inputs
WO2014018118A1 (en) * 2012-07-26 2014-01-30 Changello Enterprise Llc Force detection by an ultrasound sensor
WO2014035479A2 (en) 2012-08-30 2014-03-06 Changello Enterprise Llc Auto-baseline determination for force sensing
US9178509B2 (en) 2012-09-28 2015-11-03 Apple Inc. Ultra low travel keyboard
US9983715B2 (en) 2012-12-17 2018-05-29 Apple Inc. Force detection in touch devices using piezoelectric sensors
TWI486868B (en) * 2012-12-26 2015-06-01 Giga Byte Tech Co Ltd Electrionic device with shortcut function and control method thereof
KR102301592B1 (en) 2012-12-29 2021-09-10 애플 인크. Device, method, and graphical user interface for navigating user interface hierachies
WO2014105279A1 (en) 2012-12-29 2014-07-03 Yknots Industries Llc Device, method, and graphical user interface for switching between user interfaces
AU2013368445B8 (en) 2012-12-29 2017-02-09 Apple Inc. Device, method, and graphical user interface for determining whether to scroll or select contents
JP6158947B2 (en) 2012-12-29 2017-07-05 アップル インコーポレイテッド Device, method and graphical user interface for transitioning between relationships from touch input to display output
EP2939095B1 (en) 2012-12-29 2018-10-03 Apple Inc. Device, method, and graphical user interface for moving a cursor according to a change in an appearance of a control icon with simulated three-dimensional characteristics
JP6093877B2 (en) 2012-12-29 2017-03-08 アップル インコーポレイテッド Device, method, and graphical user interface for foregoing generation of tactile output for multi-touch gestures
US20140184513A1 (en) * 2012-12-31 2014-07-03 Nvidia Corporation Softkey magnification on touch screen
SE541496C2 (en) * 2013-02-05 2019-10-22 Contour Design Inc Improved pointing device
KR102117086B1 (en) * 2013-03-08 2020-06-01 삼성디스플레이 주식회사 Terminal and method for controlling thereof
US9952703B2 (en) 2013-03-15 2018-04-24 Apple Inc. Force sensing of inputs through strain analysis
WO2014194192A1 (en) 2013-05-30 2014-12-04 David Andrews Multi-dimensional trackpad
KR102179056B1 (en) * 2013-07-19 2020-11-16 엘지전자 주식회사 Mobile terminal and control method for the mobile terminal
US9459715B1 (en) 2013-09-20 2016-10-04 Sentons Inc. Using spectral control in detecting touch input
US9619044B2 (en) * 2013-09-25 2017-04-11 Google Inc. Capacitive and resistive-pressure touch-sensitive touchpad
US9779592B1 (en) 2013-09-26 2017-10-03 Apple Inc. Geared haptic feedback element
WO2015047356A1 (en) 2013-09-27 2015-04-02 Bodhi Technology Ventures Llc Band with haptic actuators
WO2015047343A1 (en) 2013-09-27 2015-04-02 Honessa Development Laboratories Llc Polarized magnetic actuators for haptic response
US10126817B2 (en) 2013-09-29 2018-11-13 Apple Inc. Devices and methods for creating haptic effects
CN105683865B (en) 2013-09-30 2018-11-09 苹果公司 Magnetic actuator for haptic response
DE112014004628T5 (en) 2013-10-08 2016-07-28 Tk Holdings Inc. Apparatus and method for directing haptic energy to the touch surface
US9317118B2 (en) 2013-10-22 2016-04-19 Apple Inc. Touch surface for simulating materials
WO2015066086A1 (en) 2013-10-28 2015-05-07 Changello Enterprise Llc Piezo based force sensing
US10276001B2 (en) 2013-12-10 2019-04-30 Apple Inc. Band attachment mechanism with haptic response
AU2015100011B4 (en) 2014-01-13 2015-07-16 Apple Inc. Temperature compensating transparent force sensor
DE112014006608B4 (en) 2014-04-21 2024-01-25 Apple Inc. Methods, systems and electronic devices for determining force distribution for multi-touch input devices of electronic devices
DE102015209639A1 (en) 2014-06-03 2015-12-03 Apple Inc. Linear actuator
EP3195088A2 (en) 2014-09-02 2017-07-26 Apple Inc. Haptic notifications
KR102277902B1 (en) * 2014-09-05 2021-07-15 삼성전자주식회사 Object contact pressure meter and methods of manufacturing and measuring the same
US10466826B2 (en) 2014-10-08 2019-11-05 Joyson Safety Systems Acquisition Llc Systems and methods for illuminating a track pad system
TWI586971B (en) * 2014-12-26 2017-06-11 達方電子股份有限公司 Measurement and feedback circuit, measurement and feedback keyswitch, and measurement and feedback method
CN104731151B (en) * 2015-01-20 2017-02-22 苏州达方电子有限公司 Measuration and feedback circuit, measuration and feedback button and measuration and feedback method
US10353467B2 (en) * 2015-03-06 2019-07-16 Apple Inc. Calibration of haptic devices
US9632664B2 (en) 2015-03-08 2017-04-25 Apple Inc. Devices, methods, and graphical user interfaces for manipulating user interface objects with visual and/or haptic feedback
US10095396B2 (en) 2015-03-08 2018-10-09 Apple Inc. Devices, methods, and graphical user interfaces for interacting with a control object while dragging another object
US9645732B2 (en) 2015-03-08 2017-05-09 Apple Inc. Devices, methods, and graphical user interfaces for displaying and using menus
US9990107B2 (en) 2015-03-08 2018-06-05 Apple Inc. Devices, methods, and graphical user interfaces for displaying and using menus
US10048757B2 (en) 2015-03-08 2018-08-14 Apple Inc. Devices and methods for controlling media presentation
US9639184B2 (en) * 2015-03-19 2017-05-02 Apple Inc. Touch input cursor manipulation
US9785305B2 (en) 2015-03-19 2017-10-10 Apple Inc. Touch input cursor manipulation
US10152208B2 (en) 2015-04-01 2018-12-11 Apple Inc. Devices and methods for processing touch inputs based on their intensities
US20170045981A1 (en) 2015-08-10 2017-02-16 Apple Inc. Devices and Methods for Processing Touch Inputs Based on Their Intensities
AU2016100399B4 (en) 2015-04-17 2017-02-02 Apple Inc. Contracting and elongating materials for providing input and output for an electronic device
US9830048B2 (en) 2015-06-07 2017-11-28 Apple Inc. Devices and methods for processing touch inputs with instructions in a web page
US10200598B2 (en) 2015-06-07 2019-02-05 Apple Inc. Devices and methods for capturing and interacting with enhanced digital images
US9891811B2 (en) 2015-06-07 2018-02-13 Apple Inc. Devices and methods for navigating between user interfaces
US10346030B2 (en) 2015-06-07 2019-07-09 Apple Inc. Devices and methods for navigating between user interfaces
US9674426B2 (en) 2015-06-07 2017-06-06 Apple Inc. Devices and methods for capturing and interacting with enhanced digital images
US9860451B2 (en) 2015-06-07 2018-01-02 Apple Inc. Devices and methods for capturing and interacting with enhanced digital images
US9612170B2 (en) 2015-07-21 2017-04-04 Apple Inc. Transparent strain sensors in an electronic device
US10055048B2 (en) 2015-07-31 2018-08-21 Apple Inc. Noise adaptive force touch
US9880735B2 (en) 2015-08-10 2018-01-30 Apple Inc. Devices, methods, and graphical user interfaces for manipulating user interface objects with visual and/or haptic feedback
US10416800B2 (en) 2015-08-10 2019-09-17 Apple Inc. Devices, methods, and graphical user interfaces for adjusting user interface objects
US10235035B2 (en) * 2015-08-10 2019-03-19 Apple Inc. Devices, methods, and graphical user interfaces for content navigation and manipulation
US10248308B2 (en) 2015-08-10 2019-04-02 Apple Inc. Devices, methods, and graphical user interfaces for manipulating user interfaces with physical gestures
US10007418B2 (en) * 2015-09-08 2018-06-26 Apple Inc. Device, method, and graphical user interface for enabling generation of contact-intensity-dependent interface responses
US10566888B2 (en) 2015-09-08 2020-02-18 Apple Inc. Linear actuators for use in electronic devices
US20170068413A1 (en) * 2015-09-09 2017-03-09 Microsoft Technology Licensing, Llc Providing an information set relating to a graphical user interface element on a graphical user interface
US9874965B2 (en) 2015-09-11 2018-01-23 Apple Inc. Transparent strain sensors in an electronic device
US10048811B2 (en) 2015-09-18 2018-08-14 Sentons Inc. Detecting touch input provided by signal transmitting stylus
US9886118B2 (en) 2015-09-30 2018-02-06 Apple Inc. Transparent force sensitive structures in an electronic device
US11182068B2 (en) * 2015-10-27 2021-11-23 Verizon Patent And Licensing Inc. Method and system for interacting with a touch screen
US10039080B2 (en) 2016-03-04 2018-07-31 Apple Inc. Situationally-aware alerts
US10006820B2 (en) 2016-03-08 2018-06-26 Apple Inc. Magnetic interference avoidance in resistive sensors
JP6187625B1 (en) * 2016-03-24 2017-08-30 オンキヨー株式会社 Volume adjusting device and volume adjusting program
US10209830B2 (en) 2016-03-31 2019-02-19 Apple Inc. Electronic device having direction-dependent strain elements
US10268272B2 (en) 2016-03-31 2019-04-23 Apple Inc. Dampening mechanical modes of a haptic actuator using a delay
CN109074186B (en) * 2016-04-08 2022-12-06 希迪普公司 Pressure sensor, touch input device including the same, and pressure detection method using the same
CN106020573A (en) * 2016-05-18 2016-10-12 成都吉锐时代触摸技术有限公司 Writing method for ultrasonic pen
WO2017206051A1 (en) * 2016-05-31 2017-12-07 深圳市汇顶科技股份有限公司 Method and device for use in detecting pressure
CN107710127B (en) * 2016-05-31 2020-10-20 深圳市汇顶科技股份有限公司 Method and device for detecting pressure
DK179034B1 (en) 2016-06-12 2017-09-04 Apple Inc Devices, methods, and graphical user interfaces for dynamically adjusting presentation of audio outputs
DK179297B1 (en) * 2016-06-12 2018-04-16 Apple Inc Devices, methods, and graphical user interfaces for processing intensity information associated with touch inputs
CN107491283B (en) * 2016-06-12 2020-03-27 苹果公司 Apparatus, method and graphical user interface for dynamically adjusting presentation of audio output
CN108351750B (en) 2016-06-12 2019-08-13 苹果公司 For handling equipment, method and the graphic user interface of strength information associated with touch input
US11314388B2 (en) * 2016-06-30 2022-04-26 Huawei Technologies Co., Ltd. Method for viewing application program, graphical user interface, and terminal
US10133418B2 (en) 2016-09-07 2018-11-20 Apple Inc. Force sensing in an electronic device using a single layer of strain-sensitive structures
US11500538B2 (en) * 2016-09-13 2022-11-15 Apple Inc. Keyless keyboard with force sensing and haptic feedback
CN108604145A (en) * 2016-09-17 2018-09-28 深圳市汇顶科技股份有限公司 Pressure detection method, touch chip and pressure detecting module
US10908741B2 (en) 2016-11-10 2021-02-02 Sentons Inc. Touch input detection along device sidewall
SE545355C2 (en) 2016-11-11 2023-07-18 Contour Innovations Llc Inner-sensor pointing device systems
US10296144B2 (en) 2016-12-12 2019-05-21 Sentons Inc. Touch input detection with shared receivers
US20180196567A1 (en) * 2017-01-09 2018-07-12 Microsoft Technology Licensing, Llc Pressure sensitive virtual keyboard
US10126877B1 (en) 2017-02-01 2018-11-13 Sentons Inc. Update of reference data for touch input detection
US10585522B2 (en) 2017-02-27 2020-03-10 Sentons Inc. Detection of non-touch inputs using a signature
US10444091B2 (en) 2017-04-11 2019-10-15 Apple Inc. Row column architecture for strain sensing
CN108733268A (en) * 2017-04-18 2018-11-02 成都吉锐时代触摸技术有限公司 A method of detection writing physical strength
CN108803589A (en) * 2017-04-28 2018-11-13 深圳乐动机器人有限公司 Robot virtual wall system
US10712930B2 (en) 2017-05-28 2020-07-14 International Business Machines Corporation 3D touch based user interface value pickers
US10530770B2 (en) * 2017-06-28 2020-01-07 International Business Machines Corporation Pressure-based authentication
US10216333B2 (en) 2017-06-30 2019-02-26 Microsoft Technology Licensing, Llc Phase error compensation in single correlator systems
US10622538B2 (en) 2017-07-18 2020-04-14 Apple Inc. Techniques for providing a haptic output and sensing a haptic input using a piezoelectric body
US10309846B2 (en) 2017-07-24 2019-06-04 Apple Inc. Magnetic field cancellation for strain sensors
US10775850B2 (en) 2017-07-26 2020-09-15 Apple Inc. Computer with keyboard
US11009411B2 (en) 2017-08-14 2021-05-18 Sentons Inc. Increasing sensitivity of a sensor using an encoded signal
US11580829B2 (en) 2017-08-14 2023-02-14 Sentons Inc. Dynamic feedback for haptics
US11307661B2 (en) 2017-09-25 2022-04-19 Apple Inc. Electronic device with actuators for producing haptic and audio output along a device housing
CN108073337A (en) * 2018-01-02 2018-05-25 京东方科技集团股份有限公司 Touch-control structure and its control method, display device
GB2574588A (en) * 2018-06-06 2019-12-18 Cambridge Touch Tech Ltd Pressure sensing apparatus and method
KR102139808B1 (en) 2018-03-28 2020-07-31 주식회사 하이딥 Pressure sensor comprising a pulurality of channels, touch input device including the same, and pressure detection method using the same
US20190333102A1 (en) * 2018-04-30 2019-10-31 Affle (India) Limited Method and system for hardware and software based user identification for advertisement fraud detection
JP6938426B2 (en) * 2018-05-22 2021-09-22 京セラ株式会社 Electronics
US10757491B1 (en) 2018-06-11 2020-08-25 Apple Inc. Wearable interactive audio device
US10782818B2 (en) 2018-08-29 2020-09-22 Apple Inc. Load cell array for detection of force input to an electronic device enclosure
US11334032B2 (en) 2018-08-30 2022-05-17 Apple Inc. Electronic watch with barometric vent
US11561144B1 (en) 2018-09-27 2023-01-24 Apple Inc. Wearable electronic device with fluid-based pressure sensing
US10691211B2 (en) 2018-09-28 2020-06-23 Apple Inc. Button providing force sensing and/or haptic output
US10599223B1 (en) 2018-09-28 2020-03-24 Apple Inc. Button providing force sensing and/or haptic output
JP2020119099A (en) * 2019-01-21 2020-08-06 レノボ・シンガポール・プライベート・リミテッド Touch pad and electronic apparatus
CN114399014A (en) 2019-04-17 2022-04-26 苹果公司 Wireless locatable tag
US11380470B2 (en) 2019-09-24 2022-07-05 Apple Inc. Methods to control force in reluctance actuators based on flux related parameters
TWI730474B (en) * 2019-10-25 2021-06-11 國立成功大學 Dynamic symbol feature device and operation method thereof
WO2021138077A1 (en) 2019-12-30 2021-07-08 Joyson Safety Systems Acquisition Llc Systems and methods for intelligent waveform interruption
CN111443859B (en) * 2020-03-24 2022-09-20 维沃移动通信有限公司 Touch interaction method and electronic equipment
US11977683B2 (en) 2021-03-12 2024-05-07 Apple Inc. Modular systems configured to provide localized haptic feedback using inertial actuators
US11809631B2 (en) 2021-09-21 2023-11-07 Apple Inc. Reluctance haptic engine for an electronic device
US12001666B2 (en) * 2021-10-07 2024-06-04 madVR Labs LLC Hierarchical profile based settings
US20230297229A1 (en) * 2022-03-21 2023-09-21 Shopify Inc. Systems and methods for managing graphical user interfaces on mobile devices
US12028271B2 (en) 2022-12-06 2024-07-02 Bank Of America Corporation Prioritizing messages for server processing based on monitoring and predicting server resource utilization

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040160421A1 (en) * 2001-07-04 2004-08-19 Sullivan Darius Martin Contact sensitive device
US20060071912A1 (en) * 2004-10-01 2006-04-06 Hill Nicholas P R Vibration sensing touch input device
US20060262104A1 (en) * 2005-05-19 2006-11-23 Sullivan Darius M Systems and methods for distinguishing contact-induced plate vibrations from acoustic noise-induced plate vibrations
US20090103853A1 (en) * 2007-10-22 2009-04-23 Tyler Jon Daniel Interactive Surface Optical System
US7554246B2 (en) * 2004-08-03 2009-06-30 Sony Corporation Piezoelectric composite device, method of manufacturing same, method of controlling same, input-output device, and electronic device
US20110167391A1 (en) * 2010-01-06 2011-07-07 Brian Momeyer User interface methods and systems for providing force-sensitive input

Family Cites Families (371)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3705993A (en) 1970-07-16 1972-12-12 Inst Fizica Piezoresistive transducers and devices with semiconducting films and their manufacturing process
GB1426604A (en) 1973-07-06 1976-03-03 Post Office Acoustic measurements
JPS5660550A (en) 1979-10-22 1981-05-25 Mochida Pharm Co Ltd Fiber device for infrared laser for medical treatment
JPS5723499A (en) 1980-07-19 1982-02-06 Ricoh Kk Flash power source driving system
FR2533100B1 (en) 1982-09-09 1986-06-27 Sintra Alcatel Sa METHOD AND DEVICE FOR ATTENUATING INTERFERENCE NOISE
US4488000A (en) 1982-09-30 1984-12-11 New York Institute Of Technology Apparatus for determining position and writing pressure
JPS59118040U (en) 1983-01-31 1984-08-09 アルプス電気株式会社 input device
EP0277588B1 (en) 1987-02-04 1993-03-24 Ono Pharmaceutical Co., Ltd. Prolinal derivatives
FR2630223B1 (en) 1988-04-14 1990-08-10 Asulab Sa ECHO TRACKER FOR ULTRASONIC MEASUREMENT OF THE POSITION OF A MOBILE WALL
US5008497A (en) 1990-03-22 1991-04-16 Asher David J Touch controller
US5573077A (en) * 1990-11-16 1996-11-12 Knowles; Terence J. Acoustic touch position sensor
US5074152A (en) 1990-12-24 1991-12-24 Motorola, Inc. Piezoresistive transducer with low drift output voltage
US5233873A (en) 1991-07-03 1993-08-10 Texas Instruments Incorporated Accelerometer
US5334805A (en) 1992-06-15 1994-08-02 Carroll Touch Controller for an acoustic wave touch panel
WO1995011499A1 (en) * 1993-10-18 1995-04-27 Carroll Touch, Inc. Acoustic wave touch panel for use with a non-active stylus
US5410519A (en) 1993-11-19 1995-04-25 Coastal & Offshore Pacific Corporation Acoustic tracking system
DE69416960T2 (en) 1993-12-07 1999-08-19 Seiko Epson Corp Touch panel input device and method for generating input signals for an information processing device
JPH07160355A (en) 1993-12-08 1995-06-23 Hitachi Ltd Handwritten character input device and data input method
JPH07306746A (en) 1994-03-15 1995-11-21 Fujitsu Ltd Ultrasonic coordinate input device
US6127629A (en) 1994-10-03 2000-10-03 Ford Global Technologies, Inc. Hermetically sealed microelectronic device and method of forming same
US6236391B1 (en) 1995-01-24 2001-05-22 Elo Touchsystems, Inc. Acoustic touch position sensor using a low acoustic loss transparent substrate
JP3344138B2 (en) 1995-01-30 2002-11-11 株式会社日立製作所 Semiconductor composite sensor
US5635643A (en) 1995-03-15 1997-06-03 The University Of New Mexico Identifying source of acoustic emissions in structures using lamb waves and vibration modes
US5591945A (en) 1995-04-19 1997-01-07 Elo Touchsystems, Inc. Acoustic touch position sensor using higher order horizontally polarized shear wave propagation
US5708460A (en) 1995-06-02 1998-01-13 Avi Systems, Inc. Touch screen
US6307942B1 (en) 1995-09-02 2001-10-23 New Transducers Limited Panel-form microphones
US6232960B1 (en) 1995-12-21 2001-05-15 Alfred Goldman Data input device
USRE39881E1 (en) 1996-02-15 2007-10-16 Leapfrog Enterprises, Inc. Surface position location system and method
US5739479A (en) 1996-03-04 1998-04-14 Elo Touchsystems, Inc. Gentle-bevel flat acoustic wave touch sensor
US5784054A (en) 1996-03-22 1998-07-21 Elo Toughsystems, Inc. Surface acoustic wave touchscreen with housing seal
US8421776B2 (en) * 1996-08-12 2013-04-16 Elo Touch Solutions, Inc. Acoustic condition sensor employing a plurality of mutually non-orthogonal waves
US7629969B2 (en) 1996-08-12 2009-12-08 Tyco Electronics Corporation Acoustic condition sensor employing a plurality of mutually non-orthogonal waves
BR9714435B1 (en) 1996-12-25 2010-07-27 acoustic touch device, substrate for an acoustic sensitive device and process of detecting touch on a substrate.
US5883457A (en) 1997-04-24 1999-03-16 Raychem Corporation Organic matrix for acoustic reflector array
US5912659A (en) 1997-09-03 1999-06-15 International Business Machines Corporation Graphics display pointer with integrated selection
US6507772B1 (en) 1998-01-02 2003-01-14 Perdec Pty. Ltd. Sensor device
US20030214498A1 (en) 1998-04-30 2003-11-20 David Gothard High resolution computer operated digital display system
JP4196431B2 (en) 1998-06-16 2008-12-17 パナソニック株式会社 Built-in microphone device and imaging device
US6429846B2 (en) 1998-06-23 2002-08-06 Immersion Corporation Haptic feedback for touchpads and other touch controls
JP4024933B2 (en) 1998-08-18 2007-12-19 タッチパネル・システムズ株式会社 Touch panel
GB9818827D0 (en) 1998-08-29 1998-10-21 Ncr Int Inc Surface wave touch screen
US6731270B2 (en) 1998-10-21 2004-05-04 Luidia Inc. Piezoelectric transducer for data entry device
US6535147B1 (en) 1998-11-16 2003-03-18 The Whitaker Corporation Segmented gain controller
US6254105B1 (en) 1999-04-02 2001-07-03 Elo Touchsystems, Inc. Sealing system for acoustic wave touchscreens
JP2001014093A (en) 1999-06-30 2001-01-19 Touch Panel Systems Kk Acoustic contact detecter
JP2001014094A (en) 1999-06-30 2001-01-19 Touch Panel Systems Kk Acoustic contact detecter
US6492979B1 (en) 1999-09-07 2002-12-10 Elo Touchsystems, Inc. Dual sensor touchscreen utilizing projective-capacitive and force touch sensors
US6630929B1 (en) 1999-09-29 2003-10-07 Elo Touchsystems, Inc. Adaptive frequency touchscreen controller
US6262946B1 (en) 1999-09-29 2001-07-17 The Board Of Trustees Of The Leland Stanford Junior University Capacitive micromachined ultrasonic transducer arrays with reduced cross-coupling
US6473075B1 (en) 1999-09-29 2002-10-29 Elo Touchsystems, Inc. Adaptive frequency touchscreen controller employing digital signal processing
JP2001142639A (en) 1999-11-15 2001-05-25 Pioneer Electronic Corp Touch panel device
GB9928682D0 (en) 1999-12-06 2000-02-02 Electrotextiles Comp Ltd Input apparatus and a method of generating control signals
WO2001071478A2 (en) 2000-03-22 2001-09-27 Sony Electronics Inc Data entry user interface
US6636201B1 (en) 2000-05-26 2003-10-21 Elo Touchsystems, Inc. Acoustic touchscreen having waveguided reflector arrays
US7113175B2 (en) 2000-09-26 2006-09-26 Intertact Corporation Methods and apparatus for supplying power to touch input devices in a touch sensing system
US7006081B2 (en) 2000-10-20 2006-02-28 Elo Touchsystems, Inc. Acoustic touch sensor with laminated substrate
JP4552308B2 (en) 2000-10-24 2010-09-29 パナソニック株式会社 Ultrasonic coordinate input device and interactive board
US6498590B1 (en) 2001-05-24 2002-12-24 Mitsubishi Electric Research Laboratories, Inc. Multi-user touch surface
JP2002366292A (en) 2001-06-13 2002-12-20 Matsushita Electric Ind Co Ltd Coordinate input device
US7254775B2 (en) 2001-10-03 2007-08-07 3M Innovative Properties Company Touch panel system and method for distinguishing multiple touch inputs
US7265746B2 (en) 2003-06-04 2007-09-04 Illinois Tool Works Inc. Acoustic wave touch detection circuit and method
ATE491984T1 (en) 2002-02-06 2011-01-15 Soundtouch Ltd METHOD FOR A TOUCH PAD
JP3975810B2 (en) 2002-04-05 2007-09-12 株式会社日立製作所 Optical single sideband transmitter
JP2004005415A (en) 2002-04-19 2004-01-08 Sharp Corp Input device and input/output integrated display
DE10219641A1 (en) 2002-05-02 2003-12-18 Siemens Ag Display with integrated loudspeaker and method for detecting touches of a display
US7746325B2 (en) * 2002-05-06 2010-06-29 3M Innovative Properties Company Method for improving positioned accuracy for a determined touch input
FR2841022B1 (en) 2002-06-12 2004-08-27 Centre Nat Rech Scient METHOD FOR LOCATING AN IMPACT ON A SURFACE AND DEVICE FOR IMPLEMENTING SAID METHOD
US11275405B2 (en) 2005-03-04 2022-03-15 Apple Inc. Multi-functional hand-held device
CA2397431A1 (en) 2002-08-09 2004-02-09 Andrew Lohbihler Method and apparatus for a wireless position sensing interface device employing spread spectrum technology of one or more radio transmitting devices
US20040203594A1 (en) 2002-08-12 2004-10-14 Michael Kotzin Method and apparatus for signature validation
US6990435B2 (en) 2002-10-10 2006-01-24 Harmonic Drive Systems Inc. Tactile sensing method and system
JP2004163262A (en) 2002-11-13 2004-06-10 Touch Panel Systems Kk Sound wave type contact detector
JP4090329B2 (en) 2002-11-13 2008-05-28 タッチパネル・システムズ株式会社 Acoustic wave contact detector
US7388571B2 (en) 2002-11-21 2008-06-17 Research In Motion Limited System and method of integrating a touchscreen within an LCD
US6871149B2 (en) 2002-12-06 2005-03-22 New Transducers Limited Contact sensitive device
JP4124444B2 (en) 2003-01-30 2008-07-23 富士通コンポーネント株式会社 Touch panel, input device having the same, and electronic device
US7000474B2 (en) 2003-02-28 2006-02-21 Elo Touchsystems, Inc. Acoustic device using higher order harmonic piezoelectric element
US7116315B2 (en) 2003-03-14 2006-10-03 Tyco Electronics Corporation Water tolerant touch sensor
US7119800B2 (en) 2003-06-24 2006-10-10 Tyco Electronics Corporation Acoustic touch sensor with low-profile diffractive grating transducer assembly
JP4186745B2 (en) 2003-08-01 2008-11-26 ソニー株式会社 Microphone device, noise reduction method, and recording device
JP2005092527A (en) 2003-09-17 2005-04-07 Fdk Corp Ultrasonic touch panel
FR2863070B1 (en) 2003-12-02 2006-03-03 Sensitive Object METHOD AND DEVICE FOR MAN INTERFACE - MACHINE
US7411584B2 (en) 2003-12-31 2008-08-12 3M Innovative Properties Company Touch sensitive device employing bending wave vibration sensing and excitation transducers
US7315300B2 (en) 2003-12-31 2008-01-01 3M Innovative Properties Company Touch sensitive device employing impulse reconstruction
US7277087B2 (en) * 2003-12-31 2007-10-02 3M Innovative Properties Company Touch sensing with touch down and lift off sensitivity
US7274358B2 (en) 2004-01-06 2007-09-25 Tyco Electronics Corporation Focusing-shaped reflector arrays for acoustic touchscreens
US7340960B2 (en) 2004-01-30 2008-03-11 Analatom Inc. Miniature sensor
US6856259B1 (en) 2004-02-06 2005-02-15 Elo Touchsystems, Inc. Touch sensor system to detect multiple touch events
US8010123B2 (en) 2004-02-26 2011-08-30 Research In Motion Limited Cellular communications system with mobile cellular device battery saving features based upon quality of service and access denial and related methods
KR20050092179A (en) 2004-03-15 2005-09-21 삼성전자주식회사 Heater system for use in semiconductor fabricating apparatus
US20050248548A1 (en) 2004-04-14 2005-11-10 Masahiro Tsumura Acoustic touch sensor
US7538759B2 (en) 2004-05-07 2009-05-26 Next Holdings Limited Touch panel display system with illumination and detection provided from a single edge
US7315336B2 (en) 2004-06-16 2008-01-01 Tyco Electronics Corporation Stress seal for touchscreen systems
JP2008508631A (en) 2004-08-02 2008-03-21 コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ Touch screen slider to set a floating point value
FR2874274B1 (en) 2004-08-11 2007-07-20 Sensitive Object METHOD FOR LOCATING AN IMPACT ON A SURFACE AND DEVICE FOR IMPLEMENTING SAID METHOD
TWI296383B (en) 2004-09-14 2008-05-01 Ind Tech Res Inst Surface acoustic wave touch panel and system of the same
FR2876406B1 (en) 2004-10-12 2008-05-02 Sensitive Object ACOUSTICALLY CONTROLLED ACCESS CONTROL INSTALLATION AND ACCESS CONTROL DEVICE FOR SUCH A INSTALLATION
US8334837B2 (en) 2004-11-10 2012-12-18 Nokia Corporation Method for displaying approached interaction areas
US7180401B2 (en) 2004-12-03 2007-02-20 Kulite Semiconductor Products, Ic. Personal identification apparatus using measured tactile pressure
US20060139339A1 (en) 2004-12-29 2006-06-29 Pechman Robert J Touch location determination using vibration wave packet dispersion
US8392486B2 (en) 2004-12-29 2013-03-05 Elo Touch Solutions, Inc. Method for determining the position of impacts
US7499039B2 (en) * 2005-01-10 2009-03-03 3M Innovative Properties Company Iterative method for determining touch location
JP4652826B2 (en) 2005-01-14 2011-03-16 タッチパネル・システムズ株式会社 Information input device
JP4199741B2 (en) 2005-02-25 2008-12-17 Necディスプレイソリューションズ株式会社 Wave receiver and wave reception determination method
CN101133385B (en) 2005-03-04 2014-05-07 苹果公司 Hand held electronic device, hand held device and operation method thereof
EP1707931B1 (en) 2005-03-31 2013-03-27 STMicroelectronics Srl Analog data-input device provided with a microelectromechanical pressure sensor
US8692810B2 (en) 2005-04-13 2014-04-08 Elo Touch Solutions, Inc. Method for determining the location of impacts by acoustic imaging
US7683890B2 (en) * 2005-04-28 2010-03-23 3M Innovative Properties Company Touch location determination using bending mode sensors and multiple detection techniques
US7218248B2 (en) * 2005-05-18 2007-05-15 Microsoft Corporation Input mode switching system
US8614695B2 (en) 2005-06-07 2013-12-24 Intel Corporation Ultrasonic tracking
US9019209B2 (en) 2005-06-08 2015-04-28 3M Innovative Properties Company Touch location determination involving multiple touch location processes
KR100714722B1 (en) 2005-06-17 2007-05-04 삼성전자주식회사 Apparatus and method for implementing pointing user interface using signal of light emitter
JP4549243B2 (en) 2005-07-05 2010-09-22 アルパイン株式会社 In-vehicle audio processor
US7868874B2 (en) 2005-11-15 2011-01-11 Synaptics Incorporated Methods and systems for detecting a position-based attribute of an object using digital codes
US20070126716A1 (en) 2005-11-17 2007-06-07 Jonathan Haverly Digital pen
US8494805B2 (en) 2005-11-28 2013-07-23 Orthosensor Method and system for assessing orthopedic alignment using tracking sensors
JP2007193469A (en) 2006-01-18 2007-08-02 Seiko Epson Corp Touch panel, touch input position detection method, electrooptical device and electronic apparatus
JP4412289B2 (en) * 2006-01-26 2010-02-10 セイコーエプソン株式会社 Electro-optical device and electronic apparatus
RU2406245C2 (en) 2006-02-08 2010-12-10 Квэлкомм Инкорпорейтед Spectrum formation for reducing peak to mean power ratio in wireless communication
US8312479B2 (en) 2006-03-08 2012-11-13 Navisense Application programming interface (API) for sensory events
US8139029B2 (en) 2006-03-08 2012-03-20 Navisense Method and device for three-dimensional sensing
US8405618B2 (en) * 2006-03-24 2013-03-26 Northwestern University Haptic device with indirect haptic feedback
KR20070097675A (en) 2006-03-28 2007-10-05 삼성전자주식회사 Display panel
US7764276B2 (en) 2006-04-18 2010-07-27 Schermerhorn Jerry D Touch control system and apparatus with multiple acoustic coupled substrates
JP5048271B2 (en) 2006-05-02 2012-10-17 任天堂株式会社 GAME PROGRAM AND GAME DEVICE
JP4809130B2 (en) 2006-05-31 2011-11-09 タッチパネル・システムズ株式会社 Acoustic wave type touch panel
US7538470B2 (en) 2006-06-05 2009-05-26 Zyvex Labs, Llc Monolithic nanoscale actuation
US8692809B2 (en) 2006-07-06 2014-04-08 Elo Touch Solutions, Inc. Auto-gain switching module for acoustic touch systems
KR101303881B1 (en) 2006-08-03 2013-09-04 삼성디스플레이 주식회사 Touch screen and method of manufacturing the same, and display having the same
US7747225B2 (en) 2006-10-03 2010-06-29 Motorola, Inc. Mobile assisted downlink beamforming with antenna weight feedback
KR20090088882A (en) 2006-10-25 2009-08-20 센시티브 오브젝트 Automatic gain control circuit, system including such circuit and method for automatic gain control
US7999797B2 (en) * 2006-12-26 2011-08-16 Sony Ericsson Mobile Communications Ab Detecting and locating a touch or a tap on an input surface
US20080169132A1 (en) 2007-01-03 2008-07-17 Yao Ding Multiple styli annotation system
US8094128B2 (en) 2007-01-03 2012-01-10 Apple Inc. Channel scan logic
TW200832196A (en) 2007-01-19 2008-08-01 Inventec Appliances Corp Press-type touch control pad device
US20080198145A1 (en) 2007-02-20 2008-08-21 Knowles Terrence J Acoustic wave touch actuated system
US9329719B2 (en) 2007-03-15 2016-05-03 Apple Inc. Hybrid force sensitive touch devices
WO2008130824A1 (en) 2007-04-18 2008-10-30 Luidia Inc. Pre-assembled part with an associated surface convertible to a transcription apparatus
US8243048B2 (en) 2007-04-25 2012-08-14 Elo Touch Solutions, Inc. Touchscreen for detecting multiple touches
US8493332B2 (en) 2007-06-21 2013-07-23 Elo Touch Solutions, Inc. Method and system for calibrating an acoustic touchscreen
US20090002328A1 (en) 2007-06-26 2009-01-01 Immersion Corporation, A Delaware Corporation Method and apparatus for multi-touch tactile touch panel actuator mechanisms
US8730213B2 (en) 2007-07-02 2014-05-20 Elo Touch Solutions, Inc. Method and system for detecting touch events based on redundant validation
US8378974B2 (en) 2007-07-02 2013-02-19 Elo Touch Solutions, Inc. Method and system for detecting touch events based on magnitude ratios
CN101373415B (en) 2007-08-21 2010-06-02 禾瑞亚科技股份有限公司 Surface acoustic wave type sensing apparatus of touching control panel
WO2009028680A1 (en) 2007-08-30 2009-03-05 Nec Corporation Optimum pseudo random sequence determining method, position detecting system, position detecting method, transmitter, and receiver
US8090547B2 (en) 2007-10-01 2012-01-03 Panasonic Corporation Ultrasonic measuring device and ultrasonic measuring method
DE102007052008A1 (en) 2007-10-26 2009-04-30 Andreas Steinhauser Single- or multitouch-capable touchscreen or touchpad consisting of an array of pressure sensors and production of such sensors
US8144897B2 (en) 2007-11-02 2012-03-27 Research In Motion Limited Adjusting acoustic speaker output based on an estimated degree of seal of an ear about a speaker port
US8121311B2 (en) 2007-11-05 2012-02-21 Qnx Software Systems Co. Mixer with adaptive post-filtering
US8418083B1 (en) * 2007-11-26 2013-04-09 Sprint Communications Company L.P. Applying a navigational mode to a device
US7816838B2 (en) 2007-12-11 2010-10-19 Nokia Corporation Piezoelectric force sensing
US20090160728A1 (en) 2007-12-21 2009-06-25 Motorola, Inc. Uncorrelated antennas formed of aligned carbon nanotubes
EP2073106B1 (en) 2007-12-21 2012-02-08 Tyco Electronics Services GmbH Method for determining the locations of at least two impacts
US9857872B2 (en) 2007-12-31 2018-01-02 Apple Inc. Multi-touch display screen with localized tactile feedback
GB0801396D0 (en) 2008-01-25 2008-03-05 Bisutti Giovanni Electronic apparatus
US8358277B2 (en) * 2008-03-18 2013-01-22 Microsoft Corporation Virtual keyboard based activation and dismissal
EP2104024B1 (en) 2008-03-20 2018-05-02 LG Electronics Inc. Portable terminal capable of sensing proximity touch and method for controlling screen using the same
US8156809B2 (en) 2008-03-27 2012-04-17 Immersion Corporation Systems and methods for resonance detection
US20090271004A1 (en) 2008-04-28 2009-10-29 Reese Zecchin Method and apparatus for ranging detection of gestures
US20090273583A1 (en) 2008-05-05 2009-11-05 Sony Ericsson Mobile Communications Ab Contact sensitive display
US8159469B2 (en) * 2008-05-06 2012-04-17 Hewlett-Packard Development Company, L.P. User interface for initiating activities in an electronic device
ATE523839T1 (en) 2008-05-07 2011-09-15 Tyco Electronics Services Gmbh METHOD FOR DETERMINING THE POSITION OF AN IMPACT ON THE SURFACE OF AN OBJECT
US20090309853A1 (en) 2008-06-13 2009-12-17 Polyvision Corporation Electronic whiteboard system and assembly with optical detection elements
US9030418B2 (en) * 2008-06-24 2015-05-12 Lg Electronics Inc. Mobile terminal capable of sensing proximity touch
NO328622B1 (en) 2008-06-30 2010-04-06 Tandberg Telecom As Device and method for reducing keyboard noise in conference equipment
ATE523836T1 (en) 2008-07-07 2011-09-15 Tyco Electronics Services Gmbh DEVICE FOR CHANGING THE OPERATING STATE OF A DEVICE
US8743091B2 (en) 2008-07-31 2014-06-03 Apple Inc. Acoustic multi-touch sensor panel
GB2462465B (en) * 2008-08-08 2013-02-13 Hiwave Technologies Uk Ltd Touch sensitive device
US8604364B2 (en) 2008-08-15 2013-12-10 Lester F. Ludwig Sensors, algorithms and applications for a high dimensional touchpad
US8279202B2 (en) 2008-08-25 2012-10-02 Chrontel International Ltd. Methods of making and using an apparatus for providing multi-touch sensing capability using an LCD screen without an extra touch screen layer
US10289199B2 (en) 2008-09-29 2019-05-14 Apple Inc. Haptic feedback system
GB2464117B (en) 2008-10-03 2015-01-28 Hiwave Technologies Uk Ltd Touch sensitive device
EP2175348A1 (en) 2008-10-07 2010-04-14 Sensitive Object Tactile man-machine interface with data communication interface
US20100123686A1 (en) 2008-11-19 2010-05-20 Sony Ericsson Mobile Communications Ab Piezoresistive force sensor integrated in a display
US8227956B2 (en) 2008-11-25 2012-07-24 Intel Corporation Mobile user interface with energy harvesting
US20100126273A1 (en) 2008-11-25 2010-05-27 New Jersey Institute Of Technology Flexible impact sensors and methods of making same
US20100141408A1 (en) 2008-12-05 2010-06-10 Anthony Stephen Doy Audio amplifier apparatus to drive a panel to produce both an audio signal and haptic feedback
US8686952B2 (en) 2008-12-23 2014-04-01 Apple Inc. Multi touch with multi haptics
US20100165215A1 (en) 2008-12-31 2010-07-01 Youngtack Shim Communication control systems and methods
EP2214082B1 (en) 2009-01-29 2012-08-15 Tyco Electronics Services GmbH A touch-sensing device with a touch hold function and a corresponding method
JP4856206B2 (en) 2009-03-30 2012-01-18 株式会社東芝 Wireless device
CN102388353A (en) 2009-04-09 2012-03-21 新型转换器有限公司 Touch sensitive device
KR101537706B1 (en) 2009-04-16 2015-07-20 엘지전자 주식회사 Mobile terminal and control method thereof
US8253712B2 (en) * 2009-05-01 2012-08-28 Sony Ericsson Mobile Communications Ab Methods of operating electronic devices including touch sensitive interfaces using force/deflection sensing and related devices and computer program products
JP2010263411A (en) 2009-05-07 2010-11-18 Renesas Electronics Corp Touch sensor system, and microcomputer
WO2011093837A2 (en) 2009-05-22 2011-08-04 Tyco Electronics Corporation Method and apparatus for detecting hold condition on an acoustic touch surface
US9134848B2 (en) 2009-06-03 2015-09-15 Stmicroelectronics Asia Pacific Pte Ltd Touch tracking on a touch sensitive interface
US20100311337A1 (en) 2009-06-04 2010-12-09 Psion Teklogix Inc. Multi-modal speech holster
US20100321325A1 (en) 2009-06-17 2010-12-23 Springer Gregory A Touch and display panel antennas
EP2264575B1 (en) 2009-06-19 2017-08-09 Elo Touch Solutions, Inc. Method for determining the locations of one or more impacts or touches on a surface of an object including two or more transducers
US8593415B2 (en) * 2009-06-19 2013-11-26 Lg Electronics Inc. Method for processing touch signal in mobile terminal and mobile terminal using the same
TWI467423B (en) 2009-06-22 2015-01-01 Au Optronics Corp Touch detection method
TW201101137A (en) 2009-06-29 2011-01-01 J Touch Corp Touch panel with matrix type tactile feedback
US8451237B2 (en) 2009-07-06 2013-05-28 Atmel Corporation Sensitivity control as a function of touch shape
US8310458B2 (en) 2009-07-06 2012-11-13 Research In Motion Limited Electronic device including a moveable touch-sensitive input and method of controlling same
US8378797B2 (en) 2009-07-17 2013-02-19 Apple Inc. Method and apparatus for localization of haptic feedback
KR20110007848A (en) 2009-07-17 2011-01-25 삼성전자주식회사 Apparatus and method for controlling a mobile terminal
FR2948471B1 (en) 2009-07-21 2016-02-26 Commissariat Energie Atomique METHOD AND DEVICE FOR LOCATING AT LEAST ONE TOUCH ON A TOUCH SURFACE OF AN OBJECT
WO2011011025A1 (en) 2009-07-24 2011-01-27 Research In Motion Limited Method and apparatus for a touch-sensitive display
FR2948787B1 (en) 2009-07-29 2011-09-09 Commissariat Energie Atomique DEVICE AND METHOD FOR LOCATING A LOCALLY DEFORMING CONTACT ON A DEFORMABLE TOUCH SURFACE OF AN OBJECT
JP5197521B2 (en) 2009-07-29 2013-05-15 京セラ株式会社 Input device
US9107325B1 (en) 2009-08-10 2015-08-11 The Boeing Company Systems and methods of coupling sensors to a structure
US8791907B2 (en) 2009-08-19 2014-07-29 U-Pixel Technologies Inc. Touch sensing apparatus and method using different modulated driving signals
EP2472365B1 (en) 2009-08-27 2016-10-12 Kyocera Corporation Tactile sensation imparting device and control method of tactile sensation imparting device
JP5304544B2 (en) 2009-08-28 2013-10-02 ソニー株式会社 Information processing apparatus, information processing method, and program
JP5278259B2 (en) 2009-09-07 2013-09-04 ソニー株式会社 Input device, input method, and program
US8325160B2 (en) 2009-09-11 2012-12-04 3M Innovative Properties Company Contact sensitive device for detecting temporally overlapping traces
US8436806B2 (en) 2009-10-02 2013-05-07 Research In Motion Limited Method of synchronizing data acquisition and a portable electronic device configured to perform the same
EP2315101B1 (en) 2009-10-02 2014-01-29 BlackBerry Limited A method of waking up and a portable electronic device configured to perform the same
WO2011041943A1 (en) 2009-10-09 2011-04-14 禾瑞亚科技股份有限公司 Method and device for analyzing position
US8717309B2 (en) * 2009-10-13 2014-05-06 Blackberry Limited Portable electronic device including a touch-sensitive display and method of controlling same
WO2011048433A1 (en) 2009-10-23 2011-04-28 Elliptic Laboratories As Touchless interfaces
KR20120101359A (en) 2009-10-29 2012-09-13 하이웨이브 테크놀러지스 (유케이) 리미티드 Touch sensitive device employing bending wave vibration sensors that detect touch location and provide haptic feedback
US8633916B2 (en) 2009-12-10 2014-01-21 Apple, Inc. Touch pad with force sensors and actuator feedback
US8194051B2 (en) 2009-12-15 2012-06-05 Silicon Integrated Systems Corp. Multiple fingers touch sensing method using matching algorithm
GB0922071D0 (en) 2009-12-17 2010-02-03 Wolfson Microelectronics Plc Interface
JP5345050B2 (en) 2009-12-25 2013-11-20 株式会社ワコム Indicator, position detection device and position detection method
KR101632237B1 (en) 2009-12-28 2016-06-22 삼성전자주식회사 TSP antenna of a mobile terminal
EP2524288A1 (en) 2010-01-13 2012-11-21 Elo Touch Solutions, Inc. Noise reduction in electronic device with touch sensitive surface
US8624878B2 (en) 2010-01-20 2014-01-07 Apple Inc. Piezo-based acoustic and capacitive detection
US20110260988A1 (en) 2010-01-20 2011-10-27 Northwestern University Method and apparatus for increasing magnitude and frequency of forces applied to a bare finger on a haptic surface
US9250742B1 (en) 2010-01-26 2016-02-02 Open Invention Network, Llc Method and apparatus of position tracking and detection of user input information
US8791899B1 (en) 2010-01-26 2014-07-29 Open Invention Network, Llc Method and apparatus of position tracking and detection of user input information
US20110182443A1 (en) 2010-01-26 2011-07-28 Gant Anthony W Electronic device having a contact microphone
KR101678549B1 (en) 2010-02-02 2016-11-23 삼성전자주식회사 Method and apparatus for providing user interface using surface acoustic signal, and device with the user interface
US8648815B2 (en) 2010-02-15 2014-02-11 Elo Touch Solutions, Inc. Touch panel that has an image layer and detects bending waves
US20110199342A1 (en) 2010-02-16 2011-08-18 Harry Vartanian Apparatus and method for providing elevated, indented or texturized sensations to an object near a display device or input detection using ultrasound
US8591413B2 (en) 2010-02-26 2013-11-26 Empire Technology Development Llc Echogram detection of skin conditions
US8988970B2 (en) 2010-03-12 2015-03-24 University Of Maryland Method and system for dereverberation of signals propagating in reverberative environments
KR101677621B1 (en) * 2010-03-12 2016-11-18 엘지전자 주식회사 Content control apparatus and method thereof
US8576202B2 (en) 2010-03-25 2013-11-05 Elo Touch Solutions, Inc. Bezel-less acoustic touch apparatus
EP2372509A1 (en) 2010-03-29 2011-10-05 Tyco Electronics Services GmbH Method for detecting a sustained contact and corresponding device
US8374072B2 (en) * 2010-04-07 2013-02-12 Qualcomm Incorporated Efficient zadoff-chu sequence generation
WO2011125688A1 (en) 2010-04-09 2011-10-13 Semiconductor Energy Laboratory Co., Ltd. Liquid crystal display device and method for driving the same
US8698766B2 (en) 2010-04-22 2014-04-15 Maxim Integrated Products, Inc. System integration of tactile feedback and touchscreen controller for near-zero latency haptics playout
US8736560B2 (en) 2010-05-14 2014-05-27 Blackberry Limited Electronic device including tactile touch-sensitive display and method of controlling same
US8638318B2 (en) 2010-05-28 2014-01-28 Elo Touch Solutions, Inc. Multi-layer coversheet for saw touch panel
KR101718016B1 (en) 2010-06-04 2017-03-21 엘지전자 주식회사 Mobile terminal and method for producing antenna of mobile terminal
KR20110133766A (en) 2010-06-07 2011-12-14 삼성전자주식회사 Mobile terminal and method for controlling function thereof
WO2011156447A1 (en) 2010-06-11 2011-12-15 3M Innovative Properties Company Positional touch sensor with force measurement
US20110304577A1 (en) 2010-06-11 2011-12-15 Sp Controls, Inc. Capacitive touch screen stylus
US8519982B2 (en) 2010-06-21 2013-08-27 Sony Corporation Active acoustic touch location for electronic devices
US8593398B2 (en) 2010-06-25 2013-11-26 Nokia Corporation Apparatus and method for proximity based input
US8907929B2 (en) 2010-06-29 2014-12-09 Qualcomm Incorporated Touchless sensing and gesture recognition using continuous wave ultrasound signals
US8411874B2 (en) 2010-06-30 2013-04-02 Google Inc. Removing noise from audio
EP2405332B1 (en) 2010-07-09 2013-05-29 Elo Touch Solutions, Inc. Method for determining a touch event and touch sensitive device
GB2482190A (en) 2010-07-23 2012-01-25 New Transducers Ltd Methods of generating a desired haptic sensation in a touch sensitive device
KR20120011548A (en) 2010-07-29 2012-02-08 삼성전자주식회사 Apparatus and Mehtod for Providing Feedback on User Input
KR20120012541A (en) * 2010-08-02 2012-02-10 삼성전자주식회사 Method and apparatus for operating folder in a touch device
US8917249B1 (en) 2010-08-05 2014-12-23 Amazon Technologies, Inc. Reduction of touch-sensor interference from active display
US8907930B2 (en) 2010-08-06 2014-12-09 Motorola Mobility Llc Methods and devices for determining user input location using acoustic sensing elements
JP6049990B2 (en) 2010-09-15 2016-12-21 京セラ株式会社 Portable electronic device, screen control method, and screen control program
US20120068939A1 (en) 2010-09-17 2012-03-22 Nigel Patrick Pemberton-Pigott Touch-sensitive display with depression detection and method
US9223431B2 (en) 2010-09-17 2015-12-29 Blackberry Limited Touch-sensitive display with depression detection and method
US8787599B2 (en) 2010-09-20 2014-07-22 Plantronics, Inc. Reduced microphone handling noise
US20120081337A1 (en) 2010-10-04 2012-04-05 Sony Ericsson Mobile Communications Ab Active Acoustic Multi-Touch and Swipe Detection for Electronic Devices
WO2012046890A1 (en) 2010-10-06 2012-04-12 엘지전자 주식회사 Mobile terminal, display device, and method for controlling same
US9013959B2 (en) 2010-10-13 2015-04-21 University Of Delaware Long-range acoustical positioning system on continental shelf regions
US9158958B2 (en) 2010-10-28 2015-10-13 Synaptics Incorporated Signal strength enhancement in a biometric sensor array
US8659210B2 (en) 2010-11-02 2014-02-25 Immersion Corporation Piezo based inertia actuator for high definition haptic feedback
US10120446B2 (en) 2010-11-19 2018-11-06 Apple Inc. Haptic input device
KR20120062297A (en) * 2010-12-06 2012-06-14 삼성전자주식회사 Display apparatus and user interface providing method thereof
US8543168B2 (en) 2010-12-14 2013-09-24 Motorola Mobility Llc Portable electronic device
US8681809B2 (en) 2011-01-10 2014-03-25 Qualcomm Incorporated Dynamic enabling and disabling of CLTD operation via HS SCCH orders
US20120188194A1 (en) 2011-01-20 2012-07-26 Research In Motion Limited Providing a preload apparatus in a touch sensitive electronic device
US9417696B2 (en) 2011-01-27 2016-08-16 Blackberry Limited Portable electronic device and method therefor
US8674961B2 (en) 2011-01-31 2014-03-18 National Semiconductor Corporation Haptic interface for touch screen in mobile device or other device
US9762234B2 (en) 2011-02-10 2017-09-12 Synaptics Incorporated Input device interference determination
KR101759994B1 (en) 2011-03-16 2017-07-20 엘지전자 주식회사 Mobile terminal
WO2012127471A2 (en) 2011-03-21 2012-09-27 N-Trig Ltd. System and method for authentication with a computer stylus
US20140185834A1 (en) * 2011-04-11 2014-07-03 Nokia Corporation Volume control apparatus
US8892162B2 (en) 2011-04-25 2014-11-18 Apple Inc. Vibration sensing system and method for categorizing portable device context and modifying device operation
US10198097B2 (en) 2011-04-26 2019-02-05 Sentons Inc. Detecting touch input force
US9477350B2 (en) 2011-04-26 2016-10-25 Sentons Inc. Method and apparatus for active ultrasonic touch devices
US9639213B2 (en) 2011-04-26 2017-05-02 Sentons Inc. Using multiple signals to detect touch input
US11327599B2 (en) 2011-04-26 2022-05-10 Sentons Inc. Identifying a contact type
US9189109B2 (en) 2012-07-18 2015-11-17 Sentons Inc. Detection of type of object used to provide a touch contact input
US9295082B2 (en) 2011-04-27 2016-03-22 At&T Mobility Ii Llc Distributed machine-to-machine connectivity
US20120280944A1 (en) 2011-05-05 2012-11-08 3M Innovative Properties Company Touch sensor with modular components
JP2012249097A (en) 2011-05-27 2012-12-13 Kyocera Corp Speech output device
US20120306823A1 (en) 2011-06-06 2012-12-06 Apple Inc. Audio sensors
TR201106373A2 (en) 2011-06-28 2013-01-21 Vestel Elektroni̇k Sanayi̇ Ve Ti̇caret A.Ş. Display device with touch control.
US20130050154A1 (en) 2011-06-29 2013-02-28 Benjamin T. Guy Stylus for use with touch screen computing device
US9131915B2 (en) 2011-07-06 2015-09-15 University Of New Brunswick Method and apparatus for noise cancellation
US20130050133A1 (en) * 2011-08-30 2013-02-28 Nokia Corporation Method and apparatus for precluding operations associated with accidental touch inputs
CN102981659B (en) 2011-09-06 2016-01-27 宸鸿光电科技股份有限公司 The pointer of the control system of contact panel and control method and use thereof
US20130059532A1 (en) 2011-09-07 2013-03-07 Microsoft Corporation Partially Transparent Antenna
US8823685B2 (en) 2011-10-14 2014-09-02 Elo Touch Solutions, Inc. Acoustic touch apparatus
US8681128B2 (en) 2011-10-14 2014-03-25 Elo Touch Solutions, Inc. Acoustic touch apparatus
EP2587352A1 (en) 2011-10-25 2013-05-01 austriamicrosystems AG Touch sensing system and method for operating a touch sensing system
US9449476B2 (en) 2011-11-18 2016-09-20 Sentons Inc. Localized haptic feedback
CN104169848B (en) 2011-11-18 2017-10-20 森顿斯公司 Detect touch input force
US8890852B2 (en) 2011-12-12 2014-11-18 Elo Touch Solutions, Inc. Acoustic touch signal dispersion response and minimization
JP5087723B1 (en) 2012-01-30 2012-12-05 パナソニック株式会社 Information terminal device, control method thereof, and program
US9817568B2 (en) 2012-02-29 2017-11-14 Blackberry Limited System and method for controlling an electronic device
US9201546B2 (en) 2012-03-09 2015-12-01 Elo Touch Solutions, Inc. Acoustic touch apparatus with multi-touch capability
US8692812B2 (en) 2012-03-12 2014-04-08 Elo Touch Solutions, Inc. Layered border for touch sensor device
WO2013135252A1 (en) 2012-03-15 2013-09-19 Sony Mobile Communications Ab Method for controlling a touch sensor
US20150366504A1 (en) 2014-06-20 2015-12-24 Medibotics Llc Electromyographic Clothing
US9354731B1 (en) 2012-06-20 2016-05-31 Amazon Technologies, Inc. Multi-dimension touch input
US9078066B2 (en) 2012-07-18 2015-07-07 Sentons Inc. Touch input surface speaker
US9348468B2 (en) 2013-06-07 2016-05-24 Sentons Inc. Detecting multi-touch inputs
US9524063B2 (en) 2012-07-18 2016-12-20 Sentons Inc. Detection of a number of touch contacts of a multi-touch input
WO2014018121A1 (en) 2012-07-26 2014-01-30 Changello Enterprise Llc Fingerprint-assisted force estimation
US9557845B2 (en) 2012-07-27 2017-01-31 Apple Inc. Input device for and method of communication with capacitive devices through frequency variation
US9690394B2 (en) 2012-09-14 2017-06-27 Apple Inc. Input device having extendable nib
US9639179B2 (en) 2012-09-14 2017-05-02 Apple Inc. Force-sensitive input device
US20140078086A1 (en) 2012-09-20 2014-03-20 Marvell World Trade Ltd. Augmented touch control for hand-held devices
WO2014066621A2 (en) 2012-10-26 2014-05-01 Livescribe Inc. Establishing and maintaining an authenticated connection between a smart pen and a computing device
US8994696B2 (en) 2012-10-30 2015-03-31 Elo Touch Solutions, Inc. Acoustic touch apparatus with addressable multi-touch capability
US9158411B2 (en) 2013-07-12 2015-10-13 Tactual Labs Co. Fast multi-touch post processing
US9575557B2 (en) 2013-04-19 2017-02-21 Qualcomm Incorporated Grip force sensor array for one-handed and multimodal interaction on handheld devices and methods
US9551783B2 (en) 2013-06-03 2017-01-24 Qualcomm Incorporated Display with backside ultrasonic sensor array
US8766953B1 (en) 2013-06-27 2014-07-01 Elwha Llc Tactile display driven by surface acoustic waves
KR101489813B1 (en) 2013-06-28 2015-02-04 삼성전기주식회사 Touch panel including Active stylus pen and controlling method thereof
US9366651B2 (en) 2013-07-03 2016-06-14 Matrix Sensors, Inc. Array of sensors with surface modifications
US20150009185A1 (en) 2013-07-08 2015-01-08 Corning Incorporated Touch systems and methods employing acoustic sensing in a thin cover glass
JP2016528602A (en) 2013-08-20 2016-09-15 アドヴァンスト・シリコン・ソシエテ・アノニム Capacitive touch system
US9785263B2 (en) 2013-08-22 2017-10-10 Microchip Technology Incorporated Touch screen stylus with force and/or angle sensing functionality
JP6367951B2 (en) 2013-09-27 2018-08-01 センセル インコーポレイテッドSensel,Inc. Resistive touch sensor system and method
TW201516766A (en) 2013-10-18 2015-05-01 Waltop Int Corp Stylus
WO2015066086A1 (en) 2013-10-28 2015-05-07 Changello Enterprise Llc Piezo based force sensing
CN103677339B (en) 2013-11-25 2017-07-28 泰凌微电子(上海)有限公司 The wireless communication system of time writer, electromagnetic touch reception device and both compositions
AU2015100011B4 (en) 2014-01-13 2015-07-16 Apple Inc. Temperature compensating transparent force sensor
TW201528070A (en) 2014-01-13 2015-07-16 Himax Tech Ltd Stylus without active components and an associated touch panel
US9720521B2 (en) 2014-02-21 2017-08-01 Qualcomm Incorporated In-air ultrasound pen gestures
KR101606404B1 (en) 2014-03-05 2016-03-25 주식회사 동부하이텍 A touch sensor
US20150277617A1 (en) 2014-03-28 2015-10-01 Paul Gwin Flexible sensor
US9798399B2 (en) 2014-06-02 2017-10-24 Synaptics Incorporated Side sensing for electronic devices
US10362989B2 (en) 2014-06-09 2019-07-30 Bebop Sensors, Inc. Sensor system integrated with a glove
US9558455B2 (en) 2014-07-11 2017-01-31 Microsoft Technology Licensing, Llc Touch classification
US9851845B2 (en) 2014-08-12 2017-12-26 Apple Inc. Temperature compensation for transparent force sensors
US20160062517A1 (en) 2014-09-02 2016-03-03 Apple Inc. Multi-Layer Transparent Force Sensor
US9329715B2 (en) 2014-09-11 2016-05-03 Qeexo, Co. Method and apparatus for differentiating touch screen users based on touch event analysis
US20160091308A1 (en) 2014-09-30 2016-03-31 Invensense, Inc. Microelectromechanical systems (mems) acoustic sensor-based gesture recognition
JP5723499B1 (en) 2014-10-24 2015-05-27 株式会社ワコム Sending electronic pen
TWI541686B (en) 2014-12-03 2016-07-11 矽統科技股份有限公司 Stylus and signal transmitting methods for stylus and touch panel
KR102301621B1 (en) 2015-01-16 2021-09-14 삼성전자주식회사 Stylus pen, touch penel and coordinate indicating system having the same
US20160246396A1 (en) 2015-02-20 2016-08-25 Qualcomm Incorporated Interactive touchscreen and sensor array
US10296123B2 (en) 2015-03-06 2019-05-21 Apple Inc. Reducing noise in a force signal in an electronic device
US10331242B2 (en) 2015-03-25 2019-06-25 Microsoft Technology Licensing, Llc Stylus having a plurality of operating portions configured to transmit synchronized signals
US20170020402A1 (en) 2015-05-04 2017-01-26 The Board Of Trustees Of The University Of Illinois Implantable and bioresorbable sensors
WO2016190634A1 (en) 2015-05-27 2016-12-01 Samsung Electronics Co., Ltd. Touch recognition apparatus and control method thereof
US9766727B2 (en) 2015-07-06 2017-09-19 Apple Inc. Computer stylus with antenna
US10048811B2 (en) 2015-09-18 2018-08-14 Sentons Inc. Detecting touch input provided by signal transmitting stylus
EP3174208B1 (en) 2015-11-30 2019-09-18 Nokia Technologies Oy Sensing apparatus and associated methods
EP3179335B1 (en) 2015-12-10 2020-03-04 Nxp B.V. Haptic feedback controller
US9851848B2 (en) 2015-12-14 2017-12-26 Microsoft Technology Licensing, Llc Touch sensitive device casing
US10282046B2 (en) 2015-12-23 2019-05-07 Cambridge Touch Technologies Ltd. Pressure-sensitive touch panel
US10355659B2 (en) 2016-03-11 2019-07-16 Akoustis, Inc. Piezoelectric acoustic resonator manufactured with piezoelectric thin film transfer process
US9971423B2 (en) 2016-05-03 2018-05-15 Microsoft Technology Licensing, Llc Stylus with intermittent voltage driving and sensing
WO2018023080A2 (en) 2016-07-29 2018-02-01 Apple Inc. Methodology and application of acoustic touch detection
US9870033B1 (en) 2016-08-30 2018-01-16 Apple Inc. Sensor assemblies for electronic devices
US10133418B2 (en) 2016-09-07 2018-11-20 Apple Inc. Force sensing in an electronic device using a single layer of strain-sensitive structures
CN106354328B (en) 2016-09-14 2023-11-14 宸鸿科技(厦门)有限公司 Pressure sensing module and pressure sensing touch control system
US10095342B2 (en) 2016-11-14 2018-10-09 Google Llc Apparatus for sensing user input
KR102645631B1 (en) 2016-11-22 2024-03-08 삼성디스플레이 주식회사 Bended display device
US11255737B2 (en) 2017-02-09 2022-02-22 Nextinput, Inc. Integrated digital force sensors and related methods of manufacture
US10282651B2 (en) 2017-03-23 2019-05-07 Idex Asa Sensor array system selectively configurable as a fingerprint sensor or data entry device
CN111263924B (en) 2017-03-29 2022-10-14 苹果公司 Device with integrated interface system
US10277154B2 (en) 2017-05-01 2019-04-30 Apple Inc. Closed-loop control of linear resonant actuator using back EMF data and hall sensing
US10732714B2 (en) 2017-05-08 2020-08-04 Cirrus Logic, Inc. Integrated haptic system
US20190004622A1 (en) 2017-06-28 2019-01-03 Walmart Apollo, Llc Systems, Methods, and Devices for Providing a Virtual Reality Whiteboard
CN107426434B (en) 2017-08-01 2020-04-24 京东方科技集团股份有限公司 Input device and electronic apparatus
US11009411B2 (en) 2017-08-14 2021-05-18 Sentons Inc. Increasing sensitivity of a sensor using an encoded signal
US10540023B2 (en) 2017-09-26 2020-01-21 Yong Bum Kim User interface devices for virtual reality system
US10871847B2 (en) 2017-09-29 2020-12-22 Apple Inc. Sensing force and press location in absence of touch information
CN107643852B (en) 2017-10-17 2021-02-19 厦门天马微电子有限公司 Display panel and display device
US11209927B2 (en) 2017-12-11 2021-12-28 Google Llc Apparatus for sensing user input
US20200310561A1 (en) 2019-03-29 2020-10-01 Logitech Europe S.A. Input device for use in 2d and 3d environments
US11209916B1 (en) 2020-07-30 2021-12-28 Logitech Europe S.A. Dominant hand usage for an augmented/virtual reality device

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040160421A1 (en) * 2001-07-04 2004-08-19 Sullivan Darius Martin Contact sensitive device
US7554246B2 (en) * 2004-08-03 2009-06-30 Sony Corporation Piezoelectric composite device, method of manufacturing same, method of controlling same, input-output device, and electronic device
US20060071912A1 (en) * 2004-10-01 2006-04-06 Hill Nicholas P R Vibration sensing touch input device
US20060262104A1 (en) * 2005-05-19 2006-11-23 Sullivan Darius M Systems and methods for distinguishing contact-induced plate vibrations from acoustic noise-induced plate vibrations
US20090103853A1 (en) * 2007-10-22 2009-04-23 Tyler Jon Daniel Interactive Surface Optical System
US20110167391A1 (en) * 2010-01-06 2011-07-07 Brian Momeyer User interface methods and systems for providing force-sensitive input

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See also references of EP2780783A4 *

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
AU2013259614B2 (en) * 2012-05-09 2016-08-25 Apple Inc. Device, method, and graphical user interface for providing feedback for changing activation states of a user interface object
EP3065037A4 (en) * 2014-06-11 2017-01-11 Boe Technology Group Co. Ltd. Touch device and driving method
US9696846B2 (en) 2014-06-11 2017-07-04 Boe Technology Group Co., Ltd. Touch device and method for driving the same
WO2016074919A1 (en) * 2014-11-11 2016-05-19 Fresenius Vial Sas Method for processing an input for controlling an infusion operation
WO2022075976A1 (en) * 2020-10-06 2022-04-14 Google Llc Touch sensitive audio surface
US11698701B1 (en) 2022-03-23 2023-07-11 Infineon Technologies Ag Force sensing in touch sensor by directivity control of transmit transducers

Also Published As

Publication number Publication date
US9099971B2 (en) 2015-08-04
EP2780783A1 (en) 2014-09-24
CN104169848B (en) 2017-10-20
US20160370906A9 (en) 2016-12-22
US9594450B2 (en) 2017-03-14
KR101648143B1 (en) 2016-08-16
US20130141396A1 (en) 2013-06-06
KR20140097181A (en) 2014-08-06
US20200326810A1 (en) 2020-10-15
US10055066B2 (en) 2018-08-21
US20170139499A1 (en) 2017-05-18
US20190354221A1 (en) 2019-11-21
US10162443B2 (en) 2018-12-25
KR20170134760A (en) 2017-12-06
US20130141364A1 (en) 2013-06-06
US20180373358A1 (en) 2018-12-27
KR101750300B1 (en) 2017-06-23
CN107562281A (en) 2018-01-09
US10248262B2 (en) 2019-04-02
CN104169848A (en) 2014-11-26
US20130142362A1 (en) 2013-06-06
EP2780783A4 (en) 2014-11-05
KR20160096735A (en) 2016-08-16
US20150268785A1 (en) 2015-09-24
KR20170073724A (en) 2017-06-28
CN107562281B (en) 2020-12-22
KR101850680B1 (en) 2018-04-20
KR101803261B1 (en) 2017-11-30
US11016607B2 (en) 2021-05-25
US20240103667A1 (en) 2024-03-28
US10732755B2 (en) 2020-08-04
US11829555B2 (en) 2023-11-28
EP2780783B1 (en) 2022-12-28
US10353509B2 (en) 2019-07-16
US20210318776A1 (en) 2021-10-14

Similar Documents

Publication Publication Date Title
US11829555B2 (en) Controlling audio volume using touch input force
US10877581B2 (en) Detecting touch input force
US10386968B2 (en) Method and apparatus for active ultrasonic touch devices
US11907464B2 (en) Identifying a contact type

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 12849599

Country of ref document: EP

Kind code of ref document: A1

WWE Wipo information: entry into national phase

Ref document number: 2012849599

Country of ref document: EP

ENP Entry into the national phase

Ref document number: 20147013247

Country of ref document: KR

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE