WO2017029555A2 - Device, system, and methods for entering commands or characters using a touch screen - Google Patents

Device, system, and methods for entering commands or characters using a touch screen Download PDF

Info

Publication number
WO2017029555A2
WO2017029555A2 PCT/IB2016/001256 IB2016001256W WO2017029555A2 WO 2017029555 A2 WO2017029555 A2 WO 2017029555A2 IB 2016001256 W IB2016001256 W IB 2016001256W WO 2017029555 A2 WO2017029555 A2 WO 2017029555A2
Authority
WO
WIPO (PCT)
Prior art keywords
finger
contact
touch sensitive
sensitive surface
entry
Prior art date
Application number
PCT/IB2016/001256
Other languages
French (fr)
Other versions
WO2017029555A3 (en
Inventor
Ben Etzion YARON
Nissan YARON
Original Assignee
Inpris Innovative Products Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Inpris Innovative Products Ltd filed Critical Inpris Innovative Products Ltd
Priority to CN201680061416.1A priority Critical patent/CN108780365B/en
Priority to EP16836703.5A priority patent/EP3338172A4/en
Publication of WO2017029555A2 publication Critical patent/WO2017029555A2/en
Publication of WO2017029555A3 publication Critical patent/WO2017029555A3/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • G06F3/04883Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60KARRANGEMENT OR MOUNTING OF PROPULSION UNITS OR OF TRANSMISSIONS IN VEHICLES; ARRANGEMENT OR MOUNTING OF PLURAL DIVERSE PRIME-MOVERS IN VEHICLES; AUXILIARY DRIVES FOR VEHICLES; INSTRUMENTATION OR DASHBOARDS FOR VEHICLES; ARRANGEMENTS IN CONNECTION WITH COOLING, AIR INTAKE, GAS EXHAUST OR FUEL SUPPLY OF PROPULSION UNITS IN VEHICLES
    • B60K35/00Arrangement of adaptations of instruments
    • B60K35/10
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/033Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
    • G06F3/038Control and interface arrangements therefor, e.g. drivers or device-embedded control circuitry
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • G06F3/04847Interaction techniques to control parameter settings, e.g. interaction with sliders or dials
    • B60K2360/143
    • B60K2360/146
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2203/00Indexing scheme relating to G06F3/00 - G06F3/048
    • G06F2203/048Indexing scheme relating to G06F3/048
    • G06F2203/04804Transparency, e.g. transparent or translucent windows
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2203/00Indexing scheme relating to G06F3/00 - G06F3/048
    • G06F2203/048Indexing scheme relating to G06F3/048
    • G06F2203/04808Several contacts: gestures triggering a specific function, e.g. scrolling, zooming, right-click, when the user establishes several contacts with the surface simultaneously; e.g. using several fingers or a combination of fingers and pen

Definitions

  • the present teachings are directed at devices, systems, and methods for inputting one or more commands using a touch sensitive surface for control of an electronic device.
  • a positioning mode is employed for positioning control locations (e.g., finger contact regions) based on the contact of the touch sensitive surface with three or more fingers.
  • blind typing on a touch screen display is difficult. Indeed, many individuals desiring to send messages on a mobile phone while having the phone hidden under a table or in a pocket, prefer phones having individual keys over a phone having a touch screen display. When touching keys, the operator receives continually feedback regarding the locations of the various keys (i.e., the control locations), such as by feeling spaces between keys, or by feeling a key having a different shape than a neighboring key.
  • touch screen display devices can be used as a control device, there is a need for improved methods and systems so that a touch screen device can be used as a control device in circumstances when it is required to operate blind.
  • a touch screen device typically provides visual clues to identify the position of control locations on the surface.
  • Typical control locations are defined by an application. Although the operator may be able to move the control locations, such as by dragging a window, such movements typically requires an operator to first visually identify an initial location of the control location.
  • control locations are defined when the user contacts the touch screen surface, at the same time with 3, 4, or 5 fingers of a hand.
  • the device moves or identifies (i.e., defines) the control locations based on the contact locations of the user's fingers. After contacting the touch sensitive surface with the 3, 4 or 5 fingers of a hand, some or all of the fingers may be removed and the control locations may be maintained until the user makes an input contact with the surface with 1 , 2, 3, 4, or 5 fingers for controlling a device.
  • a method of entering a command comprising the steps of: i) a processor connected to a touch sensitive surface sensing a simultaneously positioning of three or more objects (e.g., fingers) above and near (e.g., within 10 mm of) or on the touch sensitive surface at three or more different sensing locations (e.g., consistent with the positioning by three or more fingers of a user), including a positioning centered at a first finger initial sensing point (e.g., by a first finger), a positioning centered at a second finger initials sensing point (e.g., by a second finger), and a positioning centered at a third finger initial sensing point (e.g., by a third finger); ii) the processor assigning a finger location region for two or more (e.g., each) of the three or more objects, wherein each finger location region is a distinct region of the touch sensitive surface, and each finger location region includes one of the initial finger sensing points (e.g.
  • each finger location region corresponds to one of the fingers
  • the processor entering a command entry mode following the step of assigning the finger location regions (e.g., following a predetermined event such as the removal of at least one finger, or following a predetermined time interval), wherein the command entry mode includes an association having at least a first command associated with a movement of only one of the objects starting at a first finger location region and a second different command associated with a movement (e.g., the same movement or a different movement as employed with the first command) of only one of the objects starting at a second finger location region; and iv) the processor recognizing a gesture on the touch sensitive surface including sensing the movement of only one of the objects starting at the first finger location region and identifying the associated first command based on the gesture.
  • a method of entering a command comprising the steps of: identifying a simultaneously contacting on a touch sensitive surface at three or more locations (e.g., consistent with the contact by three or more fingers of a user), including a contact centered at a first finger initial contact point (e.g., by a first finger), a contact centered at a second finger initial contact point (e.g., by a second finger of a second finger), and a contact at a third finger initial contact point (e.g., by a third finger); assigning a control location (e.g., a finger contact region) for each of the three or more fingers, wherein each finger contact region is a different region of the touch sensitive surface, and each finger contact region includes one of the initial contact points (e.g.
  • each finger contact region corresponds to one of the fingers
  • recognizing the removing of the contact of the touch sensitive at the three or more locations e.g., the removal of the three or more fingers from the touch sensitive surface.
  • the process may include one or more steps of entering a control command by contacts (i.e., entry contacts) with one or more control locations.
  • the process may include recognizing the simultaneous entry contact(s) in one or more of the finger contact regions (e.g., the simultaneously contacting of one or more of the finger contact regions with the corresponding finger) each at an entry contact point; recognizing a gesture on the touch sensitive surface including a movement of the entry contact in one or more directions on the touch sensitive surface starting from the entry contact point while continuous entry contact is maintained; (e.g., from a motion on the touch sensitive surface by the user making a gesture on the touch sensitive surface with one or more of the fingers in contact with touch sensitive surface, wherein each gesture originates in the corresponding finger contact region); and identifying a command to execute based on the control locations contacted and the gestures originating in the control locations.
  • the simultaneous entry contact(s) in one or more of the finger contact regions e.g., the simultaneously contacting of one or more of the finger contact regions with the corresponding finger
  • a method of entering a command using a processor connected to a touch sensitive surface comprising the steps of: (i) the processor identifying a simultaneously contact on the touch sensitive surface at three or more contact locations (e.g., consistent with the contact by three or more fingers of a user), including a contact centered at a first finger initial contact point (e.g., by a first finger), a contact centered at a second finger initial contact point (e.g., by a second finger of a second finger), and a contact centered at a third finger initial contact point (e.g., by a third finger);(ii) the processor assigning a finger contact region for each of the three or more fingers, wherein each finger contact region is a different region of the touch sensitive surface, and each finger contact region includes one of the initial contact points (e.g.
  • each finger contact region corresponds to one of the fingers); (iii) the processor recognizing the removing of the contact of the touch sensitive surface at one or more (e.g., all) of the contact locations (e.g., the removal of the one, two, three or all of the fingers from the touch sensitive surface); (iv) following the step of recognizing the removing of contact(s) the processor recognizing the simultaneous entry contact(s) in one or more of the finger contact regions (e.g., the simultaneously contacting of one or more of the finger contact regions with the corresponding finger) each at an entry contact point, and/or the processor recognizing one or more contacts from step (i) each remaining on a finger contact region and recognizing the remaining contacts as finger entry contact(s); (v) the processor recognizing a gesture on the touch sensitive surface including a movement of the entry contact in one or more directions on the touch sensitive surface starting from the entry contact point while continuous entry contact is maintained; (e.g., from a motion on the touch sensitive surface by the user making a gesture on the touch sensitive
  • a method of entering a control command comprising the steps of: simultaneously contacting a touch sensitive surface with three or more fingers pf a user, including a contact of a first finger centered at a first finger initial contact point, a contact of a second finger centered at a second finger initial contact point, and contact of a third finger at a third finger initial contact point;.
  • each finger contact region is a different region of the touch sensitive surface, and each finger contact region includes one of the initial contact points, wherein each finger contact region corresponds to one of the fingers; removing the three or more fingers from the touch sensitive surface; contacting one of the finger contact regions with the corresponding finger and sliding said finger in one or more finger movement directions.
  • the sliding of finger may control a device, a component of a device, a communication, a display, or any combination thereof.
  • the movement direction of the finger may result in the moving of a cursor in a corresponding one or more cursor movement directions.
  • a method of entering a command comprising the steps of: simultaneously contacting a touch sensitive surface with three or more fingers of a user, including a contact of a first finger centered at a first finger initial contact point, a contact of a second finger centered at a second finger initial contact point, and contact of a third finger at a third finger initial contact point; assigning a finger contact region for each of the three or more fingers, wherein each finger contact region is a different region of the touch sensitive surface, and each finger contact region includes one of the initial contact points, wherein each finger contact region corresponds to one of the fingers; and removing the three or more fingers from the touch sensitive surface.
  • one or more gestures may be entered using the finger contact regions.
  • the entry of a gesture may include simultaneously contacting one or more of the finger contact regions with the corresponding finger; making a gesture on the touch sensitive surface with one or more of the fingers in contact with touch sensitive surface, wherein each gesture originates in the corresponding finger contact region; and identifying a control command to execute based on the finger contact regions that are contacted and the gestures originating in the finger contact regions.
  • control locations may be static or dynamic.
  • a static finger contact region may remain fixed while in a command entry mode, and may change only after ending a command entry mode and initiating a new positioning mode.
  • control locations are finger location regions.
  • Finger location regions typically are spaced apart in an arc, such as expected from the spacing of finger tips.
  • the finger location regions may be identified by the sensing of objects near the touch sensitive surface (e.g., fingers slightly above the surface, typically within 10 mm of the surface), the finger location regions preferably are finger contact regions, where the objects contact the surface, while near the touch sensitive surface,
  • control locations e.g., finger location regions or finger contact regions
  • the control locations are dynamic.
  • a finger location region may be repositioned based on the actual location of finger position or contact (e.g., within a control location) during the command entry mode.
  • a first finger location region (e.g., finger contact region) may be characterized by an initial contact area and an initial center point (e.g., a geometric center of the initial contact area).
  • the touch surface may be contacted in the initial contact area, but the center of the contact may be at a contact point offset from the initial center point.
  • the center point for the first finger contact region may then be shifted at least partially towards (e.g., entirely to) the contact point, so that the first finger contact region is characterized by a new center point different from the initial center point.
  • the offset between a center point and a center of contact in one finger contact region may also be employed for repositioning one or more different finger contact regions. Such repositioning may be employed to compensate for gradual shifting of the position of a hand on a touch sensitive surface. As such, the device adjusts to the user's hand position instead of the user adjusting the hand to a fixed position of a device.
  • Another aspect of the teachings herein is directed at a system for entry of control commands for controlling a device comprising: an entry device including a touch sensitive surface; a processor; a memory storing instruction that, when executed by the processor causes the processor to: identify a simultaneously contact on the touch sensitive surface at three or more locations (e.g., consistent with the contact by three or more fingers of a user), including a contact centered at a first finger initial contact point (e.g., by a first finger), a contact centered at a second finger initial contact point (e.g., by a second finger of a second finger), and contact at a third finger initial contact point (e.g., by a third finger); assign a control location (e.g., a finger contact region) for each of the three or more fingers, wherein each control location is a different region of the touch sensitive surface, and each control location includes one of the initial contact points (e.g.
  • each finger contact region corresponds to one of the fingers; and recognize the removing of the contact of the touch sensitive at the three or more locations (e.g., the removal of the three or more fingers from the touch sensitive surface) before or after the step of assigning the control locations.
  • the memory storing instruction, when executed, after the assignment of the control locations may also cause the processor to: recognize the simultaneous entry contact(s) in one or more of the control locations (e.g., the simultaneously contacting of one or more of the finger contact regions with the corresponding finger) each at an entry contact point; recognize a gesture on the touch sensitive surface including a movement of the entry contact in one or more directions on the touch sensitive surface starting from the entry contact point while continuous entry contact is maintained; (e.g., from a motion on the touch sensitive surface by the user making a gesture on the touch sensitive surface with one or more of the fingers in contact with touch sensitive surface, wherein each gesture originates in the corresponding finger contact region); and identify a command to be execute based on the control locations that are contacted and the gestures originating in the control locations.
  • the simultaneous entry contact(s) in one or more of the control locations e.g., the simultaneously contacting of one or more of the finger contact regions with the corresponding finger
  • each gesture originates in the corresponding finger contact region
  • teachings herein is directed at a machine readable storage medium containing instructions that when executed cause a processor of an electronic device to discern input control commands by: the processor identifying a simultaneously contact on the touch sensitive surface at three or more locations (e.g.
  • a contact centered at a first finger initial contact point e.g., by a first finger
  • a contact centered at a second finger initial contact point e.g., by a second finger of a second finger
  • contact at a third finger initial contact point e.g., by a third finger
  • the processor assigning a control location (e.g., a finger contact region) for each of the three or more finger initial contact points, wherein each control location is a different region of the touch sensitive surface, and each control location includes one of the initial contact points (e.g.
  • each finger contact region corresponds to one of the fingers
  • the processor recognizing the removing of the contact of the touch sensitive at the three or more locations (e.g., the removal of the three or more fingers from the touch sensitive surface); the processor recognizing the simultaneous entry contact(s) in one or more of the finger contact regions (e.g., the simultaneously contacting of one or more of the finger contact regions with the corresponding finger) each at an entry contact point; the processor recognizing a gesture on the touch sensitive surface including a movement of the entry contact in one or more directions on the touch sensitive surface starting from the entry contact point while continuous entry contact is maintained; (e.g., from a motion on the touch sensitive surface by the user making a gesture on the touch sensitive surface with one or more of the fingers in contact with touch sensitive surface, wherein each gesture originates in the corresponding finger contact region); and the processor identifying a command to be execute based on the control locations contacted and the gestures originating in the control locations.
  • teachings herein are directed at methods, systems and devices for unlocking one or more electronic device controls by a processor comprising: receiving an indication of simultaneous contact at multiple locations on a touch sensitive surface; determining that the multiple locations have a spacing consistent with contact of the touch sensitive surface by one or two hands of a user; assigning multiple finger contact regions based on the multiple locations of simultaneous contact; receiving indications of a sequential series of contacts each with one or more of the finger contact regions; comparing the sequential series of contacts with a predetermined sequential series (i.e., a password series), and unlocking one or more controls when the sequential series of contacts matches the predetermined sequential series.
  • a predetermined sequential series i.e., a password series
  • FIG. 1 is a drawing showing illustrative modes for a control session according to the teachings herein.
  • FIG. 2 is a drawing showing illustrative modes for a control session according to the teachings herein.
  • FIG. 3 is a drawing illustrating simultaneous contact on a touch sensitive surface at three or more locations (e.g., corresponding to the touching by three or more fingers of a hand).
  • FIG. 4 is a drawing illustrating a touch sensitive surface showing the contacts removed (e.g., the fingers removed from the surface) and showing the locations of the prior simultaneous contact.
  • FIG. 5 is a drawing illustrating simultaneous contact of a touch sensitive surface with all 5 fingers of a hand.
  • FIG. 6 is a drawing illustrating command control location (e.g., finger contact regions) assigned to the touch sensitive surface based on the locations of the simultaneous contacts.
  • FIG. 7 is an illustrative drawing of a touch sensitive surface having both base finger contact regions and secondary finger contact regions.
  • FIG. 8 is an illustrative drawing of a touch sensitive surface having both base finger contact regions and secondary finger contact regions.
  • a base finger contact may and two secondary finger contact regions may be controlled by the same finger.
  • FIG. 9 is a drawing showing features of a process of assigning initial finger contact regions.
  • FIG. 10 is an illustrative drawing showing a contact of the touch sensitive surface in one of the finger contact regions.
  • FIG. 11A is an illustrative drawing showing a sliding motion of the contact with the touch sensitive surface in a generally rightward direction.
  • FIG. 11 B is an illustrative drawing showing a sliding motion of the contact with the touch sensitive surface in a generally leftward direction.
  • FIG. 1 1C is an illustrative drawing showing a sliding motion of the contact with the touch sensitive surface in a generally upward direction (e.g., within about 45 ° or within about 30 ° of the upward direction).
  • Fig. 12 is an illustrative drawing showing features of a process of entering a command by a gesture including contacting a finger contact region.
  • FIG. 13A is an illustrative drawing showing a contact with a finger contact region that is offset from the center point of the finger contact region.
  • FIG. 13B is an illustrative drawing showing a previous (e.g. initial) finger contact region and a new finger contact region.
  • the finger contact region may be repositioned based on the direction and/or distance between the previous (e.g., initial) center of the finger contact region and the location of contact recognized on the surface.
  • FIG. 14 is an illustrative drawing showing features of a process of repositioning a finger contact region.
  • FIG. 15 is an illustrative flow diagram for entering one or more control commands. It will be appreciated that some of the steps may be eliminated and/or the order of the some of the steps may be changed. The process may also include additional steps.
  • FIG. 16 is an illustrative drawing showing a touch sensitive surface facing away from a user.
  • FIG. 17 is an illustrative drawing showing a touch sensitive surface in a vehicle facing away from a user.
  • FIG. 18 is an illustrative drawing showing a touch sensitive surface in a motor vehicle facing towards a user. The ability to use "blind" control of this touch sensitive surface may reduce the level of distraction during its operation compared with a device requiring visual observation of control locations.
  • FIGs. 19A and 19B are drawings illustrating a device mounted to a steering wheel or a steering wheel column with a touch sensitive surface facing away from the driver.
  • FIGs. 20A, 20B, 20C, and 20D are drawings of an illustrative game controller having one or more touch sensitive surfaces (e.g., on a concave surface of the game controller).
  • touch sensitive surfaces e.g., on a concave surface of the game controller.
  • the method, systems, devices and apparatus for the control sessions according to the teachings herein are generally based on a processor identifying contacts with a touch sensitive surface, resulting in the selection of one or more control command.
  • a control session 2 typically includes a mode for initial positioning of control locations 4 (e.g., finger contact regions) on a touch sensitive surface 14.
  • the mode for initial positioning of control locations preferably is employed for positioning the control locations based on the locations of simultaneous contact at three or more spaced apart locations on the touch sensitive surface (e.g., by three or more fingers contacting the surface).
  • the control session 2 includes a mode for entry of control commands 6.
  • the mode for entry of control commands is employed for selecting one or more control commands 8 based on contact(s) with the control locations on the touch sensitive surface 14.
  • the control session may include a mode for ending the control session.
  • the control session may include a mode for repositioning control locations.
  • Such a repositioning mode may be employed to compensate for shifting of a user's hand relative to the touch sensitive surface.
  • the mode for the entry of control commands 6 may include the selection of control commands 7 and a mode for repositioning control locations 8, such as illustrated by FIG. 2.
  • the ending of a control session 9 may occur after the mode for the entry of control commands 6.
  • the processor is preferably in electronic communication with the touch sensitive surface for identifying positions of objects above or in contact with the touch sensitive surface, for identifying movement of objects above or the movement of contacts on the surface, and for identifying the removal of the objects away from the touch sensitive surface (e.g., removal of one or more contacts from the surface).
  • the processor positions or assigns the control locations based on the positions of contact by a user, thus enabling "blind" interaction by a use with the touch sensitive surface.
  • a touch sensitive surface is capable of identifying the location of multiple simultaneous contacts on the surface. Each contact preferably includes a sufficient force applied to the surface as required by the touch sensitive surface to recognize a contact.
  • the touch sensitive surface may be a flat surface, may be a curved surface, or may have regions that are flat and regions that are curved.
  • the touch sensitive surface is characterized as being generally smooth and or having a generally uniform texture.
  • the touch sensitive surface may be sufficiently smooth and/or have a sufficiently uniform texture so that a user cannot identify the location of contact of the surface based on the surface topography or other tactile clues on the surfaces.
  • the touch sensitive surface may be a surface of a pure entry component or device (i.e., a component or a device that does not display images), such as a touch pad, or may be a surface of a combination entry/display component or device, such as a touch-screen display.
  • the device including the touch sensitive surface and/or a processor connected to the device preferably is capable of recognizing each of multiple contacts to the surface, the maintaining of the contact, the movement (if any) of the contact, and the termination (i.e., removal) of the contact.
  • the touch sensitive surface 14 may have a first direction 82 (e.g., an upward direction) and an opposing second direction 84 (e.g., a downward direction).
  • the touch sensitive surface may have a third direction 86 (e.g., a rightward direction) and an opposing fourth direction 88 (e.g., a leftward direction).
  • the first and second directions 82, 84 may be generally orthogonal to the third and fourth 86, 88 directions.
  • the descriptions refer to the first, second, third and fourth directions, and may have alternate meanings depending on the actual orientation of the touch sensitive surface.
  • the upward and downward directions may in actuality refer to a forward direction and a backward direction respectively.
  • the initial positioning of the control locations typically requires the simultaneous contact of a touch sensitive surface at three or more points or regions.
  • the three or more points of contact preferably are consistent with the contact by three or more fingers of a hand or hands.
  • the number of simultaneous contacts required for the initial position of the control locations may be 3 or more, 4 or more, 5 or more, 6 or more, 7 or more, 8 or more, 9 or more, or 10.
  • the number of simultaneous contacts on the touch sensitive surface is a predetermined target number or a predetermined minimum number. For example, if there is a predetermined target number of four, the positioning of the control locations will not be completed until precisely four simultaneous contacts are recognized on the touch sensitive surface. In contrast, if the predetermined minimum number of contacts of four, the positioning of the control locations will not be completed until at least four (e.g., four, five, six, eight, or more) simultaneous contacts are recognized on the touch sensitive surface.
  • FIG. 3 illustrates the simultaneous contact of a touch sensitive surface 14 at three or more spaced apart locations 16 on a touch sensitive surface of a touch sensitive device 12 (e.g., a pure input device).
  • a touch sensitive device 12 e.g., a pure input device
  • the simultaneous contacts of the touch sensitive surface preferably are spaced apart on the surface in such a manner that sequential finger locations may be assigned.
  • the positions of the contacts may relate to a natural arch (or arches) formed by the fingertips of one hand.
  • the positions of the contacts 16 may be along an arch 17, such as a natural arch corresponding to three or more adjacent fingertips.
  • the process typically includes a step of a user removing some or preferably all of the fingers from the touch sensitive surface.
  • the processor after identifying the simultaneous contact of three or more spaced apart locations, will then identify the removal (i.e., the absence of contact) at the three or more spaced apart locations).
  • the processor recognizes the removal of all contact with the touch sensitive surface, such as illustrated in FIG. 4. It will be appreciated that one or more of the fingers may remain in contact with the touch sensitive surface after the initial positioning of the control locations (provided that at least one finger has been removed from the surface, and preferably at least two fingers have been removed from the surface). Such remaining finger(s) may be employed in directly entering a control compound as discussed herein, without the need to remove said finger(s) and re-establish contact with the surface.
  • the initial positioning of the control locations may include simultaneous contact with four or more fingers (for example with all the fingers of one hand).
  • the touch sensitive surface may be a touch screen display 15.
  • the locations of initial contact, regions including the location of initial contact, or the resulting control location 18 may be displayed on the touch sensitive screen display.
  • FIG. 6 illustrates control locations after being positioned in the mode of initial positioning of the control locations.
  • Each control location 18 is spaced apart from the other control locations.
  • Each control location may be characterized by a geometric center 22.
  • the position or region of contact 16 preferably is positioned within the control location 18.
  • the geometric center 22 is preferably within the position or region of contact 16.
  • the control locations may be a region having any shape.
  • the shape of a control location may be polygonal, circular, semi-circular, rectangular, oval shaped, elliptical, egg-shaped, or square.
  • each position on the touch sensitive surface is associated with at most one control location 18.
  • the number of control locations may be more than the number of fingers that are employed when initially establishing the locations of the finger contact regions.
  • the base finger contact region may include the point of contact of the finger when defining the finger contact regions.
  • Each secondary finger contact region (if any) for the first finger is sufficiently displaced from the base finger contact region for the first finger so that the base and secondary finger contact regions do not overlap . It will be appreciated that the locations of the various finger contact regions for a first finger may depend upon different amounts of curvature of the first finger.
  • the first finger may be relatively curved when contacting the base finger contact region and relatively less curved (e.g., more outstretched when contacting a secondary finger contact region). This may be analogous to typing a "d" (relatively curved) and an "e” (relatively outstretched) on a QWERTY keyboard.
  • the first finger may be relatively less curved when contacting the base finger contact region and relatively more curved when contacting a secondary finger contact region. This may be analogous to typing a "d" (relatively less curved) and a "c" (relatively more curved) on a QWERTY keyboard.
  • Examples of a touch screen surface including base finger contact regions 72 and secondary finger contact regions 74 are shown in FIG. 7 and FIG. 8.
  • FIG. 7 there are 5 base finger contact regions (one for each finger of a hand) and 3 secondary finger contact regions (one for each of three of the fingers).
  • FIG. 8 there is one base finger contact region 72 and two secondary finger contact regions 74 for each of three fingers.
  • Each finger contact region preferably is a different region of the touch sensitive surface.
  • any location on the touch sensitive surface may correspond to at most one finger contact region at a given time. It will be appreciated that some regions of the touch sensitive surface will correspond with no finger contact region.
  • the finger contact region associated with a given location may change over time. For example, a location p on the touch sensitive surface may be associated with a one finger contact region after a first stage of positioning finger contact regions and may be associated with a different finger contact region after a later stage of positioning finger contact regions.
  • a location p on the touch sensitive surface that was initially associated with that finger contact region may no longer be associated with it (instead, p may be associated with no finger contact region or may be associated with a different finger contact region).
  • FIG. 9 Features of a process for positioning initial control locations is illustrated in FIG. 9.
  • the process may be recognized from the actions of a user of a touch sensitive device, from the actions of a device or a processor, or both.
  • the actions of the user may include: a step of a user simultaneously contacting a touch sensitive surface with three or more fingers; and after contacting the touch sensitive surface, a step of the user removing the three or more fingers from the touch sensitive surface.
  • the actions of the device or processor may include a step of recognizing a simultaneous contact on the touch sensitive surface at three or more spaced apart contact point or contact regions (e.g., consistent with the contact by fingers of a user's hand(s)); followed by a step of recognizing the removal of the contacts from the touch sensitive surface and the positioning of control locations (e.g., finger contact locations) based on the locations contacted.
  • the positioning of the control locations may occur at any time after the device or processor recognizes the simultaneous contacts.
  • the positioning of the control location may be before the removal of the contacts.
  • the positioning of the control location may occur after recognizing the removal of the contacts.
  • the positioning of the control locations may be based on the initial position of contact, or a later (e.g., a final) position of contact in the event the contact moves during the mode of positioning the control locations.
  • the touch sensitive surface may be employed in a mode for entry of control commands.
  • the touch sensitive surface may be employed for controlling multiple devices and/or controlling multiple features or functions of the device(s).
  • the processor may need to identify which device / feature / function to control with a control command.
  • the selection of the control command(s) may be based on the control location or locations contacted during a mode for entry of control commands.
  • one or any combination of the following features of a contact with the surface may be identified and used by the processor: the number of contact locations being contacted, a sliding motion of a contact, a direction of motion of a contact, the length of time of a contact, the number of sliding motions of a contact, and the removal of a contact.
  • the process may include entering a control command by a user making one or more gestures on a touch sensitive surface and a processor recognizing the gesture(s) and selecting the control command based on a predetermined association between the gestures and the control commands.
  • one or more fingers remain in contact with the touch sensitive surface during the entry of the command.
  • the gesture may be any gesture which can be recognized by the processor.
  • a gesture begins by a contact at one or more of the control locations.
  • a user may start enter a gesture by contacting the touch sensitive surface with one finger beginning in a finger contact region and, while maintaining the contact, making one or more sliding motions with the finger including sliding the finger in at least a first direction, and then removing the one finger from the touch sensitive surface.
  • a single finger may contact a finger contact region, then slide in one or any combination of the following directions: upwards, downwards, rightwards, and left wards, and then removing the finger from the touch sensitive surface. It will be appreciated that after moving in a first direction, a finger may be moved in the reverse direction prior to removing the finger from the touch sensitive surface.
  • the contact with a second finger may be a sliding contact, a tapping contact, or a contact maintained at a single location.
  • a first finger maintains in a constant location on the touch sensitive surface while a second finger makes a dynamic contact with the touch sensitive surface.
  • the entry of the command preferably ends with the removal of the first finger from the touch sensitive surface.
  • the control of a device may include a plurality of gesture entries.
  • the process may require the entry of a first control command to select a device to control and a later entry of a control command to control a function of the device.
  • the distance of movement and / or the length of contact may be employed for determining the level or extent of control of a device.
  • movement in one direction may be employed to increase the speed continuously until the end of the gesture is recognized (e.g., by removing the contact or other entry completions according to the teachings herein).
  • the rate of increase of the speed may be related to the distance of movement of the gesture.
  • the volume of a device may be controlled by the gesture originating at a control location.
  • the movement of the contact in a first direction by a first movement distance may result in the volume increasing by at a rate that correlates with the first movement distance.
  • the increase in volume may continue until the contact is removed or the completion of the gesture is otherwise identified.
  • the decrease of the volume may be similarly be accomplished by a contact in a second direction different from the first direction (e.g., an opposing direction) by a second movement distance.
  • the entry of a gesture may include a contact 40 on the touch sensitive surface at a point or region 16 within a control location 18 after the control locations have been established.
  • the point or region of contact 16 may be offset from the center of the control location 22, such as shown in FIG. 10.
  • the point or region of contact may include the center of the control location (not shown).
  • the gesture may include a movement of one or more of the contacts.
  • the movement may be a sliding movement in one or more sliding directions.
  • the gesture may include a movement of the contact 16 on the touch sensitive surface in a generally rightward direction (FIG. 1 1 A), in a generally leftward direction (FIG. 11 B), in a generally upward direction (FIG. 1 1C), or in a generally downward direction.
  • a sliding movement may include movement in multiple directions and/or movement is non-linear.
  • the entering of a command using a gesture may include multiple contacts each with a different control location. Each contact may include a movement in the same sliding direction, may include movements in different direction, or may include both contacts that are stationary and contacts that move.
  • multiple contacts during the entry of a command have simultaneous sliding motions (that are the same or different).
  • the gesture may include sliding motion of two contacts towards each other, sliding motion of two contacts away from each other, or sliding contacts that are in generally the same direction.
  • a sliding motion may be a generally small motion (e.g., within a control locations), a generally large motion (e.g., outside the control location or larger than a distance from the center to the edge of the control location).
  • the association between commands and gestures may include a first command associated with the sliding motion of only a first finger in a first direction and a second different command associated with the sliding motion of only a second finger in a second direction.
  • the first and second directions may be the same (e.g., within about 30 °, within about 20 °, within about 15 °, within about 10 °, or within about 5 °) or different (e.g., angled by more than 30 °, angled by more than 40 °, or angled by more than 50 °).
  • the entry of a control command using a gesture may include one or more of the features of FIG. 12.
  • a user may contact one or more of the control locations each with a different finger and then make a gesture on the touch sensitive surface while maintaining contact with the surface.
  • the gesture is preferably associated with a predetermined control command.
  • the process preferably includes a step of removing the fingers from the touch sensitive surface or other action for completing the entry of the gesture. After the completion of the entry of a gesture, another gesture may be entered.
  • the processor or device generally identifies contact(s) with one or more control locations (e.g. after the control locations have been free of contact). The process or device then identifies the type of gesture(s) being made from the contacts.
  • the process or device may identify the gesture as a continuous contact with the surface. After identifying a gesture, the processor or device may identify a predetermined control command associated with the entered gesture. Preferably, the control command is executed, sent, or otherwise acted upon. It will be appreciated that a control command may be associated with a single gesture or with a series of gestures. The processor or device may identify a gesture immediately upon contact, after one or more sliding motions, upon the removal of the contact from the surface, or after the completion of the entry of a gesture according to the teachings herein.
  • Each gesture entry includes a contact with one or more control locations (e.g., finger contact regions).
  • the entry of a gesture may be completed by removing all of the fingers from the touch sensitive surface.
  • the entry of a gesture may be completed (e.g., timed out by a processor) after a predetermined time limit is exceeded from the contact with the control location(s).
  • a predetermined time limit of about 0.5 seconds or more, about 1 second or more, about 1.5 seconds or more, about 2 seconds or more, about 3 seconds or more, or about 4 seconds or more.
  • the predetermined time limit typically is about 100 seconds or less, about 30 seconds or less, about 15 seconds or less, about 10 seconds or less, or about 6 seconds or less.
  • the entry of a gesture may be completed (e.g., timed out by a processor) after a predetermined limit for the number of changes in direction of a sliding motion (i.e., the predetermined direction change limit) is reached.
  • a predetermined limit for the number of changes in direction of a sliding motion i.e., the predetermined direction change limit
  • the multiple sliding directions generally occur while the contact with the touch sensitive surface is maintained.
  • a change in direction may be a change of about 15° or more, about 45° or more, about 90° or more, about 135° or more, or about 180°.
  • the gesture may be completed after the processor identifies an initial sliding motion and then a first change in the direction of the sliding (i.e., the predetermined direction change limit is one).
  • the gesture may be completed after the processor identifies a sliding motion in an initial direction, followed by a sliding motion in a second direction and then a sliding motion in a third direction (i.e., the predetermined direction change limit is two).
  • the predetermined limit may be one or more, two or more, three or more, or four or more.
  • the predetermined direction change limit will be 10 or less, or 5 or less.
  • the entry of a gesture may be completed (e.g., timed out by a processor) when the contact moves in a predetermined direction.
  • the predetermined direction may be any direction.
  • the predetermined direction may be a generally upward direction, a generally downward direction, a generally rightward direction, or a generally leftward direction.
  • the entry of a gesture may be completed (e.g., timed out by a processor) when the contact returns to the finger contact region (or a position in the finger contact region) following a sliding motion away from the finger contact region.
  • the entry of a gesture may be completed (e.g., timed out by a processor) when the contact moves in a predetermined shape.
  • the entry of the gesture may be completed when the processor recognizes a sliding motion in a shape such as an arc, a semicircle, a circle, a triangle, a rectangle, a square, a star, a letter, a number, or any combination thereof.
  • a next gesture may be entered by contacting one or more control location(s).
  • the mode for entry of control commands may include a mode for repositioning of the control locations. Instead of having a mode for repositioning of control locations, the control locations may remain fixed.
  • the touch sensitive surface may be contacted at a point or region of contact 16 inside a control location 18 but offset from the center 22 of the control location 18, such as illustrated in FIG. 13A.
  • the processor or device may reposition the control location based at least partially on the offset distance and offset direction between the center of the contact 15 and the center 22 of the control location 18.
  • the repositioning may occur every time the contact is offset from the center of the control location, or only under certain conditions. For example, the offset distance must reach a threshold value prior to repositioning the control location. As another example, the offset must occur at a sufficient frequency prior to repositioning the control location.
  • the new control location may have the same size, or a different size as the prior control location.
  • the new control location may have the same shape, or a different shape as the prior control location.
  • the new control location preferably has the same size or the same shape as the prior control location. More preferably, the new control location has the same size and shape as the prior control location.
  • FIG. 13B illustrates the repositioning of a control location, showing the position of the new location and the position of the prior location.
  • the new control location 36 may have a center 38 that is displaced from the center 22 of the prior control location 18.
  • the direction of the displacement 32 preferably is about the same as the offset direction.
  • the distance of the displacement between the prior center 22 and the new center 38 preferably is about the same as or less than the offset distance.
  • the distance between the prior center 22 and the new center 38 may be a percentage (preferably about 100% or less than 100%) of the offset distance.
  • offset distance and offset direction for one contact location may be employed for repositioning one or more of the other contact locations.
  • the reposition of the finger contact region may include one or any combination of the steps illustrated in Fig. 14.
  • the repositioning of a control location generally occurs after a contact is initiated in a control location.
  • the repositioning of the control location may occur prior to a sliding motion of the contact, after a sliding motion of the contact, or after removal of the contact from the touch sensitive surface.
  • the process may allow for repeated or continuous entry of different control commands. For example, after identifying the end of the entry of a control command, the process may be used for the entry of a subsequent control command, such as illustrated in FIG. 15. It will be appreciated that the step of the removal of the contact may be replaced by a step of identifying the end of an entry of a control command, such as described herein. It will also be appreciated that such a step of identify the end of an entry of a control command may occur (e.g., recognizing the removal of the finger contact from the touch sensitive surface) may occur after a step of identifying a command based on the gesture.
  • the touch sensitive surface may be oriented in a direction and/or location that prevents viewing of the touch sensitive surface by a user.
  • a user may be holding a device including a touch sensitive surface so that the touch sensitive surface is facing away from the user.
  • the user 62 may be holding the device 12, so that the touch sensitive surface 14 is directed away from the user's eye's 64.
  • the screen direction 52 perpendicular to and away from the touch sensitive surface 14 may partially or entirely be in a forward direction.
  • the screen display direction may be away from the user's eyes 64.
  • the user 62 may be a driver of a vehicle 66 and the touch sensitive device 14 may be mounted to the vehicle 66.
  • the touch sensitive surface is mounted so that the touch sensitive surface faces away from the eyes of the driver.
  • the touch sensitive surface may be oriented within the field of view of the user. However, operation and/or control of the device may be enhanced by setting control locations based on contact with the touch sensitive surface with three or more fingers. For example, as illustrated in FIG. 18, the touch sensitive surface may be in the field of view of the driver of a vehicle.
  • the touch sensitive surface may be mounted to, attached to, or integrated with a device positioned for contact by a user.
  • the touch sensitive surface may be located in a vehicle in proximity to a driver, and preferably at an ergonomic location.
  • the touch sensitive surface may be mounted to a steering wheel and/or a steering column, such as illustrated in FIG. 19A and 19B.
  • the devices and systems according to the teachings herein may include one or more view panels (e.g., a display panel that is not a touch sensitive display) for viewing the command functions available.
  • a display panel may be on the same device as the touch sensitive surface, but on a different location.
  • a display panel and a touch sensitive surface may be on opposing sides of a device.
  • a display panel may be on a different device as the touch sensitive surface.
  • a touch sensitive panel may be mounted on a steering wheel and a display panel may be a panel attached to or integrated with a dashboard of a vehicle.
  • a display panel 68 may be a display of a mobile phone, or a vehicle display device.
  • a view panel display may fade or turn-off after a pre-determined interval of inaction on the touch sensitive surface.
  • the device or system including the touch sensitive surface may include a control component for turning the device on or off, or for resetting the device.
  • Control component may be a switch or other component capable of executing and/or communicating an on, off, or reset function.
  • the device or systems may include one or more features for disabling the touch sensitive surface.
  • a disabling feature may be particularly useful in a vehicle when it may be desirable to disable the operation of the device and systems based on vehicle operating conditions.
  • the touch sensitive surface may be disabled when the vehicle is turning and/or has recently changed directions, is driving above certain speeds, etc.
  • the touch sensitive devices may include a grip for spreading the fingers towards different locations on the touch sensitive surface.
  • control commands may be employed for controlling an air conditioning, a radio, a window, a light, a lock, a cruise control, an application on a mobile phone, a navigation control, the location of a cursor, a mechanical device, an electronic device, operation of a land vehicle, operation of a water vehicle, operation of an air vehicle, remote operation of a vehicle or other device, a communication device, or any combination thereof.
  • a control command includes providing an API or code to enable two computer systems to interact, such as a smart phone and an automobile computer system.
  • the passcode may consist of a sequence of gestures each employing one or more contacts with the finger contact regions.
  • the touch sensitive surface is simultaneously contacted with three or more fingers for assigning the finger contact regions based on the locations of the contacts.
  • the user may then enter a passcode by contacting the finger contact regions for entering a series of gestures.
  • the passcode may be a series of 2 or more gestures, 4 or more gestures, 6 or more gestures, or 8 or more gestures.
  • Each gesture may be the same or different from the previous gesture.
  • Each gesture may employ the same or different finger contact regions as the previous gesture.
  • Each gesture may employ the same number of fingers or a different number of fingers as the previous gesture.
  • the passcode may include one gesture requiring the sliding movement of one, two, or more fingers in the same direction, the passcode may include the sliding movement of two fingers towards each other, the passcode may include the simultaneous tapping of one, two, or more finger contact regions, or any combination thereof.
  • a passcode may be entered by first contacting the touch sensitive surface simultaneously with a first, second, third, and fourth finger for assigning a finger contact region for each of the four fingers, followed by a first gesture entry of a sliding the first finger to the right, followed by a second gesture of sliding the third and fourth fingers together, followed by a third gesture of sliding the first, second, and third fingers in an upward direction, followed by a fourth gesture of tapping the surface with the fourth finger.
  • the large number of possible gestures will result in more secure passcodes and/or reduce the number of entries required to obtain a secure passcode.
  • the passcode can be entered at different locations on the touch screen surface, reducing the possibility that evidence of the touch passcode will remain on the screen after repeated entries of the passcode over time.
  • Preferred passcodes include two or more sequential gestures in different directions.
  • Preferred passcodes include two or more sequential gestures using different finger contact regions or different combinations of finger contact regions.
  • the processor may recognize the sequential contact of the touch sensitive surface at three or more locations, consistent with the contacting of the surface with three or more fingers and then assigning the finger contact locations based on the locations of contact.
  • the number of simultaneous contacts for assigning the finger contact regions may be a predetermined number, such as 3, 4, 5, 6, 7, 8, 9, or 10.
  • the processor may monitor the surface for a contact and gesture consistent with a first gesture of a predetermined passcode.
  • the system may include a mode of setting or resetting a passcode.
  • the setting or resetting of the passcode may include simultaneously contacting the touch sensitive surface with three or more fingers for assigning the finger contact regions, followed by the sequential entry of the gestures of passcode.
  • the system may require re-entry of the passcode to confirm the passcode.
  • the system may store the passcode.
  • the passcode may be encrypted by the system.
  • the system may delete prior passcodes.
  • the touch sensitive surface may be attached to a vehicle.
  • the touch screen surface in a vehicle may be proximate the driver's seat and/or the driver of the vehicle.
  • a touch screen surface for use by a passenger may be proximate the location of one or more passenger seats and/or one or more passengers of a vehicle.
  • the touch screen device in a vehicle may be connected to the steering wheel, the steering wheel column, the dashboard, a seat back, a pillar, a door, or any combination thereof.
  • the touch screen surface may be in communicative connection with one or more devices to be controlled.
  • the touch screen surface may be in communicative connection with a control processor of the vehicle.
  • the touch sensitive surface may be oriented face down (out of view of the driver or other user).
  • the touch sensitive surface may be part of a device or connected to a device adapted to track movements by multiple individually identified fingers.
  • the touch sensitive surface may be positioned so that a vehicle driver can contact the surface with one or more fingers without removing his hands from the steering wheel.
  • the touch sensitive surface may be associated with an apparatus software (e.g., an application) that identifies and/or tracks multiple contacts of the surface.
  • the apparatus software may be running on a computer system.
  • the apparatus software may be adapted to identify finger generated contact, finger generated gestures or both.
  • the apparatus software may be adapted to identify multiple (preferably three or more, four or more, or five or more) simultaneous finger generated contact, multiple finger generated gestures, or both.
  • the data entry device may be a device for controlling the operation of an application, a processor, a connected device or a remote device.
  • the data entry device may be a device for controlling (e.g., remotely controlling) a video game, a machine, a vehicle, a flying device, a robotic device, or any combination thereof.
  • the data entry device may be a game controller.
  • a handheld device, such as a game controller may have one or any combination of the features illustrated in FIG. 20A and FIG. 20B. With reference to FIG. 20A and FIG.
  • the handheld device 110 may have a forward surface 1 12 and an opposing rearward surface 1 16.
  • the forward surface may include one or more touch sensitive surfaces 14.
  • the forward surface 1 12 of the device 1 10 may include a touch sensitive surface for the fingers of a user's right hand and a second touch sensitive surface for the fingers of a user's left hand. It will be appreciated that the forward facing surface may have a single touch sensitive surface sufficiently large for receiving simultaneous contacts from fingers of both hands.
  • the touch sensitive surface 14 on the forward facing surface 1 12 of the device preferably is sufficiently large for contacting with three or more spaced apart fingers of a hand.
  • the device preferably has a sufficient number and size of touch sensitive surfaces for contact with 4 or more fingers, 6 or more fingers or 8 or more fingers.
  • the rearward surface may include one or more thumb controls, such as a button, a knob, a dial, a joystick, or a rollerball capable of being controlled by a thumb, while the fingers rest on the forward surface.
  • the device 1 10 may include one or more (e.g., two or more) thumb controls for a left hand, one or more (e.g., two or more) thumb controls 1 18 for a right hand, or both.
  • the device has side surfaces 1 14 connecting the forward and rearward surfaces.
  • the side surfaces 1 14 preferably are adapted for receiving a palm of a hand.
  • the side surfaces 114 may be rounded or otherwise curved.
  • the device 110 may include a gripping feature for assisting in the placement of the fingers and/or making the device easier to grip.
  • An example of a gripping feature 120 is shown in FIG. 20C.
  • the forward surface 112, 1 12', 1 12" may be generally concave, generally planar, or generally convex.
  • the forward surface 112" is generally convex
  • the rearward surface is generally concave, such as illustrated in FIG. 20D, resulting in a more natural placement of the side surfaces 1 14 between the thumb and fingers.
  • one or more of the contacts of the touch sensitive surface may be replaced by a sensing of the locations of the multiple fingers.
  • the locations of the fingers may be identified by light (e.g., laser light) or other forms of radiation, electrical fields, magnetic fields, darkness level (e.g., a shadow), or any combination thereof.
  • a glove or other device may be placed on one or more fingers to enable the identification of the location of the finger and/or to enhance the aforementioned observation of the location of the finger.
  • a gesture on the surface of the device may be sensed by one of the aforementioned means, with or without actual contact with the surface.
  • any such sensing occurs while the finger is at least near the touch sensitive surface (e.g., about 30 mm or less, about 10 mm or less, about 3 mm or less, or about 1 mm or less from the surface).
  • a sensing may be completed when the finger moves away from the surface.
  • the end of a mode or entry of a command may be completed when it is sensed that a finger is no longer near the touch sensitive surface (e.g., further than 1 mm, 3 mm, 10 mm, or 30 mm from the surface, or when the distance from the surface has increased by at least about 1 mm, about 3 mm, about 5 mm, about 10 mm, or about 20 mm).
  • a finger is no longer near the touch sensitive surface (e.g., further than 1 mm, 3 mm, 10 mm, or 30 mm from the surface, or when the distance from the surface has increased by at least about 1 mm, about 3 mm, about 5 mm, about 10 mm, or about 20 mm).
  • One or more of the aforementioned modes may be completed (e.g., timed out by a processor) after a predetermined time limit is exceeded from one or any combination of the following: i) one or more contacts with the touch sensitive surface; ii) the sensing of one or more objects (e.g., fingers) above the touch sensitive surface); iii) the removal of one or more contacts from the touch sensitive surface; or iv) the sensing of the movement of one or more objects (e.g., fingers) away from the touch sensitive surface.
  • the predetermined time limit if any, is about 0.5 seconds or more, about 1 second or more, about 1.5 seconds or more, about 2 seconds or more, about 3 seconds or more, or about 4 seconds or more.
  • the predetermined time limit if any, is about 100 seconds or less, about 30 seconds or less, about 15 seconds or less, about 10 seconds or less, or about 6 seconds or less.
  • a predetermined time limit may be fixed, may be adjusted (e.g., based on experience or historical values, or set by a user).
  • the transition from a mode for initial position of control locations to a mode for entry of control commands may be triggered by a trigger event.
  • the trigger event may be a predetermined time limit, such as described herein.
  • the trigger event may be the sensing of the removal of one or more objects (e.g., fingers) away from the touch sensitive surface.
  • the trigger event may require the processor to sense that exactly one of the objects (e.g., one finger) remains on or near the touch sensitive surface while the other objects are moved away from the touch sensitive surface.
  • any numerical values recited herein include all values from the lower value to the upper value in increments of one unit provided that there is a separation of at least 2 units between any lower value and any higher value.
  • the amount of a component or a value of a process variable such as, for example, temperature, pressure, time and the like is, for example, from 1 to 90, preferably from 20 to 80, more preferably from 30 to 70, it is intended that values such as 15 to 85, 22 to 68, 43 to 51 , 30 to 32 etc. are expressly enumerated in this specification. For values which are less than one, one unit is considered to be 0.0001 , 0.001 , 0.01 or 0.1 as appropriate.
  • control locations e.g., initial finger contact locations
  • First direction on touch sensitive surface e.g., upward direction
  • Second direction on touch sensitive surface e.g., downward direction

Abstract

Disclosed is a method, device, and system for entry of command using a touch sensitive surface. Instead of the device providing predefined locations for entering different commands, the device identifies the locations of three or more of the user's fingers or other objects. After determining the locations of the fingers, an entry mode allows for the entry of one or more commands based on an association between the commands and movements of different fingers and the type or direction of movement. The association may include a first command associated with the sliding motion of only a first finger in a first direction and a second different command associated with the sliding motion of only a second finger in a second direction. The first and second directions may be the same (i.e., within about 30 °, within about 20 °, within about 15 °, within about 10 °, or within about 5 °) or different (e.g., angled by more than 30 °, angled by more than 40 °, or angled by more than 50 °). Preferably the command entry mode is triggered by a trigger event. Preferably one of the fingers remain on the touch sensitive surface from the identification of the locations of the fingers through the entry of the command.

Description

DEVICE, SYSTEM, AND METHODS FOR ENTERING
COMMANDS OR CHARACTERS USING A TOUCH SCREEN
CLAIM OF PRIORITY
[01] The present application claims priority to U.S. Provisional Patent Applications 62/207,564 filed on August 20, 2015 and 62/266,916, filed on December 14, 2015. U.S. Patent Applications 14/976,005 and13/091 , 158 and U.S. Provisional Patent Applications 62/207,564 and 62/266,916 are all incorporated herein by reference in their entirety.
FIELD
[02] The present teachings are directed at devices, systems, and methods for inputting one or more commands using a touch sensitive surface for control of an electronic device. Preferably, a positioning mode is employed for positioning control locations (e.g., finger contact regions) based on the contact of the touch sensitive surface with three or more fingers. BACKGROUND
[03] The ability to control devices using touch screen displays has often resulted in more intuitive and / or simplified operation of the device. However, in many circumstances there is a need to control a device when it is not safe, not possible or otherwise not efficient to see the touch screen device. In such circumstances, the user is essentially operating blind. One approach is to use textural clues to identify various control locations. As an example, typing on a keyboard is often facilitated by a surface marking on one or more character keys for positioning fingers and reducing errors in typing, particularly when typing blind (i.e., without looking at the results of the typing). However, typical touch screens do not have textural cues on the surface of screen. Instead, the typical touch screen display relies on the user's vision to identify and locate control locations. Blind typing on a touch screen display is difficult. Indeed, many individuals desiring to send messages on a mobile phone while having the phone hidden under a table or in a pocket, prefer phones having individual keys over a phone having a touch screen display. When touching keys, the operator receives continually feedback regarding the locations of the various keys (i.e., the control locations), such as by feeling spaces between keys, or by feeling a key having a different shape than a neighboring key.
[04] Although touch screen display devices can be used as a control device, there is a need for improved methods and systems so that a touch screen device can be used as a control device in circumstances when it is required to operate blind.
[05] Instead of giving textual clues, a touch screen device typically provides visual clues to identify the position of control locations on the surface. [06] Typical control locations are defined by an application. Although the operator may be able to move the control locations, such as by dragging a window, such movements typically requires an operator to first visually identify an initial location of the control location.
[07] There are some applications which enable the entry of controls without requiring a control location to be identified. For example, when viewing photographs, various applicants allow the movement of two fingers at any location on the touch screen for enlarging, reducing, rotating, or shifting the location of the image. Here the number of possible controls is limited because the same operation occurs irrespective of which two fingers are being used and because there is no control location identified.
[08] There is a need for methods, devices and systems that allow for control locations to be defined when a user contacts a touch screen surface instead of requiring the user to identify the position of existing control location(s) (e.g., through visual or tactile clues).
SUMMARY
[09] Instead of requiring the user to identify the control locations (e.g., through visual clues), the control methods according to the present teachings allow for the control locations to be defined when the user contacts the touch screen surface, at the same time with 3, 4, or 5 fingers of a hand. Here, instead of having the user apply visual or tactile clues to recognize the control locations, the device moves or identifies (i.e., defines) the control locations based on the contact locations of the user's fingers. After contacting the touch sensitive surface with the 3, 4 or 5 fingers of a hand, some or all of the fingers may be removed and the control locations may be maintained until the user makes an input contact with the surface with 1 , 2, 3, 4, or 5 fingers for controlling a device.
[010] Disclosed is a method of entering a command comprising the steps of: i) a processor connected to a touch sensitive surface sensing a simultaneously positioning of three or more objects (e.g., fingers) above and near (e.g., within 10 mm of) or on the touch sensitive surface at three or more different sensing locations (e.g., consistent with the positioning by three or more fingers of a user), including a positioning centered at a first finger initial sensing point (e.g., by a first finger), a positioning centered at a second finger initials sensing point (e.g., by a second finger), and a positioning centered at a third finger initial sensing point (e.g., by a third finger); ii) the processor assigning a finger location region for two or more (e.g., each) of the three or more objects, wherein each finger location region is a distinct region of the touch sensitive surface, and each finger location region includes one of the initial finger sensing points (e.g. wherein each finger location region corresponds to one of the fingers); iii) the processor entering a command entry mode following the step of assigning the finger location regions (e.g., following a predetermined event such as the removal of at least one finger, or following a predetermined time interval), wherein the command entry mode includes an association having at least a first command associated with a movement of only one of the objects starting at a first finger location region and a second different command associated with a movement (e.g., the same movement or a different movement as employed with the first command) of only one of the objects starting at a second finger location region; and iv) the processor recognizing a gesture on the touch sensitive surface including sensing the movement of only one of the objects starting at the first finger location region and identifying the associated first command based on the gesture.
[01 1] Also disclosed is a method of entering a command comprising the steps of: identifying a simultaneously contacting on a touch sensitive surface at three or more locations (e.g., consistent with the contact by three or more fingers of a user), including a contact centered at a first finger initial contact point (e.g., by a first finger), a contact centered at a second finger initial contact point (e.g., by a second finger of a second finger), and a contact at a third finger initial contact point (e.g., by a third finger); assigning a control location (e.g., a finger contact region) for each of the three or more fingers, wherein each finger contact region is a different region of the touch sensitive surface, and each finger contact region includes one of the initial contact points (e.g. wherein each finger contact region corresponds to one of the fingers); and recognizing the removing of the contact of the touch sensitive at the three or more locations (e.g., the removal of the three or more fingers from the touch sensitive surface). After control locations are assigned and removal of the contacts are sensed, the process may include one or more steps of entering a control command by contacts (i.e., entry contacts) with one or more control locations. For example, the process may include recognizing the simultaneous entry contact(s) in one or more of the finger contact regions (e.g., the simultaneously contacting of one or more of the finger contact regions with the corresponding finger) each at an entry contact point; recognizing a gesture on the touch sensitive surface including a movement of the entry contact in one or more directions on the touch sensitive surface starting from the entry contact point while continuous entry contact is maintained; (e.g., from a motion on the touch sensitive surface by the user making a gesture on the touch sensitive surface with one or more of the fingers in contact with touch sensitive surface, wherein each gesture originates in the corresponding finger contact region); and identifying a command to execute based on the control locations contacted and the gestures originating in the control locations.
[012] Further disclosed is a method of entering a command using a processor connected to a touch sensitive surface comprising the steps of: (i) the processor identifying a simultaneously contact on the touch sensitive surface at three or more contact locations (e.g., consistent with the contact by three or more fingers of a user), including a contact centered at a first finger initial contact point (e.g., by a first finger), a contact centered at a second finger initial contact point (e.g., by a second finger of a second finger), and a contact centered at a third finger initial contact point (e.g., by a third finger);(ii) the processor assigning a finger contact region for each of the three or more fingers, wherein each finger contact region is a different region of the touch sensitive surface, and each finger contact region includes one of the initial contact points (e.g. wherein each finger contact region corresponds to one of the fingers); (iii) the processor recognizing the removing of the contact of the touch sensitive surface at one or more (e.g., all) of the contact locations (e.g., the removal of the one, two, three or all of the fingers from the touch sensitive surface); (iv) following the step of recognizing the removing of contact(s) the processor recognizing the simultaneous entry contact(s) in one or more of the finger contact regions (e.g., the simultaneously contacting of one or more of the finger contact regions with the corresponding finger) each at an entry contact point, and/or the processor recognizing one or more contacts from step (i) each remaining on a finger contact region and recognizing the remaining contacts as finger entry contact(s); (v) the processor recognizing a gesture on the touch sensitive surface including a movement of the entry contact in one or more directions on the touch sensitive surface starting from the entry contact point while continuous entry contact is maintained; (e.g., from a motion on the touch sensitive surface by the user making a gesture on the touch sensitive surface with one or more of the fingers in contact with touch sensitive surface, wherein each gesture originates in the corresponding finger contact region); and (vi) the processor identifying a command to be execute based on the finger contact regions that are contacted and the gestures originating in the finger contact regions.
[013] Additionally disclosed is a method of entering a control command comprising the steps of: simultaneously contacting a touch sensitive surface with three or more fingers pf a user, including a contact of a first finger centered at a first finger initial contact point, a contact of a second finger centered at a second finger initial contact point, and contact of a third finger at a third finger initial contact point;. assigning a finger contact region for each of the three or more fingers, wherein each finger contact region is a different region of the touch sensitive surface, and each finger contact region includes one of the initial contact points, wherein each finger contact region corresponds to one of the fingers; removing the three or more fingers from the touch sensitive surface; contacting one of the finger contact regions with the corresponding finger and sliding said finger in one or more finger movement directions. For example, the sliding of finger may control a device, a component of a device, a communication, a display, or any combination thereof. By way of example, the movement direction of the finger may result in the moving of a cursor in a corresponding one or more cursor movement directions.
[014] Also disclosed is a method of entering a command comprising the steps of: simultaneously contacting a touch sensitive surface with three or more fingers of a user, including a contact of a first finger centered at a first finger initial contact point, a contact of a second finger centered at a second finger initial contact point, and contact of a third finger at a third finger initial contact point; assigning a finger contact region for each of the three or more fingers, wherein each finger contact region is a different region of the touch sensitive surface, and each finger contact region includes one of the initial contact points, wherein each finger contact region corresponds to one of the fingers; and removing the three or more fingers from the touch sensitive surface. After establishing finger contact regions based on the positions of contact by the user, one or more gestures may be entered using the finger contact regions. The entry of a gesture may include simultaneously contacting one or more of the finger contact regions with the corresponding finger; making a gesture on the touch sensitive surface with one or more of the fingers in contact with touch sensitive surface, wherein each gesture originates in the corresponding finger contact region; and identifying a control command to execute based on the finger contact regions that are contacted and the gestures originating in the finger contact regions.
[015] The control locations (e.g., finger location regions or finger contact regions) may be static or dynamic. For example, a static finger contact region may remain fixed while in a command entry mode, and may change only after ending a command entry mode and initiating a new positioning mode.
[016] Preferably the control locations are finger location regions. Finger location regions typically are spaced apart in an arc, such as expected from the spacing of finger tips. Although the finger location regions may be identified by the sensing of objects near the touch sensitive surface (e.g., fingers slightly above the surface, typically within 10 mm of the surface), the finger location regions preferably are finger contact regions, where the objects contact the surface, while near the touch sensitive surface,
[017] Preferably, the control locations (e.g., finger location regions or finger contact regions) are dynamic. For example, after one or more contacts during a command entry mode, a finger location region may be repositioned based on the actual location of finger position or contact (e.g., within a control location) during the command entry mode.
[018] A first finger location region (e.g., finger contact region) may be characterized by an initial contact area and an initial center point (e.g., a geometric center of the initial contact area). During a command entry mode, the touch surface may be contacted in the initial contact area, but the center of the contact may be at a contact point offset from the initial center point. The center point for the first finger contact region may then be shifted at least partially towards (e.g., entirely to) the contact point, so that the first finger contact region is characterized by a new center point different from the initial center point. In addition to shifting the center point of the first finger contact region, the offset between a center point and a center of contact in one finger contact region may also be employed for repositioning one or more different finger contact regions. Such repositioning may be employed to compensate for gradual shifting of the position of a hand on a touch sensitive surface. As such, the device adjusts to the user's hand position instead of the user adjusting the hand to a fixed position of a device. [019] Another aspect of the teachings herein is directed at a system for entry of control commands for controlling a device comprising: an entry device including a touch sensitive surface; a processor; a memory storing instruction that, when executed by the processor causes the processor to: identify a simultaneously contact on the touch sensitive surface at three or more locations (e.g., consistent with the contact by three or more fingers of a user), including a contact centered at a first finger initial contact point (e.g., by a first finger), a contact centered at a second finger initial contact point (e.g., by a second finger of a second finger), and contact at a third finger initial contact point (e.g., by a third finger); assign a control location (e.g., a finger contact region) for each of the three or more fingers, wherein each control location is a different region of the touch sensitive surface, and each control location includes one of the initial contact points (e.g. wherein each finger contact region corresponds to one of the fingers); and recognize the removing of the contact of the touch sensitive at the three or more locations (e.g., the removal of the three or more fingers from the touch sensitive surface) before or after the step of assigning the control locations. The memory storing instruction, when executed, after the assignment of the control locations may also cause the processor to: recognize the simultaneous entry contact(s) in one or more of the control locations (e.g., the simultaneously contacting of one or more of the finger contact regions with the corresponding finger) each at an entry contact point; recognize a gesture on the touch sensitive surface including a movement of the entry contact in one or more directions on the touch sensitive surface starting from the entry contact point while continuous entry contact is maintained; (e.g., from a motion on the touch sensitive surface by the user making a gesture on the touch sensitive surface with one or more of the fingers in contact with touch sensitive surface, wherein each gesture originates in the corresponding finger contact region); and identify a command to be execute based on the control locations that are contacted and the gestures originating in the control locations.
[020] In another aspect, the teachings herein is directed at a machine readable storage medium containing instructions that when executed cause a processor of an electronic device to discern input control commands by: the processor identifying a simultaneously contact on the touch sensitive surface at three or more locations (e.g. , consistent with the contact by three or more fingers of a user), including a contact centered at a first finger initial contact point (e.g., by a first finger), a contact centered at a second finger initial contact point (e.g., by a second finger of a second finger), and contact at a third finger initial contact point (e.g., by a third finger); the processor assigning a control location (e.g., a finger contact region) for each of the three or more finger initial contact points, wherein each control location is a different region of the touch sensitive surface, and each control location includes one of the initial contact points (e.g. wherein each finger contact region corresponds to one of the fingers); the processor recognizing the removing of the contact of the touch sensitive at the three or more locations (e.g., the removal of the three or more fingers from the touch sensitive surface); the processor recognizing the simultaneous entry contact(s) in one or more of the finger contact regions (e.g., the simultaneously contacting of one or more of the finger contact regions with the corresponding finger) each at an entry contact point; the processor recognizing a gesture on the touch sensitive surface including a movement of the entry contact in one or more directions on the touch sensitive surface starting from the entry contact point while continuous entry contact is maintained; (e.g., from a motion on the touch sensitive surface by the user making a gesture on the touch sensitive surface with one or more of the fingers in contact with touch sensitive surface, wherein each gesture originates in the corresponding finger contact region); and the processor identifying a command to be execute based on the control locations contacted and the gestures originating in the control locations.
[021] In yet another aspect, the teachings herein are directed at methods, systems and devices for unlocking one or more electronic device controls by a processor comprising: receiving an indication of simultaneous contact at multiple locations on a touch sensitive surface; determining that the multiple locations have a spacing consistent with contact of the touch sensitive surface by one or two hands of a user; assigning multiple finger contact regions based on the multiple locations of simultaneous contact; receiving indications of a sequential series of contacts each with one or more of the finger contact regions; comparing the sequential series of contacts with a predetermined sequential series (i.e., a password series), and unlocking one or more controls when the sequential series of contacts matches the predetermined sequential series.
BRIEF DESCRIPTION OF DRAWINGS
[022] FIG. 1 is a drawing showing illustrative modes for a control session according to the teachings herein.
[023] FIG. 2 is a drawing showing illustrative modes for a control session according to the teachings herein.
[024] FIG. 3 is a drawing illustrating simultaneous contact on a touch sensitive surface at three or more locations (e.g., corresponding to the touching by three or more fingers of a hand).
[025] FIG. 4 is a drawing illustrating a touch sensitive surface showing the contacts removed (e.g., the fingers removed from the surface) and showing the locations of the prior simultaneous contact.
[026] FIG. 5 is a drawing illustrating simultaneous contact of a touch sensitive surface with all 5 fingers of a hand.
[027] FIG. 6 is a drawing illustrating command control location (e.g., finger contact regions) assigned to the touch sensitive surface based on the locations of the simultaneous contacts. [028] FIG. 7 is an illustrative drawing of a touch sensitive surface having both base finger contact regions and secondary finger contact regions.
[029] FIG. 8 is an illustrative drawing of a touch sensitive surface having both base finger contact regions and secondary finger contact regions. For example, a base finger contact may and two secondary finger contact regions may be controlled by the same finger.
[030] FIG. 9 is a drawing showing features of a process of assigning initial finger contact regions.
[031] FIG. 10 is an illustrative drawing showing a contact of the touch sensitive surface in one of the finger contact regions.
[032] FIG. 11A is an illustrative drawing showing a sliding motion of the contact with the touch sensitive surface in a generally rightward direction.
[033] FIG. 11 B is an illustrative drawing showing a sliding motion of the contact with the touch sensitive surface in a generally leftward direction.
[034] FIG. 1 1C is an illustrative drawing showing a sliding motion of the contact with the touch sensitive surface in a generally upward direction (e.g., within about 45 ° or within about 30 ° of the upward direction).
[035] Fig. 12 is an illustrative drawing showing features of a process of entering a command by a gesture including contacting a finger contact region.
[036] FIG. 13A is an illustrative drawing showing a contact with a finger contact region that is offset from the center point of the finger contact region.
[037] FIG. 13B is an illustrative drawing showing a previous (e.g. initial) finger contact region and a new finger contact region. The finger contact region may be repositioned based on the direction and/or distance between the previous (e.g., initial) center of the finger contact region and the location of contact recognized on the surface.
[038] FIG. 14 is an illustrative drawing showing features of a process of repositioning a finger contact region.
[039] FIG. 15 is an illustrative flow diagram for entering one or more control commands. It will be appreciated that some of the steps may be eliminated and/or the order of the some of the steps may be changed. The process may also include additional steps.
[040] FIG. 16 is an illustrative drawing showing a touch sensitive surface facing away from a user.
[041] FIG. 17 is an illustrative drawing showing a touch sensitive surface in a vehicle facing away from a user.
[042] FIG. 18 is an illustrative drawing showing a touch sensitive surface in a motor vehicle facing towards a user. The ability to use "blind" control of this touch sensitive surface may reduce the level of distraction during its operation compared with a device requiring visual observation of control locations. [043] FIGs. 19A and 19B are drawings illustrating a device mounted to a steering wheel or a steering wheel column with a touch sensitive surface facing away from the driver.
[044] FIGs. 20A, 20B, 20C, and 20D are drawings of an illustrative game controller having one or more touch sensitive surfaces (e.g., on a concave surface of the game controller). DETAILED DESCRIPTION
[045] The explanations and illustrations presented herein are intended to acquaint others skilled in the art with the invention, its principles, and its practical application. Accordingly, the specific embodiments of the present disclosure as set forth are not intended as being exhaustive or limiting. The scope of the invention should be determined with reference to the appended claims, along with the full scope of equivalents to which such claims are entitled. The disclosures of all articles and references, including patent applications and publications, are incorporated by reference for all purposes. Other combinations are also possible as will be gleaned from the following claims, which are also hereby incorporated by reference into this written description.
[046] The method, systems, devices and apparatus for the control sessions according to the teachings herein are generally based on a processor identifying contacts with a touch sensitive surface, resulting in the selection of one or more control command.
[047] With reference to FIG. 1 , a control session 2 typically includes a mode for initial positioning of control locations 4 (e.g., finger contact regions) on a touch sensitive surface 14. The mode for initial positioning of control locations preferably is employed for positioning the control locations based on the locations of simultaneous contact at three or more spaced apart locations on the touch sensitive surface (e.g., by three or more fingers contacting the surface). After the control locations are assigned, the control session 2 includes a mode for entry of control commands 6. The mode for entry of control commands is employed for selecting one or more control commands 8 based on contact(s) with the control locations on the touch sensitive surface 14. The control session may include a mode for ending the control session. The control session may include a mode for repositioning control locations. Such a repositioning mode may be employed to compensate for shifting of a user's hand relative to the touch sensitive surface. The mode for the entry of control commands 6 may include the selection of control commands 7 and a mode for repositioning control locations 8, such as illustrated by FIG. 2. The ending of a control session 9 may occur after the mode for the entry of control commands 6.
[048] Throughout the entire duration of the control session, the processor is preferably in electronic communication with the touch sensitive surface for identifying positions of objects above or in contact with the touch sensitive surface, for identifying movement of objects above or the movement of contacts on the surface, and for identifying the removal of the objects away from the touch sensitive surface (e.g., removal of one or more contacts from the surface). [049] By employing a mode for initial positioning of control locations, the processor positions or assigns the control locations based on the positions of contact by a user, thus enabling "blind" interaction by a use with the touch sensitive surface.
[050] It will be appreciated from the description herein that the methods, devices, systems, and apparatuses according to the present teachings rely on contacts with one or more touch sensitive surfaces, and more preferably with a processor receiving indications of contact with the touch sensitive surface(s).
[051] The methods, devices, and systems according to the teachings herein may employ one or more of the features described in U.S. Patent Application No. 13/091 , 158 filed on April 21 , 201 1 , and U.S. Provisional Patent Application No. 62/142, 170, filed on April 2, 2015 entitled SYSTEM, APPARATUS AND METHOD FOR VEHICLE COMMAND AND CONTROL, both incorporated herein by reference in their entirety.
[052] Touch sensitive surface
[053] The methods and systems according to the teachings herein may employ a touch sensitive surface as a component in an input device for inputting commands. As used herein, a touch sensitive surface is capable of identifying the location of multiple simultaneous contacts on the surface. Each contact preferably includes a sufficient force applied to the surface as required by the touch sensitive surface to recognize a contact. The touch sensitive surface may be a flat surface, may be a curved surface, or may have regions that are flat and regions that are curved. Preferably the touch sensitive surface is characterized as being generally smooth and or having a generally uniform texture. For example, the touch sensitive surface may be sufficiently smooth and/or have a sufficiently uniform texture so that a user cannot identify the location of contact of the surface based on the surface topography or other tactile clues on the surfaces.
[054] The touch sensitive surface may be a surface of a pure entry component or device (i.e., a component or a device that does not display images), such as a touch pad, or may be a surface of a combination entry/display component or device, such as a touch-screen display.
[055] The device including the touch sensitive surface and/or a processor connected to the device preferably is capable of recognizing each of multiple contacts to the surface, the maintaining of the contact, the movement (if any) of the contact, and the termination (i.e., removal) of the contact.
[056] An illustrative touch sensitive surface is shown in FIG. 3. With reference to FIG. 3, the touch sensitive surface 14 may have a first direction 82 (e.g., an upward direction) and an opposing second direction 84 (e.g., a downward direction). The touch sensitive surface may have a third direction 86 (e.g., a rightward direction) and an opposing fourth direction 88 (e.g., a leftward direction). The first and second directions 82, 84 may be generally orthogonal to the third and fourth 86, 88 directions. When describing the directions as upward, downward, rightward and leftward, it will be appreciated that the descriptions refer to the first, second, third and fourth directions, and may have alternate meanings depending on the actual orientation of the touch sensitive surface. For example, when a touch sensitive surface is oriented on a horizontal plane, the upward and downward directions may in actuality refer to a forward direction and a backward direction respectively.
Mode for Initial Positioning of Control Locations (e.g., Finger contact regions)
[057] The initial positioning of the control locations (e.g., finger contact regions) according to the teachings herein typically requires the simultaneous contact of a touch sensitive surface at three or more points or regions. The three or more points of contact preferably are consistent with the contact by three or more fingers of a hand or hands. The number of simultaneous contacts required for the initial position of the control locations may be 3 or more, 4 or more, 5 or more, 6 or more, 7 or more, 8 or more, 9 or more, or 10. Preferably the number of simultaneous contacts on the touch sensitive surface is a predetermined target number or a predetermined minimum number. For example, if there is a predetermined target number of four, the positioning of the control locations will not be completed until precisely four simultaneous contacts are recognized on the touch sensitive surface. In contrast, if the predetermined minimum number of contacts of four, the positioning of the control locations will not be completed until at least four (e.g., four, five, six, eight, or more) simultaneous contacts are recognized on the touch sensitive surface.
[058] FIG. 3 illustrates the simultaneous contact of a touch sensitive surface 14 at three or more spaced apart locations 16 on a touch sensitive surface of a touch sensitive device 12 (e.g., a pure input device).
[059] The simultaneous contacts of the touch sensitive surface preferably are spaced apart on the surface in such a manner that sequential finger locations may be assigned. For example, the positions of the contacts may relate to a natural arch (or arches) formed by the fingertips of one hand. With reference to FIG. 4, the positions of the contacts 16 may be along an arch 17, such as a natural arch corresponding to three or more adjacent fingertips.
[060] In the mode of the initial positioning of the control locations, the process typically includes a step of a user removing some or preferably all of the fingers from the touch sensitive surface. As such, the processor, after identifying the simultaneous contact of three or more spaced apart locations, will then identify the removal (i.e., the absence of contact) at the three or more spaced apart locations). Preferably the processor recognizes the removal of all contact with the touch sensitive surface, such as illustrated in FIG. 4. It will be appreciated that one or more of the fingers may remain in contact with the touch sensitive surface after the initial positioning of the control locations (provided that at least one finger has been removed from the surface, and preferably at least two fingers have been removed from the surface). Such remaining finger(s) may be employed in directly entering a control compound as discussed herein, without the need to remove said finger(s) and re-establish contact with the surface.
[061] With reference to FIG. 5, the initial positioning of the control locations may include simultaneous contact with four or more fingers (for example with all the fingers of one hand). As illustrated by FIG. 5, the touch sensitive surface may be a touch screen display 15. As such, the locations of initial contact, regions including the location of initial contact, or the resulting control location 18 may be displayed on the touch sensitive screen display.
[062] FIG. 6 illustrates control locations after being positioned in the mode of initial positioning of the control locations. Each control location 18 is spaced apart from the other control locations. Each control location may be characterized by a geometric center 22. The position or region of contact 16 preferably is positioned within the control location 18. The geometric center 22 is preferably within the position or region of contact 16. The control locations may be a region having any shape. For example, the shape of a control location may be polygonal, circular, semi-circular, rectangular, oval shaped, elliptical, egg-shaped, or square. With reference to FIG. 6, each position on the touch sensitive surface is associated with at most one control location 18.
[063] The number of control locations (e.g., finger contact regions) may be more than the number of fingers that are employed when initially establishing the locations of the finger contact regions. For example, there may be a base finger contact region for a first finger and one or more secondary finger contact regions for the first finger. The base finger contact region may include the point of contact of the finger when defining the finger contact regions. Each secondary finger contact region (if any) for the first finger is sufficiently displaced from the base finger contact region for the first finger so that the base and secondary finger contact regions do not overlap . It will be appreciated that the locations of the various finger contact regions for a first finger may depend upon different amounts of curvature of the first finger. For example the first finger may be relatively curved when contacting the base finger contact region and relatively less curved (e.g., more outstretched when contacting a secondary finger contact region). This may be analogous to typing a "d" (relatively curved) and an "e" (relatively outstretched) on a QWERTY keyboard. As another example, the first finger may be relatively less curved when contacting the base finger contact region and relatively more curved when contacting a secondary finger contact region. This may be analogous to typing a "d" (relatively less curved) and a "c" (relatively more curved) on a QWERTY keyboard. Examples of a touch screen surface including base finger contact regions 72 and secondary finger contact regions 74 are shown in FIG. 7 and FIG. 8. In FIG. 7, there are 5 base finger contact regions (one for each finger of a hand) and 3 secondary finger contact regions (one for each of three of the fingers). In FIG. 8, there is one base finger contact region 72 and two secondary finger contact regions 74 for each of three fingers.
[064] Each finger contact region preferably is a different region of the touch sensitive surface. As such, any location on the touch sensitive surface may correspond to at most one finger contact region at a given time. It will be appreciated that some regions of the touch sensitive surface will correspond with no finger contact region. According to the teachings herein, the finger contact region associated with a given location may change over time. For example, a location p on the touch sensitive surface may be associated with a one finger contact region after a first stage of positioning finger contact regions and may be associated with a different finger contact region after a later stage of positioning finger contact regions. As another example, after one or more entry contacts in a finger contact region that are offset from the center of that finger contact region, a location p on the touch sensitive surface that was initially associated with that finger contact region may no longer be associated with it (instead, p may be associated with no finger contact region or may be associated with a different finger contact region).
[065] Features of a process for positioning initial control locations is illustrated in FIG. 9. The process may be recognized from the actions of a user of a touch sensitive device, from the actions of a device or a processor, or both. The actions of the user may include: a step of a user simultaneously contacting a touch sensitive surface with three or more fingers; and after contacting the touch sensitive surface, a step of the user removing the three or more fingers from the touch sensitive surface. The actions of the device or processor may include a step of recognizing a simultaneous contact on the touch sensitive surface at three or more spaced apart contact point or contact regions (e.g., consistent with the contact by fingers of a user's hand(s)); followed by a step of recognizing the removal of the contacts from the touch sensitive surface and the positioning of control locations (e.g., finger contact locations) based on the locations contacted. It will be appreciated that the positioning of the control locations may occur at any time after the device or processor recognizes the simultaneous contacts. For example, the positioning of the control location may be before the removal of the contacts. As illustrated in FIG. 9, the positioning of the control location may occur after recognizing the removal of the contacts. As such, the positioning of the control locations may be based on the initial position of contact, or a later (e.g., a final) position of contact in the event the contact moves during the mode of positioning the control locations.
Mode for Entry of Control Commands
[066] After establishing initial command locations on the touch sensitive surface, the touch sensitive surface may be employed in a mode for entry of control commands.
[067] The touch sensitive surface may be employed for controlling multiple devices and/or controlling multiple features or functions of the device(s). As such, the processor may need to identify which device / feature / function to control with a control command. The selection of the control command(s) may be based on the control location or locations contacted during a mode for entry of control commands. In order to increase the number of different possible commands that can be entered, one or any combination of the following features of a contact with the surface may be identified and used by the processor: the number of contact locations being contacted, a sliding motion of a contact, a direction of motion of a contact, the length of time of a contact, the number of sliding motions of a contact, and the removal of a contact. By using different types of gestures by a user, and recognizing the gestures by a processor, the problem of needing to select a control command from a large number of different commands with relatively few fingers is solved.
[068] The process may include entering a control command by a user making one or more gestures on a touch sensitive surface and a processor recognizing the gesture(s) and selecting the control command based on a predetermined association between the gestures and the control commands.
[069] Typically one or more fingers remain in contact with the touch sensitive surface during the entry of the command.
[070] The gesture may be any gesture which can be recognized by the processor. Preferably, a gesture begins by a contact at one or more of the control locations. A user may start enter a gesture by contacting the touch sensitive surface with one finger beginning in a finger contact region and, while maintaining the contact, making one or more sliding motions with the finger including sliding the finger in at least a first direction, and then removing the one finger from the touch sensitive surface. For example, a single finger may contact a finger contact region, then slide in one or any combination of the following directions: upwards, downwards, rightwards, and left wards, and then removing the finger from the touch sensitive surface. It will be appreciated that after moving in a first direction, a finger may be moved in the reverse direction prior to removing the finger from the touch sensitive surface.
[071] Contacting the touch sensitive surface with one finger in a finger contact region and then making one or more contacts with the touch sensitive surface with a second finger, and then removing the one finger from the touch sensitive surface. The contact with a second finger may be a sliding contact, a tapping contact, or a contact maintained at a single location.
[072] A first finger maintains in a constant location on the touch sensitive surface while a second finger makes a dynamic contact with the touch sensitive surface. The entry of the command preferably ends with the removal of the first finger from the touch sensitive surface.
[073] Contacting the touch sensitive surface with two fingers, each in a different finger contact region, and moving at least one of the fingers in a first direction. Preferably moving both fingers in the same direction or moving the two fingers towards each other. [074] It will be appreciated that the control of a device may include a plurality of gesture entries. For example, the process may require the entry of a first control command to select a device to control and a later entry of a control command to control a function of the device.
[075] The distance of movement and / or the length of contact may be employed for determining the level or extent of control of a device. For example, in controlling the speed of a device, movement in one direction may be employed to increase the speed continuously until the end of the gesture is recognized (e.g., by removing the contact or other entry completions according to the teachings herein). The rate of increase of the speed may be related to the distance of movement of the gesture. As another example, the volume of a device may be controlled by the gesture originating at a control location. The movement of the contact in a first direction by a first movement distance may result in the volume increasing by at a rate that correlates with the first movement distance. The increase in volume may continue until the contact is removed or the completion of the gesture is otherwise identified. The decrease of the volume may be similarly be accomplished by a contact in a second direction different from the first direction (e.g., an opposing direction) by a second movement distance.
[076] With reference to FIG. 10, the entry of a gesture may include a contact 40 on the touch sensitive surface at a point or region 16 within a control location 18 after the control locations have been established. The point or region of contact 16 may be offset from the center of the control location 22, such as shown in FIG. 10. The point or region of contact may include the center of the control location (not shown).
[077] The gesture may include a movement of one or more of the contacts. The movement may be a sliding movement in one or more sliding directions. For example, the gesture may include a movement of the contact 16 on the touch sensitive surface in a generally rightward direction (FIG. 1 1 A), in a generally leftward direction (FIG. 11 B), in a generally upward direction (FIG. 1 1C), or in a generally downward direction. It will be appreciated that a sliding movement may include movement in multiple directions and/or movement is non-linear. The entering of a command using a gesture may include multiple contacts each with a different control location. Each contact may include a movement in the same sliding direction, may include movements in different direction, or may include both contacts that are stationary and contacts that move. Preferably, multiple contacts during the entry of a command have simultaneous sliding motions (that are the same or different). For example, the gesture may include sliding motion of two contacts towards each other, sliding motion of two contacts away from each other, or sliding contacts that are in generally the same direction. A sliding motion may be a generally small motion (e.g., within a control locations), a generally large motion (e.g., outside the control location or larger than a distance from the center to the edge of the control location). [078] The association between commands and gestures may include a first command associated with the sliding motion of only a first finger in a first direction and a second different command associated with the sliding motion of only a second finger in a second direction. The first and second directions may be the same (e.g., within about 30 °, within about 20 °, within about 15 °, within about 10 °, or within about 5 °) or different (e.g., angled by more than 30 °, angled by more than 40 °, or angled by more than 50 °).
[079] The entry of a control command using a gesture may include one or more of the features of FIG. 12. A user may contact one or more of the control locations each with a different finger and then make a gesture on the touch sensitive surface while maintaining contact with the surface. The gesture is preferably associated with a predetermined control command. The process preferably includes a step of removing the fingers from the touch sensitive surface or other action for completing the entry of the gesture. After the completion of the entry of a gesture, another gesture may be entered. During the entering of a control command using a gesture, the processor or device generally identifies contact(s) with one or more control locations (e.g. after the control locations have been free of contact). The process or device then identifies the type of gesture(s) being made from the contacts. The process or device may identify the gesture as a continuous contact with the surface. After identifying a gesture, the processor or device may identify a predetermined control command associated with the entered gesture. Preferably, the control command is executed, sent, or otherwise acted upon. It will be appreciated that a control command may be associated with a single gesture or with a series of gestures. The processor or device may identify a gesture immediately upon contact, after one or more sliding motions, upon the removal of the contact from the surface, or after the completion of the entry of a gesture according to the teachings herein.
[080] Each gesture entry includes a contact with one or more control locations (e.g., finger contact regions). The entry of a gesture may be completed by removing all of the fingers from the touch sensitive surface.
[081] The entry of a gesture may be completed (e.g., timed out by a processor) after a predetermined time limit is exceeded from the contact with the control location(s). For example, the entry of a gesture may be completed after continuous contact with the touch screen surface for a predetermined time limit of about 0.5 seconds or more, about 1 second or more, about 1.5 seconds or more, about 2 seconds or more, about 3 seconds or more, or about 4 seconds or more. The predetermined time limit, if any, typically is about 100 seconds or less, about 30 seconds or less, about 15 seconds or less, about 10 seconds or less, or about 6 seconds or less.
[082] The entry of a gesture may be completed (e.g., timed out by a processor) after a predetermined limit for the number of changes in direction of a sliding motion (i.e., the predetermined direction change limit) is reached. Here, the multiple sliding directions generally occur while the contact with the touch sensitive surface is maintained. As used herein, a change in direction may be a change of about 15° or more, about 45° or more, about 90° or more, about 135° or more, or about 180°. For example, the gesture may be completed after the processor identifies an initial sliding motion and then a first change in the direction of the sliding (i.e., the predetermined direction change limit is one). As another example, the gesture may be completed after the processor identifies a sliding motion in an initial direction, followed by a sliding motion in a second direction and then a sliding motion in a third direction (i.e., the predetermined direction change limit is two). The predetermined limit may be one or more, two or more, three or more, or four or more. Typically, the predetermined direction change limit will be 10 or less, or 5 or less.
[083] The entry of a gesture may be completed (e.g., timed out by a processor) when the contact moves in a predetermined direction. The predetermined direction may be any direction. For example, the predetermined direction may be a generally upward direction, a generally downward direction, a generally rightward direction, or a generally leftward direction.
[084] The entry of a gesture may be completed (e.g., timed out by a processor) when the contact returns to the finger contact region (or a position in the finger contact region) following a sliding motion away from the finger contact region.
[085] The entry of a gesture may be completed (e.g., timed out by a processor) when the contact moves in a predetermined shape. For example, the entry of the gesture may be completed when the processor recognizes a sliding motion in a shape such as an arc, a semicircle, a circle, a triangle, a rectangle, a square, a star, a letter, a number, or any combination thereof.
[086] After a gesture is completed (e.g., after a processor determines that the entry of a gesture is completed), a next gesture may be entered by contacting one or more control location(s).
[087] During the mode for entry of control commands, contacts with the control locations may deviate from the center of the control locations. It will be appreciated that such deviations may be systematic and/or represent a shift in a user's perception of the control location. To compensate for such changes, the mode for entry of control commands may include a mode for repositioning of the control locations. Instead of having a mode for repositioning of control locations, the control locations may remain fixed.
[088] Mode for Repositioning Control Locations
[089] During the mode for entry of a control command, the touch sensitive surface may be contacted at a point or region of contact 16 inside a control location 18 but offset from the center 22 of the control location 18, such as illustrated in FIG. 13A. [090] The processor or device may reposition the control location based at least partially on the offset distance and offset direction between the center of the contact 15 and the center 22 of the control location 18.
[091] It will be appreciated that the repositioning may occur every time the contact is offset from the center of the control location, or only under certain conditions. For example, the offset distance must reach a threshold value prior to repositioning the control location. As another example, the offset must occur at a sufficient frequency prior to repositioning the control location.
[092] The new control location may have the same size, or a different size as the prior control location. The new control location may have the same shape, or a different shape as the prior control location. The new control location preferably has the same size or the same shape as the prior control location. More preferably, the new control location has the same size and shape as the prior control location.
[093] FIG. 13B illustrates the repositioning of a control location, showing the position of the new location and the position of the prior location. The new control location 36 may have a center 38 that is displaced from the center 22 of the prior control location 18. The direction of the displacement 32 preferably is about the same as the offset direction. The distance of the displacement between the prior center 22 and the new center 38 preferably is about the same as or less than the offset distance. For example, the distance between the prior center 22 and the new center 38 may be a percentage (preferably about 100% or less than 100%) of the offset distance.
[094] It will be appreciated that the offset distance and offset direction for one contact location may be employed for repositioning one or more of the other contact locations.
[095] The reposition of the finger contact region may include one or any combination of the steps illustrated in Fig. 14. The repositioning of a control location generally occurs after a contact is initiated in a control location. The repositioning of the control location may occur prior to a sliding motion of the contact, after a sliding motion of the contact, or after removal of the contact from the touch sensitive surface.
[096] The process may allow for repeated or continuous entry of different control commands. For example, after identifying the end of the entry of a control command, the process may be used for the entry of a subsequent control command, such as illustrated in FIG. 15. It will be appreciated that the step of the removal of the contact may be replaced by a step of identifying the end of an entry of a control command, such as described herein. It will also be appreciated that such a step of identify the end of an entry of a control command may occur (e.g., recognizing the removal of the finger contact from the touch sensitive surface) may occur after a step of identifying a command based on the gesture.
[097] Orientation of Touch sensitive surface [098] The touch sensitive surface may be oriented in a direction and/or location that prevents viewing of the touch sensitive surface by a user. For example a user may be holding a device including a touch sensitive surface so that the touch sensitive surface is facing away from the user. With reference to FIG. 16, the user 62 may be holding the device 12, so that the touch sensitive surface 14 is directed away from the user's eye's 64. For example, the screen direction 52 perpendicular to and away from the touch sensitive surface 14 may partially or entirely be in a forward direction. As such, the screen display direction may be away from the user's eyes 64. With reference to FIG. 17, the user 62 may be a driver of a vehicle 66 and the touch sensitive device 14 may be mounted to the vehicle 66. Here, the touch sensitive surface is mounted so that the touch sensitive surface faces away from the eyes of the driver.
[099] The touch sensitive surface may be oriented within the field of view of the user. However, operation and/or control of the device may be enhanced by setting control locations based on contact with the touch sensitive surface with three or more fingers. For example, as illustrated in FIG. 18, the touch sensitive surface may be in the field of view of the driver of a vehicle.
[0100] The touch sensitive surface may be mounted to, attached to, or integrated with a device positioned for contact by a user. For example, the touch sensitive surface may be located in a vehicle in proximity to a driver, and preferably at an ergonomic location. As an example, the touch sensitive surface may be mounted to a steering wheel and/or a steering column, such as illustrated in FIG. 19A and 19B.
[0101] The devices and systems according to the teachings herein may include one or more view panels (e.g., a display panel that is not a touch sensitive display) for viewing the command functions available. A display panel may be on the same device as the touch sensitive surface, but on a different location. For example, a display panel and a touch sensitive surface may be on opposing sides of a device. A display panel may be on a different device as the touch sensitive surface. For example, a touch sensitive panel may be mounted on a steering wheel and a display panel may be a panel attached to or integrated with a dashboard of a vehicle. With reference to FIG. 19B, a display panel 68 may be a display of a mobile phone, or a vehicle display device. A view panel display may fade or turn-off after a pre-determined interval of inaction on the touch sensitive surface.
[0102] The device or system including the touch sensitive surface may include a control component for turning the device on or off, or for resetting the device. Control component may be a switch or other component capable of executing and/or communicating an on, off, or reset function.
[0103] The device or systems may include one or more features for disabling the touch sensitive surface. Such a disabling feature may be particularly useful in a vehicle when it may be desirable to disable the operation of the device and systems based on vehicle operating conditions. For example, the touch sensitive surface may be disabled when the vehicle is turning and/or has recently changed directions, is driving above certain speeds, etc.
[0104] The touch sensitive devices according to the teachings herein may include a grip for spreading the fingers towards different locations on the touch sensitive surface.
[0105] The control commands according to the teachings herein may be employed for controlling an air conditioning, a radio, a window, a light, a lock, a cruise control, an application on a mobile phone, a navigation control, the location of a cursor, a mechanical device, an electronic device, operation of a land vehicle, operation of a water vehicle, operation of an air vehicle, remote operation of a vehicle or other device, a communication device, or any combination thereof. As used herein, a control command includes providing an API or code to enable two computer systems to interact, such as a smart phone and an automobile computer system.
[0106] Entry of a passcode
[0107] The devices, methods, systems and apparatus according to the teachings herein may be employed for entering a passcode for unlocking a device. Here, the passcode may consist of a sequence of gestures each employing one or more contacts with the finger contact regions. When unlocking the device, the touch sensitive surface is simultaneously contacted with three or more fingers for assigning the finger contact regions based on the locations of the contacts. The user may then enter a passcode by contacting the finger contact regions for entering a series of gestures. By way of example, the passcode may be a series of 2 or more gestures, 4 or more gestures, 6 or more gestures, or 8 or more gestures. Each gesture may be the same or different from the previous gesture. Each gesture may employ the same or different finger contact regions as the previous gesture. Each gesture may employ the same number of fingers or a different number of fingers as the previous gesture. For example, the passcode may include one gesture requiring the sliding movement of one, two, or more fingers in the same direction, the passcode may include the sliding movement of two fingers towards each other, the passcode may include the simultaneous tapping of one, two, or more finger contact regions, or any combination thereof. For purposes of illustration, a passcode may be entered by first contacting the touch sensitive surface simultaneously with a first, second, third, and fourth finger for assigning a finger contact region for each of the four fingers, followed by a first gesture entry of a sliding the first finger to the right, followed by a second gesture of sliding the third and fourth fingers together, followed by a third gesture of sliding the first, second, and third fingers in an upward direction, followed by a fourth gesture of tapping the surface with the fourth finger. It will be appreciated that the large number of possible gestures will result in more secure passcodes and/or reduce the number of entries required to obtain a secure passcode. It will also be appreciated that the passcode can be entered at different locations on the touch screen surface, reducing the possibility that evidence of the touch passcode will remain on the screen after repeated entries of the passcode over time.
[0108] Preferred passcodes include two or more sequential gestures in different directions. Preferred passcodes include two or more sequential gestures using different finger contact regions or different combinations of finger contact regions.
[0109] When entering a passcode, the processor may recognize the sequential contact of the touch sensitive surface at three or more locations, consistent with the contacting of the surface with three or more fingers and then assigning the finger contact locations based on the locations of contact. The number of simultaneous contacts for assigning the finger contact regions may be a predetermined number, such as 3, 4, 5, 6, 7, 8, 9, or 10. After the processor identifies that the multiple simultaneous contacts are removed from the surface, the processor may monitor the surface for a contact and gesture consistent with a first gesture of a predetermined passcode.
[0110] The system may include a mode of setting or resetting a passcode. The setting or resetting of the passcode may include simultaneously contacting the touch sensitive surface with three or more fingers for assigning the finger contact regions, followed by the sequential entry of the gestures of passcode. The system may require re-entry of the passcode to confirm the passcode. The system may store the passcode. The passcode may be encrypted by the system. The system may delete prior passcodes.
[0111] The touch sensitive surface may be attached to a vehicle. The touch screen surface in a vehicle may be proximate the driver's seat and/or the driver of the vehicle. A touch screen surface for use by a passenger may be proximate the location of one or more passenger seats and/or one or more passengers of a vehicle. For example, the touch screen device in a vehicle may be connected to the steering wheel, the steering wheel column, the dashboard, a seat back, a pillar, a door, or any combination thereof.
[0112] The touch screen surface may be in communicative connection with one or more devices to be controlled. For example, the touch screen surface may be in communicative connection with a control processor of the vehicle.
[0113] The touch sensitive surface may be oriented face down (out of view of the driver or other user). The touch sensitive surface may be part of a device or connected to a device adapted to track movements by multiple individually identified fingers.
[0114] The touch sensitive surface may be positioned so that a vehicle driver can contact the surface with one or more fingers without removing his hands from the steering wheel.
[01 15] The touch sensitive surface may be associated with an apparatus software (e.g., an application) that identifies and/or tracks multiple contacts of the surface. The apparatus software may be running on a computer system. The apparatus software may be adapted to identify finger generated contact, finger generated gestures or both. The apparatus software may be adapted to identify multiple (preferably three or more, four or more, or five or more) simultaneous finger generated contact, multiple finger generated gestures, or both.
[0116] Data Entry Device / Game Controller
[0117] The methods, systems, and apparatus according to the teachings herein may be employed in a data entry device, such as a handheld data entry device. For example, the data entry device may be a device for controlling the operation of an application, a processor, a connected device or a remote device. By way of illustration, the data entry device may be a device for controlling (e.g., remotely controlling) a video game, a machine, a vehicle, a flying device, a robotic device, or any combination thereof. As an example, the data entry device may be a game controller. A handheld device, such as a game controller may have one or any combination of the features illustrated in FIG. 20A and FIG. 20B. With reference to FIG. 20A and FIG. 20B, the handheld device 110 may have a forward surface 1 12 and an opposing rearward surface 1 16. The forward surface may include one or more touch sensitive surfaces 14. With reference to FIG. 20B, the forward surface 1 12 of the device 1 10 may include a touch sensitive surface for the fingers of a user's right hand and a second touch sensitive surface for the fingers of a user's left hand. It will be appreciated that the forward facing surface may have a single touch sensitive surface sufficiently large for receiving simultaneous contacts from fingers of both hands. The touch sensitive surface 14 on the forward facing surface 1 12 of the device preferably is sufficiently large for contacting with three or more spaced apart fingers of a hand. The device preferably has a sufficient number and size of touch sensitive surfaces for contact with 4 or more fingers, 6 or more fingers or 8 or more fingers. The rearward surface may include one or more thumb controls, such as a button, a knob, a dial, a joystick, or a rollerball capable of being controlled by a thumb, while the fingers rest on the forward surface. With reference to FIG. 20A, the device 1 10 may include one or more (e.g., two or more) thumb controls for a left hand, one or more (e.g., two or more) thumb controls 1 18 for a right hand, or both. The device has side surfaces 1 14 connecting the forward and rearward surfaces. The side surfaces 1 14 preferably are adapted for receiving a palm of a hand. The side surfaces 114 may be rounded or otherwise curved. The device 110 may include a gripping feature for assisting in the placement of the fingers and/or making the device easier to grip. An example of a gripping feature 120 is shown in FIG. 20C. The forward surface 112, 1 12', 1 12" may be generally concave, generally planar, or generally convex. Preferably, the forward surface 112" is generally convex, and the rearward surface is generally concave, such as illustrated in FIG. 20D, resulting in a more natural placement of the side surfaces 1 14 between the thumb and fingers.
[0118] It will be appreciated that during any of the aforementioned modes (e.g., a mode for initial position of control locations, a mode for repositioning of control locations, or a mode for entry of a control command), one or more of the contacts of the touch sensitive surface (e.g., by contact with one or multiple fingers) may be replaced by a sensing of the locations of the multiple fingers. For example, the locations of the fingers may be identified by light (e.g., laser light) or other forms of radiation, electrical fields, magnetic fields, darkness level (e.g., a shadow), or any combination thereof. It will be appreciated that a glove or other device may be placed on one or more fingers to enable the identification of the location of the finger and/or to enhance the aforementioned observation of the location of the finger. Similarly, a gesture on the surface of the device may be sensed by one of the aforementioned means, with or without actual contact with the surface. Preferably any such sensing occurs while the finger is at least near the touch sensitive surface (e.g., about 30 mm or less, about 10 mm or less, about 3 mm or less, or about 1 mm or less from the surface). A sensing may be completed when the finger moves away from the surface. For example, the end of a mode or entry of a command may be completed when it is sensed that a finger is no longer near the touch sensitive surface (e.g., further than 1 mm, 3 mm, 10 mm, or 30 mm from the surface, or when the distance from the surface has increased by at least about 1 mm, about 3 mm, about 5 mm, about 10 mm, or about 20 mm).
[0119] One or more of the aforementioned modes may be completed (e.g., timed out by a processor) after a predetermined time limit is exceeded from one or any combination of the following: i) one or more contacts with the touch sensitive surface; ii) the sensing of one or more objects (e.g., fingers) above the touch sensitive surface); iii) the removal of one or more contacts from the touch sensitive surface; or iv) the sensing of the movement of one or more objects (e.g., fingers) away from the touch sensitive surface. Preferably, the predetermined time limit, if any, is about 0.5 seconds or more, about 1 second or more, about 1.5 seconds or more, about 2 seconds or more, about 3 seconds or more, or about 4 seconds or more. Preferably the predetermined time limit, if any, is about 100 seconds or less, about 30 seconds or less, about 15 seconds or less, about 10 seconds or less, or about 6 seconds or less. A predetermined time limit may be fixed, may be adjusted (e.g., based on experience or historical values, or set by a user).
[0120] The transition from a mode for initial position of control locations to a mode for entry of control commands may be triggered by a trigger event. The trigger event may be a predetermined time limit, such as described herein. The trigger event may be the sensing of the removal of one or more objects (e.g., fingers) away from the touch sensitive surface. The trigger event may require the processor to sense that exactly one of the objects (e.g., one finger) remains on or near the touch sensitive surface while the other objects are moved away from the touch sensitive surface.
[0121] Any numerical values recited herein include all values from the lower value to the upper value in increments of one unit provided that there is a separation of at least 2 units between any lower value and any higher value. As an example, if it is stated that the amount of a component or a value of a process variable such as, for example, temperature, pressure, time and the like is, for example, from 1 to 90, preferably from 20 to 80, more preferably from 30 to 70, it is intended that values such as 15 to 85, 22 to 68, 43 to 51 , 30 to 32 etc. are expressly enumerated in this specification. For values which are less than one, one unit is considered to be 0.0001 , 0.001 , 0.01 or 0.1 as appropriate. These are only examples of what is specifically intended and all possible combinations of numerical values between the lowest value and the highest value enumerated are to be considered to be expressly stated in this application in a similar manner. As can be seen, the teaching of amounts expressed as "parts by weight" herein also contemplates the same ranges expressed in terms of percent by weight. Thus, an expression in the Detailed Description of the Invention of a range in terms of at " x parts by weight of the resulting polymeric blend composition" also contemplates a teaching of ranges of same recited amount of "x" in percent by weight of the resulting polymeric blend composition."
[0122] Unless otherwise stated, all ranges include both endpoints and all numbers between the endpoints. The use of "about" or "approximately" in connection with a range applies to both ends of the range. Thus, "about 20 to 30" is intended to cover "about 20 to about 30", inclusive of at least the specified endpoints.
[0123] The disclosures of all articles and references, including patent applications and publications, are incorporated by reference for all purposes. The term "consisting essentially of" to describe a combination shall include the elements, ingredients, components or steps identified, and such other elements ingredients, components or steps that do not materially affect the basic and novel characteristics of the combination. The use of the terms "comprising" or "including" to describe combinations of elements, ingredients, components or steps herein also contemplates embodiments that consist essentially of the elements, ingredients, components or steps. By use of the term "may" herein, it is intended that any described attributes that "may" be included are optional.
[0124] Plural elements, ingredients, components or steps can be provided by a single integrated element, ingredient, component or step. Alternatively, a single integrated element, ingredient, component or step might be divided into separate plural elements, ingredients, components or steps. The disclosure of "a" or "one" to describe an element, ingredient, component or step is not intended to foreclose additional elements, ingredients, components or steps.
[0125] It is understood that the above description is intended to be illustrative and not restrictive. Many embodiments as well as many applications besides the examples provided will be apparent to those of skill in the art upon reading the above description. The scope of the invention should, therefore, be determined not with reference to the above description, but should instead be determined with reference to the appended claims, along with the full scope of equivalents to which such claims are entitled. The disclosures of all articles and references, including patent applications and publications, are incorporated by reference for all purposes. The omission in the following claims of any aspect of subject matter that is disclosed herein is not a disclaimer of such subject matter, nor should it be regarded that the inventors did not consider such subject matter to be part of the disclosed inventive subject matter.
[0126] 2 Control Session
[0127] 4 Mode for initial positioning of control locations (e.g., initial finger contact locations)
[0128] 6 Mode for entering control commands
[0129] 7 Control command
[0130] 8 Mode for repositioning control locations
[0131] 9 Ending a control session
[0132] 10 Simultaneous contact of touch sensitive surface
[0133] 12 Touch sensitive device
[0134] 14 Touch sensitive surface
[0135] 16 Point / region of contact of touch sensitive surface
[0136] 17 Arch (e.g., natural arch of spaced apart fingertips)
[0137] 18 Finger contact region
[0138] 20 Removal of some or all of the fingers from the touch sensitive surface
[0139] 22 Center of the finger contact region
[0140] 24 Finger
[0141] 30 Assignment of finger contact regions
[0142] 32 Direction for changing the finger contact region
[0143] 34 Gesture direction or sliding direction while contacting the surface
[0144] 36 New finger contact region
[0145] 38 New center of finger contact region
[0146] 40 Command entry contact (contacting a touch sensitive surface after establishing finger contact regions).
[0147] 52 View direction of the touch sensitive surface (e.g., normal to the plane of the surface)
[0148] 62 User of touch sensitive surface
[0149] 64 User's eyes
[0150] 66 Automotive vehicle
[0151] 68 Display panel
[0152] 72 base finger contact regions
[0153] 74 secondary finger contact regions
[0154] 82 First direction on touch sensitive surface (e.g., upward direction) [0155] 84 Second direction on touch sensitive surface (e.g., downward direction)
[0156] 86 Third direction on touch sensitive surface (e.g., rightward direction)
[0157] 88 Fourth direction on touch sensitive surface (e.g., leftward direction)
[0158] 90 Mode for setting initial control locations (e.g., initial finger contact regions)
[0159] 92 Mode for entering control commands
[0160] 94 Mode for repositioning control locations
[0161] 1 10 Game controller
[0162] 1 12 Forward Surface (e.g., away from user)
[0163] 1 14 Side Surface
[0164] 1 16 Rearward Surface (e.g., towards user)
[0165] 1 18 Thumb Control
[0166] 120 Gripping Feature

Claims

What is claimed is
Claim 1 : A method of entering a command comprising the steps of:
i. a processor connected to a touch sensitive surface sensing a simultaneously positioning of three or more objects above and near or on the touch sensitive surface at three or more different sensing locations, including a positioning centered at a first finger initial sensing point, a positioning centered at a second finger initials sensing point, and a positioning centered at a third finger initial sensing point;
ii. the processor assigning a finger location region for two or more (e.g., each) of the three or more objects, wherein each finger location region is a distinct region of the touch sensitive surface, and each finger location region includes one of the initial finger sensing points;
iii. the processor entering a command entry mode following the step of assigning the finger location regions (e.g., following a transitioning event), wherein the command entry mode includes an association having at least a first command associated with a movement of only one of the objects starting at a first finger location region and a second command different from the first command associated with the a movement of only one of the objects starting at a second finger location region; and
iv. the processor recognizing a gesture on the touch sensitive surface including sensing the movement of only one of the objects starting at the first finger location region and identifying the associated first command based on the gesture.
Claim 2. The method of claim 1 , wherein the method includes the processor recognizing a gesture on the touch sensitive surface including the movement of only one of the objects starting at the second finger location region and identifying the associated second command based on the gesture.
Claim 3. The method of claim 1 or 2, wherein the command entry mode is started after a predetermined time interval from the processor sensing the simultaneous positioning of the three or more objects.
Claim 4. The method of claim 1 or 2, wherein the command entry mode is started after the processor identifies the removal of one or more of the objects.
Claim 5. The method of claim 4, wherein the command entry mode is started after the processor identifies the removal of all except for one of the objects.
Claim 6. The method of claim 5, wherein the process senses a continuous contact between one of the objects and the touch sensitive surface from the sensing of the positioning of the object to the sensing the gesture made with the object.
Claim 7. The method of claim 6, wherein a third entry command different from the first and second entry commands is associated with a different movement of only one of the objects starting at the first finger location region.
Claim 8. The method of claim 1 or 2, wherein the touch sensitive surface is a surface of a pure entry device (i.e., the touch sensitive surface is not a display surface).
Claim 9. The method of claim 1 or 2, wherein the touch sensitive surface is oriented so that the surface faces away from a user.
Claim 10. The method of claim 1 or 2, wherein the method includes executing the control command for controlling a device.
Claim 11. The method of claim 1 or 2, wherein during the command entry mode the processor recognizing an entry positioning of an object within the first finger location region centered at a first finger offset sensing point different from the first finger initial sensing point, and the method includes a step of relocating the first finger location region.
Claim 12. The method of claim 11 , wherein the first finger location region has a geometric center and the step of relocating the first finger relocation region includes moving the geometric center of the first finger location region towards the first finger offset sensing point (e.g., the geometric center of the first finger location region may become the first finger offset sensing point).
Claim 13. The method of claim 1 or 2, wherein the method includes one or more of the following features:
i) the three or more objects are fingers;
ii) the step of sensing the simultaneous position occurs when the objects are 10 mm or less from the touch sensitive surface;
iii) the three or more different sensing locations are consistent with the positioning of three or more fingers of a hand;
iv) the processor assigns a finger location region for each of the objects;
v) each finger location region corresponds to only one of the fingers;
vi) the command entry mode is started after a transition event; or
vii) the transition event is either the processor sensing the removal of all except one of the objects from the touch sensitive surface or a predetermined time interval is reached after the sensing of the positioning of the objects.
Claim 14. A method of entering a control command comprising the steps of:
i. identifying a simultaneously contact on a touch sensitive surface at three or more different contact locations (e.g., consistent with the contact by three or more fingers of a user), including a contact centered at a first finger initial contact point (e.g., by a first finger), a contact centered at a second finger initial contact point (e.g., by a second finger of a second finger), and a contact centered at a third finger initial contact point (e.g., by a third finger);
ii. assigning a finger contact region for each of the three or more fingers, wherein each finger contact region is a different region of the touch sensitive surface, and each finger contact region includes one of the initial contact points (e.g. wherein each finger contact region corresponds to one of the fingers);
iii. recognizing the removing of the contact of the touch sensitive at one or more (e.g., all) of contact locations (e.g., the removal of one, two, three or all of the fingers from the touch sensitive surface); and
iv. recognizing an entry contact at one of the finger contact regions followed by a sliding movement of the entry contact in one or more contact movement directions, and moving a cursor on a display in a corresponding one or more cursor movement directions.
Claim 15. The method of claim 14, wherein the method includes:
the processor recognizing
i) a single contact on the touch sensitive surface in a single one of the finger contact regions,
ii) a movement of the single contact in one or more directions for selecting an application to control, and
iii) removal of the single contact from the touch sensitive surface;
the processor selecting the application to control based on the finger contact region and the movement (before or after the removal of the single contact);
the processor recognizing:
i) a different single contact on the touch sensitive surface in a different one of the finger contact regions,
ii) a movement of the different single contact in one or more directions for selecting a control of the application, and
iii) a removal of the different single contact from the touch sensitive surface; and the processor selecting the control of the application based on the location of the different single contact and the movement of the different single contact (before or after the removal of the different single contact).
Claim 16. The method of claim 1 or 2, wherein the entry command controls a device selected from a radio, a phone, a telecommunication device, a heating and/or air conditioning system (i.e., a HVAC system), a motor, an internet connection, an internet application, a video game, and a light.
Claim 17. The method of claim 1 or 2, wherein each position on the touch sensitive surface is within at most one of the finger location regions.
Claim 18. A system for entry of control commands for controlling a device comprising:
an entry device including a touch sensitive surface;
a processor connected to the touch sensitive surface;
a memory storing instruction that, when executed by the processor causes the processor to: i. sensing a simultaneously positioning of three or more objects (e.g., fingers) above and near or on the touch sensitive surface at three or more different sensing locations (e.g., consistent with the positioning by three or more fingers of a user), including a sensing centered at a first finger initial sensing point (e.g., by a first finger), a sensing centered at a second finger initial sensing point (e.g., by a second finger), and a sensing centered at a third finger initial sensing point (e.g., by a third finger);
ii. assign a finger location region for two or more (e.g., for each) of the three or more objects, wherein each finger location region is a distinct region of the touch sensitive surface, and each finger location region includes one of the initial sensing points (e.g. wherein each finger location region corresponds to one of the fingers);
iii. enter a command entry mode following a predetermined event, wherein the command entry mode includes an association of at least a first entry command associated with a movement of only one of the objects starting at a first finger location region and a second different entry command associated with generally the same movement of a different one of the objects starting at a second finger location region; and
iv. recognizing a gesture on the touch sensitive surface including sensing the movement of only one of the objects starting at the first finger location region and identifying the associated entry command based on the finger location and the movement.
Claim 19. A system for entry of control commands for controlling a device comprising:
an entry device including a touch sensitive surface;
a processor;
a memory storing instruction that, when executed by the processor causes the processor to: i. identify a simultaneously contact on the touch sensitive surface at three or more different contact locations (e.g., consistent with the contact by three or more fingers of a user), including a contact centered at a first finger initial contact point (e.g., by a first finger), a contact centered at a second finger initial contact point (e.g., by a second finger of a second finger), and a contact centered at a third finger initial contact point (e.g., by a third finger);
ii. assign a finger contact region for each of the three or more fingers, wherein each finger contact region is a different region of the touch sensitive surface, and each finger contact region includes one of the initial contact points (e.g. wherein each finger contact region corresponds to one of the fingers); iii. recognize the removing of the contact of the touch sensitive at one or more, (e.g., two, three, or all) of the contact locations (e.g., the removal of the one, two, three, or more fingers, or all of the fingers from the touch sensitive surface) before or after step ii (preferably after step ii);
iv. recognize the simultaneous entry contact(s) in one or more of the finger contact regions (e.g., the simultaneously contacting of one or more of the finger contact regions with the corresponding finger) each at an entry contact point;
v. recognize a gesture on the touch sensitive surface including a movement of the entry contact in one or more directions on the touch sensitive surface starting from the entry contact point while continuous entry contact is maintained; (e.g., from a motion on the touch sensitive surface by the user making a gesture on the touch sensitive surface with one or more of the fingers in contact with touch sensitive surface, wherein each gesture originates in the corresponding finger contact region); and
vi. identify a command to be execute based on the finger contact regions that are contacted and the gestures originating in the finger contact regions.
Claim 20. A machine readable storage medium containing instructions that when executed cause a processor of an electronic device to discern input control commands by: i. the processor identifying a simultaneously contact on the touch sensitive surface at three or more different contact locations (e.g., consistent with the contact by three or more fingers of a user), including a contact centered at a first finger initial contact point (e.g., by a first finger), a contact centered at a second finger initial contact point (e.g., by a second finger of a second finger), and contact at a third finger initial contact point (e.g., by a third finger);
ii. the processor assigning a finger contact region for each of the three or more fingers, wherein each finger contact region is a different region of the touch sensitive surface, and each finger contact region includes one of the initial contact points (e.g. wherein each finger contact region corresponds to one of the fingers); iii. the processor recognizing the removing of the contact of the touch sensitive at one or more (e.g., two or more, three or more, or all) of the contact locations (e.g., the removal of one, two, three, or all of the three or more fingers from the touch sensitive surface);
iv. the processor recognizing the simultaneous entry contact(s) in one or more of the finger contact regions (e.g., the simultaneously contacting of one or more of the finger contact regions with the corresponding finger) each at an entry contact point; v. the processor recognizing a gesture on the touch sensitive surface including a movement of the entry contact in one or more directions on the touch sensitive surface starting from the entry contact point while continuous entry contact is maintained; (e.g., from a motion on the touch sensitive surface by the user making a gesture on the touch sensitive surface with one or more of the fingers in contact with touch sensitive surface, wherein each gesture originates in the corresponding finger contact region); and
vi. the processor identifying a command to be execute based on the finger contact regions that are contacted and the gestures originating in the finger contact regions.
Claim 21. A method for unlocking one or more electronic device controls by a processor comprising:
i. receiving an indication of simultaneous contact at multiple locations on a touch sensitive surface;
ii. determining that the multiple locations have a spacing consistent with contact of the touch sensitive surface by one or two hands of a user;
iii. assigning multiple finger contact regions based on the multiple locations of simultaneous contact;
iv. receiving indications of a sequential series of contacts each with one or more of the finger contact regions;
v. comparing the sequential series of contacts with a predetermined sequential series (i.e., a password series) and unlocking one or more controls when the sequential series of contacts matches the predetermined sequential series.
Claim 22. A method of entering a command using a processor connected to a touch sensitive surface comprising the steps of:
i. the processor sensing a simultaneously positioning of three or more objects (e.g., fingers) above and near (e.g., within 10 mm of) the touch sensitive surface at three or more different sensing locations (e.g., consistent with the positioning by three or more fingers of a user), including a positioning centered at a first finger initial sensing point (e.g., by a first finger), a positioning centered at a second finger initials sensing point (e.g., by a second finger), and a positioning centered at a third finger initial sensing point (e.g., by a third finger);
ii. the processor assigning a finger location region (e.g., a finger sensing region) for each of the three or more objects, wherein each finger location region (e.g., finger sensing region) is a different region of the touch sensitive surface, and each finger sensing region includes one of the initial finger sensing points (e.g. wherein each finger location region corresponds to one of the fingers); iii. the processor recognizing the movement away from the touch sensitive surface of the object at one or more of the sensing locations (e.g., at two or more of the sensing locations or at all of the three or more sensing locations);
iv. the processor recognizing the simultaneous entry contact(s) in one or more of the finger contact regions (e.g., the simultaneously contacting of one or more of the finger contact regions with the corresponding finger) each at an entry contact point;
v. the processor recognizing a gesture on the touch sensitive surface including a movement of the entry contact in one or more directions on the touch sensitive surface starting from the entry contact point while continuous entry contact is maintained; (e.g., from a motion on the touch sensitive surface by the user making a gesture on the touch sensitive surface with one or more of the fingers in contact with touch sensitive surface, wherein each gesture originates in the corresponding finger contact region); and vi. the processor identifying a command to be execute based on the finger contact regions that are contacted and the gestures originating in the finger contact regions.
Claim 23. A method of entering a command using a processor connected to a touch sensitive surface comprising the steps of:
i. the processor identifying a simultaneously contact on the touch sensitive surface at three or more locations (e.g., consistent with the contact by three or more fingers of a user), including a contact centered at a first finger initial contact point (e.g., by a first finger), a contact centered at a second finger initial contact point (e.g., by a second finger of a second finger), and contact at a third finger initial contact point (e.g., by a third finger);
ii. the processor assigning a finger contact region for each of the three or more fingers, wherein each finger contact region is a different region of the touch sensitive surface, and each finger contact region includes one of the initial contact points (e.g. wherein each finger contact region corresponds to one of the fingers);
iii. the processor recognizing the removing of the contact of the touch sensitive at the three or more locations (e.g., the removal of the three or more fingers from the touch sensitive surface);
iv. the processor recognizing the simultaneous entry contact(s) in one or more of the finger contact regions (e.g., the simultaneously contacting of one or more of the finger contact regions with the corresponding finger) each at an entry contact point;
v. the processor recognizing a gesture on the touch sensitive surface including a movement of the entry contact in one or more directions on the touch sensitive surface starting from the entry contact point while continuous entry contact is maintained; (e.g., from a motion on the touch sensitive surface by the user making a gesture on the touch sensitive surface with one or more of the fingers in contact with touch sensitive surface, wherein each gesture originates in the corresponding finger contact region); and vi. the processor identifying a command to be execute based on the finger contact regions that are contacted and the gestures originating in the finger contact regions.
Claim 24. A method of entering a control command comprising the steps of:
i. identifying a simultaneously contact on a touch sensitive surface at three or more locations (e.g., consistent with the contact by three or more fingers of a user), including a contact centered at a first finger initial contact point (e.g., by a first finger), a contact centered at a second finger initial contact point (e.g., by a second finger of a second finger), and contact at a third finger initial contact point (e.g., by a third finger);
ii. assigning a finger contact region for each of the three or more fingers, wherein each finger contact region is a different region of the touch sensitive surface, and each finger contact region includes one of the initial contact points (e.g. wherein each finger contact region corresponds to one of the fingers);
iii. recognizing the removing of the contact of the touch sensitive at the three or more locations (e.g., the removal of the three or more fingers from the touch sensitive surface); and
iv. recognizing an entry contact at one of the finger contact regions followed by a sliding movement of the entry contact in one or more contact movement directions, and moving a cursor on a display in a corresponding one or more cursor movement directions.
Claim 25. The method of any of claims 1 , 2, 21 , 22, 23, or 24, wherein the method includes: the processor recognizing
i) a single contact on the touch sensitive surface in a single one of the finger contact regions,
ii) a movement of the single contact in one or more directions for selecting an application to control, and
iii) removal of the single contact from the touch sensitive surface;
the processor selecting the application to control based on the finger contact region and the movement (before or after the removal of the single contact);
the processor recognizing:
Claim 26 The method of any of claims 1 , 2, 21 , 22, 23, or 24, wherein the method is a method for unlocking one or more electronic devices.
PCT/IB2016/001256 2015-08-20 2016-08-22 Device, system, and methods for entering commands or characters using a touch screen WO2017029555A2 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN201680061416.1A CN108780365B (en) 2015-08-20 2016-08-22 Apparatus, system and method for inputting commands or characters using a touch screen
EP16836703.5A EP3338172A4 (en) 2015-08-20 2016-08-22 Device, system, and methods for entering commands or characters using a touch screen

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US201562207564P 2015-08-20 2015-08-20
US62/207,564 2015-08-20
US201562266916P 2015-12-14 2015-12-14
US62/266,916 2015-12-14

Publications (2)

Publication Number Publication Date
WO2017029555A2 true WO2017029555A2 (en) 2017-02-23
WO2017029555A3 WO2017029555A3 (en) 2017-04-27

Family

ID=58050803

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/IB2016/001256 WO2017029555A2 (en) 2015-08-20 2016-08-22 Device, system, and methods for entering commands or characters using a touch screen

Country Status (3)

Country Link
EP (1) EP3338172A4 (en)
CN (1) CN108780365B (en)
WO (1) WO2017029555A2 (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10120567B2 (en) 2015-04-02 2018-11-06 Inpris Innovative Products From Israel Ltd System, apparatus and method for vehicle command and control
US10146428B2 (en) 2011-04-21 2018-12-04 Inpris Innovative Products From Israel Ltd Device, system, and methods for entering commands or characters using a touch screen
DE102018100196A1 (en) * 2018-01-05 2019-07-11 Bcs Automotive Interface Solutions Gmbh Method for operating a human-machine interface and human-machine interface
US11449167B2 (en) 2017-06-26 2022-09-20 Inpris Innovative Products Fromisrael, Ltd Systems using dual touch and sound control, and methods thereof

Family Cites Families (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7030861B1 (en) * 2001-02-10 2006-04-18 Wayne Carl Westerman System and method for packing multi-touch gestures onto a hand
US8619048B2 (en) * 2008-08-08 2013-12-31 Moonsun Io Ltd. Method and device of stroke based user input
US9405404B2 (en) * 2010-03-26 2016-08-02 Autodesk, Inc. Multi-touch marking menus and directional chording gestures
US20110292268A1 (en) * 2010-05-26 2011-12-01 T-Mobile Usa, Inc. Multi-region touchpad device
US9235340B2 (en) * 2011-02-18 2016-01-12 Microsoft Technology Licensing, Llc Modal touch input
US9261972B2 (en) * 2011-04-21 2016-02-16 Inpris Innovative Products Ltd Ergonomic motion detection for receiving character input to electronic devices
US8970519B2 (en) * 2012-02-01 2015-03-03 Logitech Europe S.A. System and method for spurious signal detection and compensation on an input device
US20130194235A1 (en) * 2012-02-01 2013-08-01 Logitec Europe S.A. Multi-sensor input device
US9075462B2 (en) * 2012-12-10 2015-07-07 Sap Se Finger-specific input on touchscreen devices

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10146428B2 (en) 2011-04-21 2018-12-04 Inpris Innovative Products From Israel Ltd Device, system, and methods for entering commands or characters using a touch screen
US10120567B2 (en) 2015-04-02 2018-11-06 Inpris Innovative Products From Israel Ltd System, apparatus and method for vehicle command and control
US11449167B2 (en) 2017-06-26 2022-09-20 Inpris Innovative Products Fromisrael, Ltd Systems using dual touch and sound control, and methods thereof
DE102018100196A1 (en) * 2018-01-05 2019-07-11 Bcs Automotive Interface Solutions Gmbh Method for operating a human-machine interface and human-machine interface

Also Published As

Publication number Publication date
EP3338172A2 (en) 2018-06-27
EP3338172A4 (en) 2019-07-03
CN108780365B (en) 2020-07-14
CN108780365A (en) 2018-11-09
WO2017029555A3 (en) 2017-04-27

Similar Documents

Publication Publication Date Title
US10146428B2 (en) Device, system, and methods for entering commands or characters using a touch screen
EP2541385B1 (en) Information processing apparatus, information processing method, program and remote control system
US10120567B2 (en) System, apparatus and method for vehicle command and control
US9111076B2 (en) Mobile terminal and control method thereof
EP2509335B1 (en) Remote operation device, remote operation system, remote operation method, and program
EP3338172A2 (en) Device, system, and methods for entering commands or characters using a touch screen
US9703375B2 (en) Operating device that can be operated without keys
US20170060343A1 (en) Field analysis for flexible computer inputs
US9189094B2 (en) Display control apparatus and display system with pointer correction
US9811200B2 (en) Touch input device, vehicle including the touch input device, and method for controlling the touch input device
US9335822B2 (en) Method and system for providing haptic effects based on haptic context information
JP5556398B2 (en) Information processing apparatus, information processing method, and program
KR101685891B1 (en) Controlling apparatus using touch input and controlling method of the same
US20110025718A1 (en) Information input device and information input method
EP2671136A2 (en) Correcting typing mistake based on probabilities of intended contact for non-contacted keys
JP5374564B2 (en) Drawing apparatus, drawing control method, and drawing control program
CN105607770B (en) Touch input device and vehicle including the same
EP2474890A1 (en) Virtual keyboard configuration putting fingers in rest positions on a multitouch screen, calibrating key positions thereof
US20190212910A1 (en) Method for operating a human-machine interface and human-machine interface
US20200241742A1 (en) Touch input device and vehicle including the same
CN110045815A (en) For running the method and man-machine interface of man-machine interface
US20180150136A1 (en) Motor vehicle operator control device with touchscreen operation
US20140292689A1 (en) Input device, input method, and recording medium
US20160154488A1 (en) Integrated controller system for vehicle
CN110780732A (en) Input system based on space positioning and finger clicking

Legal Events

Date Code Title Description
NENP Non-entry into the national phase

Ref country code: DE

WWE Wipo information: entry into national phase

Ref document number: 2016836703

Country of ref document: EP

121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 16836703

Country of ref document: EP

Kind code of ref document: A2