US20200225799A1 - Gaze detection interlock feature for touch screen devices - Google Patents

Gaze detection interlock feature for touch screen devices Download PDF

Info

Publication number
US20200225799A1
US20200225799A1 US16/249,472 US201916249472A US2020225799A1 US 20200225799 A1 US20200225799 A1 US 20200225799A1 US 201916249472 A US201916249472 A US 201916249472A US 2020225799 A1 US2020225799 A1 US 2020225799A1
Authority
US
United States
Prior art keywords
user
gaze
display screen
processor
touch
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US16/249,472
Other versions
US10761648B2 (en
Inventor
Michael D. Marra
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Individual
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to US16/249,472 priority Critical patent/US10761648B2/en
Publication of US20200225799A1 publication Critical patent/US20200225799A1/en
Application granted granted Critical
Publication of US10761648B2 publication Critical patent/US10761648B2/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/041Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
    • G06F3/0416Control or interface arrangements specially adapted for digitisers
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • G06F3/013Eye tracking input arrangements
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures

Definitions

  • the present disclosure relates generally to a method for controlling user interaction with a touch screen device. More particularly, it relates to a method for controlling user interaction with a touch screen device, such as a smart phone or tablet, where the device, if unlocked, only responds to the user's screen touches when the user is looking at the device screen.
  • a touch screen device such as a smart phone or tablet
  • the gaze-detection feature prevents unwanted device actions which are commonly encountered when a user is simply trying to handle the device.
  • smart phones and tablet devices all have a security locking feature, where the device may be configured to lock automatically after a user-defined period of inactivity, and the device can be instantly locked by a user command or button push.
  • touch screen access to the apps is disabled, and only a very limited subset of touch screen commands are available—such as entering a passcode to unlock the phone, and placing an emergency (“911”) phone call.
  • a common problem with touch screen devices is that, when the device is unlocked, almost any touch of the screen will trigger an action. Even an accidental or inadvertent touch by the user, such as when the user is attempting to set the device down on a desk or a vehicle console, will trigger an unintended action. For example, if the user has just ended a phone call (and the smart phone is unlocked and in the phone app), an accidental screen touch might place an unwanted phone call to someone on the recent call list. As another example, if the user has started navigation using a GPS app, and accidentally touches the screen while setting the phone down, this could cause the map to zoom or pan in an undesired way, change the destination data, or end navigation. These unintended actions at a minimum are frustrating to the user, and may even be distracting and dangerous in the context of a driving situation.
  • a touch screen device user is forced to either lock the device after using it, or carefully handle the device to avoid touching any part of the display screen.
  • Most users find both of these workarounds to be inconvenient and annoying.
  • locking the device defeats the purpose of using GPS navigation apps and many other apps, and is therefore not a viable workaround.
  • Facial recognition systems are known to be available on some touch screen devices. These facial recognition systems employ a user-facing or “front-facing” camera (that is, a camera having its lens on the display screen side of the device) to recognize the user based on a previously-defined facial image map, where the facial recognition can be used as a form of security authentication in addition to or in lieu of a passcode. Some existing facial recognition systems also include gaze detection technology which recognizes whether the user's eyes are open and the user's attention (visual gaze) is directed towards the device screen.
  • gaze detection technology which recognizes whether the user's eyes are open and the user's attention (visual gaze) is directed towards the device screen.
  • the present disclosure describes a touch screen device and a method of controlling the device wherein the device responds to user touch screen inputs only when the device is unlocked and the user's visual gaze is directed at the device screen.
  • the device employs a front-facing camera and gaze detection technology to determine whether the user is looking directly at the device display screen. If the user is not looking directly at the display screen, touch input commands are ignored by the device, even if the device is unlocked, thereby preventing unintended application actions due to incidental screen contact.
  • the gaze detection interlock feature may be enabled or disabled through device configuration settings. The gaze detection interlock feature, if enabled, may respond only to the primary device user's face and visual gaze, or may respond when any person's visual gaze is directed to the device display screen.
  • FIG. 1 is an illustration of a smart phone device in a user's hand, where the thumb of the hand is making unintended incidental contact with the touch screen, as known in the art;
  • FIG. 2 is an illustration of a smart phone device in a user's hand, where several parts of the hand are making unintended incidental contact with the touch screen, as known in the art;
  • FIG. 3 is an illustration of a smart phone device in a user's hand, where the user is carefully gripping the device around the edge of the case in order to avoid making unintended incidental contact with the touch screen, as known in the art;
  • FIG. 4 is an illustration of a smart phone device in a user's hand, where the user's visual gaze is directed at the display screen, causing the device to respond to touch screen inputs, according to an embodiment of the present disclosure
  • FIG. 5 is an illustration of a smart phone device in a user's hand, where the user's visual gaze is not directed at the display screen, causing the device to ignore touch screen inputs, according to an embodiment of the present disclosure
  • FIG. 6 is a flowchart diagram of a method for controlling user interaction with the touch screen device of FIGS. 4 and 5 , according to an embodiment of the present disclosure.
  • Smart phones, tablet devices and personal computers have become an important part of everyday life.
  • One feature which has made these computing devices so powerful and easy to use is the touch screen display.
  • Virtually all smart phones and tablet devices now include a touch screen display.
  • Many laptop computers, desktop computer monitors, gaming systems and video display devices such as televisions also now include touch screen displays.
  • the gaze detection interlock feature of the present disclosure is particularly applicable to smart phones because of the manner in which smart phones are handled by users. For this reason, the following detailed discussion of the disclosed techniques is based on smart phone usage scenarios. However, it is to be understood that the disclosed techniques are applicable to all of the device types mentioned above, and any other type of touch screen device.
  • apps applications
  • data belonging to the user/owner smart phones and tablet devices have a security locking feature.
  • touch screen access to the apps is disabled, and only a very limited subset of touch screen commands are available—such as entering a passcode to unlock the device, and placing an emergency (“911”) phone call.
  • Phones and tablets typically may be configured to lock automatically after a user-defined period of inactivity, and the devices can be instantly locked by a user command or button push.
  • a common problem with smart phones is that, when the device is unlocked, almost any touch of the screen will trigger an action. Even an accidental or inadvertent touch by the user, such as when the user is attempting to set the phone down on a desk or a vehicle console, will trigger an unintended action.
  • FIG. 1 is an illustration of a smart phone 100 in a user's hand 110 , where the thumb of the hand 110 is making unintended incidental contact with a touch screen 102 .
  • the user is holding the smart phone 100 in a manner commonly used after entering information with the thumb (such as typing a text message, placing or ending a phone call, or operating any other app).
  • the thumb can inadvertently contact the touch screen 102 at a location 112 .
  • the thumb touch at the location 112 will cause an unintended application action if the smart phone 100 is unlocked.
  • FIG. 2 is an illustration of the smart phone 100 in the user's hand 110 , where the thumb and fingers of the hand 110 are making unintended incidental contact with the touch screen 102 .
  • the user is holding the smart phone 100 in a manner commonly used for securely gripping any item in the hand 110 .
  • the fingers can inadvertently contact the touch screen 102 at a location 202 or 204 , and/or the thumb can inadvertently contact the touch screen 102 at a location 206 .
  • a screen touch at any of the locations 202 / 204 / 206 will cause an unintended application action if the smart phone 100 is unlocked.
  • Unintended application actions of the type illustrated in FIGS. 1 and 2 can include launching and executing commands and operations in any application installed on the smart phone 100 .
  • an accidental screen touch might place an unwanted phone call to someone on the recent call list.
  • the user has started navigation using a GPS app, and accidentally touches the screen 102 while setting the phone 100 down in the vehicle console, this could cause the map to zoom or pan in an undesired way, change the navigation destination data, or end navigation.
  • the possibilities for unintended application actions are endless—including other examples such as accidentally starting music or video streaming, accidentally executing a transaction in a mobile banking app, accidentally making a purchase from a retailer app or website, etc.
  • a smart phone user is forced to either lock the smart phone 100 after using it, or carefully handle the phone 100 to avoid touching any part of the touch screen 102 .
  • FIG. 3 is an illustration of the smart phone 100 in the user's hand 110 , where the user is gripping the phone 100 around the edge of the case in order to avoid making incidental contact with the touch screen 102 .
  • Carefully gripping the smart phone 100 in the manner shown in FIG. 3 is one way to avoid the unintended application actions described above.
  • the smart phone 100 can also be held with the hand 110 open and flat, palm facing up, in order to avoid incidental screen contact. Most users find these types of careful gripping techniques to be inconvenient and annoying.
  • grasping the smart phone 100 without touching the screen 102 results in an inherently insecure grip, causing phones to be dropped and often damaged. This problem has given rise to the so-called “pop-socket” device which can be attached to the back of the phone 100 .
  • locking the smart phone 100 is another way to avoid unintended application actions.
  • locking the phone 100 defeats the purpose of using GPS navigation apps and many other apps, and is therefore not a viable workaround in many situations.
  • none of the workarounds described above actually resolve the real problem, which is that the smart phone 100 actively responds to user touches of the screen 102 when the user does not want it to do so.
  • the present disclosure describes a technique for controlling a smart phone or other touch screen device wherein the device only responds to user touch inputs when the user is looking directly at the touch screen.
  • the disclosed techniques are based on the premise that, if the user wants the touch screen device to respond to touch commands, the user will be looking at the touch screen, and conversely, if the user is not looking at the screen, then the user does not want the touch screen device to respond to touch commands.
  • Facial recognition systems are known to be available on some touch screen devices. These facial recognition systems employ a user-facing or “front-facing” camera (that is, a camera having its lens on the display screen side of the device) to recognize the user based on a previously-defined facial image map.
  • the facial recognition can be used as a form of security authentication in addition to, or in lieu of, a passcode.
  • Some existing facial recognition systems also include gaze detection technology which recognizes whether the user's eyes are open and the user's attention (visual gaze) is directed towards the device screen.
  • a typical gaze detection subsystem includes a front-facing camera and possibly other sensors providing images and data to the device's processor which runs an algorithm to determine whether the user is looking directly at the screen. This gaze detection technology can be employed to control the touch screen device operation.
  • the disclosed technique which might be called a gaze detection interlock feature, causes the touch screen device to respond to user inputs only when the user's visual gaze is directed toward the screen.
  • FIG. 4 is an illustration of a smart phone 400 being held by a user 410 , where the user's visual gaze is directed at a display screen 402 , causing the smart phone 400 to respond to touch screen inputs, according to an embodiment of the present disclosure.
  • the smart phone 400 employs a front-facing camera 404 and/or other sensors 406 to provide images and/or geometric shape data describing the user's face.
  • the other sensors 406 may include any type of sensor suitable for providing images or shape data of the user's face—including but not limited to another visual-light camera, an infrared camera, an ultrasonic sensor, a LiDAR sensor, etc.
  • the images and/or shape data are processed using known gaze detection technology (for example, the locations of the irises and pupils within the eye outline) to determine whether the user 410 is looking directly at the screen 402 .
  • the user 410 is looking directly at the screen 402 , as indicated by dashed lines 414 .
  • Existing gaze detection technology can easily and reliably determine that the user's gaze is directed to the screen 402 and not, for example, at a distant scene 420 . Because the user's gaze is directed to the touch screen 402 , the smart phone 400 responds normally to any touch of the screen 402 .
  • the smart phone 400 is shown in a hand 412 of the user 410 , with the thumb of the hand 412 being used in typical fashion to input touch commands on the screen 402 .
  • FIG. 5 is an illustration of the smart phone 400 being held by a user 510 , where the user's visual gaze is not directed at the display screen 402 , causing the smart phone 400 to ignore touch screen inputs, according to an embodiment of the present disclosure.
  • the user 510 is looking at a distant scene 520 as indicated by dashed lines 514 .
  • the gaze detection subsystem on the smart phone 400 can readily determine that the user's gaze is directed to the distant scene 520 and not at the touch screen 402 .
  • the gaze detection interlock feature is activated, meaning that the smart phone 400 ignores any touch of the screen 402 , such as by the thumb of the hand 512 . As illustrated in FIG. 5 , the gaze detection interlock feature causes touch input commands to be ignored by the smart phone 400 , even if the 400 is unlocked, thereby preventing unintended application actions due to incidental screen contact.
  • the gaze detection interlock feature may be enabled or disabled through device configuration settings. That is, in a settings menu on the phone 400 , the user may enable the gaze detection interlock feature (causing the phone 400 to behave as described in the discussion of FIGS. 5 and 6 ), or the user may disable the gaze detection interlock feature (causing the phone 400 to always respond to user touch inputs when the phone 400 is unlocked, regardless of whether the user is looking at the screen). Enabling and disabling the gaze detection interlock feature is also preferably possible via voice command, to make it easy for the device owner to turn the feature on or off in a hands-free manner.
  • the gaze detection interlock feature may respond only to the primary device user's face and visual gaze, or may respond when any person's visual gaze is directed to the device display screen. Responding only to the primary device user's face and visual gaze (as determined by the facial recognition security feature discussed earlier) may be advantageous, for example, if the user wants to ensure that he or she is the only person who can control the device for security reasons. Responding when any person's visual gaze is directed to the device display screen may be advantageous, for example, when a vehicle driver wants to hand the phone to a passenger to enter a navigation destination. These options can also be defined via device configuration settings, either temporary or permanent.
  • FIG. 6 is a flowchart diagram 600 of a method for controlling user interaction with a touch screen device such as the smart phone 400 of FIGS. 4 and 5 , according to an embodiment of the present disclosure.
  • a touch screen device such as the smart phone 400 in an unlocked state is provided.
  • the touch screen device determines the direction of the user's visual gaze. This is done using the gaze detection subsystem discussed earlier, including at least one camera communicating with the device processor.
  • decision diamond 606 it is determined whether the user's visual gaze is directed at the device screen. If the user's visual gaze is directed at the device screen, then at box 608 the touch screen device responds to user screen touches in a normal fashion. If the user's visual gaze is not directed at the device screen, then at box 610 the touch screen device ignores user screen touches, even if the device is unlocked.
  • the determination of the direction of the user's gaze must be repeated periodically on an ongoing basis by the touch screen device, because the user may be looking at the device screen in one moment, and not looking at the device screen in the next moment, or vice versa.
  • the process loops back to the box 604 to once again determine the direction of the user's gaze.
  • the re-determination of the direction of the user's gaze may be performed at regular time intervals, such as every one-tenth of a second or every one second, or based on other criteria or operating system considerations.
  • the gaze detection interlock feature provides advantages not found in prior art systems. For example, existing facial recognition systems are used only to unlock a device, and do not consider whether the user's gaze is directed to the screen after the device is unlocked. Also, existing eye tracking applications may be used to control a device in lieu of actual touch commands, but do not provide the interlock feature of the present disclosure.
  • the several and various features and process steps discussed herein to describe the disclosed methods may be referring to operations performed by a computer, a processor or other electronic calculating device that manipulates and/or transforms data using electrical phenomenon.
  • this refers to the processor or processors in the smart phone 400 , said processors performing calculations as part of the gaze detection subsystem, running a device operating system and executing applications, and also controlling touch screen operation via the disclosed gaze detection interlock feature.
  • processors and electronic devices may employ various volatile and/or non-volatile memories including non-transitory computer-readable medium with an executable program stored thereon including various code or executable instructions able to be performed by the computer or processor, where the memory and/or computer-readable medium may include all forms and types of memory and other computer-readable media.
  • the disclosed gaze detection interlock feature for controlling user interaction with a touch screen device provides a way to prevent unintended application actions, thereby reducing user frustration and increasing user satisfaction with electronic devices.
  • the disclosed techniques are implementable without adding any new hardware or cost to modern multi-camera touch screen devices.

Abstract

A touch screen device and method of controlling the device wherein the device responds to user touch screen inputs only when the device is unlocked and the user's visual gaze is directed at the device screen. The device employs a front-facing camera and gaze detection technology to determine whether the user is looking directly at the device display screen. If the user is not looking at the screen, touch input commands are ignored by the device, even if the device is unlocked, thereby preventing unintended application actions due to incidental screen contact. The gaze detection interlock feature may be enabled or disabled through configuration settings. The gaze detection interlock feature, if enabled, may respond only to the primary device user's face and visual gaze, or may respond when any person's visual gaze is directed to the device screen.

Description

    BACKGROUND Field
  • The present disclosure relates generally to a method for controlling user interaction with a touch screen device. More particularly, it relates to a method for controlling user interaction with a touch screen device, such as a smart phone or tablet, where the device, if unlocked, only responds to the user's screen touches when the user is looking at the device screen. The gaze-detection feature prevents unwanted device actions which are commonly encountered when a user is simply trying to handle the device.
  • Discussion of the Related Art
  • The capabilities of smart phones and tablet devices have increased dramatically since they were first introduced. Modern designs for almost all such devices now employ a touch screen user interface, where virtually the entire screen surface is used both for graphical display and for receiving user commands via touch. The touch screen interface offers application (“app”) designers almost unlimited flexibility in designing apps—from simple apps such as phone dialing and text messaging to complex and graphically-intensive apps such as gaming and virtual reality.
  • Because they provide access to sensitive personal information and data belonging to the user/owner, smart phones and tablet devices all have a security locking feature, where the device may be configured to lock automatically after a user-defined period of inactivity, and the device can be instantly locked by a user command or button push. When a device is locked, touch screen access to the apps is disabled, and only a very limited subset of touch screen commands are available—such as entering a passcode to unlock the phone, and placing an emergency (“911”) phone call.
  • A common problem with touch screen devices is that, when the device is unlocked, almost any touch of the screen will trigger an action. Even an accidental or inadvertent touch by the user, such as when the user is attempting to set the device down on a desk or a vehicle console, will trigger an unintended action. For example, if the user has just ended a phone call (and the smart phone is unlocked and in the phone app), an accidental screen touch might place an unwanted phone call to someone on the recent call list. As another example, if the user has started navigation using a GPS app, and accidentally touches the screen while setting the phone down, this could cause the map to zoom or pan in an undesired way, change the destination data, or end navigation. These unintended actions at a minimum are frustrating to the user, and may even be distracting and dangerous in the context of a driving situation.
  • In order to avoid the unintended actions described above, a touch screen device user is forced to either lock the device after using it, or carefully handle the device to avoid touching any part of the display screen. Most users find both of these workarounds to be inconvenient and annoying. Furthermore, locking the device defeats the purpose of using GPS navigation apps and many other apps, and is therefore not a viable workaround.
  • Facial recognition systems are known to be available on some touch screen devices. These facial recognition systems employ a user-facing or “front-facing” camera (that is, a camera having its lens on the display screen side of the device) to recognize the user based on a previously-defined facial image map, where the facial recognition can be used as a form of security authentication in addition to or in lieu of a passcode. Some existing facial recognition systems also include gaze detection technology which recognizes whether the user's eyes are open and the user's attention (visual gaze) is directed towards the device screen.
  • In view of the circumstances described above, there is an opportunity to incorporate visual gaze detection into touch screen device operation in order to prevent unintended device actions.
  • SUMMARY
  • The present disclosure describes a touch screen device and a method of controlling the device wherein the device responds to user touch screen inputs only when the device is unlocked and the user's visual gaze is directed at the device screen. The device employs a front-facing camera and gaze detection technology to determine whether the user is looking directly at the device display screen. If the user is not looking directly at the display screen, touch input commands are ignored by the device, even if the device is unlocked, thereby preventing unintended application actions due to incidental screen contact. The gaze detection interlock feature may be enabled or disabled through device configuration settings. The gaze detection interlock feature, if enabled, may respond only to the primary device user's face and visual gaze, or may respond when any person's visual gaze is directed to the device display screen.
  • Additional features of the present disclosure will become apparent from the following description and appended claims, taken in conjunction with the accompanying drawings.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 is an illustration of a smart phone device in a user's hand, where the thumb of the hand is making unintended incidental contact with the touch screen, as known in the art;
  • FIG. 2 is an illustration of a smart phone device in a user's hand, where several parts of the hand are making unintended incidental contact with the touch screen, as known in the art;
  • FIG. 3 is an illustration of a smart phone device in a user's hand, where the user is carefully gripping the device around the edge of the case in order to avoid making unintended incidental contact with the touch screen, as known in the art;
  • FIG. 4 is an illustration of a smart phone device in a user's hand, where the user's visual gaze is directed at the display screen, causing the device to respond to touch screen inputs, according to an embodiment of the present disclosure;
  • FIG. 5 is an illustration of a smart phone device in a user's hand, where the user's visual gaze is not directed at the display screen, causing the device to ignore touch screen inputs, according to an embodiment of the present disclosure; and
  • FIG. 6 is a flowchart diagram of a method for controlling user interaction with the touch screen device of FIGS. 4 and 5, according to an embodiment of the present disclosure.
  • DETAILED DESCRIPTION OF THE EMBODIMENTS
  • The following discussion of the embodiments of the disclosure directed to a touch screen device and a method of controlling the device including a gaze detection interlock feature is merely exemplary in nature, and is in no way intended to limit the disclosure or its applications or uses.
  • Smart phones, tablet devices and personal computers have become an important part of everyday life. One feature which has made these computing devices so powerful and easy to use is the touch screen display. Virtually all smart phones and tablet devices now include a touch screen display. Many laptop computers, desktop computer monitors, gaming systems and video display devices such as televisions also now include touch screen displays. The gaze detection interlock feature of the present disclosure is particularly applicable to smart phones because of the manner in which smart phones are handled by users. For this reason, the following detailed discussion of the disclosed techniques is based on smart phone usage scenarios. However, it is to be understood that the disclosed techniques are applicable to all of the device types mentioned above, and any other type of touch screen device.
  • Because they provide access to sensitive personal information, applications (“apps”) and data belonging to the user/owner, smart phones and tablet devices have a security locking feature. When a device is locked, touch screen access to the apps is disabled, and only a very limited subset of touch screen commands are available—such as entering a passcode to unlock the device, and placing an emergency (“911”) phone call. Phones and tablets typically may be configured to lock automatically after a user-defined period of inactivity, and the devices can be instantly locked by a user command or button push.
  • A common problem with smart phones is that, when the device is unlocked, almost any touch of the screen will trigger an action. Even an accidental or inadvertent touch by the user, such as when the user is attempting to set the phone down on a desk or a vehicle console, will trigger an unintended action.
  • FIG. 1 is an illustration of a smart phone 100 in a user's hand 110, where the thumb of the hand 110 is making unintended incidental contact with a touch screen 102. In FIG. 1, the user is holding the smart phone 100 in a manner commonly used after entering information with the thumb (such as typing a text message, placing or ending a phone call, or operating any other app). Unfortunately for users, when holding the smart phone 100 in this manner, and then attempting to set the phone 100 down in a vehicle console or on a desktop or other surface, the thumb can inadvertently contact the touch screen 102 at a location 112. The thumb touch at the location 112 will cause an unintended application action if the smart phone 100 is unlocked.
  • FIG. 2 is an illustration of the smart phone 100 in the user's hand 110, where the thumb and fingers of the hand 110 are making unintended incidental contact with the touch screen 102. In FIG. 2, the user is holding the smart phone 100 in a manner commonly used for securely gripping any item in the hand 110. Unfortunately, when holding the smart phone 100 in this manner, whether simply carrying the smart phone 100 or attempting to place the phone 100 on a surface or in a pocket or purse, the fingers can inadvertently contact the touch screen 102 at a location 202 or 204, and/or the thumb can inadvertently contact the touch screen 102 at a location 206. A screen touch at any of the locations 202/204/206 will cause an unintended application action if the smart phone 100 is unlocked.
  • Unintended application actions of the type illustrated in FIGS. 1 and 2 can include launching and executing commands and operations in any application installed on the smart phone 100.
  • For example, if the user has just ended a phone call (and the smart phone 100 is unlocked and displaying the phone app), an accidental screen touch might place an unwanted phone call to someone on the recent call list. As another example, if the user has started navigation using a GPS app, and accidentally touches the screen 102 while setting the phone 100 down in the vehicle console, this could cause the map to zoom or pan in an undesired way, change the navigation destination data, or end navigation. The possibilities for unintended application actions are endless—including other examples such as accidentally starting music or video streaming, accidentally executing a transaction in a mobile banking app, accidentally making a purchase from a retailer app or website, etc.
  • These unintended actions, at a minimum, are frustrating to the user. Furthermore, these unintended actions may create costs or consequences for the user (accidental purchase of an item, for example), and may even be distracting and dangerous in the context of a driving situation.
  • In order to avoid the unintended application actions described above, a smart phone user is forced to either lock the smart phone 100 after using it, or carefully handle the phone 100 to avoid touching any part of the touch screen 102.
  • FIG. 3 is an illustration of the smart phone 100 in the user's hand 110, where the user is gripping the phone 100 around the edge of the case in order to avoid making incidental contact with the touch screen 102.
  • Carefully gripping the smart phone 100 in the manner shown in FIG. 3 is one way to avoid the unintended application actions described above. The smart phone 100 can also be held with the hand 110 open and flat, palm facing up, in order to avoid incidental screen contact. Most users find these types of careful gripping techniques to be inconvenient and annoying. Furthermore, grasping the smart phone 100 without touching the screen 102 results in an inherently insecure grip, causing phones to be dropped and often damaged. This problem has given rise to the so-called “pop-socket” device which can be attached to the back of the phone 100.
  • As mentioned above, locking the smart phone 100 is another way to avoid unintended application actions. However, locking the phone 100 defeats the purpose of using GPS navigation apps and many other apps, and is therefore not a viable workaround in many situations. In fact, none of the workarounds described above actually resolve the real problem, which is that the smart phone 100 actively responds to user touches of the screen 102 when the user does not want it to do so.
  • According to the following discussion, the present disclosure describes a technique for controlling a smart phone or other touch screen device wherein the device only responds to user touch inputs when the user is looking directly at the touch screen. The disclosed techniques are based on the premise that, if the user wants the touch screen device to respond to touch commands, the user will be looking at the touch screen, and conversely, if the user is not looking at the screen, then the user does not want the touch screen device to respond to touch commands.
  • Facial recognition systems are known to be available on some touch screen devices. These facial recognition systems employ a user-facing or “front-facing” camera (that is, a camera having its lens on the display screen side of the device) to recognize the user based on a previously-defined facial image map. The facial recognition can be used as a form of security authentication in addition to, or in lieu of, a passcode.
  • Some existing facial recognition systems also include gaze detection technology which recognizes whether the user's eyes are open and the user's attention (visual gaze) is directed towards the device screen. A typical gaze detection subsystem includes a front-facing camera and possibly other sensors providing images and data to the device's processor which runs an algorithm to determine whether the user is looking directly at the screen. This gaze detection technology can be employed to control the touch screen device operation. The disclosed technique, which might be called a gaze detection interlock feature, causes the touch screen device to respond to user inputs only when the user's visual gaze is directed toward the screen.
  • FIG. 4 is an illustration of a smart phone 400 being held by a user 410, where the user's visual gaze is directed at a display screen 402, causing the smart phone 400 to respond to touch screen inputs, according to an embodiment of the present disclosure. The smart phone 400 employs a front-facing camera 404 and/or other sensors 406 to provide images and/or geometric shape data describing the user's face. The other sensors 406 may include any type of sensor suitable for providing images or shape data of the user's face—including but not limited to another visual-light camera, an infrared camera, an ultrasonic sensor, a LiDAR sensor, etc. The images and/or shape data are processed using known gaze detection technology (for example, the locations of the irises and pupils within the eye outline) to determine whether the user 410 is looking directly at the screen 402.
  • In FIG. 4, the user 410 is looking directly at the screen 402, as indicated by dashed lines 414. Existing gaze detection technology can easily and reliably determine that the user's gaze is directed to the screen 402 and not, for example, at a distant scene 420. Because the user's gaze is directed to the touch screen 402, the smart phone 400 responds normally to any touch of the screen 402. The smart phone 400 is shown in a hand 412 of the user 410, with the thumb of the hand 412 being used in typical fashion to input touch commands on the screen 402.
  • FIG. 5 is an illustration of the smart phone 400 being held by a user 510, where the user's visual gaze is not directed at the display screen 402, causing the smart phone 400 to ignore touch screen inputs, according to an embodiment of the present disclosure. In FIG. 5, although holding the phone 400 in a hand 512, the user 510 is looking at a distant scene 520 as indicated by dashed lines 514. The gaze detection subsystem on the smart phone 400 can readily determine that the user's gaze is directed to the distant scene 520 and not at the touch screen 402. Because the user's gaze is not directed to the touch screen 402, the gaze detection interlock feature is activated, meaning that the smart phone 400 ignores any touch of the screen 402, such as by the thumb of the hand 512. As illustrated in FIG. 5, the gaze detection interlock feature causes touch input commands to be ignored by the smart phone 400, even if the 400 is unlocked, thereby preventing unintended application actions due to incidental screen contact.
  • The gaze detection interlock feature may be enabled or disabled through device configuration settings. That is, in a settings menu on the phone 400, the user may enable the gaze detection interlock feature (causing the phone 400 to behave as described in the discussion of FIGS. 5 and 6), or the user may disable the gaze detection interlock feature (causing the phone 400 to always respond to user touch inputs when the phone 400 is unlocked, regardless of whether the user is looking at the screen). Enabling and disabling the gaze detection interlock feature is also preferably possible via voice command, to make it easy for the device owner to turn the feature on or off in a hands-free manner.
  • The gaze detection interlock feature, if enabled, may respond only to the primary device user's face and visual gaze, or may respond when any person's visual gaze is directed to the device display screen. Responding only to the primary device user's face and visual gaze (as determined by the facial recognition security feature discussed earlier) may be advantageous, for example, if the user wants to ensure that he or she is the only person who can control the device for security reasons. Responding when any person's visual gaze is directed to the device display screen may be advantageous, for example, when a vehicle driver wants to hand the phone to a passenger to enter a navigation destination. These options can also be defined via device configuration settings, either temporary or permanent.
  • FIG. 6 is a flowchart diagram 600 of a method for controlling user interaction with a touch screen device such as the smart phone 400 of FIGS. 4 and 5, according to an embodiment of the present disclosure. At box 602, a touch screen device (such as the smart phone 400) in an unlocked state is provided. At box 604, the touch screen device determines the direction of the user's visual gaze. This is done using the gaze detection subsystem discussed earlier, including at least one camera communicating with the device processor. At decision diamond 606, it is determined whether the user's visual gaze is directed at the device screen. If the user's visual gaze is directed at the device screen, then at box 608 the touch screen device responds to user screen touches in a normal fashion. If the user's visual gaze is not directed at the device screen, then at box 610 the touch screen device ignores user screen touches, even if the device is unlocked.
  • The determination of the direction of the user's gaze must be repeated periodically on an ongoing basis by the touch screen device, because the user may be looking at the device screen in one moment, and not looking at the device screen in the next moment, or vice versa. Thus, from both the box 608 and the box 610, the process loops back to the box 604 to once again determine the direction of the user's gaze. The re-determination of the direction of the user's gaze may be performed at regular time intervals, such as every one-tenth of a second or every one second, or based on other criteria or operating system considerations.
  • The gaze detection interlock feature provides advantages not found in prior art systems. For example, existing facial recognition systems are used only to unlock a device, and do not consider whether the user's gaze is directed to the screen after the device is unlocked. Also, existing eye tracking applications may be used to control a device in lieu of actual touch commands, but do not provide the interlock feature of the present disclosure.
  • As will be well understood by those skilled in the art, the several and various features and process steps discussed herein to describe the disclosed methods may be referring to operations performed by a computer, a processor or other electronic calculating device that manipulates and/or transforms data using electrical phenomenon. In particular, this refers to the processor or processors in the smart phone 400, said processors performing calculations as part of the gaze detection subsystem, running a device operating system and executing applications, and also controlling touch screen operation via the disclosed gaze detection interlock feature. Those processors and electronic devices may employ various volatile and/or non-volatile memories including non-transitory computer-readable medium with an executable program stored thereon including various code or executable instructions able to be performed by the computer or processor, where the memory and/or computer-readable medium may include all forms and types of memory and other computer-readable media.
  • The disclosed gaze detection interlock feature for controlling user interaction with a touch screen device provides a way to prevent unintended application actions, thereby reducing user frustration and increasing user satisfaction with electronic devices. The disclosed techniques are implementable without adding any new hardware or cost to modern multi-camera touch screen devices.
  • The foregoing discussion discloses and describes merely exemplary embodiments of the present disclosure. One skilled in the art will readily recognize from such discussion and from the accompanying drawings and claims that various changes, modifications and variations can be made therein without departing from the spirit and scope of the disclosure as defined in the following claims.

Claims (20)

1. A method for controlling a touch screen device, said method comprising:
providing the device having a display screen responsive to touch inputs by a user, at least one front-facing camera configured to detect a gaze direction of the user's eyes, and a processor with memory;
unlocking the device, by the user, to enable interaction with one or more application programs executable by the processor;
determining, by the processor analyzing images from the camera, whether the user's gaze is directed toward the display screen;
responding to the touch inputs to perform application operations, by the processor, when the device is unlocked and the user's gaze is directed toward the display screen; and
ignoring the touch inputs on the display screen, by the processor, when the user's gaze is not directed toward the display screen.
2. The method according to claim 1 wherein, in determining whether the user's gaze is directed toward the display screen, the processor also uses images or data from one or more of second camera or a front-facing sensor.
3. The method according to claim 2 wherein the front-facing sensor includes an infrared sensor, an ultrasound sensor or a LiDAR sensor.
4. The method according to claim 1 wherein the touch screen device is a smart phone, a tablet device or a laptop computer.
5. A method for controlling a touch screen device, said method comprising:
providing the device having a display screen responsive to touch inputs by a user, at least one front-facing camera configured to detect a gaze direction of the user's eyes, and a processor with memory;
unlocking the device, by the user, to enable interaction with one or more application programs executable by the processor;
determining, by the processor analyzing images from the camera, whether the user's gaze is directed toward the display screen;
responding to the touch inputs to perform application operations, by the processor, when the device is unlocked and the user's gaze is directed toward the display screen; and
ignoring the touch inputs on the display screen, by the processor when a gaze detection interlock feature is enabled, when the user's gaze is not directed toward the display screen.
6. The method according to claim 5 wherein, in determining whether the user's gaze is directed toward the display screen, the processor also uses images or data from one or more of second front-facing camera or a front-facing sensor.
7. The method according to claim 6 wherein the front-facing sensor includes an infrared sensor, an ultrasound sensor or a LiDAR sensor.
8. The method according to claim 5 wherein the gaze detection interlock feature is enabled or disabled by the user via a configuration setting of the touch screen device, including a configuration setting which is controllable via voice command.
9. The method according to claim 5 wherein the gaze detection interlock feature is configured to recognize a difference between a primary user of the touch screen device and another user of the touch screen device.
10. The method according to claim 9 wherein the gaze detection interlock feature is configured to also ignore the touch inputs on the display screen when another user's gaze is directed toward the display screen.
11. The method according to claim 9 wherein the gaze detection interlock feature is configured to ignore the touch inputs on the display screen only when neither the primary user's gaze nor another user's gaze is directed toward the display screen.
12. The method according to claim 5 wherein determining whether the user's gaze is directed toward the display screen is repeated periodically when the touch screen device is unlocked.
13. The method according to claim 5 wherein the touch screen device is a smart phone, a tablet device or a laptop computer.
14. An electronic device comprising:
a display screen responsive to touch inputs by a user;
a front-facing camera or sensor configured to detect a gaze direction of the user's eyes; and
a processor with memory, said processor being in communication with the camera or sensor and receiving the touch inputs from the display screen, where the memory contains one or more application programs for execution by the processor, and where the processor is configured to execute from the memory a gaze detection interlock algorithm performing steps of;
determining whether the user's gaze is directed toward the display screen based on images or data from the camera or sensor,
responding to the touch inputs to perform application operations when the device is unlocked and the user's gaze is directed toward the display screen, and
ignoring the touch inputs when the user's gaze is not directed toward the display screen.
15. The device according to claim 14 further comprising an additional front-facing camera or sensor, wherein the processor also uses images or data from the additional front-facing camera or sensor in determining whether the user's gaze is directed toward the display screen.
16. The device according to claim 15 wherein the additional front-facing camera or sensor includes a visual image camera, an infrared sensor, an ultrasound sensor or a LiDAR sensor.
17. The device according to claim 14 wherein the gaze detection interlock algorithm is enabled or disabled by the user via a configuration setting of the touch screen device, including a configuration setting which is controllable via voice command.
18. The device according to claim 14 wherein the gaze detection interlock algorithm is configured to recognize an identity of a primary user of the touch screen device versus another user of the touch screen device, and the identity is usable by the algorithm in determining whether to respond to the touch inputs or ignore the touch inputs.
19. The device according to claim 14 wherein determining whether the user's gaze is directed toward the display screen is repeated periodically when the touch screen device is unlocked.
20. The device according to claim 14 wherein the touch screen device is a smart phone, a tablet device or a laptop computer.
US16/249,472 2019-01-16 2019-01-16 Gaze detection interlock feature for touch screen devices Active US10761648B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US16/249,472 US10761648B2 (en) 2019-01-16 2019-01-16 Gaze detection interlock feature for touch screen devices

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US16/249,472 US10761648B2 (en) 2019-01-16 2019-01-16 Gaze detection interlock feature for touch screen devices

Publications (2)

Publication Number Publication Date
US20200225799A1 true US20200225799A1 (en) 2020-07-16
US10761648B2 US10761648B2 (en) 2020-09-01

Family

ID=71517618

Family Applications (1)

Application Number Title Priority Date Filing Date
US16/249,472 Active US10761648B2 (en) 2019-01-16 2019-01-16 Gaze detection interlock feature for touch screen devices

Country Status (1)

Country Link
US (1) US10761648B2 (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11303649B2 (en) * 2019-05-30 2022-04-12 International Business Machines Corporation Maintaining electronic communications session continuity during session inactivity
JP2023020879A (en) * 2021-07-30 2023-02-09 功憲 末次 Inappropriate use control system and inappropriate use control program
JP2023060890A (en) * 2021-07-30 2023-04-28 功憲 末次 Inappropriate use control system and inappropriate use control program
JP7331281B2 (en) 2021-07-30 2023-08-22 功憲 末次 Improper Use Control System and Improper Use Control Program
JP7331282B2 (en) 2021-07-30 2023-08-22 功憲 末次 Improper Use Control System and Improper Use Control Program
US11886317B2 (en) 2021-07-30 2024-01-30 Katsunori SUETSUGU Inappropriate use control system and inappropriate use control program

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060022959A1 (en) * 2001-07-09 2006-02-02 Geaghan Bernard O Touch screen with selective touch sources
US8913004B1 (en) * 2010-03-05 2014-12-16 Amazon Technologies, Inc. Action based device control
US20150035776A1 (en) * 2012-03-23 2015-02-05 Ntt Docomo, Inc. Information terminal, method for controlling input acceptance, and program for controlling input acceptance
US20150338914A1 (en) * 2013-11-01 2015-11-26 Intel Corporation Gaze-assisted touchscreen inputs
US20160275314A1 (en) * 2014-04-28 2016-09-22 Sony Corporation Operating a display of a user equipment
US20180088665A1 (en) * 2016-09-26 2018-03-29 Lenovo (Singapore) Pte. Ltd. Eye tracking selection validation
US9946371B2 (en) * 2014-10-16 2018-04-17 Qualcomm Incorporated System and method for using touch orientation to distinguish between users of a touch panel
US20190187786A1 (en) * 2017-12-15 2019-06-20 International Business Machines Corporation Preventing unintended input

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060022959A1 (en) * 2001-07-09 2006-02-02 Geaghan Bernard O Touch screen with selective touch sources
US8913004B1 (en) * 2010-03-05 2014-12-16 Amazon Technologies, Inc. Action based device control
US20150035776A1 (en) * 2012-03-23 2015-02-05 Ntt Docomo, Inc. Information terminal, method for controlling input acceptance, and program for controlling input acceptance
US20150338914A1 (en) * 2013-11-01 2015-11-26 Intel Corporation Gaze-assisted touchscreen inputs
US20160275314A1 (en) * 2014-04-28 2016-09-22 Sony Corporation Operating a display of a user equipment
US9946371B2 (en) * 2014-10-16 2018-04-17 Qualcomm Incorporated System and method for using touch orientation to distinguish between users of a touch panel
US20180088665A1 (en) * 2016-09-26 2018-03-29 Lenovo (Singapore) Pte. Ltd. Eye tracking selection validation
US20190187786A1 (en) * 2017-12-15 2019-06-20 International Business Machines Corporation Preventing unintended input

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11303649B2 (en) * 2019-05-30 2022-04-12 International Business Machines Corporation Maintaining electronic communications session continuity during session inactivity
JP2023020879A (en) * 2021-07-30 2023-02-09 功憲 末次 Inappropriate use control system and inappropriate use control program
JP2023060890A (en) * 2021-07-30 2023-04-28 功憲 末次 Inappropriate use control system and inappropriate use control program
JP7282235B2 (en) 2021-07-30 2023-05-26 功憲 末次 Improper Use Control System and Improper Use Control Program
JP7331281B2 (en) 2021-07-30 2023-08-22 功憲 末次 Improper Use Control System and Improper Use Control Program
JP7331282B2 (en) 2021-07-30 2023-08-22 功憲 末次 Improper Use Control System and Improper Use Control Program
JP7331283B2 (en) 2021-07-30 2023-08-22 功憲 末次 Improper Use Control System and Improper Use Control Program
US11886317B2 (en) 2021-07-30 2024-01-30 Katsunori SUETSUGU Inappropriate use control system and inappropriate use control program

Also Published As

Publication number Publication date
US10761648B2 (en) 2020-09-01

Similar Documents

Publication Publication Date Title
US10761648B2 (en) Gaze detection interlock feature for touch screen devices
US20220100841A1 (en) Authenticated device used to unlock another device
CN109753159B (en) Method and apparatus for controlling electronic device
US9733752B2 (en) Mobile terminal and control method thereof
US10956734B2 (en) Electronic device providing iris recognition based on proximity and operating method thereof
US8717393B2 (en) System and method for controlling a display of a mobile device
US20230014800A1 (en) User interface for accessing an account
US20130342672A1 (en) Using gaze determination with device input
US8938612B1 (en) Limited-access state for inadvertent inputs
AU2017291584B2 (en) Method for recognizing iris based on user intention and electronic device for the same
EP3259701B1 (en) Biometric setup that runs in the background
US20170123587A1 (en) Method and device for preventing accidental touch of terminal with touch screen
US9536073B2 (en) Device-based application security
US20180088665A1 (en) Eye tracking selection validation
KR102320072B1 (en) Electronic device and method for controlling of information disclosure thereof
WO2023030238A1 (en) Secure input method and apparatus
WO2023125238A1 (en) Unlocking control method and apparatus, electronic device, and readable storage medium
US20220221932A1 (en) Controlling a function via gaze detection
WO2021004352A1 (en) Interface interaction method and terminal
JP6132245B2 (en) Providing calls based on the detected direction
US9948767B2 (en) Rejecting or accepting a phone call
KR102660224B1 (en) Electronic device and method of providing notification information thereof
WO2022037247A1 (en) Device, method and system for operating device
CN107632754B (en) Key control method based on fingerprint identification module, mobile terminal and storage medium
CN112162810A (en) Message display method and device and electronic equipment

Legal Events

Date Code Title Description
FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: MICROENTITY

FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO MICRO (ORIGINAL EVENT CODE: MICR); ENTITY STATUS OF PATENT OWNER: MICROENTITY

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STCF Information on status: patent grant

Free format text: PATENTED CASE

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, MICRO ENTITY (ORIGINAL EVENT CODE: M3551); ENTITY STATUS OF PATENT OWNER: MICROENTITY

Year of fee payment: 4