US20200225799A1 - Gaze detection interlock feature for touch screen devices - Google Patents
Gaze detection interlock feature for touch screen devices Download PDFInfo
- Publication number
- US20200225799A1 US20200225799A1 US16/249,472 US201916249472A US2020225799A1 US 20200225799 A1 US20200225799 A1 US 20200225799A1 US 201916249472 A US201916249472 A US 201916249472A US 2020225799 A1 US2020225799 A1 US 2020225799A1
- Authority
- US
- United States
- Prior art keywords
- user
- gaze
- display screen
- processor
- touch
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000001514 detection method Methods 0.000 title claims abstract description 39
- 238000000034 method Methods 0.000 claims abstract description 32
- 230000000007 visual effect Effects 0.000 claims abstract description 23
- 230000015654 memory Effects 0.000 claims description 8
- 230000003993 interaction Effects 0.000 claims description 7
- 238000002604 ultrasonography Methods 0.000 claims 3
- 230000009471 action Effects 0.000 abstract description 21
- 230000001815 facial effect Effects 0.000 description 12
- 210000003813 thumb Anatomy 0.000 description 9
- 230000008859 change Effects 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 210000003811 finger Anatomy 0.000 description 2
- 238000004364 calculation method Methods 0.000 description 1
- 229910003460 diamond Inorganic materials 0.000 description 1
- 239000010432 diamond Substances 0.000 description 1
- 230000003203 everyday effect Effects 0.000 description 1
- 210000000554 iris Anatomy 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000008569 process Effects 0.000 description 1
- 210000001747 pupil Anatomy 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/041—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
- G06F3/0416—Control or interface arrangements specially adapted for digitisers
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
- G06F3/013—Eye tracking input arrangements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
Definitions
- the present disclosure relates generally to a method for controlling user interaction with a touch screen device. More particularly, it relates to a method for controlling user interaction with a touch screen device, such as a smart phone or tablet, where the device, if unlocked, only responds to the user's screen touches when the user is looking at the device screen.
- a touch screen device such as a smart phone or tablet
- the gaze-detection feature prevents unwanted device actions which are commonly encountered when a user is simply trying to handle the device.
- smart phones and tablet devices all have a security locking feature, where the device may be configured to lock automatically after a user-defined period of inactivity, and the device can be instantly locked by a user command or button push.
- touch screen access to the apps is disabled, and only a very limited subset of touch screen commands are available—such as entering a passcode to unlock the phone, and placing an emergency (“911”) phone call.
- a common problem with touch screen devices is that, when the device is unlocked, almost any touch of the screen will trigger an action. Even an accidental or inadvertent touch by the user, such as when the user is attempting to set the device down on a desk or a vehicle console, will trigger an unintended action. For example, if the user has just ended a phone call (and the smart phone is unlocked and in the phone app), an accidental screen touch might place an unwanted phone call to someone on the recent call list. As another example, if the user has started navigation using a GPS app, and accidentally touches the screen while setting the phone down, this could cause the map to zoom or pan in an undesired way, change the destination data, or end navigation. These unintended actions at a minimum are frustrating to the user, and may even be distracting and dangerous in the context of a driving situation.
- a touch screen device user is forced to either lock the device after using it, or carefully handle the device to avoid touching any part of the display screen.
- Most users find both of these workarounds to be inconvenient and annoying.
- locking the device defeats the purpose of using GPS navigation apps and many other apps, and is therefore not a viable workaround.
- Facial recognition systems are known to be available on some touch screen devices. These facial recognition systems employ a user-facing or “front-facing” camera (that is, a camera having its lens on the display screen side of the device) to recognize the user based on a previously-defined facial image map, where the facial recognition can be used as a form of security authentication in addition to or in lieu of a passcode. Some existing facial recognition systems also include gaze detection technology which recognizes whether the user's eyes are open and the user's attention (visual gaze) is directed towards the device screen.
- gaze detection technology which recognizes whether the user's eyes are open and the user's attention (visual gaze) is directed towards the device screen.
- the present disclosure describes a touch screen device and a method of controlling the device wherein the device responds to user touch screen inputs only when the device is unlocked and the user's visual gaze is directed at the device screen.
- the device employs a front-facing camera and gaze detection technology to determine whether the user is looking directly at the device display screen. If the user is not looking directly at the display screen, touch input commands are ignored by the device, even if the device is unlocked, thereby preventing unintended application actions due to incidental screen contact.
- the gaze detection interlock feature may be enabled or disabled through device configuration settings. The gaze detection interlock feature, if enabled, may respond only to the primary device user's face and visual gaze, or may respond when any person's visual gaze is directed to the device display screen.
- FIG. 1 is an illustration of a smart phone device in a user's hand, where the thumb of the hand is making unintended incidental contact with the touch screen, as known in the art;
- FIG. 2 is an illustration of a smart phone device in a user's hand, where several parts of the hand are making unintended incidental contact with the touch screen, as known in the art;
- FIG. 3 is an illustration of a smart phone device in a user's hand, where the user is carefully gripping the device around the edge of the case in order to avoid making unintended incidental contact with the touch screen, as known in the art;
- FIG. 4 is an illustration of a smart phone device in a user's hand, where the user's visual gaze is directed at the display screen, causing the device to respond to touch screen inputs, according to an embodiment of the present disclosure
- FIG. 5 is an illustration of a smart phone device in a user's hand, where the user's visual gaze is not directed at the display screen, causing the device to ignore touch screen inputs, according to an embodiment of the present disclosure
- FIG. 6 is a flowchart diagram of a method for controlling user interaction with the touch screen device of FIGS. 4 and 5 , according to an embodiment of the present disclosure.
- Smart phones, tablet devices and personal computers have become an important part of everyday life.
- One feature which has made these computing devices so powerful and easy to use is the touch screen display.
- Virtually all smart phones and tablet devices now include a touch screen display.
- Many laptop computers, desktop computer monitors, gaming systems and video display devices such as televisions also now include touch screen displays.
- the gaze detection interlock feature of the present disclosure is particularly applicable to smart phones because of the manner in which smart phones are handled by users. For this reason, the following detailed discussion of the disclosed techniques is based on smart phone usage scenarios. However, it is to be understood that the disclosed techniques are applicable to all of the device types mentioned above, and any other type of touch screen device.
- apps applications
- data belonging to the user/owner smart phones and tablet devices have a security locking feature.
- touch screen access to the apps is disabled, and only a very limited subset of touch screen commands are available—such as entering a passcode to unlock the device, and placing an emergency (“911”) phone call.
- Phones and tablets typically may be configured to lock automatically after a user-defined period of inactivity, and the devices can be instantly locked by a user command or button push.
- a common problem with smart phones is that, when the device is unlocked, almost any touch of the screen will trigger an action. Even an accidental or inadvertent touch by the user, such as when the user is attempting to set the phone down on a desk or a vehicle console, will trigger an unintended action.
- FIG. 1 is an illustration of a smart phone 100 in a user's hand 110 , where the thumb of the hand 110 is making unintended incidental contact with a touch screen 102 .
- the user is holding the smart phone 100 in a manner commonly used after entering information with the thumb (such as typing a text message, placing or ending a phone call, or operating any other app).
- the thumb can inadvertently contact the touch screen 102 at a location 112 .
- the thumb touch at the location 112 will cause an unintended application action if the smart phone 100 is unlocked.
- FIG. 2 is an illustration of the smart phone 100 in the user's hand 110 , where the thumb and fingers of the hand 110 are making unintended incidental contact with the touch screen 102 .
- the user is holding the smart phone 100 in a manner commonly used for securely gripping any item in the hand 110 .
- the fingers can inadvertently contact the touch screen 102 at a location 202 or 204 , and/or the thumb can inadvertently contact the touch screen 102 at a location 206 .
- a screen touch at any of the locations 202 / 204 / 206 will cause an unintended application action if the smart phone 100 is unlocked.
- Unintended application actions of the type illustrated in FIGS. 1 and 2 can include launching and executing commands and operations in any application installed on the smart phone 100 .
- an accidental screen touch might place an unwanted phone call to someone on the recent call list.
- the user has started navigation using a GPS app, and accidentally touches the screen 102 while setting the phone 100 down in the vehicle console, this could cause the map to zoom or pan in an undesired way, change the navigation destination data, or end navigation.
- the possibilities for unintended application actions are endless—including other examples such as accidentally starting music or video streaming, accidentally executing a transaction in a mobile banking app, accidentally making a purchase from a retailer app or website, etc.
- a smart phone user is forced to either lock the smart phone 100 after using it, or carefully handle the phone 100 to avoid touching any part of the touch screen 102 .
- FIG. 3 is an illustration of the smart phone 100 in the user's hand 110 , where the user is gripping the phone 100 around the edge of the case in order to avoid making incidental contact with the touch screen 102 .
- Carefully gripping the smart phone 100 in the manner shown in FIG. 3 is one way to avoid the unintended application actions described above.
- the smart phone 100 can also be held with the hand 110 open and flat, palm facing up, in order to avoid incidental screen contact. Most users find these types of careful gripping techniques to be inconvenient and annoying.
- grasping the smart phone 100 without touching the screen 102 results in an inherently insecure grip, causing phones to be dropped and often damaged. This problem has given rise to the so-called “pop-socket” device which can be attached to the back of the phone 100 .
- locking the smart phone 100 is another way to avoid unintended application actions.
- locking the phone 100 defeats the purpose of using GPS navigation apps and many other apps, and is therefore not a viable workaround in many situations.
- none of the workarounds described above actually resolve the real problem, which is that the smart phone 100 actively responds to user touches of the screen 102 when the user does not want it to do so.
- the present disclosure describes a technique for controlling a smart phone or other touch screen device wherein the device only responds to user touch inputs when the user is looking directly at the touch screen.
- the disclosed techniques are based on the premise that, if the user wants the touch screen device to respond to touch commands, the user will be looking at the touch screen, and conversely, if the user is not looking at the screen, then the user does not want the touch screen device to respond to touch commands.
- Facial recognition systems are known to be available on some touch screen devices. These facial recognition systems employ a user-facing or “front-facing” camera (that is, a camera having its lens on the display screen side of the device) to recognize the user based on a previously-defined facial image map.
- the facial recognition can be used as a form of security authentication in addition to, or in lieu of, a passcode.
- Some existing facial recognition systems also include gaze detection technology which recognizes whether the user's eyes are open and the user's attention (visual gaze) is directed towards the device screen.
- a typical gaze detection subsystem includes a front-facing camera and possibly other sensors providing images and data to the device's processor which runs an algorithm to determine whether the user is looking directly at the screen. This gaze detection technology can be employed to control the touch screen device operation.
- the disclosed technique which might be called a gaze detection interlock feature, causes the touch screen device to respond to user inputs only when the user's visual gaze is directed toward the screen.
- FIG. 4 is an illustration of a smart phone 400 being held by a user 410 , where the user's visual gaze is directed at a display screen 402 , causing the smart phone 400 to respond to touch screen inputs, according to an embodiment of the present disclosure.
- the smart phone 400 employs a front-facing camera 404 and/or other sensors 406 to provide images and/or geometric shape data describing the user's face.
- the other sensors 406 may include any type of sensor suitable for providing images or shape data of the user's face—including but not limited to another visual-light camera, an infrared camera, an ultrasonic sensor, a LiDAR sensor, etc.
- the images and/or shape data are processed using known gaze detection technology (for example, the locations of the irises and pupils within the eye outline) to determine whether the user 410 is looking directly at the screen 402 .
- the user 410 is looking directly at the screen 402 , as indicated by dashed lines 414 .
- Existing gaze detection technology can easily and reliably determine that the user's gaze is directed to the screen 402 and not, for example, at a distant scene 420 . Because the user's gaze is directed to the touch screen 402 , the smart phone 400 responds normally to any touch of the screen 402 .
- the smart phone 400 is shown in a hand 412 of the user 410 , with the thumb of the hand 412 being used in typical fashion to input touch commands on the screen 402 .
- FIG. 5 is an illustration of the smart phone 400 being held by a user 510 , where the user's visual gaze is not directed at the display screen 402 , causing the smart phone 400 to ignore touch screen inputs, according to an embodiment of the present disclosure.
- the user 510 is looking at a distant scene 520 as indicated by dashed lines 514 .
- the gaze detection subsystem on the smart phone 400 can readily determine that the user's gaze is directed to the distant scene 520 and not at the touch screen 402 .
- the gaze detection interlock feature is activated, meaning that the smart phone 400 ignores any touch of the screen 402 , such as by the thumb of the hand 512 . As illustrated in FIG. 5 , the gaze detection interlock feature causes touch input commands to be ignored by the smart phone 400 , even if the 400 is unlocked, thereby preventing unintended application actions due to incidental screen contact.
- the gaze detection interlock feature may be enabled or disabled through device configuration settings. That is, in a settings menu on the phone 400 , the user may enable the gaze detection interlock feature (causing the phone 400 to behave as described in the discussion of FIGS. 5 and 6 ), or the user may disable the gaze detection interlock feature (causing the phone 400 to always respond to user touch inputs when the phone 400 is unlocked, regardless of whether the user is looking at the screen). Enabling and disabling the gaze detection interlock feature is also preferably possible via voice command, to make it easy for the device owner to turn the feature on or off in a hands-free manner.
- the gaze detection interlock feature may respond only to the primary device user's face and visual gaze, or may respond when any person's visual gaze is directed to the device display screen. Responding only to the primary device user's face and visual gaze (as determined by the facial recognition security feature discussed earlier) may be advantageous, for example, if the user wants to ensure that he or she is the only person who can control the device for security reasons. Responding when any person's visual gaze is directed to the device display screen may be advantageous, for example, when a vehicle driver wants to hand the phone to a passenger to enter a navigation destination. These options can also be defined via device configuration settings, either temporary or permanent.
- FIG. 6 is a flowchart diagram 600 of a method for controlling user interaction with a touch screen device such as the smart phone 400 of FIGS. 4 and 5 , according to an embodiment of the present disclosure.
- a touch screen device such as the smart phone 400 in an unlocked state is provided.
- the touch screen device determines the direction of the user's visual gaze. This is done using the gaze detection subsystem discussed earlier, including at least one camera communicating with the device processor.
- decision diamond 606 it is determined whether the user's visual gaze is directed at the device screen. If the user's visual gaze is directed at the device screen, then at box 608 the touch screen device responds to user screen touches in a normal fashion. If the user's visual gaze is not directed at the device screen, then at box 610 the touch screen device ignores user screen touches, even if the device is unlocked.
- the determination of the direction of the user's gaze must be repeated periodically on an ongoing basis by the touch screen device, because the user may be looking at the device screen in one moment, and not looking at the device screen in the next moment, or vice versa.
- the process loops back to the box 604 to once again determine the direction of the user's gaze.
- the re-determination of the direction of the user's gaze may be performed at regular time intervals, such as every one-tenth of a second or every one second, or based on other criteria or operating system considerations.
- the gaze detection interlock feature provides advantages not found in prior art systems. For example, existing facial recognition systems are used only to unlock a device, and do not consider whether the user's gaze is directed to the screen after the device is unlocked. Also, existing eye tracking applications may be used to control a device in lieu of actual touch commands, but do not provide the interlock feature of the present disclosure.
- the several and various features and process steps discussed herein to describe the disclosed methods may be referring to operations performed by a computer, a processor or other electronic calculating device that manipulates and/or transforms data using electrical phenomenon.
- this refers to the processor or processors in the smart phone 400 , said processors performing calculations as part of the gaze detection subsystem, running a device operating system and executing applications, and also controlling touch screen operation via the disclosed gaze detection interlock feature.
- processors and electronic devices may employ various volatile and/or non-volatile memories including non-transitory computer-readable medium with an executable program stored thereon including various code or executable instructions able to be performed by the computer or processor, where the memory and/or computer-readable medium may include all forms and types of memory and other computer-readable media.
- the disclosed gaze detection interlock feature for controlling user interaction with a touch screen device provides a way to prevent unintended application actions, thereby reducing user frustration and increasing user satisfaction with electronic devices.
- the disclosed techniques are implementable without adding any new hardware or cost to modern multi-camera touch screen devices.
Abstract
Description
- The present disclosure relates generally to a method for controlling user interaction with a touch screen device. More particularly, it relates to a method for controlling user interaction with a touch screen device, such as a smart phone or tablet, where the device, if unlocked, only responds to the user's screen touches when the user is looking at the device screen. The gaze-detection feature prevents unwanted device actions which are commonly encountered when a user is simply trying to handle the device.
- The capabilities of smart phones and tablet devices have increased dramatically since they were first introduced. Modern designs for almost all such devices now employ a touch screen user interface, where virtually the entire screen surface is used both for graphical display and for receiving user commands via touch. The touch screen interface offers application (“app”) designers almost unlimited flexibility in designing apps—from simple apps such as phone dialing and text messaging to complex and graphically-intensive apps such as gaming and virtual reality.
- Because they provide access to sensitive personal information and data belonging to the user/owner, smart phones and tablet devices all have a security locking feature, where the device may be configured to lock automatically after a user-defined period of inactivity, and the device can be instantly locked by a user command or button push. When a device is locked, touch screen access to the apps is disabled, and only a very limited subset of touch screen commands are available—such as entering a passcode to unlock the phone, and placing an emergency (“911”) phone call.
- A common problem with touch screen devices is that, when the device is unlocked, almost any touch of the screen will trigger an action. Even an accidental or inadvertent touch by the user, such as when the user is attempting to set the device down on a desk or a vehicle console, will trigger an unintended action. For example, if the user has just ended a phone call (and the smart phone is unlocked and in the phone app), an accidental screen touch might place an unwanted phone call to someone on the recent call list. As another example, if the user has started navigation using a GPS app, and accidentally touches the screen while setting the phone down, this could cause the map to zoom or pan in an undesired way, change the destination data, or end navigation. These unintended actions at a minimum are frustrating to the user, and may even be distracting and dangerous in the context of a driving situation.
- In order to avoid the unintended actions described above, a touch screen device user is forced to either lock the device after using it, or carefully handle the device to avoid touching any part of the display screen. Most users find both of these workarounds to be inconvenient and annoying. Furthermore, locking the device defeats the purpose of using GPS navigation apps and many other apps, and is therefore not a viable workaround.
- Facial recognition systems are known to be available on some touch screen devices. These facial recognition systems employ a user-facing or “front-facing” camera (that is, a camera having its lens on the display screen side of the device) to recognize the user based on a previously-defined facial image map, where the facial recognition can be used as a form of security authentication in addition to or in lieu of a passcode. Some existing facial recognition systems also include gaze detection technology which recognizes whether the user's eyes are open and the user's attention (visual gaze) is directed towards the device screen.
- In view of the circumstances described above, there is an opportunity to incorporate visual gaze detection into touch screen device operation in order to prevent unintended device actions.
- The present disclosure describes a touch screen device and a method of controlling the device wherein the device responds to user touch screen inputs only when the device is unlocked and the user's visual gaze is directed at the device screen. The device employs a front-facing camera and gaze detection technology to determine whether the user is looking directly at the device display screen. If the user is not looking directly at the display screen, touch input commands are ignored by the device, even if the device is unlocked, thereby preventing unintended application actions due to incidental screen contact. The gaze detection interlock feature may be enabled or disabled through device configuration settings. The gaze detection interlock feature, if enabled, may respond only to the primary device user's face and visual gaze, or may respond when any person's visual gaze is directed to the device display screen.
- Additional features of the present disclosure will become apparent from the following description and appended claims, taken in conjunction with the accompanying drawings.
-
FIG. 1 is an illustration of a smart phone device in a user's hand, where the thumb of the hand is making unintended incidental contact with the touch screen, as known in the art; -
FIG. 2 is an illustration of a smart phone device in a user's hand, where several parts of the hand are making unintended incidental contact with the touch screen, as known in the art; -
FIG. 3 is an illustration of a smart phone device in a user's hand, where the user is carefully gripping the device around the edge of the case in order to avoid making unintended incidental contact with the touch screen, as known in the art; -
FIG. 4 is an illustration of a smart phone device in a user's hand, where the user's visual gaze is directed at the display screen, causing the device to respond to touch screen inputs, according to an embodiment of the present disclosure; -
FIG. 5 is an illustration of a smart phone device in a user's hand, where the user's visual gaze is not directed at the display screen, causing the device to ignore touch screen inputs, according to an embodiment of the present disclosure; and -
FIG. 6 is a flowchart diagram of a method for controlling user interaction with the touch screen device ofFIGS. 4 and 5 , according to an embodiment of the present disclosure. - The following discussion of the embodiments of the disclosure directed to a touch screen device and a method of controlling the device including a gaze detection interlock feature is merely exemplary in nature, and is in no way intended to limit the disclosure or its applications or uses.
- Smart phones, tablet devices and personal computers have become an important part of everyday life. One feature which has made these computing devices so powerful and easy to use is the touch screen display. Virtually all smart phones and tablet devices now include a touch screen display. Many laptop computers, desktop computer monitors, gaming systems and video display devices such as televisions also now include touch screen displays. The gaze detection interlock feature of the present disclosure is particularly applicable to smart phones because of the manner in which smart phones are handled by users. For this reason, the following detailed discussion of the disclosed techniques is based on smart phone usage scenarios. However, it is to be understood that the disclosed techniques are applicable to all of the device types mentioned above, and any other type of touch screen device.
- Because they provide access to sensitive personal information, applications (“apps”) and data belonging to the user/owner, smart phones and tablet devices have a security locking feature. When a device is locked, touch screen access to the apps is disabled, and only a very limited subset of touch screen commands are available—such as entering a passcode to unlock the device, and placing an emergency (“911”) phone call. Phones and tablets typically may be configured to lock automatically after a user-defined period of inactivity, and the devices can be instantly locked by a user command or button push.
- A common problem with smart phones is that, when the device is unlocked, almost any touch of the screen will trigger an action. Even an accidental or inadvertent touch by the user, such as when the user is attempting to set the phone down on a desk or a vehicle console, will trigger an unintended action.
-
FIG. 1 is an illustration of asmart phone 100 in a user'shand 110, where the thumb of thehand 110 is making unintended incidental contact with atouch screen 102. InFIG. 1 , the user is holding thesmart phone 100 in a manner commonly used after entering information with the thumb (such as typing a text message, placing or ending a phone call, or operating any other app). Unfortunately for users, when holding thesmart phone 100 in this manner, and then attempting to set thephone 100 down in a vehicle console or on a desktop or other surface, the thumb can inadvertently contact thetouch screen 102 at alocation 112. The thumb touch at thelocation 112 will cause an unintended application action if thesmart phone 100 is unlocked. -
FIG. 2 is an illustration of thesmart phone 100 in the user'shand 110, where the thumb and fingers of thehand 110 are making unintended incidental contact with thetouch screen 102. InFIG. 2 , the user is holding thesmart phone 100 in a manner commonly used for securely gripping any item in thehand 110. Unfortunately, when holding thesmart phone 100 in this manner, whether simply carrying thesmart phone 100 or attempting to place thephone 100 on a surface or in a pocket or purse, the fingers can inadvertently contact thetouch screen 102 at alocation touch screen 102 at alocation 206. A screen touch at any of thelocations 202/204/206 will cause an unintended application action if thesmart phone 100 is unlocked. - Unintended application actions of the type illustrated in
FIGS. 1 and 2 can include launching and executing commands and operations in any application installed on thesmart phone 100. - For example, if the user has just ended a phone call (and the
smart phone 100 is unlocked and displaying the phone app), an accidental screen touch might place an unwanted phone call to someone on the recent call list. As another example, if the user has started navigation using a GPS app, and accidentally touches thescreen 102 while setting thephone 100 down in the vehicle console, this could cause the map to zoom or pan in an undesired way, change the navigation destination data, or end navigation. The possibilities for unintended application actions are endless—including other examples such as accidentally starting music or video streaming, accidentally executing a transaction in a mobile banking app, accidentally making a purchase from a retailer app or website, etc. - These unintended actions, at a minimum, are frustrating to the user. Furthermore, these unintended actions may create costs or consequences for the user (accidental purchase of an item, for example), and may even be distracting and dangerous in the context of a driving situation.
- In order to avoid the unintended application actions described above, a smart phone user is forced to either lock the
smart phone 100 after using it, or carefully handle thephone 100 to avoid touching any part of thetouch screen 102. -
FIG. 3 is an illustration of thesmart phone 100 in the user'shand 110, where the user is gripping thephone 100 around the edge of the case in order to avoid making incidental contact with thetouch screen 102. - Carefully gripping the
smart phone 100 in the manner shown inFIG. 3 is one way to avoid the unintended application actions described above. Thesmart phone 100 can also be held with thehand 110 open and flat, palm facing up, in order to avoid incidental screen contact. Most users find these types of careful gripping techniques to be inconvenient and annoying. Furthermore, grasping thesmart phone 100 without touching thescreen 102 results in an inherently insecure grip, causing phones to be dropped and often damaged. This problem has given rise to the so-called “pop-socket” device which can be attached to the back of thephone 100. - As mentioned above, locking the
smart phone 100 is another way to avoid unintended application actions. However, locking thephone 100 defeats the purpose of using GPS navigation apps and many other apps, and is therefore not a viable workaround in many situations. In fact, none of the workarounds described above actually resolve the real problem, which is that thesmart phone 100 actively responds to user touches of thescreen 102 when the user does not want it to do so. - According to the following discussion, the present disclosure describes a technique for controlling a smart phone or other touch screen device wherein the device only responds to user touch inputs when the user is looking directly at the touch screen. The disclosed techniques are based on the premise that, if the user wants the touch screen device to respond to touch commands, the user will be looking at the touch screen, and conversely, if the user is not looking at the screen, then the user does not want the touch screen device to respond to touch commands.
- Facial recognition systems are known to be available on some touch screen devices. These facial recognition systems employ a user-facing or “front-facing” camera (that is, a camera having its lens on the display screen side of the device) to recognize the user based on a previously-defined facial image map. The facial recognition can be used as a form of security authentication in addition to, or in lieu of, a passcode.
- Some existing facial recognition systems also include gaze detection technology which recognizes whether the user's eyes are open and the user's attention (visual gaze) is directed towards the device screen. A typical gaze detection subsystem includes a front-facing camera and possibly other sensors providing images and data to the device's processor which runs an algorithm to determine whether the user is looking directly at the screen. This gaze detection technology can be employed to control the touch screen device operation. The disclosed technique, which might be called a gaze detection interlock feature, causes the touch screen device to respond to user inputs only when the user's visual gaze is directed toward the screen.
-
FIG. 4 is an illustration of asmart phone 400 being held by auser 410, where the user's visual gaze is directed at adisplay screen 402, causing thesmart phone 400 to respond to touch screen inputs, according to an embodiment of the present disclosure. Thesmart phone 400 employs a front-facingcamera 404 and/orother sensors 406 to provide images and/or geometric shape data describing the user's face. Theother sensors 406 may include any type of sensor suitable for providing images or shape data of the user's face—including but not limited to another visual-light camera, an infrared camera, an ultrasonic sensor, a LiDAR sensor, etc. The images and/or shape data are processed using known gaze detection technology (for example, the locations of the irises and pupils within the eye outline) to determine whether theuser 410 is looking directly at thescreen 402. - In
FIG. 4 , theuser 410 is looking directly at thescreen 402, as indicated by dashedlines 414. Existing gaze detection technology can easily and reliably determine that the user's gaze is directed to thescreen 402 and not, for example, at adistant scene 420. Because the user's gaze is directed to thetouch screen 402, thesmart phone 400 responds normally to any touch of thescreen 402. Thesmart phone 400 is shown in ahand 412 of theuser 410, with the thumb of thehand 412 being used in typical fashion to input touch commands on thescreen 402. -
FIG. 5 is an illustration of thesmart phone 400 being held by auser 510, where the user's visual gaze is not directed at thedisplay screen 402, causing thesmart phone 400 to ignore touch screen inputs, according to an embodiment of the present disclosure. InFIG. 5 , although holding thephone 400 in ahand 512, theuser 510 is looking at adistant scene 520 as indicated by dashedlines 514. The gaze detection subsystem on thesmart phone 400 can readily determine that the user's gaze is directed to thedistant scene 520 and not at thetouch screen 402. Because the user's gaze is not directed to thetouch screen 402, the gaze detection interlock feature is activated, meaning that thesmart phone 400 ignores any touch of thescreen 402, such as by the thumb of thehand 512. As illustrated inFIG. 5 , the gaze detection interlock feature causes touch input commands to be ignored by thesmart phone 400, even if the 400 is unlocked, thereby preventing unintended application actions due to incidental screen contact. - The gaze detection interlock feature may be enabled or disabled through device configuration settings. That is, in a settings menu on the
phone 400, the user may enable the gaze detection interlock feature (causing thephone 400 to behave as described in the discussion ofFIGS. 5 and 6 ), or the user may disable the gaze detection interlock feature (causing thephone 400 to always respond to user touch inputs when thephone 400 is unlocked, regardless of whether the user is looking at the screen). Enabling and disabling the gaze detection interlock feature is also preferably possible via voice command, to make it easy for the device owner to turn the feature on or off in a hands-free manner. - The gaze detection interlock feature, if enabled, may respond only to the primary device user's face and visual gaze, or may respond when any person's visual gaze is directed to the device display screen. Responding only to the primary device user's face and visual gaze (as determined by the facial recognition security feature discussed earlier) may be advantageous, for example, if the user wants to ensure that he or she is the only person who can control the device for security reasons. Responding when any person's visual gaze is directed to the device display screen may be advantageous, for example, when a vehicle driver wants to hand the phone to a passenger to enter a navigation destination. These options can also be defined via device configuration settings, either temporary or permanent.
-
FIG. 6 is a flowchart diagram 600 of a method for controlling user interaction with a touch screen device such as thesmart phone 400 ofFIGS. 4 and 5 , according to an embodiment of the present disclosure. Atbox 602, a touch screen device (such as the smart phone 400) in an unlocked state is provided. Atbox 604, the touch screen device determines the direction of the user's visual gaze. This is done using the gaze detection subsystem discussed earlier, including at least one camera communicating with the device processor. Atdecision diamond 606, it is determined whether the user's visual gaze is directed at the device screen. If the user's visual gaze is directed at the device screen, then atbox 608 the touch screen device responds to user screen touches in a normal fashion. If the user's visual gaze is not directed at the device screen, then atbox 610 the touch screen device ignores user screen touches, even if the device is unlocked. - The determination of the direction of the user's gaze must be repeated periodically on an ongoing basis by the touch screen device, because the user may be looking at the device screen in one moment, and not looking at the device screen in the next moment, or vice versa. Thus, from both the
box 608 and thebox 610, the process loops back to thebox 604 to once again determine the direction of the user's gaze. The re-determination of the direction of the user's gaze may be performed at regular time intervals, such as every one-tenth of a second or every one second, or based on other criteria or operating system considerations. - The gaze detection interlock feature provides advantages not found in prior art systems. For example, existing facial recognition systems are used only to unlock a device, and do not consider whether the user's gaze is directed to the screen after the device is unlocked. Also, existing eye tracking applications may be used to control a device in lieu of actual touch commands, but do not provide the interlock feature of the present disclosure.
- As will be well understood by those skilled in the art, the several and various features and process steps discussed herein to describe the disclosed methods may be referring to operations performed by a computer, a processor or other electronic calculating device that manipulates and/or transforms data using electrical phenomenon. In particular, this refers to the processor or processors in the
smart phone 400, said processors performing calculations as part of the gaze detection subsystem, running a device operating system and executing applications, and also controlling touch screen operation via the disclosed gaze detection interlock feature. Those processors and electronic devices may employ various volatile and/or non-volatile memories including non-transitory computer-readable medium with an executable program stored thereon including various code or executable instructions able to be performed by the computer or processor, where the memory and/or computer-readable medium may include all forms and types of memory and other computer-readable media. - The disclosed gaze detection interlock feature for controlling user interaction with a touch screen device provides a way to prevent unintended application actions, thereby reducing user frustration and increasing user satisfaction with electronic devices. The disclosed techniques are implementable without adding any new hardware or cost to modern multi-camera touch screen devices.
- The foregoing discussion discloses and describes merely exemplary embodiments of the present disclosure. One skilled in the art will readily recognize from such discussion and from the accompanying drawings and claims that various changes, modifications and variations can be made therein without departing from the spirit and scope of the disclosure as defined in the following claims.
Claims (20)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US16/249,472 US10761648B2 (en) | 2019-01-16 | 2019-01-16 | Gaze detection interlock feature for touch screen devices |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US16/249,472 US10761648B2 (en) | 2019-01-16 | 2019-01-16 | Gaze detection interlock feature for touch screen devices |
Publications (2)
Publication Number | Publication Date |
---|---|
US20200225799A1 true US20200225799A1 (en) | 2020-07-16 |
US10761648B2 US10761648B2 (en) | 2020-09-01 |
Family
ID=71517618
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/249,472 Active US10761648B2 (en) | 2019-01-16 | 2019-01-16 | Gaze detection interlock feature for touch screen devices |
Country Status (1)
Country | Link |
---|---|
US (1) | US10761648B2 (en) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11303649B2 (en) * | 2019-05-30 | 2022-04-12 | International Business Machines Corporation | Maintaining electronic communications session continuity during session inactivity |
JP2023020879A (en) * | 2021-07-30 | 2023-02-09 | 功憲 末次 | Inappropriate use control system and inappropriate use control program |
JP2023060890A (en) * | 2021-07-30 | 2023-04-28 | 功憲 末次 | Inappropriate use control system and inappropriate use control program |
JP7331281B2 (en) | 2021-07-30 | 2023-08-22 | 功憲 末次 | Improper Use Control System and Improper Use Control Program |
JP7331282B2 (en) | 2021-07-30 | 2023-08-22 | 功憲 末次 | Improper Use Control System and Improper Use Control Program |
US11886317B2 (en) | 2021-07-30 | 2024-01-30 | Katsunori SUETSUGU | Inappropriate use control system and inappropriate use control program |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20060022959A1 (en) * | 2001-07-09 | 2006-02-02 | Geaghan Bernard O | Touch screen with selective touch sources |
US8913004B1 (en) * | 2010-03-05 | 2014-12-16 | Amazon Technologies, Inc. | Action based device control |
US20150035776A1 (en) * | 2012-03-23 | 2015-02-05 | Ntt Docomo, Inc. | Information terminal, method for controlling input acceptance, and program for controlling input acceptance |
US20150338914A1 (en) * | 2013-11-01 | 2015-11-26 | Intel Corporation | Gaze-assisted touchscreen inputs |
US20160275314A1 (en) * | 2014-04-28 | 2016-09-22 | Sony Corporation | Operating a display of a user equipment |
US20180088665A1 (en) * | 2016-09-26 | 2018-03-29 | Lenovo (Singapore) Pte. Ltd. | Eye tracking selection validation |
US9946371B2 (en) * | 2014-10-16 | 2018-04-17 | Qualcomm Incorporated | System and method for using touch orientation to distinguish between users of a touch panel |
US20190187786A1 (en) * | 2017-12-15 | 2019-06-20 | International Business Machines Corporation | Preventing unintended input |
-
2019
- 2019-01-16 US US16/249,472 patent/US10761648B2/en active Active
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20060022959A1 (en) * | 2001-07-09 | 2006-02-02 | Geaghan Bernard O | Touch screen with selective touch sources |
US8913004B1 (en) * | 2010-03-05 | 2014-12-16 | Amazon Technologies, Inc. | Action based device control |
US20150035776A1 (en) * | 2012-03-23 | 2015-02-05 | Ntt Docomo, Inc. | Information terminal, method for controlling input acceptance, and program for controlling input acceptance |
US20150338914A1 (en) * | 2013-11-01 | 2015-11-26 | Intel Corporation | Gaze-assisted touchscreen inputs |
US20160275314A1 (en) * | 2014-04-28 | 2016-09-22 | Sony Corporation | Operating a display of a user equipment |
US9946371B2 (en) * | 2014-10-16 | 2018-04-17 | Qualcomm Incorporated | System and method for using touch orientation to distinguish between users of a touch panel |
US20180088665A1 (en) * | 2016-09-26 | 2018-03-29 | Lenovo (Singapore) Pte. Ltd. | Eye tracking selection validation |
US20190187786A1 (en) * | 2017-12-15 | 2019-06-20 | International Business Machines Corporation | Preventing unintended input |
Cited By (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11303649B2 (en) * | 2019-05-30 | 2022-04-12 | International Business Machines Corporation | Maintaining electronic communications session continuity during session inactivity |
JP2023020879A (en) * | 2021-07-30 | 2023-02-09 | 功憲 末次 | Inappropriate use control system and inappropriate use control program |
JP2023060890A (en) * | 2021-07-30 | 2023-04-28 | 功憲 末次 | Inappropriate use control system and inappropriate use control program |
JP7282235B2 (en) | 2021-07-30 | 2023-05-26 | 功憲 末次 | Improper Use Control System and Improper Use Control Program |
JP7331281B2 (en) | 2021-07-30 | 2023-08-22 | 功憲 末次 | Improper Use Control System and Improper Use Control Program |
JP7331282B2 (en) | 2021-07-30 | 2023-08-22 | 功憲 末次 | Improper Use Control System and Improper Use Control Program |
JP7331283B2 (en) | 2021-07-30 | 2023-08-22 | 功憲 末次 | Improper Use Control System and Improper Use Control Program |
US11886317B2 (en) | 2021-07-30 | 2024-01-30 | Katsunori SUETSUGU | Inappropriate use control system and inappropriate use control program |
Also Published As
Publication number | Publication date |
---|---|
US10761648B2 (en) | 2020-09-01 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10761648B2 (en) | Gaze detection interlock feature for touch screen devices | |
US20220100841A1 (en) | Authenticated device used to unlock another device | |
CN109753159B (en) | Method and apparatus for controlling electronic device | |
US9733752B2 (en) | Mobile terminal and control method thereof | |
US10956734B2 (en) | Electronic device providing iris recognition based on proximity and operating method thereof | |
US8717393B2 (en) | System and method for controlling a display of a mobile device | |
US20230014800A1 (en) | User interface for accessing an account | |
US20130342672A1 (en) | Using gaze determination with device input | |
US8938612B1 (en) | Limited-access state for inadvertent inputs | |
AU2017291584B2 (en) | Method for recognizing iris based on user intention and electronic device for the same | |
EP3259701B1 (en) | Biometric setup that runs in the background | |
US20170123587A1 (en) | Method and device for preventing accidental touch of terminal with touch screen | |
US9536073B2 (en) | Device-based application security | |
US20180088665A1 (en) | Eye tracking selection validation | |
KR102320072B1 (en) | Electronic device and method for controlling of information disclosure thereof | |
WO2023030238A1 (en) | Secure input method and apparatus | |
WO2023125238A1 (en) | Unlocking control method and apparatus, electronic device, and readable storage medium | |
US20220221932A1 (en) | Controlling a function via gaze detection | |
WO2021004352A1 (en) | Interface interaction method and terminal | |
JP6132245B2 (en) | Providing calls based on the detected direction | |
US9948767B2 (en) | Rejecting or accepting a phone call | |
KR102660224B1 (en) | Electronic device and method of providing notification information thereof | |
WO2022037247A1 (en) | Device, method and system for operating device | |
CN107632754B (en) | Key control method based on fingerprint identification module, mobile terminal and storage medium | |
CN112162810A (en) | Message display method and device and electronic equipment |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: MICROENTITY |
|
FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO MICRO (ORIGINAL EVENT CODE: MICR); ENTITY STATUS OF PATENT OWNER: MICROENTITY |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, MICRO ENTITY (ORIGINAL EVENT CODE: M3551); ENTITY STATUS OF PATENT OWNER: MICROENTITY Year of fee payment: 4 |