US20190324531A1 - System and method for using gaze control to control electronic switches and machinery - Google Patents
System and method for using gaze control to control electronic switches and machinery Download PDFInfo
- Publication number
- US20190324531A1 US20190324531A1 US16/029,808 US201816029808A US2019324531A1 US 20190324531 A1 US20190324531 A1 US 20190324531A1 US 201816029808 A US201816029808 A US 201816029808A US 2019324531 A1 US2019324531 A1 US 2019324531A1
- Authority
- US
- United States
- Prior art keywords
- gaze
- user
- interaction
- action
- control
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
- G06F3/013—Eye tracking input arrangements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F21/00—Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
- G06F21/30—Authentication, i.e. establishing the identity or authorisation of security principals
- G06F21/31—User authentication
- G06F21/32—User authentication using biometric data, e.g. fingerprints, iris scans or voiceprints
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/16—Sound input; Sound output
- G06F3/167—Audio in a user interface, e.g. using voice commands for navigating, audio feedback
Definitions
- the present invention is related generally to gaze recognition and specifically using gaze recognition to control switches and machinery.
- FIG. 1 is a block diagram of a first illustrative system for using gaze to control an electrical and/or mechanical element
- FIG. 2 is a diagram that shows a process for using gaze to control an electrical and/or mechanical element using gaze quadrants
- FIG. 3 is a flow diagram of a process for using gaze to control an electrical and/or mechanical element
- FIG. 4 is a flow diagram of a process for using a gaze sequence
- FIG. 5 is a flow diagram of a process for identifying a gaze quadrant.
- FIG. 1 is a block diagram of a first illustrative system 100 for using gaze to control electrical and/or mechanical element(s) 104 .
- the first illustrative system 100 comprises camera(s) 101 , a microprocessor 102 , a memory 103 , and the electrical and/or mechanical element(s) 104 .
- the camera(s) 101 can be or may include any type of camera 101 that can capture an image, such as a two dimensional camera, an acoustic camera (a three dimensional camera), an inferred camera, a digital camera, a video camera, and/or the like.
- the camera(s) 101 may send a video stream, a picture, and/or the like.
- the camera(s) 101 may also capture an audio stream associated with the video stream or picture (e.g., from a microphone (not shown)).
- the microprocessor 102 can be or may or can include any known microprocessor 102 , such as a microcontroller, an application specific microprocessor 102 , a digital signaling processor, a multi-core processor, and/or the like.
- the microprocessor 102 interfaces with the camera(s) 101 , the memory 103 , and the electrical and/or mechanical element(s) 104 .
- the microprocessor 102 may interface with the camera(s) 101 , the memory 103 , and the electrical and/or mechanical element(s) 104 either directly (e.g., in the same device and/or on the same circuit board), via a computer network, via wiring, and/or the like.
- the camera(s) 101 and/or the electrical and/or mechanical element(s) 104 may be connected to the microprocessor 102 /memory 103 via a computer network.
- the microprocessor 102 may comprise multiple processors.
- the microprocessor 102 may comprise a microprocessor 102 in the camera 101 or a touch screen device along with a central microprocessor 102 .
- the camera(s) 101 , the microprocessor 102 , the memory 103 , and the electrical and/or mechanical element(s) 104 may be in the same device.
- the camera(s) 101 may be external from the microprocessor 102 /memory 103 /electrical and/or mechanical element(s) 104 .
- the camera 101 may be an external camera 101 attached to a computer monitor, an array of cameras 101 embedded in a monitor, a camera 101 at a door, an array of cameras 101 in a room, and/or the like.
- the memory 103 can be or may include, a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing.
- RAM random access memory
- ROM read-only memory
- EPROM or Flash memory erasable programmable read-only memory
- CD-ROM compact disc read-only memory
- optical storage device a magnetic storage device, or any suitable combination of the foregoing.
- the memory 103 further comprises a gaze application 105 .
- the gaze application 105 can detect and manage gaze information received from the camera(s) 101 .
- the gaze application 105 can initiate different actions based on detecting and managing gaze.
- the gaze application 105 can track the gaze or eye of a human and/or an animal.
- the detection of the gaze interaction may be used in situations where a user's hands are occupied, where the user 110 does not have hands, where a device is out of arms reach (e.g., a light in a ceiling), where there is no space on the device that allows integration of touch interaction, where speech interaction is not possible, and/or the like.
- the gaze application 105 can control a gaze interaction that may be a primary or a supplementary gaze interaction, which can be used in combination with other types of interactions, such as speech interactions, touch interactions (e.g., a user 110 touching a touch screen or button), gesture interactions, and/or the like.
- a speech interaction may follow a gaze interaction; in this case, the gaze of a user 110 looking at a device tells the device to turn on a microphone and then wait/listen for a speech interaction.
- the gaze application 105 can turn off the microphone when no gaze detected, or in other words, when the user 110 is not looking at the device.
- the gaze application 105 can be used not only to activate a switch or a device, but also can be used to deactivate the switch or device by measuring either the point of gaze (where one is looking) and/or the motion of an eye relative to the head.
- One variant uses video images from which the eye position is extracted.
- Other embodiments may use search coils, eye-attached tracking (e.g., a special contact lens with an embedded mirror or magnetic field sensor), electric potential measurement (e.g., using electrodes positioned around the eyes of the user 110 to generate an electrooculogram signal), and/or the like.
- the gaze application 105 can use a video-based eye-tracker.
- the camera 101 focuses on one or both eyes and records eye movement as the user 110 looks a particular object.
- the gaze application 105 can use eye-trackers to track the center of the pupil and infrared/near-infrared non-collimated light to create corneal reflections (CR).
- CR corneal reflections
- a vector between the pupil center and the corneal reflections may be used to compute the point of regard on the surface of the eye or the gaze direction of the eye.
- a simple calibration procedure of the user 110 may be needed before using the gaze application 103 .
- the gaze application 105 may support various types of infrared/near-infrared (also known as active light) eye-tracking techniques.
- the gaze application 105 may support bright-pupil and dark-pupil techniques. Differences between the two may be based on the location of an illumination source with respect to the camera 101 . If the illumination is coaxial with the optical path, then the eye acts as a retroreflector as the light reflects off the retina creating a bright pupil effect similar to red eye. If the illumination source is offset from the optical path, then the pupil appears dark because the retroreflection from the retina is directed away from the camera 101 .
- Bright-pupil tracking creates greater iris/pupil contrast, thus allowing more robust eye-tracking with all types of iris pigmentation.
- bright-pupil tracking greatly reduces interference caused by eyelashes and other obscuring features.
- Bright-pupil tracking also allows tracking in lighting conditions ranging from total darkness to very bright light.
- bright-pupil techniques are not as effective for tracking gaze in outdoors conditions and where extraneous inferred sources interfere with bright-pupil tracking.
- the gaze application 105 may use a process known as passive light tracking.
- Passive light tracking uses visible light to illuminate things which may cause distractions to users 110 .
- Another challenge with this passive light tracking is that the contrast of the pupil is less than in active light methods, therefore, the center of iris is used for calculating the gaze vector instead. This calculation needs to detect the boundary of the iris and the white sclera (limbus tracking).
- passive light tracking presents another challenge, such as detecting vertical eye movements due to obstruction of eyelids.
- the gaze application 105 may use an eye-tracking device.
- an eye tracking device may be a head-mounted device, a chin rest (request the user's head to be stable), an attached device (attached to the user 110 ) that remotely and automatically tracks head motion.
- These types of devices typically use a sampling rate of at least 30 Hz. Although a sampling rate of 50/60 Hz is more common, many video-based eye trackers run at 240, 350 or even 1000/1250 Hz, speeds in order to capture fixational eye movements and/or correctly measure saccade dynamics.
- Eye movements are typically divided into fixations and saccades (i.e., when the eye gaze pauses in a certain position and when it moves to another position respectively).
- the resulting series of fixations and saccades is called a scan-path.
- Smooth pursuit describes the eye following a moving object.
- Fixational eye movements may include micro saccades: small, involuntary saccades that occur during attempted fixation. Most information from the eye is made available during a fixation or smooth pursuit, but not during a saccade.
- the central one or two degrees of the visual angle that area of the visual field which falls on the fovea) provides the bulk of the visual information; the input from larger eccentricities (the periphery) has less resolution and little to no color, even though contrast and movement is detected better in peripheral vision.
- fixations last for around 200 milliseconds (ms) during the reading of linguistic text, and 350 ms during the viewing of a scene. Preparing a saccade towards a new goal takes around 200 ms.
- Scan-paths are useful for analyzing cognitive intent, interest, and salience. Other biological factors (some as simple as gender) may affect the scan-path as well. For example, eye tracking in human-computer interaction (HCI) typically investigates the scan-path for usability purposes and/or as a method of input in gaze-contingent displays (also known as gaze-based interfaces).
- HCI human-computer interaction
- Eye-trackers typically measure the rotation of the eye with respect to some frame of reference.
- the frame of reference is usually tied to the measuring system.
- the measuring system is head-mounted (e.g., as with Electrooculography (EOG) or a video-based system mounted to a helmet) then the eye-in-head angles are measured using the frame of reference.
- EOG Electrooculography
- the user's head must be kept in a constant position or its movements must be tracked as well.
- head direction is added to eye-in-head direction to determine the gaze direction.
- gaze application 105 uses a table-mounted device (e.g., as with scleral search coils or table-mounted camera 101 (“remote”) systems), gaze angles are measured directly in world coordinates. Typically, in these situations head movements are prohibited. For example, the head position is fixed using a bite bar or a forehead support. In this embodiment, a head-centered reference frame is identical to a world-centered reference frame. Or colloquially, the eye-in-head position directly determines the gaze direction.
- a table-mounted device e.g., as with scleral search coils or table-mounted camera 101 (“remote”) systems
- gaze angles are measured directly in world coordinates.
- head movements are prohibited.
- the head position is fixed using a bite bar or a forehead support.
- a head-centered reference frame is identical to a world-centered reference frame.
- the eye-in-head position directly determines the gaze direction.
- the gaze application 105 may be a distributed gaze application.
- the gaze application 105 may be in multiple devices that communicate via a computer network, such as the Internet.
- the gaze application 105 may be in a distributed access system that manages door access in multiple buildings.
- the electrical and/or mechanical elements 104 can be or may include a variety of elements, such as, a computer system, an electronic switch, an electronic control (e.g., a volume control), a relay, a system that sends electronic messages, an alarm system, a door lock, a disk space (e.g., grant access to the disk space), a dog/cat feeder, a vehicle lock, a garage door opener, a vehicle starter, a lighting system, an individual light, an electrode (e.g., for moving a muscle), a listening device (e.g., Amazon's Alexa®), a printer, a scanner, a computer, a laptop, a note pad, a heating/air system, a sprinkler system, an individual sprinkler, a temperature gauge, a card scanner, a biometric scanner, a sensor, a camera 101 , and/or the like.
- a computer system an electronic switch, an electronic control (e.g., a volume control), a relay, a system that sends electronic
- the first illustrative system 100 also shows a user 110 .
- the user 110 may be any person that has eyes that can be used to detect gaze.
- the user 110 is shown to be in view of the camera(s) 101 .
- an animal may also be within view of the camera(s) 101 instead of the user 110 (or along with the user 110 ).
- FIG. 2 is a diagram 200 that shows a process for using gaze to control an electrical and/or mechanical element 104 using gaze quadrants 201 A- 201 N (also identified by the numbers 1-9).
- the diagram 200 comprises the gaze quadrants 201 A- 201 N and the user 110 (a person).
- the gaze quadrants 201 A- 201 N are used to represent where a user 110 is gazing (view angles from the center gaze).
- a top gaze quadrant (the area of gaze quadrants 201 A- 201 C), a center gaze quadrant (the area of gaze quadrants 201 D- 201 F), and a bottom gaze quadrant (the area of gaze quadrants 201 G- 201 N).
- the user 110 is looking directly at the camera 101 as shown in step 210 A, the user 110 is gazing at gaze quadrant 201 E (the center gaze). If the same user 110 then looks up as shown in step 210 B, the user 110 is gazing at gaze quadrant 201 B.
- the gaze angle 211 changes a number of degrees, which is detected by the gaze application 105 (e.g., using one or more of the processes described above).
- a different gaze angle e.g., a left gaze angle and an up gaze angle in this example
- the gaze of the user 110 may be coupled with a time period. For example, the user 110 may have to look at the gaze quadrant 201 for a specific time period (e.g., 2 seconds).
- each of the gaze quadrants 201 A- 201 N can have a specific action associated with the gaze quadrant 201 .
- the action may be to ask the user 110 if he wants to place an order for a product from a specific service provider (e.g., Amazon®). The user 110 then can then say yes or no. If the user 110 says yes, the user 110 can then proceed and place the order.
- a specific service provider e.g., Amazon®
- the action may be to ask the user 110 if he wants to place an order using Google®. The user 110 can then say yes or no. Likewise, different actions may occur when the user 110 gazes at the various other gaze quadrants 201 .
- the gaze quadrants 201 A- 201 N may be for associated actions where the gaze quadrants 201 A- 201 N change based on different contexts of the associated actions.
- gaze quadrant 201 E (the center gaze) may be used to turn on a radio, which then changes the context of the gaze quadrants 201 A- 201 N.
- the gaze quadrants 201 A- 201 D and 201 F- 201 N may then be used to the select particular radio stations, particular playlists, and/or albums while the gaze quadrant 201 E (the center gaze) is used to turn off the radio.
- gaze quadrants 201 A- 201 N may be used for completely unrelated actions.
- gaze quadrant 201 A may be to turn on the lights and gaze quadrant 201 B may be used to place a voice call to a particular user 110 .
- the user 110 may be provided with an audible and/or visible indication of the action associated with the gaze quadrant 201 .
- the radio example from above where the gaze quadrants 201 A- 201 D and 201 F- 201 N represent individual radio stations when the user 110 is gazing in the gaze quadrant 201 A, the radio says “country 107.9, select?” The user 110 could then say “select” to select the radio station or change the gaze angle to a different gaze quadrant 201 .
- the radio says “rock 97.5, select?” The user 110 could then say “select” to select the radio station.
- the radio says “turn off the radio?” The user 110 could the say “yes” and turn off the radio.
- the gaze quadrants 201 A- 201 N may be used by another party (a party who is not being viewed by the camera 101 ).
- the gaze application 105 may be part of an alarm system that is used to detect unauthorized access to a building or facility (a restricted area).
- a silent alarm is generated (e.g., a security guard's screen or phone flashes/vibrates).
- an audible alarm sounds may also use other types of input, such as, facial recognition, biometrics, and/or the like.
- the gaze quadrants 201 A- 201 N can be used to detect a gaze sequence 220 by the user 110 .
- the gaze quadrants 201 A- 201 N can be used to detect the gaze sequence 220 .
- the gaze sequence 220 is where the user 110 gazes using the gaze sequence 1->4->5->8->9 (represented by the numbers at the top of the gaze quadrants 201 A- 201 N) to initiate an action.
- the gaze sequence 220 may be used to authenticate a user 110 (e.g., where there is no display or keyboard).
- the gaze quadrants 201 A- 201 N may represent a non-displayed keypad (representing the numbers 1-9).
- the user 110 can select the gaze quadrant 201 by looking at the gaze quadrate 201 (an angle from center) for a time period. This may be coupled with a voice input. For example, the user 110 may say “select” when looking at the gaze quadrant 201 A or “done” when entering the gaze sequence 220 .
- the gaze quadrants 201 A- 201 N are actually printed on a wall (or displayed on a device).
- the gaze quadrants 201 A- 201 N could be printed on a wall to look similar to FIG. 200 (without any text or numbers) where the camera 101 is located in the center of gaze quadrant 201 E.
- the gaze quadrants 201 A- 201 N could have text or numbers (e.g., 1-9 as shown in FIG. 2 ).
- the gaze sequence 220 represents the Personal Identification Number (PIN) of 1->4->5->8->9.
- PIN Personal Identification Number
- This type of authentication could be coupled with audible sounds (e.g., “select”) as the gaze moves from one gaze quadrant 201 to another gaze quadrant 201 .
- the audible sound would typically not state the numbers because another person could be listening and hear the PIN.
- the gaze quadrants 201 A- 201 N act like keypad without actually using a mechanical keypad.
- the gaze sequence 220 may be coupled with facial recognition. For example, the user's face is compared to a stored face print of the user 110 . Both the face print and the gaze sequence 220 have to match.
- authentication process may use multiple gaze sequences 220 from multiple users 110 (coupled with face prints) to provide an action. For example, access to a secure area may only be granted based on two people providing separate gaze sequences/facial recognition together (or in sequence).
- the gaze quadrants 201 A- 201 N could each have at least one camera 101 in each gaze quadrant 201 .
- the gaze sequence 220 is detected when the user 110 is directly looking at the particular camera 101 in the gaze quadrant 201 .
- the gaze quadrants 201 A- 201 N could be shown on an optical input screen (essentially an array of cameras 101 that captures gaze over the gaze quadrants 201 A- 201 N of the optical input screen).
- the optical input screen captures the direct gaze of the user 110 for each gaze quadrant 201 A- 201 N.
- the cameras 101 in each gaze quadrant 201 may be in various places.
- the cameras 101 may be two or more cameras 101 attached to the top of monitor, embedded in the top of the monitor, placed around the monitor (e.g., one on the top, right side, left side, and bottom), around a room, around a doorway, by a door frame, and/or the like.
- the gaze application 105 could use additional authentication metrics, such as, a password, a digital certificate, a biometric (e.g., a fingerprint scan, an iris (eye) scan, a palm print scan, a hearing range test, and/or the like) to authenticate the user 110 .
- additional authentication metrics such as, a password, a digital certificate, a biometric (e.g., a fingerprint scan, an iris (eye) scan, a palm print scan, a hearing range test, and/or the like) to authenticate the user 110 .
- the gaze application 105 may use different time periods to detect the gaze sequence 220 . For example, there may be a two second gaze period where the user 110 looks at the specific gaze quadrant 201 . When the user 110 is done with the gaze sequence 220 , the user 110 can look for a longer period of time (e.g., 4 seconds) on the last gaze quadrant 201 to end the gaze sequence 220 entry (or look away for a period of time).
- a longer period of time e.g. 4 seconds
- FIG. 3 is a flow diagram of a process for using gaze to control an electrical and/or mechanical element 104 .
- the camera(s) 101 and the gaze application 105 are stored-program-controlled entities, such as a computer or microprocessor 102 , which performs the method of FIGS. 3-5 and the processes described herein by executing program instructions stored in a computer readable storage medium, such as a memory 103 (i.e., a computer memory, a hard disk, and/or the like).
- a computer readable storage medium such as a memory 103 (i.e., a computer memory, a hard disk, and/or the like).
- FIGS. 3-5 are shown in a specific order, one of skill in the art would recognize that the steps in FIGS. 3-5 may be implemented in different orders and/or be implemented in a multi-threaded environment. Moreover, various steps may be omitted or added based on implementation.
- the process starts in step 300 .
- the gaze application 105 determines, in step 302 , if a gaze interaction is detected.
- the gaze interaction is with one or more of the cameras 101 .
- the gaze interaction can be where the user 110 is looking directly at the camera 101 or at an angle (e.g., as described in FIG. 2 ) from directly looking at the camera 101 .
- the gaze application 105 determines, in step 308 , if the process is complete. If the process is complete in step 308 , the process ends in step 310 . Otherwise, the process goes back to step 302 .
- the gaze application 105 determines, in step 304 , if there is an action associated with the gaze interaction. If there is not an action associated with the gaze interaction in step 304 , the process goes to step 308 . Otherwise, if the gaze application 105 identifies the associated action for controlling the electrical and/or mechanical element 104 in step 304 , the gaze application 105 the initiates the action for controlling the electrical and/or mechanical element 104 in step 306 . The process then goes to step 308 .
- the electrical and/or mechanical element(s) 104 is a listening device with a camera 101 (e.g., an Amazon® Alexa).
- a camera 101 e.g., an Amazon® Alexa
- the listening device is turned on and asks the user 110 if he/she wants to place an order or asks a question.
- the advantage to this approach is that the listening device is only active for the period of time when the user 110 want to make an order or ask a question.
- the listening device is not constantly listening, thus addressing privacy concerns.
- the user 110 can gaze at the listening device for a period of time to turn off the listening device (or based on a time period where no input is received).
- the gaze detection of FIG. 3 can be used in a variety ways in a variety of devices/systems, such as, opening/closing a garage door, turning on/off a specific light or group of lights, turning on/off a specific sprinkler or group of sprinklers, alarming an alarm system, changing a temperature setting of a heating/air system, turning on/off the heating/air system, opening/closing a door, securing a house, activating a computer, initiating a call, turning on/off a printer/scanner, and/or the like.
- the gaze detection can be coupled with other types of user input, such as gestures, voice, touch, and/or the like.
- a heating/air system may be turned on by the user 110 gazing at a controller for the heating/air system.
- the user 110 then may turn the temperature up or down based on a hand gesture.
- the user 110 may turn down the heat setting by moving their hand in a downward motion or turn up the heat setting by moving their hand in an upward motion.
- both gestures and gaze events may be required to initiate an action.
- the gaze coupled with the user 110 moving their hand up may be required to turn the hearing system temperature setting up.
- the gaze/gesture events may also use other types of input, such as voice and touch.
- the gaze detection can be coupled with touch and/or voice input.
- a user's gaze may turn on a device and then the user 110 can press a button, touch a touch screen, speak a verbal command, and/or the like to implement a specific action.
- FIG. 4 is a flow diagram of a process for using a gaze sequence 220 .
- the process of FIG. 4 goes between steps 304 and 306 of FIG. 3 .
- the process of FIG. 4 is based on the discussion of the gaze sequence 220 described in FIG. 2 .
- the gaze sequence 220 is a series of gaze interactions that are typically used as part of a user authentication process.
- the gaze sequence 220 may be used for a non-authentication actions, such as multi-function control process.
- a first gaze sequence 220 may be used to turn on/off a first light set and a second gaze sequence 220 may be used to turn on/off a second light set.
- the gaze detection in step 302 of FIG. 2 may switch modes based on context. For example, when the user 110 first accesses a system, the gaze interaction of step 302 may be for detecting a gaze sequence 220 for authenticating a user 110 . After the user 110 is authenticated, the gaze interaction of step 302 may switch to detecting a different type of gaze interaction. For example, the gaze interaction may cause a lighting system to turn on/off.
- the gaze application 105 determines if the gaze interaction is supposed to be for a gaze sequence 220 in step 400 . If the gaze application 105 is not looking for a gaze sequence 220 in step 400 , the process goes to step 306 . Otherwise, if the gaze application 105 is looking for a gaze sequence 220 in step 400 , the gaze application 105 determines, in step 402 , if the entered gaze sequence 220 matches a stored gaze sequence (e.g., a stored PIN). If the gaze sequence matches, in step 402 , the process goes to step 306 and initiates the action (e.g., to authenticate the user 110 and grant access to a secure room).
- a stored gaze sequence e.g., a stored PIN
- the gaze application 105 may optionally notify the user 110 of the invalid gaze sequence 220 in step 404 .
- the system may say “invalid gaze sequence entered. Try again.” The process then goes to step 308 .
- FIG. 5 is a flow diagram of a process for identifying a gaze quadrant 201 .
- FIG. 5 is an exemplary embodiment of step 306 of FIG. 3 where multiple gaze quadrants 201 are used (versus only a single gaze quadrant 201 ).
- the gaze application 105 After determining that there is an action associated with the gaze interaction in step 304 , the gaze application 105 identifies the gaze quadrant 201 that the user 110 is looking at in step 500 .
- the gaze application 105 identifies the action based on the gaze quadrant 201 for controlling the electrical and/or mechanical element 104 in step 502 .
- the process then goes to step 308 .
- certain components of the system can be located remotely, at distant portions of a distributed network, such as a LAN and/or the Internet, or within a dedicated system.
- a distributed network such as a LAN and/or the Internet
- the components of the system can be combined into one or more devices, such as a server, communication device, or collocated on a particular node of a distributed network, such as an analog and/or digital telecommunications network, a packet-switched network, or a circuit-switched network.
- the components of the system can be arranged at any location within a distributed network of components without affecting the operation of the system.
- the various links connecting the elements can be wired or wireless links, or any combination thereof, or any other known or later developed element(s) that is capable of supplying and/or communicating data to and from the connected elements.
- These wired or wireless links can also be secure links and may be capable of communicating encrypted information.
- Transmission media used as links can be any suitable carrier for electrical signals, including coaxial cables, copper wire, and fiber optics, and may take the form of acoustic or light waves, such as those generated during radio-wave and infra-red data communications.
- the systems and methods of this disclosure can be implemented in conjunction with a special purpose computer, a programmed microprocessor 102 or microcontroller and peripheral integrated circuit element(s), an ASIC or other integrated circuit, a digital signal processor, a hard-wired electronic or logic circuit such as discrete element circuit, a programmable logic device or gate array such as PLD, PLA, FPGA, PAL, special purpose computer, any comparable means, or the like.
- a special purpose computer e.g., cellular, Internet enabled, digital, analog, hybrids, and others
- other hardware known in the art e.g.
- processors e.g., a single or multiple microprocessors 102
- memory 103 nonvolatile storage
- input devices input devices
- output devices input devices
- alternative software implementations including, but not limited to, distributed processing or component/object distributed processing, parallel processing, or virtual machine processing can also be constructed to implement the methods described herein.
- the disclosed methods may be readily implemented in conjunction with software using object or object-oriented software development environments that provide portable source code that can be used on a variety of computer or workstation platforms.
- the disclosed system may be implemented partially or fully in hardware using standard logic circuits or VLSI design. Whether software or hardware is used to implement the systems in accordance with this disclosure is dependent on the speed and/or efficiency requirements of the system, the particular function, and the particular software or hardware systems or microprocessor 102 or microcomputer systems being utilized.
- the disclosed methods may be partially implemented in software that can be stored on a storage medium, executed on programmed general-purpose computer with the cooperation of a controller and memory 103 , a special purpose computer, a microprocessor 102 , or the like.
- the systems and methods of this disclosure can be implemented as a program embedded on a personal computer such as an applet, JAVA® or CGI script, as a resource residing on a server or computer workstation, as a routine embedded in a dedicated measurement system, system component, or the like.
- the system can also be implemented by physically incorporating the system and/or method into a software and/or hardware system.
- the present disclosure in various embodiments, configurations, and aspects, includes components, methods, processes, systems and/or apparatus substantially as depicted and described herein, including various embodiments, sub-combinations, and subsets thereof. Those of skill in the art will understand how to make and use the systems and methods disclosed herein after understanding the present disclosure.
- the present disclosure in various embodiments, configurations, and aspects, includes providing devices and processes in the absence of items not depicted and/or described herein or in various embodiments, configurations, or aspects hereof, including in the absence of such items as may have been used in previous devices or processes, e.g., for improving performance, achieving ease, and/or reducing cost of implementation.
- Embodiments include a device comprising: a microprocessor; a computer readable medium, coupled with the microprocessor and comprising microprocessor readable and executable instructions that program the microprocessor to: detect a gaze interaction from a user or animal based on information received from a camera; in response to detecting the gaze interaction, identify an action for controlling one or more of an electrical element and a mechanical element; and initiate the action to control the one or more of the electrical element and the mechanical element.
- aspects of the above device include wherein the camera is in the device, wherein the action is to power on the device, and wherein the microprocessor readable and executable instructions further program the microprocessor to ask if the user wants to speak a voice command that controls the device.
- aspects of the above device include wherein the gaze interaction is based on gaze information received from a plurality of different cameras and wherein the gaze interaction is a sequence of gaze interactions with the plurality of different cameras that is used for at least one of: a user authentication process and a multi-function control process.
- aspects of the above device include wherein the sequence of gaze interactions with the plurality of different cameras is used for the user authentication process.
- aspects of the above device include wherein the sequence of gaze interactions with the plurality of different cameras is used for the multi-function control process.
- the gaze interaction comprises a plurality of gaze quadrants that are used to initiate a plurality of different actions to control the one or more of the electrical element and the mechanical element.
- aspects of the above device include wherein the plurality of gaze quadrants are used as part of a gaze sequence for authenticating a user.
- aspects of the above device include wherein the gaze interaction is coupled with one or more gestures in order to initiate the action to control the one or more of the electrical element and the mechanical element.
- aspects of the above device include wherein detecting the gaze interaction is used to control an alarm system and wherein the action is to initiate different alarm actions based on whether the detected gaze interaction is a direct gaze interaction or an indirect gaze interaction.
- aspects of the above device include wherein the gaze interaction is coupled with one or more touch events to initiate the action to control the one or more of the electrical element and the mechanical element.
- aspects of the above device include wherein the gaze interaction is coupled with one or more voice commands to initiate the action to control the one or more of the electrical element and the mechanical element.
- Embodiments include a method comprising: detecting, by a microprocessor, a gaze interaction from a user or animal based on information received from a camera; in response to detecting the gaze interaction, identifying, by the microprocessor, an action for controlling one or more of an electrical element and a mechanical element; and initiating, by the microprocessor, the action to control the one or more of the electrical element and the mechanical element.
- aspects of the above method include wherein the camera is in the device, wherein the action is to power on the device, and further comprising, asking, by the microprocessor, if the user wants to speak a voice command that controls the device.
- aspects of the above method include wherein the gaze interaction is based on gaze information received from a plurality of different cameras and wherein the gaze interaction is a sequence of gaze interactions with the plurality of different cameras that is used for at least one of: a user authentication process and a multi-function control process.
- aspects of the above method include wherein the sequence of gaze interactions with the plurality of different cameras is used for the user authentication process.
- aspects of the above method include wherein the gaze interaction comprises a plurality of gaze quadrants that are used to initiate a plurality of different actions to control the one or more of the electrical element and the mechanical element.
- aspects of the above method include wherein the plurality of gaze quadrants are used as part of a gaze sequence for authenticating a user.
- aspects of the above method include wherein the gaze interaction is coupled with one or more gestures in order to initiate the action to control the one or more of the electrical element and the mechanical element.
- aspects of the above method include wherein detecting the gaze interaction is used to control an alarm system and wherein the action is to initiate different alarm actions based on whether the detected gaze interaction is a direct gaze interaction or an indirect gaze interaction.
- aspects of the above method include wherein the gaze interaction is coupled with one or more touch events to initiate the action to control the one or more of the electrical element and the mechanical element.
- each of the expressions “at least one of A, B and C,” “at least one of A, B, or C,” “one or more of A, B, and C,” “one or more of A, B, or C,” “A, B, and/or C,” and “A, B, or C” means A alone, B alone, C alone, A and B together, A and C together, B and C together, or A, B and C together.
- automated refers to any process or operation, which is typically continuous or semi-continuous, done without material human input when the process or operation is performed.
- a process or operation can be automatic, even though performance of the process or operation uses material or immaterial human input, if the input is received before performance of the process or operation.
- Human input is deemed to be material if such input influences how the process or operation will be performed. Human input that consents to the performance of the process or operation is not deemed to be “material.”
- aspects of the present disclosure may take the form of an embodiment that is entirely hardware, an embodiment that is entirely software (including firmware, resident software, micro-code, etc.) or an embodiment combining software and hardware aspects that may all generally be referred to herein as a “circuit,” “module,” or “system.” Any combination of one or more computer-readable medium(s) may be utilized.
- the computer-readable medium may be a computer-readable signal medium or a computer-readable storage medium.
- a computer-readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing. More specific examples (a non-exhaustive list) of the computer-readable storage medium would include the following: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing.
- a computer-readable storage medium may be any tangible medium that can contain or store a program for use by or in connection with an instruction execution system, apparatus, or device.
- a computer-readable signal medium may include a propagated data signal with computer-readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated signal may take any of a variety of forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof.
- a computer-readable signal medium may be any computer-readable medium that is not a computer-readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device.
- Program code embodied on a computer-readable medium may be transmitted using any appropriate medium, including, but not limited to, wireless, wireline, optical fiber cable, RF, etc., or any suitable combination of the foregoing.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- Computer Security & Cryptography (AREA)
- Software Systems (AREA)
- Computer Hardware Design (AREA)
- Multimedia (AREA)
- General Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Health & Medical Sciences (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
A system that detects a gaze interaction from a user or animal based on information received from a camera. For example, the user looks directly at the camera or is looking at an angle toward the camera. In response to detecting the gaze interaction, the system identifies an associated action for controlling one or more of an electrical element and a mechanical element. For example, the action may be to control lighting system. The action is then initiated to control the one or more of the electrical element and the mechanical element.
Description
- The present application claims the benefits of and priority, under 35 U.S.C. § 119(e), to U.S. Provisional Application Ser. No. 62/659,506, filed on Apr. 18, 2018, entitled “METHOD TO USE HUMAN AND ANIMAL GAZE TO CONTROL ELECTRONIC SWITCHES AND MACHINERY.” The entire disclosure of the application listed above is hereby incorporated by reference, in its entirety, for all that it teaches and for all purposes.
- The present invention is related generally to gaze recognition and specifically using gaze recognition to control switches and machinery.
- Currently, there are a couple of ways humans interact with machines/electrical components. The most common way is by touching or a tactile input device, such as a button or touchscreen. Another way to interact with machines/electrical components that is growing in popularity is use speech recognition to translate the speech into commands thereby enabling the machine/electrical component to understand human speech. These types of systems can have physical limitations of where a device is incapable of providing voice and or touch interaction. For example, the device may not have an area that can provide tactile input or does not have digital voice processing capabilities. For example, a small sensor may be unable to provide a tactile input or voice processing capabilities (e.g., supporting a microphone/speaker) due to its size. Alternatively, a device may not be easily reachable due to distance, thus limiting any tactile input and/or voice access.
-
FIG. 1 is a block diagram of a first illustrative system for using gaze to control an electrical and/or mechanical element; -
FIG. 2 is a diagram that shows a process for using gaze to control an electrical and/or mechanical element using gaze quadrants; -
FIG. 3 is a flow diagram of a process for using gaze to control an electrical and/or mechanical element; -
FIG. 4 is a flow diagram of a process for using a gaze sequence; and -
FIG. 5 is a flow diagram of a process for identifying a gaze quadrant. -
FIG. 1 is a block diagram of a firstillustrative system 100 for using gaze to control electrical and/or mechanical element(s) 104. The firstillustrative system 100 comprises camera(s) 101, amicroprocessor 102, amemory 103, and the electrical and/or mechanical element(s) 104. - The camera(s) 101 can be or may include any type of
camera 101 that can capture an image, such as a two dimensional camera, an acoustic camera (a three dimensional camera), an inferred camera, a digital camera, a video camera, and/or the like. The camera(s) 101 may send a video stream, a picture, and/or the like. The camera(s) 101 may also capture an audio stream associated with the video stream or picture (e.g., from a microphone (not shown)). - The
microprocessor 102 can be or may or can include any knownmicroprocessor 102, such as a microcontroller, an applicationspecific microprocessor 102, a digital signaling processor, a multi-core processor, and/or the like. Themicroprocessor 102 interfaces with the camera(s) 101, thememory 103, and the electrical and/or mechanical element(s) 104. Themicroprocessor 102 may interface with the camera(s) 101, thememory 103, and the electrical and/or mechanical element(s) 104 either directly (e.g., in the same device and/or on the same circuit board), via a computer network, via wiring, and/or the like. For example, the camera(s) 101 and/or the electrical and/or mechanical element(s) 104 may be connected to themicroprocessor 102/memory 103 via a computer network. Themicroprocessor 102 may comprise multiple processors. For example, themicroprocessor 102 may comprise amicroprocessor 102 in thecamera 101 or a touch screen device along with acentral microprocessor 102. - The camera(s) 101, the
microprocessor 102, thememory 103, and the electrical and/or mechanical element(s) 104 may be in the same device. Alternatively, the camera(s) 101 may be external from themicroprocessor 102/memory 103/electrical and/or mechanical element(s) 104. For example, thecamera 101 may be anexternal camera 101 attached to a computer monitor, an array ofcameras 101 embedded in a monitor, acamera 101 at a door, an array ofcameras 101 in a room, and/or the like. - The
memory 103 can be or may include, a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. - The
memory 103 further comprises agaze application 105. Thegaze application 105 can detect and manage gaze information received from the camera(s) 101. Thegaze application 105 can initiate different actions based on detecting and managing gaze. Thegaze application 105 can track the gaze or eye of a human and/or an animal. The detection of the gaze interaction may be used in situations where a user's hands are occupied, where theuser 110 does not have hands, where a device is out of arms reach (e.g., a light in a ceiling), where there is no space on the device that allows integration of touch interaction, where speech interaction is not possible, and/or the like. - The
gaze application 105 can control a gaze interaction that may be a primary or a supplementary gaze interaction, which can be used in combination with other types of interactions, such as speech interactions, touch interactions (e.g., auser 110 touching a touch screen or button), gesture interactions, and/or the like. For example, a speech interaction may follow a gaze interaction; in this case, the gaze of auser 110 looking at a device tells the device to turn on a microphone and then wait/listen for a speech interaction. In another embodiment, thegaze application 105 can turn off the microphone when no gaze detected, or in other words, when theuser 110 is not looking at the device. - The
gaze application 105 can be used not only to activate a switch or a device, but also can be used to deactivate the switch or device by measuring either the point of gaze (where one is looking) and/or the motion of an eye relative to the head. There are a number of methods for measuring eye movement that may be employed by thegaze application 105. One variant uses video images from which the eye position is extracted. Other embodiments may use search coils, eye-attached tracking (e.g., a special contact lens with an embedded mirror or magnetic field sensor), electric potential measurement (e.g., using electrodes positioned around the eyes of theuser 110 to generate an electrooculogram signal), and/or the like. - In addition, the
gaze application 105 can use a video-based eye-tracker. Thecamera 101 focuses on one or both eyes and records eye movement as theuser 110 looks a particular object. Thegaze application 105 can use eye-trackers to track the center of the pupil and infrared/near-infrared non-collimated light to create corneal reflections (CR). A vector between the pupil center and the corneal reflections may be used to compute the point of regard on the surface of the eye or the gaze direction of the eye. In this embodiment, a simple calibration procedure of theuser 110 may be needed before using thegaze application 103. - The
gaze application 105 may support various types of infrared/near-infrared (also known as active light) eye-tracking techniques. For example, thegaze application 105 may support bright-pupil and dark-pupil techniques. Differences between the two may be based on the location of an illumination source with respect to thecamera 101. If the illumination is coaxial with the optical path, then the eye acts as a retroreflector as the light reflects off the retina creating a bright pupil effect similar to red eye. If the illumination source is offset from the optical path, then the pupil appears dark because the retroreflection from the retina is directed away from thecamera 101. - Bright-pupil tracking creates greater iris/pupil contrast, thus allowing more robust eye-tracking with all types of iris pigmentation. In addition, bright-pupil tracking greatly reduces interference caused by eyelashes and other obscuring features. Bright-pupil tracking also allows tracking in lighting conditions ranging from total darkness to very bright light. However, bright-pupil techniques are not as effective for tracking gaze in outdoors conditions and where extraneous inferred sources interfere with bright-pupil tracking.
- In another embodiment, the
gaze application 105 may use a process known as passive light tracking. Passive light tracking uses visible light to illuminate things which may cause distractions tousers 110. Another challenge with this passive light tracking is that the contrast of the pupil is less than in active light methods, therefore, the center of iris is used for calculating the gaze vector instead. This calculation needs to detect the boundary of the iris and the white sclera (limbus tracking). In addition, passive light tracking presents another challenge, such as detecting vertical eye movements due to obstruction of eyelids. - In some embodiments the
gaze application 105 may use an eye-tracking device. For example, an eye tracking device may be a head-mounted device, a chin rest (request the user's head to be stable), an attached device (attached to the user 110) that remotely and automatically tracks head motion. These types of devices typically use a sampling rate of at least 30 Hz. Although a sampling rate of 50/60 Hz is more common, many video-based eye trackers run at 240, 350 or even 1000/1250 Hz, speeds in order to capture fixational eye movements and/or correctly measure saccade dynamics. - Eye movements are typically divided into fixations and saccades (i.e., when the eye gaze pauses in a certain position and when it moves to another position respectively). The resulting series of fixations and saccades is called a scan-path. Smooth pursuit describes the eye following a moving object. Fixational eye movements may include micro saccades: small, involuntary saccades that occur during attempted fixation. Most information from the eye is made available during a fixation or smooth pursuit, but not during a saccade. The central one or two degrees of the visual angle (that area of the visual field which falls on the fovea) provides the bulk of the visual information; the input from larger eccentricities (the periphery) has less resolution and little to no color, even though contrast and movement is detected better in peripheral vision. Hence, the locations of fixations or smooth pursuit along a scan-path shows what information loci on the stimulus was processed during an eye-tracking session. On average, fixations last for around 200 milliseconds (ms) during the reading of linguistic text, and 350 ms during the viewing of a scene. Preparing a saccade towards a new goal takes around 200 ms.
- Scan-paths are useful for analyzing cognitive intent, interest, and salience. Other biological factors (some as simple as gender) may affect the scan-path as well. For example, eye tracking in human-computer interaction (HCI) typically investigates the scan-path for usability purposes and/or as a method of input in gaze-contingent displays (also known as gaze-based interfaces).
- Eye-trackers typically measure the rotation of the eye with respect to some frame of reference. The frame of reference is usually tied to the measuring system. Thus, if the measuring system is head-mounted (e.g., as with Electrooculography (EOG) or a video-based system mounted to a helmet) then the eye-in-head angles are measured using the frame of reference. In order to deduce the line of sight in world coordinates, the user's head must be kept in a constant position or its movements must be tracked as well. In these embodiments, head direction is added to eye-in-head direction to determine the gaze direction.
- If the
gaze application 105 uses a table-mounted device (e.g., as with scleral search coils or table-mounted camera 101 (“remote”) systems), gaze angles are measured directly in world coordinates. Typically, in these situations head movements are prohibited. For example, the head position is fixed using a bite bar or a forehead support. In this embodiment, a head-centered reference frame is identical to a world-centered reference frame. Or colloquially, the eye-in-head position directly determines the gaze direction. - In one embodiment, the
gaze application 105 may be a distributed gaze application. Thegaze application 105 may be in multiple devices that communicate via a computer network, such as the Internet. For example, thegaze application 105 may be in a distributed access system that manages door access in multiple buildings. - The electrical and/or
mechanical elements 104 can be or may include a variety of elements, such as, a computer system, an electronic switch, an electronic control (e.g., a volume control), a relay, a system that sends electronic messages, an alarm system, a door lock, a disk space (e.g., grant access to the disk space), a dog/cat feeder, a vehicle lock, a garage door opener, a vehicle starter, a lighting system, an individual light, an electrode (e.g., for moving a muscle), a listening device (e.g., Amazon's Alexa®), a printer, a scanner, a computer, a laptop, a note pad, a heating/air system, a sprinkler system, an individual sprinkler, a temperature gauge, a card scanner, a biometric scanner, a sensor, acamera 101, and/or the like. - The first
illustrative system 100 also shows auser 110. Theuser 110 may be any person that has eyes that can be used to detect gaze. InFIG. 1 , theuser 110 is shown to be in view of the camera(s) 101. Although not shown inFIG. 1 , an animal may also be within view of the camera(s) 101 instead of the user 110 (or along with the user 110). -
FIG. 2 is a diagram 200 that shows a process for using gaze to control an electrical and/ormechanical element 104 usinggaze quadrants 201A-201N (also identified by the numbers 1-9). The diagram 200 comprises the gaze quadrants 201A-201N and the user 110 (a person). The gaze quadrants 201A-201N are used to represent where auser 110 is gazing (view angles from the center gaze). InFIG. 2 , there are nineillustrative gaze quadrants 201A-201N. However, in other embodiments, there may be more or less gaze quadrants 201. For example, there may only be three gaze quadrants (a top gaze quadrant (the area ofgaze quadrants 201A-201C), a center gaze quadrant (the area ofgaze quadrants 201D-201F), and a bottom gaze quadrant (the area ofgaze quadrants 201G-201N). - If the
user 110 is looking directly at thecamera 101 as shown instep 210A, theuser 110 is gazing atgaze quadrant 201E (the center gaze). If thesame user 110 then looks up as shown instep 210B, theuser 110 is gazing atgaze quadrant 201B. In this example, the gaze angle 211 changes a number of degrees, which is detected by the gaze application 105 (e.g., using one or more of the processes described above). Likewise, when theuser 110 gazes at the gaze quadrant 201C (or any other gaze quadrant 201), a different gaze angle (e.g., a left gaze angle and an up gaze angle in this example) is determined (e.g., based on a difference from the center gaze) to identify that theuser 110 is looking at the gaze quadrant 201C. The gaze of theuser 110 may be coupled with a time period. For example, theuser 110 may have to look at the gaze quadrant 201 for a specific time period (e.g., 2 seconds). - When the
user 110 gazes at one of the gaze quadrants 201 (e.g., for the specific time period), there is an action associated with the gaze quadrant 201. Each of the gaze quadrants 201A-201N can have a specific action associated with the gaze quadrant 201. For example, when theuser 110 gazes at the gaze quadrant 210E instep 210A, the action may be to ask theuser 110 if he wants to place an order for a product from a specific service provider (e.g., Amazon®). Theuser 110 then can then say yes or no. If theuser 110 says yes, theuser 110 can then proceed and place the order. If theuser 110 gazes that thegaze quadrant 201B as shown instep 210B, the action may be to ask theuser 110 if he wants to place an order using Google®. Theuser 110 can then say yes or no. Likewise, different actions may occur when theuser 110 gazes at the various other gaze quadrants 201. - The gaze quadrants 201A-201N may be for associated actions where the gaze quadrants 201A-201N change based on different contexts of the associated actions. For example,
gaze quadrant 201E (the center gaze) may be used to turn on a radio, which then changes the context of the gaze quadrants 201A-201N. For example, the gaze quadrants 201A-201D and 201F-201N may then be used to the select particular radio stations, particular playlists, and/or albums while thegaze quadrant 201E (the center gaze) is used to turn off the radio. - Alternatively, the gaze quadrants 201A-201N may be used for completely unrelated actions. For example,
gaze quadrant 201A may be to turn on the lights andgaze quadrant 201B may be used to place a voice call to aparticular user 110. - In one embodiment, as the
user 110 changes the gaze angle from one gaze quadrant 201 to another gaze quadrant 201, theuser 110 may be provided with an audible and/or visible indication of the action associated with the gaze quadrant 201. For example, using the radio example from above where the gaze quadrants 201A-201D and 201F-201N represent individual radio stations, when theuser 110 is gazing in thegaze quadrant 201A, the radio says “country 107.9, select?” Theuser 110 could then say “select” to select the radio station or change the gaze angle to a different gaze quadrant 201. For example, when theuser 110 changes the gaze angle to gaze atgaze quadrant 201B, the radio says “rock 97.5, select?” Theuser 110 could then say “select” to select the radio station. When theuser 110 gazes at thegaze quadrant 201E, the radio says “turn off the radio?” Theuser 110 could the say “yes” and turn off the radio. - In one embodiment, the gaze quadrants 201A-201N may be used by another party (a party who is not being viewed by the camera 101). For example, the
gaze application 105 may be part of an alarm system that is used to detect unauthorized access to a building or facility (a restricted area). When anunauthorized user 110 enters the restricted area and theuser 110 does not gaze directly at thecamera 101, a silent alarm is generated (e.g., a security guard's screen or phone flashes/vibrates). When theunauthorized user 110 looks directly at thecamera 101, an audible alarm sounds. The alarm system may also use other types of input, such as, facial recognition, biometrics, and/or the like. - In one embodiment, the gaze quadrants 201A-201N can be used to detect a
gaze sequence 220 by theuser 110. For example, as shown inFIG. 2 , the gaze quadrants 201A-201N can be used to detect thegaze sequence 220. Thegaze sequence 220 is where theuser 110 gazes using the gaze sequence 1->4->5->8->9 (represented by the numbers at the top of the gaze quadrants 201A-201N) to initiate an action. For example, thegaze sequence 220 may be used to authenticate a user 110 (e.g., where there is no display or keyboard). The gaze quadrants 201A-201N may represent a non-displayed keypad (representing the numbers 1-9). Theuser 110 can select the gaze quadrant 201 by looking at the gaze quadrate 201 (an angle from center) for a time period. This may be coupled with a voice input. For example, theuser 110 may say “select” when looking at thegaze quadrant 201A or “done” when entering thegaze sequence 220. - In one embodiment, the gaze quadrants 201A-201N are actually printed on a wall (or displayed on a device). For example, the gaze quadrants 201A-201N could be printed on a wall to look similar to
FIG. 200 (without any text or numbers) where thecamera 101 is located in the center ofgaze quadrant 201E. Alternatively, the gaze quadrants 201A-201N could have text or numbers (e.g., 1-9 as shown inFIG. 2 ). In this example, thegaze sequence 220 represents the Personal Identification Number (PIN) of 1->4->5->8->9. This type of authentication could be coupled with audible sounds (e.g., “select”) as the gaze moves from one gaze quadrant 201 to another gaze quadrant 201. In this embodiment, the audible sound would typically not state the numbers because another person could be listening and hear the PIN. In this embodiment, the gaze quadrants 201A-201N act like keypad without actually using a mechanical keypad. - In one embodiment, the
gaze sequence 220 may be coupled with facial recognition. For example, the user's face is compared to a stored face print of theuser 110. Both the face print and thegaze sequence 220 have to match. In one embodiment, authentication process may usemultiple gaze sequences 220 from multiple users 110 (coupled with face prints) to provide an action. For example, access to a secure area may only be granted based on two people providing separate gaze sequences/facial recognition together (or in sequence). - In one embodiment, the gaze quadrants 201A-201N could each have at least one
camera 101 in each gaze quadrant 201. In this embodiment, thegaze sequence 220 is detected when theuser 110 is directly looking at theparticular camera 101 in the gaze quadrant 201. For example, the gaze quadrants 201A-201N could be shown on an optical input screen (essentially an array ofcameras 101 that captures gaze over the gaze quadrants 201A-201N of the optical input screen). The optical input screen captures the direct gaze of theuser 110 for eachgaze quadrant 201A-201N. In one embodiment, thecameras 101 in each gaze quadrant 201 may be in various places. For example, thecameras 101 may be two ormore cameras 101 attached to the top of monitor, embedded in the top of the monitor, placed around the monitor (e.g., one on the top, right side, left side, and bottom), around a room, around a doorway, by a door frame, and/or the like. - In the embodiments that use the
gaze sequence 220 for authenticating auser 110, thegaze application 105 could use additional authentication metrics, such as, a password, a digital certificate, a biometric (e.g., a fingerprint scan, an iris (eye) scan, a palm print scan, a hearing range test, and/or the like) to authenticate theuser 110. - The
gaze application 105 may use different time periods to detect thegaze sequence 220. For example, there may be a two second gaze period where theuser 110 looks at the specific gaze quadrant 201. When theuser 110 is done with thegaze sequence 220, theuser 110 can look for a longer period of time (e.g., 4 seconds) on the last gaze quadrant 201 to end thegaze sequence 220 entry (or look away for a period of time). -
FIG. 3 is a flow diagram of a process for using gaze to control an electrical and/ormechanical element 104. Illustratively, the camera(s) 101 and thegaze application 105 are stored-program-controlled entities, such as a computer ormicroprocessor 102, which performs the method ofFIGS. 3-5 and the processes described herein by executing program instructions stored in a computer readable storage medium, such as a memory 103 (i.e., a computer memory, a hard disk, and/or the like). Although the methods described inFIGS. 3-5 are shown in a specific order, one of skill in the art would recognize that the steps inFIGS. 3-5 may be implemented in different orders and/or be implemented in a multi-threaded environment. Moreover, various steps may be omitted or added based on implementation. - The process starts in
step 300. Thegaze application 105 determines, instep 302, if a gaze interaction is detected. The gaze interaction is with one or more of thecameras 101. The gaze interaction can be where theuser 110 is looking directly at thecamera 101 or at an angle (e.g., as described inFIG. 2 ) from directly looking at thecamera 101. If a gaze interaction is not detected instep 302, thegaze application 105 determines, instep 308, if the process is complete. If the process is complete instep 308, the process ends instep 310. Otherwise, the process goes back tostep 302. - If the gaze interaction is detected in
step 302, thegaze application 105 determines, instep 304, if there is an action associated with the gaze interaction. If there is not an action associated with the gaze interaction instep 304, the process goes to step 308. Otherwise, if thegaze application 105 identifies the associated action for controlling the electrical and/ormechanical element 104 instep 304, thegaze application 105 the initiates the action for controlling the electrical and/ormechanical element 104 instep 306. The process then goes to step 308. - To further illustrate the process of
FIG. 3 , consider the following non-limiting examples. While the non-limiting examples are discussed in regard toFIG. 3 , one of skill in the art would clearly understand that these illustrative examples will also work with the embodiments discusses inFIGS. 1, 2, 4, and 5 . - In one embodiment, assume that the electrical and/or mechanical element(s) 104 is a listening device with a camera 101 (e.g., an Amazon® Alexa). When the
user 110 gazes at thecamera 101 in the listening device, the listening device is turned on and asks theuser 110 if he/she wants to place an order or asks a question. The advantage to this approach is that the listening device is only active for the period of time when theuser 110 want to make an order or ask a question. The listening device is not constantly listening, thus addressing privacy concerns. In a similar manner, theuser 110 can gaze at the listening device for a period of time to turn off the listening device (or based on a time period where no input is received). - The gaze detection of
FIG. 3 can be used in a variety ways in a variety of devices/systems, such as, opening/closing a garage door, turning on/off a specific light or group of lights, turning on/off a specific sprinkler or group of sprinklers, alarming an alarm system, changing a temperature setting of a heating/air system, turning on/off the heating/air system, opening/closing a door, securing a house, activating a computer, initiating a call, turning on/off a printer/scanner, and/or the like. - In one embodiment, the gaze detection can be coupled with other types of user input, such as gestures, voice, touch, and/or the like. For example, a heating/air system may be turned on by the
user 110 gazing at a controller for the heating/air system. Theuser 110 then may turn the temperature up or down based on a hand gesture. For example, theuser 110 may turn down the heat setting by moving their hand in a downward motion or turn up the heat setting by moving their hand in an upward motion. In some embodiments, both gestures and gaze events may be required to initiate an action. For example, the gaze coupled with theuser 110 moving their hand up may be required to turn the hearing system temperature setting up. The gaze/gesture events may also use other types of input, such as voice and touch. - Alternatively, the gaze detection can be coupled with touch and/or voice input. For example, a user's gaze may turn on a device and then the
user 110 can press a button, touch a touch screen, speak a verbal command, and/or the like to implement a specific action. -
FIG. 4 is a flow diagram of a process for using agaze sequence 220. The process ofFIG. 4 goes betweensteps FIG. 3 . The process ofFIG. 4 is based on the discussion of thegaze sequence 220 described inFIG. 2 . As discussed inFIG. 2 , thegaze sequence 220 is a series of gaze interactions that are typically used as part of a user authentication process. - However, in one embodiment, the
gaze sequence 220 may be used for a non-authentication actions, such as multi-function control process. For example, afirst gaze sequence 220 may be used to turn on/off a first light set and asecond gaze sequence 220 may be used to turn on/off a second light set. - The gaze detection in
step 302 ofFIG. 2 may switch modes based on context. For example, when theuser 110 first accesses a system, the gaze interaction ofstep 302 may be for detecting agaze sequence 220 for authenticating auser 110. After theuser 110 is authenticated, the gaze interaction ofstep 302 may switch to detecting a different type of gaze interaction. For example, the gaze interaction may cause a lighting system to turn on/off. - After determining that there is an action associated with the detected gaze interaction in
step 304, thegaze application 105 determines if the gaze interaction is supposed to be for agaze sequence 220 instep 400. If thegaze application 105 is not looking for agaze sequence 220 instep 400, the process goes to step 306. Otherwise, if thegaze application 105 is looking for agaze sequence 220 instep 400, thegaze application 105 determines, instep 402, if the enteredgaze sequence 220 matches a stored gaze sequence (e.g., a stored PIN). If the gaze sequence matches, instep 402, the process goes to step 306 and initiates the action (e.g., to authenticate theuser 110 and grant access to a secure room). If thegaze sequence 220 does not match instep 402, thegaze application 105 may optionally notify theuser 110 of theinvalid gaze sequence 220 instep 404. For example, the system may say “invalid gaze sequence entered. Try again.” The process then goes to step 308. -
FIG. 5 is a flow diagram of a process for identifying a gaze quadrant 201.FIG. 5 is an exemplary embodiment ofstep 306 ofFIG. 3 where multiple gaze quadrants 201 are used (versus only a single gaze quadrant 201). After determining that there is an action associated with the gaze interaction instep 304, thegaze application 105 identifies the gaze quadrant 201 that theuser 110 is looking at instep 500. Thegaze application 105 identifies the action based on the gaze quadrant 201 for controlling the electrical and/ormechanical element 104 instep 502. The process then goes to step 308. - Any of the steps, functions, and operations discussed herein can be performed continuously and automatically.
- To avoid unnecessarily obscuring the present disclosure, the preceding description omits a number of known structures and devices. This omission is not to be construed as a limitation of the scope of the claimed disclosure. Specific details are set forth to provide an understanding of the present disclosure. It should, however, be appreciated that the present disclosure may be practiced in a variety of ways beyond the specific detail set forth herein.
- Furthermore, while the exemplary embodiments illustrated herein show the various components of the system collocated, certain components of the system can be located remotely, at distant portions of a distributed network, such as a LAN and/or the Internet, or within a dedicated system. Thus, it should be appreciated, that the components of the system can be combined into one or more devices, such as a server, communication device, or collocated on a particular node of a distributed network, such as an analog and/or digital telecommunications network, a packet-switched network, or a circuit-switched network. It will be appreciated from the preceding description, and for reasons of computational efficiency, that the components of the system can be arranged at any location within a distributed network of components without affecting the operation of the system.
- Furthermore, it should be appreciated that the various links connecting the elements can be wired or wireless links, or any combination thereof, or any other known or later developed element(s) that is capable of supplying and/or communicating data to and from the connected elements. These wired or wireless links can also be secure links and may be capable of communicating encrypted information. Transmission media used as links, for example, can be any suitable carrier for electrical signals, including coaxial cables, copper wire, and fiber optics, and may take the form of acoustic or light waves, such as those generated during radio-wave and infra-red data communications.
- While the flowcharts have been discussed and illustrated in relation to a particular sequence of events, it should be appreciated that changes, additions, and omissions to this sequence can occur without materially affecting the operation of the disclosed embodiments, configuration, and aspects.
- A number of variations and modifications of the disclosure can be used. It would be possible to provide for some features of the disclosure without providing others.
- In yet another embodiment, the systems and methods of this disclosure can be implemented in conjunction with a special purpose computer, a
programmed microprocessor 102 or microcontroller and peripheral integrated circuit element(s), an ASIC or other integrated circuit, a digital signal processor, a hard-wired electronic or logic circuit such as discrete element circuit, a programmable logic device or gate array such as PLD, PLA, FPGA, PAL, special purpose computer, any comparable means, or the like. In general, any device(s) or means capable of implementing the methodology illustrated herein can be used to implement the various aspects of this disclosure. Exemplary hardware that can be used for the present disclosure includes computers, handheld devices, telephones (e.g., cellular, Internet enabled, digital, analog, hybrids, and others), and other hardware known in the art. Some of these devices include processors (e.g., a single or multiple microprocessors 102),memory 103, nonvolatile storage, input devices, and output devices. Furthermore, alternative software implementations including, but not limited to, distributed processing or component/object distributed processing, parallel processing, or virtual machine processing can also be constructed to implement the methods described herein. - In yet another embodiment, the disclosed methods may be readily implemented in conjunction with software using object or object-oriented software development environments that provide portable source code that can be used on a variety of computer or workstation platforms. Alternatively, the disclosed system may be implemented partially or fully in hardware using standard logic circuits or VLSI design. Whether software or hardware is used to implement the systems in accordance with this disclosure is dependent on the speed and/or efficiency requirements of the system, the particular function, and the particular software or hardware systems or
microprocessor 102 or microcomputer systems being utilized. - In yet another embodiment, the disclosed methods may be partially implemented in software that can be stored on a storage medium, executed on programmed general-purpose computer with the cooperation of a controller and
memory 103, a special purpose computer, amicroprocessor 102, or the like. In these instances, the systems and methods of this disclosure can be implemented as a program embedded on a personal computer such as an applet, JAVA® or CGI script, as a resource residing on a server or computer workstation, as a routine embedded in a dedicated measurement system, system component, or the like. The system can also be implemented by physically incorporating the system and/or method into a software and/or hardware system. - Although the present disclosure describes components and functions implemented in the embodiments with reference to particular standards and protocols, the disclosure is not limited to such standards and protocols. Other similar standards and protocols not mentioned herein are in existence and are considered to be included in the present disclosure. Moreover, the standards and protocols mentioned herein, and other similar standards and protocols not mentioned herein, are periodically superseded by faster or more effective equivalents having essentially the same functions. Such replacement standards and protocols having the same functions are considered equivalents included in the present disclosure.
- The present disclosure, in various embodiments, configurations, and aspects, includes components, methods, processes, systems and/or apparatus substantially as depicted and described herein, including various embodiments, sub-combinations, and subsets thereof. Those of skill in the art will understand how to make and use the systems and methods disclosed herein after understanding the present disclosure. The present disclosure, in various embodiments, configurations, and aspects, includes providing devices and processes in the absence of items not depicted and/or described herein or in various embodiments, configurations, or aspects hereof, including in the absence of such items as may have been used in previous devices or processes, e.g., for improving performance, achieving ease, and/or reducing cost of implementation.
- The foregoing discussion of the disclosure has been presented for purposes of illustration and description. The foregoing is not intended to limit the disclosure to the form or forms disclosed herein. In the foregoing Detailed Description for example, various features of the disclosure are grouped together in one or more embodiments, configurations, or aspects for the purpose of streamlining the disclosure. The features of the embodiments, configurations, or aspects of the disclosure may be combined in alternate embodiments, configurations, or aspects other than those discussed above. This method of disclosure is not to be interpreted as reflecting an intention that the claimed disclosure requires more features than are expressly recited in each claim. Rather, as the following claims reflect, inventive aspects lie in less than all features of a single foregoing disclosed embodiment, configuration, or aspect. Thus, the following claims are hereby incorporated into this Detailed Description, with each claim standing on its own as a separate preferred embodiment of the disclosure.
- Moreover, though the description of the disclosure has included description of one or more embodiments, configurations, or aspects and certain variations and modifications, other variations, combinations, and modifications are within the scope of the disclosure, e.g., as may be within the skill and knowledge of those in the art, after understanding the present disclosure. It is intended to obtain rights, which include alternative embodiments, configurations, or aspects to the extent permitted, including alternate, interchangeable and/or equivalent structures, functions, ranges, or steps to those claimed, whether or not such alternate, interchangeable and/or equivalent structures, functions, ranges, or steps are disclosed herein, and without intending to publicly dedicate any patentable subject matter.
- Embodiments include a device comprising: a microprocessor; a computer readable medium, coupled with the microprocessor and comprising microprocessor readable and executable instructions that program the microprocessor to: detect a gaze interaction from a user or animal based on information received from a camera; in response to detecting the gaze interaction, identify an action for controlling one or more of an electrical element and a mechanical element; and initiate the action to control the one or more of the electrical element and the mechanical element.
- Aspects of the above device include wherein the camera is in the device, wherein the action is to power on the device, and wherein the microprocessor readable and executable instructions further program the microprocessor to ask if the user wants to speak a voice command that controls the device.
- Aspects of the above device include wherein the gaze interaction is based on gaze information received from a plurality of different cameras and wherein the gaze interaction is a sequence of gaze interactions with the plurality of different cameras that is used for at least one of: a user authentication process and a multi-function control process.
- Aspects of the above device include wherein the sequence of gaze interactions with the plurality of different cameras is used for the user authentication process.
- Aspects of the above device include wherein the sequence of gaze interactions with the plurality of different cameras is used for the multi-function control process.
- Aspects of the above device include wherein the gaze interaction comprises a plurality of gaze quadrants that are used to initiate a plurality of different actions to control the one or more of the electrical element and the mechanical element.
- Aspects of the above device include wherein the plurality of gaze quadrants are used as part of a gaze sequence for authenticating a user.
- Aspects of the above device include wherein the gaze interaction is coupled with one or more gestures in order to initiate the action to control the one or more of the electrical element and the mechanical element.
- Aspects of the above device include wherein detecting the gaze interaction is used to control an alarm system and wherein the action is to initiate different alarm actions based on whether the detected gaze interaction is a direct gaze interaction or an indirect gaze interaction.
- Aspects of the above device include wherein the gaze interaction is coupled with one or more touch events to initiate the action to control the one or more of the electrical element and the mechanical element.
- Aspects of the above device include wherein the gaze interaction is coupled with one or more voice commands to initiate the action to control the one or more of the electrical element and the mechanical element.
- Embodiments include a method comprising: detecting, by a microprocessor, a gaze interaction from a user or animal based on information received from a camera; in response to detecting the gaze interaction, identifying, by the microprocessor, an action for controlling one or more of an electrical element and a mechanical element; and initiating, by the microprocessor, the action to control the one or more of the electrical element and the mechanical element.
- Aspects of the above method include wherein the camera is in the device, wherein the action is to power on the device, and further comprising, asking, by the microprocessor, if the user wants to speak a voice command that controls the device.
- Aspects of the above method include wherein the gaze interaction is based on gaze information received from a plurality of different cameras and wherein the gaze interaction is a sequence of gaze interactions with the plurality of different cameras that is used for at least one of: a user authentication process and a multi-function control process.
- Aspects of the above method include wherein the sequence of gaze interactions with the plurality of different cameras is used for the user authentication process.
- Aspects of the above method include wherein the gaze interaction comprises a plurality of gaze quadrants that are used to initiate a plurality of different actions to control the one or more of the electrical element and the mechanical element.
- Aspects of the above method include wherein the plurality of gaze quadrants are used as part of a gaze sequence for authenticating a user.
- Aspects of the above method include wherein the gaze interaction is coupled with one or more gestures in order to initiate the action to control the one or more of the electrical element and the mechanical element.
- Aspects of the above method include wherein detecting the gaze interaction is used to control an alarm system and wherein the action is to initiate different alarm actions based on whether the detected gaze interaction is a direct gaze interaction or an indirect gaze interaction.
- Aspects of the above method include wherein the gaze interaction is coupled with one or more touch events to initiate the action to control the one or more of the electrical element and the mechanical element.
- The phrases “at least one,” “one or more,” “or,” and “and/or” are open-ended expressions that are both conjunctive and disjunctive in operation. For example, each of the expressions “at least one of A, B and C,” “at least one of A, B, or C,” “one or more of A, B, and C,” “one or more of A, B, or C,” “A, B, and/or C,” and “A, B, or C” means A alone, B alone, C alone, A and B together, A and C together, B and C together, or A, B and C together.
- The term “a” or “an” entity refers to one or more of that entity. As such, the terms “a” (or “an”), “one or more,” and “at least one” can be used interchangeably herein. It is also to be noted that the terms “comprising,” “including,” and “having” can be used interchangeably.
- The term “automatic” and variations thereof, as used herein, refers to any process or operation, which is typically continuous or semi-continuous, done without material human input when the process or operation is performed. However, a process or operation can be automatic, even though performance of the process or operation uses material or immaterial human input, if the input is received before performance of the process or operation. Human input is deemed to be material if such input influences how the process or operation will be performed. Human input that consents to the performance of the process or operation is not deemed to be “material.”
- Aspects of the present disclosure may take the form of an embodiment that is entirely hardware, an embodiment that is entirely software (including firmware, resident software, micro-code, etc.) or an embodiment combining software and hardware aspects that may all generally be referred to herein as a “circuit,” “module,” or “system.” Any combination of one or more computer-readable medium(s) may be utilized. The computer-readable medium may be a computer-readable signal medium or a computer-readable storage medium.
- A computer-readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing. More specific examples (a non-exhaustive list) of the computer-readable storage medium would include the following: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In the context of this document, a computer-readable storage medium may be any tangible medium that can contain or store a program for use by or in connection with an instruction execution system, apparatus, or device.
- A computer-readable signal medium may include a propagated data signal with computer-readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated signal may take any of a variety of forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof. A computer-readable signal medium may be any computer-readable medium that is not a computer-readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device. Program code embodied on a computer-readable medium may be transmitted using any appropriate medium, including, but not limited to, wireless, wireline, optical fiber cable, RF, etc., or any suitable combination of the foregoing.
- The terms “determine,” “calculate,” “compute,” and variations thereof, as used herein, are used interchangeably and include any type of methodology, process, mathematical operation or technique.
Claims (20)
1. A device comprising:
a microprocessor;
a computer readable medium, coupled with the microprocessor and comprising microprocessor readable and executable instructions that program the microprocessor to:
detect a gaze interaction from a user or animal based on information received from a camera;
in response to detecting the gaze interaction, identify an action for controlling one or more of an electrical element and a mechanical element; and
initiate the action to control the one or more of the electrical element and the mechanical element.
2. The device of claim 1 , wherein the camera is in the device, wherein the action is to power on the device, and wherein the microprocessor readable and executable instructions further program the microprocessor to ask if the user wants to speak a voice command that controls the device.
3. The device of claim 1 , wherein the gaze interaction is based on gaze information received from a plurality of different cameras and wherein the gaze interaction is a sequence of gaze interactions with the plurality of different cameras that is used for at least one of: a user authentication process and a multi-function control process.
4. The device of claim 3 , wherein the sequence of gaze interactions with the plurality of different cameras is used for the user authentication process.
5. The device of claim 3 , wherein the sequence of gaze interactions with the plurality of different cameras is used for the multi-function control process.
6. The device of claim 1 , wherein the gaze interaction comprises a plurality of gaze quadrants that are used to initiate a plurality of different actions to control the one or more of the electrical element and the mechanical element.
7. The device of claim 6 , wherein the plurality of gaze quadrants are used as part of a gaze sequence for authenticating a user.
8. The device of claim 1 , wherein the gaze interaction is coupled with one or more gestures in order to initiate the action to control the one or more of the electrical element and the mechanical element.
9. The device of claim 1 , wherein detecting the gaze interaction is used to control an alarm system and wherein the action is to initiate different alarm actions based on whether the detected gaze interaction is a direct gaze interaction or an indirect gaze interaction.
10. The device of claim 1 , wherein the gaze interaction is coupled with one or more touch events to initiate the action to control the one or more of the electrical element and the mechanical element.
11. The device of claim 1 , wherein the gaze interaction is coupled with one or more voice commands to initiate the action to control the one or more of the electrical element and the mechanical element.
12. A method comprising:
detecting, by a microprocessor, a gaze interaction from a user or animal based on information received from a camera;
in response to detecting the gaze interaction, identifying, by the microprocessor, an action for controlling one or more of an electrical element and a mechanical element; and
initiating, by the microprocessor, the action to control the one or more of the electrical element and the mechanical element.
13. The method of claim 12 , wherein the camera is in the device, wherein the action is to power on the device, and further comprising, asking, by the microprocessor, if the user wants to speak a voice command that controls the device.
14. The method of claim 12 , wherein the gaze interaction is based on gaze information received from a plurality of different cameras and wherein the gaze interaction is a sequence of gaze interactions with the plurality of different cameras that is used for at least one of: a user authentication process and a multi-function control process.
15. The method of claim 14 , wherein the sequence of gaze interactions with the plurality of different cameras is used for the user authentication process.
16. The method of claim 12 , wherein the gaze interaction comprises a plurality of gaze quadrants that are used to initiate a plurality of different actions to control the one or more of the electrical element and the mechanical element.
17. The method of claim 16 , wherein the plurality of gaze quadrants are used as part of a gaze sequence for authenticating a user.
18. The method of claim 12 , wherein the gaze interaction is coupled with one or more gestures in order to initiate the action to control the one or more of the electrical element and the mechanical element.
19. The method of claim 12 , wherein detecting the gaze interaction is used to control an alarm system and wherein the action is to initiate different alarm actions based on whether the detected gaze interaction is a direct gaze interaction or an indirect gaze interaction.
20. The method of claim 12 , wherein the gaze interaction is coupled with one or more touch events to initiate the action to control the one or more of the electrical element and the mechanical element.
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US16/029,808 US20190324531A1 (en) | 2018-04-18 | 2018-07-09 | System and method for using gaze control to control electronic switches and machinery |
US16/570,748 US10936060B2 (en) | 2018-04-18 | 2019-09-13 | System and method for using gaze control to control electronic switches and machinery |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201862659506P | 2018-04-18 | 2018-04-18 | |
US16/029,808 US20190324531A1 (en) | 2018-04-18 | 2018-07-09 | System and method for using gaze control to control electronic switches and machinery |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/570,748 Continuation US10936060B2 (en) | 2018-04-18 | 2019-09-13 | System and method for using gaze control to control electronic switches and machinery |
Publications (1)
Publication Number | Publication Date |
---|---|
US20190324531A1 true US20190324531A1 (en) | 2019-10-24 |
Family
ID=63491829
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/029,808 Abandoned US20190324531A1 (en) | 2018-04-18 | 2018-07-09 | System and method for using gaze control to control electronic switches and machinery |
US16/570,748 Active US10936060B2 (en) | 2018-04-18 | 2019-09-13 | System and method for using gaze control to control electronic switches and machinery |
Family Applications After (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/570,748 Active US10936060B2 (en) | 2018-04-18 | 2019-09-13 | System and method for using gaze control to control electronic switches and machinery |
Country Status (2)
Country | Link |
---|---|
US (2) | US20190324531A1 (en) |
WO (1) | WO2019202355A1 (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10936060B2 (en) | 2018-04-18 | 2021-03-02 | Flex Ltd. | System and method for using gaze control to control electronic switches and machinery |
US20220048387A1 (en) * | 2020-08-12 | 2022-02-17 | Hyundai Motor Company | Vehicle and method of controlling the same |
US11789554B2 (en) * | 2020-07-29 | 2023-10-17 | Motorola Mobility Llc | Task invocation based on control actuation, fingerprint detection, and gaze detection |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
TW202301083A (en) | 2021-06-28 | 2023-01-01 | 見臻科技股份有限公司 | Optical system providing accurate eye-tracking and related method |
Family Cites Families (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20060192775A1 (en) | 2005-02-25 | 2006-08-31 | Microsoft Corporation | Using detected visual cues to change computer system operating states |
WO2013033842A1 (en) | 2011-09-07 | 2013-03-14 | Tandemlaunch Technologies Inc. | System and method for using eye gaze information to enhance interactions |
US9092600B2 (en) | 2012-11-05 | 2015-07-28 | Microsoft Technology Licensing, Llc | User authentication on augmented reality display device |
US20150227735A1 (en) | 2014-02-13 | 2015-08-13 | Robert Chappell | System and method for eye tracking authentication |
KR102223728B1 (en) * | 2014-06-20 | 2021-03-05 | 엘지전자 주식회사 | Mobile terminal and method for controlling the same |
US10228904B2 (en) | 2014-11-12 | 2019-03-12 | Lenovo (Singapore) Pte. Ltd. | Gaze triggered voice recognition incorporating device velocity |
US9812126B2 (en) * | 2014-11-28 | 2017-11-07 | Microsoft Technology Licensing, Llc | Device arbitration for listening devices |
US9990921B2 (en) * | 2015-12-09 | 2018-06-05 | Lenovo (Singapore) Pte. Ltd. | User focus activated voice recognition |
WO2019202355A1 (en) | 2018-04-18 | 2019-10-24 | Flex Ltd. | System and method for using gaze control to control electronic switches and machinery |
-
2018
- 2018-07-09 WO PCT/IB2018/000858 patent/WO2019202355A1/en active Application Filing
- 2018-07-09 US US16/029,808 patent/US20190324531A1/en not_active Abandoned
-
2019
- 2019-09-13 US US16/570,748 patent/US10936060B2/en active Active
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10936060B2 (en) | 2018-04-18 | 2021-03-02 | Flex Ltd. | System and method for using gaze control to control electronic switches and machinery |
US11789554B2 (en) * | 2020-07-29 | 2023-10-17 | Motorola Mobility Llc | Task invocation based on control actuation, fingerprint detection, and gaze detection |
US20220048387A1 (en) * | 2020-08-12 | 2022-02-17 | Hyundai Motor Company | Vehicle and method of controlling the same |
US11667196B2 (en) * | 2020-08-12 | 2023-06-06 | Hyundai Motor Company | Vehicle and method of controlling the same |
Also Published As
Publication number | Publication date |
---|---|
WO2019202355A1 (en) | 2019-10-24 |
US20200081528A1 (en) | 2020-03-12 |
US10936060B2 (en) | 2021-03-02 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10936060B2 (en) | System and method for using gaze control to control electronic switches and machinery | |
CN110874129B (en) | Display system | |
JP6722272B2 (en) | User identification and/or authentication using gaze information | |
CN106471419B (en) | Management information is shown | |
EP3035656A1 (en) | Method and apparatus for controlling an electronic device | |
US11119573B2 (en) | Pupil modulation as a cognitive control signal | |
US10345902B1 (en) | Method and apparatus for maintaining a secure head-mounted display session | |
KR20150112621A (en) | Mobile terminal and control method thereof | |
WO2015011703A1 (en) | Method and system for touchless activation of a device | |
EP3593191B1 (en) | In-game reactions to interruptions | |
WO2016114496A1 (en) | Method for providing user interface through head mounted display using eye recognition and bio-signal, apparatus using same, and computer readable recording medium | |
US11188154B2 (en) | Context dependent projection of holographic objects | |
US11516214B2 (en) | Authenticated device assisted user authentication | |
US11632258B1 (en) | Recognizing and mitigating displays of unacceptable and unhealthy behavior by participants of online video meetings | |
US20240164672A1 (en) | Stress detection | |
KR101728707B1 (en) | Method and program for controlling electronic device by wearable glass device | |
US12112441B2 (en) | Content transformations based on reflective object recognition | |
KR20160053391A (en) | System, method and application for confirmation of identity by wearable glass device | |
WO2016010328A1 (en) | Information processing system and method using wearable device | |
CN109144263A (en) | Social householder method, device, storage medium and wearable device | |
US20240319789A1 (en) | User interactions and eye tracking with text embedded elements | |
US20240212272A1 (en) | Interactions based on mirror detection and context awareness | |
US20230418372A1 (en) | Gaze behavior detection | |
US20240221301A1 (en) | Extended reality assistance based on user understanding | |
US20230351676A1 (en) | Transitioning content in views of three-dimensional environments using alternative positional constraints |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: FLEX LTD., SINGAPORE Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:ERASLAN, MESUT GORKEM;REEL/FRAME:046362/0422 Effective date: 20180629 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |