RU2017103121A3 - - Google Patents

Download PDF

Info

Publication number
RU2017103121A3
RU2017103121A3 RU2017103121A RU2017103121A RU2017103121A3 RU 2017103121 A3 RU2017103121 A3 RU 2017103121A3 RU 2017103121 A RU2017103121 A RU 2017103121A RU 2017103121 A RU2017103121 A RU 2017103121A RU 2017103121 A3 RU2017103121 A3 RU 2017103121A3
Authority
RU
Russia
Application number
RU2017103121A
Other versions
RU2017103121A (ru
RU2693308C2 (ru
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed filed Critical
Publication of RU2017103121A publication Critical patent/RU2017103121A/ru
Publication of RU2017103121A3 publication Critical patent/RU2017103121A3/ru
Application granted granted Critical
Publication of RU2693308C2 publication Critical patent/RU2693308C2/ru

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • G06F3/013Eye tracking input arrangements
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/0304Detection arrangements using opto-electronic means
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • G06F3/04842Selection of displayed objects or displayed text elements
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/10Image acquisition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation
    • G06V40/165Detection; Localisation; Normalisation using facial parts and geometric relationships
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/172Classification, e.g. identification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/18Eye characteristics, e.g. of the iris
    • G06V40/19Sensors therefor
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2203/00Indexing scheme relating to G06F3/00 - G06F3/048
    • G06F2203/038Indexing scheme relating to G06F3/038
    • G06F2203/0381Multimodal input, i.e. interface arrangements enabling the user to issue commands by simultaneous use of input devices of different nature, e.g. voice plus gesture on digitizer
RU2017103121A 2014-08-01 2015-07-29 Активация органов управления на основе отражения RU2693308C2 (ru)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US14/449,435 US9645641B2 (en) 2014-08-01 2014-08-01 Reflection-based control activation
US14/449,435 2014-08-01
PCT/US2015/042550 WO2016018957A1 (en) 2014-08-01 2015-07-29 Reflection-based control activation

Publications (3)

Publication Number Publication Date
RU2017103121A RU2017103121A (ru) 2018-08-01
RU2017103121A3 true RU2017103121A3 (ru) 2018-12-20
RU2693308C2 RU2693308C2 (ru) 2019-07-02

Family

ID=53836836

Family Applications (1)

Application Number Title Priority Date Filing Date
RU2017103121A RU2693308C2 (ru) 2014-08-01 2015-07-29 Активация органов управления на основе отражения

Country Status (11)

Country Link
US (1) US9645641B2 (ru)
EP (1) EP3175325B1 (ru)
JP (1) JP6898234B2 (ru)
KR (1) KR102392437B1 (ru)
CN (1) CN107077200B (ru)
AU (1) AU2015296666B2 (ru)
BR (1) BR112017001326A2 (ru)
CA (1) CA2955072C (ru)
MX (1) MX2017001514A (ru)
RU (1) RU2693308C2 (ru)
WO (1) WO2016018957A1 (ru)

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2014181403A1 (ja) * 2013-05-08 2014-11-13 富士通株式会社 入力装置及び入力プログラム
US10620748B2 (en) * 2014-10-22 2020-04-14 Telefonaktiebolaget Lm Ericsson (Publ) Method and device for providing a touch-based user interface
JP6780315B2 (ja) * 2016-06-22 2020-11-04 カシオ計算機株式会社 投影装置、投影システム、投影方法及びプログラム
US11423879B2 (en) * 2017-07-18 2022-08-23 Disney Enterprises, Inc. Verbal cues for high-speed control of a voice-enabled device
US11080514B2 (en) * 2019-09-13 2021-08-03 Citrix Systems, Inc. Facial recognition based auto zoom
JP7446913B2 (ja) * 2020-05-21 2024-03-11 キヤノン株式会社 電子機器、電子機器の制御方法、プログラム

Family Cites Families (27)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH05204526A (ja) * 1992-01-23 1993-08-13 Nec Corp 視点監視入力装置
JPH1097376A (ja) * 1996-09-20 1998-04-14 Canon Inc 視線操作装置
US6088018A (en) 1998-06-11 2000-07-11 Intel Corporation Method of using video reflection in providing input data to a computer system
RU2316051C2 (ru) * 2005-01-12 2008-01-27 Самсунг Электроникс Ко., Лтд. Способ и система автоматической проверки присутствия лица живого человека в биометрических системах безопасности
JP2008118633A (ja) * 2006-10-11 2008-05-22 Victor Co Of Japan Ltd 電子機器の制御装置及び制御方法
US8726194B2 (en) * 2007-07-27 2014-05-13 Qualcomm Incorporated Item selection using enhanced control
US9002073B2 (en) * 2007-09-01 2015-04-07 Eyelock, Inc. Mobile identity platform
US20110273369A1 (en) 2010-05-10 2011-11-10 Canon Kabushiki Kaisha Adjustment of imaging property in view-dependent rendering
CN103347437B (zh) * 2011-02-09 2016-06-08 苹果公司 3d映射环境中的凝视检测
US8625847B2 (en) * 2011-03-21 2014-01-07 Blackberry Limited Login method based on direction of gaze
US8594374B1 (en) 2011-03-30 2013-11-26 Amazon Technologies, Inc. Secure device unlock with gaze calibration
KR101773845B1 (ko) 2011-05-16 2017-09-01 삼성전자주식회사 휴대용 단말기에서 입력 처리 방법 및 장치
US8885882B1 (en) * 2011-07-14 2014-11-11 The Research Foundation For The State University Of New York Real time eye tracking for human computer interaction
US9083532B2 (en) * 2012-03-06 2015-07-14 Ebay Inc. Physiological response PIN entry
JP5456817B2 (ja) * 2012-03-23 2014-04-02 ヤフー株式会社 表示制御装置、表示制御方法、情報表示システム及びプログラム
AU2013205535B2 (en) 2012-05-02 2018-03-15 Samsung Electronics Co., Ltd. Apparatus and method of controlling mobile terminal based on analysis of user's face
TWI590098B (zh) 2012-05-09 2017-07-01 劉鴻達 以臉部表情為輸入的控制系統
US9952663B2 (en) * 2012-05-10 2018-04-24 Umoove Services Ltd. Method for gesture-based operation control
US20130342672A1 (en) 2012-06-25 2013-12-26 Amazon Technologies, Inc. Using gaze determination with device input
PT106430B (pt) * 2012-07-03 2018-08-07 Cesar Augusto Dos Santos Silva Sistema para medição da distância interpupilar usando um dispositivo equipado com um ecrã e uma câmara
KR20140073730A (ko) * 2012-12-06 2014-06-17 엘지전자 주식회사 이동 단말기 및 이동 단말기 제어방법
US9274608B2 (en) * 2012-12-13 2016-03-01 Eyesight Mobile Technologies Ltd. Systems and methods for triggering actions based on touch-free gesture detection
KR102012254B1 (ko) * 2013-04-23 2019-08-21 한국전자통신연구원 이동 단말기를 이용한 사용자 응시점 추적 방법 및 그 장치
RU2522848C1 (ru) * 2013-05-14 2014-07-20 Федеральное государственное бюджетное учреждение "Национальный исследовательский центр "Курчатовский институт" Способ управления устройством с помощью глазных жестов в ответ на стимулы
GB2514603B (en) * 2013-05-30 2020-09-23 Tobii Ab Gaze-controlled user interface with multimodal input
TW201518979A (zh) * 2013-11-15 2015-05-16 Utechzone Co Ltd 手持式眼控接目裝置、及密碼輸入裝置、方法以及電腦可讀取記錄媒體及電腦程式產品
US20150227735A1 (en) * 2014-02-13 2015-08-13 Robert Chappell System and method for eye tracking authentication

Also Published As

Publication number Publication date
CN107077200A (zh) 2017-08-18
US20160034030A1 (en) 2016-02-04
AU2015296666A1 (en) 2017-02-09
EP3175325A1 (en) 2017-06-07
BR112017001326A2 (pt) 2017-11-14
JP2017526096A (ja) 2017-09-07
CA2955072C (en) 2022-02-22
CN107077200B (zh) 2020-01-21
EP3175325B1 (en) 2020-11-18
MX2017001514A (es) 2017-05-09
WO2016018957A1 (en) 2016-02-04
CA2955072A1 (en) 2016-02-04
AU2015296666B2 (en) 2020-04-30
JP6898234B2 (ja) 2021-07-07
US9645641B2 (en) 2017-05-09
RU2017103121A (ru) 2018-08-01
KR20170040222A (ko) 2017-04-12
RU2693308C2 (ru) 2019-07-02
KR102392437B1 (ko) 2022-04-28

Similar Documents

Publication Publication Date Title
BR112016027568A2 (ru)
BR112016022165A2 (ru)
BR112016020985A2 (ru)
BR112016025940A2 (ru)
BR112016013819A2 (ru)
BR112016020441A2 (ru)
BR112016018810A2 (ru)
BR112016030654A2 (ru)
BR112016017130A2 (ru)
RU2017103121A3 (ru)
BR112016022729A2 (ru)
BR112016018609A2 (ru)
BR112016009588A2 (ru)
BR112016025943A2 (ru)
CN302742118S (ru)
CN302738765S (ru)
CN302704777S (ru)
CN302917028S (ru)
CN302872983S (ru)
CN302708909S (ru)
CN302746860S (ru)
CN302746584S (ru)
CN302745881S (ru)
CN302743003S (ru)
CN302721986S (ru)