WO2011134106A1 - Procédé et dispositif d'identification d'entrées utilisateur - Google Patents

Procédé et dispositif d'identification d'entrées utilisateur Download PDF

Info

Publication number
WO2011134106A1
WO2011134106A1 PCT/CN2010/000593 CN2010000593W WO2011134106A1 WO 2011134106 A1 WO2011134106 A1 WO 2011134106A1 CN 2010000593 W CN2010000593 W CN 2010000593W WO 2011134106 A1 WO2011134106 A1 WO 2011134106A1
Authority
WO
WIPO (PCT)
Prior art keywords
user
user input
center
choices
input
Prior art date
Application number
PCT/CN2010/000593
Other languages
English (en)
Other versions
WO2011134106A8 (fr
Inventor
Peng Qin
Original Assignee
Thomson Licensing
Shangguan, Sinan
Du, Lin
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Thomson Licensing, Shangguan, Sinan, Du, Lin filed Critical Thomson Licensing
Priority to PCT/CN2010/000593 priority Critical patent/WO2011134106A1/fr
Publication of WO2011134106A1 publication Critical patent/WO2011134106A1/fr
Publication of WO2011134106A8 publication Critical patent/WO2011134106A8/fr

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/041Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
    • G06F3/042Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by opto-electronic means
    • G06F3/0425Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by opto-electronic means using a single imaging device like a video camera for tracking the absolute position of a single or a plurality of objects with respect to an imaged reference surface, e.g. video camera imaging a display or a projection screen, a table or a wall surface, on which a computer generated image is displayed or projected
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • G06F3/04883Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text

Definitions

  • the present invention relates to user interface, and more particularly, relating to a method and a device for identifying user inputs.
  • the user-machine interface which is also called user interface (Ul)
  • U user interface
  • the Ul system comprises three main components, i.e. processing unit, display unit and input unit.
  • the TV is used as display device for displaying the electronic program guide (EPG)
  • the STB is used as the processing device for outputting EPG to the TV and processing user's inputs
  • the STB remote is used as input device.
  • the user uses the STB remote to input instruction, e.g. channel up/down, volume up/down, viewing next channel information etc., with the help of EPG displayed on the TV.
  • a method for identifying user inputs in a system comprising the steps of obtaining a first user input; obtaining a second user input; and determining which choice among the set of the at least two explicit or implicit choices is chosen based on the position relationship between said first user input and said second user input.
  • a device for identifying user inputs comprises a communication module configured to received user inputs captured by at least one camera; and a processing module configured to, in response to a set of at least two explicit or implicit choices arranged in a first portion of display area of a display device, determine which choice is chosen based on position relationship between a first user input and a second user input, wherein said first user input and said second user input are obtained through said communication module.
  • the aspect of present invention provides an effective and efficient input method for a user without needing complex training each time the user changes his position.
  • Fig. 1 is a diagram illustrating a system for user inputs according to an embodiment of present invention
  • Fig. 2 is a diagram illustrating an example of a menu comprising 4 options according to the present embodiment
  • Fig. 3 is a diagram illustrating an example of a trajectory of the infrared emitter in a camera according to the present embodiment
  • Fig. 4A, 4B, 4C, 4D, 4E, 4F, 4G, 4H and 4I are diagrams illustrating examples of possible menus according to the present embodiment
  • Fig. 5 is a flow chart illustrating a user input method according to the present embodiment
  • Fig. 6A, 6B, 6C and 6D are screen snapshots illustrating an example of steps of getting gravity center from raw data when a user presses button of the infrared emitter in front of the cameras according to the present embodiment
  • Fig. 7 is a diagram illustrating an example of coordinates system, in which the center of the first input and the occurrence position of the second input are shown, according to the present embodiment.
  • the present invention aims to provide an input method for users, and make user feel more like operating an object in realistic world, for example, pressing a particular button among several buttons displayed on a display device, unlike the traditional operation in prior art that the user uses a remote to move selection pointer to a visual button on screen that he wants to execute and then presses 'OK' button on the remote.
  • the method of the present invention uses two consecutive inputs of a user for generating a computer-comprehensible instruction, wherein the latter input is recognized and converted to the instruction based on the former input, so that once a user moves his position relative to the display device, the accuracy of input is increased without need him to perform the complex training/recalibration process.
  • the system 100 comprises a display device 101 , two cameras 102, 103 mounted on the display device 101 , a processing device 104 and an infrared emitter (not shown).
  • the display device 101 is a PC display or a TV
  • the processing device is a PC main unit or a STB
  • the two cameras are mounted on the left-top and right-top of the TV, separately.
  • some components may be integrated into one device according to actual implementation, for example, the cameras 102, 103, the display device 101 and the processing device 104 are integrated into a single device.
  • the two cameras can be mounted on other place other than the left-top and right-top of the TV, e.g. the left-bottom and right-bottom of the TV, or on the desk supporting the TV.
  • the display device 101 is used to display information/prompt to user.
  • the displayed information/prompt may relate to a later user instruction, the later user instruction is derived from user's two inputs by using infrared emitter.
  • the displayed information/prompt is a menu comprising 4 options for user to choose, and the Fig. 2 shows one possible implementation.
  • the two cameras 102, 103 are used to receive and recognize the input of the infrared emitter.
  • the user holds the infrared emitter, with button being pressed, to stroke from left to right, and each camera will record the trajectory of the infrared emitter as shown in the Fig. 3.
  • light filters can be attached to the surface of the lens of the cameras.
  • the infrared emitter is used for emitting infrared light when a button thereon is pressed. When the button is kept pressed, the infrared light is kept on. So the user can use the infrared emitter to make gesture inputs with the button pressed.
  • the gesture may comprises press, push, pull, stroke up, stroke down, stroke left, stroke right, stroke circle, stroke arc etc.
  • horizontality as X-axis
  • Y-axis verticality
  • Z-axis is perpendicular to both X-axis and Y-axis.
  • Push Z-axis values have a big change, for example relative over
  • Pull Z-axis values have a big change, for example relative over
  • Stroke Up Not push or pull gesture and Y-axis values have a bigger change than X-axis. The values are relative decreased.
  • Stroke Down Not push or pull gesture and Y-axis values have a bigger change than X-axis.
  • the values are relative increased.
  • Stroke Left Not push or pull gesture and X-axis values have a bigger change than Y-axis. The values are relative decreased.
  • Stroke Right Not push or pull gesture and X-axis values have a bigger change than X-axis. The values are relative increased. Stroke Circle Not push or pull gesture. If we set one virtual coordinates base on the center of this gesture's image, the trace will cross with the four virtual axes.
  • Stroke Arc Not push or pull gesture. If we set on virtual coordinates system base on the center of this gesture's image, the trace will cross with three continuous virtual axes.
  • the processing device 104 comprises functions of
  • each menu displayed on the display device 101 may comprise two or more buttons/ choices/ options for user to choose;
  • Fig. 5 is a flow chart illustrating a user input method according to the present embodiment.
  • a menu having two or more buttons is displayed on the display device 101.
  • Two or more buttons are arranged in such a way that each button is put into a position around the center of the display area for the menu and the positions of all buttons are distinct from each other.
  • the display are for the menu is the whole screen of the display device 101 , and consequently, the center of the display are for the menu is the center of the screen of display device 101 , but in a few cases, for example, as shown in the Fig. 4I, the display area for menu is a rectangle area on the right-top of the TV screen.
  • the reason for making that "positions of all buttons are distinct from each other" is to make it easy and accurate for user to make gesture input in the air.
  • Fig. 4A, 4B, 4C, 4D, 4E, 4F, 4G, 4H and 41 shows some examples of such arrangement that make positions of all buttons distinct from each other.
  • the shapes of the buttons do no matter and the shape can be circle, rectangle, triangle, square etc.
  • the distinct positions comprise left-top, left-center, left-bottom, bottom-center, right-bottom, right-center, right-top and top-center relative to the center point of the display area for the menu.
  • a default menu is set either by the manufacture or by the user, and the user knows the locations/ positions of all buttons of the default menu, thus he can perform the input without needing the menu being displayed on the display device 101.
  • a first user input is obtained.
  • the user holds the infrared emitter, with button being pressed, to make a gesture of stroking from left to right.
  • the purpose of the first input is to help both the user and the processing device 104 to obtain information about a center.
  • the center is used for assisting the user to make a second input and assisting the processing device 104 to determine which button the user intends to press.
  • center of the first user input is defined as the reference center, with reference to which the second input is made. For the user, he needs to estimate the center of his first input. As for straight line, circle, arc line and dot input, it's easy to estimate a rough center.
  • the center of his first input After he estimates the center of his first input, he is able to make reference to the estimated center and make a second gesture (e.g. a press action) in the spatial space corresponding to a button he wants to press. Still taking the menu in the Fig. 3 as example, assuming the option 1 is what the user wants to press, the user just needs to make the second gesture input in a spatial space locating left-top to the estimated center in the vertical plain.
  • the center of the first input is roughly estimated, because the buttons/ options are arranged distinct in location from each other, it's hard for the user to make mistaken/ wrong input. And consequentially, the input accuracy can be guaranteed.
  • the center of the first input is calculated by average value of all points that are collected by the two cameras.
  • the coordinates of the center of the first input is obtained as (Xcr, Ycr).
  • the start point of one gesture is where user press the button of infrared emitter
  • the stop point of one gesture is where user releases the button of infrared emitter.
  • camera will collect the raw image of this infrared emitter's output infrared light.
  • gravity center to represent every infrared point. And store these gravity centers to buffer.
  • the center can be calculated as below:
  • a second user input is obtained following the obtaining of the first input.
  • a restriction on time can be added between the obtaining of the first input and the second input, i.e. if the second input is not obtained within a predefined time period, e.g. 5 second, after obtaining the first input, the second input is deemed invalid.
  • the second input is a press action or a pull action.
  • second input can also be a short stroke, a small circle, and long stroke and big circle are also possible. We use its center for the determination step 504.
  • the processing device 104 determines which button is chosen based on the first input and the second input. To be specific, in this example, it is determined based on the position relationship between the center of the first input and the occurrence position of the second input.
  • a menu having four buttons as shown in the Fig. 2 is used. Therefore, we only need to determine the position relationship is which one in a group containing left-top, right-top, left-bottom and right-bottom. It can be easily realized in coordinates system. Assuming the camera's coordinates origin (0, 0) is the left-top corner as shown in the Fig. 7. After the first input gesture, we get the center (X0, Y0).
  • the processing device 104 needs to determine the position relationship is which one in group containing left-top, center-top, right-top, right-center, right-bottom, center-bottom, left- bottom and left-center. It can be realized by averagely dividing the area into 9 blocks, wherein the center of the area and the center of the first input are overlapped.
  • the processing device 104 determines and executes the corresponding instruction based on mapping information between the buttons of the menu and computer comprehensible instructions.
  • the method of present invention is used for selecting a portion of an image, wherein, the image is divided into several selectable portions, and the division of image into portions can be referenced to the arrangement of the buttons as described above.
  • the gesture type of first input is fixed to a particular one, e.g. a left-to-right straight line, or a particular group of ones.
  • the merit of usage of fixed gesture type is that it makes coexist of input method of present invention and traditional gesture input method possible, because the fixed gesture type differentiated from other being-used gesture is carefully selected to indicate the input method of present invention will be used.
  • a computer comprehensible instruction is generated by two consecutive gesture inputs, i.e. a former input that is used to obtain the reference center and a latter input that is used to locate which button is chosen.
  • it doesn't necessarily need the input of the former input.
  • the user wants the processing device to consecutively execute two or more instruction while he is at a same place, e.g. sitting on a sofa. He can make several gesture inputs, the first one of which is used as the former input and the succeeding ones are used as latter inputs. For example, when the user wants the computer to execute 4 instructions, he just needs to make 5 gesture inputs.
  • the processing device uses the center of previous gesture input as the reference center when a user makes a gesture input.
  • the menu is caused to be displayed upon a user's gesture input, for example, a pull gesture. We can store the occurrence position of the pull gesture input as the reference center during the life-span of this menu. Therefore, during the lifespan of the menu, only one gesture input is enough to cause the processing device to generate an instruction for execution.
  • a prompt that the second user input is expected to be inputted is displayed on the screen of the display device after the step 502. And in one example, the prompt may be displayed in top-center of the screen.
  • the infrared emitter it's not limited to the infrared emitter. It can be extended to only use gesture recognition method. According to a variant of present embodiment, it's not limited to the stereo cameras. It can be extended to only use one camera which have built-in deep sensor.
  • the principle of the present invention is: use the relative positions of press gesture and other gesture's center to determine which event is triggered. It is appreciated that a person skilled in art can contemplate other variants or implementation after reading the description, and the variants and implementations shall falls in the scope of the principle of the present invention.

Landscapes

  • Engineering & Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • User Interface Of Digital Computer (AREA)
  • Position Input By Displaying (AREA)

Abstract

L'invention concerne un procédé permettant d'identifier des entrées utilisateur dans un système, ledit système comprenant un dispositif d'affichage et au moins une caméra pour capturer des entrées utilisateur. En réponse à un ensemble d'au moins deux choix explicites ou implicites disposé dans une première partie d'une zone d'affichage dudit dispositif d'affichage, le procédé comprend les étapes consistant à obtenir une première entrée utilisateur ; obtenir une seconde entrée utilisateur ; et déterminer quel choix parmi l'ensemble des deux choix explicites ou implicites ou plus est sélectionné d'après la relation de position entre ladite première entrée utilisateur et ladite seconde entrée utilisateur.
PCT/CN2010/000593 2010-04-29 2010-04-29 Procédé et dispositif d'identification d'entrées utilisateur WO2011134106A1 (fr)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/CN2010/000593 WO2011134106A1 (fr) 2010-04-29 2010-04-29 Procédé et dispositif d'identification d'entrées utilisateur

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2010/000593 WO2011134106A1 (fr) 2010-04-29 2010-04-29 Procédé et dispositif d'identification d'entrées utilisateur

Publications (2)

Publication Number Publication Date
WO2011134106A1 true WO2011134106A1 (fr) 2011-11-03
WO2011134106A8 WO2011134106A8 (fr) 2012-01-19

Family

ID=44860729

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2010/000593 WO2011134106A1 (fr) 2010-04-29 2010-04-29 Procédé et dispositif d'identification d'entrées utilisateur

Country Status (1)

Country Link
WO (1) WO2011134106A1 (fr)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH0981362A (ja) * 1995-09-12 1997-03-28 Casio Comput Co Ltd データ入力装置
JP2006139615A (ja) * 2004-11-12 2006-06-01 Access Co Ltd 表示装置、メニュー表示プログラムおよびタブ表示プログラム
US20090289904A1 (en) * 2008-05-20 2009-11-26 Tae Jin Park Electronic device with touch device and method of executing functions thereof

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH0981362A (ja) * 1995-09-12 1997-03-28 Casio Comput Co Ltd データ入力装置
JP2006139615A (ja) * 2004-11-12 2006-06-01 Access Co Ltd 表示装置、メニュー表示プログラムおよびタブ表示プログラム
US20090289904A1 (en) * 2008-05-20 2009-11-26 Tae Jin Park Electronic device with touch device and method of executing functions thereof

Also Published As

Publication number Publication date
WO2011134106A8 (fr) 2012-01-19

Similar Documents

Publication Publication Date Title
US9329714B2 (en) Input device, input assistance method, and program
JP6390799B2 (ja) 入力装置、入力方法、及びプログラム
WO2019033957A1 (fr) Procédé et système de détermination de position d'interaction, support de stockage et terminal intelligent
WO2011142317A1 (fr) Dispositif de reconnaissance de gestes, procédé, programme et support lisible par ordinateur sur lequel le programme est stocké
EP2040156A2 (fr) Traitement de l'image
JP6344530B2 (ja) 入力装置、入力方法、及びプログラム
JP6062416B2 (ja) 情報入力装置及び情報表示方法
US9804667B2 (en) Electronic apparatus
JP2013137613A (ja) 情報処理プログラム、情報処理装置、情報処理システム、および、情報処理方法
EP2816456A1 (fr) Dispositif de traitement d'informations, procédé de traitement d'informations et programme d'ordinateur
US20140132725A1 (en) Electronic device and method for determining depth of 3d object image in a 3d environment image
US11199946B2 (en) Information processing apparatus, control method, and program
US9400575B1 (en) Finger detection for element selection
JP2012008808A (ja) 情報処理装置
KR101807516B1 (ko) 모션 인식을 통한 디지털 기기 제어 장치 및 방법
US20150052477A1 (en) Enlargement and reduction of data with a stylus
KR101321274B1 (ko) 두대의 카메라와 광원을 이용한 포인터를 사용하지 않는 가상 터치 장치
US20160054860A1 (en) Input device
WO2021004413A1 (fr) Dispositif d'entrée portatif et procédé et appareil de commande d'extinction pour icône d'indication d'un dispositif d'entrée portatif
EP3088991B1 (fr) Dispositif portable et procédé permettant une interaction utilisateur
WO2011134106A1 (fr) Procédé et dispositif d'identification d'entrées utilisateur
JP5080409B2 (ja) 情報端末装置
JP6686319B2 (ja) 画像投影装置及び画像表示システム
JP2018097443A (ja) 入力システム、入力プログラム
KR101272458B1 (ko) 포인터를 사용하지 않는 가상 터치 장치 및 방법

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 10850439

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 10850439

Country of ref document: EP

Kind code of ref document: A1