WO2018018624A1 - 一种可穿戴设备的手势输入方法及可穿戴设备 - Google Patents

一种可穿戴设备的手势输入方法及可穿戴设备 Download PDF

Info

Publication number
WO2018018624A1
WO2018018624A1 PCT/CN2016/092399 CN2016092399W WO2018018624A1 WO 2018018624 A1 WO2018018624 A1 WO 2018018624A1 CN 2016092399 W CN2016092399 W CN 2016092399W WO 2018018624 A1 WO2018018624 A1 WO 2018018624A1
Authority
WO
WIPO (PCT)
Prior art keywords
area
gesture
wearable device
recognition
track
Prior art date
Application number
PCT/CN2016/092399
Other languages
English (en)
French (fr)
Inventor
杨波
金庆浩
张泽狮
贺真
卢睿
Original Assignee
华为技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 华为技术有限公司 filed Critical 华为技术有限公司
Priority to US16/321,714 priority Critical patent/US20210117078A1/en
Priority to EP16910204.3A priority patent/EP3486747A4/en
Priority to CN201680049793.3A priority patent/CN108027648A/zh
Priority to PCT/CN2016/092399 priority patent/WO2018018624A1/zh
Publication of WO2018018624A1 publication Critical patent/WO2018018624A1/zh

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • G06F3/04883Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/0304Detection arrangements using opto-electronic means
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F1/00Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
    • G06F1/16Constructional details or arrangements
    • G06F1/1613Constructional details or arrangements for portable computers
    • G06F1/163Wearable computers, e.g. on a belt
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/017Gesture based interaction, e.g. based on a set of recognized hand gestures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/033Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
    • G06F3/0346Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of the device orientation or free movement in a 3D space, e.g. 3D mice, 6-DOF [six degrees of freedom] pointers using gyroscopes, accelerometers or tilt-sensors

Definitions

  • the present invention relates to the field of terminal technologies, and in particular, to a gesture input method and a wearable device of a wearable device.
  • the input of information generally adopts the following methods: 1. By setting common phrases And the emoticon icon, when the information is replied, directly click on the list to realize the input of the information; 2. Through the voice input, the voice recognition is input into the text and text information through the built-in voice system; 3. The wearable device is made as large as possible With a larger operating area, you can design a larger touch screen or more buttons.
  • mode 1 the expression is limited and the utility is low; for mode 2, the use environment is limited; for mode 3, the wearable device is large in size, resulting in insufficient wearing comfort, and is limited to the wearable mode, and can only be limited. The size is increased so that the operating area is still small.
  • the embodiment of the invention provides a gesture input method for a wearable device and a wearable device to solve the problem that the existing wearable device uses the common language reply expression limitation, the voice input use environment is limited, and the device size is large but the operation area is still small. The problem.
  • the first aspect of the present invention provides a gesture input method for a wearable device.
  • an area may be selected as a track recognition area in the scan area of the sensor, and the The track recognition area consists of two parts, one part of which is used to actually identify the motion track.
  • a gesture recognition area and a gesture sensing area disposed on the gesture recognition area, and then collecting a motion track of the user gesture motion in the track recognition area by the sensor, and finally, according to the area information of the track recognition area, the area information may be the coordinates of the track recognition area The information is used to identify the motion track collected in the gesture recognition area as the input information.
  • the trajectory recognition area is set by the sensor, and the position and size of the trajectory identification area are related to the scanning area of the sensor, and the trajectory recognition area is scanned by the sensor continuously, thereby obtaining gesture recognition in the trajectory recognition area.
  • the motion trajectory of the user's gesture motion in the area, and then the motion trajectory is identified based on the coordinate system information of the trajectory recognition area, that is, the position of the motion trajectory in the coordinate system is calculated, and then the motion trajectory in the coordinate system is calculated.
  • the position is identified as the input information.
  • the size can be set by itself, so that the size of the operation area of the input information can also be defined by itself, thereby solving the existing
  • the technical operation area has limited problems, and due to the gesture input method, the input accuracy and the applicable environment range are greatly improved.
  • a display screen may be set on the wearable device, and the motion track in the track recognition area is synchronously mapped to On the display screen, the position in the track recognition area is set corresponding to the position in the display screen, so that the gesture movement is more intuitive, and the information input is realized by the gesture movement.
  • the sensor may only recognize the motion trajectory of the gesture recognition area, and it is not necessary to identify the motion trajectory in the entire trajectory recognition area. At this time, when identifying the input information, first according to the trajectory identification area The area information determines the area information of the gesture recognition area, and then the motion track is identified according to the area information of the gesture recognition area. It can be seen that only the motion track in the gesture recognition area needs to be directly recognized, and Improve the efficiency of gesture recognition.
  • the information identifying the motion track as the input by two-dimensional recognition may first determine the recognition reference surface according to the region information of the gesture recognition area, and the identification The reference plane is used as a reference plane for final two-dimensional recognition, and then the motion trajectory in the gesture recognition area is projected onto the recognition reference plane, and finally the content in the recognition reference plane is recognized to achieve the recognition effect on the input information.
  • the gesture recognition area generally considers the three-dimensional space of the motion trajectory, that is, it is impossible to input only on one plane, it is designed as a three-dimensional space.
  • the identification reference plane is set to cooperate with the recognition of the motion trajectory, and the efficiency of the trajectory recognition is improved.
  • the gesture sensing area can also recognize the motion track, but the motion track in the gesture sensing area does not need to be recognized, and only needs to display the cursor on the corresponding position of the display screen based on the position of the motion track. can. Can give users a more friendly input prompt.
  • the gesture recognition area and the gesture sensing area are in contact, and both areas are located above the wearable device, so that the motion track passes through the gesture sensing area and reaches the gesture recognition area, and the gesture sensing area is equivalent to The writing area of the writing operation, and the gesture recognition area is equivalent to the writing area of the writing operation, so that the setting of the gesture recognition area and the gesture sensing area can simulate the real pen area and the pen writing area, since there is no need to adapt to the new one. Input habits can improve the input efficiency of the gesture input method.
  • the gesture recognition area and the gesture sensing area may also be located not only above the wearable device, but also diagonally above or to the side of the wearable device as long as it is within the scanning range of the sensor. This method can enhance the scalability of the gesture input method.
  • the gesture recognition area may be set as an input mode in addition to being disposed above the wearable device, and the gesture recognition area may be set as an identification surface. Specifically, when the recognition is performed, the gesture may be identified according to the gesture.
  • the area information of the area identifies the motion trajectory within the first recognition plane as the input information, and the first recognition surface may be a plane or a curved surface.
  • the first recognition surface may coincide with the surface of the object, so that the user can perform the writing operation on the surface of the object as the input information, and the achievability of the gesture input method can be improved, and the setting of the plane or the surface can satisfy different
  • the shape of the object surface enhances the scalability of the gesture input method.
  • the surface of the object may move relative to the wearable device, thereby causing an angle between the first recognition surface and the surface of the object, thereby affecting the gesture input, and a preset recognition surface may be set at this time.
  • the preset recognition surface is an identification surface when the first recognition surface coincides with the surface of the object, that is, an optimal recognition surface, and when the relative movement between the wearable device and the object occurs, the first recognition surface and the preset recognition surface An angle is generated between the two. When the angle exceeds the preset angle threshold, the effect of the gesture input is deteriorated, and a reminder is issued to prompt that the first recognition surface needs to be adjusted.
  • the wearable device is a wrist wearable device
  • the first identification surface is located at the Wear the wearer's hand on the back of the device. Since the setting of the back of the hand is sufficient as the writing area, the first identification surface is directly disposed on the back of the hand, which is convenient for the user to perform gesture input.
  • the senor may be a sensor group consisting of more than one sensor, and the type of sensor may be an ultrasonic sensor or an infrared sensor. It can be understood that the combination of multiple sensors can improve the accuracy of the trajectory acquisition, thereby improving the recognition efficiency, and the ultrasonic sensor or the infrared sensor can be selected according to actual conditions.
  • a second aspect of the embodiments of the present invention further provides a wearable device, where the wearable device includes two parts: a sensor and an identification module, wherein the sensor is configured to collect a motion track of a user gesture motion in the track recognition area, and the track recognition area An area of the scan area that is preset in the sensor, the track recognition area includes a gesture recognition area and a gesture sensing area disposed above the gesture recognition area; and the identification module is configured to identify the area according to the track
  • the area information identifies the motion trajectory in the gesture recognition area as input information, and the area information includes coordinate system information of the trajectory identification area.
  • the track recognition area is set based on the scan area of the sensor, the size thereof can be set by itself, so that the size of the operation area of the input information can also be defined by itself, which solves the limited operation area of the prior art.
  • the wearable device since the sensor scanning area is invisible, in order to facilitate the user to perform gesture motion in an accurate area, the wearable device further includes a display screen for moving the track of the track recognition area.
  • a synchronization map is mapped into the display screen, the track identification area corresponding to an area within the display screen.
  • the sensor may only recognize the motion trajectory of the gesture recognition area, and it is not necessary to identify the motion trajectory in the entire trajectory recognition area. At this time, the sensor actually collects the motion in the gesture recognition area in the trajectory recognition area. a trajectory; the identification module first determines the area information of the gesture recognition area according to the area information of the trajectory identification area; and then identifies the motion trajectory in the gesture recognition area as an input according to the area information of the gesture recognition area. information. Since the identified area becomes smaller, the recognition efficiency can be improved.
  • both are identified by two-dimensional, such as characters and texts.
  • Word or the like so the actual recognition process of the recognition module may be: first, determining the recognition reference plane according to the region information of the gesture recognition area; secondly, projecting the motion trajectory in the gesture recognition area to the recognition reference plane; The motion trajectory projected into the recognition reference plane is identified as input information. Since the gesture recognition area generally considers the three-dimensional spatiality of the motion trajectory, that is, it is impossible to input only on one plane, it is designed as a three-dimensional space to accommodate the three-dimensional motion trajectory, and when the motion trajectory is actually recognized, It also needs to be converted into two-dimensional, so this recognition reference plane is set to match the recognition of the motion trajectory, and the efficiency of the trajectory recognition is improved.
  • the gesture sensing area can also recognize the motion track, but the motion track in the gesture sensing area does not need to be recognized, and only needs to display the cursor on the corresponding position of the display screen based on the position of the motion track.
  • the sensor is further configured to detect a motion track in the gesture sensing area;
  • the display screen is further configured to display a cursor in a current position of the motion track corresponding to the gesture sensing area in the display screen. . In this way, the user can be given a more friendly input prompt.
  • the gesture recognition area and the gesture sensing area are in contact, and both areas are located above the wearable device, so that the motion track passes through the gesture sensing area and reaches the gesture recognition area, and the gesture sensing area is equivalent to The writing area of the writing operation, and the gesture recognition area is equivalent to the writing area of the writing operation, so that the setting of the gesture recognition area and the gesture sensing area can simulate the real pen area and the pen writing area, since there is no need to adapt to the new one. Input habits can improve the input efficiency of the gesture input method.
  • the gesture recognition area may be set as an input mode in addition to being disposed above the wearable device, and the gesture recognition area may be an identification surface according to the gesture recognition area.
  • the first identification surface is established by the area information.
  • the identification module is configured to: identify the motion track in the first recognition surface as the input information according to the area information of the gesture recognition area, where the first recognition surface is a plane Or a surface.
  • the first recognition surface may coincide with the surface of the object, so that the user can perform the writing operation on the surface of the object as the input information, and the achievability of the gesture input method can be improved, and the setting of the plane or the surface can satisfy different
  • the shape of the object surface enhances the scalability of the gesture input method.
  • the wearable device further includes a prompting module, configured to issue a first recognition surface adjustment prompt when determining that an angle between the first identification surface and the preset recognition surface exceeds a preset angle threshold.
  • a prompting module configured to issue a first recognition surface adjustment prompt when determining that an angle between the first identification surface and the preset recognition surface exceeds a preset angle threshold.
  • the wearable device is a wrist wearable device with a first identification surface located on the back of the wearer's hand of the wearable device. Since the setting of the back of the hand is sufficient as the writing area, the first identification surface is directly disposed on the back of the hand, which is convenient for the user to perform gesture input.
  • the senor may be a sensor group consisting of more than one sensor, and the type of sensor may be an ultrasonic sensor or an infrared sensor. It can be understood that the combination of multiple sensors can improve the accuracy of the trajectory acquisition, thereby improving the recognition efficiency, and the ultrasonic sensor or the infrared sensor can be selected according to actual conditions.
  • a third aspect of the embodiments of the present invention further provides a wearable device, including: a sensor, a processor, and a memory, wherein the sensor is configured to collect a motion track of a user gesture motion, where the memory is used to store a processor to perform Instructions and data, the processor for executing the instructions in the memory to implement information identifying motion trajectories within the gesture recognition area as input.
  • a wearable device including: a sensor, a processor, and a memory, wherein the sensor is configured to collect a motion track of a user gesture motion, where the memory is used to store a processor to perform Instructions and data, the processor for executing the instructions in the memory to implement information identifying motion trajectories within the gesture recognition area as input.
  • FIG. 1 is a schematic diagram of linear distance collection in a gesture input method of a wearable device according to an embodiment of the present invention
  • FIG. 2 is a schematic diagram of plane distance collection in a gesture input method of a wearable device according to an embodiment of the present invention
  • FIG. 3 is a schematic diagram of spatial distance collection in a gesture input method of a wearable device according to an embodiment of the present invention
  • FIG. 4 is a diagram of an embodiment of a gesture input method of a wearable device according to an embodiment of the present invention.
  • FIG. 5 is a diagram of another embodiment of a gesture input method of a wearable device according to an embodiment of the present invention.
  • FIG. 6 is a diagram of another embodiment of a gesture input method of a wearable device according to an embodiment of the present invention.
  • FIG. 7 is a diagram of another embodiment of a gesture input method of a wearable device according to an embodiment of the present invention.
  • FIG. 8 is a diagram of another embodiment of a gesture input method of a wearable device according to an embodiment of the present invention.
  • FIG. 9 is a diagram of another embodiment of a gesture input method of a wearable device according to an embodiment of the present invention.
  • FIG. 10 is a diagram of another embodiment of a gesture input method of a wearable device according to an embodiment of the present invention.
  • FIG. 11 is a diagram of another embodiment of a gesture input method of a wearable device according to an embodiment of the present invention.
  • FIG. 12 is a diagram of another embodiment of a gesture input method of a wearable device according to an embodiment of the present invention.
  • FIG. 13 is a diagram showing an embodiment of a gesture input method of a wearable device according to an embodiment of the present invention.
  • Figure 14 is a diagram showing an embodiment of a wearable device according to an embodiment of the present invention.
  • Figure 15 is a diagram showing an embodiment of a wearable device in accordance with an embodiment of the present invention.
  • the embodiment of the invention provides a gesture input method for a wearable device and a wearable device to solve the problem that the existing wearable device uses the common language reply expression limitation, the voice input use environment is limited, and the device size is large but the operation area is still small. The problem. .
  • the input of information of the wearable device is an important reason hindering the development of the wearable device. Due to the inconvenience of input, the wearable device is still a tool for assisting other portable terminals, such as a secondary mobile phone, a secondary tablet, and in fact, The performance of wearable devices has been able to replace these portable terminals in some functions, such as sending text messages instead of mobile phones, and performing some text input instead of tablets.
  • Some of the current input methods for example, quick reply of commonly used content, are only fixed within a preset Content, the editing of the content needs to be edited through the terminal communicatively connected with the wearable device, but these are generally simple common words, can only be used as a quick reply, the expression is too restrictive, and the practicality is low;
  • the voice input method is adopted, but due to the characteristics of the voice input, the noisy environment is inferior in use, the public environment takes into consideration the need of privacy, is not easy to use, has a large limitation, and further, the wearable device is made larger.
  • a large dial is set on a smart watch, but due to the limitations of the characteristics of the watch product, even if it is as large as possible, it is still relatively small compared to a terminal device such as a mobile phone, and cannot be equipped with a sufficiently large touch screen or enough. Press the button to enter the content.
  • the embodiment of the present invention recognizes the input information by means of the motion track of the sensor scanning gesture motion, which is to add a sensor group capable of scanning the motion track on the basis of the existing wearable device, and
  • the trajectory capture in the space is taken as the input trajectory of the user, and the trajectories are identified to obtain corresponding input characters.
  • the sensor group in the embodiment of the present invention mainly adopts an ultrasonic sensor of an ultrasonic technology or an infrared sensor of an infrared technology or a sensor group composed of both, by setting a measurement reference point and basic coordinate information, and then capturing a gesture at the coordinate. Tracking of the trajectory can be achieved by the motion trajectory in the area. The principle of collecting trajectories in one-dimensional, two-dimensional and three-dimensional situations is explained below.
  • FIG. 1 is a schematic diagram of linear distance acquisition in a gesture input method of a wearable device according to an embodiment of the present invention.
  • the sensor group in FIG. 1 may only include the sensor A, that is, at least A sensor can realize the distance collection in the one-dimensional case, and the infrared light or the ultrasonic wave is emitted through the sensor A, and the time of returning the infrared light or the ultrasonic wave after encountering the obstacle is measured, and the sensor A to the measured object can be calculated according to the return time.
  • the distance between the sensor A and the measured object at two different t1 and t2 times is recorded, and the distance that the measured object moves in the linear direction during the t2-t1 period can be calculated.
  • FIG. 2 is a schematic diagram of plane distance acquisition in the gesture input method of the wearable device according to the embodiment of the present invention, under the premise that the X-axis and the Y-axis are determined in the plane, Adding a sensor B to the sensor group based on Figure 1, that is, using at least two passes The sensor can realize the acquisition of the distance in the two-dimensional situation, and the distance L between the sensor A and the sensor B is known.
  • the sensor A and the sensor B respectively measure the relationship with the object to be measured.
  • the distances L1 and L2 and measure the distance between the sensor A and the sensor B and the measured object at different times t1 and t2, so that the L11 and L21 data at time t1 and the L12 and L22 data at time t2 combine the distance L
  • the offset of the measured object in the X-axis and Y-axis directions on the plane can be calculated, and the motion of the measured object in the plane can be determined by continuous measurement.
  • FIG. 3 is a schematic diagram of spatial distance acquisition in the gesture input method of the wearable device according to the embodiment of the present invention, which is determined by X-axis, Y-axis and Z-axis in space.
  • a sensor C is added to the sensor group on the basis of FIG. 2, that is, the distance in the three-dimensional situation can be collected by using at least three sensors, and the sensor A, the sensor B and the sensor C are not collinear, and the sensor The distance between A and sensor B, between sensor A and sensor C, and between sensor B and sensor C is known.
  • the distance between sensor A, sensor B and sensor C is L1, L2 and L3, respectively.
  • the fixed distance can be used to calculate the offset of the measured object in the X-axis, Y-axis and Z-axis directions in the three-dimensional space from t1 to t2.
  • the continuous measurement can determine the motion track of the measured object in three-dimensional space. trace.
  • the sensor A may be disposed at the intersection of the X/Z plane and the X/Y plane, the sensor B is disposed in the X/Y plane, and the sensor C is disposed in the X/Z plane.
  • the offset of the measured object in the X-axis and Y-axis directions from t1 to t2 can be measured.
  • the sensor A and sensor C can be measured during the period from t1 to t2.
  • the offset of the object in the X-axis and Z-axis directions thereby obtaining the offset of the measured object on the X-axis, Y-axis, and Z-axis.
  • FIG. 4 is a diagram of an embodiment of a gesture input method of a wearable device according to an embodiment of the present invention.
  • the method may include:
  • the track recognition area is an area disposed in a scan area preset by the sensor, and the track recognition area includes a gesture recognition area and a gesture sensing area disposed above the gesture recognition area.
  • the number of sensors varies according to the minimum number of the foregoing three situations.
  • the preset scanning area is first set according to the situation, and further, due to the irregularity of the area that the sensor can collect, For the accuracy of the acquisition of different regions, the region with relatively high acquisition accuracy can be selected as the trajectory recognition region in the scanning area of the sensor for trajectory recognition.
  • setting the track recognition area to include a gesture recognition area and a gesture sensing area disposed above the gesture recognition area can simulate a pen and a pen down operation in a real writing situation, wherein the gesture sensing area can simulate a pen operation area, The gesture recognition area can simulate a pen down writing operation area.
  • the senor may be a sensor group composed of a plurality of sensors, and the sensor in the sensor group may mainly adopt an ultrasonic sensor of an ultrasonic technology or an infrared sensor of an infrared technology or a sensor group composed of both, by measuring infrared
  • the distance of the movement of the measured object in the time when the infrared light or the ultrasonic wave returns from the time of emission to the object to be measured is negligible with respect to the moving distance during the movement of the object to be measured.
  • the gesture input mode can adopt the writing mode, which is determined by its characteristics, and there are two kinds of pen-writing operations and pen-down operations, and only the part of the pen-down operation is the part that actually needs to be recognized, so the sensor
  • the track recognition area needs to divide two different areas for the two situations, that is, the area corresponding to the pen operation and the area corresponding to the pen down operation, and the area of the pen down operation is called the gesture recognition area, and all the areas in the gesture recognition area
  • the motion trajectory is used as the basis for the information that is subsequently recognized as input; and the gesture sensing area can be used as the pen operation area.
  • the gesture sensing area and the gesture recognition area can also be combined to implement a mouse-like input operation, such as the operation of the fingertip entering the gesture sensing area as a left click operation of the mouse, continuously in a preset short time. Entering the gesture sensing area twice as a left-click double-click operation, the operation of entering the gesture recognition area is the right-click operation of the mouse, and the fingertip maintaining the preset time in the gesture sensing area is a long-press operation or an outgoing menu operation, and more At the same time, the fingertips enter the gesture sensing area as a multi-touch operation. It can be seen that since two regions are set, multiple types of input operations can be performed by a reasonable definition of the two regions.
  • the input operation may also be for a touch screen, that is, a touch operation such as a click touch, a double tap touch, a long press touch, or a multi touch input.
  • FIG. 13 is a diagram showing an embodiment of a gesture input method of a wearable device according to an embodiment of the present invention.
  • the user displayed on the display screen of the wearable device in FIG. 13 is input on the back of the user's hand as the track recognition area.
  • Information, and the content in the display screen corresponds to the content of the track recognition area.
  • the display is on the display
  • the cursor corresponding to the position is displayed. If the current moment of the motion track is located in the gesture recognition area, in addition to displaying the cursor on the corresponding position on the display screen, the recognized writing track is also displayed on the display screen, as shown in the figure. Show, the dotted line indicates the motion track in the gesture sensing area, and the thick solid line indicates the motion track in the gesture recognition area, the display of the wearable device
  • the screen is also mapped out of the trajectory of the gesture recognition area, in addition, the position of the finger gesture sensing area corresponding to the position of the cursor within the display screen.
  • the wearable device can be given feedback for operation through changes in the state of the buoy graphic such as a cursor on the display screen, for example, using different sizes, different transparency, different degrees of falsehood.
  • the buoy pattern of the same gives operation feedback, of course, sound or vibration can also be used as the operation feedback to distinguish the pen and the pen down action, assist the user operation, and enhance the user's input experience.
  • the entire track recognition area may be used as the gesture recognition area, and the non-track recognition area is used as the area for the pen operation.
  • the area in which the pen operation is performed is also the track recognition area, as long as the motion track is still in the area, the input is not yet completed.
  • the delay time since there is no track recognition area as the sensing condition, a feasible way is to increase the delay time.
  • the purpose of the gesture sensing area setting is to display a cursor in the area of the wearable device corresponding to the current position of the motion track in the gesture sensing area as long as the motion track is detected in the area.
  • the gesture sensing area can also be used as the pen operation area, that is, during the writing process, the writing process is not interrupted as long as the gesture sensing area is not left. .
  • the gesture recognition area and the gesture sensing area may be in contact with each other, and since the track identification area is determined by the scanning area of the sensor, it may be disposed in any orientation of the wearable device, such as a wearable device.
  • the position of the track recognition area can be set according to actual needs, and in addition, for different wearable devices, the track recognition area
  • the size can also be set to meet the needs of different users for the size of the writing area.
  • the wearable device is a wristband or a watch
  • the writing area thereof can be directly disposed on the back of the hand of the hand wearing the wristband or the watch
  • the sensor on the wristband or the watch first covers the scanning area to the back of the hand
  • the scanning area pre-defines an area similar to the back surface of the hand through the coordinate system.
  • the back of the user's hand can also be scanned in real time, and a gesture recognition area corresponding to the back of the hand is determined from the scanning area by the scanned back data of the hand.
  • the area can be spatially coincident with the back of the hand, and then a gesture sensing area having a shape similar to the gesture recognition area can be disposed on the gesture recognition area.
  • the gesture recognition area and the gesture sensing area may also be set according to the surface of the object with reference to the above manner.
  • the gesture recognition area may be disposed at a certain distance in front of the user, and the height of the area may be set in the upper body of the user for the convenience of the user.
  • FIG. 5 is a diagram of another embodiment of a gesture input method of a wearable device according to an embodiment of the present invention
  • FIG. 6 is a schematic diagram of a gesture input method of a wearable device according to an embodiment of the present invention.
  • FIG. 7 is another embodiment of a gesture input method of a wearable device according to an embodiment of the present invention
  • FIG. 8 is another embodiment of a gesture input method of the wearable device according to an embodiment of the present invention.
  • the light-colored area portion of the track recognition area 1 is the gesture sensing area 11
  • the dark area portion is the gesture recognition area 12
  • the gesture sensing area 11 and the gesture recognition area 12 are in contact.
  • the gesture recognition area 12 and the gesture sensing area 11 are wearable in FIGS. 5 and 6.
  • the cursor is displayed on the display screen 21 of the wearable device 2.
  • the corresponding position of the user's finger in the gesture sensing area 11 is displayed, and when the user's finger enters the gesture recognition area 12, not only the corresponding position of the user's finger in the gesture sensing area 11 but also the corresponding position of the user's finger in the gesture sensing area 11 is displayed.
  • the process of the user's current writing is displayed.
  • FIG. 9 is a diagram of another embodiment of a gesture input method of a wearable device according to an embodiment of the present invention, wherein a track recognition area may be located above a side of the wearable device, and the track recognition area may be simultaneously Set multiple locations so that users can operate smoothly in each application scenario.
  • step 401 is specifically:
  • a motion trajectory in the gesture recognition area within the trajectory recognition area is acquired by the sensor.
  • the trajectory collected by the sensor is mainly the trajectory in the gesture recognition area, rather than the trajectory of the entire trajectory recognition area, which can reduce a part of the calculation amount, thereby improving the performance of the trajectory recognition.
  • the area information includes coordinate system information of the track identification area.
  • the area information may include the coordinate system information of the trajectory identification area, and the gesture recognition area It is an area set inside the track recognition area. After the coordinate system information of the track recognition area is determined, the coordinate system information of the gesture recognition area can be determined, so that the direction and offset of the motion track are in the coordinate system. The basis is calculated.
  • step 402 may specifically be:
  • the following may specifically include the following content:
  • the recognition reference plane can be used as the paper surface in the paper writing process, and is also the surface written by the user during the writing process. Since the gesture recognition area is a three-dimensional space, after obtaining the motion track in the space, The motion trajectory needs to be processed to obtain the trajectory on the recognition reference plane, that is, the motion trajectory that is finally required.
  • the motion trajectory projected into the recognition reference plane can be identified, thereby identifying the desired input information through the motion trajectory.
  • the track identification area may be implemented as follows.
  • the track identification area includes a first identification surface that is established according to the coordinate system, and the information that identifies the motion track as the input according to the area information of the track identification area includes:
  • the motion trajectory in the first recognition surface is identified as input information according to the region information of the first recognition surface, and the first recognition surface is a plane or a curved surface.
  • the portion of the trajectory recognition area that is actually recognized is not a region but a surface, that is, a first recognition surface, and the first recognition surface may be a plane or a curved surface, and the method is different from the foregoing division gesture.
  • a suspended gesture input mode of the sensing area and the gesture recognition area which can project the surface actually used for recognition onto the surface of an object, thereby obtaining a more realistic writing experience.
  • the first identification surface may be disposed on the surface of some objects.
  • the wearable device is a smart watch or a smart bracelet
  • the first identification surface when the first identification surface is set on the back of the user's hand
  • the first recognition surface serves as a reference surface for writing, and when the distance between the pen tip and the first recognition surface is greater than 0 mm, it is regarded as a pen operation, when the pen tip and the When the distance of an identification surface is less than or equal to 0 mm, it is indicated as a pen down operation. Since the back of the human hand is a flexible surface, the distance between the nib and the first recognition surface is less than 0 mm.
  • a predetermined margin can also be reserved, such as when the nib and the When the distance of an identification surface is greater than 2 mm, it is regarded as a pen operation, and when the distance between the pen tip and the first recognition surface is less than or equal to 2 mm, it is a pen down operation, and the margin can be set by itself according to the shape of the surface of the object. And setting the size of the first recognition surface.
  • the track identification area may further be provided with a second identification area, and the first identification area of the second identification area is in contact with each other.
  • the function of the second recognition area is similar to the function of the gesture sensing area in the foregoing case.
  • it can be used as a pen area, on the other hand, the position of the current gesture input in the input area can be located, and the reaction is displayed on the wearable device. The position of the cursor on the screen.
  • a preset identification surface may be disposed. Since the wearable device uses an input operation during the wearing process, the object may inevitably have a certain angle with respect to the wearable device. Moving, the size of the first recognition surface projected on the object will change. If the movement amplitude is too large, the identification of the input trajectory may be inaccurate, so a preset recognition surface is set, when the angle of the first recognition surface is After the angle of the recognition surface exceeds a certain value, an adjustment prompt is issued to ensure that the first recognition surface can perform an input operation in a high-precision input environment.
  • the calculation of the angle between the first identification surface and the preset recognition surface can be performed in various ways. One of the following is introduced. For one object surface, the movement of different angles of the wearable device leads to the first The size of the area projected by the recognition surface on the surface of the object changes, and the angle between the angle of the current first recognition surface and the fixed preset identification surface can be calculated according to the degree of the size change.
  • FIG. 10 is a schematic diagram of another embodiment of a gesture input method of a wearable device according to an embodiment of the present invention
  • FIG. 11 is a schematic diagram of a gesture input method of a wearable device according to an embodiment of the present invention
  • FIG. 12 is a diagram showing another embodiment of a gesture input method of a wearable device according to an embodiment of the present invention.
  • the wearable device 3 is a wrist wearable device
  • the first identification surface 41 of the track recognition area 4 is located on the back of the wearer of the wearable device 3;
  • the track recognition area 4 further includes a second identification area 42 that is in contact with the first identification surface 41, in which case the user operates in the second identification area 42.
  • the trajectory is mapped to the display screen of the wearable device 3, the current position of the trajectory is mapped to the position on the display screen as the cursor 31, and further, as long as the back of the hand is not contacted in the second identification area 42, It is regarded as a pen-drawing operation, and only touching the back of the hand, that is, equivalent to contacting the first identifying surface 41, is regarded as a pen-down operation;
  • FIG. 12 shows an illustration of the angle between the aforementioned identification reference surface 43 and the first identification surface 41 and the identification reference surface 43.
  • the identification reference plane may be an X/Y plane perpendicular to the Z axis, as long as If the angle between the first identification surface 41 and the X/Y plane is within a preset range, the prompt of the adjustment of the first identification surface 41 is not issued in the correct area, and if the first identification surface 41 and the X/Y plane are When the angle exceeds the preset range, a prompt for adjusting the first recognition surface 41 is issued, indicating that the user needs to adjust the back of the opponent to achieve the best writing effect.
  • FIG. 14 is a diagram of an embodiment of the wearable device according to the embodiment of the present invention.
  • the wearable device 14 includes:
  • the sensor 1401 is configured to collect a motion track of a user gesture motion in the track recognition area, where the track recognition area is an area disposed in a scan area preset by the sensor, where the track recognition area includes a gesture recognition area and is disposed at the A gesture sensing area above the gesture recognition area.
  • the sensor 1401 can implement step 401 in the embodiment shown in FIG.
  • the senor 1401 may be a sensor group composed of a plurality of sensors, and the type of the sensor may be an ultrasonic sensor or an infrared sensor.
  • the specific plurality of sensors 1401 constitute a sensor group similar to the description of step 401 in the embodiment shown in FIG. , will not repeat them here.
  • the gesture recognition area and the gesture sensing area are set to correspond to the writing and writing operations of the actual writing process. For details, refer to the description of the step 401 in the embodiment shown in FIG. 4, and details are not described herein again.
  • the gesture recognition area and the gesture sensing area are similar to the description of step 401 in the embodiment shown in FIG. 4 except that the gesture recognition area and the gesture detection area are implemented, and an input operation similar to a mouse or a touch screen can also be implemented.
  • the identification module 1402 is configured to: the gesture recognition area according to the area information of the track identification area
  • the motion track within is identified as input information, and the area information includes coordinate system information of the track recognition area.
  • the working process of the wearable device in the embodiment of the present invention may be that the sensor 1401 first collects the motion track of the user's gesture motion in the track recognition area, especially in the gesture recognition area, and then recognizes the motion track by the recognition module 1402.
  • the area is the area information of the track recognition area and the motion track in the gesture recognition area recognizes the motion track as input information, and completes the process of information input.
  • the wearable device further includes:
  • the display screen 1403 is configured to synchronously map the motion track of the track recognition area into the display screen 1403, and the track identification area corresponds to an area in the display screen.
  • the number of display screens 1403 is mainly used to display the cursor in the gesture sensing area, and the cursor reflects the user gesture motion in the gesture sensing area.
  • the senor 1401 is specifically configured to: collect motion trajectories in the gesture recognition area in the trajectory recognition area;
  • the identification module 1402 is specifically configured to: determine area information of the gesture recognition area according to the area information of the track recognition area; and identify the motion track in the gesture recognition area as an input according to the area information of the gesture recognition area. Information.
  • the information for determining the final input is the motion track in the gesture recognition area, so the sensor 1401 can only collect the gesture recognition.
  • the motion track in the area is sufficient, and the identification module 1402 determines the area information of the gesture recognition area, that is, the coordinate system information of the gesture recognition area, based on the area information of the track recognition area, and passes the The coordinate system information identifies the motion trajectory in the gesture recognition area, and finally obtains the input information.
  • the identifying module 1402 is specifically configured to:
  • a motion trajectory projected into the recognition reference plane is identified as input information.
  • the identification reference surface is first determined, and the identification reference surface can be used as the paper surface in the paper writing process, and is also the default writing surface of the user during the writing process, due to gesture recognition.
  • the area is a three-dimensional space. After obtaining the motion trajectory in the space, the motion trajectory needs to be processed to obtain the trajectory on the recognition reference plane, that is, the final required motion trajectory. After determining the recognition reference plane, the motion trajectory in the gesture recognition area is projected back to the recognition reference plane to obtain a projection trajectory on the recognition reference plane, and finally the projection trajectory is identified to obtain a final input. Information.
  • the gesture recognition area and the gesture sensing area set by the area scanned by the sensor 1401 in the embodiment of the present invention may be a dangling setting, or the gesture recognition area may be a face on the object. The following are explained separately:
  • the gesture recognition area and the gesture sensing area may be suspended.
  • the gesture recognition area is in contact with the gesture sensing area, and the gesture recognition area and the gesture sensing area are located above the wearable device.
  • This setting method does not need to consider the medium carrying the gesture recognition area and the gesture sensing area, and can be applied to various occasions.
  • the gesture recognition area is located on the object.
  • the gesture recognition area is a first identification surface that is established according to the area information of the gesture recognition area, and the identification module 1402 is specifically configured to:
  • the motion track in the first recognition surface is identified as input information according to the area information of the gesture recognition area, and the first recognition surface is a plane or a curved surface.
  • the gesture recognition area may be an identification surface disposed on the surface of the object, the recognition surface substantially coincides with the surface of the object, so that the user can perform an input operation by writing on the surface of the object, which can greatly
  • the first identification surface is similar to the description of the step 402 in the embodiment shown in FIG. 4, and the information description for the first identification surface can be referred to the step 402 in the embodiment shown in FIG.
  • the wearable device is an example of writing a smart watch or a smart bracelet, and details are not described herein again.
  • the preset identification surface may be corrected, and optionally, the wearable The device 14 may further include a prompting module 1404 for issuing a first recognition surface adjustment prompt when it is determined that the angle between the first identification surface and the preset identification surface exceeds a preset angle threshold.
  • the wearable device uses an input operation during the wearing process, the object inevitably moves at a certain angle with respect to the wearable device, and the size of the first recognition surface projected on the object changes. If the moving range is too large, the identification of the input track may be inaccurate. Therefore, a preset identification surface is set. When the angle between the angle of the first recognition surface and the preset recognition surface exceeds a certain value, an adjustment prompt is issued, thereby ensuring the first
  • the recognition surface is capable of performing input operations in a high-precision input environment.
  • FIG. 15 is a diagram of an embodiment of a wearable device according to an embodiment of the present invention, wherein the wearable device 15 may include at least one connected to the bus.
  • a processor 1501, at least one sensor 1502, and at least one memory 1503, the wearable device according to the embodiment of the present invention may have more or less components than those shown in FIG. 15, and two or more components may be combined. Or, there may be different component configurations or arrangements, each component being implemented in hardware, software, or a combination of hardware and software, including one or more signal processing and/or application specific integrated circuits.
  • the processor 1501 can implement the functions of the identification module 1402 and the prompting module 1404 in the embodiment shown in FIG. 14, and the sensor 1502 can implement the embodiment shown in FIG.
  • the function of the sensor 1401 is used to store instructions and data to be executed by the processor 1501, and the processor 1501 is configured to execute the instructions in the memory to implement information identifying the motion track in the gesture recognition area as input.
  • the wearable device may further include a display screen 1504 that can implement the function of the display screen 1403 in FIG.
  • the disclosed system, apparatus, and method may be implemented in other manners.
  • the device embodiments described above are merely illustrative.
  • the division of the unit is only a logical function division.
  • there may be another division manner for example, multiple units or components may be combined or Can be integrated into another system, or Some features can be ignored or not executed.
  • the mutual coupling or direct coupling or communication connection shown or discussed may be an indirect coupling or communication connection through some interface, device or unit, and may be in an electrical, mechanical or other form.
  • the units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, may be located in one place, or may be distributed to multiple network units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution of the embodiment.
  • each functional unit in each embodiment of the present invention may be integrated into one processing unit, or each unit may exist physically separately, or two or more units may be integrated into one unit.
  • the above integrated unit can be implemented in the form of hardware or in the form of a software functional unit.
  • the integrated unit if implemented in the form of a software functional unit and sold or used as a standalone product, may be stored in a computer readable storage medium.
  • the technical solution of the present invention which is essential or contributes to the prior art, or all or part of the technical solution, may be embodied in the form of a software product stored in a storage medium.
  • a number of instructions are included to cause a computer device (which may be a personal computer, server, or network device, etc.) to perform all or part of the steps of the methods described in various embodiments of the present invention.
  • the foregoing storage medium includes: a U disk, a mobile hard disk, a read-only memory (ROM), a random access memory (RAM), a magnetic disk, or an optical disk, and the like. .

Abstract

本发明涉及终端技术领域,具体涉及一种可穿戴设备的手势输入方法及可穿戴设备。该收拾输入方法包括,通过传感器采集轨迹识别区内用户手势运动的运动轨迹,轨迹识别区为设置于传感器预设的扫描区域内的区域,轨迹识别区包括手势识别区和设于手势识别区之上的手势感应区;根据轨迹识别区的区域信息将手势识别区内的运动轨迹识别为输入的信息,区域信息包括轨迹识别区的坐标系信息。在该轨迹识别区内的手势识别区内的用户手势运动的运动轨迹,计算该运动轨迹在该坐标系中的位置,而后将坐标系中运动轨迹的位置识别为输入的信息,解决了现有技术操作区域有限问题,并且由于采用手势输入的方式,输入准确度和适用环境范围会大大提高。

Description

一种可穿戴设备的手势输入方法及可穿戴设备 技术领域
本发明涉及终端技术领域,具体涉及一种可穿戴设备的手势输入方法及可穿戴设备。
背景技术
随着系统级芯片(System on Chip,SOC)的高速发展,其体积越来越来越小的同时却具有越来越高的性能,并且能够同时集成多种功能,由此带动了智能的可穿戴设备的发展,一些可穿戴设备甚至集成了一些移动操作系统,如安卓操作系统和IOS操作系统,在一些情况下能够取代手机或者平板电脑的功能。
目前的可穿戴设备受限于结构大小的限制,操作区域有限,既无法配备较大的触摸屏,也无法配备较多的按键,导致信息的输入一般采用如下几种方式,1、通过设置常用短语和表情图标,在信息回复时直接从列表中点选实现信息的输入;2、通过语音输入,通过内置的语音系统将语音识别成文文字信息输入;3、将可穿戴设备尽量做大,以适配更大操作区域,从而能够设计更大的触摸屏或者更多的按键。
然而,对于方式1,表达受限,实用性低;对于方式2,使用环境受限;对于方式3,可穿戴设备尺寸大,导致佩戴舒适性不足,并且受限于可穿戴方式,仅能有限的增大尺寸,从而操作区域仍然偏小。
发明内容
本发明实施例提供了一种可穿戴设备的手势输入方法及可穿戴设备来解决现有可穿戴设备采用常用语回复表达受限、语音输入使用环境受限以及设备尺寸大但是操作区域依然较小的问题。
有鉴于此,本发明第一方面提供一种可穿戴设备的手势输入方法,该方法中,由于可穿戴设备内设有传感器,可在传感器的扫描区域内选取一区域作为轨迹识别区,且该轨迹识别区包括两个部分,一部分作为实际识别运动轨迹的 手势识别区和设置在手势识别区上的手势感应区,而后通过传感器采集此轨迹识别区内用户手势运动的运动轨迹,最后根据轨迹识别区的区域信息,该区域信息可以是轨迹识别区的坐标系信息,将手势识别区内采集的运动轨迹识别为输入的信息。
可以看出,依靠传感器设置轨迹识别区,该轨迹识别区的位置及大小与传感器的扫描区域相关,通过传感器不停的对轨迹识别区进行扫描,从而得出在该轨迹识别区内的手势识别区内的用户手势运动的运动轨迹,而后以该轨迹识别区的坐标系信息为基础对该运动轨迹进行识别,即计算该运动轨迹在该坐标系中的位置,而后将坐标系中运动轨迹的位置识别为输入的信息,由于该轨迹识别区是以传感器的扫描区域为基础设定的,因此其大小可以自行设定,从而使得输入信息的操作区域的大小也可自行定义,解决了现有技术操作区域有限问题,并且由于采用手势输入的方式,输入准确度和适用环境范围会大大提高。
在一些实施例中,由于传感器扫描区域是不可见的,为了更便于用户在准确的区域内进行手势运动,可在可穿戴设备上设置显示屏,且将轨迹识别区内的运动轨迹同步映射至该显示屏上,且将轨迹识别区内的位置与所述显示屏内的位置设置为相对应,从而使得手势运动更为直观,更有利于通过手势运动实现信息的输入。
在一些实施例中,传感器可以仅仅识别手势识别区的运动轨迹,而不必要对整个轨迹识别区内的运动轨迹进行识别,此时,在识别输入的信息时,首先根据所述轨迹识别区的区域信息确定所述手势识别区的区域信息,而后根据所述手势识别区的区域信息对此运动轨迹进行识别即可,可以看出,仅需要直接识别手势识别区内的运动轨迹即可,能够提高手势识别的效率。
在一些实施例中,对于信息的输入来说,都是通过二维识别,如字符和文字等,将运动轨迹识别为输入的信息可以首先根据手势识别区的区域信息确定识别基准面,该识别基准面作为最终进行二维识别的参考平面,之后再将手势识别区内的运动轨迹投影至该识别基准面,最后再识别该识别基准面内的内容达到对输入的信息的识别的效果。由于手势识别区一般而言考虑到运动轨迹的三维空间性,即不可能仅在一个平面上输入,因此会设计成一个三维的空间, 以容纳三维的运动轨迹,而在实际识别运动轨迹时,又需要转换为二维,因此设置此识别基准面用以配合运动轨迹的识别,提高轨迹识别的效率。
在一些实施例中,手势感应区也可以对运动轨迹进行识别,不过在手势感应区的运动轨迹无需进行识别,仅需要以该运动轨迹的位置为基础在显示屏的对应位置上显示出光标即可。能够给予用户更为友好的输入提示。
在一些实施例中,手势识别区和手势感应区是相接触的,且两区域都位于可穿戴设备的上方,因此运动轨迹在穿过手势感应区就会抵达手势识别区,手势感应区相当于书写操作的提笔区域,而手势识别区相当于书写操作的落笔书写区域,从而手势识别区和手势感应区相接触的设置能够模拟真实的提笔区域和落笔书写区域,由于无需重新适应新的输入习惯,能够提高手势输入方法的输入效率。
在一些实施例中,手势识别区和手势感应区也可以不仅在可穿戴设备的上方,而可以位于可穿戴设备的斜上方或侧方,只要是在传感器的扫描范围之内即可。此方式能够增强该手势输入方法的可扩展性。
在一些实施例中,手势识别区除了设置在可穿戴设备的上方作为悬空的输入方式之外,还可将该手势识别区设置为识别面,具体的,在识别时,会根据所述手势识别区的区域信息将在第一识别面内的运动轨迹识别为输入的信息,且该第一识别面可以是平面或者曲面。该第一识别面可以与物体的表面的重合,从而使得用户在物体表面上进行书写操作能够被识别为输入的信息,能够提高手势输入方法的可实现性,此外平面或曲面的设置能够满足不同形状的物体表面,增强手势输入方法的可扩展性。
在一些实施例中,物体表面相对于可穿戴设备之间会发生运动,从而会导致第一识别面与物体表面之间产生夹角,从而会影响手势输入,此时可设置一预设识别面,该预设识别面为第一识别面与物体表面重合时的识别面,即为最佳识别面,当可穿戴设备与物体之间发生相对运动时,第一识别面会与该预设识别面之间产生夹角,当此夹角超过预设的角度阈值时,手势输入的效果会变差,此时会发出提醒,以提示需要调整第一识别面。
在一些实施例中,可穿戴设备为手腕可穿戴设备,第一识别面位于所述可 穿戴设备的穿戴者的手背上。此设置由于手背作为书写区域大小足够,直接将第一识别面设置在手背上,能够便于用户进行手势输入。
在一些实施例中,传感器可以是一个以上的传感器组成的传感器组,且传感器的类型可以是超声波传感器或者红外传感器。可以理解,多个传感器组合的方式能够提高轨迹采集精确度,从而提升识别效率,超声波传感器或者红外传感器可以视实际情况进行选择。
本发明实施例第二方面还提供一种可穿戴设备,该可穿戴设备包括传感器和识别模块两个部分,其中,传感器用于采集轨迹识别区内用户手势运动的运动轨迹,所述轨迹识别区为设置于所述传感器预设的扫描区域内的区域,所述轨迹识别区包括手势识别区和设于所述手势识别区之上的手势感应区;而识别模块用于根据所述轨迹识别区的区域信息将所述手势识别区内的运动轨迹识别为输入的信息,所述区域信息包括所述轨迹识别区的坐标系信息。
可以理解,由于该轨迹识别区是以传感器的扫描区域为基础设定的,因此其大小可以自行设定,从而使得输入信息的操作区域的大小也可自行定义,解决了现有技术操作区域有限问题,并且由于采用手势输入的方式,输入准确度和适用环境范围会大大提高。
在一些实施例中,由于传感器扫描区域是不可见的,为了更便于用户在准确的区域内进行手势运动,所述可穿戴设备还包括,显示屏,用于将所述轨迹识别区的运动轨迹同步映射至所述显示屏内,所述轨迹识别区与所述显示屏内的区域相对应。从而使得手势运动更为直观,更有利于通过手势运动实现信息的输入。
在一些实施例中,传感器可以仅仅识别手势识别区的运动轨迹,而不必要对整个轨迹识别区内的运动轨迹进行识别,此时传感器实际采集的是轨迹识别区内的手势识别区内的运动轨迹;此时识别模块首先根据所述轨迹识别区的区域信息确定所述手势识别区的区域信息;而后根据所述手势识别区的区域信息将所述手势识别区内的运动轨迹识别为输入的信息。由于识别的区域变小,能够提高识别效率。
在一些实施例中,对于信息的输入来说,都是通过二维识别,如字符和文 字等,因此识别模块实际的识别过程可以是,首先根据所述手势识别区的区域信息确定识别基准面;其次,将所述手势识别区内的运动轨迹投影至所述识别基准面;最后将投影至所述识别基准面内的运动轨迹识别为输入的信息。由于手势识别区一般而言考虑到运动轨迹的三维空间性,即不可能仅在一个平面上输入,因此会设计成一个三维的空间,以容纳三维的运动轨迹,而在实际识别运动轨迹时,又需要转换为二维,因此设置此识别基准面用以配合运动轨迹的识别,提高轨迹识别的效率。
在一些实施例中,手势感应区也可以对运动轨迹进行识别,不过在手势感应区的运动轨迹无需进行识别,仅需要以该运动轨迹的位置为基础在显示屏的对应位置上显示出光标即可,此时,传感器还具体用于检测所述手势感应区内的运动轨迹;所述显示屏还用于在所述显示屏内对应所述手势感应区内的运动轨迹的当前位置显示出光标。通过此方式能够给予用户更为友好的输入提示。
在一些实施例中,手势识别区和手势感应区是相接触的,且两区域都位于可穿戴设备的上方,因此运动轨迹在穿过手势感应区就会抵达手势识别区,手势感应区相当于书写操作的提笔区域,而手势识别区相当于书写操作的落笔书写区域,从而手势识别区和手势感应区相接触的设置能够模拟真实的提笔区域和落笔书写区域,由于无需重新适应新的输入习惯,能够提高手势输入方法的输入效率。
在一些实施例中,手势识别区除了设置在可穿戴设备的上方作为悬空的输入方式之外,还可将该手势识别区设置为识别面,所述手势识别区为根据所述手势识别区的区域信息建立的第一识别面,此时识别模块具体用于:根据所述手势识别区的区域信息将在第一识别面内的运动轨迹识别为输入的信息,所述第一识别面为平面或者曲面。该第一识别面可以与物体的表面的重合,从而使得用户在物体表面上进行书写操作能够被识别为输入的信息,能够提高手势输入方法的可实现性,此外平面或曲面的设置能够满足不同形状的物体表面,增强手势输入方法的可扩展性。
在一些实施例中,物体表面相对于可穿戴设备之间会发生运动,从而会导致第一识别面与物体表面之间产生夹角,从而会影响手势输入,为解决此问题, 可穿戴设备还包括提示模块,用于当确定所述第一识别面与预设识别面的夹角超出预设的角度阈值时,发出第一识别面调整提示。当可穿戴设备与物体之间发生相对运动时,第一识别面会与该预设识别面之间产生夹角,当此夹角超过预设的角度阈值时,手势输入的效果会变差,此时会发出提醒,以提示需要调整第一识别面。
在一些实施例中,可穿戴设备为手腕可穿戴设备,第一识别面位于所述可穿戴设备的穿戴者的手背上。此设置由于手背作为书写区域大小足够,直接将第一识别面设置在手背上,能够便于用户进行手势输入。
在一些实施例中,传感器可以是一个以上的传感器组成的传感器组,且传感器的类型可以是超声波传感器或者红外传感器。可以理解,多个传感器组合的方式能够提高轨迹采集精确度,从而提升识别效率,超声波传感器或者红外传感器可以视实际情况进行选择。
本发明实施例第三方面还提供一种可穿戴设备,该可穿戴设备包括传感器、处理器和内存所述传感器用于采集用户手势运动的运动轨迹,所述内存用于存储处理器要执行的指令和数据,所述处理器用于执行所述内存中的指令以实现将手势识别区内的运动轨迹识别为输入的信息。
附图说明
图1是本发明实施例的可穿戴设备的手势输入方法中的直线距离采集示意图;
图2是本发明实施例的可穿戴设备的手势输入方法中的平面距离采集示意图;
图3是本发明实施例的可穿戴设备的手势输入方法中的空间距离采集示意图;
图4是本发明实施例的可穿戴设备的手势输入方法的一个实施例图;
图5为本发明实施例的可穿戴设备的手势输入方法的另一个实施例图;
图6为本发明实施例的可穿戴设备的手势输入方法的另一个实施例图;
图7为本发明实施例的可穿戴设备的手势输入方法的另一个实施例图;
图8为本发明实施例的可穿戴设备的手势输入方法的另一个实施例图;
图9为本发明实施例的可穿戴设备的手势输入方法的另一个实施例图;
图10为本发明实施例的可穿戴设备的手势输入方法的另一个实施例图;
图11为本发明实施例的可穿戴设备的手势输入方法的另一个实施例图;
图12为本发明实施例的可穿戴设备的手势输入方法的另一个实施例图;
图13是本发明实施例的可穿戴设备的手势输入方法的一个实施例图;
图14是本发明实施例的可穿戴设备的一个实施例图;
图15是本发明实施例的可穿戴设备的一个实施例图。
具体实施方式
本发明实施例提供了一种可穿戴设备的手势输入方法及可穿戴设备来解决现有可穿戴设备采用常用语回复表达受限、语音输入使用环境受限以及设备尺寸大但是操作区域依然较小的问题。。
为了使本技术领域的人员更好地理解本发明方案,下面将结合本发明实施例中的附图,对本发明实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例仅仅是本发明一部分的实施例,而不是全部的实施例。基于本发明中的实施例,本领域普通技术人员在没有做出创造性劳动前提下所获得的所有其他实施例,都应当属于本发明保护的范围。
以下分别进行详细说明。
本发明的说明书和权利要求书及上述附图中的术语“第一”、“第二”、“第三”“第四”等(如果存在)是用于区别类似的对象,而不必用于描述特定的顺序或先后次序。应该理解这样使用的数据在适当情况下可以互换,以便这里描述的实施例能够以除了在这里图示或描述的内容以外的顺序实施。
可穿戴设备的信息的输入是阻碍可穿戴设备发展的一个重要原因,由于输入不便使得可穿戴设备目前仍是一种辅助其他便携终端的工具,例如辅助手机,辅助平板电脑,而事实上,该可穿戴设备的性能在一些功能上已经能够取代这些便携终端,例如代替手机发送短信,代替平板电脑进行一些文字输入。
目前的一些输入方式例如,常用内容快速回复,仅仅是预先设定的固定内 容,内容的编辑上需要通过与该可穿戴设备通信连接的终端进行编辑,但是这些一般都是较为简单的常用语,仅能作为快速回复使用,表达限制性太大,实用性较低;又例如采用语音输入的方式,但是由于语音输入的特性,嘈杂的环境使用效果较差,公共环境考虑到隐私性的需要,也不便于使用,局限性大,此外,将可穿戴设备做的较大,例如智能手表上设置一个较大的表盘,但是由于手表产品特性的限制,即便尽可能做大,相对于手机等终端类设备,依旧是较小,也无法配备足够大的触摸屏或者配备足够多按键进行内容输入。
上述问题均是由于可穿戴设备本身设计空间局促所造成的,因此如何能在较为局促的空间的前提下设计一套符合这种情形的信息输入方式对于可穿戴设备来说,是尤为重要的。
未解决上述问题,本发明实施例采取传感器扫描手势运动的运动轨迹的方式识别出输入的信息,此方式在于在现有的可穿戴设备的基础上添加能够扫描运动轨迹的传感器组,通过对预设空间内的轨迹捕捉作为用户的输入轨迹,对这些轨迹进行识别从而得到相应的输入字符。
本发明实施例中的传感器组主要采用超声波技术的超声波传感器或者红外技术的红外传感器或者二者均采用而组成的传感器组,通过设置测量基准点及基本的坐标信息,而后在捕捉手势在该坐标区域内的运动轨迹,即可实现轨迹的捕捉。下面分别对一维、二维和三维情形下采集轨迹的原理进行说明。
对于情形,即一维线性情形,请参阅图1,图1是本发明实施例的可穿戴设备的手势输入方法中的直线距离采集示意图,图1中传感器组可仅包括传感器A,即最少采用一个传感器即可实现一维情形中的距离的采集,通过传感器A发射出红外光线或者超声波,测量该红外光线或者超声波遇到障碍后返回的时间,根据返回时间可计算出传感器A到被测物体之间的距离,通过记录并得到两次不同的t1时刻和t2时刻传感器A与被测物体之间的距离,从而可以计算出被测物体在t2-t1时间段内沿直线方向移动的距离。
对于情形二,即二维平面情形,请参阅图2,图2是本发明实施例的可穿戴设备的手势输入方法中的平面距离采集示意图,在平面内X轴和Y轴确定的前提下,在图1的基础上传感器组中增加一个传感器B,即最少采用两个传 感器即可实现二维情形中的距离的采集,并且传感器A和传感器B之间的距离L已知,在测量被测物体时,由传感器A和传感器B分别测量与被测物体之间的距离L1和L2,并测量在不同的t1时刻和t2时刻传感器A和传感器B与被测物体之间的距离,从而通过t1时刻的L11和L21数据和t2时刻的L12和L22数据结合距离L即可计算出被测物体在平面上的X轴和Y轴方向上的偏移量,通过连续测量即可确定出被测物体在该平面内的运动轨迹。
对于情形散三,即三维空间情形,请参阅图3,图3是本发明实施例的可穿戴设备的手势输入方法中的空间距离采集示意图,在空间内X轴、Y轴和Z轴确定的前提下,在图2的基础上传感器组中增加一个传感器C,即最少采用三个传感器即可实现三维情形中的距离的采集,并且传感器A、传感器B和传感器C三者不共线,传感器A与传感器B之间、传感器A与传感器C以及传感器B与传感器C之间的距离是已知的,传感器A、传感器B和传感器C之间的距离分别为L1、L2和L3,测量在不同的t1时刻和t2时刻传感器A、传感器B和传感器C与被测物体之间的距离,从而通过t1时刻的L11、L21和L31数据和t2时刻的L12、L22和L32数据结合各传感器之间的固定距离即可计算出被测物体在t1至t2时间段内在三维空间中X轴、Y轴和Z轴方向上的偏移量,连续测量即可确定出被测物体在三维空间中的运动轨迹。
需要说明的是,为了便于计算,可将传感器A设置于X/Z平面和X/Y平面的交界处,传感器B设置于X/Y平面内,传感器C设置于X/Z平面内,此时通过传感器A和传感器B即可测得t1到t2时间段内被测物体在X轴和Y轴方向上的偏移量,通过传感器A和传感器C即可测得t1到t2时间段内被测物体在X轴和Z轴方向上的偏移量,从而得到被测物体在X轴、Y轴和Z轴上的偏移量。
具体的,请参阅图4,图4是本发明实施例的可穿戴设备的手势输入方法的一个实施例图,该方法可包括:
401、通过传感器采集轨迹识别区内用户手势运动的运动轨迹。
其中,所述轨迹识别区为设置于所述传感器预设的扫描区域内的区域,所述轨迹识别区包括手势识别区和设于所述手势识别区之上的手势感应区。
可以理解的是,传感器的数量根据前述三种情形需要的最少数量有所不同,在测量时首先会根据情形对预设的扫描区域进行设置,此外,由于传感器能够采集的区域的不规则性以及对不同区域的采集的精准度的不同,可在传感器的扫描区域中选取采集精准度相对较高的区域作为轨迹识别区进行轨迹识别。此外,将轨迹识别区设置为包括手势识别区和设于所述手势识别区之上的手势感应区能够模拟真实书写情况中的提笔和落笔操作,其中手势感应区可以模拟提笔操作区域,手势识别区可以模拟落笔书写操作区域。
可选的,该传感器可以是多个传感器构成的传感器组,而传感器组中的传感器可主要采用超声波技术的超声波传感器或者红外技术的红外传感器或者二者均采用而组成的传感器组,通过测算红外光线或者超声波在发出后碰到障碍返回的时间计算传感器到障碍物,即被测物体的之间的距离,由于红外光线和超声波的传输速度很快,远高于被测物体的运动速度,因此相对于被测物体的运动过程中的运动距离,在红外光线或者超声波从发射到碰到被测物体返回的时间内的被测物体运动的距离可以忽略不计。
需要说明的是,手势输入方式可以采用的是书写的方式,由其特性决定,会有提笔操作和落笔操作两种,而只有落笔操作的部分才是实际需要进行识别的部分,因此传感器的轨迹识别区需要为两种情形划分出两种不同的区域,即对应提笔操作的区域和对应落笔操作的区域,而落笔操作的区域称之为手势识别区,在该手势识别区内的所有运动轨迹都作为后续识别为输入的信息的基础;而手势感应区则可作为提笔操作区域。
需要说明的是,设置手势感应区和手势识别区还可配合实现类似鼠标的输入操作,如指尖进入手势感应区的操作作为鼠标的左键单击操作,在预设的较短时间内连续进入两次手势感应区则作为左键双击操作,进入手势识别区的操作为鼠标的右键单击操作,指尖在手势感应区内维持预设的时间则为长按操作或者呼出菜单操作,多个指尖同时进入手势感应区可作为多点触控操作。可以看出,由于设置了两个区域,通过两个区域的合理定义能够多种类型的输入操作。
可以理解的是,除了能够模拟鼠标的左键单击、左键双击和右键单击之外, 该输入操作也可以是针对触摸屏的,即可以是单击触摸、双击触摸、长按触摸或者是多点触摸输入等触摸操作。
可选的,在输入过程中,为了便于用户能够识别正在输入的信息,可将运动轨迹同步映射至可穿戴设备的显示屏上,且轨迹识别区与所述显示屏上的区域相对应,请参阅图13,图13是本发明实施例的可穿戴设备的手势输入方法的一个实施例图,图13中可穿戴设备的显示屏上显示的用户在作为轨迹识别区的用户的手背上输入的信息,且显示屏中的内容与轨迹识别区的内容相对应,由于在轨迹识别区中划分手势感应区和手势识别区的情形下,若运动轨迹当前时刻位于手势感应区内,则在显示屏上显示出对应位置的光标,若运动轨迹当前时刻位于手势识别区内,则除了在显示屏上对应位置显示光标之外,还会识别出的书写轨迹也显示在显示屏上,如图中所示,虚线表示在手势感应区内的运动轨迹,而粗实线表示在手势识别区内的运动轨迹,可穿戴设备的显示屏上也映射出手势识别区内的运动轨迹,此外,手势感应区内的手指位置对应显示屏内的光标的位置。
可以理解的是,为了增强用户的输入体验,当进行书写操作时,穿戴设备可以通过显示屏上的光标等浮标图形状态的变化给予用于操作反馈,例如采用不同大小、不同透明度、不同虚实程度等的浮标图形给予操作反馈,当然也可以采用声音或振动作为区分提笔与落笔动作的操作反馈,辅助用户操作,增强用户的输入体验。
需要说明的是,对于手势操作来说,除了使用指尖进行操作,还可以采用与指尖类似的柱状物体进行操作。
当然,除了上述将轨迹识别区中划分出两个区域之外,也可将整个轨迹识别区作为手势识别区,而非轨迹识别区作为提笔操作的区域。这两种情形都需要对是否为连续输入进行确定,在前一种分区域的情形中,由于提笔操作的区域也为轨迹识别区,只要运动轨迹仍在该区域内则表示输入尚未完成,而对于后一种情形,由于没有轨迹识别区作为感应条件,一个可行的方式是增加延时时间,在运动轨迹离开轨迹识别区达到预设时间即确定输入操作完成,而在预设之间内又再次进入轨迹识别区,则表示输入尚未完成。
可选的,手势感应区设置的目的在于只要在该区域内检测到运动轨迹,并在可穿戴设备的显示屏内对应所述手势感应区内的运动轨迹的当前位置显示出光标。当然,也可将该手势感应区作为提笔操作区域,即书写过程中,只要未离开手势感应区则判定书写过程并未中断。。
可选的,上述手势识别区和手势感应区之间可以是相接触的,并且由于轨迹识别区是由传感器的扫描区域确定,因此可以设置在可穿戴设备的任意方位上,例如可穿戴设备的上方,该可穿戴设备的上方的周围,例如用户的手背上,或者是手臂上,该轨迹识别区的位置可根据实际需要进行位置设定,此外,针对不同的可穿戴设备,该轨迹识别区的大小也可自行设定,以满足不同用户对书写区域大小的需求。
举例来说,可穿戴设备为手环或者手表,其书写区域可以直接设置在穿戴该手环或者手表的手的手背上,手环或者手表上的传感器首先将扫描区域覆盖到手背,并在该扫描区域内通过坐标系预定义一个与手背表面类似的区域,当然,也可以实时扫描用户的手背,通过扫描出的手背数据从扫描区域中确定出一个对应该手背的手势识别区,该手势识别区在空间上能够与手背基本重合,而后可在该手势识别区上设置一形状与该手势识别区类似的手势感应区。
当然,对于其他物体,也可参照上述方式按照物体的表面设置手势识别区和手势感应区。
又举例来说,可穿戴设备为眼镜或者类似头戴可穿戴设备,则可将手势识别区设置在用户身前一定距离处,且为了便于用户操作,可将该区域的高度设置在用户的上半身某处,此外,由于这类可穿戴设备会具有一个透明的显示屏,可直接在该透明的显示屏上设置一区域图标,穿过该区域图标内部看到的空间作为轨迹识别区。
举例来说,请参阅图5至图8,图5为本发明实施例的可穿戴设备的手势输入方法的另一个实施例图,图6为本发明实施例的可穿戴设备的手势输入方法的另一个实施例图,图7为本发明实施例的可穿戴设备的手势输入方法的另一个实施例图,图8为本发明实施例的可穿戴设备的手势输入方法的另一个实施例图,在图5至图8中,轨迹识别区1的浅色区域部分为手势感应区11, 而深色区域部分则为手势识别区12,且该手势感应区11和手势识别区12是相接触的,此外,该手势识别区12和手势感应区11在图5和图6中位于可穿戴设备2的上方,而在图7和图8中则位于可穿戴设备2的侧上方,当用户的手指进入手势感应区11时,在可穿戴设备2的显示屏21上会以光标的方式显示出用户手指在手势感应区11的对应位置,而当用户的手指进入手势识别区12时,在可穿戴设备2的显示屏21上不仅会显示出用户手指在手势感应区11的对应位置,而且会显示出用户当前书写的过程。
此外,请参阅图9,图9是本发明实施例的可穿戴设备的手势输入方法的另一个实施例图,其中,轨迹识别区可位于可穿戴设备的侧上方,并且该轨迹识别区可以同时设置多个位置,以便于用户在各个应用场景下均能够顺利操作。
在上述情形下,步骤401具体为:
通过传感器采集所述轨迹识别区内的手势识别区内的运动轨迹。
可以看出,在此情形下,传感器采集的轨迹主要是手势识别区内的轨迹,而非整个轨迹识别区的轨迹,此方式能够减少一部分计算量,从而提高轨迹识别的性能。
402、根据轨迹识别区的区域信息将所述手势识别区内的运动轨迹识别为输入的信息。
其中,所述区域信息包括所述轨迹识别区的坐标系信息。
可以理解的是,为了能够正确识别出运动轨迹中的方向即偏移量信息,需要为轨迹识别区预先设置区域信息,该区域信息可以包括该轨迹识别区的坐标系信息,而由于手势识别区是设置在轨迹识别区内部的一个区域,在确定出轨迹识别区的坐标系信息后,便可以确定出手势识别区的坐标系信息,从而运动轨迹的方向以及偏移量均以该坐标系为基础进行计算。
可选的,若将轨迹识别区中划分出手势识别区和手势感应区的情形下,步骤402具体可为:
根据轨迹识别区的区域信息确定手势识别区的区域信息;
根据所述手势识别区的区域信息将所述手势识别区内的运动轨迹识别为 输入的信息。可以看出,此时无需识别整个轨迹识别区,而是仅需要识别轨迹识别区内的手势识别区即可完成输入的信息的识别,并且由于仅识别一个区域,识别干扰性较小,使得识别的准确性和速率都有所增强。
对于上述手势识别区的具体识别过程,具体可包括以下内容:
a、根据所述手势识别区的区域信息确定识别基准面。
可以理解的是,该识别基准面可以作为纸质书写过程中的纸面,也是用户在书写过程中默认书写的面,由于手势识别区为一个三维空间,在获得该空间内的运动轨迹后,需要对该运动轨迹进行处理才可得到在识别基准面上的轨迹,即最终需要的运动轨迹。
b、将所述手势识别区内的运动轨迹投影至所述识别基准面。
其中,为了使得手势识别区中的运动轨迹能够准确被识别成输入的信息,因此需要将该运动轨迹投影至识别基准面上,从而得到在该识别基准面上的精确的运动轨迹。
c、将投影至所述识别基准面内的运动轨迹识别为输入的信息。
在完成前述步骤a和步骤b后,即可对投影至所述识别基准面内的运动轨迹进行识别,从而通过该运动轨迹识别出想要的输入的信息。
需要说明的是,除了上述采用从轨迹识别区中划分出手势感应区和手势识别区的方式之外,轨迹识别区还可采用如下方式实现。
可选的,述轨迹识别区包括根据所述坐标系建立的第一识别面,所述根据所述轨迹识别区的区域信息将所述运动轨迹识别为输入的信息包括:
根据所述第一识别面的区域信息将在第一识别面内的运动轨迹识别为输入的信息,所述第一识别面为平面或者曲面。
即此情形下,轨迹识别区中真正作为识别的部分并非是一个区域而是一个面,即第一识别面,且该第一识别面可以是平面或者曲面,此方式目的在于区别于前述划分手势感应区和手势识别区的悬空的手势输入方式,此方式可以将真正用作识别的面投影至某一物体的表面,从而获得更为真实的书写体验。
需要说明的是,该第一识别面可设置在一些的物体的表面上。举例来说,当可穿戴设备为智能手表或者智能手环,当该第一识别面设置在用户的手背 上,用户的指尖或者其他类似柱状的物体作为笔尖,该第一识别面作为书写的参照面,当笔尖与第一识别面的距离大于0毫米时,视为提笔操作,当笔尖与第一识别面的距离小于等于0毫米时,则表示为落笔操作。由于人体手背是柔性表面,因此会出现笔尖与第一识别面的距离小于0毫米的情形,此外,该在设置提笔操作的距离时,也可预留预定的余量,如当笔尖与第一识别面的距离大于2毫米时,视为提笔操作,而笔尖与第一识别面的距离小于等于2毫米时,则为落笔操作,该余量可自行设定,可根据物体表面的形状以及第一识别面的大小进行设定。
此外,可选的,类似前述手势识别区和手势感应区的组合,该轨迹识别区中除了第一识别面之外,还可设置第二识别区,该第二识别区域第一识别面相接触,该第二识别区的作用与前述情形中的手势感应区的功能类似,一方面能够作为提笔区域,另一方面能够定位当前手势输入在输入区域内的位置,反应在可穿戴设备上即显示屏上的光标位置。
可选的,除了第一识别面之外,还可设置一个预设识别面,由于可穿戴设备是在佩戴过程中使用输入操作,因此物体相对于可穿戴设备在使用时难免会发生一定角度的移动,导致投影在物体上的第一识别面的大小会发生变化,若移动幅度过大,可能导致输入轨迹的识别不准,因此设置一个预设识别面,当第一识别面的角度与预设识别面夹角超过一定数值后,发出调整提示,从而保证第一识别面在能够在高精度的输入环境下进行输入操作。
其中,对于第一识别面与预设识别面的夹角的计算可以有多种方式,下面以其中一种进行介绍,对于一个物体表面来说,可穿戴设备的不同角度的移动会导致第一识别面在物体表面上投影出的区域大小发生改变,根据大小改变的程度即可计算出当前第一识别面的角度相较于固定的预设识别面之间的夹角。
举例来说,请参阅图10至图12,图10为本发明实施例的可穿戴设备的手势输入方法的另一个实施例图,图11为本发明实施例的可穿戴设备的手势输入方法的另一个实施例图,图12为本发明实施例的可穿戴设备的手势输入方法的另一个实施例图。其中,如图10所示,穿戴设备3为手腕可穿戴设备,轨迹识别区4的第一识别面41位于所述可穿戴设备3的穿戴者的手背上;
此外,如图11所示,该轨迹识别区4还包括第二识别区42,该第二识别区42与第一识别面41相接触,此情形下,用户在第二识别区42内的操作的轨迹均会映射至可穿戴设备3的显示屏上,该轨迹的当前位置会映射至显示屏上作为光标31的位置,此外,只要在该第二识别区42内不接触手背的情况下,都视为提笔操作,只有接触到手背,即相当于接触了第一识别面41才被认作是落笔操作;
另外,图12中给出了前述识别基准面43以及第一识别面41与该识别基准面43之间夹角的示意,图中识别基准面可以是垂直于Z轴的X/Y平面,只要该第一识别面41与X/Y平面的夹角在预设范围内,则由于处于正确的区域不发出第一识别面41调整的提示,而若第一识别面41与X/Y平面的夹角超出预设的范围,则发出第一识别面41调整的提示,指示用户需要对手背进行调整,以达到最佳的书写效果。
上面对本发明实施例的可穿戴设备的手势输入方法进行了介绍,下面对本发明实施例的可穿戴设备进行介绍,请参阅图14,图14是本发明实施例的可穿戴设备的一个实施例图,该可穿戴设备14包括:
传感器1401,用于采集轨迹识别区内用户手势运动的运动轨迹,所述轨迹识别区为设置于所述传感器预设的扫描区域内的区域,所述轨迹识别区包括手势识别区和设于所述手势识别区之上的手势感应区。
其中,该传感器1401能够实现图4所示实施例中步骤401。
此外,该传感器1401可以是多个传感器构成的传感器组,传感器的类型可以是超声波传感器或者是红外传感器,具体的多个传感器1401构成传感器组与图4所示实施例中针对步骤401的说明类似,此处不再赘述。
需要说明的是,设置手势识别区和手势感应区是为了对应实际书写过程的提笔和落笔操作,具体可参见图4所示实施例中针对步骤401的说明类似,此处不再赘述。当然,手势识别区和手势感应区除了实现提笔和落笔操作外,与图4所示实施例中针对步骤401的说明类似,还可以配合实现类似鼠标或者触摸屏的输入操作。
识别模块1402,用于根据所述轨迹识别区的区域信息将所述手势识别区 内的运动轨迹识别为输入的信息,所述区域信息包括所述轨迹识别区的坐标系信息。
可以看出,本发明实施例的可穿戴设备的工作过程可以是先由传感器1401采集轨迹识别区内尤其是手势识别区内的用户手势运动的运动轨迹,而后再由识别模块1402识别运动轨迹的基础为轨迹识别区的区域信息以及在手势识别区内的运动轨迹将运动轨迹识别成输入的信息,完成信息输入的过程。可选的,所述可穿戴设备还包括:
显示屏1403,用于将所述轨迹识别区的运动轨迹同步映射至所述显示屏1403内,所述轨迹识别区与所述显示屏内的区域相对应。
可以理解的是,与图4所实施例中针对步骤401的说明类似,次数的显示屏1403主要用于配合手势感应区显示出光标,该光标反应出的是在手势感应区内的用户手势运动的当前位置,且由于手势感应区与该显示屏1403具有位置对应关系,因此在手势感应区内的当前位置也会对应显示在显示屏1403上。
可选的,所述传感器1401具体用于:采集所述轨迹识别区内的手势识别区内的运动轨迹;
所述识别模块1402具体用于:根据所述轨迹识别区的区域信息确定所述手势识别区的区域信息;根据所述手势识别区的区域信息将所述手势识别区内的运动轨迹识别为输入的信息。
可以看出,与图4所示实施例中步骤402的说明类似,在实际进行运动轨迹识别时,用于确定最终输入的信息为手势识别区内的运动轨迹,因此传感器1401可以仅采集手势识别区内的运动轨迹即可,而识别模块1402在具有轨迹识别区的区域信息的基础上,会根据此区域信息确定出手势识别区的区域信息,即手势识别区的坐标系信息,并通过该坐标系信息对手势识别区内的运动轨迹进行识别,最终获得输入的信息。
可选的,与图4所示实施例中针对步骤402的替换步骤a至步骤c类似,识别模块1402具体用于:
根据所述手势识别区的区域信息确定识别基准面;
将所述手势识别区内的运动轨迹投影至所述识别基准面;
将投影至所述识别基准面内的运动轨迹识别为输入的信息。
可以看出,在识别模块1402的识别过程中,首先会确定出识别基准面,该识别基准面可以作为纸质书写过程中的纸面,也是用户在书写过程中默认书写的面,由于手势识别区为一个三维空间,在获得该空间内的运动轨迹后,需要对该运动轨迹进行处理才可得到在识别基准面上的轨迹,即最终需要的运动轨迹。在确定出识别基准面后,变回将手势识别区内的运动轨迹向该识别基准面进行投影,以获得在该识别基准面上的投影轨迹,最后对该投影轨迹进行识别,得到最终的输入的信息。
需要说明的是,本发明实施例中的通过传感器1401扫描的区域设立的手势识别区和手势感应区可以是悬空设置,也可以是手势识别区是位于物体上的一个面。下面分别进行说明:
一、手势识别区和手势感应区可以是悬空设置。
可选的,所述手势识别区和所述手势感应区相接触,所述手势识别区和所述手势感应区位于所述可穿戴设备的上方。此设置方式无需考虑承载该手势识别区和手势感应区的媒介,可以应用于各种场合。
二、手势识别区是位于物体上。
可选的,所述手势识别区为根据所述手势识别区的区域信息建立的第一识别面,所述识别模块1402具体用于:
根据所述手势识别区的区域信息将在第一识别面内的运动轨迹识别为输入的信息,所述第一识别面为平面或者曲面。
可以看出,在此情形下,手势识别区可以是设置在物体表面上的一个识别面,该识别面基本与物体的表面重合,从而使得用户能在物体表面上通过书写进行输入操作,能够大大增强书写的真实感,第一识别面的与图4所示实施例中针对步骤402的说明类似,此外,针对该第一识别面的信息说明可参见图4所示实施例中针对步骤402的说明中以可穿戴设备为智能手表或者智能手环的书写实例,此处不再赘述。
可选的,由于可穿戴设备与物体表面之间往往会发生相对运动,可能影响到手势输入的准确性,可以通过设置预设识别面进行矫正,可选的,该可穿戴 设备14还可包括提示模块1404,用于当确定所述第一识别面与预设识别面的夹角超出预设的角度阈值时,发出第一识别面调整提示。
具体的,由于可穿戴设备是在佩戴过程中使用输入操作,因此物体相对于可穿戴设备在使用时难免会发生一定角度的移动,导致投影在物体上的第一识别面的大小会发生变化,若移动幅度过大,可能导致输入轨迹的识别不准,因此设置一个预设识别面,当第一识别面的角度与预设识别面夹角超过一定数值后,发出调整提示,从而保证第一识别面在能够在高精度的输入环境下进行输入操作。
下面对本发明实施例中训练设备的结构进行描述,请参阅图15,图15是本发明实施例的可穿戴设备的一个实施例图,其中,可穿戴设备15可包括均与总线相连接的至少一个处理器1501、至少一个传感器1502和至少一个内存1503,本发明实施例涉及的可穿戴设备可以具有比图15所示出的更多或更少的部件,可以组合两个或更多个部件,或者可以具有不同的部件配置或设置,各个部件可以在包括一个或多个信号处理和/或专用集成电路在内的硬件、软件或硬件和软件的组合实现。
具体的,对于图14所示的实施例来说,该处理器1501能实现图14所示实施例中的识别模块1402和提示模块1404的功能,该传感器1502能实现图14所示实施例中的传感器1401的功能,该内存1503用于存储处理器1501要执行的指令和数据,处理器1501用于执行所述内存中的指令以实现将手势识别区内的运动轨迹识别为输入的信息。其中,可穿戴设备还可以包括显示屏1504,该显示屏1504能实现图14中显示屏1403的功能。
所属领域的技术人员可以清楚地了解到,为描述的方便和简洁,上述描述的系统,装置和单元的具体工作过程,可以参考前述方法实施例中的对应过程,在此不再赘述。
在本申请所提供的几个实施例中,应该理解到,所揭露的系统,装置和方法,可以通过其它的方式实现。例如,以上所描述的装置实施例仅仅是示意性的,例如,所述单元的划分,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分方式,例如多个单元或组件可以结合或者可以集成到另一个系统,或 一些特征可以忽略,或不执行。另一点,所显示或讨论的相互之间的耦合或直接耦合或通信连接可以是通过一些接口,装置或单元的间接耦合或通信连接,可以是电性,机械或其它的形式。
所述作为分离部件说明的单元可以是或者也可以不是物理上分开的,作为单元显示的部件可以是或者也可以不是物理单元,即可以位于一个地方,或者也可以分布到多个网络单元上。可以根据实际的需要选择其中的部分或者全部单元来实现本实施例方案的目的。
另外,在本发明各个实施例中的各功能单元可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中。上述集成的单元既可以采用硬件的形式实现,也可以采用软件功能单元的形式实现。
所述集成的单元如果以软件功能单元的形式实现并作为独立的产品销售或使用时,可以存储在一个计算机可读取存储介质中。基于这样的理解,本发明的技术方案本质上或者说对现有技术做出贡献的部分或者该技术方案的全部或部分可以以软件产品的形式体现出来,该计算机软件产品存储在一个存储介质中,包括若干指令用以使得一台计算机设备(可以是个人计算机,服务器,或者网络设备等)执行本发明各个实施例所述方法的全部或部分步骤。而前述的存储介质包括:U盘、移动硬盘、只读存储器(ROM,Read-Only Memory)、随机存取存储器(RAM,Random Access Memory)、磁碟或者光盘等各种可以存储程序代码的介质。
以上所述,以上实施例仅用以说明本发明的技术方案,而非对其限制;尽管参照前述实施例对本发明进行了详细的说明,本领域的普通技术人员应当理解:其依然可以对前述各实施例所记载的技术方案进行修改,或者对其中部分技术特征进行等同替换;而这些修改或者替换,并不使相应技术方案的本质脱离本发明各实施例技术方案的精神和范围。

Claims (21)

  1. 一种可穿戴设备的手势输入方法,其特征在于,所述方法包括:
    通过传感器采集轨迹识别区内用户手势运动的运动轨迹,所述轨迹识别区为设置于所述传感器预设的扫描区域内的区域,所述轨迹识别区包括手势识别区和设于所述手势识别区之上的手势感应区;
    根据所述轨迹识别区的区域信息将所述手势识别区内的运动轨迹识别为输入的信息,所述区域信息包括所述轨迹识别区的坐标系信息。
  2. 根据权利要1所述的可穿戴设备的手势输入方法,其特征在于,所述可穿戴设备还包括显示屏,所述方法还包括:
    将所述轨迹识别区的运动轨迹同步映射至所述显示屏内,所述轨迹识别区与所述显示屏内的区域相对应。
  3. 根据权利要求1或2所述的可穿戴设备的手势输入方法,其特征在于,所述通过传感器采集轨迹识别区内用户手势运动的运动轨迹具体为:
    通过传感器采集所述轨迹识别区内的手势识别区内的运动轨迹;
    所述根据所述轨迹识别区的区域信息将所述运动轨迹识别为输入的信息具体为:
    根据所述轨迹识别区的区域信息确定所述手势识别区的区域信息;
    根据所述手势识别区的区域信息将所述手势识别区内的运动轨迹识别为输入的信息。
  4. 根据权利要求3所述的可穿戴设备的手势输入方法,其特征在于,所述根据所述手势识别区的区域信息将所述手势识别区内的运动轨迹识别为输入的信息包括:
    根据所述手势识别区的区域信息确定识别基准面;
    将所述手势识别区内的运动轨迹投影至所述识别基准面;
    将投影至所述识别基准面内的运动轨迹识别为输入的信息。
  5. 根据权利要求3或4所述的可穿戴设备的手势输入方法,其特征在于:所述方法还包括:
    当所述手势感应区检测到运动轨迹,所述可穿戴设备的显示屏内对应所述 手势感应区内的运动轨迹的当前位置显示出光标。
  6. 根据权利要求5所述的可穿戴设备的手势输入方法,其特征在于:所述手势识别区和所述手势感应区相接触,所述手势识别区和所述手势感应区位于所述可穿戴设备的上方。
  7. 根据权利要求1或2所述的可穿戴设备的手势输入方法,其特征在于,所述手势识别区为根据所述手势识别区的区域信息建立的第一识别面,根据所述轨迹识别区的区域信息将所述手势识别区内的运动轨迹识别为输入的信息包括:
    根据所述手势识别区的区域信息将在第一识别面内的运动轨迹识别为输入的信息,所述第一识别面为平面或者曲面。
  8. 根据权利要求7所述的可穿戴设备的手势输入方法,其特征在于,所述方法还包括:
    当确定所述第一识别面与预设识别面的夹角超出预设的角度阈值时,发出第一识别面调整提示。
  9. 根据权利要求7或8中任一项所述的可穿戴设备的手势输入方法,其特征在于,所述可穿戴设备为手腕可穿戴设备,所述第一识别面位于所述可穿戴设备的穿戴者的手背上。
  10. 根据权利要求1至9中任一项所述的可穿戴设备的手势输入方法,其特征在于,所述传感器为一个以上的超声波传感器或者红外传感器。
  11. 一种可穿戴设备,其特征在于,包括:
    传感器,用于采集轨迹识别区内用户手势运动的运动轨迹,所述轨迹识别区为设置于所述传感器预设的扫描区域内的区域,所述轨迹识别区包括手势识别区和设于所述手势识别区之上的手势感应区;
    识别模块,用于根据所述轨迹识别区的区域信息将所述手势识别区内的运动轨迹识别为输入的信息,所述区域信息包括所述轨迹识别区的坐标系信息。
  12. 根据权利要求11所述的可穿戴设备,其特征在于,所述可穿戴设备还包括:
    显示屏,用于将所述轨迹识别区的运动轨迹同步映射至所述显示屏内,所 述轨迹识别区与所述显示屏内的区域相对应。
  13. 根据权利要求11或12所述的可穿戴设备,其特征在于,所述传感器具体用于:
    采集所述轨迹识别区内的手势识别区内的运动轨迹;
    所述识别模块具体用于:
    根据所述轨迹识别区的区域信息确定所述手势识别区的区域信息;
    根据所述手势识别区的区域信息将所述手势识别区内的运动轨迹识别为输入的信息。
  14. 根据权利要求13所述的可穿戴设备,其特征在于,所述识别模块具体用于:
    根据所述手势识别区的区域信息确定识别基准面;
    将所述手势识别区内的运动轨迹投影至所述识别基准面;
    将投影至所述识别基准面内的运动轨迹识别为输入的信息。
  15. 根据权利要求13所述的可穿戴设备,其特征在于,所述传感器还具体用于检测所述手势感应区内的运动轨迹;
    所述显示屏还用于在所述显示屏内对应所述手势感应区内的运动轨迹的当前位置显示出光标。
  16. 根据权利要求15所述的可穿戴设备,其特征在于,所述手势识别区和所述手势感应区相接触,所述手势识别区和所述手势感应区位于所述可穿戴设备的上方。
  17. 根据权利要求11或12所述的可穿戴设备,其特征在于,所述手势识别区为根据所述手势识别区的区域信息建立的第一识别面,所述识别模块具体用于:
    根据所述手势识别区的区域信息将在第一识别面内的运动轨迹识别为输入的信息,所述第一识别面为平面或者曲面。
  18. 根据权利要求17所述的可穿戴设备,其特征在于,所述可穿戴设备还包括:
    提示模块,用于当确定所述第一识别面与预设识别面的夹角超出预设的角 度阈值时,发出第一识别面调整提示。
  19. 根据权利要求17或18所述的可穿戴设备,其特征在于,所述可穿戴设备为手腕可穿戴设备,所述第一识别面位于所述可穿戴设备的穿戴者的手背上。
  20. 根据权利要求11至19中任一项所述的可穿戴设备,其特征在于,所述传感器为一个以上的超声波传感器或者红外传感器。
  21. 一种可穿戴设备,其特征在于,包括传感器、处理器和内存所述传感器用于采集用户手势运动的运动轨迹,所述内存用于存储处理器要执行的指令和数据,所述处理器用于执行所述内存中的指令以实现将手势识别区内的运动轨迹识别为输入的信息。
PCT/CN2016/092399 2016-07-29 2016-07-29 一种可穿戴设备的手势输入方法及可穿戴设备 WO2018018624A1 (zh)

Priority Applications (4)

Application Number Priority Date Filing Date Title
US16/321,714 US20210117078A1 (en) 2016-07-29 2016-07-29 Gesture Input Method for Wearable Device and Wearable Device
EP16910204.3A EP3486747A4 (en) 2016-07-29 2016-07-29 GESTURE INPUT METHOD FOR PORTABLE DEVICE, AND PORTABLE DEVICE
CN201680049793.3A CN108027648A (zh) 2016-07-29 2016-07-29 一种可穿戴设备的手势输入方法及可穿戴设备
PCT/CN2016/092399 WO2018018624A1 (zh) 2016-07-29 2016-07-29 一种可穿戴设备的手势输入方法及可穿戴设备

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2016/092399 WO2018018624A1 (zh) 2016-07-29 2016-07-29 一种可穿戴设备的手势输入方法及可穿戴设备

Publications (1)

Publication Number Publication Date
WO2018018624A1 true WO2018018624A1 (zh) 2018-02-01

Family

ID=61015913

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2016/092399 WO2018018624A1 (zh) 2016-07-29 2016-07-29 一种可穿戴设备的手势输入方法及可穿戴设备

Country Status (4)

Country Link
US (1) US20210117078A1 (zh)
EP (1) EP3486747A4 (zh)
CN (1) CN108027648A (zh)
WO (1) WO2018018624A1 (zh)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110209268A (zh) * 2019-04-24 2019-09-06 努比亚技术有限公司 穿戴式设备控制方法、穿戴式设备及计算机可读存储介质
CN110989844A (zh) * 2019-12-16 2020-04-10 广东小天才科技有限公司 一种基于超声波的输入方法、手表、系统及存储介质

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11461907B2 (en) * 2019-02-15 2022-10-04 EchoPixel, Inc. Glasses-free determination of absolute motion
USD962929S1 (en) * 2020-04-03 2022-09-06 Embr Labs Inc. Wearable device
CN111580656B (zh) * 2020-05-08 2023-07-18 安徽华米信息科技有限公司 可穿戴设备及其控制方法、装置
JP1695382S (zh) * 2020-09-28 2021-09-21
CN113672093B (zh) * 2021-08-27 2024-02-02 歌尔科技有限公司 一种手势识别方法、装置、穿戴设备主体及穿戴设备
CN113934307B (zh) * 2021-12-16 2022-03-18 佛山市霖云艾思科技有限公司 一种根据手势和场景开启电子设备的方法

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6313825B1 (en) * 1998-12-28 2001-11-06 Gateway, Inc. Virtual input device
US20070121097A1 (en) * 2005-11-29 2007-05-31 Navisense, Llc Method and system for range measurement
CN102782612A (zh) * 2010-02-24 2012-11-14 诺基亚公司 姿势控制

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8624836B1 (en) * 2008-10-24 2014-01-07 Google Inc. Gesture-based small device input
EP2256592A1 (en) * 2009-05-18 2010-12-01 Lg Electronics Inc. Touchless control of an electronic device
KR101534282B1 (ko) * 2014-05-07 2015-07-03 삼성전자주식회사 포터블 디바이스의 사용자 입력 방법 및 상기 사용자 입력 방법이 수행되는 포터블 디바이스
CN105159539B (zh) * 2015-09-10 2018-06-01 京东方科技集团股份有限公司 可穿戴设备的触控响应方法、装置及可穿戴设备
CN105302298B (zh) * 2015-09-17 2017-05-31 深圳市国华识别科技开发有限公司 空中书写断笔系统和方法

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6313825B1 (en) * 1998-12-28 2001-11-06 Gateway, Inc. Virtual input device
US20070121097A1 (en) * 2005-11-29 2007-05-31 Navisense, Llc Method and system for range measurement
CN102782612A (zh) * 2010-02-24 2012-11-14 诺基亚公司 姿势控制

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See also references of EP3486747A4 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110209268A (zh) * 2019-04-24 2019-09-06 努比亚技术有限公司 穿戴式设备控制方法、穿戴式设备及计算机可读存储介质
CN110989844A (zh) * 2019-12-16 2020-04-10 广东小天才科技有限公司 一种基于超声波的输入方法、手表、系统及存储介质

Also Published As

Publication number Publication date
CN108027648A (zh) 2018-05-11
EP3486747A1 (en) 2019-05-22
US20210117078A1 (en) 2021-04-22
EP3486747A4 (en) 2019-05-22

Similar Documents

Publication Publication Date Title
WO2018018624A1 (zh) 一种可穿戴设备的手势输入方法及可穿戴设备
US20210271340A1 (en) Gesture recognition devices and methods
US11614793B2 (en) Precision tracking of user interaction with a virtual input device
CN105159539B (zh) 可穿戴设备的触控响应方法、装置及可穿戴设备
EP3090331B1 (en) Systems with techniques for user interface control
KR100811015B1 (ko) 가상 입력 장치를 사용하여 데이터를 입력하는 방법 및 장치
TWI599922B (zh) 包含具有兩種以上自由度之使用者介面之電子設備,該使用者介面包含一觸碰敏感表面及非接觸式偵測構件
US20060028457A1 (en) Stylus-Based Computer Input System
US20150323998A1 (en) Enhanced user interface for a wearable electronic device
CN107430450B (zh) 用于生成输入的设备和方法
WO2019062243A1 (zh) 触摸操作的识别方法、装置及电子设备
CN103713737A (zh) 用于智能眼镜的虚拟键盘系统
US20130120250A1 (en) Gesture recognition system and method
US9262012B2 (en) Hover angle
WO2021227628A1 (zh) 一种电子设备及其交互方法
Guimbretière et al. Bimanual marking menu for near surface interactions
US20230057020A1 (en) Meeting interaction system
CN205050078U (zh) 一种可穿戴设备
WO2013039205A1 (ja) 電子筆記装置及び電子筆記処理プログラム
CN106598422B (zh) 混合操控方法及操控系统和电子设备
CN113515228A (zh) 一种虚拟标尺显示方法以及相关设备
KR101564089B1 (ko) 손동작 인식을 이용한 프리젠테이션 실행시스템
CN110570930B (zh) 一种基于平板电脑的医疗影像勾画方法及系统
JP2010086367A (ja) 位置情報入力装置、位置情報入力方法、プログラム、情報処理システム、及び電子装置
JP2024024440A (ja) ペン状態検出回路及び方法、並びに入力システム

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 16910204

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

ENP Entry into the national phase

Ref document number: 2016910204

Country of ref document: EP

Effective date: 20190215