US20180260034A1 - Text input method and device based on gesture recognition, and storage medium - Google Patents

Text input method and device based on gesture recognition, and storage medium Download PDF

Info

Publication number
US20180260034A1
US20180260034A1 US15/761,034 US201615761034A US2018260034A1 US 20180260034 A1 US20180260034 A1 US 20180260034A1 US 201615761034 A US201615761034 A US 201615761034A US 2018260034 A1 US2018260034 A1 US 2018260034A1
Authority
US
United States
Prior art keywords
writing
target object
trace
plane
writing plane
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US15/761,034
Other versions
US10725552B2 (en
Inventor
Guohua Liu
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
SHENZHEN PRTEK Co Ltd
Original Assignee
SHENZHEN PRTEK Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by SHENZHEN PRTEK Co Ltd filed Critical SHENZHEN PRTEK Co Ltd
Publication of US20180260034A1 publication Critical patent/US20180260034A1/en
Assigned to SHENZHEN PRTEK CO. LTD. reassignment SHENZHEN PRTEK CO. LTD. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: LIU, GUOHUA
Application granted granted Critical
Publication of US10725552B2 publication Critical patent/US10725552B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/017Gesture based interaction, e.g. based on a set of recognized hand gestures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/0304Detection arrangements using opto-electronic means
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • G06F3/04883Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • G06F3/167Audio in a user interface, e.g. using voice commands for navigating, audio feedback
    • G06K9/00355
    • G06K9/00416
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/22Character recognition characterised by the type of writing
    • G06V30/228Character recognition characterised by the type of writing of three-dimensional handwriting, e.g. writing in the air
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/32Digital ink
    • G06V30/333Preprocessing; Feature extraction
    • G06V30/347Sampling; Contour coding; Stroke extraction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • G06V40/28Recognition of hand or arm movements, e.g. recognition of deaf sign language
    • G06K2209/011
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/28Character recognition specially adapted to the type of the alphabet, e.g. Latin alphabet
    • G06V30/287Character recognition specially adapted to the type of the alphabet, e.g. Latin alphabet of Kanji, Hiragana or Katakana characters

Definitions

  • the present disclosure is related to the computer recognition technology, and in particular related to a method, a device and a storage medium of inputting text based on hand gesture recognition.
  • a method, a device and a storage medium of inputting text based on hand gesture recognition are provided.
  • a method of inputting text based on hand gesture recognition includes:
  • a device of inputting text based on hand gesture recognition including a processor; and a memory storing instructions, which, when executed by the processor causes the processor to perform steps including:
  • At least one non-transitory computer-readable storage medium storing computer-readable instructions that, when executed by at least one processors, cause the at least one processor to perform steps including:
  • the target object is determined according to the image information, and the writing plane and the movement trace of the target object are acquired, the movement trace of the target object in the writing plane is extracted as the writing trace to be output, the character recognition is performed according to the writing trace, and the character recognition result is output, such that even if the movement trace of the target object in the air in consecutive, the trace extraction module can still extract the writing trace in the writing plane, which is equivalent to a strokes distinguishing process and facilitates the recognition of the character recognition module and improves the precision of recognition.
  • FIG. 1 is a schematic diagram of the device of inputting text based on hand gesture recognition according to an embodiment
  • FIG. 2 is a block diagram of the device of inputting text based on hand gesture recognition according to an embodiment
  • FIG. 3 is a block diagram of a target acquiring module according to the embodiment shown in FIG. 2 ;
  • FIG. 4 is a block diagram of a trace extraction module according to the embodiment shown in FIG. 2 ;
  • FIG. 5 is a schematic diagram of a writing in the air according to an embodiment
  • FIG. 6 is a flowchart of a method of inputting text based on hand gesture recognition according to an embodiment
  • FIG. 7 is a flowchart of a method of inputting text based on hand gesture recognition according to another embodiment.
  • FIG. 8 is a schematic diagram of an application of a method of inputting text based on hand gesture recognition according to an embodiment.
  • FIG. 1 is a schematic diagram of the device of inputting text based on hand gesture recognition according to an embodiment.
  • the device of inputting text based on hand gesture recognition includes a processor, a storage medium, a Random Access Memory (RAM) and an input/output (I/O) interface connected through device bus.
  • An operation device, a database and computer-readable instructions are stored in the storage medium of the device, the database is configured to store data, such as storing image information, movement trace and writing trace and so on.
  • the processor of the device is configured to provide computation and control capability to support the operation of the entire device.
  • the RAM of the terminal provides an operation environment for the computer-readable instructions in the storage medium.
  • the I/O interface of the device is configured to connect other equipment.
  • FIG. 2 is a block diagram of the device of inputting text based on hand gesture recognition according to an embodiment.
  • the device of inputting text based on hand gesture recognition includes an image collection module 110 , a target acquiring module 120 , a trace extraction module 130 , a character recognition module 140 , and a display module 150 .
  • the image collection module 110 is configured to collect image information containing a target object.
  • the target acquiring module 120 is configured to determine the target object according to the image information, and acquire a writing plane and a movement trace of the target object.
  • the trace extraction module 130 is configured to extract the movement trace of the target object in the writing plane as a writing trace, outputting the writing trace.
  • the character recognition module 140 is configured to perform character recognition according to the writing trace, and output a character recognition result.
  • the image collection module 110 is a camera that is configured to collect image information containing a target object.
  • the target object can be a finger of a user, or a stick-like object held by the user, such as a pencil.
  • the image collection module 110 will collect the image information containing the pencil.
  • the device of inputting text based on hand gesture recognition also includes a hand gesture recognition module and/or a voice recognition module.
  • the hand gesture recognition module is configured to recognize the hand gesture of the user and output a hand gesture recognition result, and the hand gesture recognition result is to start or end a writing status.
  • the hand gesture recognition module is configured to recognize the hand gesture of the user and output a hand gesture recognition result, and the hand gesture recognition result is to start or end a writing status.
  • a corresponding hand gesture can be made within the visible area of the image collection module 110 .
  • the hand gesture recognition module will automatically adapt the visible area of the image collection module 110 for the collection of hand gesture images (such as focusing on the hand of the user and zooming in with the lens), and a hand gesture recognition result will be output by the hand gesture recognition module, i.e., starting writing status.
  • the user can also start the writing status by way of voice directly, such as say “start writing” to a microphone for collecting voice, then the voice recognition module will output a voice recognition result, i.e., starting the writing status.
  • a corresponding hand gesture can be made within the visible area of the image collection module 110 , the hand gesture recognition module will output a hand gesture recognition result, i.e., ending the writing status.
  • the user can also end the writing status by way of voice directly, such as say “end writing” to the microphone for collecting voice, then the voice recognition module will output a voice recognition result, i.e., ending the writing status.
  • the target acquiring module 120 is configured to determine the target object according to the image information, and acquire a writing plane and a movement trace of the target object.
  • the target acquiring module 120 includes an image processing unit 122 , a time calculation unit 124 , and a writing plane determination unit 126 .
  • the image processing unit 122 is configured to perform image characteristic extraction to the image information to determine the target object.
  • the image characteristics herein mainly refer to acquiring image contours of the digital image by an edge extraction. Typically, a place where the gray scale value changes drastically is defined an edge. If the target object is a finger or a pencil or other rod-like object, then the edge of the image is a very critical characteristic. By extracting the image characteristics, the target object can be determined swiftly.
  • the time calculation unit 124 is configured to count a pause duration of the target object at a position. In particular, when the target object is determined, a tracking of the target object can be achieved. When the target object is static, the time calculating unit 124 will start timing which is stopped when the target object starts moving again, the duration elapsed therebetween is the pause duration of the target object at the position.
  • the writing plane determination unit 126 is configured to acquire a first position where the pause duration of the target object reaches a preset duration for a first time after starting the writing status, and acquire a second position where the pause duration of the target object reaches the preset duration for a second time; and defining the movement trace of the target object from the first position to the second position as a first stroke, defining a plane where the first stroke is located as a reference writing plane, and defining a plane away from the reference writing plane within a preset distance range as a writing plane.
  • the first position where the pause duration of the target object reaches the preset duration for the first time is defined A
  • the second position where the pause duration of the target object reaches the preset duration for the second time is defined B
  • a connecting line between A and B constitutes a reference writing plane.
  • the writing plane determination unit 126 can also use other manners to define the reference writing plane, such as by a particular hand gesture to start the writing status, and regard the plane where one end for writing of the target object located as the reference writing plane.
  • the writing status is started by voice, then the target object stay for a period of time at the position about to write, and the plane where the target object stayed is defined the reference writing plane.
  • the writing status is started by voice, gesture, remote control or any other means, the target object moves forward or downward rectilinearly at a certain speed to a position and stay for a while, the plane where such position located is defined the reference writing plane.
  • the trace of the target object in the air turned and stayed in a position after an “L”-like trace is drew
  • the plane where such a position is drew is defined the reference writing plane.
  • the determination of whether an “L”-like trace is drew by the target object is mainly to track the target object to acquire the movement trace of the target object, and then determine if the movement trace is “L”-like.
  • the reference writing plane can be defined by a combination of the foregoing manners.
  • the plane whose intersection angle in relation to the writing plane is within a preset angle range is defined as the writing plane, which equivalent to regarding the writing plane as a set of planes, and the movement traces in all of the planes in the set are then extracted, so as to facilitate the improvement of the precision of character recognition.
  • the preset duration is two seconds, i.e., the user will control the target object to pause over two seconds both in starting and ending the stroke. It should be understood that, the preset duration can be set as one second, two seconds, or three seconds etc., no limitation is intended herein.
  • the trace extraction module 130 is configured to extract the movement trace of the target object in the writing plane as a writing trace, outputting the writing trace.
  • the trace extraction module 130 includes a trace extraction unit 132 and a character output unit 134 .
  • the trace extraction unit 132 is configured to extract the movement trace of the target object in the writing plane as a writing trace.
  • the character output unit 134 is configured to output the writing trace with the characters as the unit.
  • the target object pause for a certain period of time upon completing of the input of a character.
  • the threshold of interval to be four seconds
  • the pause duration of the target object exceeds four seconds, it is considered that the user has completed the input of a character.
  • the character can be the Chinese character, the letter, the digit, the symbol, or other individual characters input by the user.
  • the character recognition module 140 is configured to perform character recognition according to the writing trace, and output a character recognition result.
  • the character recognition module 140 will compare the writing trace (i.e., the output character) output by the trace extraction module 130 to the character in the preset character base and find out from the character base a character with the highest similarity to the writing trace as the character recognition result to output.
  • the display module 150 is configured to display the character recognition result and/or the writing trace. Referring to FIG. 5 for details, the display module 150 includes a writing display zone 152 and a recognition zone 154 .
  • the writing display zone 152 will display the writing trace output by the trace extraction module 130
  • the recognition display zone 154 will display the character recognition result output by the character recognition module 140 .
  • the display module 150 displays the writing trace in real time, and when the writing trace is output with the characters as the unit to the character recognition module 140 , the display module 150 will then display the character recognition result output by the character recognition module 140 .
  • the display module 150 includes a target object detection unit and a display unit.
  • the target object detection unit is configured to detect if the target object is departing the writing plane in an angle between 45 and 90 degrees in relation to the reference writing plane, or to detect if the target object is departing the writing plane at a preset speed; if yes, then the display unit stops displaying the writing trace.
  • the target object detection unit is further configured to detect if the target object is arriving the writing plane in the angle between 45 and 90 degrees in relation to the reference writing plane, or to detect if the target object is arriving the writing plane at the preset speed, or detecting if the target object is arriving the writing plane in an “L”-like movement trace; if yes, the target object detection unit starts displaying the writing trace.
  • To determine if an “L”-like trace is drew by the target object it is mainly to track the target object to acquire the movement trace of the target object, and then determine if the movement trace is “L”-like.
  • a size adjustment module is included, which is configured to adjust the size of the character according to a distance between one end of the target object proximately to the reference writing plane and the reference writing plane.
  • the end of the target object near the reference writing plane can be understood as the pen point, the farther the pen point away from the reference writing plane, the thinner the stroke of the character.
  • the nearer the pen point away from the reference writing plane the thicker the stroke of the character.
  • the size adjustment module can also be configured to effect an adjustment of the displayed character from being thin to being thick according to the extension of the pause in the first down stroke. Also, the farther the target object is away from the reference writing plane, the thinner the stroke at the up stroke. When the distance target object is away from the reference writing plane amounts to a certain value, it is considered the “up stroke” is finished and a process of distinguishing the strokes is performed.
  • it can be defined a “down stroke” that the target object arrives the writing plane in an “L”-like movement trace, and a “up stroke” that the target object departs the writing plane in an “L”-like movement trace.
  • a “down stroke” that the target object arrives the writing plane in an “L”-like movement trace
  • a “up stroke” that the target object departs the writing plane in an “L”-like movement trace.
  • FIG. 6 is a flowchart of a method of inputting text based on hand gesture recognition according to an embodiment.
  • a method of inputting text based on hand gesture recognition includes:
  • step S 110 image information containing a target object is collected
  • the target object can be a finger of a user, or a stick-like object held by the user, such as a pencil.
  • step S 120 the target object is determined according to the image information, and a writing plane and a movement trace of the target object are acquired.
  • a size of the character is adjusted according to a distance between one end of the target object proximately to the reference writing plane and the reference writing plane.
  • step S 130 the movement trace of the target object in the writing plane is extracted as the writing trace, the writing trace is output.
  • the output is performed with the characters as the unit in step S 130 .
  • step S 140 character recognition is performed according to the writing trace, and a character recognition result is output.
  • FIG. 7 is a flowchart of a method of inputting text based on hand gesture recognition according to another embodiment.
  • a method of inputting text based on hand gesture recognition includes:
  • step S 210 a hand gesture of a user is recognized and a writing status is started according to a hand gesture recognition result.
  • step S 210 can also be that a voice of a user is recognized and the writing status is started according to a voice recognition result.
  • step S 220 image information containing a target object is collected.
  • step S 230 the target object is determined according to the image information, and a writing plane and a movement trace of the target object are acquired.
  • step S 240 the movement trace of the target object in the writing plane is extracted as the writing trace, the writing trace is output.
  • step S 250 character recognition is performed according to the writing trace, and a character recognition result is output.
  • step S 260 the character recognition result and the writing trace are displayed.
  • the writing trace is first displayed and then the character recognition result is displayed, it is because the writing trace is displayed in real time. In an embodiment, it can only display the character recognition result or the writing trace.
  • step of displaying the writing trace in step S 260 includes:
  • the target object is arriving the writing plane in the angle between 45 and 90 degrees in relation to the reference writing plane, or it is detected if the target object is arriving the writing plane at the preset speed, or detecting if the target object is arriving the writing plane in an “L”-like movement trace; if the target object is arriving the writing plane in the angle between 45 and 90 degrees in relation to the reference writing plane or the target object is departing the writing plane at the preset speed or the target object is arriving the writing plane in the “L”-like movement trace, start displaying the writing trace.
  • step S 270 a hand gesture of a user is recognized and a writing status is ended according to a hand gesture recognition result.
  • step S 270 can also be that a voice of a user is recognized and the writing status is ended according to a voice recognition result.
  • the target object is a finger that writes a Chinese character “ ” in the air.
  • the user shows his/her finger in the visible range of the camera, and starts the writing status by voice.
  • it is the down stroke the finger of the user moving from position “0” to the first stroke starting point position “1”, from position “1” to position “2” is a first stroke, and the plane where the first stroke locates is defined the writing plane.
  • From position “2” to position “3” is an up stroke, moving from position “3” to position “4” is a movement outside the writing plane.
  • From position “4” to position “5” is a second down stroke, and from position “5” to position “6” is a second stroke.
  • From position “6” to position “7” is the up stroke, moving from position “7” to position “8” is a movement outside the writing plane.
  • From position “8” to position “9” is a third stroke, from position “9” to position “10” is a third stroke, from position “10” to position “11” is the up stroke, and the input of a character is completed.
  • the finger of the user will make a pause at position “1”, position “2”, position “5”, position “6”, position “9” and position “10”, although the movement trace of the finger of the user from position “0” to position “11” is consecutive, yet as only the movement trace in the writing plane is extracted as writing trace, this amounts to a strokes distinguishing process, which can improve the recognition precision of the input character, facilitates the user to perform inputting text, and helps to train the kids with good a writing manner.
  • the target object is determined according to the image information, and the writing plane and the movement trace of the target object are acquired, the movement trace of the target object in the writing plane is extracted as the writing trace to be output, the character recognition is performed according to the writing trace, and the character recognition result is output, such that even if the movement trace of the target object in the air in consecutive, yet the trace extraction module will only extract the writing trace in the writing plane, which is equivalent to a strokes distinguishing process and facilitates the recognition of the character recognition module and improves the precision of recognition.
  • At least one non-transitory computer-readable storage medium which stores computer-readable instructions that, when executed by at least one processors, cause the at least one processor to perform steps as follow:
  • Image information containing a target object is collected
  • the target object is determined according to the image information, and a writing plane and a movement trace of the target object are acquired;
  • the movement trace of the target object in the writing plane is extracted as a writing trace, outputting the writing trace;
  • Character recognition is performed according to the writing trace, and a character recognition result is output.
  • the processes of the methods in the above embodiments can be, in full or in part, implemented by computer programs instructing underlying hardware, the programs can be stored in a computer-readable storage medium, the program can include the processes in the embodiments of the various methods when it is being executed.
  • the storage medium can be a disk, a CD, a Read-Only Memory (ROM) and other non-transitory storage mediums or Random Access Memory (RAM) and so on.

Abstract

A text input method based on gesture recognition, comprising: collecting image information containing a target object; determining the target object according to the image information, and acquiring a writing surface and a movement trace of the target object; extracting the movement trace of the target object in the writing surface as a writing trace and outputting same; and performing character recognition according to the writing trace, and outputting a character recognition result.

Description

    CROSS-REFERENCED APPLICATIONS
  • This application is a US national phase application under 35 U.S.C. § 371 of PCT/CN2016/096501 filed on Aug. 24, 2016, and titled TEXT INPUT METHOD AND DEVICE BASED ON GESTURE RECOGNITION AND STORAGE MEDIUM, which claims priority to Chinese Patent Application No. 201510593632.8 entitled “DEVICE AND METHOD FOR DISTINGUISHING STROKES IN AIR” and filed on Sep. 17, 2015, the contents of both being expressly incorporated by reference herein in its entirety.
  • TECHNICAL FIELD
  • The present disclosure is related to the computer recognition technology, and in particular related to a method, a device and a storage medium of inputting text based on hand gesture recognition.
  • BACKGROUND
  • With the development of computer science, more and more people are accustomed to effect the expression of text messages by a computer. Widely used inputting text means nowadays includes the keyboard, the touchscreen, the writing pad and so on, those means of inputting text have their respective disadvantages. Accordingly, designing a more naturalized, convenient and effective way of writing text in the air is important as a field of research.
  • However, as for a computer vision, the trace of a text writing in the air is consecutive, although the conventional text recognition technology is readily capable of recognizing the text consecutively written, the precision of recognition is yet to be improved.
  • SUMMARY
  • According to the various embodiments of the present disclosure, a method, a device and a storage medium of inputting text based on hand gesture recognition are provided.
  • A method of inputting text based on hand gesture recognition, the method includes:
  • collecting image information containing a target object;
  • determining the target object according to the image information, and acquiring a writing plane and a movement trace of the target object;
  • extracting the movement trace of the target object in the writing plane as a writing trace, outputting the writing trace; and
  • performing character recognition according to the writing trace, and outputting a character recognition result.
  • A device of inputting text based on hand gesture recognition, the device including a processor; and a memory storing instructions, which, when executed by the processor causes the processor to perform steps including:
  • collecting image information containing a target object;
  • determining the target object according to the image information, and acquiring a writing plane and a movement trace of the target object;
  • extracting the movement trace of the target object in the writing plane as a writing trace, outputting the writing trace; and
  • performing character recognition according to the writing trace, and outputting a character recognition result.
  • At least one non-transitory computer-readable storage medium storing computer-readable instructions that, when executed by at least one processors, cause the at least one processor to perform steps including:
  • collecting image information containing a target object;
  • determining the target object according to the image information, and acquiring a writing plane and a movement trace of the target object;
  • extracting the movement trace of the target object in the writing plane as a writing trace, outputting the writing trace; and
  • performing character recognition according to the writing trace, and outputting a character recognition result.
  • According to the foregoing method of inputting text based on hand gesture recognition, the target object is determined according to the image information, and the writing plane and the movement trace of the target object are acquired, the movement trace of the target object in the writing plane is extracted as the writing trace to be output, the character recognition is performed according to the writing trace, and the character recognition result is output, such that even if the movement trace of the target object in the air in consecutive, the trace extraction module can still extract the writing trace in the writing plane, which is equivalent to a strokes distinguishing process and facilitates the recognition of the character recognition module and improves the precision of recognition.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • To illustrate the technical solutions according to the embodiments of the present disclosure or in the prior art more clearly, the accompanying drawings for describing the embodiments or the prior art are introduced briefly in the following. Apparently, the accompanying drawings in the following description are only some embodiments of the present disclosure, and persons of ordinary skill in the art can derive other drawings from the accompanying drawings without creative efforts.
  • FIG. 1 is a schematic diagram of the device of inputting text based on hand gesture recognition according to an embodiment;
  • FIG. 2 is a block diagram of the device of inputting text based on hand gesture recognition according to an embodiment;
  • FIG. 3 is a block diagram of a target acquiring module according to the embodiment shown in FIG. 2;
  • FIG. 4 is a block diagram of a trace extraction module according to the embodiment shown in FIG. 2;
  • FIG. 5 is a schematic diagram of a writing in the air according to an embodiment;
  • FIG. 6 is a flowchart of a method of inputting text based on hand gesture recognition according to an embodiment;
  • FIG. 7 is a flowchart of a method of inputting text based on hand gesture recognition according to another embodiment; and
  • FIG. 8 is a schematic diagram of an application of a method of inputting text based on hand gesture recognition according to an embodiment.
  • DETAILED DESCRIPTION OF THE EMBODIMENTS
  • Embodiments of the disclosure are described more fully hereinafter with reference to the accompanying drawings. Preferable embodiments are presented in the drawings. The various embodiments of the disclosure may, however, be embodied in many different forms and should not be construed as limited to the embodiments set forth herein. Rather, these embodiments are provided so that this disclosure will be thorough and complete, and will fully convey the scope of the disclosure to those skilled in the art.
  • Unless otherwise defined, all terms used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this disclosure belongs. The terms used herein is for the purpose of describing the embodiments of the present disclosure, as opposed to limiting thereto. The language “and/or” used in the disclosure refers to any and all combinations of the one or multiple items listed.
  • FIG. 1 is a schematic diagram of the device of inputting text based on hand gesture recognition according to an embodiment.
  • In the present embodiment, the device of inputting text based on hand gesture recognition includes a processor, a storage medium, a Random Access Memory (RAM) and an input/output (I/O) interface connected through device bus. An operation device, a database and computer-readable instructions are stored in the storage medium of the device, the database is configured to store data, such as storing image information, movement trace and writing trace and so on. The processor of the device is configured to provide computation and control capability to support the operation of the entire device. The RAM of the terminal provides an operation environment for the computer-readable instructions in the storage medium. The I/O interface of the device is configured to connect other equipment.
  • FIG. 2 is a block diagram of the device of inputting text based on hand gesture recognition according to an embodiment.
  • The internal structure of the device is corresponding to the structure shown in FIG. 1, all or a part of each of the module below can be implemented by way of software, hardware or a combination thereof. As shown in FIG. 2, in an embodiment, the device of inputting text based on hand gesture recognition includes an image collection module 110, a target acquiring module 120, a trace extraction module 130, a character recognition module 140, and a display module 150. The image collection module 110 is configured to collect image information containing a target object. The target acquiring module 120 is configured to determine the target object according to the image information, and acquire a writing plane and a movement trace of the target object. The trace extraction module 130 is configured to extract the movement trace of the target object in the writing plane as a writing trace, outputting the writing trace. The character recognition module 140 is configured to perform character recognition according to the writing trace, and output a character recognition result.
  • The image collection module 110 is a camera that is configured to collect image information containing a target object. In an embodiment, the target object can be a finger of a user, or a stick-like object held by the user, such as a pencil.
  • For example, when the user needs to write in the air using the pencil as the target object, it is only required to hold the pencil and place the pencil in a visible area of the camera, then the image collection module 110 will collect the image information containing the pencil.
  • In an embodiment, the device of inputting text based on hand gesture recognition also includes a hand gesture recognition module and/or a voice recognition module. The hand gesture recognition module is configured to recognize the hand gesture of the user and output a hand gesture recognition result, and the hand gesture recognition result is to start or end a writing status. The hand gesture recognition module is configured to recognize the hand gesture of the user and output a hand gesture recognition result, and the hand gesture recognition result is to start or end a writing status.
  • In particular, when the user needs to start a writing status, a corresponding hand gesture can be made within the visible area of the image collection module 110. When the hand of the user appears in the visible area of the image collection module 110, the hand gesture recognition module will automatically adapt the visible area of the image collection module 110 for the collection of hand gesture images (such as focusing on the hand of the user and zooming in with the lens), and a hand gesture recognition result will be output by the hand gesture recognition module, i.e., starting writing status. It should be understood that the user can also start the writing status by way of voice directly, such as say “start writing” to a microphone for collecting voice, then the voice recognition module will output a voice recognition result, i.e., starting the writing status.
  • Correspondingly, when the user needs to end the writing status, a corresponding hand gesture can be made within the visible area of the image collection module 110, the hand gesture recognition module will output a hand gesture recognition result, i.e., ending the writing status. It should be understood that the user can also end the writing status by way of voice directly, such as say “end writing” to the microphone for collecting voice, then the voice recognition module will output a voice recognition result, i.e., ending the writing status.
  • The target acquiring module 120 is configured to determine the target object according to the image information, and acquire a writing plane and a movement trace of the target object.
  • Referring to FIG. 3, in an embodiment, the target acquiring module 120 includes an image processing unit 122, a time calculation unit 124, and a writing plane determination unit 126.
  • In particular, the image processing unit 122 is configured to perform image characteristic extraction to the image information to determine the target object. The image characteristics herein mainly refer to acquiring image contours of the digital image by an edge extraction. Typically, a place where the gray scale value changes drastically is defined an edge. If the target object is a finger or a pencil or other rod-like object, then the edge of the image is a very critical characteristic. By extracting the image characteristics, the target object can be determined swiftly.
  • The time calculation unit 124 is configured to count a pause duration of the target object at a position. In particular, when the target object is determined, a tracking of the target object can be achieved. When the target object is static, the time calculating unit 124 will start timing which is stopped when the target object starts moving again, the duration elapsed therebetween is the pause duration of the target object at the position.
  • The writing plane determination unit 126 is configured to acquire a first position where the pause duration of the target object reaches a preset duration for a first time after starting the writing status, and acquire a second position where the pause duration of the target object reaches the preset duration for a second time; and defining the movement trace of the target object from the first position to the second position as a first stroke, defining a plane where the first stroke is located as a reference writing plane, and defining a plane away from the reference writing plane within a preset distance range as a writing plane.
  • For example, when the writing status is started, the first position where the pause duration of the target object reaches the preset duration for the first time is defined A, the second position where the pause duration of the target object reaches the preset duration for the second time is defined B, then a connecting line between A and B constitutes a reference writing plane. It should be understood that the movement trace from target objects A to B is the very first stroke of the character the user input. The determination of the reference writing plane, in fact is in combination with the spatial position where the first stroke located, and the writing range of the human arm.
  • In addition, it should be understood that, in alternative embodiments, the writing plane determination unit 126 can also use other manners to define the reference writing plane, such as by a particular hand gesture to start the writing status, and regard the plane where one end for writing of the target object located as the reference writing plane. For another example, the writing status is started by voice, then the target object stay for a period of time at the position about to write, and the plane where the target object stayed is defined the reference writing plane. For a third example, the writing status is started by voice, gesture, remote control or any other means, the target object moves forward or downward rectilinearly at a certain speed to a position and stay for a while, the plane where such position located is defined the reference writing plane. For a fourth example, the trace of the target object in the air turned and stayed in a position after an “L”-like trace is drew, the plane where such a position is drew is defined the reference writing plane. The determination of whether an “L”-like trace is drew by the target object is mainly to track the target object to acquire the movement trace of the target object, and then determine if the movement trace is “L”-like. The reference writing plane can be defined by a combination of the foregoing manners.
  • After the reference writing plane is determined, in view that it is impossible that every stroke fall into a same plane during writing in the air, i.e., every movement trace of the target object will have a respective intersection angle in relation to the reference writing plane, the plane whose intersection angle in relation to the writing plane is within a preset angle range (e.g., within 30 degrees) is defined as the writing plane, which equivalent to regarding the writing plane as a set of planes, and the movement traces in all of the planes in the set are then extracted, so as to facilitate the improvement of the precision of character recognition.
  • In an embodiment, the preset duration is two seconds, i.e., the user will control the target object to pause over two seconds both in starting and ending the stroke. It should be understood that, the preset duration can be set as one second, two seconds, or three seconds etc., no limitation is intended herein.
  • The trace extraction module 130 is configured to extract the movement trace of the target object in the writing plane as a writing trace, outputting the writing trace.
  • Referring to FIG. 4, in an embodiment, the trace extraction module 130 includes a trace extraction unit 132 and a character output unit 134.
  • The trace extraction unit 132 is configured to extract the movement trace of the target object in the writing plane as a writing trace.
  • The character output unit 134 is configured to output the writing trace with the characters as the unit. In particular, in the process of writing by the user, the target object pause for a certain period of time upon completing of the input of a character. In an embodiment, assuming the threshold of interval to be four seconds, when the pause duration of the target object exceeds four seconds, it is considered that the user has completed the input of a character. The character can be the Chinese character, the letter, the digit, the symbol, or other individual characters input by the user.
  • The character recognition module 140 is configured to perform character recognition according to the writing trace, and output a character recognition result. In particular, the character recognition module 140 will compare the writing trace (i.e., the output character) output by the trace extraction module 130 to the character in the preset character base and find out from the character base a character with the highest similarity to the writing trace as the character recognition result to output.
  • The display module 150 is configured to display the character recognition result and/or the writing trace. Referring to FIG. 5 for details, the display module 150 includes a writing display zone 152 and a recognition zone 154. The writing display zone 152 will display the writing trace output by the trace extraction module 130, the recognition display zone 154 will display the character recognition result output by the character recognition module 140.
  • In an embodiment, the display module 150 displays the writing trace in real time, and when the writing trace is output with the characters as the unit to the character recognition module 140, the display module 150 will then display the character recognition result output by the character recognition module 140.
  • In particular, the display module 150 includes a target object detection unit and a display unit. The target object detection unit is configured to detect if the target object is departing the writing plane in an angle between 45 and 90 degrees in relation to the reference writing plane, or to detect if the target object is departing the writing plane at a preset speed; if yes, then the display unit stops displaying the writing trace. The target object detection unit is further configured to detect if the target object is arriving the writing plane in the angle between 45 and 90 degrees in relation to the reference writing plane, or to detect if the target object is arriving the writing plane at the preset speed, or detecting if the target object is arriving the writing plane in an “L”-like movement trace; if yes, the target object detection unit starts displaying the writing trace. To determine if an “L”-like trace is drew by the target object, it is mainly to track the target object to acquire the movement trace of the target object, and then determine if the movement trace is “L”-like.
  • Additionally, a size adjustment module is included, which is configured to adjust the size of the character according to a distance between one end of the target object proximately to the reference writing plane and the reference writing plane.
  • In particular, the end of the target object near the reference writing plane can be understood as the pen point, the farther the pen point away from the reference writing plane, the thinner the stroke of the character. In contrast, the nearer the pen point away from the reference writing plane, the thicker the stroke of the character.
  • Furthermore, if a stroke that the target object arrives at the reference writing plane in an angle close to 90 degrees and at a speed close to a speed threshold and stayed for a certain period of time is defined to be a “down stroke”; and a stroke that departs from the reference writing plane in an angle close to 90 degrees and at a speed close to the speed threshold is defined to be a “up stroke”, then the size adjustment module can also be configured to effect an adjustment of the displayed character from being thin to being thick according to the extension of the pause in the first down stroke. Also, the farther the target object is away from the reference writing plane, the thinner the stroke at the up stroke. When the distance target object is away from the reference writing plane amounts to a certain value, it is considered the “up stroke” is finished and a process of distinguishing the strokes is performed.
  • Alternatively, in an embodiment, it can be defined a “down stroke” that the target object arrives the writing plane in an “L”-like movement trace, and a “up stroke” that the target object departs the writing plane in an “L”-like movement trace. In particular, after approaching the writing plane, it takes a sharp turn at near 90 degrees, and starts moving in the writing plane, then it is determined that it is an “L”-like movement trace that arrived at the writing plane; in the opposite, the target object, while moving in the writing plane, takes a sharp turn at near 90 degrees in a sudden and leaves the writing plane, then it is determined it is an “L”-like movement trace that departed from the writing plane.
  • FIG. 6 is a flowchart of a method of inputting text based on hand gesture recognition according to an embodiment.
  • A method of inputting text based on hand gesture recognition includes:
  • In step S110, image information containing a target object is collected;
  • In an embodiment, the target object can be a finger of a user, or a stick-like object held by the user, such as a pencil.
  • In step S120, the target object is determined according to the image information, and a writing plane and a movement trace of the target object are acquired.
  • In an embodiment, in step S120, a size of the character is adjusted according to a distance between one end of the target object proximately to the reference writing plane and the reference writing plane.
  • In step S130: the movement trace of the target object in the writing plane is extracted as the writing trace, the writing trace is output.
  • In an embodiment, the output is performed with the characters as the unit in step S130.
  • In step S140, character recognition is performed according to the writing trace, and a character recognition result is output.
  • FIG. 7 is a flowchart of a method of inputting text based on hand gesture recognition according to another embodiment.
  • A method of inputting text based on hand gesture recognition, the method includes:
  • In step S210: a hand gesture of a user is recognized and a writing status is started according to a hand gesture recognition result.
  • In an embodiment, step S210 can also be that a voice of a user is recognized and the writing status is started according to a voice recognition result.
  • In step S220: image information containing a target object is collected.
  • In step S230: the target object is determined according to the image information, and a writing plane and a movement trace of the target object are acquired.
  • In step S240: the movement trace of the target object in the writing plane is extracted as the writing trace, the writing trace is output.
  • In step S250, character recognition is performed according to the writing trace, and a character recognition result is output.
  • In step S260, the character recognition result and the writing trace are displayed.
  • In particular, the writing trace is first displayed and then the character recognition result is displayed, it is because the writing trace is displayed in real time. In an embodiment, it can only display the character recognition result or the writing trace.
  • In an embodiment, the step of displaying the writing trace in step S260 includes:
  • It is detected if the target object is departing the writing plane in an angle between 45 and 90 degrees in relation to the reference writing plane, or it is detected if the target object is departing the writing plane at a preset speed; if yes, stop displaying the writing trace; and
  • It is detected if the target object is arriving the writing plane in the angle between 45 and 90 degrees in relation to the reference writing plane, or it is detected if the target object is arriving the writing plane at the preset speed, or detecting if the target object is arriving the writing plane in an “L”-like movement trace; if the target object is arriving the writing plane in the angle between 45 and 90 degrees in relation to the reference writing plane or the target object is departing the writing plane at the preset speed or the target object is arriving the writing plane in the “L”-like movement trace, start displaying the writing trace.
  • In step S270: a hand gesture of a user is recognized and a writing status is ended according to a hand gesture recognition result.
  • In an embodiment, step S270 can also be that a voice of a user is recognized and the writing status is ended according to a voice recognition result.
  • The principle of the method of inputting text based on hand gesture recognition in the foregoing embodiment will be described through specific an application scenario where, as an example, the target object is a finger that writes a Chinese character “
    Figure US20180260034A1-20180913-P00001
    ” in the air.
  • The user shows his/her finger in the visible range of the camera, and starts the writing status by voice. As illustrated in FIG. 8, it is the down stroke the finger of the user moving from position “0” to the first stroke starting point position “1”, from position “1” to position “2” is a first stroke, and the plane where the first stroke locates is defined the writing plane. From position “2” to position “3” is an up stroke, moving from position “3” to position “4” is a movement outside the writing plane. From position “4” to position “5” is a second down stroke, and from position “5” to position “6” is a second stroke. From position “6” to position “7” is the up stroke, moving from position “7” to position “8” is a movement outside the writing plane. From position “8” to position “9” is a third stroke, from position “9” to position “10” is a third stroke, from position “10” to position “11” is the up stroke, and the input of a character is completed. The finger of the user will make a pause at position “1”, position “2”, position “5”, position “6”, position “9” and position “10”, although the movement trace of the finger of the user from position “0” to position “11” is consecutive, yet as only the movement trace in the writing plane is extracted as writing trace, this amounts to a strokes distinguishing process, which can improve the recognition precision of the input character, facilitates the user to perform inputting text, and helps to train the kids with good a writing manner.
  • According to the foregoing method of inputting text based on hand gesture recognition, the target object is determined according to the image information, and the writing plane and the movement trace of the target object are acquired, the movement trace of the target object in the writing plane is extracted as the writing trace to be output, the character recognition is performed according to the writing trace, and the character recognition result is output, such that even if the movement trace of the target object in the air in consecutive, yet the trace extraction module will only extract the writing trace in the writing plane, which is equivalent to a strokes distinguishing process and facilitates the recognition of the character recognition module and improves the precision of recognition.
  • In an embodiment, at least one non-transitory computer-readable storage medium is also provided, which stores computer-readable instructions that, when executed by at least one processors, cause the at least one processor to perform steps as follow:
  • Image information containing a target object is collected;
  • The target object is determined according to the image information, and a writing plane and a movement trace of the target object are acquired;
  • The movement trace of the target object in the writing plane is extracted as a writing trace, outputting the writing trace; and
  • Character recognition is performed according to the writing trace, and a character recognition result is output.
  • A person skilled in the art should understand that the processes of the methods in the above embodiments can be, in full or in part, implemented by computer programs instructing underlying hardware, the programs can be stored in a computer-readable storage medium, the program can include the processes in the embodiments of the various methods when it is being executed. The storage medium can be a disk, a CD, a Read-Only Memory (ROM) and other non-transitory storage mediums or Random Access Memory (RAM) and so on.
  • The different technical features of the above embodiments can have various combinations which are not described for the purpose of brevity. Nevertheless, to the extent the combining of the different technical features do not conflict with each other, all such combinations must be regarded as being within the scope of the disclosure.
  • The foregoing implementations are merely specific embodiments of the present disclosure, and are not intended to limit the protection scope of the present disclosure. It should be noted that any variation or replacement readily figured out by persons skilled in the art within the technical scope disclosed in the present disclosure shall all fall into the protection scope of the present disclosure. Therefore, the protection scope of the present disclosure shall be subject to the protection scope of the claims.

Claims (13)

What is claimed is:
1. A method of inputting text based on hand gesture recognition, comprising:
collecting image information containing a target object;
determining the target object according to the image information, and acquiring a writing plane and a movement trace of the target object;
extracting the movement trace of the target object in the writing plane as a writing trace, outputting the writing trace; and
performing character recognition according to the writing trace, and outputting a character recognition result.
2. The method of claim 1, further comprising:
recognizing a hand gesture or a voice of a user, starting a writing status according to a hand gesture recognition result or a voice recognition result; and
recognizing a hand gesture or a voice of the user, ending the writing status according to a hand gesture recognition result or a voice recognition result;
3. The method according to claim 2, wherein the step of determining the target object according to the image information, and acquiring the writing plane and the movement trace of the target object comprises:
performing image characteristic extraction to the image information to determine the target object;
counting a pause duration of the target object at a position;
acquiring a first position where the pause duration of the target object reaches a preset duration for a first time after starting the writing status, and acquiring a second position where the pause duration of the target object reaches the preset duration for a second time; and defining the movement trace of the target object from the first position to the second position as a first stroke, defining a plane where the first stroke is located as a reference writing plane, and defining a plane away from the reference writing plane within a preset distance range as a writing plane;
wherein the step of extracting the movement trace of the target object in the writing plane as the writing trace, outputting the writing trace comprises:
extracting the movement trace of the target object in the writing plane as the writing trace; and
outputting the writing trace with a character as an unit.
4. The method of claim 3, further comprising:
adjusting a size of the character according to a distance between one end of the target object proximately to the reference writing plane and the reference writing plane.
5. The method of claim 1, wherein after the step of recognizing the character according to the writing trace and outputting a character recognition result, the method further comprises:
displaying the character recognition result and/or the writing trace.
6. The method of claim 5, wherein the step of displaying the writing trace comprises:
detecting if the target object is departing the writing plane in an angle between 45 and 90 degrees in relation to the reference writing plane, or detecting if the target object is departing the writing plane at a preset speed; if the target object is departing the writing plane in the angle between 45 and 90 degrees in relation to the reference writing plane or the target object is departing the writing plane at the preset speed, stop displaying the writing trace; and
detecting if the target object is arriving the writing plane in the angle between 45 and 90 degrees in relation to the reference writing plane, or detecting if the target object is arriving the writing plane at the preset speed, or detecting if the target object is arriving the writing plane in an “L”-like movement trace; if the target object is arriving the writing plane in the angle between 45 and 90 degrees in relation to the reference writing plane or the target object is departing the writing plane at the preset speed or the target object is arriving the writing plane in the “L”-like movement trace, start displaying the writing trace.
7. A device of inputting text based on hand gesture recognition, the device comprising a processor; and a memory storing instructions, which, when executed by the processor causes the processor to perform steps comprising:
collecting image information containing a target object;
determining the target object according to the image information, and acquiring a writing plane and a movement trace of the target object;
extracting the movement trace of the target object in the writing plane as a writing trace, outputting the writing trace; and
performing character recognition according to the writing trace, and outputting a character recognition result.
8. The device of claim 7, wherein the processor further executes the instructions to perform steps of:
recognizing a hand gesture or a voice of a user and outputting a hand gesture recognition result or a voice recognition result starting a writing status according to; and
recognizing a hand gesture or a voice of the user and outputting a hand gesture recognition result or a voice recognition result ending a writing status according to.
9. The device according to claim 8, wherein when the instructions are executed by the processor, the step of determining the target object according to the image information, and acquiring the writing plane and the movement trace of the target object comprises:
performing image characteristic extraction to the image information to determine the target object;
counting a pause duration of the target object at a position;
acquiring a first position where the pause duration of the target object reaches a preset duration for a first time after starting the writing status, and acquiring a second position where the pause duration of the target object reaches the preset duration for a second time; and defining the movement trace of the target object from the first position to the second position as a first stroke, defining a plane where the first stroke is located as a reference writing plane, and defining a plane away from the reference writing plane within a preset distance range as a writing plane;
the step of extracting the movement trace of the target object in the writing plane as the writing trace, outputting the writing trace comprises:
extracting the movement trace of the target object in the writing plane as the writing trace; and
outputting the writing trace with a character as a unit.
10. The device of claim 9, wherein the processor further executes the instructions to perform steps of:
adjusting a size of the character according to a distance between one end of the target object proximately to the reference writing plane and the reference writing plane.
11. The device of claim 7, wherein, after the step of recognizing the character according to the writing trace and outputting the character recognition result, the processor further executes the instructions to perform steps of:
displaying the character recognition result and/or the writing trace.
12. The device of claim 11, wherein when the instructions are executed, the step of displaying the character recognition result and/or the writing trace comprises:
detecting if the target object is departing the writing plane in an angle between 45 and 90 degrees in relation to the reference writing plane, or detecting if the target object is departing the writing plane at a preset speed; if the target object is departing the writing plane in the angle between 45 and 90 degrees in relation to the reference writing plane or the target object is departing the writing plane at the preset speed, stop displaying the writing trace; and
detecting if the target object is arriving the writing plane in the angle between 45 and 90 degrees in relation to the reference writing plane, or detecting if the target object is arriving the writing plane at the preset speed, or detecting if the target object is arriving the writing plane in an “L”-like movement trace; if the target object is arriving the writing plane in the angle between 45 and 90 degrees in relation to the reference writing plane or the target object is departing the writing plane at the preset speed or the target object is arriving the writing plane in the “L”-like movement trace, start displaying the writing trace.
13. At least one non-transitory computer-readable storage medium storing computer-readable instructions that, when executed by at least one processors, cause the at least one processor to perform steps comprising:
collecting image information containing a target object;
determining the target object according to the image information, and acquiring a writing plane and a movement trace of the target object;
extracting the movement trace of the target object in the writing plane as a writing trace, outputting the writing trace; and
performing character recognition according to the writing trace, and outputting a character recognition result.
US15/761,034 2015-09-17 2016-08-24 Text input method and device based on gesture recognition, and storage medium Active 2037-07-09 US10725552B2 (en)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
CN201510593632.8A CN105302298B (en) 2015-09-17 2015-09-17 Sky-writing breaks a system and method
CN201510593632.8 2015-09-17
CN201510593632 2015-09-17
PCT/CN2016/096501 WO2017045517A1 (en) 2015-09-17 2016-08-24 Text input method and device based on gesture recognition, and storage medium

Publications (2)

Publication Number Publication Date
US20180260034A1 true US20180260034A1 (en) 2018-09-13
US10725552B2 US10725552B2 (en) 2020-07-28

Family

ID=55199650

Family Applications (1)

Application Number Title Priority Date Filing Date
US15/761,034 Active 2037-07-09 US10725552B2 (en) 2015-09-17 2016-08-24 Text input method and device based on gesture recognition, and storage medium

Country Status (5)

Country Link
US (1) US10725552B2 (en)
EP (1) EP3340014A4 (en)
JP (1) JP6695963B2 (en)
CN (1) CN105302298B (en)
WO (1) WO2017045517A1 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109409361A (en) * 2018-09-26 2019-03-01 广东小天才科技有限公司 Searching method, pen, storage medium and search system based on pen
US20190235643A1 (en) * 2016-06-28 2019-08-01 Nikon Corporation Control device, display device, program, and detection method
CN113093913A (en) * 2021-04-20 2021-07-09 北京乐学帮网络技术有限公司 Test question processing method and device, electronic equipment and storage medium

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105302298B (en) * 2015-09-17 2017-05-31 深圳市国华识别科技开发有限公司 Sky-writing breaks a system and method
CN106125954B (en) * 2016-06-13 2018-07-20 北京小米移动软件有限公司 Input method based on terminal movement locus and device
WO2018018624A1 (en) * 2016-07-29 2018-02-01 华为技术有限公司 Gesture input method for wearable device, and wearable device
CN107562203A (en) * 2017-09-14 2018-01-09 北京奇艺世纪科技有限公司 A kind of input method and device
CN111553326B (en) * 2020-05-29 2023-04-18 上海依图网络科技有限公司 Hand motion recognition method and device, electronic equipment and storage medium
CN113391703A (en) * 2021-06-16 2021-09-14 咙咙信息技术(沈阳)有限公司 System for operating air writing based on media application

Family Cites Families (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090183929A1 (en) * 2005-06-08 2009-07-23 Guanglie Zhang Writing system with camera
TW200816062A (en) * 2006-09-29 2008-04-01 Ibm Video-based handwriting input method and apparatus
US7929002B2 (en) * 2007-08-10 2011-04-19 Hewlett-Packard Development Company, L.P. System and method of projecting an image on a physical substrate using a virtual image and a virtual plane
CN101354608A (en) * 2008-09-04 2009-01-28 中兴通讯股份有限公司 Method and system for implementing video input
CN101739118A (en) * 2008-11-06 2010-06-16 大同大学 Video handwriting character inputting device and method thereof
JP2010146266A (en) * 2008-12-18 2010-07-01 Seiko Epson Corp Display device and program
US11012732B2 (en) 2009-06-25 2021-05-18 DISH Technologies L.L.C. Voice enabled media presentation systems and methods
CN102279668B (en) 2010-06-08 2014-12-10 株式会社理光 Information interaction method and system
KR101262700B1 (en) * 2011-08-05 2013-05-08 삼성전자주식회사 Method for Controlling Electronic Apparatus based on Voice Recognition and Motion Recognition, and Electric Apparatus thereof
CN102360263A (en) 2011-09-26 2012-02-22 中兴通讯股份有限公司 Method implemented by taking three-dimensional moving track as input and mobile terminal
JP5866949B2 (en) * 2011-10-11 2016-02-24 大日本印刷株式会社 Electronic pen system and program
US10591998B2 (en) * 2012-10-03 2020-03-17 Rakuten, Inc. User interface device, user interface method, program, and computer-readable information storage medium
US9741085B2 (en) * 2013-03-14 2017-08-22 Artificial Intelligence Research Group Limited System and method of encoding content and an image
CN103226388B (en) * 2013-04-07 2016-05-04 华南理工大学 A kind of handwriting sckeme based on Kinect
JP2016525235A (en) 2013-06-25 2016-08-22 トムソン ライセンシングThomson Licensing Method and device for character input
US10691878B2 (en) * 2014-02-28 2020-06-23 Ricoh Co., Ltd. Presenting associations of strokes with content
CN105302298B (en) 2015-09-17 2017-05-31 深圳市国华识别科技开发有限公司 Sky-writing breaks a system and method

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20190235643A1 (en) * 2016-06-28 2019-08-01 Nikon Corporation Control device, display device, program, and detection method
US11635827B2 (en) * 2016-06-28 2023-04-25 Nikon Corporation Control device, display device, program, and detection method
CN109409361A (en) * 2018-09-26 2019-03-01 广东小天才科技有限公司 Searching method, pen, storage medium and search system based on pen
CN113093913A (en) * 2021-04-20 2021-07-09 北京乐学帮网络技术有限公司 Test question processing method and device, electronic equipment and storage medium

Also Published As

Publication number Publication date
JP6695963B2 (en) 2020-05-20
CN105302298B (en) 2017-05-31
WO2017045517A1 (en) 2017-03-23
JP2018535469A (en) 2018-11-29
EP3340014A4 (en) 2019-04-24
EP3340014A1 (en) 2018-06-27
US10725552B2 (en) 2020-07-28
CN105302298A (en) 2016-02-03

Similar Documents

Publication Publication Date Title
US10725552B2 (en) Text input method and device based on gesture recognition, and storage medium
US8970696B2 (en) Hand and indicating-point positioning method and hand gesture determining method used in human-computer interaction system
EP3258423B1 (en) Handwriting recognition method and apparatus
US9953216B2 (en) Systems and methods for performing actions in response to user gestures in captured images
KR102460737B1 (en) Method, apparatus, apparatus and computer readable storage medium for public handwriting recognition
WO2020215565A1 (en) Hand image segmentation method and apparatus, and computer device
TWI659331B (en) Screen capture method and device for smart terminal
WO2022166243A1 (en) Method, apparatus and system for detecting and identifying pinching gesture
CN105205454A (en) System and method for capturing target object automatically
US20170156589A1 (en) Method of identification based on smart glasses
CN106774850B (en) Mobile terminal and interaction control method thereof
US10133945B2 (en) Sketch misrecognition correction system based on eye gaze monitoring
US10395101B2 (en) Interest degree determination device, interest Degree determination method, and non-transitory computer-readable recording medium
US10031619B2 (en) Probabilistic palm rejection using spatiotemporal touch features and iterative classification
Aggarwal et al. Online handwriting recognition using depth sensors
KR20150106823A (en) Gesture recognition apparatus and control method of gesture recognition apparatus
CN112749646A (en) Interactive point-reading system based on gesture recognition
CN107450717B (en) Information processing method and wearable device
CN114821753B (en) Eye movement interaction system based on visual image information
CN110858291A (en) Character segmentation method and device
US9405375B2 (en) Translation and scale invariant features for gesture recognition
CN111432131B (en) Photographing frame selection method and device, electronic equipment and storage medium
CN110390281B (en) Sign language recognition system based on sensing equipment and working method thereof
Wong et al. Virtual touchpad: Hand gesture recognition for smartphone with depth camera
CN113449652A (en) Positioning method and device based on biological feature recognition

Legal Events

Date Code Title Description
FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY

FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO SMALL (ORIGINAL EVENT CODE: SMAL); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

AS Assignment

Owner name: SHENZHEN PRTEK CO. LTD., CHINA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:LIU, GUOHUA;REEL/FRAME:052364/0500

Effective date: 20180315

STPP Information on status: patent application and granting procedure in general

Free format text: AWAITING TC RESP., ISSUE FEE NOT PAID

STPP Information on status: patent application and granting procedure in general

Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED

STCF Information on status: patent grant

Free format text: PATENTED CASE

FEPP Fee payment procedure

Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY