WO2013075466A1 - Character input method, device and terminal based on image sensing module - Google Patents

Character input method, device and terminal based on image sensing module Download PDF

Info

Publication number
WO2013075466A1
WO2013075466A1 PCT/CN2012/075103 CN2012075103W WO2013075466A1 WO 2013075466 A1 WO2013075466 A1 WO 2013075466A1 CN 2012075103 W CN2012075103 W CN 2012075103W WO 2013075466 A1 WO2013075466 A1 WO 2013075466A1
Authority
WO
WIPO (PCT)
Prior art keywords
fingertip
information
image
running track
character
Prior art date
Application number
PCT/CN2012/075103
Other languages
French (fr)
Chinese (zh)
Inventor
辛静
Original Assignee
中兴通讯股份有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 中兴通讯股份有限公司 filed Critical 中兴通讯股份有限公司
Publication of WO2013075466A1 publication Critical patent/WO2013075466A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/017Gesture based interaction, e.g. based on a set of recognized hand gestures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/32Digital ink

Definitions

  • the present invention relates to the field of mobile terminals, and in particular, to a character input method, apparatus, and terminal based on an image sensing module. Background technique
  • mobile communication is undergoing rapid development, and thus, the functional requirements for human-computer interaction functions of mobile terminals (such as mobile phones) are becoming higher and higher.
  • mobile terminals such as mobile phones
  • Text editing and input technology has dramatically changed the way people communicate since its birth. With this technology, people can edit text messages, enter characters, and more. At present, there are many ways to input characters, but in view of the fact that many character input methods have more or less limitations or defects.
  • the input of characters is edited by means of a button.
  • a button that does not stop with a thumb.
  • the user needs to repeatedly repeat the button action, and the input efficiency is low, and Due to long-term button presses, even when it is serious, it may cause health problems similar to "threbitis", which greatly affects the health of users.
  • the current handheld terminals are equipped with a resistive screen or a capacitive screen
  • writing is performed on the screen of the terminal by means of a finger or a stylus, by pressure or capacitive sensing between the finger or the stylus and the screen.
  • the method requires the user's terminal to configure a resistive screen or a capacitive screen, and a mobile phone without a resistive screen or a capacitive screen cannot use the method for character input, so the method has certain limitations.
  • some handheld terminals can control the sensor to sense various actions made by the user to achieve the purpose of character input.
  • the target tracking is performed by the sensor, and the user is detected.
  • the character input is performed by the action; or the target tracking is performed in the case where the user uses the highlighter of the special device, and the action made by the user is detected to achieve the same purpose of character input.
  • these methods are used to implement character input, it is necessary to make these portable terminals additionally configure corresponding hardware, which makes the cost higher, and requires additional equipment when performing character input, thereby making the user's use extremely inconvenient. .
  • the main object of the present invention is to provide a character input method, device and terminal based on an image sensing module, which can meet the personalized requirements of the user for the handheld terminal, and at the same time realize the fast and accurate character. Input.
  • the present invention adopts the following technical solutions:
  • a character input method based on an image sensing module comprising:
  • the pre-trained character model library is queried according to the running track information, and the running track information is converted into corresponding characters.
  • the character input method based on the image sensing module further includes:
  • Tracking of the fingertip of the object is performed according to the gesture contour model.
  • the method also includes:
  • Binarizing the acquired object gesture contour image information and/or,
  • the tracking of the fingertip of the object according to the gesture contour model is: mapping a rectangular region of the fingertip coarsely positioned of the reference gesture contour model to the extracted object gesture contour image, and obtaining the finger of the object gesture contour image The rectangular area where the pointed position is located;
  • the finger contour segment with the largest edge curvature is obtained as the fingertip of the object.
  • the running track information of the fingertip of the acquiring object is:
  • the ( ) refers to the real-time coordinate information of the fingertip of the object at the time t in real time.
  • the character model with the largest likelihood value is obtained as the target character model, wherein the likelihood value of each character model is calculated by using the Verbit algorithm;
  • a character input device based on an image sensing module the device comprising:
  • An image sensing module configured to collect a video frame including a fingertip of the object
  • the object fingertip running track information acquiring module is configured to acquire the running track information of the fingertip of the object through the video frame collected by the image sensing module;
  • the character conversion module is configured to query the pre-trained character model library according to the running track information, and convert the running track information into corresponding characters.
  • the character input module based on the image sensing module further includes:
  • An object gesture contour image information collecting module configured to collect object hand contour image information by using an image sensing module
  • a gesture contour model obtaining module configured to query a pre-trained gesture contour model library according to the object gesture contour image information, and obtain a matching gesture contour model, so that the object fingertip running track information acquiring module performs the object according to the gesture contour model Tracking of the fingertips, obtaining the running track information of the fingertips of the object.
  • the character input module based on the image sensing module further includes:
  • the image processing module is configured to process the object gesture contour image information collected by the object gesture contour image information collection module as follows:
  • the object fingertip running track information acquiring module performs tracking of the fingertip of the object according to the gesture contour model:
  • the information about the running track of the fingertip of the object obtained by the object fingertip trajectory information acquiring module by the image sensing module is:
  • the character conversion module queries the pre-trained character model library according to the running track information, so as to convert the running track information into corresponding characters:
  • the character model with the largest likelihood value is obtained as the target character model, wherein the likelihood value of each character model is calculated by using the Verbit algorithm;
  • the running track information is converted into corresponding characters according to the target character model.
  • a terminal comprising the image sensor module based character input device, the device comprising:
  • An image sensing module configured to collect a video frame including a fingertip of the object
  • the object fingertip running track information acquiring module is configured to acquire the running track information of the fingertip of the object through the video frame collected by the image sensing module;
  • the character conversion module is configured to query the pre-trained character model library according to the running track information, and convert the running track information into corresponding characters.
  • the image sensing module provided by the present invention is based on The character input method, device and terminal of the block rely on the image sensing module (such as the camera device) provided by the terminal to collect images, extract the gesture contour, match the hand contour with the existing gesture contour model, and recognize the gesture model to the object.
  • the fingertip is coarsely positioned, and then the fingertip of the finger is accurately positioned according to the bending rate of the contour of the finger. Predict the approximate position of the fingertip at the next moment, capture the movement trajectory of the fingertip, calculate the tangential angle at different moments, and accumulate the tangential angle changes over a period of time to obtain the trajectory of the fingertip of the object during the time period.
  • the obtained object fingertip trajectory is matched with the pre-stored character model library to generate corresponding characters.
  • the image input module-based character input method, device and terminal provided by the invention can meet the personalized requirements of the user for the handheld terminal, and at the same time realize fast and accurate input of characters.
  • FIG. 1 is a schematic flow chart of a character input method based on an image sensing module according to an embodiment of the present invention
  • FIG. 2 is a detailed flow chart of a character input method based on an image sensing module according to a preferred embodiment of the present invention
  • FIG. 3 is a schematic structural diagram of a character input device based on an image sensing module according to an embodiment of the present invention. detailed description
  • the basic idea of the present invention is: acquiring the running track information of the fingertip of the object; querying the pre-trained character model library according to the running track information, and converting the running track information into corresponding characters.
  • a character input method based on an image sensing module includes the following steps:
  • the image sensing module is configured to collect fingertip screen information of a user's finger, and extract user's fingertip running track information according to the collected fingertip screen information, for example, the image sensing module.
  • a camera device that is configured for both regular mobile phones and smartphones.
  • the method before performing all the steps, further includes the following steps: S1001: Acquire an object gesture contour image information, such as the object, by using an image sensing module.
  • Gesture contour image information includes but is not limited to the following: Only the index finger is straight, other fingers are clenched; only the middle finger is straight, other fingers are clenched; the thumb and forefinger are straight at the same time, other fingers are clenched.
  • S1002 Query a pre-trained gesture contour model library according to the object gesture contour image information, and obtain a matching gesture contour model.
  • the method before querying the pre-trained gesture contour model library according to the object gesture contour image information to obtain a matching gesture contour model, the method further includes:
  • the grayscale image has only the luminance information of the spatial sampling point, it can be represented by a numerical value.
  • the content is richer, but the presentation is more complicated.
  • the object gesture contour image information of the color image can be binarized and converted into a grayscale image, which reduces complexity and reduces memory resources.
  • image sharpening processing can be performed on the object gesture contour image information, and the present invention uses the Roberts gradient sharpening method to enhance the image edge. In addition to this, in order to reduce noise, it is also possible to perform image smoothing processing on the subject gesture profile image information.
  • the step of tracking the fingertip of the object according to the gesture contour model includes:
  • S10031 Map a rectangular area of the fingertip coarsely positioned of the reference gesture contour model to the extracted object gesture contour image, and obtain a rectangular area where the fingertip position of the target gesture contour image is located.
  • the method for acquiring the running track information of the fingertip of the object by the image sensing module specifically includes the following steps:
  • the ( ⁇ , refers to the real-time coordinate information of the fingertip of the object at real time t.
  • the method for querying the pre-trained character model library according to the running track information to convert the running track information into corresponding characters includes:
  • the character model with the largest likelihood value is the target character model, wherein the likelihood value of each character model is calculated by the Verbit algorithm.
  • FIG. 2 is a schematic flowchart of a character input method based on an image sensing module according to a preferred embodiment of the present invention. The process specifically includes the following steps:
  • the selection of the gesture model directly affects the recognition effect.
  • the present invention selects a common gesture model as a reference model, for example, but not limited to the following: Only the index finger is straight, and other fingers are clenched; Only the middle finger is straight, the other fingers are clenched; the thumb and forefinger are straight at the same time, and other fingers are clenched.
  • the initial position of the fingertip is predefined. For example, in the case where only the index finger is straight, the initial position of the fingertip is defined at the top of the index finger, and the fingertip position is included with a rectangular frame of a certain size, which is the area where the subsequent fingertip is accurately positioned.
  • Video images are typically analyzed using inter-frame or intra-frame information.
  • the spatial information such as the color and brightness of the object should be combined for video segmentation.
  • the embodiment of the present invention adopts a time-space joint segmentation method, and comprehensively utilizes inter-frame motion information in a time domain and spatial skin color and luminance information, and simultaneously performs time and space segmentation methods to extract an accurate gesture edge.
  • the spatial domain segmentation the initial segmentation region with accurate semantics is obtained, and the motion region of the image is obtained by time domain segmentation, and the discontinuous edge is connected to obtain the gesture contour.
  • the grayscale image has only the luminance information of the spatial sampling point, and thus can be represented by a numerical value.
  • the content is richer, but the representation is more complicated.
  • the color image needs to be binarized and converted into a grayscale image, which reduces complexity and reduces memory resources.
  • the present invention uses the Roberts gradient sharpening method to enhance the edges of the image.
  • the binary image can be smoothed.
  • the gesture contour is matched with the gesture contour model.
  • the extracted gesture contour is matched with the gesture contour model, and the matching process is: performing panning matching on the extracted gesture contour image with each reference gesture contour model, calculating a matching value, and selecting a gesture contour model with the largest matching value as the target to be identified. . Then, the rectangular area of the fingertip coarsely positioned by the gesture contour model is mapped onto the extracted gesture contour, and the rectangular area where the fingertip position of the gesture contour is located is obtained. In order to improve the matching degree, the extracted gesture outline can be scaled S305, precise positioning of the fingertips;
  • the contours of the fingers in the rectangular area are equally spaced, and the bending rate of each curve is calculated separately, and the maximum curve of the bending rate is taken as the precise position of the fingertip.
  • the embodiment of the present invention selects a Kalman filter to track the position of the fingertip.
  • the two main stages of the Kalman filter are prediction and updating. The equations are:
  • the user generally has a medium-quality frame in the middle of the writing process, and the starting point has many meaningless frames, which affects the target detection. Therefore, according to the empirical value, the fourth frame is used as the starting point. The starting point of the track.
  • the character input operation is considered to be ended.
  • the present invention quantifies the tangent angle, for example, every 15 is quantized into one direction, that is, a uniform quantization method using 24 feature vectors.
  • the change of the tangent angle at different times constitutes a finger Sharp trajectory.
  • the obtained trajectory is matched with the trained character model, and the model with the greatest likelihood is selected as the target character model.
  • the probability problem is involved.
  • the present invention uses the Verbit algorithm to find the likelihood value of each model, and determines the maximum likelihood as the final target.
  • the running track information is finally converted into corresponding characters according to the target character model.
  • an embodiment of the present invention further provides a character input device based on an image sensing module, where the device includes:
  • the image sensing module 10 is configured to collect a video frame including a fingertip of the object
  • the object fingertip trajectory information acquiring module 20 is configured to acquire the trajectory information of the fingertip of the object by using the video frame acquired by the image sensing module.
  • the object fingertip trajectory information acquiring module 20 acquires the object.
  • the trajectory information of the fingertip includes the following specific steps: acquiring an object gesture contour image through the image sensing module, performing time domain and spatial domain joint segmentation and binarization processing on the object gesture contour image; and performing gesture by panning, zooming, rotating, etc.
  • the contour is matched with the stored gesture contour reference model, and the gesture contour reference model with the highest similarity is selected, and the fingertip region of the reference model is mapped to the fingertip region of the object model; the finger contour bending rate is calculated at equal intervals in the fingertip region, The maximum bending rate is used as the precise position of the fingertip; the tangential angle at different times is calculated, and the tangential angle changes over a period of time are accumulated to obtain the running track information of the fingertip of the object.
  • the character conversion module 30 is configured to query the pre-trained character model library according to the running track information to convert the running track information into corresponding characters.
  • the character input device based on the image sensing module further includes: an object gesture contour image information collecting module 40, configured to collect object gesture contour image information through the image sensing module;
  • a gesture contour model obtaining module 50 configured to query according to the object gesture contour image information
  • the pre-trained gesture contour model library obtains a matching gesture contour model, so that the object fingertip running track information acquiring module can track the object fingertip according to the gesture contour model, and acquire the running track information of the object fingertip.
  • the image sensor module-based character input device further includes: an image processing module 60, configured to process the object gesture contour image information collected by the object gesture contour image information collection module 40 as follows:
  • the step of the object fingertip trajectory information acquiring module 20 performing tracking of the fingertip of the object according to the gesture contour model includes:
  • the method for acquiring the trajectory information of the fingertip of the object by the target finger trajectory information acquiring module 20 through the video frame collected by the image sensing module includes:
  • f ⁇ '
  • 1 , ⁇ 1 refers to the coordinates of the object's fingertip at the last t-i time Information
  • the () refers to real-time coordinate information of the fingertip of the object at real time t.
  • the method for the character conversion module 30 to query the pre-trained character model library according to the running track information to convert the running track information into corresponding characters includes:
  • the embodiment of the present invention further provides a terminal, which includes the image input module based character input device as described above.
  • the device includes:
  • the image sensing module 10 is configured to collect a video frame including a fingertip of the object
  • the object fingertip running track information acquiring module 20 is configured to acquire the running track information of the fingertip of the object through the video frame collected by the image sensing module;
  • the character conversion module 30 is configured to query the pre-trained character model library according to the running track information to convert the running track information into corresponding characters.
  • the image sensing module can be a common camera device. Therefore, for the terminal provided by the present invention, since the image can be acquired by the camera device provided by the terminal, the gesture contour of the user is extracted, and the hand contour and the existing hand are saved.
  • the gesture contour model is matched, and the gesture model is recognized to perform coarse positioning on the fingertip of the object, and then the fingertip of the finger is accurately positioned according to the bending rate of the contour of the finger. Predict the approximate position of the fingertip at the next moment, capture the movement trajectory of the fingertip, calculate the tangential angle at different moments, and accumulate the tangential angle changes over a period of time to obtain the trajectory of the fingertip of the object during the time period.
  • the obtained object fingertip trajectory is matched with the pre-stored character model library to generate corresponding characters.
  • the image input module-based character input method, device and terminal provided by the invention can meet the personalized requirements of the user for the handheld terminal, and at the same time realize fast and accurate input of characters.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Multimedia (AREA)
  • Human Computer Interaction (AREA)
  • User Interface Of Digital Computer (AREA)
  • Image Analysis (AREA)

Abstract

Disclosed are a character input method, device and terminal based on an image sensing module. The method includes: acquiring motion trace information about an object fingertip; and querying a pre-trained character model library according to the motion trace information, and converting the motion trace information into a corresponding character. The character input method, device and terminal based on an image sensing module provided in the present invention can meet the personalized demand of a user for a handheld terminal, and at the same time can realize rapid and accurate input of characters.

Description

一种基于图像传感模块的字符输入方法、 装置及终端 技术领域  Character input method, device and terminal based on image sensing module
本发明涉及移动终端领域, 尤其涉及一种基于图像传感模块的字符输 入方法、 装置及终端。 背景技术  The present invention relates to the field of mobile terminals, and in particular, to a character input method, apparatus, and terminal based on an image sensing module. Background technique
目前, 移动通信在进行着突飞猛进的发展, 由此, 人们对移动终端(例 如手机) 的人机交互功能的功能要求也随之越来越高。 手机作为人们日常 生活以及工作中必不可少的通讯工具, 目前也越来越多的向着智能化、 人 性化升级和发展, 以不断满足人们日益挑剔的个性化需求。  At present, mobile communication is undergoing rapid development, and thus, the functional requirements for human-computer interaction functions of mobile terminals (such as mobile phones) are becoming higher and higher. As an indispensable communication tool in people's daily life and work, mobile phones are increasingly being upgraded and developed intelligently and ergonomically to meet the increasingly demanding individual needs of people.
文字编辑和输入技术, 自从其诞生之日起便极大地改变了人们的通信 方式, 通过该技术, 人们可以编辑短信, 输入字符等。 目前, 对于字符的 输入方式有多种, 但鉴于多种字符输入方法都或多或少的具有一定的局限 性或缺陷。  Text editing and input technology has dramatically changed the way people communicate since its birth. With this technology, people can edit text messages, enter characters, and more. At present, there are many ways to input characters, but in view of the fact that many character input methods have more or less limitations or defects.
最初, 字符(或称文本) 的输入是靠按键编辑的, 在编辑字符的过程 中, 需要利用大拇指不停的按键, 采用该方法, 用户需要不断的重复按键 动作, 输入效率较低, 且由于长期的按键, 严重时甚至会引发类似于 "拇 指炎" 健康问题, 极大地影响到用户的健康。  Initially, the input of characters (or text) is edited by means of a button. In the process of editing characters, it is necessary to use a button that does not stop with a thumb. With this method, the user needs to repeatedly repeat the button action, and the input efficiency is low, and Due to long-term button presses, even when it is serious, it may cause health problems similar to "threbitis", which greatly affects the health of users.
另外, 由于目前大多数的手持终端配置了电阻屏或电容屏, 在输入字 符时, 借助手指或手写笔在终端的屏幕上进行书写, 通过手指或手写笔与 屏幕之间的压力或电容感应进行字符的识别以及字符的生成, 从而输入相 应的字符。 但该方法需要用户的终端配置电阻屏或电容屏, 而没有配置电 阻屏或电容屏的手机则无法采用该方法进行字符输入, 因此该方法具有一 定的局限性。 除此之外, 还有一些手持终端通过控制传感器感应用户做出的各种动 作来达到字符输入的目的, 例如在用户佩戴了特殊手套的情形下, 通过传 感器进行目标跟踪, 检测用户做出的动作而进行字符输入; 或在用户使用 特殊装置的荧光笔的情形下进行目标跟踪, 检测用户做出的动作而同样的 达到字符输入的目的。 但是采用这些方法实现字符输入时, 都需要使得这 些可持终端额外配置相应的硬件, 从而使得其成本较高, 并且在进行字符 输入时由于需要使用额外的设备, 因此使得用户的使用极不方便。 In addition, since most of the current handheld terminals are equipped with a resistive screen or a capacitive screen, when a character is input, writing is performed on the screen of the terminal by means of a finger or a stylus, by pressure or capacitive sensing between the finger or the stylus and the screen. The recognition of characters and the generation of characters, thereby inputting corresponding characters. However, the method requires the user's terminal to configure a resistive screen or a capacitive screen, and a mobile phone without a resistive screen or a capacitive screen cannot use the method for character input, so the method has certain limitations. In addition, some handheld terminals can control the sensor to sense various actions made by the user to achieve the purpose of character input. For example, when the user wears special gloves, the target tracking is performed by the sensor, and the user is detected. The character input is performed by the action; or the target tracking is performed in the case where the user uses the highlighter of the special device, and the action made by the user is detected to achieve the same purpose of character input. However, when these methods are used to implement character input, it is necessary to make these portable terminals additionally configure corresponding hardware, which makes the cost higher, and requires additional equipment when performing character input, thereby making the user's use extremely inconvenient. .
为此, 如何提供一种个性化、 输入快捷化的字符输入方法便逐渐受到 了人们的普遍关注。 发明内容  To this end, how to provide a personalized, input-driven character input method has gradually received widespread attention. Summary of the invention
有鉴于此, 本发明的主要目的在于提供一种基于图像传感模块的字符 输入方法、 装置及终端, 能够满足用户对于手持终端所提出的个性化需求, 同时还能实现字符的快捷、 准确的输入。  In view of this, the main object of the present invention is to provide a character input method, device and terminal based on an image sensing module, which can meet the personalized requirements of the user for the handheld terminal, and at the same time realize the fast and accurate character. Input.
为了达到本发明的目的, 本发明采用以下技术方案:  In order to achieve the object of the present invention, the present invention adopts the following technical solutions:
一种基于图像传感模块的字符输入方法, 包括:  A character input method based on an image sensing module, comprising:
获取对象指尖的运行轨迹信息;  Obtaining the running track information of the fingertip of the object;
依据所述运行轨迹信息查询预先训练出的字符模型库, 将所述运行轨 迹信息转化为相应的字符。  The pre-trained character model library is queried according to the running track information, and the running track information is converted into corresponding characters.
优选地, 所述获取对象指尖的运行轨迹信息之前, 所述基于图像传感 模块的字符输入方法还包括:  Preferably, before the acquiring the running track information of the fingertip, the character input method based on the image sensing module further includes:
通过图像传感模块采集对象手势轮廓图像信息;  Acquiring object gesture contour image information by using an image sensing module;
依据所述对象手势轮廓图像信息查询预先训练出的手势轮廓模型库, 获取匹配的手势轮廓模型;  Querying the pre-trained gesture contour model library according to the object gesture contour image information, and acquiring a matching gesture contour model;
依据所述手势轮廓模型进行对象指尖的跟踪。  Tracking of the fingertip of the object is performed according to the gesture contour model.
优选地, 所述依据所述对象手势轮廓图像信息查询预先训练出的手势 轮廓模型库之前, 该方法还包括: Preferably, the querying the pre-trained gesture according to the object gesture contour image information Before contouring the model library, the method also includes:
对采集到的对象手势轮廓图像信息进行二值化处理; 和 /或,  Binarizing the acquired object gesture contour image information; and/or,
对采集到的对象手势轮廓图像信息进行图象锐化处理; 和 /或, 对采集到的对象手势轮廓图像信息进行图象平滑处理。  Performing image sharpening processing on the collected object gesture contour image information; and/or performing image smoothing processing on the collected object gesture contour image information.
优选地, 所述依据所述手势轮廓模型进行对象指尖的跟踪为: 将参考 的手势轮廓模型的指尖粗定位的矩形区域映射到提取的对象手势轮廓图像 上, 获得对象手势轮廓图像的指尖位置所在的矩形区域;  Preferably, the tracking of the fingertip of the object according to the gesture contour model is: mapping a rectangular region of the fingertip coarsely positioned of the reference gesture contour model to the extracted object gesture contour image, and obtaining the finger of the object gesture contour image The rectangular area where the pointed position is located;
对所述矩形区域内的手指轮廓进行等间距划分, 并计算每段手指轮廓 片段的边缘弯曲率;  Equally dividing the contours of the fingers in the rectangular area, and calculating the edge bending rate of each segment of the finger contour;
获取边缘弯曲率最大的手指轮廓片段为对象指尖。  The finger contour segment with the largest edge curvature is obtained as the fingertip of the object.
优选地, 所述获取对象指尖的运行轨迹信息为:  Preferably, the running track information of the fingertip of the acquiring object is:
依据卡尔曼滤波器进行对象指尖的跟踪预测;  Tracking prediction of the fingertip of the object according to the Kalman filter;
从图像传感模块采集的视频帧中获取对象指尖的起始坐标信息, 之后 平均每隔至少一帧采集一次对象指尖的实时坐标信息;  Obtaining the starting coordinate information of the fingertip of the object from the video frame collected by the image sensing module, and then collecting the real-time coordinate information of the fingertip of the object at least once every at least one frame;
计算所述实时坐标信息与上一次坐标信息的切线角度 ^ ,并依据所述切 线角度的变化获取对象指尖的运行轨迹信息, 其中, 所述切线角度 ^的计算 为: ; 其中, 所述 (CJ 1)是指对象指尖在上一次 t-i时刻的坐标
Figure imgf000005_0001
Calculating a tangential angle ^ of the real-time coordinate information and the previous coordinate information, and acquiring trajectory information of the fingertip of the object according to the change of the tangential angle, wherein the tangential angle ^ is calculated as: CJ 1 ) refers to the coordinates of the fingertip of the object at the last ti time.
Figure imgf000005_0001
信息, 所述 ( )是指对象指尖在实时 t时刻的实时坐标信息。 Information, the ( ) refers to the real-time coordinate information of the fingertip of the object at the time t in real time.
优选地, 所述依据所述运行轨迹信息查询预先训练出的字符模型库, 将所述运行轨迹信息转化为相应的字符为:  Preferably, the querying the pre-trained character model library according to the running track information, and converting the running track information into corresponding characters:
依据所述运行轨迹信息查询预先训练出的字符模型库, 并将所述运行 轨迹信息与所有的字符模型进行匹配;  Querying the pre-trained character model library according to the running track information, and matching the running track information with all character models;
获取似然值最大的字符模型为目标字符模型, 其中, 采用 Verbit算法 计算每个字符模型的似然值;  The character model with the largest likelihood value is obtained as the target character model, wherein the likelihood value of each character model is calculated by using the Verbit algorithm;
依据所述目标字符模型将所述运行轨迹信息转化为相应的字符。 一种基于图像传感模块的字符输入装置, 该装置包括: The running track information is converted into corresponding characters according to the target character model. A character input device based on an image sensing module, the device comprising:
图像传感模块, 用于采集包含对象指尖的视频帧;  An image sensing module, configured to collect a video frame including a fingertip of the object;
对象指尖运行轨迹信息获取模块, 用于通过图像传感模块采集的视频 帧获取对象指尖的运行轨迹信息;  The object fingertip running track information acquiring module is configured to acquire the running track information of the fingertip of the object through the video frame collected by the image sensing module;
字符转化模块, 用于依据所述运行轨迹信息查询预先训练出的字符模 型库, 将所述运行轨迹信息转化为相应的字符。  The character conversion module is configured to query the pre-trained character model library according to the running track information, and convert the running track information into corresponding characters.
优选地, 所述基于图像传感模块的字符输入装置还包括:  Preferably, the character input module based on the image sensing module further includes:
对象手势轮廓图像信息采集模块, 用于通过图像传感模块采集对象手 势轮廓图像信息;  An object gesture contour image information collecting module, configured to collect object hand contour image information by using an image sensing module;
手势轮廓模型获取模块, 用于依据所述对象手势轮廓图像信息查询预 先训练出的手势轮廓模型库, 获取匹配的手势轮廓模型, 使得对象指尖运 行轨迹信息获取模块依据所述手势轮廓模型进行对象指尖的跟踪, 获取对 象指尖的运行轨迹信息。  a gesture contour model obtaining module, configured to query a pre-trained gesture contour model library according to the object gesture contour image information, and obtain a matching gesture contour model, so that the object fingertip running track information acquiring module performs the object according to the gesture contour model Tracking of the fingertips, obtaining the running track information of the fingertips of the object.
优选地, 所述基于图像传感模块的字符输入装置还包括:  Preferably, the character input module based on the image sensing module further includes:
图像处理模块, 用于对对象手势轮廓图像信息采集模块采集到的对象 手势轮廓图像信息进行如下处理:  The image processing module is configured to process the object gesture contour image information collected by the object gesture contour image information collection module as follows:
对采集到的对象手势轮廓图像信息进行二值化处理; 和 /或, 对采集到的对象手势轮廓图像信息进行图象锐化处理; 和 /或, 对采集到的对象手势轮廓图像信息进行图象平滑处理。  Performing binarization processing on the collected object gesture contour image information; and/or performing image sharpening processing on the collected object gesture contour image information; and/or, performing mapping on the collected object gesture contour image information Like smoothing.
优选地, 所述对象指尖运行轨迹信息获取模块依据所述手势轮廓模型 进行对象指尖的跟踪为:  Preferably, the object fingertip running track information acquiring module performs tracking of the fingertip of the object according to the gesture contour model:
将参考的手势轮廓模型的指尖粗定位的矩形区域映射到提取的对象手 势轮廓图像上, 获得对象手势轮廓图像的指尖位置所在的矩形区域;  Mapping a rectangular area of the fingertip coarsely positioned of the reference gesture contour model to the extracted object hand contour image, and obtaining a rectangular area where the fingertip position of the object gesture contour image is located;
对所述矩形区域内的手指轮廓进行等间距划分, 并计算每段手指轮廓 片段的边缘弯曲率; 获取边缘弯曲率最大的手指轮廓片段为对象指尖。 优选地, 所述对象 指尖运行轨迹信息获取模块通过图像传感模块采集的视频帧获取对象指尖 的运行轨迹信息为: Equally dividing the contours of the fingers in the rectangular area, and calculating the edge bending rate of each segment of the finger contour; The finger contour segment with the largest edge curvature is obtained as the fingertip of the object. Preferably, the information about the running track of the fingertip of the object obtained by the object fingertip trajectory information acquiring module by the image sensing module is:
依据卡尔曼滤波器进行对象指尖的跟踪预测;  Tracking prediction of the fingertip of the object according to the Kalman filter;
从图像传感模块采集的视频帧中获取对象指尖的起始坐标信息, 之后 平均每隔至少一帧采集一次对象指尖的实时坐标信息;  Obtaining the starting coordinate information of the fingertip of the object from the video frame collected by the image sensing module, and then collecting the real-time coordinate information of the fingertip of the object at least once every at least one frame;
计算所述实时坐标信息与上一次坐标信息的切线角度 ^ ,并依据所述切 线角度的变化获取对象指尖的运行轨迹信息, 其中, 所述切线角度 ^的计算 为: 0 = 4^; 其中, 所述 1, }^1)是指对象指尖在上一次 t-i时刻的坐标 信息, 所述 ( )是指对象指尖在实时 t时刻的实时坐标信息。 Calculating a tangential angle ^ of the real-time coordinate information and the previous coordinate information, and acquiring running trajectory information of the fingertip of the object according to the change of the tangential angle, wherein the tangential angle ^ is calculated as: 0 = 4^; The 1 , }^ 1 ) refers to the coordinate information of the fingertip of the object at the last time ti, and the ( ) refers to the real-time coordinate information of the fingertip of the object at the time t in real time.
优选地, 所述字符转化模块依据所述运行轨迹信息查询预先训练出的 字符模型库, 以将所述运行轨迹信息转化为相应的字符为:  Preferably, the character conversion module queries the pre-trained character model library according to the running track information, so as to convert the running track information into corresponding characters:
依据所述运行轨迹信息查询预先训练出的字符模型库, 并将所述运行 轨迹信息与所有的字符模型进行匹配;  Querying the pre-trained character model library according to the running track information, and matching the running track information with all character models;
获取似然值最大的字符模型为目标字符模型, 其中, 采用 Verbit算法 计算每个字符模型的似然值;  The character model with the largest likelihood value is obtained as the target character model, wherein the likelihood value of each character model is calculated by using the Verbit algorithm;
依据所述目标字符模型将所述运行轨迹信息转化为相应的字符。  The running track information is converted into corresponding characters according to the target character model.
一种终端, 其包括所述的基于图像传感模块的字符输入装置, 所述装 置包括:  A terminal comprising the image sensor module based character input device, the device comprising:
图像传感模块, 用于采集包含对象指尖的视频帧;  An image sensing module, configured to collect a video frame including a fingertip of the object;
对象指尖运行轨迹信息获取模块, 用于通过图像传感模块采集的视频 帧获取对象指尖的运行轨迹信息;  The object fingertip running track information acquiring module is configured to acquire the running track information of the fingertip of the object through the video frame collected by the image sensing module;
字符转化模块, 用于依据所述运行轨迹信息查询预先训练出的字符模 型库, 将所述运行轨迹信息转化为相应的字符。  The character conversion module is configured to query the pre-trained character model library according to the running track information, and convert the running track information into corresponding characters.
通过上述本发明的技术方案可以看出, 本发明提供的基于图像传感模 块的字符输入方法、 装置及终端依靠终端自带的图像传感模块(例如摄像 头装置)采集图像, 提取手势轮廓, 将手轮廓与已存的手势轮廓模型匹配, 识别出手势模型, 以对对象指尖进行粗定位, 然后根据手指轮廓的弯曲率 对对象指尖进行精确定位。 预测对象指尖下一时刻出现的大致位置, 捕获 指尖的运动轨迹, 计算不同时刻的切线角度, 将一段时间内的切线角度变 化累加, 便可得到该时间段内对象指尖运动轨迹。 将得到的对象指尖运动 轨迹与预存的字符模型库进行匹配, 从而生成相应的字符。 本发明提供的 基于图像传感模块的字符输入方法、 装置及终端, 能够满足用户对于手持 终端所提出的个性化需求, 同时还能实现字符的快捷、 准确的输入。 附图说明 It can be seen from the above technical solution of the present invention that the image sensing module provided by the present invention is based on The character input method, device and terminal of the block rely on the image sensing module (such as the camera device) provided by the terminal to collect images, extract the gesture contour, match the hand contour with the existing gesture contour model, and recognize the gesture model to the object. The fingertip is coarsely positioned, and then the fingertip of the finger is accurately positioned according to the bending rate of the contour of the finger. Predict the approximate position of the fingertip at the next moment, capture the movement trajectory of the fingertip, calculate the tangential angle at different moments, and accumulate the tangential angle changes over a period of time to obtain the trajectory of the fingertip of the object during the time period. The obtained object fingertip trajectory is matched with the pre-stored character model library to generate corresponding characters. The image input module-based character input method, device and terminal provided by the invention can meet the personalized requirements of the user for the handheld terminal, and at the same time realize fast and accurate input of characters. DRAWINGS
此处所说明的附图用来提供对本发明的进一步理解, 构成本发明的一 部分, 本发明的示意性实施例及其说明用于解释本发明, 并不构成对本发 明的不当限定。 在附图中:  The drawings are intended to provide a further understanding of the present invention, and are intended to be a part of the invention. In the drawing:
图 1是本发明一实施例提供的基于图像传感模块的字符输入方法流程 示意图;  1 is a schematic flow chart of a character input method based on an image sensing module according to an embodiment of the present invention;
图 2是本发明一较佳实施例提供的基于图像传感模块的字符输入方法 详细流程示意图;  2 is a detailed flow chart of a character input method based on an image sensing module according to a preferred embodiment of the present invention;
图 3是本发明一实施例提供的基于图像传感模块的字符输入装置结构 示意图。 具体实施方式  FIG. 3 is a schematic structural diagram of a character input device based on an image sensing module according to an embodiment of the present invention. detailed description
本发明的基本思想是: 获取对象指尖的运行轨迹信息; 依据所述运行 轨迹信息查询预先训练出的字符模型库, 将所述运行轨迹信息转化为相应 的字符。  The basic idea of the present invention is: acquiring the running track information of the fingertip of the object; querying the pre-trained character model library according to the running track information, and converting the running track information into corresponding characters.
为了使本发明所要解决的技术问题、 技术方案及有益效果更加清楚、 明白, 以下结合附图和实施例, 对本发明进行进一步详细说明。 应当理解, 此处所描述的具体实施例仅仅用以解释本发明, 并不用于限定本发明。 In order to make the technical problems, technical solutions and beneficial effects to be solved by the present invention clearer, It is to be understood that the present invention will be further described in detail below with reference to the drawings and embodiments. It is understood that the specific embodiments described herein are merely illustrative of the invention and are not intended to limit the invention.
如图 1 所示, 本发明一实施例提供的一种基于图像传感模块的字符输 入方法, 该方法包括如下步驟:  As shown in FIG. 1, a character input method based on an image sensing module according to an embodiment of the present invention includes the following steps:
S101、 通过图像传感模块获取对象指尖的运行轨迹信息;  S101. Acquire, by using an image sensing module, information about a running track of an object fingertip;
在该步驟中, 所述图像传感模块用于采集用户的手指的指尖画面信息, 并依据所述采集到的指尖画面信息提取用户的指尖运行轨迹信息, 例如所 述图像传感模块为普通手机或智能手机均配置了的摄像头装置。  In this step, the image sensing module is configured to collect fingertip screen information of a user's finger, and extract user's fingertip running track information according to the collected fingertip screen information, for example, the image sensing module. A camera device that is configured for both regular mobile phones and smartphones.
具体实施时, 在该步驟中, 包括如下具体步驟:  In the specific implementation, in this step, the following specific steps are included:
a、 通过图像传感模块获取对象手势轮廓图像, 对对象手势轮廓图像进 行时域和空域联合分割、 二值化处理;  a. acquiring an object gesture contour image through the image sensing module, and performing joint segmentation and binarization processing on the object gesture contour image in time domain and spatial domain;
b、 通过平移、 缩放、 旋转等方法将手势轮廓与存储的手势轮廓参考模 型匹配, 选择相似度最大的手势轮廓参考模型, 将该参考模型的指尖区域 映射到对象模型指尖区域;  b. matching the gesture contour with the stored gesture contour reference model by panning, zooming, rotating, etc., selecting the gesture contour reference model with the largest similarity, and mapping the fingertip region of the reference model to the fingertip region of the object model;
C、 在该指尖区域内等间隔计算手指轮廓弯曲率, 弯曲率最大的作为指 尖的精确位置;  C. Calculate the bending rate of the finger contour at equal intervals in the fingertip region, and the bending position is the precise position of the fingertip;
d、 计算不同时刻的切线角度, 将一段时间内的切线角度变化累加, 获 得对象指尖的运行轨迹信息。  d. Calculate the tangent angle at different moments, accumulate the tangent angle changes over a period of time, and obtain the running track information of the fingertip of the object.
S102、 依据所述运行轨迹信息查询预先训练出的字符模型库, 以将所 述运行轨迹信息转化为相应的字符。  S102. Query a pre-trained character model library according to the running track information, so as to convert the running track information into corresponding characters.
在该步驟中, 依据上述采集到的用户的指尖运行轨迹信息与预先建立 好的字符模型库进行匹配, 在找到对应的字符模型时, 以将所述运行轨迹 信息翻译成对应的字符信息, 从而为字符输入提供依据。  In this step, according to the collected fingertip running track information of the user, matching with the pre-established character model library, when the corresponding character model is found, the running track information is translated into corresponding character information, This provides a basis for character input.
优选实施方式下, 在执行所有步驟之前, 该方法还包括以下步驟: S1001、 通过图像传感模块采集对象手势轮廓图像信息, 例如所述对象 手势轮廓图像信息包括但不限于以下几种: 只有食指伸直, 其他手指握拳; 只有中指伸直, 其他手指握拳; 拇指和食指同时伸直, 其他手指握拳等。 In a preferred embodiment, before performing all the steps, the method further includes the following steps: S1001: Acquire an object gesture contour image information, such as the object, by using an image sensing module. Gesture contour image information includes but is not limited to the following: Only the index finger is straight, other fingers are clenched; only the middle finger is straight, other fingers are clenched; the thumb and forefinger are straight at the same time, other fingers are clenched.
S1002、依据所述对象手势轮廓图像信息查询预先训练出的手势轮廓模 型库, 获取匹配的手势轮廓模型。  S1002: Query a pre-trained gesture contour model library according to the object gesture contour image information, and obtain a matching gesture contour model.
S1003、 依据所述手势轮廓模型进行对象指尖的跟踪。  S1003. Track the object fingertip according to the gesture contour model.
优选实施方式下, 在依据所述对象手势轮廓图像信息查询预先训练出 的手势轮廓模型库, 以获取匹配的手势轮廓模型之前, 该方法还包括: In a preferred embodiment, before querying the pre-trained gesture contour model library according to the object gesture contour image information to obtain a matching gesture contour model, the method further includes:
( 1 )对采集到的对象手势轮廓图像信息进行二值化处理; 和 /或,(1) performing binarization processing on the acquired object gesture contour image information; and/or,
( 2 )对采集到的对象手势轮廓图像信息进行图象锐化处理; 和 /或, ( 3 )对采集到的对象手势轮廓图像信息进行图象平滑处理。 (2) performing image sharpening processing on the collected object gesture contour image information; and/or, (3) performing image smoothing processing on the collected object gesture contour image information.
以同时对采集到的对象手势轮廓图像信息进行二值化处理、 图象锐化 处理以及图象平滑处理为例, 由于灰度图像只有空间采样点的亮度信息, 因而可以用一个数值来表示。 而对于彩色图像, 包含的内容更丰富, 但表 示更为复杂。 但是在本发明中不关心且没有必要保留图像的色彩信息, 因 此可以将彩色图像的对象手势轮廓图像信息进行二值化处理, 转化为灰度 图像, 降低复杂度, 减少内存资源。 同时, 为了突出手势轮廓的边缘, 可 以对对象手势轮廓图像信息进行图象锐化处理,本发明采用 Roberts梯度锐 化方法增强图像边缘。 除此之外, 为了降低噪声, 还可以可对对象手势轮 廓图像信息进行图象平滑处理。  Taking the binarization processing, the image sharpening processing, and the image smoothing processing of the collected object gesture contour image information at the same time as an example, since the grayscale image has only the luminance information of the spatial sampling point, it can be represented by a numerical value. For color images, the content is richer, but the presentation is more complicated. However, in the present invention, it is not necessary to maintain the color information of the image, so that the object gesture contour image information of the color image can be binarized and converted into a grayscale image, which reduces complexity and reduces memory resources. At the same time, in order to highlight the edge of the gesture outline, image sharpening processing can be performed on the object gesture contour image information, and the present invention uses the Roberts gradient sharpening method to enhance the image edge. In addition to this, in order to reduce noise, it is also possible to perform image smoothing processing on the subject gesture profile image information.
优选实施方式下, 在所述步驟 S1003 中, 依据所述手势轮廓模型进行 对象指尖的跟踪的步驟包括:  In a preferred embodiment, in the step S1003, the step of tracking the fingertip of the object according to the gesture contour model includes:
S10031、 将该参考的手势轮廓模型的指尖粗定位的矩形区域映射到提 取的对象手势轮廓图像上, 获得对象手势轮廓图像的指尖位置所在的矩形 区域。  S10031: Map a rectangular area of the fingertip coarsely positioned of the reference gesture contour model to the extracted object gesture contour image, and obtain a rectangular area where the fingertip position of the target gesture contour image is located.
S10032, 对所述矩形区域内的手指轮廓进行等间距划分, 并计算每段 手指轮廓片段的边缘弯曲率。 S10032, equally dividing a finger contour in the rectangular area, and calculating each segment The edge bend rate of the finger contour segment.
S 10033、 获取边缘弯曲率最大的手指轮廓片段为对象指尖。  S 10033. Obtain a finger contour segment with the largest edge bending rate as the object fingertip.
在所述步驟 S101中, 通过图像传感模块获取对象指尖的运行轨迹信息 的方法具体包括如下步驟:  In the step S101, the method for acquiring the running track information of the fingertip of the object by the image sensing module specifically includes the following steps:
S 1011、 依据卡尔曼滤波器进行对象指尖的跟踪预测。  S 1011: Tracking prediction of an object fingertip according to a Kalman filter.
51012、 从图像传感模块采集的视频帧中获取对象指尖的起始坐标信 息, 之后平均每隔至少一帧采集一次对象指尖的实时坐标信息。  51012. Acquire a starting coordinate information of an object fingertip from a video frame collected by the image sensing module, and then collect real-time coordinate information of the fingertip of the object once every at least one frame.
51013、 计算所述实时坐标信息与上一次坐标信息的切线角度^ 并依 据所述切线角度的变化获取对象指尖的运行轨迹信息, 其中, 所述切线角 度 ø的计算数学式如下: , 其中, 所述 (e -1)是指对象指尖在上一次 t-i时刻的坐标
Figure imgf000011_0001
51013. Calculate a tangential angle of the real-time coordinate information and the previous coordinate information, and obtain operation trajectory information of the fingertip of the object according to the change of the tangential angle, where the calculation formula of the tangential angle ø is as follows: The (e - 1 ) refers to the coordinates of the fingertip of the object at the last ti time.
Figure imgf000011_0001
信息, 所述 ( ί, 是指对象指尖在实时 t时刻的实时坐标信息。 Information, the ( ί, refers to the real-time coordinate information of the fingertip of the object at real time t.
在所述步驟 S102中, 依据所述运行轨迹信息查询预先训练出的字符模 型库, 以将所述运行轨迹信息转化为相应的字符的方法包括:  In the step S102, the method for querying the pre-trained character model library according to the running track information to convert the running track information into corresponding characters includes:
51021、 依据所述运行轨迹信息查询预先训练出的字符模型库, 并将所 述运行轨迹信息与所有的字符模型进行匹配。  51021. Query a pre-trained character model library according to the running track information, and match the running track information with all character models.
51022、获取似然值最大的字符模型为目标字符模型,其中,采用 Verbit 算法计算每个字符模型的似然值。  51022. The character model with the largest likelihood value is the target character model, wherein the likelihood value of each character model is calculated by the Verbit algorithm.
51023、 依据所述目标字符模型将所述运行轨迹信息转化为相应的字 付。  51023. Convert the running track information into a corresponding word payment according to the target character model.
如图 2所示, 图 2是本发明一较佳实施例提供的基于图像传感模块的 字符输入方法详细流程示意图, 所述流程具体包括如下步驟:  As shown in FIG. 2, FIG. 2 is a schematic flowchart of a character input method based on an image sensing module according to a preferred embodiment of the present invention. The process specifically includes the following steps:
S301、 手势模型选取以及指尖粗定位;  S301, gesture model selection, and fingertip coarse positioning;
手势模型的选取直接影响识别效果。 本发明选取常见的手势模型作为 参考模型, 例如包括但不限于以下几种: 只有食指伸直, 其他手指握拳; 只有中指伸直, 其他手指握拳; 拇指和食指同时伸直, 其他手指握拳等。 对于不同的手势模型, 分别预定义指尖的初始位置。 例如, 对于只有食指 伸直的情况, 将指尖的初始位置定义在食指最上方, 并用一定大小的矩形 框将指尖位置包含进去, 该矩形区域作为后续指尖精确定位的区域。 The selection of the gesture model directly affects the recognition effect. The present invention selects a common gesture model as a reference model, for example, but not limited to the following: Only the index finger is straight, and other fingers are clenched; Only the middle finger is straight, the other fingers are clenched; the thumb and forefinger are straight at the same time, and other fingers are clenched. For different gesture models, the initial position of the fingertip is predefined. For example, in the case where only the index finger is straight, the initial position of the fingertip is defined at the top of the index finger, and the fingertip position is included with a rectangular frame of a certain size, which is the area where the subsequent fingertip is accurately positioned.
5302、 手势轮廓提取;  5302, gesture contour extraction;
一般利用帧间或帧内的信息对视频图像进行分析。 但由于场景中存在 噪声, 单个的视频分割方法并不能准确地逼近物体的边缘。 因此要准确的 提取手势轮廓, 要联合物体的颜色、 亮度等空间信息进行视频分割。 例如 本发明实施例采用基于时 -空联合分割法, 综合利用时间域的帧间运动信息 和空间的肤色、 亮度信息, 同时进行时间和空间的分割方法, 提取准确的 手势边缘。 通过空域分割, 获取具有准确语义的初始分割区域, 通过时域 分割获取图像的运动区域, 连接不连续边缘, 得到手势轮廓。  Video images are typically analyzed using inter-frame or intra-frame information. However, due to the presence of noise in the scene, a single video segmentation method does not accurately approximate the edge of the object. Therefore, to accurately extract the gesture outline, the spatial information such as the color and brightness of the object should be combined for video segmentation. For example, the embodiment of the present invention adopts a time-space joint segmentation method, and comprehensively utilizes inter-frame motion information in a time domain and spatial skin color and luminance information, and simultaneously performs time and space segmentation methods to extract an accurate gesture edge. Through the spatial domain segmentation, the initial segmentation region with accurate semantics is obtained, and the motion region of the image is obtained by time domain segmentation, and the discontinuous edge is connected to obtain the gesture contour.
5303、 图像预处理;  5303, image preprocessing;
灰度图像只有空间采样点的亮度信息, 因而可以用一个数值来表示。 而对于彩色图像, 包含的内容更丰富, 但表示更为复杂。 但是在该发明实 施例中, 没有必要保留图像的色彩信息, 因此需要将彩色图像进行二值化 处理, 转化为灰度图像, 降低复杂度, 减少内存资源。 为突出手的边缘, 对图像进行锐化, 本发明采用 Roberts梯度锐化方法增强图像边缘。 为降低 噪声, 可对二值图像进行平滑处理。  The grayscale image has only the luminance information of the spatial sampling point, and thus can be represented by a numerical value. For color images, the content is richer, but the representation is more complicated. However, in the embodiment of the invention, it is not necessary to preserve the color information of the image. Therefore, the color image needs to be binarized and converted into a grayscale image, which reduces complexity and reduces memory resources. To highlight the edges of the hand and sharpen the image, the present invention uses the Roberts gradient sharpening method to enhance the edges of the image. To reduce noise, the binary image can be smoothed.
S304、 手势轮廓与手势轮廓模型匹配;  S304. The gesture contour is matched with the gesture contour model.
将提取的手势轮廓与手势轮廓模型进行匹配, 匹配的过程为: 将提取 的手势轮廓图像与每个参考手势轮廓模型进行平移匹配, 计算匹配值, 选 取匹配值最大的手势轮廓模型作为要识别目标。 然后将该手势轮廓模型的 指尖粗定位的矩形区域映射到提取的手势轮廓上, 便得到手势轮廓的指尖 位置所在的矩形区域。 为了提高匹配度, 可以将提取的手势轮廓进行缩放 S305、 指尖精确定位; The extracted gesture contour is matched with the gesture contour model, and the matching process is: performing panning matching on the extracted gesture contour image with each reference gesture contour model, calculating a matching value, and selecting a gesture contour model with the largest matching value as the target to be identified. . Then, the rectangular area of the fingertip coarsely positioned by the gesture contour model is mapped onto the extracted gesture contour, and the rectangular area where the fingertip position of the gesture contour is located is obtained. In order to improve the matching degree, the extracted gesture outline can be scaled S305, precise positioning of the fingertips;
对矩形区域内的手指轮廓进行等间隔的划分, 分别计算每一段曲线的 弯曲率, 弯曲率最大曲线的作为指尖的精确位置。  The contours of the fingers in the rectangular area are equally spaced, and the bending rate of each curve is calculated separately, and the maximum curve of the bending rate is taken as the precise position of the fingertip.
S306、 运动目标预测跟踪  S306, moving target prediction tracking
在视频图像中, 相邻帧的时间间隔较小, 可以认为是匀速运动, 因此 可以将运动状态的变化描述成动态线性系统。 考虑到实时性要求, 本发明 实施例选择卡尔曼滤波器实现对指尖位置进行跟踪。 卡尔曼滤波器最主要 的两个阶段是预测和更新, 方程分别为:  In video images, the time interval of adjacent frames is small and can be considered as uniform motion, so the change of motion state can be described as a dynamic linear system. In view of real-time requirements, the embodiment of the present invention selects a Kalman filter to track the position of the fingertip. The two main stages of the Kalman filter are prediction and updating. The equations are:
S (n) = AS (n - \) S (n) = AS (n - \)
D{n) = AD{n - \)A
Figure imgf000013_0001
D{n) = AD{n - \)A
Figure imgf000013_0001
5307、 运动目标起点获取;  5307, the starting point of the moving target is obtained;
在本发明实施例的具体实施过程中, 用户一般在书写过程中其中间帧 是质量较好的帧, 而起点有很多无意义帧, 影响目标检测, 因此根据经验 值, 从第 4帧开始作为轨迹的起始点。  In the specific implementation process of the embodiment of the present invention, the user generally has a medium-quality frame in the middle of the writing process, and the starting point has many meaningless frames, which affects the target detection. Therefore, according to the empirical value, the fourth frame is used as the starting point. The starting point of the track.
5308、 运动目标终点获取;  5308, the target of the moving target is obtained;
在本发明实施例中, 当运动目标(即指尖)在 2秒内没有任何运动时, 则认为字符输入操作结束。  In the embodiment of the present invention, when the moving target (i.e., the fingertip) does not have any motion within 2 seconds, the character input operation is considered to be ended.
5309、 运动轨迹特征获取;  5309, acquisition of motion trajectory features;
获取相邻两帧的坐标后, 计算切线角度, 设 t-1 时刻指尖点的坐标为 — ^ό—1 ) , t时刻指尖点的坐标是 ( , 此时有切线角度^ f ^。 为了 简化计算, 本发明将切线角度进行量化, 例如, 每 15度量化为一个方向, 即采用 24个特征矢量的均匀量化方法。 不同时刻的切线角度变化组成了指 尖的运动轨迹。 After obtaining the coordinates of two adjacent frames, calculate the tangent angle, set the coordinates of the fingertip point at time t-1 to be -^ό- 1 , and the coordinate of the fingertip point at time t is (, there is a tangent angle ^f^). In order to simplify the calculation, the present invention quantifies the tangent angle, for example, every 15 is quantized into one direction, that is, a uniform quantization method using 24 feature vectors. The change of the tangent angle at different times constitutes a finger Sharp trajectory.
S310、 动态轨迹识别。  S310, dynamic track recognition.
将得到的轨迹与训练好的字符模型相匹配, 选择似然度最大的模型作 为目标字符模型。 其中涉及概率问题, 本发明采用 Verbit算法求出每个模 型的似然值, 将似然度最大的确定为最终目标。 最终依据所述目标字符模 型将所述运行轨迹信息转化为相应的字符。  The obtained trajectory is matched with the trained character model, and the model with the greatest likelihood is selected as the target character model. Among them, the probability problem is involved. The present invention uses the Verbit algorithm to find the likelihood value of each model, and determines the maximum likelihood as the final target. The running track information is finally converted into corresponding characters according to the target character model.
如图 3 所示, 本发明实施例还提供了一种基于图像传感模块的字符输 入装置, 所述装置包括:  As shown in FIG. 3, an embodiment of the present invention further provides a character input device based on an image sensing module, where the device includes:
图像传感模块 10, 用于采集包含对象指尖的视频帧;  The image sensing module 10 is configured to collect a video frame including a fingertip of the object;
对象指尖运行轨迹信息获取模块 20, 用于通过图像传感模块采集的视 频帧获取对象指尖的运行轨迹信息, 在本发明实施例中, 所述对象指尖运 行轨迹信息获取模块 20获取对象指尖的运行轨迹信息包括如下具体步驟: 通过图像传感模块获取对象手势轮廓图像, 对对象手势轮廓图像进行 时域和空域联合分割、 二值化处理; 通过平移、 缩放、 旋转等方法将手势 轮廓与存储的手势轮廓参考模型匹配, 选择相似度最大的手势轮廓参考模 型, 将该参考模型的指尖区域映射到对象模型指尖区域; 在该指尖区域内 等间隔计算手指轮廓弯曲率, 弯曲率最大的作为指尖的精确位置; 计算不 同时刻的切线角度, 将一段时间内的切线角度变化累加, 获得对象指尖的 运行轨迹信息。  The object fingertip trajectory information acquiring module 20 is configured to acquire the trajectory information of the fingertip of the object by using the video frame acquired by the image sensing module. In the embodiment of the present invention, the object fingertip trajectory information acquiring module 20 acquires the object. The trajectory information of the fingertip includes the following specific steps: acquiring an object gesture contour image through the image sensing module, performing time domain and spatial domain joint segmentation and binarization processing on the object gesture contour image; and performing gesture by panning, zooming, rotating, etc. The contour is matched with the stored gesture contour reference model, and the gesture contour reference model with the highest similarity is selected, and the fingertip region of the reference model is mapped to the fingertip region of the object model; the finger contour bending rate is calculated at equal intervals in the fingertip region, The maximum bending rate is used as the precise position of the fingertip; the tangential angle at different times is calculated, and the tangential angle changes over a period of time are accumulated to obtain the running track information of the fingertip of the object.
字符转化模块 30, 用于依据所述运行轨迹信息查询预先训练出的字符 模型库, 以将所述运行轨迹信息转化为相应的字符。  The character conversion module 30 is configured to query the pre-trained character model library according to the running track information to convert the running track information into corresponding characters.
优选实施方式下, 所述基于图像传感模块的字符输入装置还包括: 对象手势轮廓图像信息采集模块 40, 用于通过图像传感模块采集对象 手势轮廓图像信息;  In a preferred embodiment, the character input device based on the image sensing module further includes: an object gesture contour image information collecting module 40, configured to collect object gesture contour image information through the image sensing module;
手势轮廓模型获取模块 50, 用于依据所述对象手势轮廓图像信息查询 预先训练出的手势轮廓模型库, 获取匹配的手势轮廓模型, 以使得对象指 尖运行轨迹信息获取模块能够依据所述手势轮廓模型进行对象指尖的跟 踪, 获取对象指尖的运行轨迹信息。 a gesture contour model obtaining module 50, configured to query according to the object gesture contour image information The pre-trained gesture contour model library obtains a matching gesture contour model, so that the object fingertip running track information acquiring module can track the object fingertip according to the gesture contour model, and acquire the running track information of the object fingertip.
优选实施方式下, 所述基于图像传感模块的字符输入装置还包括: 图像处理模块 60,用于对对象手势轮廓图像信息采集模块 40采集到的 对象手势轮廓图像信息进行如下处理:  In a preferred embodiment, the image sensor module-based character input device further includes: an image processing module 60, configured to process the object gesture contour image information collected by the object gesture contour image information collection module 40 as follows:
( 1 )对采集到的对象手势轮廓图像信息进行二值化处理; 和 /或, (1) performing binarization processing on the acquired object gesture contour image information; and/or,
( 2 )对采集到的对象手势轮廓图像信息进行图象锐化处理; 和 /或, ( 3 )对采集到的对象手势轮廓图像信息进行图象平滑处理。 (2) performing image sharpening processing on the collected object gesture contour image information; and/or, (3) performing image smoothing processing on the collected object gesture contour image information.
其中, 所述对象指尖运行轨迹信息获取模块 20依据所述手势轮廓模型 进行对象指尖的跟踪的步驟包括:  The step of the object fingertip trajectory information acquiring module 20 performing tracking of the fingertip of the object according to the gesture contour model includes:
( 1 )将该参考的手势轮廓模型的指尖粗定位的矩形区域映射到提取的 对象手势轮廓图像上, 获得对象手势轮廓图像的指尖位置所在的矩形区域; (1) mapping a rectangular area of the fingertip coarsely positioned of the reference gesture contour model to the extracted object gesture contour image, and obtaining a rectangular area where the fingertip position of the object gesture contour image is located;
( 2 )对所述矩形区域内的手指轮廓进行等间距划分, 并计算每段手指 轮廓片段的边缘弯曲率; (2) equally dividing the contours of the fingers in the rectangular area, and calculating the edge bending rate of each of the finger contour segments;
( 3 )获取边缘弯曲率最大的手指轮廓片段为对象指尖。  (3) Obtain the finger contour segment with the largest edge bending rate as the fingertip of the object.
另外, 所述对象指尖运行轨迹信息获取模块 20通过图像传感模块采集 的视频帧获取对象指尖的运行轨迹信息的方法包括:  In addition, the method for acquiring the trajectory information of the fingertip of the object by the target finger trajectory information acquiring module 20 through the video frame collected by the image sensing module includes:
( 1 )依据卡尔曼滤波器进行对象指尖的跟踪预测;  (1) Tracking prediction of the fingertip of the object according to the Kalman filter;
( 2 )从图像传感模块采集的视频帧中获取对象指尖的起始坐标信息, 之后平均每隔至少一帧采集一次对象指尖的实时坐标信息;  (2) acquiring the starting coordinate information of the fingertip of the object from the video frame collected by the image sensing module, and then collecting the real-time coordinate information of the fingertip of the object at least once every at least one frame;
( 3 )计算所述实时坐标信息与上一次坐标信息的切线角度^ 并依据 所述切线角度的变化获取对象指尖的运行轨迹信息, 其中, 所述切线角度 ^ 的计算数学式如下:  (3) calculating a tangential angle of the real-time coordinate information and the previous coordinate information, and acquiring trajectory information of the fingertip of the object according to the change of the tangential angle, wherein the calculation formula of the tangential angle ^ is as follows:
^= f ^ ' 其中, 所述 1,}^1)是指对象指尖在上一次 t- i时刻的坐标 信息, 所述 ( )是指对象指尖在实时 t时刻的实时坐标信息。 ^= f ^ ' where 1 , }^ 1 ) refers to the coordinates of the object's fingertip at the last t-i time Information, the () refers to real-time coordinate information of the fingertip of the object at real time t.
所述字符转化模块 30依据所述运行轨迹信息查询预先训练出的字符模 型库, 以将所述运行轨迹信息转化为相应的字符的方法包括:  The method for the character conversion module 30 to query the pre-trained character model library according to the running track information to convert the running track information into corresponding characters includes:
( 1 )依据所述运行轨迹信息查询预先训练出的字符模型库, 并将所述 运行轨迹信息与所有的字符模型进行匹配;  (1) querying the pre-trained character model library according to the running track information, and matching the running track information with all character models;
( 2 )获取似然值最大的字符模型为目标字符模型, 其中, 采用 Verbit 算法计算每个字符模型的似然值;  (2) obtaining the character model with the largest likelihood value as the target character model, wherein the Verbit algorithm is used to calculate the likelihood value of each character model;
( 3 )依据所述目标字符模型将所述运行轨迹信息转化为相应的字符。 相应地, 本发明实施例还提供了一种终端, 其包括如上所述的基于图 像传感模块的字符输入装置, 参考图 3, 所述装置包括:  (3) converting the running track information into corresponding characters according to the target character model. Correspondingly, the embodiment of the present invention further provides a terminal, which includes the image input module based character input device as described above. Referring to FIG. 3, the device includes:
图像传感模块 10, 用于采集包含对象指尖的视频帧;  The image sensing module 10 is configured to collect a video frame including a fingertip of the object;
对象指尖运行轨迹信息获取模块 20, 用于通过图像传感模块采集的视 频帧获取对象指尖的运行轨迹信息;  The object fingertip running track information acquiring module 20 is configured to acquire the running track information of the fingertip of the object through the video frame collected by the image sensing module;
字符转化模块 30, 用于依据所述运行轨迹信息查询预先训练出的字符 模型库, 以将所述运行轨迹信息转化为相应的字符。  The character conversion module 30 is configured to query the pre-trained character model library according to the running track information to convert the running track information into corresponding characters.
在该终端中, 所述图像传感模块可以为普通摄像头装置, 因此对于本 发明所提供的终端, 由于能够依靠终端自带的摄像头装置采集图像, 提取 用户的手势轮廓, 将手轮廓与已存的手势轮廓模型匹配, 识别出手势模型, 以对对象指尖进行粗定位, 然后根据手指轮廓的弯曲率对对象指尖进行精 确定位。 预测对象指尖下一时刻出现的大致位置, 捕获指尖的运动轨迹, 计算不同时刻的切线角度, 将一段时间内的切线角度变化累加, 便可得到 该时间段内对象指尖运动轨迹。 将得到的对象指尖运动轨迹与预存的字符 模型库进行匹配, 从而生成相应的字符。 本发明提供的基于图像传感模块 的字符输入方法、 装置及终端, 能够满足用户对于手持终端所提出的个性 化需求, 同时还能实现字符的快捷、 准确的输入。 上述说明示出并描述了本发明的一个优选实施例, 但如前所述, 应当 理解本发明并非局限于本文所披露的形式, 不应看作是对其他实施例的排 除, 而可用于各种其他组合、 修改和环境, 并能够在本文所述发明构想范 围内, 通过上述教导或相关领域的技术或知识进行改动。 而本领域人员所 进行的改动和变化不脱离本发明的精神和范围, 则都应在本发明所附权利 要求的保护范围内。 In the terminal, the image sensing module can be a common camera device. Therefore, for the terminal provided by the present invention, since the image can be acquired by the camera device provided by the terminal, the gesture contour of the user is extracted, and the hand contour and the existing hand are saved. The gesture contour model is matched, and the gesture model is recognized to perform coarse positioning on the fingertip of the object, and then the fingertip of the finger is accurately positioned according to the bending rate of the contour of the finger. Predict the approximate position of the fingertip at the next moment, capture the movement trajectory of the fingertip, calculate the tangential angle at different moments, and accumulate the tangential angle changes over a period of time to obtain the trajectory of the fingertip of the object during the time period. The obtained object fingertip trajectory is matched with the pre-stored character model library to generate corresponding characters. The image input module-based character input method, device and terminal provided by the invention can meet the personalized requirements of the user for the handheld terminal, and at the same time realize fast and accurate input of characters. The above description shows and describes a preferred embodiment of the present invention, but as described above, it should be understood that the present invention is not limited to the forms disclosed herein, and should not be construed as Other combinations, modifications, and environments are possible and can be modified by the teachings or related art or knowledge within the scope of the inventive concept described herein. All changes and modifications made by those skilled in the art are intended to be within the scope of the appended claims.

Claims

权利要求书 Claim
1、一种基于图像传感模块的字符输入方法,其特征在于,该方法包括: 获取对象指尖的运行轨迹信息;  A character input method based on an image sensing module, the method comprising: acquiring operation track information of a fingertip of an object;
依据所述运行轨迹信息查询预先训练出的字符模型库, 将所述运行轨 迹信息转化为相应的字符。  The pre-trained character model library is queried according to the running track information, and the running track information is converted into corresponding characters.
2、 如权利要求 1所述的基于图像传感模块的字符输入方法, 其特征在 于, 所述获取对象指尖的运行轨迹信息之前, 该方法还包括:  2. The image input module-based character input method according to claim 1, wherein before the acquiring the trajectory information of the fingertip of the object, the method further comprises:
通过图像传感模块采集对象手势轮廓图像信息;  Acquiring object gesture contour image information by using an image sensing module;
依据所述对象手势轮廓图像信息查询预先训练出的手势轮廓模型库, 获取匹配的手势轮廓模型;  Querying the pre-trained gesture contour model library according to the object gesture contour image information, and acquiring a matching gesture contour model;
依据所述手势轮廓模型进行对象指尖的跟踪。  Tracking of the fingertip of the object is performed according to the gesture contour model.
3、 如权利要求 2所述的基于图像传感模块的字符输入方法, 其特征在 于, 所述依据所述对象手势轮廓图像信息查询预先训练出的手势轮廓模型 库之前, 该方法还包括:  The image sensor module-based character input method according to claim 2, wherein the method further comprises: before querying the pre-trained gesture contour model library according to the object gesture contour image information, the method further comprises:
对采集到的对象手势轮廓图像信息进行二值化处理; 和 /或,  Binarizing the acquired object gesture contour image information; and/or,
对采集到的对象手势轮廓图像信息进行图象锐化处理; 和 /或, 对采集到的对象手势轮廓图像信息进行图象平滑处理。  Performing image sharpening processing on the collected object gesture contour image information; and/or performing image smoothing processing on the collected object gesture contour image information.
4、 如权利要求 2所述的基于图像传感模块的字符输入方法, 其特征在 于, 所述依据所述手势轮廓模型进行对象指尖的跟踪为:  4. The image input module-based character input method according to claim 2, wherein the tracking of the object fingertip according to the gesture contour model is:
将参考的手势轮廓模型的指尖粗定位的矩形区域映射到提取的对象手 势轮廓图像上, 获得对象手势轮廓图像的指尖位置所在的矩形区域;  Mapping a rectangular area of the fingertip coarsely positioned of the reference gesture contour model to the extracted object hand contour image, and obtaining a rectangular area where the fingertip position of the object gesture contour image is located;
对所述矩形区域内的手指轮廓进行等间距划分, 并计算每段手指轮廓 片段的边缘弯曲率;  Equally dividing the contours of the fingers in the rectangular area, and calculating the edge bending rate of each segment of the finger contour;
获取边缘弯曲率最大的手指轮廓片段为对象指尖。  The finger contour segment with the largest edge curvature is obtained as the fingertip of the object.
5、 如权利要求 1所述的基于图像传感模块的字符输入方法, 其特征在 于, 所述获取对象指尖的运行轨迹信息为: 5. The image sensing module based character input method according to claim 1, wherein the character input method is The information about the running track of the fingertip of the acquiring object is:
依据卡尔曼滤波器进行对象指尖的跟踪预测;  Tracking prediction of the fingertip of the object according to the Kalman filter;
从图像传感模块采集的视频帧中获取对象指尖的起始坐标信息, 之后 平均每隔至少一帧采集一次对象指尖的实时坐标信息;  Obtaining the starting coordinate information of the fingertip of the object from the video frame collected by the image sensing module, and then collecting the real-time coordinate information of the fingertip of the object at least once every at least one frame;
计算所述实时坐标信息与上一次坐标信息的切线角度^并依据所述切 线角度的变化获取对象指尖的运行轨迹信息, 其中, 所述切线角度 0的计算 为: 其中, 所述 (CJ 1)是指对象指尖在上一次 t-i时刻的坐标
Figure imgf000019_0001
Calculating a tangential angle of the real-time coordinate information and the previous coordinate information, and acquiring trajectory information of the fingertip of the object according to the change of the tangential angle, wherein the tangential angle 0 is calculated as: where (CJ 1 ) refers to the coordinates of the fingertip of the object at the last ti time
Figure imgf000019_0001
信息, 所述 , 是指对象指尖在实时 t时刻的实时坐标信息。 Information, said, refers to the real-time coordinate information of the fingertip of the object at real time t.
6、 如权利要求 1所述的基于图像传感模块的字符输入方法, 其特征在 于, 所述依据所述运行轨迹信息查询预先训练出的字符模型库, 将所述运 行轨迹信息转化为相应的字符为:  The image input module-based character input method according to claim 1, wherein the querying the pre-trained character model library according to the running track information, converting the running track information into corresponding The characters are:
依据所述运行轨迹信息查询预先训练出的字符模型库, 并将所述运行 轨迹信息与所有的字符模型进行匹配;  Querying the pre-trained character model library according to the running track information, and matching the running track information with all character models;
获取似然值最大的字符模型为目标字符模型, 其中, 采用 Verbit算法 计算每个字符模型的似然值;  The character model with the largest likelihood value is obtained as the target character model, wherein the likelihood value of each character model is calculated by using the Verbit algorithm;
依据所述目标字符模型将所述运行轨迹信息转化为相应的字符。  The running track information is converted into corresponding characters according to the target character model.
7、一种基于图像传感模块的字符输入装置,其特征在于,该装置包括: 图像传感模块, 用于采集包含对象指尖的视频帧;  A character input device based on an image sensing module, the device comprising: an image sensing module, configured to collect a video frame including a fingertip of the object;
对象指尖运行轨迹信息获取模块, 用于通过图像传感模块采集的视频 帧获取对象指尖的运行轨迹信息;  The object fingertip running track information acquiring module is configured to acquire the running track information of the fingertip of the object through the video frame collected by the image sensing module;
字符转化模块, 用于依据所述运行轨迹信息查询预先训练出的字符模 型库, 将所述运行轨迹信息转化为相应的字符。  The character conversion module is configured to query the pre-trained character model library according to the running track information, and convert the running track information into corresponding characters.
8、 如权利要求 7所述的基于图像传感模块的字符输入装置, 其特征在 于, 该装置还包括:  8. The image sensor module-based character input device of claim 7, wherein the device further comprises:
对象手势轮廓图像信息采集模块, 用于通过图像传感模块采集对象手 势轮廓图像信息; Object gesture contour image information acquisition module, used for collecting object hands through image sensing module Potential contour image information;
手势轮廓模型获取模块, 用于依据所述对象手势轮廓图像信息查询预 先训练出的手势轮廓模型库, 获取匹配的手势轮廓模型, 使得对象指尖运 行轨迹信息获取模块依据所述手势轮廓模型进行对象指尖的跟踪, 获取对 象指尖的运行轨迹信息。  a gesture contour model obtaining module, configured to query a pre-trained gesture contour model library according to the object gesture contour image information, and obtain a matching gesture contour model, so that the object fingertip running track information acquiring module performs the object according to the gesture contour model Tracking of the fingertips, obtaining the running track information of the fingertips of the object.
9、 如权利要求 8所述的基于图像传感模块的字符输入装置, 其特征在 于, 该装置还包括:  9. The image sensor module-based character input device of claim 8, wherein the device further comprises:
图像处理模块, 用于对对象手势轮廓图像信息采集模块采集到的对象 手势轮廓图像信息进行如下处理:  The image processing module is configured to process the object gesture contour image information collected by the object gesture contour image information collection module as follows:
对采集到的对象手势轮廓图像信息进行二值化处理; 和 /或,  Binarizing the acquired object gesture contour image information; and/or,
对采集到的对象手势轮廓图像信息进行图象锐化处理; 和 /或, 对采集到的对象手势轮廓图像信息进行图象平滑处理。  Performing image sharpening processing on the collected object gesture contour image information; and/or performing image smoothing processing on the collected object gesture contour image information.
10、 如权利要求 8所述的基于图像传感模块的字符输入装置, 其特征 在于, 所述对象指尖运行轨迹信息获取模块依据所述手势轮廓模型进行对 象指尖的跟踪为:  The image sensor module-based character input device according to claim 8, wherein the object fingertip trajectory information acquisition module performs tracking of the fingertip according to the gesture contour model as:
将参考的手势轮廓模型的指尖粗定位的矩形区域映射到提取的对象手 势轮廓图像上, 获得对象手势轮廓图像的指尖位置所在的矩形区域;  Mapping a rectangular area of the fingertip coarsely positioned of the reference gesture contour model to the extracted object hand contour image, and obtaining a rectangular area where the fingertip position of the object gesture contour image is located;
对所述矩形区域内的手指轮廓进行等间距划分, 并计算每段手指轮廓 片段的边缘弯曲率;  Equally dividing the contours of the fingers in the rectangular area, and calculating the edge bending rate of each segment of the finger contour;
获取边缘弯曲率最大的手指轮廓片段为对象指尖。  The finger contour segment with the largest edge curvature is obtained as the fingertip of the object.
11、 如权利要求 8 所述的基于图像传感模块的字符输入装置, 其特征 在于, 所述对象指尖运行轨迹信息获取模块通过图像传感模块采集的视频 帧获取对象指尖的运行轨迹信息为:  The image sensor module-based character input device according to claim 8, wherein the object fingertip trajectory information acquisition module acquires the trajectory information of the fingertip of the object through the video frame acquired by the image sensing module. For:
依据卡尔曼滤波器进行对象指尖的跟踪预测;  Tracking prediction of the fingertip of the object according to the Kalman filter;
从图像传感模块采集的视频帧中获取对象指尖的起始坐标信息, 之后 平均每隔至少一帧采集一次对象指尖的实时坐标信息; Obtain the starting coordinate information of the fingertip of the object from the video frame captured by the image sensing module, and then Real-time coordinate information of the fingertip of the object is collected once every at least one frame;
计算所述实时坐标信息与上一次坐标信息的切线角度^并依据所述切 线角度的变化获取对象指尖的运行轨迹信息, 其中, 所述切线角度 0的计算 为: 其中, 所述 (CJ 1)是指对象指尖在上一次 t-i时刻的坐标
Figure imgf000021_0001
Calculating a tangential angle of the real-time coordinate information and the previous coordinate information, and acquiring trajectory information of the fingertip of the object according to the change of the tangential angle, wherein the tangential angle 0 is calculated as: where (CJ 1 ) refers to the coordinates of the fingertip of the object at the last ti time
Figure imgf000021_0001
信息, 所述 , 是指对象指尖在实时 t时刻的实时坐标信息。 Information, said, refers to the real-time coordinate information of the fingertip of the object at real time t.
12、 如权利要求 8所述的基于图像传感模块的字符输入装置, 其特征 在于, 所述字符转化模块依据所述运行轨迹信息查询预先训练出的字符模 型库, 以将所述运行轨迹信息转化为相应的字符为:  The character input module of the image sensing module according to claim 8, wherein the character conversion module queries the pre-trained character model library according to the running track information to use the running track information Convert to the corresponding characters as:
依据所述运行轨迹信息查询预先训练出的字符模型库, 并将所述运行 轨迹信息与所有的字符模型进行匹配;  Querying the pre-trained character model library according to the running track information, and matching the running track information with all character models;
获取似然值最大的字符模型为目标字符模型, 其中, 采用 Verbit算法 计算每个字符模型的似然值;  The character model with the largest likelihood value is obtained as the target character model, wherein the likelihood value of each character model is calculated by using the Verbit algorithm;
依据所述目标字符模型将所述运行轨迹信息转化为相应的字符。  The running track information is converted into corresponding characters according to the target character model.
13、 一种终端, 其特征在于, 该终端包括权利要求 7至 12任一项所述 的基于图像传感模块的字符输入装置。  A terminal, characterized in that the terminal comprises the image sensing module based character input device according to any one of claims 7 to 12.
PCT/CN2012/075103 2011-11-23 2012-05-04 Character input method, device and terminal based on image sensing module WO2013075466A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201110376331.1 2011-11-23
CN2011103763311A CN102520790A (en) 2011-11-23 2011-11-23 Character input method based on image sensing module, device and terminal

Publications (1)

Publication Number Publication Date
WO2013075466A1 true WO2013075466A1 (en) 2013-05-30

Family

ID=46291742

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2012/075103 WO2013075466A1 (en) 2011-11-23 2012-05-04 Character input method, device and terminal based on image sensing module

Country Status (2)

Country Link
CN (1) CN102520790A (en)
WO (1) WO2013075466A1 (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110865767A (en) * 2019-11-20 2020-03-06 深圳传音控股股份有限公司 Application program running method, device, equipment and storage medium
CN111311590A (en) * 2020-03-06 2020-06-19 通控研究院(安徽)有限公司 Turnout switch point tightness detection method based on image detection technology
CN111382598A (en) * 2018-12-27 2020-07-07 北京搜狗科技发展有限公司 Identification method and device and electronic equipment
CN111627039A (en) * 2020-05-09 2020-09-04 北京小狗智能机器人技术有限公司 Interaction system and interaction method based on image recognition
CN113591822A (en) * 2021-10-08 2021-11-02 广州市简筱网络科技有限公司 Special crowd gesture interaction information consultation and recognition system

Families Citing this family (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103064530B (en) * 2012-12-31 2017-03-08 华为技术有限公司 input processing method and device
CN103941846B (en) * 2013-01-18 2017-11-24 联想(北京)有限公司 Electronic equipment and input method
CN103235836A (en) * 2013-05-07 2013-08-07 西安电子科技大学 Method for inputting information through mobile phone
CN103577843B (en) * 2013-11-22 2016-06-22 中国科学院自动化研究所 A kind of aerial hand-written character string recognition methods
CN104123008B (en) * 2014-07-30 2017-11-03 哈尔滨工业大学深圳研究生院 A kind of man-machine interaction method and system based on static gesture
CN104834412B (en) * 2015-05-13 2018-02-23 深圳市蓝晨科技股份有限公司 A kind of touch terminal based on contactless gesture identification
CN105094544B (en) * 2015-07-16 2020-03-03 百度在线网络技术(北京)有限公司 Method and device for acquiring characters
JP6741008B2 (en) * 2015-08-19 2020-08-19 ソニー株式会社 Information processing apparatus, information processing method, and program
CN105975934B (en) * 2016-05-05 2022-03-04 中国人民解放军63908部队 Dynamic gesture recognition method and system for augmented reality auxiliary maintenance
CN107643821B (en) * 2016-07-22 2021-07-27 北京搜狗科技发展有限公司 Input control method and device and electronic equipment
CN109344793B (en) * 2018-10-19 2021-03-16 北京百度网讯科技有限公司 Method, apparatus, device and computer readable storage medium for recognizing handwriting in the air
CN112357703B (en) * 2020-10-27 2022-09-13 广州广日电梯工业有限公司 Elevator control system and elevator control method

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101593022A (en) * 2009-06-30 2009-12-02 华南理工大学 A kind of quick human-computer interaction of following the tracks of based on finger tip
CN101739118A (en) * 2008-11-06 2010-06-16 大同大学 Video handwriting character inputting device and method thereof
CN102063618A (en) * 2011-01-13 2011-05-18 中科芯集成电路股份有限公司 Dynamic gesture identification method in interactive system
CN102184021A (en) * 2011-05-27 2011-09-14 华南理工大学 Television man-machine interaction method based on handwriting input and fingertip mouse

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE602004006190T8 (en) * 2003-03-31 2008-04-10 Honda Motor Co., Ltd. Device, method and program for gesture recognition

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101739118A (en) * 2008-11-06 2010-06-16 大同大学 Video handwriting character inputting device and method thereof
CN101593022A (en) * 2009-06-30 2009-12-02 华南理工大学 A kind of quick human-computer interaction of following the tracks of based on finger tip
CN102063618A (en) * 2011-01-13 2011-05-18 中科芯集成电路股份有限公司 Dynamic gesture identification method in interactive system
CN102184021A (en) * 2011-05-27 2011-09-14 华南理工大学 Television man-machine interaction method based on handwriting input and fingertip mouse

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
GUO, KANGDE ET AL.: "3D Fingertip Tracking Algorithm Based on Computer Vision", JOURNAL OF COMPUTER RESEARCH AND DEVELOPMENT, 15 June 2010 (2010-06-15), pages 1017 *

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111382598A (en) * 2018-12-27 2020-07-07 北京搜狗科技发展有限公司 Identification method and device and electronic equipment
CN111382598B (en) * 2018-12-27 2024-05-24 北京搜狗科技发展有限公司 Identification method and device and electronic equipment
CN110865767A (en) * 2019-11-20 2020-03-06 深圳传音控股股份有限公司 Application program running method, device, equipment and storage medium
CN111311590A (en) * 2020-03-06 2020-06-19 通控研究院(安徽)有限公司 Turnout switch point tightness detection method based on image detection technology
CN111311590B (en) * 2020-03-06 2023-11-21 通控研究院(安徽)有限公司 Switch point contact degree detection method based on image detection technology
CN111627039A (en) * 2020-05-09 2020-09-04 北京小狗智能机器人技术有限公司 Interaction system and interaction method based on image recognition
CN113591822A (en) * 2021-10-08 2021-11-02 广州市简筱网络科技有限公司 Special crowd gesture interaction information consultation and recognition system
CN113591822B (en) * 2021-10-08 2022-02-08 广州市简筱网络科技有限公司 Special crowd gesture interaction information consultation and recognition system

Also Published As

Publication number Publication date
CN102520790A (en) 2012-06-27

Similar Documents

Publication Publication Date Title
WO2013075466A1 (en) Character input method, device and terminal based on image sensing module
EP3258423B1 (en) Handwriting recognition method and apparatus
CN102231093B (en) Screen locating control method and device
CN106774850B (en) Mobile terminal and interaction control method thereof
CN107885327B (en) Fingertip detection method based on Kinect depth information
TWI382352B (en) Video based handwritten character input device and method thereof
CN103294996A (en) 3D gesture recognition method
Fang et al. A novel video-based system for in-air signature verification
CN107357414B (en) Click action recognition method and device
WO2018076182A1 (en) Image capturing method and device
WO2015090075A1 (en) Touch screen terminal and method and system for identifying hand gestures of touch screen terminal
CN107450717B (en) Information processing method and wearable device
TW201317843A (en) Virtual mouse driving apparatus and virtual mouse simulation method
CN103677254A (en) Methods and apparatus for documenting a procedure
CN112286360A (en) Method and apparatus for operating a mobile device
KR101433543B1 (en) Gesture-based human-computer interaction method and system, and computer storage media
WO2016145827A1 (en) Terminal control method and device
CN101739118A (en) Video handwriting character inputting device and method thereof
JP6051991B2 (en) Signal processing apparatus, signal processing method, and signal processing program
WO2012049645A1 (en) Method and apparatus for determining motion with normalized representations of movement
CN113961067B (en) Non-contact doodling drawing method and recognition interaction system based on deep learning
TWI426420B (en) Hand track identification system and method thereof
US9952671B2 (en) Method and apparatus for determining motion
Meng et al. Building smart cameras on mobile tablets for hand gesture recognition
TWI775128B (en) Gesture control device and control method thereof

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 12852341

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 12852341

Country of ref document: EP

Kind code of ref document: A1