WO2019000817A1 - Control method and electronic equipment for hand gesture recognition - Google Patents

Control method and electronic equipment for hand gesture recognition Download PDF

Info

Publication number
WO2019000817A1
WO2019000817A1 PCT/CN2017/112312 CN2017112312W WO2019000817A1 WO 2019000817 A1 WO2019000817 A1 WO 2019000817A1 CN 2017112312 W CN2017112312 W CN 2017112312W WO 2019000817 A1 WO2019000817 A1 WO 2019000817A1
Authority
WO
WIPO (PCT)
Prior art keywords
arm
image
specific user
gesture input
gesture
Prior art date
Application number
PCT/CN2017/112312
Other languages
French (fr)
Chinese (zh)
Inventor
丁琦城
吴梦
Original Assignee
联想(北京)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 联想(北京)有限公司 filed Critical 联想(北京)有限公司
Publication of WO2019000817A1 publication Critical patent/WO2019000817A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • G06V40/28Recognition of hand or arm movements, e.g. recognition of deaf sign language
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/017Gesture based interaction, e.g. based on a set of recognized hand gestures

Definitions

  • the present invention relates to the technical field of gesture recognition, and more particularly to a gesture recognition control method capable of distinguishing gesture input of a specific user and an electronic device to which the method is applied.
  • gesture recognition modules have been added to more and more electronic devices.
  • the user's experience can be improved by triggering the corresponding instruction by the recognized gesture.
  • the electronic device regards the gesture of another person as a gesture of a specific user, thereby causing an erroneous reaction.
  • a gesture recognition control method for an electronic device, the method comprising: acquiring a first image when a gesture input is performed, the first image including a hand and at least a part of an arm An image; based on the first image, determining whether the gesture input is made by a specific user; and if the determination result is yes, performing recognition on the gesture input, otherwise ignoring the gesture input.
  • the method according to an embodiment of the present invention may further include: pre-storing a three-dimensional model of a specific user's arm for performing a gesture operation; wherein based on the first image, determining whether the gesture input is made by a specific user
  • the step further includes: determining whether the arm included in the first image is a left arm or a right arm, and decomposing different parts of the arm in the first image, extracting features of different parts; and acquiring features of different parts of the arm A comparison is made with features in the corresponding three-dimensional model and it is determined whether the gesture input is made by the particular user.
  • the step of pre-storing a three-dimensional model of a specific user's arm for performing a gesture operation further comprises: acquiring a plurality of arms of the specific user in a plurality of gesture operations Image For each of the plurality of arm images, different parts of the arm are decomposed and features of different parts are extracted; and a plurality of features of the same part are fused for the same arm, and a three-dimensional model for the arm is obtained.
  • the step of comparing the acquired features of the different parts of the arm with the features in the corresponding three-dimensional model further comprises: determining the size and the different parts of the arm in the first image Whether the difference between the corresponding sizes in the three-dimensional model is less than a predetermined threshold, obtaining a first determination result; and/or determining whether an angle between the forearm and the boom in the first image is determined based on the three-dimensional model Within a predetermined range, obtaining a second determination result, wherein determining whether the gesture input is made by the specific user comprises: determining whether the gesture input is based on the first determination result and/or the second determination result Made by the specific user.
  • the method according to an embodiment of the present invention may further include: after determining that the gesture input is made by a specific user, determining Whether the user's undo operation is received; after determining that the gesture input is made by a specific user, determining whether the recognized gesture input is completely unrelated to the current task being executed by the electronic device; determining whether the first image is There are other people appearing; after determining that the gesture input is not made by a specific user, determining whether to receive the repeated input of the gesture; based on at least one of the above determination results, determining whether the gesture input is made by a specific user Whether the determination is correct to update the corresponding three-dimensional model with the arm features in the first image.
  • an electronic device includes: an image acquisition module, configured to acquire a first image when a gesture input is performed, the first image including an image of a hand and at least a portion of an arm; and a memory And a processor configured to, when executing the program stored on the memory, implement a function of: determining, based on the first image, whether the gesture input is made by a specific user; and if the determination result is Yes, the recognition is performed on the gesture input, otherwise the gesture input is ignored.
  • the memory is further configured to pre-store a three-dimensional model of an arm of a specific user for performing a gesture operation; wherein the processor is configured to execute the program to further Implementing a function of determining whether the arm included in the first image is a left arm or a right arm, and decomposing different parts of the arm in the first image, extracting features of different parts; and acquiring features of different parts of the arm A comparison is made with features in the corresponding three-dimensional model and it is determined whether the gesture input is made by the particular user.
  • the processor is configured to execute the program to further implement the following functions:
  • a plurality of features of the same portion are fused for the same arm, and a three-dimensional model for the arm is obtained.
  • the processor is configured to execute the program to further implement The function of determining whether a difference between a size of a different part of the arm in the first image and a corresponding size in the three-dimensional model is less than a predetermined threshold, obtaining a first determination result; and/or determining the first image Whether the angle between the middle forearm and the boom is within a predetermined range determined based on the three-dimensional model, obtaining a second determination result; and determining the gesture input based on the first determination result and/or the second determination result Whether it is made by the specific user.
  • the processor is configured to execute the program to further implement a function of: determining whether the user's revocation is received after determining that the gesture input is made by a specific user After determining that the gesture input is made by a specific user, determining whether the recognized gesture input is completely unrelated to the current task being executed by the electronic device; determining whether another person appears in the first image; Determining whether the gesture input is not made by a specific user, determining whether a repeated input of the gesture is received; and determining, based on at least one of the above determination results, whether the determination as to whether the gesture input is made by a specific user is correct And updating the corresponding three-dimensional model with the arm feature in the first image.
  • the accuracy of a specific user's own gesture recognition is improved by image processing of the entire arm instead of a single palm.
  • the modeling process is completed and perfected by using the data determined as the arm of the specific user during use, so that the more the three-dimensional model of the arm is used, the higher the accuracy.
  • no unnecessary device is added or the user's use burden is increased (for example, the user is actively selected whether the gesture is made by himself), thereby reducing the cost and improving the cost. user experience.
  • FIG. 1 is a flowchart illustrating a procedure of a gesture recognition control method according to an embodiment of the present invention
  • FIG. 2 is a functional block diagram illustrating a configuration of an electronic device according to an embodiment of the present invention.
  • the techniques of this disclosure may be implemented in the form of hardware and/or software (including firmware, microcode, etc.). Additionally, the techniques of this disclosure may take the form of a computer program product on a computer readable medium storing instructions for use by or in connection with an instruction execution system.
  • a computer readable medium can be any medium that can contain, store, communicate, propagate or transport the instructions.
  • a computer readable medium can include, but is not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, device, or propagation medium.
  • the computer readable medium include: a magnetic storage device such as a magnetic tape or a hard disk (HDD); an optical storage device such as a compact disk (CD-ROM); a memory such as a random access memory (RAM) or a flash memory; and/or a wired /Wireless communication link.
  • a magnetic storage device such as a magnetic tape or a hard disk (HDD)
  • an optical storage device such as a compact disk (CD-ROM)
  • a memory such as a random access memory (RAM) or a flash memory
  • RAM random access memory
  • the gesture recognition control method is applied to an electronic device.
  • the electronic device can be a cell phone, a tablet, a smart TV, or a wearable device.
  • the gesture recognition control method includes the following steps.
  • step S101 a first image when a gesture input is performed is acquired, the first image including an image of a hand and at least a part of an arm.
  • step S102 based on the first image, it is determined whether the gesture input is made by a specific user.
  • step S102 If the result of the determination in step S102 is YES, the process proceeds to step S103.
  • step S103 the gesture is lost Into the execution identification. Then, based on the recognized gesture, the corresponding function is activated.
  • step S104 the gesture input is ignored. That is, the gesture input is not recognized and responded.
  • the gesture recognition control method when the user performs gesture input, not only the hand image but also at least part of the arm image is acquired, compared with the prior art.
  • the image acquisition module eg, depth camera
  • the gesture recognition control method can determine whether a gesture is made for a specific user based on the arm image, avoiding the erroneous recognition of gestures of other users that should not be recognized.
  • the gesture recognition method in an application scenario where the electronic device is a mobile phone, a tablet computer, a smart TV, or the like, in the gesture recognition method according to the prior art, a specific user can be verified by additional facial recognition, and only the hand is collected after the verification is passed.
  • Image for gesture recognition That is to say, in the gesture recognition method according to the related art, it is possible to confirm whether it is a specific user by face recognition, but it is not possible to confirm whether the current gesture is made by the identified specific user.
  • the gesture recognition control method according to the present invention it is possible to confirm whether the gesture input is made by a specific user, thereby avoiding a misjudgment caused when a specific user's face is recognized but the gesture is made by another person.
  • head-mounted display devices will be the most promising interactive interface rendering devices in augmented reality display devices.
  • the electronic device is such a head mounted display device
  • the frequency and severity of such misrecognition will be higher.
  • determining whether the gesture input is made by a specific user may include, but is not limited to, the following manners: (1) determining whether the gesture is satisfied by a specific user based on the position of the arm in the first image. Rationality; (2) Based on the three-dimensional model of the arm to determine whether the similarity with the gesture input by the specific user is satisfied.
  • step S102 in FIG. 1 may further include: confirming whether it is a specific user based on a relative relationship between the arm and the verified face in the first image. The gesture made. If the distance between the arm and the face exceeds a predetermined threshold, or if the angle between the arm and the face exceeds a predetermined angular range, then the relative relationship between the arm and the face is unreasonable, and it can be determined that the gesture input is not Made by a specific user. On the other hand, if it is not judged that the relative relationship between the arm and the face is unreasonable, then the gesture input is temporarily considered to be made by a specific user.
  • step S102 in FIG. 1 may further include first determining whether the arm in which the gesture input is performed in the first image is the left arm or the right arm.
  • the left arm it is determined whether the left arm appears in the right area, and if so, the position of the arm is unreasonable, and it can be determined that the gesture input is not made by a specific user. On the other hand, if it is not judged that the positional relationship of the arm is unreasonable, it is temporarily considered that the gesture input is made by a specific user. Similarly, if it is determined to be the right arm, it is determined whether the right arm is present in the left area, and if so, the position of the arm is unreasonable, and it can be determined that the gesture input is not made by a specific user. On the other hand, if it is not judged that the positional relationship of the arm is unreasonable, it is temporarily considered that the gesture input is made by a specific user.
  • the step of pre-storing the three-dimensional model for the arm of the specific user further comprises: acquiring a plurality of arm images of the specific user in the multiple gesture operation; and decomposing the arm for each of the plurality of arm images Different parts, and extracting features of different parts; and merging multiple features of the same part for the same arm, and obtaining a three-dimensional model for the arm.
  • the recognition range of the depth camera and the machine learning algorithm are used to record the characteristics of the specific user's arm at various angles, and the whole body is learned and modeled, not only It is only the recognition of finger movements.
  • the degree of integrity of the arm that can be obtained is different.
  • the three-dimensional data of all the arms is gradually obtained. After a certain amount of data learning, that is, after acquiring a sufficient arm image of a specific user in multiple gesture operations, a model for a specific user's arm can be established.
  • Different parts of the arm include: hand, forearm, and big arm.
  • the three-dimensional model of the arm includes: size information of the forearm and the big arm, angular range of the forearm and the big arm, and characteristic images on the forearm and the big arm (eg, personalized joints, bumps, etc.).
  • the model can be used to determine if it is a particular user's own gesture.
  • Step S102 in FIG. 1 may further include: determining whether the arm included in the first image is a left arm or a right arm, and decomposing different parts of the arm in the first image, extracting features of different parts; and acquiring The features of the different parts of the arm are compared to the features in the corresponding three-dimensional model and it is determined whether the gesture input is made by the particular user.
  • the step of comparing the acquired features of different parts of the arm with the features in the corresponding three-dimensional model further includes:
  • determining whether the gesture input is made by the specific user comprises determining whether the gesture input is made by the specific user based on the first determination result and/or the second determination result. Specifically, if there is a judgment result of the first judgment result and/or the second judgment result, it is determined that the gesture input is not made by the specific user, otherwise it is determined that the gesture input is performed by the specific user. Out.
  • the two methods described above for determining whether the gesture input is made by a particular user may be used separately. Alternatively, these two methods can also be used serially. For example, first judge in terms of rationality and then judge in terms of similarity. Since the calculation amount of the rationality judgment is small, the rationality is used as the screening condition, and the calculation amount based on the judgment of the three-dimensional model of the arm can be effectively reduced.
  • the principle of "doubt is never" is adopted. That is, as long as it is not certain that the gesture input is not made by a particular user, then the gesture input is temporarily considered to be made by a particular user. For example, when it is not judged to be unreasonable based on the first image, or when it is not possible to determine whether the gesture input is made by a specific user based on the current arm three-dimensional model, then the gesture input is temporarily considered to be made by a specific user.
  • the gesture recognition control method after determining whether the gesture input is made by a specific user based on the first image, Further includes a verification step.
  • the verifying step may include the following processing:
  • the gesture input After determining that the gesture input is made by a specific user, it is determined whether the user's undo operation is received. In general, if the user's undo operation is received after determining that the gesture input is made by a particular user, then the gesture input is most likely not made by a particular user, ie, the probability of a false positive is greater.
  • a repeated input of the gesture is received. In general, if a repeated input of the gesture is received after determining that the gesture input is not made by a particular user, then the gesture input is likely to be made by a particular user, ie, the probability of a false positive is greater.
  • the arm features in the first image are added to the corresponding three-dimensional model.
  • the corresponding three-dimensional model is modified based on the arm features in the first image.
  • the corresponding three-dimensional model is modified to exclude the arm feature in the three-dimensional model, that is, if the arm feature is detected again, a gesture that is not made by a specific user is determined.
  • the electronic device 200 includes an image acquisition module 201, a memory 202, and a processor 203.
  • the image acquisition module 201 is configured to acquire a first image when the gesture input is performed, wherein the first image includes an image of the hand and at least a part of the arm.
  • the image acquisition module 201 can be implemented by a depth camera.
  • the memory 202 is used to store programs.
  • the processor 203 is configured to implement the following functions when executing the program stored on the memory 202:
  • determining whether the gesture input is made by a specific user may include, but is not limited to, the following manners: (1) determining whether the gesture is satisfied by a specific user based on the position of the arm in the first image. Rationality (2) Based on the three-dimensional model of the arm, it is judged whether the similarity with the gesture input by the specific user is satisfied.
  • the processor 203 can execute the program to further implement a function of confirming based on a relative relationship between the arm and the verified face in the first image. Whether it is a gesture made by a specific user. If the distance between the arm and the face exceeds a predetermined threshold, or if the angle between the arm and the face exceeds a predetermined angular range, then the relative relationship between the arm and the face is unreasonable, and it can be determined that the gesture input is not by a specific user. Made. On the other hand, if it is not judged that the relative relationship between the arm and the face is unreasonable, then the gesture input is temporarily considered to be made by a specific user.
  • the electronic device is a head mounted display device
  • a specific user in this case, a wearer
  • the processor 203 can execute the program to further implement a function of first determining whether the arm in which the gesture input is made in the first image is the left arm or the right arm.
  • the left arm it is determined whether the left arm appears in the right area, and if so, the position of the arm is unreasonable, and it can be determined that the gesture input is not made by a specific user. On the other hand, if it is not judged that the positional relationship of the arm is unreasonable, it is temporarily considered that the gesture input is made by a specific user. Similarly, if it is determined to be the right arm, it is determined whether the right arm is present in the left area, and if so, the position of the arm is unreasonable, and it can be determined that the gesture input is not made by a specific user. On the other hand, if it is not judged that the positional relationship of the arm is unreasonable, it is temporarily considered that the gesture input is made by a specific user.
  • the memory is further configured to pre-store a three-dimensional model of the arm of the specific user for performing the gesture operation.
  • two three-dimensional models of the arms of the specific user may be stored in advance so that the gesture of any arm of the user can be recognized.
  • the processor is configured to execute the program to further implement a function of: decomposing an arm of each of the plurality of arm images of the specific user acquired in the plurality of gesture operations by the image acquisition unit Different parts, and extracting features of different parts; and merging multiple features of the same part for the same arm, and obtaining a three-dimensional model for the arm.
  • the model can be used to determine if it is a particular user's own gesture.
  • the processor is configured to execute the program to further implement the following functions:
  • the acquired features of different parts of the arm are compared with features in the corresponding three-dimensional model, and it is determined whether the gesture input is made by the specific user.
  • comparing the acquired features of the different parts of the arm with the features in the corresponding three-dimensional model, and determining whether the gesture input is processed by the specific user further comprises:
  • the gesture input is made by the specific user. Specifically, if there is a judgment result of the first judgment result and/or the second judgment result, it is determined that the gesture input is not made by the specific user, otherwise it is determined that the gesture input is performed by the specific user. Out.
  • the processor is configured to execute the program to further implement the following functions:
  • the arm features in the first image are added to the corresponding three-dimensional model.
  • the corresponding three-dimensional model is modified based on the arm features in the first image.
  • the corresponding three-dimensional model is modified to exclude the arm feature in the three-dimensional model, that is, if the arm feature is detected again, a gesture that is not made by a specific user is determined.
  • Processor 203 may include a general purpose microprocessor, an instruction set processor, and/or a related chipset, in accordance with an embodiment of the present disclosure. And/or a dedicated microprocessor (eg, an application specific integrated circuit (ASIC)), and the like. Processor 203 may also include an onboard memory for caching purposes.
  • the processor 203 may be a single processing unit or a plurality of processing units for performing different actions of the method flow according to the embodiment of the present disclosure described with reference to FIG.
  • Memory 202 can be any medium that can contain, store, communicate, propagate, or transport the instructions.
  • memory 202 can include, but is not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, device, or propagation medium.
  • Specific examples of the memory 202 include: a magnetic storage device such as a magnetic tape or a hard disk (HDD); an optical storage device such as a compact disk (CD-ROM); a memory such as a random access memory (RAM) or a flash memory; and/or a wired/wireless Communication link.
  • the gesture recognition control method and electronic apparatus according to the present invention have been described in detail with reference to FIGS. 1 and 2.
  • the accuracy of a specific user's own gesture recognition is improved by image processing of the entire arm instead of a single palm.
  • the modeling process is completed and perfected by using the data determined as the arm of the specific user during use, so that the more the three-dimensional model of the arm is used, the higher the accuracy.
  • no unnecessary redundancy is added or the user's use burden is increased (for example, the user is actively selected whether the gesture is made by himself), thereby reducing costs and improving The user experience.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Health & Medical Sciences (AREA)
  • Psychiatry (AREA)
  • Social Psychology (AREA)
  • Multimedia (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

Disclosed are a control method and an electronic equipment for hand gesture recognition. The hand gesture recognition control method is applicable to the electronic equipment, wherein the method comprises: acquiring a first image from when hand gesture input is carried out, the first image comprising an image of a hand and at least part of an arm; on the basis of the first image, determining whether the hand gesture input is performed by a specific user; and if the determination result is yes, then performing recognition on the hand gesture input; if not, ignoring the hand gesture input.

Description

手势识别控制方法和电子设备Gesture recognition control method and electronic device 技术领域Technical field
本发明涉及手势识别的技术领域,更具体地说,涉及能够区分特定用户的手势输入的手势识别控制方法和应用所述方法的电子设备。The present invention relates to the technical field of gesture recognition, and more particularly to a gesture recognition control method capable of distinguishing gesture input of a specific user and an electronic device to which the method is applied.
背景技术Background technique
目前,在越来越多的电子设备中增加了手势识别模块。通过识别出的手势来触发对应的指令,能够提升用户的使用体验。在这样的电子设备的使用过程中,可能仅希望对于特定用户的手势输入进行识别。然而,除了能够获取特定用户的用于手势输入的手部图像之外,还经常能够获取到特定用户以外的其他用户的手部图像。从而,会出现电子设备将他人的手势当作特定用户的手势的情况,进而导致错误的反应。Currently, gesture recognition modules have been added to more and more electronic devices. The user's experience can be improved by triggering the corresponding instruction by the recognized gesture. During use of such an electronic device, it may only be desirable to identify the gesture input for a particular user. However, in addition to being able to acquire a hand image for a gesture input by a specific user, it is often possible to acquire a hand image of a user other than a specific user. Thus, there is a case where the electronic device regards the gesture of another person as a gesture of a specific user, thereby causing an erroneous reaction.
作为一种可能的解决方案,通过使用数据手套或者带有传感器的控制器来表明特定用户的数据来源。但这种方案的缺点在于,虽然可以有效地区分是否是特定用户的手势,但是增加了更多的附件(手套或者控制器),一方面提高了成本,一方面也降低了用户的便利性和总体体验。As a possible solution, data sources for specific users are indicated by using data gloves or controllers with sensors. However, the disadvantage of this scheme is that although it can effectively distinguish whether it is a gesture of a specific user, more accessories (gloves or controllers) are added, which increases the cost on the one hand and reduces the convenience of the user on the other hand. Overall experience.
发明内容Summary of the invention
鉴于以上情形,期望提供在不增加多余设备或不增加用户负担的情况下,能够区分特定用户和其他用户的手势的手势识别控制方法和设备。In view of the above circumstances, it is desirable to provide a gesture recognition control method and apparatus capable of distinguishing gestures of specific users and other users without adding redundant equipment or increasing user burden.
根据本发明的一个方面,提供了一种手势识别控制方法,应用于一电子设备,所述方法包括:获取进行手势输入时的第一图像,所述第一图像包括手部和至少部分手臂的图像;基于所述第一图像,判断所述手势输入是否由特定用户做出;以及如果判断结果为是,则对所述手势输入执行识别,否则忽略所述手势输入。According to an aspect of the present invention, a gesture recognition control method is provided for an electronic device, the method comprising: acquiring a first image when a gesture input is performed, the first image including a hand and at least a part of an arm An image; based on the first image, determining whether the gesture input is made by a specific user; and if the determination result is yes, performing recognition on the gesture input, otherwise ignoring the gesture input.
优选地,根据本发明实施例的方法可以进一步包括:预先存储特定用户的用于进行手势操作的手臂的三维模型;其中基于所述第一图像,判断所述手势输入是否由特定用户做出的步骤进一步包括:确定所述第一图像中包括的手臂为左臂还是右臂,并且在所述第一图像中分解手臂的不同部位,提取不同部位的特征;以及将获取的手臂不同部位的特征与对应的三维模型中的特征进行比较,并判断所述手势输入是否由所述特定用户做出。Preferably, the method according to an embodiment of the present invention may further include: pre-storing a three-dimensional model of a specific user's arm for performing a gesture operation; wherein based on the first image, determining whether the gesture input is made by a specific user The step further includes: determining whether the arm included in the first image is a left arm or a right arm, and decomposing different parts of the arm in the first image, extracting features of different parts; and acquiring features of different parts of the arm A comparison is made with features in the corresponding three-dimensional model and it is determined whether the gesture input is made by the particular user.
优选地,在根据本发明实施例的方法中,所述预先存储特定用户的用于进行手势操作的手臂的三维模型的步骤进一步包括:获取所述特定用户在多次手势操作中的多个手臂图像; 针对所述多个手臂图像中的每一个,分解手臂的不同部位,并提取不同部位的特征;以及针对同一手臂融合相同部位的多个特征,并获得针对该手臂的三维模型。Preferably, in the method according to the embodiment of the present invention, the step of pre-storing a three-dimensional model of a specific user's arm for performing a gesture operation further comprises: acquiring a plurality of arms of the specific user in a plurality of gesture operations Image For each of the plurality of arm images, different parts of the arm are decomposed and features of different parts are extracted; and a plurality of features of the same part are fused for the same arm, and a three-dimensional model for the arm is obtained.
优选地,在根据本发明实施例的方法中,将获取的手臂不同部位的特征与对应的三维模型中的特征进行比较的步骤进一步包括:判断所述第一图像中手臂不同部位的尺寸与所述三维模型中的对应尺寸之间的差值是否小于预定阈值,获得第一判断结果;和/或判断所述第一图像中前臂与大臂之间的角度是否处于基于所述三维模型确定的预定范围之内,获得第二判断结果,其中判断所述手势输入是否由所述特定用户做出的步骤包括:基于所述第一判断结果和/或第二判断结果,确定所述手势输入是否由所述特定用户做出。Preferably, in the method according to the embodiment of the invention, the step of comparing the acquired features of the different parts of the arm with the features in the corresponding three-dimensional model further comprises: determining the size and the different parts of the arm in the first image Whether the difference between the corresponding sizes in the three-dimensional model is less than a predetermined threshold, obtaining a first determination result; and/or determining whether an angle between the forearm and the boom in the first image is determined based on the three-dimensional model Within a predetermined range, obtaining a second determination result, wherein determining whether the gesture input is made by the specific user comprises: determining whether the gesture input is based on the first determination result and/or the second determination result Made by the specific user.
优选地,在基于所述第一图像,判断出所述手势输入是否由特定用户做出之后,根据本发明实施例的方法可以进一步包括:在确定所述手势输入由特定用户做出之后,判断是否接收到用户的撤销操作;在确定所述手势输入由特定用户做出之后,判断识别出的手势输入是否与所述电子设备正在执行的当前任务完全不相关;判断所述第一图像中是否有其他人出现;在确定所述手势输入不是由特定用户做出之后,判断是否接收到所述手势的重复输入;基于以上判断结果中的至少一个,确定关于所述手势输入是否由特定用户做出的判断是否正确,以利用所述第一图像中的手臂特征更新对应的三维模型。Preferably, after determining whether the gesture input is made by a specific user based on the first image, the method according to an embodiment of the present invention may further include: after determining that the gesture input is made by a specific user, determining Whether the user's undo operation is received; after determining that the gesture input is made by a specific user, determining whether the recognized gesture input is completely unrelated to the current task being executed by the electronic device; determining whether the first image is There are other people appearing; after determining that the gesture input is not made by a specific user, determining whether to receive the repeated input of the gesture; based on at least one of the above determination results, determining whether the gesture input is made by a specific user Whether the determination is correct to update the corresponding three-dimensional model with the arm features in the first image.
根据本发明的另一方面,提供了一种电子设备,包括:图像采集模组,用于获取进行手势输入时的第一图像,所述第一图像包括手部和至少部分手臂的图像;存储器,用于存储程序;以及处理器,用于当执行所述存储器上存储的程序时实现如下功能:基于所述第一图像,判断所述手势输入是否由特定用户做出;以及如果判断结果为是,则对所述手势输入执行识别,否则忽略所述手势输入。According to another aspect of the present invention, an electronic device includes: an image acquisition module, configured to acquire a first image when a gesture input is performed, the first image including an image of a hand and at least a portion of an arm; and a memory And a processor configured to, when executing the program stored on the memory, implement a function of: determining, based on the first image, whether the gesture input is made by a specific user; and if the determination result is Yes, the recognition is performed on the gesture input, otherwise the gesture input is ignored.
优选地,在根据本发明实施例的设备中,所述存储器进一步被配置为预先存储特定用户的用于进行手势操作的手臂的三维模型;其中所述处理器被配置为执行所述程序以进一步实现如下功能:确定所述第一图像中包括的手臂为左臂还是右臂,并且在所述第一图像中分解手臂的不同部位,提取不同部位的特征;以及将获取的手臂不同部位的特征与对应的三维模型中的特征进行比较,并判断所述手势输入是否由所述特定用户做出。Preferably, in the device according to an embodiment of the present invention, the memory is further configured to pre-store a three-dimensional model of an arm of a specific user for performing a gesture operation; wherein the processor is configured to execute the program to further Implementing a function of determining whether the arm included in the first image is a left arm or a right arm, and decomposing different parts of the arm in the first image, extracting features of different parts; and acquiring features of different parts of the arm A comparison is made with features in the corresponding three-dimensional model and it is determined whether the gesture input is made by the particular user.
优选地,在根据本发明实施例的设备中,所述处理器被配置为执行所述程序以进一步实现如下功能:Preferably, in the device according to an embodiment of the invention, the processor is configured to execute the program to further implement the following functions:
针对所述图像采集单元获取的所述特定用户在多次手势操作中的多个手臂图像中的每一个,分解手臂的不同部位,并提取不同部位的特征;以及Separating different parts of the arm and extracting features of different parts for each of the plurality of arm images of the specific user acquired by the image acquisition unit in the plurality of gesture operations;
针对同一手臂融合相同部位的多个特征,并获得针对该手臂的三维模型。A plurality of features of the same portion are fused for the same arm, and a three-dimensional model for the arm is obtained.
优选地,在根据本发明实施例的设备中,所述处理器被配置为执行所述程序以进一步实 现如下功能:判断所述第一图像中手臂不同部位的尺寸与所述三维模型中的对应尺寸之间的差值是否小于预定阈值,获得第一判断结果;和/或判断所述第一图像中前臂与大臂之间的角度是否处于基于所述三维模型确定的预定范围之内,获得第二判断结果;以及基于所述第一判断结果和/或第二判断结果,确定所述手势输入是否由所述特定用户做出。Preferably, in the device according to an embodiment of the invention, the processor is configured to execute the program to further implement The function of determining whether a difference between a size of a different part of the arm in the first image and a corresponding size in the three-dimensional model is less than a predetermined threshold, obtaining a first determination result; and/or determining the first image Whether the angle between the middle forearm and the boom is within a predetermined range determined based on the three-dimensional model, obtaining a second determination result; and determining the gesture input based on the first determination result and/or the second determination result Whether it is made by the specific user.
优选地,在根据本发明实施例的设备中,所述处理器被配置为执行所述程序以进一步实现如下功能:在确定所述手势输入由特定用户做出之后,判断是否接收到用户的撤销操作;在确定所述手势输入由特定用户做出之后,判断识别出的手势输入是否与所述电子设备正在执行的当前任务完全不相关;判断所述第一图像中是否有其他人出现;在确定所述手势输入不是由特定用户做出之后,判断是否接收到所述手势的重复输入;以及基于以上判断结果中的至少一个,确定关于所述手势输入是否由特定用户做出的判断是否正确,以利用所述第一图像中的手臂特征更新对应的三维模型。Preferably, in the device according to an embodiment of the present invention, the processor is configured to execute the program to further implement a function of: determining whether the user's revocation is received after determining that the gesture input is made by a specific user After determining that the gesture input is made by a specific user, determining whether the recognized gesture input is completely unrelated to the current task being executed by the electronic device; determining whether another person appears in the first image; Determining whether the gesture input is not made by a specific user, determining whether a repeated input of the gesture is received; and determining, based on at least one of the above determination results, whether the determination as to whether the gesture input is made by a specific user is correct And updating the corresponding three-dimensional model with the arm feature in the first image.
在根据本发明的手势识别控制方法和电子设备中,通过对整个手臂而非单个手掌的图像处理来提高特定用户自身手势识别的准确率。并且,利用使用过程中判定为特定用户的手臂的数据来完成和完善建模过程,从而手臂的三维模型使用得越多,则其准确率就越高。另外,在根据本发明的手势识别控制方法和电子设备中,没有增加多余的设备或者增加用户的使用负担(例如,使用户主动选择该手势是否由自己做出),从而降低了成本并提高了用户体验。In the gesture recognition control method and electronic device according to the present invention, the accuracy of a specific user's own gesture recognition is improved by image processing of the entire arm instead of a single palm. Moreover, the modeling process is completed and perfected by using the data determined as the arm of the specific user during use, so that the more the three-dimensional model of the arm is used, the higher the accuracy. In addition, in the gesture recognition control method and the electronic device according to the present invention, no unnecessary device is added or the user's use burden is increased (for example, the user is actively selected whether the gesture is made by himself), thereby reducing the cost and improving the cost. user experience.
附图说明DRAWINGS
图1是图示根据本发明实施例的手势识别控制方法的过程的流程图;以及1 is a flowchart illustrating a procedure of a gesture recognition control method according to an embodiment of the present invention;
图2是图示根据本发明实施例的电子设备的配置的功能性框图。FIG. 2 is a functional block diagram illustrating a configuration of an electronic device according to an embodiment of the present invention.
具体实施方式Detailed ways
下面将参照附图对本发明的各个优选的实施方式进行描述。提供以下参照附图的描述,以帮助对由权利要求及其等价物所限定的本发明的示例实施方式的理解。其包括帮助理解的各种具体细节,但它们只能被看作是示例性的。因此,本领域技术人员将认识到,可对这里描述的实施方式进行各种改变和修改,而不脱离本发明的范围和精神。而且,为了使说明书更加清楚简洁,将省略对本领域熟知功能和构造的详细描述。Various preferred embodiments of the present invention will now be described with reference to the accompanying drawings. The following description with reference to the accompanying drawings will be understood to It includes various specific details to help understanding, but they can only be considered as exemplary. Accordingly, it will be appreciated by those skilled in the art that various modifications and changes may be made to the embodiments described herein without departing from the scope and spirit of the invention. Further, detailed descriptions of well-known functions and constructions in the art are omitted for clarity and conciseness.
在此使用的术语仅仅是为了描述具体实施例,而并非意在限制本公开。在此使用的术语“包括”、“包含”等表明了所述特征、步骤、操作和/或部件的存在,但是并不排除存在或添加一个或多个其他特征、步骤、操作或部件。The terminology used herein is for the purpose of describing the particular embodiments, The use of the terms "comprising", "comprising" or "an"
在此使用的所有术语(包括技术和科学术语)具有本领域技术人员通常所理解的含义, 除非另外定义。应注意,这里使用的术语应解释为具有与本说明书的上下文相一致的含义,而不应以理想化或过于刻板的方式来解释。All terms (including technical and scientific terms) used herein have the meaning commonly understood by one of ordinary skill in the art. Unless otherwise defined. It should be noted that the terms used herein are to be interpreted as having a meaning consistent with the context of the present specification and should not be interpreted in an ideal or too rigid manner.
在使用类似于“A、B和C等中至少一个”这样的表述的情况下,一般来说应该按照本领域技术人员通常理解该表述的含义来予以解释(例如,“具有A、B和C中至少一个的系统”应包括但不限于单独具有A、单独具有B、单独具有C、具有A和B、具有A和C、具有B和C、和/或具有A、B、C的系统等)。在使用类似于“A、B或C等中至少一个”这样的表述的情况下,一般来说应该按照本领域技术人员通常理解该表述的含义来予以解释(例如,“具有A、B或C中至少一个的系统”应包括但不限于单独具有A、单独具有B、单独具有C、具有A和B、具有A和C、具有B和C、和/或具有A、B、C的系统等)。本领域技术人员还应理解,实质上任意表示两个或更多可选项目的转折连词和/或短语,无论是在说明书、权利要求书还是附图中,都应被理解为给出了包括这些项目之一、这些项目任一方、或两个项目的可能性。例如,短语“A或B”应当被理解为包括“A”或“B”、或“A和B”的可能性。Where an expression similar to "at least one of A, B, and C, etc." is used, it should generally be interpreted in accordance with the meaning of the expression as commonly understood by those skilled in the art (for example, "having A, B, and C" "Systems of at least one of" shall include, but are not limited to, systems having A alone, B alone, C alone, A and B, A and C, B and C, and/or A, B, C, etc. ). Where an expression similar to "at least one of A, B or C, etc." is used, it should generally be interpreted according to the meaning of the expression as commonly understood by those skilled in the art (for example, "having A, B or C" "Systems of at least one of" shall include, but are not limited to, systems having A alone, B alone, C alone, A and B, A and C, B and C, and/or A, B, C, etc. ). Those skilled in the art will also appreciate that transitional conjunctions and/or phrases that are arbitrarily arbitrarily representing two or more optional items, whether in the specification, claims, or drawings, are to be construed as The possibility of one of the projects, either or both of these projects. For example, the phrase "A or B" should be understood to include the possibility of "A" or "B", or "A and B."
附图中示出了一些方框图和/或流程图。应理解,方框图和/或流程图中的一些方框或其组合可以由计算机程序指令来实现。这些计算机程序指令可以提供给通用计算机、专用计算机或其他可编程数据处理装置的处理器,从而这些指令在由该处理器执行时可以创建用于实现这些方框图和/或流程图中所说明的功能/操作的装置。Some block diagrams and/or flowcharts are shown in the drawings. It will be understood that some blocks or combinations of the block diagrams and/or flowcharts can be implemented by computer program instructions. These computer program instructions may be provided to a general purpose computer, a special purpose computer or a processor of other programmable data processing apparatus such that when executed by the processor, the instructions may be used to implement the functions illustrated in the block diagrams and/or flowcharts. / operating device.
因此,本公开的技术可以硬件和/或软件(包括固件、微代码等)的形式来实现。另外,本公开的技术可以采取存储有指令的计算机可读介质上的计算机程序产品的形式,该计算机程序产品可供指令执行系统使用或者结合指令执行系统使用。在本公开的上下文中,计算机可读介质可以是能够包含、存储、传送、传播或传输指令的任意介质。例如,计算机可读介质可以包括但不限于电、磁、光、电磁、红外或半导体系统、装置、器件或传播介质。计算机可读介质的具体示例包括:磁存储装置,如磁带或硬盘(HDD);光存储装置,如光盘(CD-ROM);存储器,如随机存取存储器(RAM)或闪存;和/或有线/无线通信链路。Thus, the techniques of this disclosure may be implemented in the form of hardware and/or software (including firmware, microcode, etc.). Additionally, the techniques of this disclosure may take the form of a computer program product on a computer readable medium storing instructions for use by or in connection with an instruction execution system. In the context of the present disclosure, a computer readable medium can be any medium that can contain, store, communicate, propagate or transport the instructions. For example, a computer readable medium can include, but is not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, device, or propagation medium. Specific examples of the computer readable medium include: a magnetic storage device such as a magnetic tape or a hard disk (HDD); an optical storage device such as a compact disk (CD-ROM); a memory such as a random access memory (RAM) or a flash memory; and/or a wired /Wireless communication link.
首先,将参照图1描述根据本发明的手势识别控制方法。所述手势识别控制方法应用于一电子设备。例如,所述电子设备可以是手机、平板电脑、智能电视或穿戴式设备。如图1所示,所述手势识别控制方法包括如下步骤。First, a gesture recognition control method according to the present invention will be described with reference to FIG. 1. The gesture recognition control method is applied to an electronic device. For example, the electronic device can be a cell phone, a tablet, a smart TV, or a wearable device. As shown in FIG. 1, the gesture recognition control method includes the following steps.
首先,在步骤S101,获取进行手势输入时的第一图像,所述第一图像包括手部和至少部分手臂的图像。First, in step S101, a first image when a gesture input is performed is acquired, the first image including an image of a hand and at least a part of an arm.
然后,在步骤S102,基于所述第一图像,判断所述手势输入是否由特定用户做出。Then, in step S102, based on the first image, it is determined whether the gesture input is made by a specific user.
如果在步骤S102判断结果为是,则处理进行到步骤S103。在步骤S103,对所述手势输 入执行识别。然后,基于识别出的手势,启动对应的功能。另一方面,如果在步骤S102判断结果为否,则处理进行到步骤S104。在步骤S104,忽略所述手势输入。也就是说,不对所述手势输入进行识别和响应。If the result of the determination in step S102 is YES, the process proceeds to step S103. In step S103, the gesture is lost Into the execution identification. Then, based on the recognized gesture, the corresponding function is activated. On the other hand, if the result of the determination in step S102 is NO, the process proceeds to step S104. At step S104, the gesture input is ignored. That is, the gesture input is not recognized and responded.
可见,在根据本发明的手势识别控制方法中,与现有技术相比,当用户进行手势输入时,不仅仅采集手部图像,而且还采集至少部分手臂图像。取决于用户的手臂姿态和/或手臂与图像采集模组(如,深度摄像头)的位置关系,可能能够获取全部手臂的图像,也可能仅能够获取包括部分手臂的图像。从而,根据本发明的手势识别控制方法能够基于手臂图像判断手势是否为特定用户做出,避免发生错误地识别本不应被识别的其他用户的手势。It can be seen that in the gesture recognition control method according to the present invention, when the user performs gesture input, not only the hand image but also at least part of the arm image is acquired, compared with the prior art. Depending on the user's arm posture and/or the positional relationship of the arm to the image acquisition module (eg, depth camera), it may be possible to capture images of all of the arms, or only images that include part of the arm. Thus, the gesture recognition control method according to the present invention can determine whether a gesture is made for a specific user based on the arm image, avoiding the erroneous recognition of gestures of other users that should not be recognized.
例如,在电子设备为手机、平板电脑、智能电视等的应用场景下,在根据现有技术的手势识别方法中,可以通过额外的面部识别来验证特定用户,并在验证通过后仅采集手部图像以进行手势识别。也就是说,在根据现有技术的手势识别方法中,可以通过面部识别来确认是否是特定用户,但不能确认当前的手势是否由识别到的特定用户做出。相比之下,通过根据本发明的手势识别控制方法,能够确认手势输入是否由特定用户做出,从而避免了当识别出特定用户的面部但手势却由其他人做出时引起的误判。For example, in an application scenario where the electronic device is a mobile phone, a tablet computer, a smart TV, or the like, in the gesture recognition method according to the prior art, a specific user can be verified by additional facial recognition, and only the hand is collected after the verification is passed. Image for gesture recognition. That is to say, in the gesture recognition method according to the related art, it is possible to confirm whether it is a specific user by face recognition, but it is not possible to confirm whether the current gesture is made by the identified specific user. In contrast, with the gesture recognition control method according to the present invention, it is possible to confirm whether the gesture input is made by a specific user, thereby avoiding a misjudgment caused when a specific user's face is recognized but the gesture is made by another person.
另外,随着虚拟现实和增强现实技术的逐渐成熟,头戴式显示设备将会是增强现实显示设备中最具有市场前景的交互界面呈现设备。在电子设备为这样的头戴式显示设备的应用场景下,在使用过程中,除了能够获取用于手势识别的佩戴者的手部图像之外,还可能获取佩戴者以外的其他用户的手部图像。从而,会出现头戴式显示设备将他人的手势当作佩戴者的手势的情况,进而导致错误的反应。在将来多人同时佩戴头戴式显示设备进行合作的场景中,这种错误识别的频率和严重程度会更高。通过根据本发明的手势识别控制方法,能够确认手势输入是否由佩戴者做出,有效避免了将其他头戴式显示设备的佩戴者的手势误判为当前头戴式显示设备的佩戴者的手势的情况。In addition, with the gradual maturity of virtual reality and augmented reality technology, head-mounted display devices will be the most promising interactive interface rendering devices in augmented reality display devices. In an application scenario where the electronic device is such a head mounted display device, in addition to being able to acquire a hand image of the wearer for gesture recognition, it is also possible to acquire the hand of another user other than the wearer. image. Therefore, there is a case where the head-mounted display device regards the gesture of another person as the gesture of the wearer, thereby causing an erroneous reaction. In the scene where many people wear head-mounted display devices at the same time in the future, the frequency and severity of such misrecognition will be higher. With the gesture recognition control method according to the present invention, it can be confirmed whether the gesture input is made by the wearer, and the gesture of mistaking the gesture of the wearer of the other head mounted display device as the wearer of the current head mounted display device is effectively avoided. Case.
基于所述第一图像,判断所述手势输入是否由特定用户做出的方式可以包括但不限于以下方式:(1)基于手臂在第一图像中的位置判断是否满足特定用户做出该手势输入的合理性;(2)基于手臂的三维模型判断是否满足与特定用户做出该手势输入的相似性。Based on the first image, determining whether the gesture input is made by a specific user may include, but is not limited to, the following manners: (1) determining whether the gesture is satisfied by a specific user based on the position of the arm in the first image. Rationality; (2) Based on the three-dimensional model of the arm to determine whether the similarity with the gesture input by the specific user is satisfied.
接下来,将分别对这两种方式进行说明。Next, the two methods will be explained separately.
(1)基于手臂在第一图像中的位置判断是否满足特定用户做出该手势输入的合理性(1) judging whether the satisfaction of the specific user making the gesture input is based on the position of the arm in the first image
例如,在电子设备为手机、平板电脑、智能电视等的情况下,图1中的步骤S102可以进一步包括:基于第一图像中手臂和经验证的面部之间的相对关系来确认是否是特定用户做出的手势。如果手臂与面部之间的距离超过预定阈值,或者如果手臂与面部之间的角度超过预定角度范围,那么说明手臂与面部之间的相对关系不合理,进而可以确定所述手势输入并非 由特定用户做出。另一方面,如果没有判断出手臂与面部之间的相对关系不合理,那么暂且认为所述手势输入由特定用户做出。For example, in the case where the electronic device is a mobile phone, a tablet computer, a smart TV, etc., step S102 in FIG. 1 may further include: confirming whether it is a specific user based on a relative relationship between the arm and the verified face in the first image. The gesture made. If the distance between the arm and the face exceeds a predetermined threshold, or if the angle between the arm and the face exceeds a predetermined angular range, then the relative relationship between the arm and the face is unreasonable, and it can be determined that the gesture input is not Made by a specific user. On the other hand, if it is not judged that the relative relationship between the arm and the face is unreasonable, then the gesture input is temporarily considered to be made by a specific user.
例如,在电子设备为头戴式显示设备的情况下,基于第一图像中手臂的位置来确认是否是特定用户(在这种情况下,为佩戴者)做出的手势。具体来讲,由于通过佩戴在头部的头戴式显示设备上的摄像头来拍摄佩戴者的手臂图像,因此在所获得的第一图像中,左臂图像不可能位于右侧区域,且右臂图像不可能位于左侧区域。因此,图1中的步骤S102可以进一步包括:首先确定第一图像中进行手势输入的手臂为左臂还是右臂。如果判断为左臂,那么继续判断左臂是否出现在右侧区域,如果是,则说明手臂的位置不合理,进而可以确定所述手势输入并非由特定用户做出。另一方面,如果没有判断出手臂的位置关系不合理,则暂且认为所述手势输入由特定用户做出。同样地,如果判断为右臂,那么继续判断右臂是否出现在左侧区域,如果是,则说明手臂的位置不合理,进而可以确定所述手势输入并非由特定用户做出。另一方面,如果没有判断出手臂的位置关系不合理,则暂且认为所述手势输入由特定用户做出。For example, in the case where the electronic device is a head mounted display device, it is confirmed based on the position of the arm in the first image whether it is a gesture made by a specific user (in this case, a wearer). Specifically, since the image of the wearer's arm is photographed by the camera worn on the head mounted display device of the head, in the obtained first image, the left arm image may not be located in the right side region, and the right arm The image cannot be in the left area. Therefore, step S102 in FIG. 1 may further include first determining whether the arm in which the gesture input is performed in the first image is the left arm or the right arm. If it is determined to be the left arm, it is determined whether the left arm appears in the right area, and if so, the position of the arm is unreasonable, and it can be determined that the gesture input is not made by a specific user. On the other hand, if it is not judged that the positional relationship of the arm is unreasonable, it is temporarily considered that the gesture input is made by a specific user. Similarly, if it is determined to be the right arm, it is determined whether the right arm is present in the left area, and if so, the position of the arm is unreasonable, and it can be determined that the gesture input is not made by a specific user. On the other hand, if it is not judged that the positional relationship of the arm is unreasonable, it is temporarily considered that the gesture input is made by a specific user.
(2)基于手臂的三维模型判断是否满足与特定用户做出该手势输入的相似性(2) Based on the three-dimensional model of the arm to determine whether the similarity with the gesture input by the specific user is satisfied
通过这种方式来判断是否满足与特定用户做出该手势输入的相似性,需要预先存储所述特定用户的用于进行手势操作的手臂的三维模型。这里,可以预先存储所述特定用户的双臂的两个三维模型,从而用户的任意手臂的手势都能够被识别。或者,也可以仅存储所述特定用户的一个手臂的三维模型,从而仅能够识别用户的特定手臂的手势。In this way, it is judged whether the similarity with the gesture input by the specific user is satisfied, and it is necessary to store in advance the three-dimensional model of the arm of the specific user for performing the gesture operation. Here, two three-dimensional models of the arms of the specific user may be stored in advance so that the gesture of any arm of the user can be recognized. Alternatively, it is also possible to store only the three-dimensional model of one arm of the specific user, so that only the gesture of the specific arm of the user can be recognized.
其中,所述预先存储针对特定用户的手臂的三维模型的步骤进一步包括:获取所述特定用户在多次手势操作中的多个手臂图像;针对所述多个手臂图像中的每一个,分解手臂的不同部位,并提取不同部位的特征;以及针对同一手臂融合相同部位的多个特征,并获得针对该手臂的三维模型。Wherein the step of pre-storing the three-dimensional model for the arm of the specific user further comprises: acquiring a plurality of arm images of the specific user in the multiple gesture operation; and decomposing the arm for each of the plurality of arm images Different parts, and extracting features of different parts; and merging multiple features of the same part for the same arm, and obtaining a three-dimensional model for the arm.
在根据本发明的手势识别控制方法中,利用深度摄像头大概120度的识别范围和机器学习算法,记录特定用户的手臂在各个角度出现时的特征,对手臂进行整体的学习和建模,而不仅仅是对手指动作的识别。这里,需要指出的是,在利用深度摄像头获得手势操作的过程中,由于手臂放置的位置和方向不同,因此能够获得的手臂的完整度是不同的。但随着手势使用次数的累积,会逐渐获得全部手臂的三维数据。经过一定量的数据学习,即在获取了特定用户在多次手势操作中的足够的手臂图像之后,即可建立针对特定用户的手臂的模型。手臂的不同部位包括:手、前臂、大臂。手臂的三维模型包括:前臂和大臂的尺寸信息、前臂与大臂的角度范围、前臂和大臂上的特征图像(如,个性化的关节、凸起等)等。当出现一个新的手臂位置时,可以利用模型来判断是不是特定用户自身的手势。 In the gesture recognition control method according to the present invention, the recognition range of the depth camera and the machine learning algorithm are used to record the characteristics of the specific user's arm at various angles, and the whole body is learned and modeled, not only It is only the recognition of finger movements. Here, it should be noted that in the process of obtaining a gesture operation using the depth camera, since the position and direction in which the arm is placed are different, the degree of integrity of the arm that can be obtained is different. However, as the number of gestures is accumulated, the three-dimensional data of all the arms is gradually obtained. After a certain amount of data learning, that is, after acquiring a sufficient arm image of a specific user in multiple gesture operations, a model for a specific user's arm can be established. Different parts of the arm include: hand, forearm, and big arm. The three-dimensional model of the arm includes: size information of the forearm and the big arm, angular range of the forearm and the big arm, and characteristic images on the forearm and the big arm (eg, personalized joints, bumps, etc.). When a new arm position appears, the model can be used to determine if it is a particular user's own gesture.
图1中的步骤S102可以进一步包括:确定所述第一图像中包括的手臂为左臂还是右臂,并且在所述第一图像中分解手臂的不同部位,提取不同部位的特征;以及将获取的手臂不同部位的特征与对应的三维模型中的特征进行比较,并判断所述手势输入是否由所述特定用户做出。Step S102 in FIG. 1 may further include: determining whether the arm included in the first image is a left arm or a right arm, and decomposing different parts of the arm in the first image, extracting features of different parts; and acquiring The features of the different parts of the arm are compared to the features in the corresponding three-dimensional model and it is determined whether the gesture input is made by the particular user.
其中,将获取的手臂不同部位的特征与对应的三维模型中的特征进行比较的步骤进一步包括:The step of comparing the acquired features of different parts of the arm with the features in the corresponding three-dimensional model further includes:
判断所述第一图像中手臂不同部位的尺寸与所述三维模型中的对应尺寸之间的差值是否小于预定阈值,获得第一判断结果;和/或Determining whether a difference between a size of a different part of the arm in the first image and a corresponding size in the three-dimensional model is less than a predetermined threshold, obtaining a first determination result; and/or
判断所述第一图像中前臂与大臂之间的角度是否处于基于所述三维模型确定的预定范围之内,获得第二判断结果。Determining whether the angle between the forearm and the boom in the first image is within a predetermined range determined based on the three-dimensional model, obtaining a second determination result.
其中,判断所述手势输入是否由所述特定用户做出的步骤包括:基于所述第一判断结果和/或第二判断结果,确定所述手势输入是否由所述特定用户做出。具体来讲,如果第一判断结果和/或第二判断结果中存在否的判断结果,那么确定所述手势输入并非由所述特定用户做出,否则确定所述手势输入由所述特定用户做出。Wherein, determining whether the gesture input is made by the specific user comprises determining whether the gesture input is made by the specific user based on the first determination result and/or the second determination result. Specifically, if there is a judgment result of the first judgment result and/or the second judgment result, it is determined that the gesture input is not made by the specific user, otherwise it is determined that the gesture input is performed by the specific user. Out.
在上文中描述的用以判断所述手势输入是否由特定用户做出的两种方式可以分别单独使用。或者,这两种方式也可以串行地使用。例如,首先在合理性方面判断然后在相似性方面判断。由于合理性判断的运算量较小,因此以合理性作为筛选条件,可以有效降低基于手臂三维模型的判断的运算量。The two methods described above for determining whether the gesture input is made by a particular user may be used separately. Alternatively, these two methods can also be used serially. For example, first judge in terms of rationality and then judge in terms of similarity. Since the calculation amount of the rationality judgment is small, the rationality is used as the screening condition, and the calculation amount based on the judgment of the three-dimensional model of the arm can be effectively reduced.
从以上的描述可以看出,在根据本发明的手势识别控制方法中,采取“疑判从无”的原则。也就是说,只要不是确切地肯定手势输入不是特定用户做出,那么就暂且认为手势输入是由特定用户做出的。例如,当基于第一图像没有判断出明显的不合理时,或者当基于目前的手臂三维模型不能确定手势输入是否由特定用户做出时,那么就暂且认为手势输入是由特定用户做出的。As can be seen from the above description, in the gesture recognition control method according to the present invention, the principle of "doubt is never" is adopted. That is, as long as it is not certain that the gesture input is not made by a particular user, then the gesture input is temporarily considered to be made by a particular user. For example, when it is not judged to be unreasonable based on the first image, or when it is not possible to determine whether the gesture input is made by a specific user based on the current arm three-dimensional model, then the gesture input is temporarily considered to be made by a specific user.
因此,例如,当手臂尺寸与特定用户的手臂尺寸相仿的其他用户进行手势输入时,可能会发生误判的情况。为了避免这种情况的发生和进一步提高判定的准确性,在根据本发明的手势识别控制方法中,在基于所述第一图像,判断出所述手势输入是否由特定用户做出之后,还可以进一步包括验证步骤。Therefore, for example, when other users who have an arm size similar to the arm size of a specific user perform gesture input, a misjudgment may occur. In order to avoid the occurrence of such a situation and further improve the accuracy of the determination, in the gesture recognition control method according to the present invention, after determining whether the gesture input is made by a specific user based on the first image, Further includes a verification step.
具体来讲,所述验证步骤可以包括如下处理:Specifically, the verifying step may include the following processing:
(1)在确定所述手势输入由特定用户做出之后,判断是否接收到用户的撤销操作。一般而言,如果在确定所述手势输入由特定用户做出之后接收到用户的撤销操作,那么所述手势输入很有可能并非由特定用户做出,即:误判的概率较大。 (1) After determining that the gesture input is made by a specific user, it is determined whether the user's undo operation is received. In general, if the user's undo operation is received after determining that the gesture input is made by a particular user, then the gesture input is most likely not made by a particular user, ie, the probability of a false positive is greater.
(2)在确定所述手势输入由特定用户做出之后,判断识别出的手势输入是否与所述电子设备正在执行的当前任务完全不相关。一般而言,如果确定由特定用户做出的手势输入与所述电子设备正在执行的当前任务完全不相关,那么所述手势输入很有可能并非由特定用户做出,即:误判的概率较大。(2) After determining that the gesture input is made by a specific user, it is determined whether the recognized gesture input is completely unrelated to the current task being executed by the electronic device. In general, if it is determined that the gesture input made by a particular user is completely unrelated to the current task being performed by the electronic device, then the gesture input is likely not to be made by a particular user, ie, the probability of a false positive is greater than Big.
(3)判断所述第一图像中是否有其他人出现。一般而言,如果第一图像中有其他人出现,那么所述手势输入并非由特定用户做出的概率增加,即:误判的概率增大。但是,由于第一图像中有其他人出现并不一定会发生误判,因此该判断仅能作为其他条件的辅助参考,不能单独作为是否发生误判的依据。(3) determining whether another person appears in the first image. In general, if there are other people in the first image, the probability that the gesture input is not made by a particular user increases, that is, the probability of false positives increases. However, since other people in the first image do not necessarily have a misjudgment, the judgment can only serve as a supplementary reference for other conditions, and cannot be used alone as a basis for misjudgment.
(4)在确定所述手势输入不是由特定用户做出之后,判断是否接收到所述手势的重复输入。一般而言,如果在确定所述手势输入不是由特定用户做出之后接收到所述手势的重复输入,那么所述手势输入很有可能由特定用户做出,即:误判的概率较大。(4) After determining that the gesture input is not made by a specific user, it is determined whether a repeated input of the gesture is received. In general, if a repeated input of the gesture is received after determining that the gesture input is not made by a particular user, then the gesture input is likely to be made by a particular user, ie, the probability of a false positive is greater.
基于以上判断结果中的至少一个,确定关于所述手势输入是否由特定用户做出的判断是否正确,以利用所述第一图像中的手臂特征更新对应的三维模型。Based on at least one of the above determination results, it is determined whether the determination as to whether the gesture input is made by a specific user is correct to update the corresponding three-dimensional model with the arm feature in the first image.
具体来讲,如果确定关于所述手势输入是否由特定用户做出的判断正确,则将所述第一图像中的手臂特征添加到对应的三维模型中。In particular, if it is determined whether the determination as to whether the gesture input is made by a particular user is correct, the arm features in the first image are added to the corresponding three-dimensional model.
另一方面,如果确定关于所述手势输入是否由特定用户做出的判断错误,则基于所述第一图像中的手臂特征,修改对应的三维模型。修改对应的三维模型为在该三维模型中将该手臂特征作为排除条件,即:如果再次检测到该手臂特征,则确定不是由特定用户做出的手势。On the other hand, if a determination is made as to whether the gesture input is made by a particular user, the corresponding three-dimensional model is modified based on the arm features in the first image. The corresponding three-dimensional model is modified to exclude the arm feature in the three-dimensional model, that is, if the arm feature is detected again, a gesture that is not made by a specific user is determined.
通过利用使用过程中判定为特定用户的手臂的数据来完善建模的过程,从而所述三维模型使用得越多,其准确率就越高。By using the data determined to be the arm of a particular user during use to complete the process of modeling, the more the three-dimensional model is used, the higher its accuracy.
在上文中,已经参照图1详细描述了根据本发明的手势识别控制方法的具体过程。接下来,将参照图2描述应用了根据本发明的手势识别控制方法的电子设备。In the above, the specific process of the gesture recognition control method according to the present invention has been described in detail with reference to FIG. Next, an electronic device to which the gesture recognition control method according to the present invention is applied will be described with reference to FIG.
如图2所示,所述电子设备200包括:图像采集模组201、存储器202和处理器203。As shown in FIG. 2, the electronic device 200 includes an image acquisition module 201, a memory 202, and a processor 203.
图像采集模组201用于获取进行手势输入时的第一图像,其中,第一图像包括手部和至少部分手臂的图像。通常地,可以通过深度摄像头来实现图像采集模组201。The image acquisition module 201 is configured to acquire a first image when the gesture input is performed, wherein the first image includes an image of the hand and at least a part of the arm. Generally, the image acquisition module 201 can be implemented by a depth camera.
存储器202用于存储程序。The memory 202 is used to store programs.
处理器203用于当执行所述存储器202上存储的程序时实现如下功能:The processor 203 is configured to implement the following functions when executing the program stored on the memory 202:
基于所述第一图像,判断所述手势输入是否由特定用户做出;以及Determining whether the gesture input is made by a specific user based on the first image;
如果判断结果为是,则对所述手势输入执行识别,否则忽略所述手势输入。If the result of the determination is yes, recognition is performed on the gesture input, otherwise the gesture input is ignored.
基于所述第一图像,判断所述手势输入是否由特定用户做出的方式可以包括但不限于以下方式:(1)基于手臂在第一图像中的位置判断是否满足特定用户做出该手势输入的合理性; (2)基于手臂的三维模型判断是否满足与特定用户做出该手势输入的相似性。Based on the first image, determining whether the gesture input is made by a specific user may include, but is not limited to, the following manners: (1) determining whether the gesture is satisfied by a specific user based on the position of the arm in the first image. Rationality (2) Based on the three-dimensional model of the arm, it is judged whether the similarity with the gesture input by the specific user is satisfied.
接下来,将分别对这两种方式进行说明。Next, the two methods will be explained separately.
(1)基于手臂在第一图像中的位置判断是否满足特定用户做出该手势输入的合理性(1) judging whether the satisfaction of the specific user making the gesture input is based on the position of the arm in the first image
例如,在电子设备为手机、平板电脑、智能电视等的情况下,处理器203可以执行所述程序以进一步实现如下功能:基于第一图像中手臂和经验证的面部之间的相对关系来确认是否是特定用户做出的手势。如果手臂与面部之间的距离超过预定阈值,或者如果手臂与面部之间的角度超过预定角度范围,那么说明手臂与面部之间的相对关系不合理,进而可以确定所述手势输入并非由特定用户做出。另一方面,如果没有判断出手臂与面部之间的相对关系不合理,那么暂且认为所述手势输入由特定用户做出。For example, in the case where the electronic device is a mobile phone, a tablet, a smart TV, etc., the processor 203 can execute the program to further implement a function of confirming based on a relative relationship between the arm and the verified face in the first image. Whether it is a gesture made by a specific user. If the distance between the arm and the face exceeds a predetermined threshold, or if the angle between the arm and the face exceeds a predetermined angular range, then the relative relationship between the arm and the face is unreasonable, and it can be determined that the gesture input is not by a specific user. Made. On the other hand, if it is not judged that the relative relationship between the arm and the face is unreasonable, then the gesture input is temporarily considered to be made by a specific user.
例如,在电子设备为头戴式显示设备的情况下,基于第一图像中手臂的位置来确认是否是特定用户(在这种情况下,为佩戴者)做出的手势。具体来讲,由于通过佩戴在头部的头戴式显示设备上的摄像头来拍摄佩戴者的手臂图像,因此在所获得的第一图像中,左臂图像不可能位于右侧区域,且右臂图像不可能位于左侧区域。因此,处理器203可以执行所述程序以进一步实现如下功能:首先确定第一图像中进行手势输入的手臂为左臂还是右臂。如果判断为左臂,那么继续判断左臂是否出现在右侧区域,如果是,则说明手臂的位置不合理,进而可以确定所述手势输入并非由特定用户做出。另一方面,如果没有判断出手臂的位置关系不合理,则暂且认为所述手势输入由特定用户做出。同样地,如果判断为右臂,那么继续判断右臂是否出现在左侧区域,如果是,则说明手臂的位置不合理,进而可以确定所述手势输入并非由特定用户做出。另一方面,如果没有判断出手臂的位置关系不合理,则暂且认为所述手势输入由特定用户做出。For example, in the case where the electronic device is a head mounted display device, it is confirmed based on the position of the arm in the first image whether it is a gesture made by a specific user (in this case, a wearer). Specifically, since the image of the wearer's arm is photographed by the camera worn on the head mounted display device of the head, in the obtained first image, the left arm image may not be located in the right side region, and the right arm The image cannot be in the left area. Accordingly, the processor 203 can execute the program to further implement a function of first determining whether the arm in which the gesture input is made in the first image is the left arm or the right arm. If it is determined to be the left arm, it is determined whether the left arm appears in the right area, and if so, the position of the arm is unreasonable, and it can be determined that the gesture input is not made by a specific user. On the other hand, if it is not judged that the positional relationship of the arm is unreasonable, it is temporarily considered that the gesture input is made by a specific user. Similarly, if it is determined to be the right arm, it is determined whether the right arm is present in the left area, and if so, the position of the arm is unreasonable, and it can be determined that the gesture input is not made by a specific user. On the other hand, if it is not judged that the positional relationship of the arm is unreasonable, it is temporarily considered that the gesture input is made by a specific user.
(2)基于手臂的三维模型判断是否满足与特定用户做出该手势输入的相似性(2) Based on the three-dimensional model of the arm to determine whether the similarity with the gesture input by the specific user is satisfied
通过这种方式来判断是否满足与特定用户做出该手势输入的相似性,所述存储器进一步被配置为预先存储特定用户的用于进行手势操作的手臂的三维模型。这里,可以预先存储所述特定用户的双臂的两个三维模型,从而用户的任意手臂的手势都能够被识别。或者,也可以仅存储所述特定用户的一个手臂的三维模型,从而仅能够识别用户的特定手臂的手势。In this manner, it is judged whether the similarity with the gesture input by the specific user is satisfied, and the memory is further configured to pre-store a three-dimensional model of the arm of the specific user for performing the gesture operation. Here, two three-dimensional models of the arms of the specific user may be stored in advance so that the gesture of any arm of the user can be recognized. Alternatively, it is also possible to store only the three-dimensional model of one arm of the specific user, so that only the gesture of the specific arm of the user can be recognized.
其中,所述处理器被配置为执行所述程序以进一步实现如下功能:针对所述图像采集单元获取的所述特定用户在多次手势操作中的多个手臂图像中的每一个,分解手臂的不同部位,并提取不同部位的特征;以及针对同一手臂融合相同部位的多个特征,并获得针对该手臂的三维模型。Wherein the processor is configured to execute the program to further implement a function of: decomposing an arm of each of the plurality of arm images of the specific user acquired in the plurality of gesture operations by the image acquisition unit Different parts, and extracting features of different parts; and merging multiple features of the same part for the same arm, and obtaining a three-dimensional model for the arm.
当出现一个新的手臂位置时,可以利用模型来判断是不是特定用户自身的手势。具体地,所述处理器被配置为执行所述程序以进一步实现如下功能: When a new arm position appears, the model can be used to determine if it is a particular user's own gesture. Specifically, the processor is configured to execute the program to further implement the following functions:
确定所述第一图像中包括的手臂为左臂还是右臂,并且在所述第一图像中分解手臂的不同部位,提取不同部位的特征;以及Determining whether the arm included in the first image is a left arm or a right arm, and decomposing different parts of the arm in the first image to extract features of different parts;
将获取的手臂不同部位的特征与对应的三维模型中的特征进行比较,并判断所述手势输入是否由所述特定用户做出。The acquired features of different parts of the arm are compared with features in the corresponding three-dimensional model, and it is determined whether the gesture input is made by the specific user.
其中,将获取的手臂不同部位的特征与对应的三维模型中的特征进行比较,并判断所述手势输入是否由所述特定用户做出的处理进一步包括:Wherein, comparing the acquired features of the different parts of the arm with the features in the corresponding three-dimensional model, and determining whether the gesture input is processed by the specific user further comprises:
判断所述第一图像中手臂不同部位的尺寸与所述三维模型中的对应尺寸之间的差值是否小于预定阈值,获得第一判断结果;和/或Determining whether a difference between a size of a different part of the arm in the first image and a corresponding size in the three-dimensional model is less than a predetermined threshold, obtaining a first determination result; and/or
判断所述第一图像中前臂与大臂之间的角度是否处于基于所述三维模型确定的预定范围之内,获得第二判断结果;以及Determining whether an angle between the forearm and the boom in the first image is within a predetermined range determined based on the three-dimensional model, obtaining a second determination result;
基于所述第一判断结果和/或第二判断结果,确定所述手势输入是否由所述特定用户做出。具体来讲,如果第一判断结果和/或第二判断结果中存在否的判断结果,那么确定所述手势输入并非由所述特定用户做出,否则确定所述手势输入由所述特定用户做出。Based on the first determination result and/or the second determination result, it is determined whether the gesture input is made by the specific user. Specifically, if there is a judgment result of the first judgment result and/or the second judgment result, it is determined that the gesture input is not made by the specific user, otherwise it is determined that the gesture input is performed by the specific user. Out.
从以上的描述可以看出,在根据本发明的电子设备中,采取“疑判从无”的原则。例如,当手臂尺寸与特定用户的手臂尺寸相仿的其他用户进行手势输入时,可能会发生误判的情况。为了避免这种情况的发生和进一步提高判定的准确性,在根据本发明的电子设备中,所述处理器被配置为执行所述程序以进一步实现如下功能:As can be seen from the above description, in the electronic device according to the present invention, the principle of "doubt is never" is taken. For example, when other users who have an arm size similar to the arm size of a particular user make a gesture input, a misjudgment may occur. In order to avoid this occurrence and further improve the accuracy of the determination, in the electronic device according to the present invention, the processor is configured to execute the program to further implement the following functions:
(1)在确定所述手势输入由特定用户做出之后,判断是否接收到用户的撤销操作;(1) determining whether the user's revocation operation is received after determining that the gesture input is made by a specific user;
(2)在确定所述手势输入由特定用户做出之后,判断识别出的手势输入是否与所述电子设备正在执行的当前任务完全不相关;(2) after determining that the gesture input is made by a specific user, determining whether the recognized gesture input is completely unrelated to the current task being executed by the electronic device;
(3)判断所述第一图像中是否有其他人出现;(3) determining whether another person appears in the first image;
(4)在确定所述手势输入不是由特定用户做出之后,判断是否接收到所述手势的重复输入;以及(4) determining whether a repeated input of the gesture is received after determining that the gesture input is not made by a specific user;
基于以上判断结果中的至少一个,确定关于所述手势输入是否由特定用户做出的判断是否正确。Based on at least one of the above determination results, it is determined whether the determination as to whether the gesture input is made by a specific user is correct.
具体来讲,如果确定关于所述手势输入是否由特定用户做出的判断正确,则将所述第一图像中的手臂特征添加到对应的三维模型中。In particular, if it is determined whether the determination as to whether the gesture input is made by a particular user is correct, the arm features in the first image are added to the corresponding three-dimensional model.
另一方面,如果确定关于所述手势输入是否由特定用户做出的判断错误,则基于所述第一图像中的手臂特征,修改对应的三维模型。修改对应的三维模型为在该三维模型中将该手臂特征作为排除条件,即:如果再次检测到该手臂特征,则确定不是由特定用户做出的手势。On the other hand, if a determination is made as to whether the gesture input is made by a particular user, the corresponding three-dimensional model is modified based on the arm features in the first image. The corresponding three-dimensional model is modified to exclude the arm feature in the three-dimensional model, that is, if the arm feature is detected again, a gesture that is not made by a specific user is determined.
根据本公开实施例,处理器203可以包括通用微处理器、指令集处理器和/或相关芯片组 和/或专用微处理器(例如,专用集成电路(ASIC)),等等。处理器203还可以包括用于缓存用途的板载存储器。处理器203可以是用于执行参考图1描述的根据本公开实施例的方法流程的不同动作的单一处理单元或者是多个处理单元。 Processor 203 may include a general purpose microprocessor, an instruction set processor, and/or a related chipset, in accordance with an embodiment of the present disclosure. And/or a dedicated microprocessor (eg, an application specific integrated circuit (ASIC)), and the like. Processor 203 may also include an onboard memory for caching purposes. The processor 203 may be a single processing unit or a plurality of processing units for performing different actions of the method flow according to the embodiment of the present disclosure described with reference to FIG.
存储器202,例如可以是能够包含、存储、传送、传播或传输指令的任意介质。例如,存储器202可以包括但不限于电、磁、光、电磁、红外或半导体系统、装置、器件或传播介质。存储器202的具体示例包括:磁存储装置,如磁带或硬盘(HDD);光存储装置,如光盘(CD-ROM);存储器,如随机存取存储器(RAM)或闪存;和/或有线/无线通信链路。由于根据本发明的电子设备的配置完全与之前描述的手势识别控制方法中的各步骤对应,因此为了避免冗余起见,在针对电子设备的描述中,并未展开很多细节。但是,本领域的技术人员应该理解,针对手势识别控制方法描述的内容可以完全类似地应用于电子设备。 Memory 202, for example, can be any medium that can contain, store, communicate, propagate, or transport the instructions. For example, memory 202 can include, but is not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, device, or propagation medium. Specific examples of the memory 202 include: a magnetic storage device such as a magnetic tape or a hard disk (HDD); an optical storage device such as a compact disk (CD-ROM); a memory such as a random access memory (RAM) or a flash memory; and/or a wired/wireless Communication link. Since the configuration of the electronic device according to the present invention completely corresponds to the steps in the previously described gesture recognition control method, in order to avoid redundancy, in the description for the electronic device, many details are not developed. However, those skilled in the art will appreciate that the content described for the gesture recognition control method can be applied analogously to electronic devices.
迄今为止,已经参照图1和图2详细描述了根据本发明的手势识别控制方法和电子设备。在根据本发明的手势识别控制方法和电子设备中,通过对整个手臂而非单个手掌的图像处理来提高特定用户自身手势识别的准确率。并且,利用使用过程中判定为特定用户的手臂的数据来完成和完善建模过程,从而手臂的三维模型使用得越多,则其准确率就越高。另外,在根据本发明的手势识别控制方法和电子设备中,没有增加多余的没备或者增加用户的使用负担(例如,使用户主动选择该手势是否由自己做出),从而降低了成本并提高了用户体验。Heretofore, the gesture recognition control method and electronic apparatus according to the present invention have been described in detail with reference to FIGS. 1 and 2. In the gesture recognition control method and electronic device according to the present invention, the accuracy of a specific user's own gesture recognition is improved by image processing of the entire arm instead of a single palm. Moreover, the modeling process is completed and perfected by using the data determined as the arm of the specific user during use, so that the more the three-dimensional model of the arm is used, the higher the accuracy. In addition, in the gesture recognition control method and the electronic device according to the present invention, no unnecessary redundancy is added or the user's use burden is increased (for example, the user is actively selected whether the gesture is made by himself), thereby reducing costs and improving The user experience.
需要说明的是,在本说明书中,术语“包括”、“包含”或者其任何其他变体意在涵盖非排他性的包含,从而使得包括一系列要素的过程、方法、物品或者设备不仅包括那些要素,而且还包括没有明确列出的其他要素,或者是还包括为这种过程、方法、物品或者设备所固有的要素。在没有更多限制的情况下,由语句“包括……”限定的要素,并不排除在包括所述要素的过程、方法、物品或者设备中还存在另外的相同要素。It should be noted that, in this specification, the terms "including", "comprising", or any other variations thereof are intended to encompass a non-exclusive inclusion, such that a process, method, article, or device that comprises a And also includes other elements not explicitly listed, or elements that are inherent to such a process, method, item, or device. An element that is defined by the phrase "comprising", without limiting the invention, does not exclude the presence of additional elements in the process, method, article, or device.
最后,还需要说明的是,上述一系列处理不仅包括以这里所述的顺序按时间序列执行的处理,而且包括并行或分别地、而不是按时间顺序执行的处理。Finally, it should also be noted that the series of processes described above include not only processes that are performed in time series in the order described herein, but also processes that are performed in parallel or separately, rather than in chronological order.
通过以上的实施方式的描述,本领域的技术人员可以清楚地了解到本发明可借助软件加必需的硬件平台的方式来实现,当然也可以全部通过软件来实施。基于这样的理解,本发明的技术方案对背景技术做出贡献的全部或者部分可以以软件产品的形式体现出来,该计算机软件产品可以存储在存储介质中,如ROM/RAM、磁碟、光盘等,包括若干指令用以使得一台计算机设备(可以是个人计算机,服务器,或者网络设备等)执行本发明各个实施例或者实施例的某些部分所述的方法。Through the description of the above embodiments, those skilled in the art can clearly understand that the present invention can be implemented by means of software plus a necessary hardware platform, and of course, all can be implemented by software. Based on such understanding, all or part of the technical solution of the present invention contributing to the background art may be embodied in the form of a software product, which may be stored in a storage medium such as a ROM/RAM, a magnetic disk, an optical disk, or the like. A number of instructions are included to cause a computer device (which may be a personal computer, server, or network device, etc.) to perform the methods described in various embodiments of the present invention or in some portions of the embodiments.
以上对本发明进行了详细介绍,本文中应用了具体个例对本发明的原理及实施方式进行了阐述,以上实施例的说明只是用于帮助理解本发明的方法及其核心思想;同时,对于本领 域的一般技术人员,依据本发明的思想,在具体实施方式及应用范围上均会有改变之处,综上所述,本说明书内容不应理解为对本发明的限制。 The present invention has been described in detail above, and the principles and embodiments of the present invention have been described with reference to specific examples. The description of the above embodiments is only for helping to understand the method of the present invention and its core ideas; The description of the present invention is not intended to limit the scope of the present invention.

Claims (10)

  1. 一种手势识别控制方法,应用于一电子设备,所述方法包括:A gesture recognition control method is applied to an electronic device, and the method includes:
    获取进行手势输入时的第一图像,所述第一图像包括手部和至少部分手臂的图像;Obtaining a first image when the gesture input is performed, the first image including an image of the hand and at least a portion of the arm;
    基于所述第一图像,判断所述手势输入是否由特定用户做出;以及Determining whether the gesture input is made by a specific user based on the first image;
    如果判断结果为是,则对所述手势输入执行识别,否则忽略所述手势输入。If the result of the determination is yes, recognition is performed on the gesture input, otherwise the gesture input is ignored.
  2. 根据权利要求1所述的方法,其中:The method of claim 1 wherein:
    所述方法进一步包括:预先存储特定用户的用于进行手势操作的手臂的三维模型;The method further includes pre-storing a three-dimensional model of a specific user's arm for performing a gesture operation;
    所述基于所述第一图像,判断所述手势输入是否由特定用户做出包括:The determining, based on the first image, whether the gesture input is made by a specific user comprises:
    确定所述第一图像中包括的手臂为左臂还是右臂,并且在所述第一图像中分解手臂的不同部位,提取不同部位的特征;以及Determining whether the arm included in the first image is a left arm or a right arm, and decomposing different parts of the arm in the first image to extract features of different parts;
    将获取的手臂不同部位的特征与对应的三维模型中的特征进行比较,并判断所述手势输入是否由所述特定用户做出。The acquired features of different parts of the arm are compared with features in the corresponding three-dimensional model, and it is determined whether the gesture input is made by the specific user.
  3. 根据权利要求2所述的方法,其中,所述预先存储特定用户的用于进行手势操作的手臂的三维模型包括:The method according to claim 2, wherein said three-dimensional model of an arm for performing a gesture operation of a specific user in advance comprises:
    获取所述特定用户在多次手势操作中的多个手臂图像;Obtaining a plurality of arm images of the specific user in a plurality of gesture operations;
    针对所述多个手臂图像中的每一个,分解手臂的不同部位,并提取不同部位的特征;以及Decomposing different parts of the arm for each of the plurality of arm images and extracting features of the different parts;
    针对同一手臂融合相同部位的多个特征,并获得针对该手臂的三维模型。A plurality of features of the same portion are fused for the same arm, and a three-dimensional model for the arm is obtained.
  4. 根据权利要求2所述的方法,其中:The method of claim 2 wherein:
    所述将获取的手臂不同部位的特征与对应的三维模型中的特征进行比较包括:The comparing the acquired features of different parts of the arm with the features in the corresponding three-dimensional model includes:
    判断所述第一图像中手臂不同部位的尺寸与所述三维模型中的对应尺寸之间的差值是否小于预定阈值,获得第一判断结果;和/或Determining whether a difference between a size of a different part of the arm in the first image and a corresponding size in the three-dimensional model is less than a predetermined threshold, obtaining a first determination result; and/or
    判断所述第一图像中前臂与大臂之间的角度是否处于基于所述三维模型确定的预定范围之内,获得第二判断结果,Determining whether an angle between the forearm and the boom in the first image is within a predetermined range determined based on the three-dimensional model, obtaining a second determination result,
    所述判断所述手势输入是否由所述特定用户做出包括:Determining whether the gesture input is made by the specific user comprises:
    基于所述第一判断结果和/或第二判断结果,确定所述手势输入是否由所述特定用户做出。Based on the first determination result and/or the second determination result, it is determined whether the gesture input is made by the specific user.
  5. 根据权利要求1所述的方法,其中,所述基于所述第一图像,判断出所述手势输入是否由特定用户做出之后,进一步包括:The method of claim 1, wherein the determining, after the gesture input is made by a specific user based on the first image, further comprises:
    在确定所述手势输入由特定用户做出之后,判断是否接收到用户的撤销操作;After determining that the gesture input is made by a specific user, determining whether the user's undo operation is received;
    在确定所述手势输入由特定用户做出之后,判断识别出的手势输入是否与所述电子设备 正在执行的当前任务完全不相关;After determining that the gesture input is made by a specific user, determining whether the recognized gesture input is related to the electronic device The current task being executed is completely irrelevant;
    判断所述第一图像中是否有其他人出现;Determining whether another person appears in the first image;
    在确定所述手势输入不是由特定用户做出之后,判断是否接收到所述手势的重复输入;After determining that the gesture input is not made by a specific user, determining whether a repeated input of the gesture is received;
    基于以上判断结果中的至少一个,确定关于所述手势输入是否由特定用户做出的判断是否正确,以利用所述第一图像中的手臂特征更新对应的三维模型。Based on at least one of the above determination results, it is determined whether the determination as to whether the gesture input is made by a specific user is correct to update the corresponding three-dimensional model with the arm feature in the first image.
  6. 一种电子设备,包括:An electronic device comprising:
    图像采集模组,用于获取进行手势输入时的第一图像,所述第一图像包括手部和至少部分手臂的图像;An image acquisition module, configured to acquire a first image when the gesture input is performed, where the first image includes an image of the hand and at least a portion of the arm;
    存储器,用于存储程序;以及a memory for storing programs;
    处理器,用于当执行所述存储器上存储的程序时实现如下功能:a processor for performing the following functions when executing a program stored on the memory:
    基于所述第一图像,判断所述手势输入是否由特定用户做出;以及Determining whether the gesture input is made by a specific user based on the first image;
    如果判断结果为是,则对所述手势输入执行识别,否则忽略所述手势输入。If the result of the determination is yes, recognition is performed on the gesture input, otherwise the gesture input is ignored.
  7. 根据权利要求6所述的设备,其中:The device of claim 6 wherein:
    所述存储器进一步被配置为预先存储特定用户的用于进行手势操作的手臂的三维模型;The memory is further configured to pre-store a three-dimensional model of an arm of a specific user for performing a gesture operation;
    所述处理器被配置为执行所述程序以进一步实现如下功能:The processor is configured to execute the program to further implement the following functions:
    确定所述第一图像中包括的手臂为左臂还是右臂,并且在所述第一图像中分解手臂的不同部位,提取不同部位的特征;以及Determining whether the arm included in the first image is a left arm or a right arm, and decomposing different parts of the arm in the first image to extract features of different parts;
    将获取的手臂不同部位的特征与对应的三维模型中的特征进行比较,并判断所述手势输入是否由所述特定用户做出。The acquired features of different parts of the arm are compared with features in the corresponding three-dimensional model, and it is determined whether the gesture input is made by the specific user.
  8. 根据权利要求7所述的设备,其中,所述处理器被配置为执行所述程序以进一步实现如下功能:The apparatus of claim 7, wherein the processor is configured to execute the program to further implement the following functions:
    针对所述图像采集单元获取的所述特定用户在多次手势操作中的多个手臂图像中的每一个,分解手臂的不同部位,并提取不同部位的特征;以及Separating different parts of the arm and extracting features of different parts for each of the plurality of arm images of the specific user acquired by the image acquisition unit in the plurality of gesture operations;
    针对同一手臂融合相同部位的多个特征,并获得针对该手臂的三维模型。A plurality of features of the same portion are fused for the same arm, and a three-dimensional model for the arm is obtained.
  9. 根据权利要求7所述的设备,其中所述处理器被配置为执行所述程序以进一步实现如下功能:The apparatus of claim 7, wherein the processor is configured to execute the program to further implement the following functions:
    判断所述第一图像中手臂不同部位的尺寸与所述三维模型中的对应尺寸之间的差值是否小于预定阈值,获得第一判断结果;和/或Determining whether a difference between a size of a different part of the arm in the first image and a corresponding size in the three-dimensional model is less than a predetermined threshold, obtaining a first determination result; and/or
    判断所述第一图像中前臂与大臂之间的角度是否处于基于所述三维模型确定的预定范围之内,获得第二判断结果;以及Determining whether an angle between the forearm and the boom in the first image is within a predetermined range determined based on the three-dimensional model, obtaining a second determination result;
    基于所述第一判断结果和/或第二判断结果,确定所述手势输入是否由所述特定用户做出。 Based on the first determination result and/or the second determination result, it is determined whether the gesture input is made by the specific user.
  10. 根据权利要求6所述的设备,所述处理器被配置为执行所述程序以进一步实现如下功能:The apparatus of claim 6, the processor being configured to execute the program to further implement the following functions:
    在确定所述手势输入由特定用户做出之后,判断是否接收到用户的撤销操作;After determining that the gesture input is made by a specific user, determining whether the user's undo operation is received;
    在确定所述手势输入由特定用户做出之后,判断识别出的手势输入是否与所述电子设备正在执行的当前任务完全不相关;After determining that the gesture input is made by a specific user, determining whether the recognized gesture input is completely unrelated to the current task being executed by the electronic device;
    判断所述第一图像中是否有其他人出现;Determining whether another person appears in the first image;
    在确定所述手势输入不是由特定用户做出之后,判断是否接收到所述手势的重复输入;以及After determining that the gesture input is not made by a specific user, determining whether a repeated input of the gesture is received;
    基于以上判断结果中的至少一个,确定关于所述手势输入是否由特定用户做出的判断是否正确,以利用所述第一图像中的手臂特征更新对应的三维模型。 Based on at least one of the above determination results, it is determined whether the determination as to whether the gesture input is made by a specific user is correct to update the corresponding three-dimensional model with the arm feature in the first image.
PCT/CN2017/112312 2017-06-29 2017-11-22 Control method and electronic equipment for hand gesture recognition WO2019000817A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201710517121.7 2017-06-29
CN201710517121.7A CN107273869B (en) 2017-06-29 2017-06-29 Gesture recognition control method and electronic equipment

Publications (1)

Publication Number Publication Date
WO2019000817A1 true WO2019000817A1 (en) 2019-01-03

Family

ID=60070780

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2017/112312 WO2019000817A1 (en) 2017-06-29 2017-11-22 Control method and electronic equipment for hand gesture recognition

Country Status (2)

Country Link
CN (1) CN107273869B (en)
WO (1) WO2019000817A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115143722A (en) * 2021-03-31 2022-10-04 青岛海尔电冰箱有限公司 Control method of refrigerating device

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107273869B (en) * 2017-06-29 2020-04-24 联想(北京)有限公司 Gesture recognition control method and electronic equipment
CN109614953A (en) * 2018-12-27 2019-04-12 华勤通讯技术有限公司 A kind of control method based on image recognition, mobile unit and storage medium
CN114153308B (en) * 2020-09-08 2023-11-21 阿里巴巴集团控股有限公司 Gesture control method, gesture control device, electronic equipment and computer readable medium
CN112333511A (en) * 2020-09-27 2021-02-05 深圳Tcl新技术有限公司 Control method, device and equipment of smart television and computer readable storage medium
CN115220566A (en) * 2021-04-19 2022-10-21 北京有竹居网络技术有限公司 Gesture recognition method, device, equipment, medium and computer program product

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2016163068A1 (en) * 2015-04-07 2016-10-13 Sony Corporation Information processing apparatus, information processing method, and program
CN106484098A (en) * 2015-08-31 2017-03-08 柯尼卡美能达美国研究所有限公司 The real time interactive operation system and method for user interface
CN106537173A (en) * 2014-08-07 2017-03-22 谷歌公司 Radar-based gesture recognition
CN107273869A (en) * 2017-06-29 2017-10-20 联想(北京)有限公司 Gesture identification control method and electronic equipment

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7957762B2 (en) * 2007-01-07 2011-06-07 Apple Inc. Using ambient light sensor to augment proximity sensor output
US8726194B2 (en) * 2007-07-27 2014-05-13 Qualcomm Incorporated Item selection using enhanced control
CN104808788B (en) * 2015-03-18 2017-09-01 北京工业大学 A kind of method that non-contact gesture manipulates user interface

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106537173A (en) * 2014-08-07 2017-03-22 谷歌公司 Radar-based gesture recognition
WO2016163068A1 (en) * 2015-04-07 2016-10-13 Sony Corporation Information processing apparatus, information processing method, and program
CN106484098A (en) * 2015-08-31 2017-03-08 柯尼卡美能达美国研究所有限公司 The real time interactive operation system and method for user interface
CN107273869A (en) * 2017-06-29 2017-10-20 联想(北京)有限公司 Gesture identification control method and electronic equipment

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115143722A (en) * 2021-03-31 2022-10-04 青岛海尔电冰箱有限公司 Control method of refrigerating device
CN115143722B (en) * 2021-03-31 2023-10-24 青岛海尔电冰箱有限公司 Control method of refrigerating device

Also Published As

Publication number Publication date
CN107273869B (en) 2020-04-24
CN107273869A (en) 2017-10-20

Similar Documents

Publication Publication Date Title
WO2019000817A1 (en) Control method and electronic equipment for hand gesture recognition
US10296102B1 (en) Gesture and motion recognition using skeleton tracking
WO2017152794A1 (en) Method and device for target tracking
Sharp et al. Accurate, robust, and flexible real-time hand tracking
WO2018028546A1 (en) Key point positioning method, terminal, and computer storage medium
EP3113114B1 (en) Image processing method and device
CN106462242B (en) Use the user interface control of eye tracking
CN106845335B (en) Gesture recognition method and device for virtual reality equipment and virtual reality equipment
CN104635920B (en) The control method of gesture recognition device and gesture recognition device
WO2016127437A1 (en) Live body face verification method and system, and computer program product
WO2022166243A1 (en) Method, apparatus and system for detecting and identifying pinching gesture
US8897490B2 (en) Vision-based user interface and related method
US9671873B2 (en) Device interaction with spatially aware gestures
KR101745651B1 (en) System and method for recognizing hand gesture
JP2020518051A (en) Face posture detection method, device and storage medium
US11335025B2 (en) Method and device for joint point detection
WO2018103416A1 (en) Method and device for detecting facial image
WO2017113407A1 (en) Gesture recognition method and apparatus, and electronic device
WO2020124993A1 (en) Liveness detection method and apparatus, electronic device, and storage medium
US20170140215A1 (en) Gesture recognition method and virtual reality display output device
CN109740511B (en) Facial expression matching method, device, equipment and storage medium
EP3639193A1 (en) Human feedback in 3d model fitting
CN112183155B (en) Method and device for establishing action posture library, generating action posture and identifying action posture
CN114038016A (en) Sitting posture detection method, device, equipment and storage medium
CN114510142A (en) Gesture recognition method based on two-dimensional image, system thereof and electronic equipment

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17916307

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: 1205A 23.06.2020

122 Ep: pct application non-entry in european phase

Ref document number: 17916307

Country of ref document: EP

Kind code of ref document: A1