WO2018098861A1 - Gesture recognition method and device for virtual reality apparatus, and virtual reality apparatus - Google Patents

Gesture recognition method and device for virtual reality apparatus, and virtual reality apparatus Download PDF

Info

Publication number
WO2018098861A1
WO2018098861A1 PCT/CN2016/111062 CN2016111062W WO2018098861A1 WO 2018098861 A1 WO2018098861 A1 WO 2018098861A1 CN 2016111062 W CN2016111062 W CN 2016111062W WO 2018098861 A1 WO2018098861 A1 WO 2018098861A1
Authority
WO
WIPO (PCT)
Prior art keywords
current
gesture recognition
virtual reality
current user
image
Prior art date
Application number
PCT/CN2016/111062
Other languages
French (fr)
Chinese (zh)
Inventor
张茜
张绍谦
张超
Original Assignee
歌尔科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 歌尔科技有限公司 filed Critical 歌尔科技有限公司
Publication of WO2018098861A1 publication Critical patent/WO2018098861A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/017Gesture based interaction, e.g. based on a set of recognized hand gestures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • G06T19/006Mixed reality
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10028Range image; Depth image; 3D point clouds
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/107Static hand or arm
    • G06V40/117Biometrics derived from hands

Definitions

  • the present invention relates to the field of virtual reality device technologies, and in particular, to a gesture recognition method, apparatus, and virtual reality device for a virtual reality device.
  • Virtual Reality is a high-tech that has emerged in recent years.
  • Virtual reality technology will be a key technology to support a comprehensive integration of multidimensional information space combining qualitative and quantitative, perceptual knowledge and rational understanding.
  • As the speed of the Internet increases, an Internet era based on virtual reality technology is quietly coming, which will dramatically change people's production and lifestyle. It is conceivable that we can experience and interact with the virtual world through VR headsets, such as swimming in space, skydiving, and so on.
  • a gesture recognition method for a virtual reality device including a depth camera, and the gesture recognition method includes:
  • a gesture recognition apparatus for a virtual reality device including:
  • a first control module configured to control the depth camera to collect a current hand image of the current user
  • a determining module configured to determine, according to the current hand image, whether the current user performs a tapping action
  • a current feature extraction module configured to extract a current feature from the current hand image if the determination result of the determination module is YES;
  • a matching module configured to match the current feature with a reference feature in the model, and determine, according to the matching result, the button that the current user taps.
  • a virtual reality device comprising the gesture recognition device according to the second aspect of the invention.
  • a virtual reality device comprising a depth camera for acquiring an image, the memory for storing an instruction, the instruction for controlling the The processor performs the gesture recognition method according to the first aspect of the present invention.
  • a computer readable storage medium storing program code for performing the gesture recognition method according to the first aspect of the invention.
  • FIG. 1 is a flow chart of an embodiment of a gesture recognition method for a virtual reality device in accordance with the present invention
  • FIG. 2 is a block schematic diagram of an implementation structure of a gesture recognition apparatus for a virtual reality device according to the present invention
  • FIG. 3 is a block schematic diagram of an implementation structure of a virtual reality device in accordance with the present invention.
  • Reality devices include depth cameras.
  • the depth camera is also called a depth sensor or a 3D sensor.
  • it can be a TOF camera that emits modulated near-infrared light.
  • the camera calculates the distance of the object by calculating the time difference or phase difference between the light emission and reflection.
  • the three-dimensional contour of the object can be represented by images of different colors representing different distances.
  • FIG. 1 is a flow chart of an embodiment of a gesture recognition method for a virtual reality device in accordance with the present invention.
  • the gesture recognition method comprises the following steps:
  • Step S110 controlling the depth camera to collect the current hand image of the current user.
  • the TOF camera may be controlled to emit modulated near-infrared light, and the camera may reflect the distance between the current user's hand and the phase difference by calculating the light emission and reflection time difference or phase difference.
  • the three-dimensional contour of the hand can be represented by different colors to represent the current hand of different distances. The image is presented.
  • Step S120 determining, according to the current hand image, whether the current user performs a tap action, and if yes, executing step S130; if not, proceeding to step S110.
  • the curvature of each point of the hand contour in the current hand image is obtained, and the fingertip position can be calculated.
  • the curvature of each point is determined according to a certain step size.
  • the curvature of the fingertip has a certain range, and it is determined whether the curvature of each point is within the range by checking the curvature of each point.
  • the position of the fingertip can be determined, and the position of the other key points of the hand can be estimated by the morphology by the result of the gesture determination and the position of the fingertip, wherein the other key points can be, for example, a joint.
  • the depth sensor can be placed on top of the virtual reality device or anywhere in front of the user's hand.
  • the location of the depth sensor may be located on a virtual reality device.
  • the depth sensor collects the current hand image of the current user
  • the current fingertip coordinates of the current user's hands can be obtained by calculating the curvature of the hand contour point in the current hand image, and other key points of the user hand are calculated according to the user's fingertip coordinates. Point, hand fingertips and other key points are imaged by the user's hand image on top of the VR keyboard image by, for example, 3D rendering techniques.
  • the depth change of the user's finger can be confirmed to confirm whether or not the keyboard is tapped.
  • the image is clear, the resolution of the image is increased, and the viewing angle and visual distance can be adjusted appropriately.
  • the depth sensor is obtained as a depth map, the depth is different according to the distance of the hand from the depth sensor. Therefore, when the finger is raised, the depth value is small, and when the finger is dropped, the depth value is large, thereby determining whether the current user performs. Tap the action.
  • the virtual reality device further includes a display screen, and before performing step S120, the method further includes:
  • the keyboard image and the initial position of the current user's finger on the keyboard image are displayed on the display.
  • the current user uses the virtual device, for example, a head-mounted virtual reality device for character input
  • the current user can use the surrounding flat object such as a desk or the like.
  • the current user's hands can be placed according to the gesture of the physical keyboard, depth camera
  • the head obtains the current hand image and fingertip coordinates, which can be placed by the index finger fingertip normalization processing to place the left index finger on the keyboard button F, and the right hand index finger is placed on the keyboard button J, the current user can be based on the virtual reality Imaging of the device, appropriate adjustment of other finger positions, to fall to the correct initial position of each finger.
  • Step S130 extracting a current feature from the current hand image.
  • the extraction of current features may be implemented by a neural network (CNN) improved algorithm.
  • CNN neural network
  • the CNN improved algorithm obtains all the neurons of the hand according to the convolution and the feature points of the user's hand (for example, 10), and then according to the maximum pool layer and the fully connected layer, wherein the CNN improvement algorithm can be adopted, for example.
  • step S140 the current feature is matched with the reference feature in the model, and the button of the current user tap is determined according to the matching result.
  • the reference feature may be, for example, stored in the model before the virtual reality device is shipped from the factory, or may be stored by the current user before using the virtual reality device.
  • the gesture recognition method before performing step S140, the gesture recognition method further includes:
  • the reference hand image of the user's various actions of tapping the keys of the keyboard is collected, and the reference features are extracted from each reference hand image to establish a reference feature corresponding to various actions of the hand.
  • the model, the resulting model is applied to match the current feature, and if the match is successful, the button of the current user tap corresponding to the matching reference feature can be determined.
  • the gesture recognition method further includes:
  • the current position of the current user's finger on the keyboard image is displayed on the display screen.
  • the method of displaying the current position of the current user's finger on the keyboard image on the display screen may be the same as the above method of displaying the initial position.
  • the present invention can distinguish the various gestures, the left and right hand confirmation, and the fingertip coordinate acquisition by using the gesture recognition technology of the depth sensor, and the virtual keyboard can be created by the finger tapping the keyboard screen in front of the user to complete the character or the digital input.
  • the virtual keyboard can be created by the finger tapping the keyboard screen in front of the user to complete the character or the digital input.
  • it can greatly improve the flexibility of user use, and also release the physical keyboard space, reducing the user's trouble caused by complex character input, thereby improving the user experience.
  • the present invention also provides a gesture recognition apparatus for a virtual reality device
  • FIG. 2 is a block schematic diagram of an implementation structure of a gesture recognition apparatus for a virtual reality device according to the present invention.
  • the gesture recognition apparatus 200 includes a first control module 210 , a determination module 220 , a current feature extraction module 230 , and a matching module 240 .
  • the first control module 210 is configured to control the depth camera to collect the current hand image of the current user.
  • the determining module 220 is configured to determine, according to the current hand image, whether the current user performs a tapping action.
  • the current feature extraction module 230 is configured to extract the current feature from the current hand image if the determination result of the determination module is YES.
  • the matching module 240 is configured to match the current feature with the reference feature in the model, and determine a button of the current user tap according to the matching result.
  • the gesture recognition apparatus further includes a second control module and a reference feature extraction module, wherein the second control module is configured to control the depth camera to acquire the reference hand image of the reference user; and the reference feature extraction module is configured to extract the reference hand image. Refer to the feature and store the reference feature in the model.
  • the virtual reality device further includes a display screen
  • the gesture recognition device further includes a first display module configured to display the keyboard image and the initial position of the current user's finger on the keyboard image on the display screen.
  • the gesture recognition device further includes a second display module, configured to display, on the display screen, the current user's finger on the keyboard according to the button pressed by the current user determined by the matching module. Like the current location on.
  • the present invention also provides a virtual reality device.
  • the virtual reality device includes a gesture recognition apparatus 200 for a virtual reality device of the present invention.
  • the virtual reality device may be, for example, a product such as virtual reality glasses, a virtual reality helmet, or the like.
  • FIG. 3 is a block schematic diagram of an implementation of the virtual reality device in accordance with another aspect of the present invention.
  • the virtual reality device 300 includes a memory 301 and a processor 302 for storing instructions for controlling the processor 302 to operate to perform the above-described gesture recognition method for the virtual reality device.
  • the virtual reality device 300 further includes an interface device 303, an input device 304, a display device 305, a communication device 306, and the like.
  • an interface device 303 an input device 304
  • a display device 305 a display device 305
  • a communication device 306 and the like.
  • the present invention may relate only to some of the devices, such as the processor 301, the memory 302, the display device 305, and the like.
  • the communication device 306 can be wired or wirelessly communicated, for example.
  • the above interface device 303 includes, for example, a headphone jack, a USB interface, and the like.
  • the input device 304 described above may include, for example, a touch screen, a button, and the like.
  • the display device 305 described above is, for example, a liquid crystal display, a touch display, or the like.
  • the invention can be a system, method and/or computer program product.
  • the computer program product can include a computer readable storage medium loaded with various means for causing a processor to implement the present invention Computer readable program instructions.
  • the computer readable storage medium can be a tangible device that can hold and store the instructions used by the instruction execution device.
  • the computer readable storage medium can be, for example, but not limited to, an electrical storage device, a magnetic storage device, an optical storage device, an electromagnetic storage device, a semiconductor storage device, or any suitable combination of the foregoing.
  • Non-exhaustive list of computer readable storage media include: portable computer disks, hard disks, random access memory (RAM), read only memory (ROM), erasable programmable read only memory (EPROM) Or flash memory), static random access memory (SRAM), portable compact disk read only memory (CD-ROM), digital versatile disk (DVD), memory stick, floppy disk, mechanical encoding device, for example, with instructions stored thereon A raised structure in the hole card or groove, and any suitable combination of the above.
  • a computer readable storage medium as used herein is not to be interpreted as a transient signal itself, such as a radio wave or other freely propagating electromagnetic wave, an electromagnetic wave propagating through a waveguide or other transmission medium (eg, a light pulse through a fiber optic cable), or through a wire The electrical signal transmitted.
  • the computer readable program instructions described herein can be downloaded from a computer readable storage medium to various computing/processing devices or downloaded to an external computer or external storage device over a network, such as the Internet, a local area network, a wide area network, and/or a wireless network.
  • the network may include copper transmission cables, fiber optic transmissions, wireless transmissions, routers, firewalls, switches, gateway computers, and/or edge servers.
  • a network adapter card or network interface in each computing/processing device receives computer readable program instructions from the network and forwards the computer readable program instructions for storage in a computer readable storage medium in each computing/processing device .
  • Computer program instructions for performing the operations of the present invention may be assembly instructions, instruction set architecture (ISA) instructions, machine instructions, machine related instructions, microcode, firmware instructions, state setting data, or in one or more programming languages.
  • the computer readable program instructions can execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer, partly on the remote computer, or entirely on the remote computer or server. carried out.
  • the remote computer can be connected via any kind of network, including a local area network (LAN) or wide area network (WAN).
  • LAN local area network
  • WAN wide area network
  • the customized electronic circuit such as a programmable logic circuit, a field programmable gate array (FPGA), or a programmable logic array (PLA)
  • FPGA field programmable gate array
  • PLA programmable logic array
  • Computer readable program instructions are executed to implement various aspects of the present invention.
  • the computer readable program instructions can be provided to a general purpose computer, a special purpose computer, or a processor of other programmable data processing apparatus to produce a machine such that when executed by a processor of a computer or other programmable data processing apparatus Means for implementing the functions/acts specified in one or more of the blocks of the flowcharts and/or block diagrams.
  • the computer readable program instructions can also be stored in a computer readable storage medium that causes the computer, programmable data processing device, and/or other device to operate in a particular manner, such that the computer readable medium storing the instructions includes An article of manufacture that includes instructions for implementing various aspects of the functions/acts recited in one or more of the flowcharts.
  • the computer readable program instructions can also be loaded onto a computer, other programmable data processing device, or other device to perform a series of operational steps on a computer, other programmable data processing device or other device to produce a computer-implemented process.
  • instructions executed on a computer, other programmable data processing apparatus, or other device implement the functions/acts recited in one or more of the flowcharts and/or block diagrams.
  • each block in the flowchart or block diagram can represent a module, a program segment, or a portion of an instruction that includes one or more components for implementing the specified logical functions.
  • Executable instructions can also occur in a different order than those illustrated in the drawings. For example, two consecutive blocks may be executed substantially in parallel, and they may sometimes be executed in the reverse order, depending upon the functionality involved.
  • each block of the block diagrams and/or flowcharts, and combinations of blocks in the block diagrams and/or flowcharts can be implemented in a dedicated hardware-based system that performs the specified function or function. Or it can be implemented by a combination of dedicated hardware and computer instructions. It is well known to those skilled in the art that implementation by hardware, implementation by software, and implementation by a combination of software and hardware are equivalent.

Abstract

A gesture recognition method and device for a virtual reality apparatus, and a virtual reality apparatus. The gesture recognition method comprises: controlling a depth camera to capture a current hand image of a current user (S110); determining, according to the current hand image, whether the current user executes a tapping action or not (S120); if yes, extracting current features from the current hand image (S130); and matching the current features with reference features in a model, and determining, according to the matching result, a key tapped by the current user (S140). By applying a virtual keyboard to reality, the gesture recognition method improves the use flexibility of users, and also releases space of a real keyboard, and reduces the trouble of the user caused by complex character input, thereby improving user experience.

Description

用于虚拟现实设备的手势识别方法、装置及虚拟现实设备Gesture recognition method, device and virtual reality device for virtual reality device 技术领域Technical field
本发明涉及虚拟现实设备技术领域,更具体地,涉及一种用于虚拟现实设备的手势识别方法、装置及虚拟现实设备。The present invention relates to the field of virtual reality device technologies, and in particular, to a gesture recognition method, apparatus, and virtual reality device for a virtual reality device.
背景技术Background technique
虚拟现实(Virtual Reality,简称VR)是近年来出现的高新技术。虚拟现实技术将是支撑一个定性和定量相结合,感性认识和理性认识相结合的综合集成多维信息空间的关键技术。随着网络的速度的提升,基于虚拟现实技术的一个互联网时代正悄然走来,它将极大地改变人们的生产和生活方式。可以想象我们可以通过VR头戴进行游太空、跳伞等等不敢去玩又想尝试的事,实现对虚拟世界进行体验和交互作用。Virtual Reality (VR) is a high-tech that has emerged in recent years. Virtual reality technology will be a key technology to support a comprehensive integration of multidimensional information space combining qualitative and quantitative, perceptual knowledge and rational understanding. As the speed of the Internet increases, an Internet era based on virtual reality technology is quietly coming, which will dramatically change people's production and lifestyle. It is conceivable that we can experience and interact with the virtual world through VR headsets, such as swimming in space, skydiving, and so on.
目前,新兴的虚拟现实技术已渗透到办公、娱乐等领域,引起许多行业的变革。At present, the emerging virtual reality technology has penetrated into the fields of office and entertainment, which has caused changes in many industries.
发明内容Summary of the invention
根据本发明的第一方面,提供了一种用于虚拟现实设备的手势识别方法,所述虚拟现实设备包括深度摄像头,所述手势识别方法包括:According to a first aspect of the present invention, a gesture recognition method for a virtual reality device is provided, the virtual reality device including a depth camera, and the gesture recognition method includes:
控制所述深度摄像头采集当前用户的当前手部图像;Controlling the depth camera to capture a current hand image of the current user;
根据所述当前手部图像判断所述当前用户是否执行敲击动作,如是,则:Determining, according to the current hand image, whether the current user performs a tap action, and if so, then:
从所述当前手部图像中提取当前特征;Extracting a current feature from the current hand image;
将所述当前特征与模型中的参考特征进行匹配,根据匹配结果确定所述当前用户敲击的按键。Matching the current feature with a reference feature in the model, and determining a button of the current user tap based on the matching result.
根据本发明的第二方面,提供了一种用于虚拟现实设备的手势识别装置,包括: According to a second aspect of the present invention, a gesture recognition apparatus for a virtual reality device is provided, including:
第一控制模块,用于控制所述深度摄像头采集当前用户的当前手部图像;a first control module, configured to control the depth camera to collect a current hand image of the current user;
判断模块,用于根据所述当前手部图像判断所述当前用户是否执行敲击动作;a determining module, configured to determine, according to the current hand image, whether the current user performs a tapping action;
当前特征提取模块,用于在所述判断模块的判断结果为是的情况下,从所述当前手部图像中提取当前特征;以及,a current feature extraction module, configured to extract a current feature from the current hand image if the determination result of the determination module is YES;
匹配模块,用于将所述当前特征与模型中的参考特征进行匹配,根据匹配结果确定所述当前用户敲击的按键。And a matching module, configured to match the current feature with a reference feature in the model, and determine, according to the matching result, the button that the current user taps.
根据本发明的第三方面,提供了一种虚拟现实设备,包括根据本发明第二方面所述的手势识别装置。According to a third aspect of the invention, there is provided a virtual reality device comprising the gesture recognition device according to the second aspect of the invention.
根据本发明的第四方面,提供了一种虚拟现实设备,包括深度摄像头、处理器和存储器,所述深度摄像头用于采集图像,所述存储器用于存储指令,所述指令用于控制所述处理器执行根据本发明第一方面所述的手势识别方法。According to a fourth aspect of the present invention, there is provided a virtual reality device comprising a depth camera for acquiring an image, the memory for storing an instruction, the instruction for controlling the The processor performs the gesture recognition method according to the first aspect of the present invention.
根据本发明的第五方面,提供了一种计算机可读存储介质,其存储有用于执行根据本发明的第一方面所述手势识别方法的程序代码。According to a fifth aspect of the invention, there is provided a computer readable storage medium storing program code for performing the gesture recognition method according to the first aspect of the invention.
通过以下参照附图对本发明的示例性实施例的详细描述,本发明的其它特征及其优点将会变得清楚。Other features and advantages of the present invention will become apparent from the Detailed Description of the <RTIgt;
附图说明DRAWINGS
被结合在说明书中并构成说明书的一部分的附图示出了本发明的实施例,并且连同其说明一起用于解释本发明的原理。The accompanying drawings, which are incorporated in FIG
图1为根据本发明一种用于虚拟现实设备的手势识别方法的一种实施方式的流程图;1 is a flow chart of an embodiment of a gesture recognition method for a virtual reality device in accordance with the present invention;
图2为根据本发明一种用于虚拟现实设备的手势识别装置的一种实施结构的方框原理图;2 is a block schematic diagram of an implementation structure of a gesture recognition apparatus for a virtual reality device according to the present invention;
图3为根据本发明一种虚拟现实设备的一种实施结构的方框原理图。3 is a block schematic diagram of an implementation structure of a virtual reality device in accordance with the present invention.
具体实施方式 detailed description
现在将参照附图来详细描述本发明的各种示例性实施例。应注意到:除非另外具体说明,否则在这些实施例中阐述的部件和步骤的相对布置、数字表达式和数值不限制本发明的范围。Various exemplary embodiments of the present invention will now be described in detail with reference to the drawings. It should be noted that the relative arrangement of the components and steps, numerical expressions and numerical values set forth in the embodiments are not intended to limit the scope of the invention unless otherwise specified.
以下对至少一个示例性实施例的描述实际上仅仅是说明性的,决不作为对本发明及其应用或使用的任何限制。The following description of the at least one exemplary embodiment is merely illustrative and is in no way
对于相关领域普通技术人员已知的技术、方法和设备可能不作详细讨论,但在适当情况下,所述技术、方法和设备应当被视为说明书的一部分。Techniques, methods and apparatus known to those of ordinary skill in the relevant art may not be discussed in detail, but the techniques, methods and apparatus should be considered as part of the specification, where appropriate.
在这里示出和讨论的所有例子中,任何具体值应被解释为仅仅是示例性的,而不是作为限制。因此,示例性实施例的其它例子可以具有不同的值。In all of the examples shown and discussed herein, any specific values are to be construed as illustrative only and not as a limitation. Thus, other examples of the exemplary embodiments may have different values.
应注意到:相似的标号和字母在下面的附图中表示类似项,因此,一旦某一项在一个附图中被定义,则在随后的附图中不需要对其进行进一步讨论。It should be noted that similar reference numerals and letters indicate similar items in the following figures, and therefore, once an item is defined in one figure, it is not required to be further discussed in the subsequent figures.
为了解决现有技术中存在的头戴式虚拟现实设备的交互主要是通过语言、手势等,无法实现对文字的复杂处理的问题,提供了一种用于虚拟现实设备的手势识别方法,该虚拟现实设备包括深度摄像头。In order to solve the problem that the interaction of the head-mounted virtual reality device existing in the prior art is mainly through language, gesture, etc., the complex processing of the text cannot be realized, and a gesture recognition method for the virtual reality device is provided. Reality devices include depth cameras.
深度摄像头也称为深度传感器或者3D传感器,例如可以是TOF摄像头,发出经调制的近红外光,遇物体后反射,摄像头通过计算光线发射和反射时间差或相位差,来换算被拍摄物体的距离,以产生深度信息,此外再结合传统的相机拍摄,就能将物体的三维轮廓以不同颜色代表不同距离的图像呈现出来。The depth camera is also called a depth sensor or a 3D sensor. For example, it can be a TOF camera that emits modulated near-infrared light. When it is reflected by an object, the camera calculates the distance of the object by calculating the time difference or phase difference between the light emission and reflection. In order to generate depth information, in addition to the traditional camera shooting, the three-dimensional contour of the object can be represented by images of different colors representing different distances.
图1为根据本发明一种用于虚拟现实设备的手势识别方法的一种实施方式的流程图。1 is a flow chart of an embodiment of a gesture recognition method for a virtual reality device in accordance with the present invention.
根据图1所示,该手势识别方法包括以下步骤:According to FIG. 1, the gesture recognition method comprises the following steps:
步骤S110,控制深度摄像头采集当前用户的当前手部图像。Step S110, controlling the depth camera to collect the current hand image of the current user.
具体的,例如可以是控制TOF摄像头,发出经调制的近红外光,遇当前用户的手部后反射,摄像头通过计算光线发射和反射时间差或相位差,来换算当前用户手部任意位置的距离,以产生深度信息,此外再结合传统的相机拍摄,就能将手部的三维轮廓以不同颜色代表不同距离的当前手部 图像呈现出来。Specifically, for example, the TOF camera may be controlled to emit modulated near-infrared light, and the camera may reflect the distance between the current user's hand and the phase difference by calculating the light emission and reflection time difference or phase difference. In order to generate depth information, and in combination with the traditional camera shooting, the three-dimensional contour of the hand can be represented by different colors to represent the current hand of different distances. The image is presented.
步骤S120,根据该当前手部图像判断当前用户是否执行敲击动作,如是,则执行步骤S130;如否,则继续执行步骤S110。Step S120: determining, according to the current hand image, whether the current user performs a tap action, and if yes, executing step S130; if not, proceeding to step S110.
根据深度摄像头采集得到当前手部图像中手部轮廓各点的曲率,可以计算得到指尖位置。对当前手部图像中手部轮廓的每一个点根据一定的步长确定每个点的曲率,手指指尖的曲率是有一定范围的,通过查看每个点的曲率判定是否在该范围内就可确定指尖的位置,通过手势判定的结果以及指尖的位置再通过形态学可推算手部其他关键点的位置,其中,其他关键点例如可以是关节。According to the depth camera, the curvature of each point of the hand contour in the current hand image is obtained, and the fingertip position can be calculated. For each point of the hand contour in the current hand image, the curvature of each point is determined according to a certain step size. The curvature of the fingertip has a certain range, and it is determined whether the curvature of each point is within the range by checking the curvature of each point. The position of the fingertip can be determined, and the position of the other key points of the hand can be estimated by the morphology by the result of the gesture determination and the position of the fingertip, wherein the other key points can be, for example, a joint.
深度传感器可以设置在虚拟现实设备上面或者是用户手部前面的任意位置上。在本发明的一个具体实施例中,深度传感器的位置可以是设置在虚拟现实设备上。深度传感器采集到当前用户的当前手部图像后,通过当前手部图像中手部轮廓点的曲率的计算可得到当前用户的双手当前指尖坐标,根据用户指尖坐标推算出用户手部其他关键点,手部指尖和其他关键点通过例如是3D渲染技术得到用户手部图像成像在VR的键盘图像之上。之后根据当前手部图像可以确认用户手指的深度变化从而确认有无敲击键盘。The depth sensor can be placed on top of the virtual reality device or anywhere in front of the user's hand. In a specific embodiment of the invention, the location of the depth sensor may be located on a virtual reality device. After the depth sensor collects the current hand image of the current user, the current fingertip coordinates of the current user's hands can be obtained by calculating the curvature of the hand contour point in the current hand image, and other key points of the user hand are calculated according to the user's fingertip coordinates. Point, hand fingertips and other key points are imaged by the user's hand image on top of the VR keyboard image by, for example, 3D rendering techniques. Then, based on the current hand image, the depth change of the user's finger can be confirmed to confirm whether or not the keyboard is tapped.
采用3D渲染技术,使得图像清晰,调高了图像的分辨率,并可适当调整可视角度和可视距离。Using 3D rendering technology, the image is clear, the resolution of the image is increased, and the viewing angle and visual distance can be adjusted appropriately.
由于深度传感器获得是深度图,根据手部离深度传感器的远近,深度是不一样的,因此,当手指抬起时深度值较小,手指落下时深度值较大,从而可以确定当前用户是否执行敲击动作。Since the depth sensor is obtained as a depth map, the depth is different according to the distance of the hand from the depth sensor. Therefore, when the finger is raised, the depth value is small, and when the finger is dropped, the depth value is large, thereby determining whether the current user performs. Tap the action.
在本发明的一个具体实施例中,该虚拟现实设备还包括显示屏,在执行步骤S120之前还包括:In a specific embodiment of the present invention, the virtual reality device further includes a display screen, and before performing step S120, the method further includes:
在显示屏上显示出键盘图像及当前用户的手指在该键盘图像上的初始位置。The keyboard image and the initial position of the current user's finger on the keyboard image are displayed on the display.
当前用户使用该虚拟设备例如可以是头戴式虚拟现实设备进行字符输入时,为了减轻当前用户手臂的不适,当前用户可借助周围的平板的物件例如桌子等。当前用户的双手可以按照实体键盘的手势放置,深度摄像 头得到当前手部图像和指尖坐标,可以是通过食指指尖归一化处理将左手食指放置在键盘的按键F处,将右手的食指放置在键盘的按键J处,当前用户可根据虚拟现实设备的成像,适当的调整其他手指位置,从而落到各个手指的正确初始位置。When the current user uses the virtual device, for example, a head-mounted virtual reality device for character input, in order to alleviate the discomfort of the current user's arm, the current user can use the surrounding flat object such as a desk or the like. The current user's hands can be placed according to the gesture of the physical keyboard, depth camera The head obtains the current hand image and fingertip coordinates, which can be placed by the index finger fingertip normalization processing to place the left index finger on the keyboard button F, and the right hand index finger is placed on the keyboard button J, the current user can be based on the virtual reality Imaging of the device, appropriate adjustment of other finger positions, to fall to the correct initial position of each finger.
这样,能够使得当前用户清楚地看到键盘上需要按下的按键的位置,以此来确定敲击的动作,提升用户体验。In this way, the current user can clearly see the position of the button on the keyboard that needs to be pressed, thereby determining the action of the tap and improving the user experience.
步骤S130,从当前手部图像中提取当前特征。Step S130, extracting a current feature from the current hand image.
在本发明的一个具体实施例中,当前特征的提取可以是通过神经网络(CNN)改进算法实现的。CNN改进算法,通过卷积和得到用户手部的特征点(例如10几个),之后再根据最大池层和全连接层从而得到手部所有的神经元,其中,CNN改进算法例如可以是采用caffe或者是tensorflow等CNN的算法架构,再从中提取出当前特征。In a specific embodiment of the invention, the extraction of current features may be implemented by a neural network (CNN) improved algorithm. The CNN improved algorithm obtains all the neurons of the hand according to the convolution and the feature points of the user's hand (for example, 10), and then according to the maximum pool layer and the fully connected layer, wherein the CNN improvement algorithm can be adopted, for example. Caffe or the algorithm architecture of CNN such as tensorflow, and extract the current features from it.
步骤S140,将该当前特征与模型中的参考特征进行匹配,根据匹配结果确定当前用户敲击的按键。In step S140, the current feature is matched with the reference feature in the model, and the button of the current user tap is determined according to the matching result.
参考特征例如可以是在虚拟现实设备出厂前存储在模型中的,也可以是当前用户在使用该虚拟现实设备之前存储的。The reference feature may be, for example, stored in the model before the virtual reality device is shipped from the factory, or may be stored by the current user before using the virtual reality device.
在本发明的一个具体实施例中,在执行步骤S140之前,该手势识别方法还包括:In a specific embodiment of the present invention, before performing step S140, the gesture recognition method further includes:
控制该深度摄像头采集参考用户的参考手部图像;Controlling the depth camera to acquire a reference hand image of the reference user;
从所述参考手部图像中提取所述参考特征,并将所述参考特征存储在该模型中。Extracting the reference feature from the reference hand image and storing the reference feature in the model.
具体的是,采集用户关于敲击键盘各个按键的手部各种动作的参考手部图像,并从每一参考手部图像中提取参考特征,建立包含与手部各种动作对应的参考特征的模型,得到的模型会应用于与当前特征进行匹配,如果匹配成功,则能够确定与匹配成功的参考特征相对应的当前用户敲击的按键。Specifically, the reference hand image of the user's various actions of tapping the keys of the keyboard is collected, and the reference features are extracted from each reference hand image to establish a reference feature corresponding to various actions of the hand. The model, the resulting model is applied to match the current feature, and if the match is successful, the button of the current user tap corresponding to the matching reference feature can be determined.
执行完步骤S140之后,该手势识别方法还包括:After performing step S140, the gesture recognition method further includes:
根据当前用户敲击的按键,在显示屏上显示出当前用户的手指在键盘图像上的当前位置。 According to the button that the current user taps, the current position of the current user's finger on the keyboard image is displayed on the display screen.
具体的,在显示屏上显示出当前用户的手指在键盘图像上的当前位置的方法,与上述的显示初始位置的方法可以相同。Specifically, the method of displaying the current position of the current user's finger on the keyboard image on the display screen may be the same as the above method of displaying the initial position.
这样,本发明通过使用深度传感器的手势识别技术来区别各种手势、左右手确认以及手指指尖坐标获取,在用户眼前出现手指敲击键盘画面从而完成字符或者是数字的输入,就能够将虚拟键盘运用到现实中,可以很好的调高用户使用的灵活性,同时还释放了实体键盘空间,减少用户因复杂的字符输入产生的困扰,从而提高用户体验。In this way, the present invention can distinguish the various gestures, the left and right hand confirmation, and the fingertip coordinate acquisition by using the gesture recognition technology of the depth sensor, and the virtual keyboard can be created by the finger tapping the keyboard screen in front of the user to complete the character or the digital input. In the real world, it can greatly improve the flexibility of user use, and also release the physical keyboard space, reducing the user's trouble caused by complex character input, thereby improving the user experience.
本发明还提供了一种用于虚拟现实设备的手势识别装置,图2为根据本发明一种用于虚拟现实设备的手势识别装置的一种实施结构的方框原理图。The present invention also provides a gesture recognition apparatus for a virtual reality device, and FIG. 2 is a block schematic diagram of an implementation structure of a gesture recognition apparatus for a virtual reality device according to the present invention.
根据图2所示,该手势识别装置200包括第一控制模块210、判断模块220、当前特征提取模块230和匹配模块240。According to FIG. 2 , the gesture recognition apparatus 200 includes a first control module 210 , a determination module 220 , a current feature extraction module 230 , and a matching module 240 .
上述第一控制模块210用于控制深度摄像头采集当前用户的当前手部图像。The first control module 210 is configured to control the depth camera to collect the current hand image of the current user.
上述判断模块220用于根据当前手部图像判断当前用户是否执行敲击动作。The determining module 220 is configured to determine, according to the current hand image, whether the current user performs a tapping action.
上述当前特征提取模块230用于在判断模块的判断结果为是的情况下,从当前手部图像中提取当前特征。The current feature extraction module 230 is configured to extract the current feature from the current hand image if the determination result of the determination module is YES.
上述匹配模块240用于将当前特征与模型中的参考特征进行匹配,根据匹配结果确定当前用户敲击的按键。The matching module 240 is configured to match the current feature with the reference feature in the model, and determine a button of the current user tap according to the matching result.
具体的,该手势识别装置还包括第二控制模块和参考特征提取模块,第二控制模块用于控制深度摄像头采集参考用户的参考手部图像;参考特征提取模块用于从参考手部图像中提取参考特征,并将参考特征存储在模型中。Specifically, the gesture recognition apparatus further includes a second control module and a reference feature extraction module, wherein the second control module is configured to control the depth camera to acquire the reference hand image of the reference user; and the reference feature extraction module is configured to extract the reference hand image. Refer to the feature and store the reference feature in the model.
进一步地,虚拟现实设备还包括显示屏,手势识别装置还包括第一显示模块,用于在显示屏上显示出键盘图像及当前用户的手指在键盘图像上的初始位置。Further, the virtual reality device further includes a display screen, and the gesture recognition device further includes a first display module configured to display the keyboard image and the initial position of the current user's finger on the keyboard image on the display screen.
在此基础上,手势识别装置还包括第二显示模块,用于根据匹配模块确定的当前用户敲击的按键,在显示屏上显示出当前用户的手指在键盘图 像上的当前位置。On the basis of the above, the gesture recognition device further includes a second display module, configured to display, on the display screen, the current user's finger on the keyboard according to the button pressed by the current user determined by the matching module. Like the current location on.
本发明还提供了一种虚拟现实设备,根据一方面,该虚拟现实设备包括本发明的一种用于虚拟现实设备的手势识别装置200。该虚拟现实设备例如可以是虚拟现实眼镜、虚拟现实头盔等产品。The present invention also provides a virtual reality device. According to one aspect, the virtual reality device includes a gesture recognition apparatus 200 for a virtual reality device of the present invention. The virtual reality device may be, for example, a product such as virtual reality glasses, a virtual reality helmet, or the like.
图3为根据本发明另一方面的该虚拟现实设备的一种实施结构的方框原理图。3 is a block schematic diagram of an implementation of the virtual reality device in accordance with another aspect of the present invention.
根据图3所示,该虚拟现实设备300包括存储器301和处理器302,该存储器301用于存储指令,该指令用于控制处理器302进行操作以执行上述用于虚拟现实设备的手势识别方法。According to FIG. 3, the virtual reality device 300 includes a memory 301 and a processor 302 for storing instructions for controlling the processor 302 to operate to perform the above-described gesture recognition method for the virtual reality device.
除此之外,根据图3所示,该虚拟现实设备300还包括接口装置303、输入装置304、显示装置305、通信装置306等等。尽管在图3中示出了多个装置,但是,本发明可以仅涉及其中的部分装置,例如,处理器301、存储器302、显示装置305等。In addition, as shown in FIG. 3, the virtual reality device 300 further includes an interface device 303, an input device 304, a display device 305, a communication device 306, and the like. Although a plurality of devices are illustrated in FIG. 3, the present invention may relate only to some of the devices, such as the processor 301, the memory 302, the display device 305, and the like.
上述通信装置306例如能够进行有有线或无线通信。The communication device 306 can be wired or wirelessly communicated, for example.
上述接口装置303例如包括耳机插孔、USB接口等。The above interface device 303 includes, for example, a headphone jack, a USB interface, and the like.
上述输入装置304例如可以包括触摸屏、按键等。The input device 304 described above may include, for example, a touch screen, a button, and the like.
上述显示装置305例如是液晶显示屏、触摸显示屏等。The display device 305 described above is, for example, a liquid crystal display, a touch display, or the like.
上述各实施例主要重点描述与其他实施例的不同之处,但本领域技术人员应当清楚的是,上述各实施例可以根据需要单独使用或者相互结合使用。The above embodiments mainly focus on the differences from the other embodiments, but it should be apparent to those skilled in the art that the above embodiments may be used alone or in combination with each other as needed.
本说明书中的各个实施例均采用递进的方式描述,各个实施例之间相同相似的部分相互参见即可,每个实施例重点说明的都是与其他实施例的不同之处,但本领域技术人员应当清楚的是,上述各实施例可以根据需要单独使用或者相互结合使用。另外,对于装置实施例而言,由于其是与方法实施例相对应,所以描述得比较简单,相关之处参见方法实施例的对应部分的说明即可。以上所描述的系统实施例仅仅是示意性的,其中作为分离部件说明的模块可以是或者也可以不是物理上分开的。The various embodiments in the present specification are described in a progressive manner, and the same or similar parts between the various embodiments may be referred to each other, and each embodiment focuses on the differences from other embodiments, but the field It should be clear to the skilled person that the above embodiments can be used individually or in combination with each other as needed. In addition, for the device embodiment, since it corresponds to the method embodiment, the description is relatively simple, and the relevant parts can be referred to the description of the corresponding part of the method embodiment. The system embodiments described above are merely illustrative, and the modules illustrated as separate components may or may not be physically separate.
本发明可以是系统、方法和/或计算机程序产品。计算机程序产品可以包括计算机可读存储介质,其上载有用于使处理器实现本发明的各个方 面的计算机可读程序指令。The invention can be a system, method and/or computer program product. The computer program product can include a computer readable storage medium loaded with various means for causing a processor to implement the present invention Computer readable program instructions.
计算机可读存储介质可以是可以保持和存储由指令执行设备使用的指令的有形设备。计算机可读存储介质例如可以是――但不限于――电存储设备、磁存储设备、光存储设备、电磁存储设备、半导体存储设备或者上述的任意合适的组合。计算机可读存储介质的更具体的例子(非穷举的列表)包括:便携式计算机盘、硬盘、随机存取存储器(RAM)、只读存储器(ROM)、可擦式可编程只读存储器(EPROM或闪存)、静态随机存取存储器(SRAM)、便携式压缩盘只读存储器(CD-ROM)、数字多功能盘(DVD)、记忆棒、软盘、机械编码设备、例如其上存储有指令的打孔卡或凹槽内凸起结构、以及上述的任意合适的组合。这里所使用的计算机可读存储介质不被解释为瞬时信号本身,诸如无线电波或者其他自由传播的电磁波、通过波导或其他传输媒介传播的电磁波(例如,通过光纤电缆的光脉冲)、或者通过电线传输的电信号。The computer readable storage medium can be a tangible device that can hold and store the instructions used by the instruction execution device. The computer readable storage medium can be, for example, but not limited to, an electrical storage device, a magnetic storage device, an optical storage device, an electromagnetic storage device, a semiconductor storage device, or any suitable combination of the foregoing. More specific examples (non-exhaustive list) of computer readable storage media include: portable computer disks, hard disks, random access memory (RAM), read only memory (ROM), erasable programmable read only memory (EPROM) Or flash memory), static random access memory (SRAM), portable compact disk read only memory (CD-ROM), digital versatile disk (DVD), memory stick, floppy disk, mechanical encoding device, for example, with instructions stored thereon A raised structure in the hole card or groove, and any suitable combination of the above. A computer readable storage medium as used herein is not to be interpreted as a transient signal itself, such as a radio wave or other freely propagating electromagnetic wave, an electromagnetic wave propagating through a waveguide or other transmission medium (eg, a light pulse through a fiber optic cable), or through a wire The electrical signal transmitted.
这里所描述的计算机可读程序指令可以从计算机可读存储介质下载到各个计算/处理设备,或者通过网络、例如因特网、局域网、广域网和/或无线网下载到外部计算机或外部存储设备。网络可以包括铜传输电缆、光纤传输、无线传输、路由器、防火墙、交换机、网关计算机和/或边缘服务器。每个计算/处理设备中的网络适配卡或者网络接口从网络接收计算机可读程序指令,并转发该计算机可读程序指令,以供存储在各个计算/处理设备中的计算机可读存储介质中。The computer readable program instructions described herein can be downloaded from a computer readable storage medium to various computing/processing devices or downloaded to an external computer or external storage device over a network, such as the Internet, a local area network, a wide area network, and/or a wireless network. The network may include copper transmission cables, fiber optic transmissions, wireless transmissions, routers, firewalls, switches, gateway computers, and/or edge servers. A network adapter card or network interface in each computing/processing device receives computer readable program instructions from the network and forwards the computer readable program instructions for storage in a computer readable storage medium in each computing/processing device .
用于执行本发明操作的计算机程序指令可以是汇编指令、指令集架构(ISA)指令、机器指令、机器相关指令、微代码、固件指令、状态设置数据、或者以一种或多种编程语言的任意组合编写的源代码或目标代码,所述编程语言包括面向对象的编程语言—诸如Smalltalk、C++等,以及常规的过程式编程语言—诸如“C”语言或类似的编程语言。计算机可读程序指令可以完全地在用户计算机上执行、部分地在用户计算机上执行、作为一个独立的软件包执行、部分在用户计算机上部分在远程计算机上执行、或者完全在远程计算机或服务器上执行。在涉及远程计算机的情形中,远程计算机可以通过任意种类的网络—包括局域网(LAN)或广域网(WAN)—连接 到用户计算机,或者,可以连接到外部计算机(例如利用因特网服务提供商来通过因特网连接)。在一些实施例中,通过利用计算机可读程序指令的状态信息来个性化定制电子电路,例如可编程逻辑电路、现场可编程门阵列(FPGA)或可编程逻辑阵列(PLA),该电子电路可以执行计算机可读程序指令,从而实现本发明的各个方面。Computer program instructions for performing the operations of the present invention may be assembly instructions, instruction set architecture (ISA) instructions, machine instructions, machine related instructions, microcode, firmware instructions, state setting data, or in one or more programming languages. Source code or object code written in any combination, including object oriented programming languages such as Smalltalk, C++, etc., as well as conventional procedural programming languages such as the "C" language or similar programming languages. The computer readable program instructions can execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer, partly on the remote computer, or entirely on the remote computer or server. carried out. In the case of a remote computer, the remote computer can be connected via any kind of network, including a local area network (LAN) or wide area network (WAN). To the user's computer, or, can be connected to an external computer (for example, using an Internet service provider to connect via the Internet). In some embodiments, the customized electronic circuit, such as a programmable logic circuit, a field programmable gate array (FPGA), or a programmable logic array (PLA), can be customized by utilizing state information of computer readable program instructions. Computer readable program instructions are executed to implement various aspects of the present invention.
这里参照根据本发明实施例的方法、装置(系统)和计算机程序产品的流程图和/或框图描述了本发明的各个方面。应当理解,流程图和/或框图的每个方框以及流程图和/或框图中各方框的组合,都可以由计算机可读程序指令实现。Aspects of the present invention are described herein with reference to flowchart illustrations and/or block diagrams of methods, apparatus, and computer program products according to embodiments of the invention. It will be understood that each block of the flowchart illustrations and/or block diagrams, and combinations of blocks in the flowcharts and/or block diagrams can be implemented by computer readable program instructions.
这些计算机可读程序指令可以提供给通用计算机、专用计算机或其它可编程数据处理装置的处理器,从而生产出一种机器,使得这些指令在通过计算机或其它可编程数据处理装置的处理器执行时,产生了实现流程图和/或框图中的一个或多个方框中规定的功能/动作的装置。也可以把这些计算机可读程序指令存储在计算机可读存储介质中,这些指令使得计算机、可编程数据处理装置和/或其他设备以特定方式工作,从而,存储有指令的计算机可读介质则包括一个制造品,其包括实现流程图和/或框图中的一个或多个方框中规定的功能/动作的各个方面的指令。The computer readable program instructions can be provided to a general purpose computer, a special purpose computer, or a processor of other programmable data processing apparatus to produce a machine such that when executed by a processor of a computer or other programmable data processing apparatus Means for implementing the functions/acts specified in one or more of the blocks of the flowcharts and/or block diagrams. The computer readable program instructions can also be stored in a computer readable storage medium that causes the computer, programmable data processing device, and/or other device to operate in a particular manner, such that the computer readable medium storing the instructions includes An article of manufacture that includes instructions for implementing various aspects of the functions/acts recited in one or more of the flowcharts.
也可以把计算机可读程序指令加载到计算机、其它可编程数据处理装置、或其它设备上,使得在计算机、其它可编程数据处理装置或其它设备上执行一系列操作步骤,以产生计算机实现的过程,从而使得在计算机、其它可编程数据处理装置、或其它设备上执行的指令实现流程图和/或框图中的一个或多个方框中规定的功能/动作。The computer readable program instructions can also be loaded onto a computer, other programmable data processing device, or other device to perform a series of operational steps on a computer, other programmable data processing device or other device to produce a computer-implemented process. Thus, instructions executed on a computer, other programmable data processing apparatus, or other device implement the functions/acts recited in one or more of the flowcharts and/or block diagrams.
附图中的流程图和框图显示了根据本发明的多个实施例的系统、方法和计算机程序产品的可能实现的体系架构、功能和操作。在这点上,流程图或框图中的每个方框可以代表一个模块、程序段或指令的一部分,所述模块、程序段或指令的一部分包含一个或多个用于实现规定的逻辑功能的可执行指令。在有些作为替换的实现中,方框中所标注的功能也可以以不同于附图中所标注的顺序发生。例如,两个连续的方框实际上可以基本并行地执行,它们有时也可以按相反的顺序执行,这依所涉及的功能而定。 也要注意的是,框图和/或流程图中的每个方框、以及框图和/或流程图中的方框的组合,可以用执行规定的功能或动作的专用的基于硬件的系统来实现,或者可以用专用硬件与计算机指令的组合来实现。对于本领域技术人员来说公知的是,通过硬件方式实现、通过软件方式实现以及通过软件和硬件结合的方式实现都是等价的。The flowchart and block diagrams in the Figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods, and computer program products according to various embodiments of the invention. In this regard, each block in the flowchart or block diagram can represent a module, a program segment, or a portion of an instruction that includes one or more components for implementing the specified logical functions. Executable instructions. In some alternative implementations, the functions noted in the blocks may also occur in a different order than those illustrated in the drawings. For example, two consecutive blocks may be executed substantially in parallel, and they may sometimes be executed in the reverse order, depending upon the functionality involved. It is also noted that each block of the block diagrams and/or flowcharts, and combinations of blocks in the block diagrams and/or flowcharts, can be implemented in a dedicated hardware-based system that performs the specified function or function. Or it can be implemented by a combination of dedicated hardware and computer instructions. It is well known to those skilled in the art that implementation by hardware, implementation by software, and implementation by a combination of software and hardware are equivalent.
以上已经描述了本发明的各实施例,上述说明是示例性的,并非穷尽性的,并且也不限于所披露的各实施例。在不偏离所说明的各实施例的范围和精神的情况下,对于本技术领域的普通技术人员来说许多修改和变更都是显而易见的。本文中所用术语的选择,旨在最好地解释各实施例的原理、实际应用或对市场中的技术的技术改进,或者使本技术领域的其它普通技术人员能理解本文披露的各实施例。本发明的范围由所附权利要求来限定。 The embodiments of the present invention have been described above, and the foregoing description is illustrative, not limiting, and not limited to the disclosed embodiments. Numerous modifications and changes will be apparent to those skilled in the art without departing from the scope of the invention. The choice of terms used herein is intended to best explain the principles, practical applications, or technical improvements of the techniques in the <RTIgt; </ RTI> <RTIgt; </ RTI> <RTIgt; </ RTI> <RTIgt; The scope of the invention is defined by the appended claims.

Claims (11)

  1. 一种用于虚拟现实设备的手势识别方法,所述虚拟现实设备包括深度摄像头,其特征在于,所述手势识别方法包括:A gesture recognition method for a virtual reality device, the virtual reality device comprising a depth camera, wherein the gesture recognition method comprises:
    控制所述深度摄像头采集当前用户的当前手部图像;Controlling the depth camera to capture a current hand image of the current user;
    根据所述当前手部图像判断所述当前用户是否执行敲击动作,如是,则:Determining, according to the current hand image, whether the current user performs a tap action, and if so, then:
    从所述当前手部图像中提取当前特征;Extracting a current feature from the current hand image;
    将所述当前特征与模型中的参考特征进行匹配,根据匹配结果确定所述当前用户敲击的按键。Matching the current feature with a reference feature in the model, and determining a button of the current user tap based on the matching result.
  2. 根据权利要求1所述的手势识别方法,其特征在于,所述手势识别方法还包括:The gesture recognition method according to claim 1, wherein the gesture recognition method further comprises:
    控制所述深度摄像头采集参考用户的参考手部图像;Controlling the depth camera to acquire a reference hand image of the reference user;
    从所述参考手部图像中提取所述参考特征,并将所述参考特征存储在所述模型中。Extracting the reference feature from the reference hand image and storing the reference feature in the model.
  3. 根据权利要求1或2所述的手势识别方法,所述虚拟现实设备还包括显示屏,其特征在于,所述根据所述当前图像判断所述当前用户是否敲击之前还包括:The gesture recognition method according to claim 1 or 2, wherein the virtual reality device further comprises a display screen, wherein the determining, before determining whether the current user is tapping according to the current image, further comprises:
    在所述显示屏上显示出键盘图像及所述当前用户的手指在所述键盘图像上的初始位置。A keyboard image and an initial position of the current user's finger on the keyboard image are displayed on the display screen.
  4. 根据权利要求1-3中任一项所述的手势识别方法,其特征在于,所述手势识别方法还包括:The gesture recognition method according to any one of claims 1 to 3, wherein the gesture recognition method further comprises:
    根据所述当前用户敲击的按键,在所述显示屏上显示出所述当前用户的手指在所述键盘图像上的当前位置。And displaying, according to the button pressed by the current user, a current position of the current user's finger on the keyboard image on the display screen.
  5. 一种用于虚拟现实设备的手势识别装置,其特征在于,包括:A gesture recognition apparatus for a virtual reality device, comprising:
    第一控制模块,用于控制所述深度摄像头采集当前用户的当前手部图像;a first control module, configured to control the depth camera to collect a current hand image of the current user;
    判断模块,用于根据所述当前手部图像判断所述当前用户是否执行敲击动作;a determining module, configured to determine, according to the current hand image, whether the current user performs a tapping action;
    当前特征提取模块,用于在所述判断模块的判断结果为是的情况下, 从所述当前手部图像中提取当前特征;以及,a current feature extraction module, configured to: when the determination result of the determination module is yes, Extracting current features from the current hand image; and,
    匹配模块,用于将所述当前特征与模型中的参考特征进行匹配,根据匹配结果确定所述当前用户敲击的按键。And a matching module, configured to match the current feature with a reference feature in the model, and determine, according to the matching result, the button that the current user taps.
  6. 根据权利要求5所述的手势识别装置,其特征在于,所述手势识别装置还包括:The gesture recognition apparatus according to claim 5, wherein the gesture recognition apparatus further comprises:
    第二控制模块,用于控制所述深度摄像头采集参考用户的参考手部图像;a second control module, configured to control the depth camera to collect a reference hand image of the reference user;
    参考特征提取模块,用于从所述参考手部图像中提取所述参考特征,并将所述参考特征存储在所述模型中。And a reference feature extraction module, configured to extract the reference feature from the reference hand image and store the reference feature in the model.
  7. 根据权利要求5或6所述的手势识别装置,所述虚拟现实设备还包括显示屏,其特征在于,所述手势识别装置还包括:The gesture recognition device according to claim 5 or 6, wherein the virtual reality device further comprises a display screen, wherein the gesture recognition device further comprises:
    第一显示模块,用于在所述显示屏上显示出键盘图像及所述当前用户的手指在所述键盘图像上的初始位置。And a first display module, configured to display, on the display screen, a keyboard image and an initial position of the current user's finger on the keyboard image.
  8. 根据权利要求5-7中任一项所述的手势识别装置,其特征在于,所述手势识别装置还包括:The gesture recognition apparatus according to any one of claims 5-7, wherein the gesture recognition apparatus further comprises:
    第二显示模块,用于根据所述匹配模块确定的所述当前用户敲击的按键,在所述显示屏上显示出所述当前用户的手指在所述键盘图像上的当前位置。And a second display module, configured to display, according to the button of the current user tap determined by the matching module, a current position of the current user's finger on the keyboard image on the display screen.
  9. 一种虚拟现实设备,其特征在于,包括根据权利要求5-8中任一项所述的手势识别装置。A virtual reality device, comprising the gesture recognition device according to any one of claims 5-8.
  10. 一种虚拟现实设备,其特征在于,包括深度摄像头、处理器和存储器,所述深度摄像头用于采集图像,所述存储器用于存储指令,所述指令用于控制所述处理器执行所述权利要求1-4任一项所述的手势识别方法。A virtual reality device, comprising: a depth camera for acquiring an image, the memory for storing an instruction, the instruction for controlling the processor to execute the right The gesture recognition method according to any one of claims 1 to 4.
  11. 一种计算机可读存储介质,其特征在于,存储有用于执行根据权利要求1-4中任一项所述手势识别方法的程序代码。 A computer readable storage medium characterized by storing program code for performing the gesture recognition method according to any one of claims 1-4.
PCT/CN2016/111062 2016-11-29 2016-12-20 Gesture recognition method and device for virtual reality apparatus, and virtual reality apparatus WO2018098861A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201611073934.3A CN106845335B (en) 2016-11-29 2016-11-29 Gesture recognition method and device for virtual reality equipment and virtual reality equipment
CN201611073934.3 2016-11-29

Publications (1)

Publication Number Publication Date
WO2018098861A1 true WO2018098861A1 (en) 2018-06-07

Family

ID=59145422

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2016/111062 WO2018098861A1 (en) 2016-11-29 2016-12-20 Gesture recognition method and device for virtual reality apparatus, and virtual reality apparatus

Country Status (2)

Country Link
CN (1) CN106845335B (en)
WO (1) WO2018098861A1 (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110096166A (en) * 2019-04-23 2019-08-06 广东工业大学华立学院 A kind of virtual input method
CN111158476A (en) * 2019-12-25 2020-05-15 中国人民解放军军事科学院国防科技创新研究院 Key identification method, system, equipment and storage medium of virtual keyboard
CN111443831A (en) * 2020-03-30 2020-07-24 北京嘉楠捷思信息技术有限公司 Gesture recognition method and device
CN113299132A (en) * 2021-06-08 2021-08-24 上海松鼠课堂人工智能科技有限公司 Student speech skill training method and system based on virtual reality scene

Families Citing this family (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107490983A (en) * 2017-09-29 2017-12-19 中国船舶重工集团公司第七〇四研究所 A kind of emulation mode for simulating parachute jumping full experience
CN107693117B (en) * 2017-09-29 2020-06-12 苏州蓝软智能医疗科技有限公司 Auxiliary operation system and method for automatically matching 3D model and operation patient in superposition mode
CN107644631A (en) * 2017-10-13 2018-01-30 深圳市明德智慧教育科技有限公司 Method, system and the virtual reality device of music input based on virtual reality
CN109857244B (en) * 2017-11-30 2023-09-01 百度在线网络技术(北京)有限公司 Gesture recognition method and device, terminal equipment, storage medium and VR glasses
CN108052277A (en) * 2017-12-14 2018-05-18 深圳市艾德互联网络有限公司 A kind of AR positioning learning methods and device
CN108519855A (en) * 2018-04-17 2018-09-11 北京小米移动软件有限公司 Characters input method and device
CN108815845B (en) * 2018-05-15 2019-11-26 百度在线网络技术(北京)有限公司 The information processing method and device of human-computer interaction, computer equipment and readable medium
CN109508635B (en) * 2018-10-08 2022-01-07 海南师范大学 Traffic light identification method based on TensorFlow combined with multilayer CNN network
CN109598998A (en) * 2018-11-30 2019-04-09 深圳供电局有限公司 Power grid training wearable device and its exchange method based on gesture identification
CN109933190B (en) * 2019-02-02 2022-07-19 青岛小鸟看看科技有限公司 Head-mounted display equipment and interaction method thereof
CN110321174A (en) * 2019-06-25 2019-10-11 Oppo广东移动通信有限公司 A kind of starting-up method and device, equipment, storage medium
CN111766947A (en) * 2020-06-30 2020-10-13 歌尔科技有限公司 Display method, display device, wearable device and medium
CN112462937B (en) * 2020-11-23 2022-11-08 青岛小鸟看看科技有限公司 Local perspective method and device of virtual reality equipment and virtual reality equipment
CN113269089B (en) * 2021-05-25 2023-07-18 上海人工智能研究院有限公司 Real-time gesture recognition method and system based on deep learning

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102778951A (en) * 2012-06-15 2012-11-14 惠州华阳通用电子有限公司 Input device and input method employing virtual key
JP2012252584A (en) * 2011-06-03 2012-12-20 Nakayo Telecommun Inc Virtual keyboard input method
CN103105930A (en) * 2013-01-16 2013-05-15 中国科学院自动化研究所 Non-contact type intelligent inputting method based on video images and device using the same
US20140029789A1 (en) * 2012-07-30 2014-01-30 Bruno Delean Method and system for vision based interfacing with a computer
CN104246682A (en) * 2012-03-26 2014-12-24 苹果公司 Enhanced virtual touchpad and touchscreen
CN104423578A (en) * 2013-08-25 2015-03-18 何安莉 Interactive Input System And Method
CN105224069A (en) * 2014-07-03 2016-01-06 王登高 The device of a kind of augmented reality dummy keyboard input method and use the method

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2011106797A1 (en) * 2010-02-28 2011-09-01 Osterhout Group, Inc. Projection triggering through an external marker in an augmented reality eyepiece
US10262462B2 (en) * 2014-04-18 2019-04-16 Magic Leap, Inc. Systems and methods for augmented and virtual reality
US9766806B2 (en) * 2014-07-15 2017-09-19 Microsoft Technology Licensing, Llc Holographic keyboard display
CN104636725B (en) * 2015-02-04 2017-09-29 华中科技大学 A kind of gesture identification method and system based on depth image

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2012252584A (en) * 2011-06-03 2012-12-20 Nakayo Telecommun Inc Virtual keyboard input method
CN104246682A (en) * 2012-03-26 2014-12-24 苹果公司 Enhanced virtual touchpad and touchscreen
CN102778951A (en) * 2012-06-15 2012-11-14 惠州华阳通用电子有限公司 Input device and input method employing virtual key
US20140029789A1 (en) * 2012-07-30 2014-01-30 Bruno Delean Method and system for vision based interfacing with a computer
CN103105930A (en) * 2013-01-16 2013-05-15 中国科学院自动化研究所 Non-contact type intelligent inputting method based on video images and device using the same
CN104423578A (en) * 2013-08-25 2015-03-18 何安莉 Interactive Input System And Method
CN105224069A (en) * 2014-07-03 2016-01-06 王登高 The device of a kind of augmented reality dummy keyboard input method and use the method

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110096166A (en) * 2019-04-23 2019-08-06 广东工业大学华立学院 A kind of virtual input method
CN111158476A (en) * 2019-12-25 2020-05-15 中国人民解放军军事科学院国防科技创新研究院 Key identification method, system, equipment and storage medium of virtual keyboard
CN111158476B (en) * 2019-12-25 2023-05-23 中国人民解放军军事科学院国防科技创新研究院 Key recognition method, system, equipment and storage medium of virtual keyboard
CN111443831A (en) * 2020-03-30 2020-07-24 北京嘉楠捷思信息技术有限公司 Gesture recognition method and device
CN113299132A (en) * 2021-06-08 2021-08-24 上海松鼠课堂人工智能科技有限公司 Student speech skill training method and system based on virtual reality scene

Also Published As

Publication number Publication date
CN106845335A (en) 2017-06-13
CN106845335B (en) 2020-03-17

Similar Documents

Publication Publication Date Title
WO2018098861A1 (en) Gesture recognition method and device for virtual reality apparatus, and virtual reality apparatus
CN109635621B (en) System and method for recognizing gestures based on deep learning in first-person perspective
US11093045B2 (en) Systems and methods to augment user interaction with the environment outside of a vehicle
CN107810465B (en) System and method for generating a drawing surface
US9811650B2 (en) User authentication system and method
US8823642B2 (en) Methods and systems for controlling devices using gestures and related 3D sensor
US10120454B2 (en) Gesture recognition control device
JP6013583B2 (en) Method for emphasizing effective interface elements
JP6090140B2 (en) Information processing apparatus, information processing method, and program
US20160054791A1 (en) Navigating augmented reality content with a watch
US20150185825A1 (en) Assigning a virtual user interface to a physical object
US9578399B2 (en) Remote sensor access and queuing
US20150187108A1 (en) Augmented reality content adapted to changes in real world space geometry
US10254847B2 (en) Device interaction with spatially aware gestures
US20150187137A1 (en) Physical object discovery
US11714540B2 (en) Remote touch detection enabled by peripheral device
US11709593B2 (en) Electronic apparatus for providing a virtual keyboard and controlling method thereof
US20220276698A1 (en) IMU for Touch Detection
KR20210033394A (en) Electronic apparatus and controlling method thereof
US20150185851A1 (en) Device Interaction with Self-Referential Gestures
Bhowmik Natural and intuitive user interfaces with perceptual computing technologies
US20240096043A1 (en) Display method, apparatus, electronic device and storage medium for a virtual input device
US20170003872A1 (en) Touch-encoded keyboard
KR20240025593A (en) Method and device for dynamically selecting an action modality for an object
CN114923418A (en) Point selection based measurement

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 16922696

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 16922696

Country of ref document: EP

Kind code of ref document: A1