CN107463873B - A real-time gesture analysis and evaluation method and system based on RGBD depth sensor - Google Patents

A real-time gesture analysis and evaluation method and system based on RGBD depth sensor Download PDF

Info

Publication number
CN107463873B
CN107463873B CN201710523575.5A CN201710523575A CN107463873B CN 107463873 B CN107463873 B CN 107463873B CN 201710523575 A CN201710523575 A CN 201710523575A CN 107463873 B CN107463873 B CN 107463873B
Authority
CN
China
Prior art keywords
palm
node
frame
initial image
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
CN201710523575.5A
Other languages
Chinese (zh)
Other versions
CN107463873A (en
Inventor
梁华刚
易生
孙凯
李怀德
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Changan University
Original Assignee
Changan University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Changan University filed Critical Changan University
Priority to CN201710523575.5A priority Critical patent/CN107463873B/en
Publication of CN107463873A publication Critical patent/CN107463873A/en
Application granted granted Critical
Publication of CN107463873B publication Critical patent/CN107463873B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Health & Medical Sciences (AREA)
  • Psychiatry (AREA)
  • Social Psychology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Image Analysis (AREA)

Abstract

本发明公开了一种基于RGBD深度传感器的实时手势分析与评价的方法与系统,包括火车驾驶员手掌的静态手势识别与评价系统和火车驾驶员手臂的动态手势识别与评价系统,火车驾驶员手掌的静态手势识别与评价系统包括确定手掌掌心位置模块、提取手掌区域图像模块、去噪模块、手势识别及评价模块;火车驾驶员手臂的动态手势识别与评价系统包括手臂骨骼节点运动序列提取模块、动态手势最优匹配模块、手臂动态手势评价模块。本发明对于环境背景、光照具有很强的鲁棒性,而且在手掌手势检测时采用基于掌心节点的手势像素搜索,提高了手掌手势的检测效果;实时监督驾驶员手势以确保火车行驶安全,还可以避免人为的对火车驾驶员手势监控,减轻人力资源消耗。

Figure 201710523575

The invention discloses a method and system for real-time gesture analysis and evaluation based on RGBD depth sensors, including a static gesture recognition and evaluation system for the palm of a train driver and a dynamic gesture recognition and evaluation system for the arm of the train driver. The static gesture recognition and evaluation system includes the palm position determination module, the palm area image extraction module, the denoising module, the gesture recognition and evaluation module; the dynamic gesture recognition and evaluation system of the train driver's arm includes the arm skeleton node motion sequence extraction module, Dynamic gesture optimal matching module, arm dynamic gesture evaluation module. The present invention has strong robustness to environmental background and lighting, and adopts gesture pixel search based on palm nodes in palm gesture detection, which improves the detection effect of palm gestures; supervises the driver's gestures in real time to ensure the safety of trains, and also It can avoid artificial gesture monitoring of train drivers and reduce the consumption of human resources.

Figure 201710523575

Description

一种基于RGBD深度传感器的实时手势分析与评价方法与系统A real-time gesture analysis and evaluation method and system based on RGBD depth sensor

技术领域technical field

本发明属于图像处理技术领域,具体涉及一种基于RGBD深度传感器的实时手势分析与评价方法与系统。The invention belongs to the technical field of image processing, and in particular relates to a real-time gesture analysis and evaluation method and system based on an RGBD depth sensor.

背景技术Background technique

手势识别技术作为未来人机交互系统的关键技术之一不仅具有重要的研究价值,而且具有广泛的应用前景。目前,传统的手势识别方法通常是对输入的二维图像进行手势检测并加以识别,然而这类方法对于输入的图像比较敏感,当背景简单、环境光照影响较小时手势检测与识别效果较好,然而当背景复杂、光照变化较大时手势检测与识别效果急剧下降,应用范围受到限制。近几年,为了克服传统手势识别方法的缺陷,三维图像传感设备越来越受到人们的青睐,这类设备不仅可以获取到RGB图像,而且还可以获取到图像的深度数据,从而可以避免复杂的环境背景、光照变化等对手势识别的影响。As one of the key technologies of future human-computer interaction systems, gesture recognition technology not only has important research value, but also has broad application prospects. At present, the traditional gesture recognition method usually detects and recognizes the gesture of the input two-dimensional image. However, this method is more sensitive to the input image, and the gesture detection and recognition effect is better when the background is simple and the influence of ambient light is small. However, when the background is complex and the illumination changes greatly, the effect of gesture detection and recognition drops sharply, and the application scope is limited. In recent years, in order to overcome the shortcomings of traditional gesture recognition methods, 3D image sensing devices have become more and more popular. Such devices can not only obtain RGB images, but also obtain depth data of images, so as to avoid complicated The influence of environmental background, lighting changes, etc. on gesture recognition.

目前,手势在交通领域应用广泛,例如火车驾驶员在火车行驶前、行驶中、车内仪器检测情况等需要用手势演示出来,以确保火车的安全行驶,防止事故的发生,交警则通过一系列的手势来保障道路交通安全畅通等。然而,由于火车驾驶员、交警等人员工作环境较为复杂,光线变化极大,传统方法无法对驾驶员手势有效地进行识别与评价,而且驾驶员标准手势众多,不仅包含手臂上的动态手势,还包括手掌区域的手势,进一步加大了手势识别的难度。传统的手势识别方法主要包括两个步骤,首先在二维图像上检测出手掌与手臂,其次是对检测出来的手势进行识别。一般说来,手势检测的好坏直接影响着手势识别的效果,通常背景环境和光照会对手势检测有影响,而且,在二维图像上检测出手掌与手臂也是一大难点,传统做法是训练大量手势样本获得分类器,然而,人手是复杂的变形体,手势具有多样性、多义性、时间上的差异等特点,很难训练得到理想的手势分类器,从而传统手势识别方法无法应用于驾驶员等手势识别当中。At present, gestures are widely used in the field of transportation. For example, the train driver needs to use gestures to demonstrate the situation before the train runs, while the train is running, and the detection of the in-vehicle instruments, so as to ensure the safe running of the train and prevent accidents. The traffic police use a series of gestures to ensure the safety and smooth flow of road traffic. However, due to the complex working environment of train drivers, traffic police and other personnel, and the great changes in light, traditional methods cannot effectively identify and evaluate driver gestures, and there are many standard driver gestures, including not only dynamic gestures on the arm, but also dynamic gestures on the arm. Including gestures in the palm area further increases the difficulty of gesture recognition. The traditional gesture recognition method mainly includes two steps, the first is to detect the palm and arm on the two-dimensional image, and the second is to recognize the detected gesture. Generally speaking, the quality of gesture detection directly affects the effect of gesture recognition. Usually the background environment and lighting will have an impact on gesture detection. Moreover, it is also difficult to detect palms and arms on two-dimensional images. The traditional method is to train a large number of Gesture samples are used to obtain a classifier. However, the human hand is a complex deformation body, and gestures have the characteristics of diversity, ambiguity, and time difference. It is difficult to train an ideal gesture classifier, so traditional gesture recognition methods cannot be applied to driving. During the gesture recognition of the staff and so on.

发明内容SUMMARY OF THE INVENTION

本发明的目的是提供一种基于RGBD深度传感器的实时手势分析与评价系统,解决了在复杂背景和光照条件下检测手掌与手臂的问题,并且能够对火车驾驶员、交警等工作人员的各种手势进行分析并评价,具有广阔的应用前景。The purpose of the present invention is to provide a real-time gesture analysis and evaluation system based on RGBD depth sensor, which solves the problem of detecting palms and arms under complex background and lighting conditions, and is able to detect various problems of train drivers, traffic police and other staff. Gestures are analyzed and evaluated, which has broad application prospects.

本发明所采用的技术方案是,The technical scheme adopted in the present invention is,

一种基于RGBD深度传感器的实时手掌手势分析与评价方法,包括以下步骤:A real-time palm gesture analysis and evaluation method based on RGBD depth sensor, comprising the following steps:

步骤1,利用RGBD传感器在一段时间内获取T帧初始图像,所述T帧初始图像中的每一帧初始图像均包括手掌节点、手腕节点和手肘节点,确定T帧初始图像的手掌节点坐标;Step 1, using an RGBD sensor to acquire an initial image of T frames within a period of time, each frame of initial image in the initial T frame image includes a palm node, a wrist node and an elbow node, and determine the palm node coordinates of the T frame initial image. ;

包括:include:

步骤11,从T帧初始图像中任选一帧初始图像作为当前帧初始图像,通过初始手掌 节点得到当前帧初始图像中手掌节点P的坐标

Figure GDA0001399495770000021
Step 11, select a frame of initial image from the initial image of T frames as the initial image of the current frame, and obtain the coordinates of the palm node P in the initial image of the current frame through the initial palm node
Figure GDA0001399495770000021

其中,M表示区域圆内白色像素点的个数,M为大于等于1的自然数,xi表示第i个像素点的横坐标,yi表示第i个像素点的纵坐标,zi表示第i个像素点到RGBD传感器之间的距离;Among them, M represents the number of white pixels in the area circle, M is a natural number greater than or equal to 1, xi represents the abscissa of the ith pixel, yi represents the ordinate of the ith pixel, and zi represents the ith pixel The distance between i pixels and the RGBD sensor;

所述区域圆为以初始手掌节点P1为圆心,以初始手掌节点到手腕节点之间的距离为半径的圆;The area circle is a circle whose center is the initial palm node P1, and the radius is the distance between the initial palm node and the wrist node;

所述手掌节点坐标P是在以RGBD传感器中心为原点,以水平方向为X轴,竖直方向为Y轴,传感器指向驾驶员方向为Z轴的坐标系下;The palm node coordinate P is in a coordinate system with the center of the RGBD sensor as the origin, the horizontal direction as the X axis, the vertical direction as the Y axis, and the direction of the sensor pointing to the driver as the Z axis;

步骤12,重复步骤11,直至T帧初始图像中的每一帧初始图像都被作为当前帧初始图像,得到T帧初始图像的手掌节点坐标;Step 12, repeating step 11, until each frame of the initial image in the T frame initial image is used as the current frame initial image, to obtain the palm node coordinates of the T frame initial image;

步骤2,根据T帧初始图像的手掌节点坐标,提取T帧初始图像的手掌区域图像,包括:Step 2, according to the palm node coordinates of the initial image of the T frame, extract the palm area image of the initial image of the T frame, including:

步骤21,从T帧初始图像中任选一帧初始图像作为当前初始图像,提取该当前初始图像的方法包括:Step 21, select one frame of initial image as the current initial image from the T frame initial images, and the method for extracting the current initial image includes:

以当前初始图像的手掌节点为中心,在宽度为W,高度为H的矩形区域内进行手掌像素点搜索,将满足式(2)的手掌像素点放入手掌像素点集合Sk中,即得到当前手掌区域图像;Taking the palm node of the current initial image as the center, search for palm pixels in a rectangular area with a width of W and a height of H, and put the palm pixels that satisfy the formula (2) into the palm pixel set S k , that is, we get Current palm area image;

Figure GDA0001399495770000031
Figure GDA0001399495770000031

式(2)中,k=1,2,...,W×H,k表示搜寻的次数,dp表示当当前初始图像的手掌节点到RGBD传感器之间的距离,gi表示第i个像素点,di表示矩形区域内第i个像素点到RGBD传感器之间的距离,abs(dp-di,j)表示当前初始图像的手掌节点与矩形区域内第i个像素点到RGBD传感器之间的距离之差的绝对值,threshold表示阈值,25≤threshold≤35,Sk表示第k次搜索到的手势像素点集合,Sk-1表示第k-1次搜索到的手势像素点集合;In formula (2), k=1,2,...,W×H, k represents the number of searches, d p represents the distance between the palm node of the current initial image and the RGBD sensor, and gi represents the i-th Pixel point, d i represents the distance from the i-th pixel in the rectangular area to the RGBD sensor, abs(d p -d i,j ) represents the palm node of the current initial image and the i-th pixel in the rectangular area to the RGBD sensor The absolute value of the difference between the distances between the sensors, threshold represents the threshold, 25≤threshold≤35, S k represents the set of gesture pixels searched for the kth time, and S k-1 represents the gesture pixels searched for the k-1th time point set;

Figure GDA0001399495770000041
其中(xw,yw)和(xe,ye)分别表示当前初始图像的手掌节点中手腕节点以及手肘节点的坐标;
Figure GDA0001399495770000041
where (x w , y w ) and (x e , y e ) represent the coordinates of the wrist node and the elbow node in the palm node of the current initial image, respectively;

步骤22,重复步骤22,直至提取到T帧初始图像的手掌区域图像;Step 22, repeat step 22, until the palm area image of the T frame initial image is extracted;

步骤3,对T帧初始图像的手掌区域图像中的每一帧手掌区域图像进行膨胀和腐蚀操作,得到T帧去噪后的手掌区域图像;Step 3, performing expansion and erosion operations on each frame of the palm area image in the palm area image of the T frame initial image, to obtain a denoised palm area image of the T frame;

步骤4,通过神经网络识别T帧去噪后的手掌区域图像中的手掌区域的手势,并通过式(1)得到识别出的手掌区域的手势的分数PpalmStep 4: Identify the gestures in the palm area in the denoised palm area image of the T frame through the neural network, and obtain the recognized gesture score P palm in the palm area by formula (1):

Figure GDA0001399495770000042
Figure GDA0001399495770000042

式(1)中,

Figure GDA0001399495770000046
为将T帧去噪后的手掌区域图像输入到神经网络中后神经网络第t帧的输出结果,T为去噪后的手掌区域图像的总帧数,round(·)表示取整数。In formula (1),
Figure GDA0001399495770000046
It is the output result of the t-th frame of the neural network after inputting the denoised palm region image of T frame into the neural network, where T is the total number of frames of the denoised palm region image, and round( ) represents an integer.

一种基于RGBD深度传感器的实时手臂手势分析与评价方法,包括以下步骤:A real-time arm gesture analysis and evaluation method based on RGBD depth sensor, comprising the following steps:

步骤1,利用RGBD传感器在一段时间内获取T帧初始图像,从T帧初始图像中任选一帧初始图像作为第t帧初始图像,提取第t帧初始图像的手臂骨骼节点运动序列;Step 1, use the RGBD sensor to obtain the initial image of T frames within a period of time, select a frame of initial image from the initial image of the T frame as the initial image of the t frame, and extract the arm bone node motion sequence of the initial image of the t frame;

包括:include:

所述第t帧初始图像包括初始手掌节点P1 t、手腕节点P2 t、手肘节点P3 t、肩节点P4 t和肩中心节点Ps t,通过式(3)得到节点Pn t到肩中心节点Ps t的距离Dsn tThe initial image of the t-th frame includes the initial palm node P 1 t , the wrist node P 2 t , the elbow node P 3 t , the shoulder node P 4 t and the shoulder center node P s t , and the node P n is obtained by formula (3). Distance D sn t from t to shoulder center node P s t :

Figure GDA0001399495770000045
Figure GDA0001399495770000045

式(3)中,n=1,2,3,4,t=1,2,...,T,Dsn t表示第t帧初始图像中节点Pn t到肩中心节点Ps t的距离,T为初始图像的总帧数,xn t,yn t,zn t分别表示第t帧初始图像中手掌节点、手腕节点、手肘节点、肩节点的坐标值;xs t,ys t,zs t表示第t帧初始图像中肩中心节点的坐标;In formula (3), n=1, 2, 3, 4, t=1, 2,..., T, D sn t represents the distance from node P n t to the shoulder center node P s t in the initial image of the t-th frame. distance, T is the total number of frames of the initial image, x n t , y n t , and z n t respectively represent the coordinate values of the palm node, wrist node, elbow node, and shoulder node in the initial image of the t-th frame; x s t , y s t , z s t represent the coordinates of the shoulder center node in the initial image of the t-th frame;

即可得到初始图像中手臂骨骼节点的运动序列Dsn=(Dsn 1,Dsn 2,...,Dsn t,...,Dsn T),所述手臂骨骼节点包括手掌节点、手腕节点、手肘节点、肩节点;The motion sequence D sn =(D sn 1 ,D sn 2 ,...,D sn t ,...,D sn T ) of the arm skeleton nodes in the initial image can be obtained, where the arm skeleton nodes include palm nodes, Wrist node, elbow node, shoulder node;

步骤2,在驾驶员标准动态手势样本的运动序列库中找到与初始图像中手臂骨骼节点的运动序列之间的对应点距离之和最小的驾驶员标准动态手势样本的运动序列;Step 2, find the motion sequence of the driver's standard dynamic gesture sample with the smallest sum of the corresponding point distances between the motion sequences of the arm skeleton nodes in the initial image in the motion sequence library of the driver's standard dynamic gesture samples;

步骤3,通过式(4)得到手臂动态手势的得分ParmStep 3, the score P arm of the arm dynamic gesture is obtained by formula (4):

Figure GDA0001399495770000051
Figure GDA0001399495770000051

式(4)中,α为标准手势序列样本之间的DTW距离平均值,

Figure GDA0001399495770000052
Da,Db表示驾驶员标准动态手势样本的运动序列库中的任一运动序列,a=1,2,...,N,b=1,2,...,N,a≠b,N为驾驶员标准动态手势样本的运动序列库中运动序列的总数。In formula (4), α is the average DTW distance between standard gesture sequence samples,
Figure GDA0001399495770000052
D a , D b represent any motion sequence in the motion sequence library of driver's standard dynamic gesture samples, a=1,2,...,N, b=1,2,...,N, a≠b , N is the total number of motion sequences in the motion sequence library of standard dynamic gesture samples of drivers.

一种基于RGBD深度传感器的实时手势分析与评价方法,包括权利要求1所述实时手掌手势分析与评价方法和权利要求2所述的实时手臂手势分析与评价方法。A real-time gesture analysis and evaluation method based on RGBD depth sensor, including the real-time palm gesture analysis and evaluation method of claim 1 and the real-time arm gesture analysis and evaluation method of claim 2.

一种基于RGBD深度传感器的实时手掌手势分析与评价系统,包括:A real-time palm gesture analysis and evaluation system based on RGBD depth sensor, including:

确定手掌掌心位置模块,用于利用RGBD传感器在一段时间内获取T帧初始图像,所述T帧初始图像中的每一帧初始图像均包括手掌节点、手腕节点和手肘节点,确定T帧初始图像的手掌节点坐标;The module for determining the palm position of the palm is used to obtain an initial T frame image by using the RGBD sensor within a period of time, and each initial image in the T frame initial image includes a palm node, a wrist node and an elbow node, and determines the initial T frame image. The palm node coordinates of the image;

包括:include:

步骤11,从T帧初始图像中任选一帧初始图像作为当前帧初始图像,通过初始手掌 节点得到当前帧初始图像中手掌节点P的坐标

Figure GDA0001399495770000061
Step 11, select a frame of initial image from the initial image of T frames as the initial image of the current frame, and obtain the coordinates of the palm node P in the initial image of the current frame through the initial palm node
Figure GDA0001399495770000061

其中,M表示区域圆内白色像素点的个数,M为大于等于1的自然数,xi表示第i个像素点的横坐标,yi表示第i个像素点的纵坐标,zi表示第i个像素点到RGBD传感器之间的距离;Among them, M represents the number of white pixels in the area circle, M is a natural number greater than or equal to 1, xi represents the abscissa of the ith pixel, yi represents the ordinate of the ith pixel, and zi represents the ith pixel The distance between i pixels and the RGBD sensor;

所述区域圆为以初始手掌节点P1为圆心,以初始手掌节点到手腕节点之间的距离为半径的圆;The area circle is a circle whose center is the initial palm node P1, and the radius is the distance between the initial palm node and the wrist node;

所述手掌节点坐标P是在以RGBD传感器中心为原点,以水平方向为X轴,竖直方向为Y轴,传感器指向驾驶员方向为Z轴的坐标系下;The palm node coordinate P is in a coordinate system with the center of the RGBD sensor as the origin, the horizontal direction as the X axis, the vertical direction as the Y axis, and the direction of the sensor pointing to the driver as the Z axis;

步骤12,重复步骤11,直至T帧初始图像中的每一帧初始图像都被作为当前帧初始图像,得到T帧初始图像的手掌节点坐标;Step 12, repeating step 11, until each frame of the initial image in the T frame initial image is used as the current frame initial image, to obtain the palm node coordinates of the T frame initial image;

提取手掌区域图像模块,用于根据T帧初始图像的手掌节点坐标,提取T帧初始图像的手掌区域图像:The palm area image extraction module is used to extract the palm area image of the T frame initial image according to the palm node coordinates of the T frame initial image:

包括:include:

步骤21,从T帧初始图像中任选一帧初始图像作为当前初始图像,提取该当前初始图像的方法包括:Step 21, select one frame of initial image as the current initial image from the T frame initial images, and the method for extracting the current initial image includes:

以当前初始图像的手掌节点为中心,在宽度为W,高度为H的矩形区域内进行手掌像素点搜索,将满足式(2)的手掌像素点放入手掌像素点集合Sk中,即得到当前手掌区域图像;Taking the palm node of the current initial image as the center, search for palm pixels in a rectangular area with a width of W and a height of H, and put the palm pixels that satisfy the formula (2) into the palm pixel set S k , that is, we get Current palm area image;

Figure GDA0001399495770000071
Figure GDA0001399495770000071

式(2)中,k=1,2,...,W×H,k表示搜寻的次数,dp表示当当前初始图像的手掌节点到RGBD传感器之间的距离,gi表示第i个像素点,di表示矩形区域内第i个像素点到RGBD传感器之间的距离,abs(dp-di,j)表示当前初始图像的手掌节点与矩形区域内第i个像素点到RGBD传感器之间的距离之差的绝对值,threshold表示阈值,25≤threshold≤35,Sk表示第k次搜索到的手势像素点集合,Sk-1表示第k-1次搜索到的手势像素点集合;In formula (2), k=1,2,...,W×H, k represents the number of searches, d p represents the distance between the palm node of the current initial image and the RGBD sensor, and gi represents the i-th Pixel point, d i represents the distance from the i-th pixel in the rectangular area to the RGBD sensor, abs(d p -d i,j ) represents the palm node of the current initial image and the i-th pixel in the rectangular area to the RGBD sensor The absolute value of the difference between the distances between the sensors, threshold represents the threshold, 25≤threshold≤35, S k represents the set of gesture pixels searched for the kth time, and S k-1 represents the gesture pixels searched for the k-1th time point set;

Figure GDA0001399495770000072
其中(xw,yw)和(xe,ye)分别表示当前初始图像的手掌节点中手腕节点以及手肘节点的坐标;
Figure GDA0001399495770000072
where (x w , y w ) and (x e , y e ) represent the coordinates of the wrist node and the elbow node in the palm node of the current initial image, respectively;

步骤22,重复步骤22,直至提取到T帧初始图像的手掌区域图像;Step 22, repeat step 22, until the palm area image of the T frame initial image is extracted;

去噪模块:用于对T帧初始图像的手掌区域图像中的每一帧手掌区域图像进行膨胀和腐蚀操作,得到T帧去噪后的手掌区域图像;Denoising module: It is used to dilate and corrode each palm area image in the palm area image of the T frame initial image, and obtain the palm area image after T frame denoising;

手势识别及评价模块:用于通过神经网络识别T帧去噪后的手掌区域图像中的手掌区域的手势,并通过式(1)得到识别出的手掌区域的手势的分数PpalmGesture recognition and evaluation module: used to recognize the gesture of the palm area in the palm area image after T frame denoising through the neural network, and obtain the recognized gesture score P palm of the palm area by formula (1):

Figure GDA0001399495770000073
Figure GDA0001399495770000073

式(1)中,

Figure GDA0001399495770000084
为将T帧去噪后的手掌区域图像输入到神经网络中后神经网络第t帧的输出结果,T为去噪后的手掌区域图像的总帧数,round(·)表示取整数。In formula (1),
Figure GDA0001399495770000084
It is the output result of the t-th frame of the neural network after inputting the denoised palm region image of T frame into the neural network, where T is the total number of frames of the denoised palm region image, and round( ) represents an integer.

一种基于RGBD深度传感器的实时手臂手势分析与评价系统,包括:A real-time arm gesture analysis and evaluation system based on RGBD depth sensor, including:

手臂骨骼节点运动序列提取模块,利用RGBD传感器在一段时间内获取T帧初始图像,从T帧初始图像中任选一帧初始图像作为第t帧初始图像,提取第t帧初始图像的手臂骨骼节点运动序列;The arm bone node motion sequence extraction module uses the RGBD sensor to obtain the T frame initial image within a period of time, selects one frame of the initial image from the T frame initial image as the t frame initial image, and extracts the arm bone node of the t frame initial image. motion sequence;

包括:include:

所述第t帧初始图像包括初始手掌节点P1 t、手腕节点P2 t、手肘节点P3 t、肩节点P4 t和肩中心节点Ps t,通过式(3)得到节点Pn t到肩中心节点Ps t的距离Dsn tThe initial image of the t-th frame includes the initial palm node P 1 t , the wrist node P 2 t , the elbow node P 3 t , the shoulder node P 4 t and the shoulder center node P s t , and the node P n is obtained by formula (3). Distance D sn t from t to shoulder center node P s t :

Figure GDA0001399495770000083
Figure GDA0001399495770000083

式(3)中,n=1,2,3,4,t=1,2,...,T,Dsn t表示第t帧初始图像中节点Pn t到肩中心节点Ps t的距离,T为初始图像的总帧数,xn t,yn t,zn t分别表示第t帧初始图像中手掌节点、手腕节点、手肘节点、肩节点的坐标值;xs t,ys t,zs t表示第t帧初始图像中肩中心节点的坐标;In formula (3), n=1, 2, 3, 4, t=1, 2,..., T, D sn t represents the distance from node P n t to the shoulder center node P s t in the initial image of the t-th frame. distance, T is the total number of frames of the initial image, x n t , y n t , and z n t respectively represent the coordinate values of the palm node, wrist node, elbow node, and shoulder node in the initial image of the t-th frame; x s t , y s t , z s t represent the coordinates of the shoulder center node in the initial image of the t-th frame;

即可得到初始图像中手臂骨骼节点的运动序列Dsn=(Dsn 1,Dsn 2,...,Dsn t,...,Dsn T),所述手臂骨骼节点包括手掌节点、手腕节点、手肘节点、肩节点;The motion sequence D sn =(D sn 1 ,D sn 2 ,...,D sn t ,...,D sn T ) of the arm skeleton nodes in the initial image can be obtained, where the arm skeleton nodes include palm nodes, Wrist node, elbow node, shoulder node;

动态手势最优匹配模块,用于在驾驶员标准动态手势样本的运动序列库中找到与初始图像中手臂骨骼节点的运动序列之间的对应点距离之和最小的驾驶员标准动态手势样本的运动序列;The dynamic gesture optimal matching module is used to find the motion of the driver's standard dynamic gesture sample with the smallest sum of the corresponding point distances between the motion sequences of the arm skeleton nodes in the initial image in the motion sequence library of the driver's standard dynamic gesture samples sequence;

手臂动态手势评价模块,用于通过式(4)得到手臂动态手势的得分ParmThe arm dynamic gesture evaluation module is used to obtain the arm dynamic gesture score P arm by formula (4):

Figure GDA0001399495770000091
Figure GDA0001399495770000091

式(4)中,α为标准手势序列样本之间的DTW距离平均值,

Figure GDA0001399495770000092
Da,Db表示驾驶员标准动态手势样本的运动序列库中的任一运动序列,a=1,2,...,N,b=1,2,...,N,a≠b,N为驾驶员标准动态手势样本的运动序列库中运动序列的总数。In formula (4), α is the average DTW distance between standard gesture sequence samples,
Figure GDA0001399495770000092
D a , D b represent any motion sequence in the motion sequence library of driver's standard dynamic gesture samples, a=1,2,...,N, b=1,2,...,N, a≠b , N is the total number of motion sequences in the motion sequence library of standard dynamic gesture samples of drivers.

一种基于RGBD深度传感器的实时手势分析与评价系统,包括权利要求4所述实时手掌手势分析与评价系统和权利要求5所述的实时手臂手势分析与评价系统。A real-time gesture analysis and evaluation system based on RGBD depth sensor, including the real-time palm gesture analysis and evaluation system of claim 4 and the real-time arm gesture analysis and evaluation system of claim 5.

本发明的有益效果是The beneficial effects of the present invention are

本发明采用RGBD深度传感器来获取二维图像的深度数据,通过相应算法能够很好的提取到手掌、手臂等关键数据,相比传统方法,本发明对于环境背景、光照具有很强的鲁棒性,而且在手掌手势检测时采用基于掌心节点的手势像素搜索,进一步提高了手掌手势的检测效果。其次,本发明能够同时识别驾驶员手掌手势与手臂动态手势,而且能够依据识别算法的输出结果对驾驶员手势进行规范性评价并给出手掌手势和手臂动态手势的得分,不仅可以实时监督驾驶员手势以确保火车行驶安全,还可以避免人为的对火车驾驶员手势监控,减轻人力资源消耗。The invention adopts the RGBD depth sensor to obtain the depth data of the two-dimensional image, and the key data such as the palm and the arm can be well extracted through the corresponding algorithm. Compared with the traditional method, the invention has strong robustness to the environmental background and illumination. , and the gesture pixel search based on the palm node is used in the palm gesture detection, which further improves the detection effect of the palm gesture. Secondly, the present invention can recognize the palm gesture of the driver and the dynamic arm gesture at the same time, and can perform normative evaluation on the driver's gesture according to the output result of the recognition algorithm and give the score of the palm gesture and the dynamic arm gesture, which can not only supervise the driver in real time Gestures can ensure the safety of trains, and can also avoid artificial monitoring of train drivers' gestures, reducing human resource consumption.

附图说明Description of drawings

图1是火车驾驶员手掌区域手势识别与评价流程图;Fig. 1 is the flow chart of gesture recognition and evaluation in the palm area of the train driver;

图2是火车驾驶员手掌与手臂各节点示意图;Fig. 2 is a schematic diagram of each node of the train driver's palm and arm;

图3是火车驾驶员手臂动态手势识别与评价流程图;Fig. 3 is the flow chart of dynamic gesture recognition and evaluation of the train driver's arm;

图4是本发明应用场景图。FIG. 4 is an application scenario diagram of the present invention.

具体实施方式Detailed ways

火车驾驶员手势通常包含手掌区域的手势以及手臂部分动态手势,下面分别对手掌区域手势和手臂动态手势识别过程分别进行详细介绍。The gesture of the train driver usually includes the gesture of the palm area and the dynamic gesture of the arm part. The following is a detailed introduction to the recognition process of the palm area gesture and the dynamic arm gesture.

实施例1Example 1

一种基于RGBD深度传感器的实时手掌手势分析与评价方法,其特征在于,包括以下步骤:A real-time palm gesture analysis and evaluation method based on RGBD depth sensor, is characterized in that, comprises the following steps:

步骤1,利用RGBD传感器在一段时间内获取T帧初始图像,所述T帧初始图像中的每一帧初始图像均包括手掌节点、手腕节点和手肘节点,确定T帧初始图像的手掌节点坐标;Step 1, using an RGBD sensor to acquire an initial image of T frames within a period of time, each frame of initial image in the initial T frame image includes a palm node, a wrist node and an elbow node, and determine the palm node coordinates of the T frame initial image. ;

包括:include:

步骤11,从T帧初始图像中任选一帧初始图像作为当前帧初始图像,通过初始手掌 节点得到当前帧初始图像中手掌节点P的坐标

Figure GDA0001399495770000101
Step 11, select a frame of initial image from the initial image of T frames as the initial image of the current frame, and obtain the coordinates of the palm node P in the initial image of the current frame through the initial palm node
Figure GDA0001399495770000101

其中,M表示区域圆内白色像素点的个数,M为大于等于1的自然数,xi表示第i个像素点的横坐标,yi表示第i个像素点的纵坐标,zi表示第i个像素点到RGBD传感器之间的距离;Among them, M represents the number of white pixels in the area circle, M is a natural number greater than or equal to 1, xi represents the abscissa of the ith pixel, yi represents the ordinate of the ith pixel, and zi represents the ith pixel The distance between i pixels and the RGBD sensor;

如图2,所述区域圆为以初始手掌节点P1为圆心,以初始手掌节点到手腕节点之间的距离为半径的圆;As shown in Figure 2, the area circle is a circle with the initial palm node P1 as the center and the distance between the initial palm node and the wrist node as the radius;

如图4,所述手掌节点坐标P是在以RGBD传感器中心为原点,以水平方向为X轴,竖直方向为Y轴,传感器指向驾驶员方向为Z轴的坐标系下;As shown in Figure 4, the palm node coordinate P is in the coordinate system with the center of the RGBD sensor as the origin, the horizontal direction as the X axis, the vertical direction as the Y axis, and the direction of the sensor pointing to the driver as the Z axis;

由于RGBD传感器跟踪人体骨骼节点时容易发生节点漂移现象,导致手掌与传感器之间的距离存在偏差,为了减小偏差,需对初始手掌节点进行校正来得到手掌节点P的准确坐标:Since the RGBD sensor is prone to node drift when tracking human skeleton nodes, resulting in a deviation in the distance between the palm and the sensor, in order to reduce the deviation, it is necessary to correct the initial palm node to obtain the accurate coordinates of the palm node P:

步骤12,重复步骤11,直至T帧初始图像中的每一帧初始图像都被作为当前帧初始图像,得到T帧初始图像的手掌节点坐标;Step 12, repeating step 11, until each frame of the initial image in the T frame initial image is used as the current frame initial image, to obtain the palm node coordinates of the T frame initial image;

步骤2,根据T帧初始图像的手掌节点坐标,提取T帧初始图像的手掌区域图像,包括:Step 2, according to the palm node coordinates of the initial image of the T frame, extract the palm area image of the initial image of the T frame, including:

步骤21,从T帧初始图像中任选一帧初始图像作为当前初始图像,提取该当前初始图像的方法包括:Step 21, select one frame of initial image as the current initial image from the T frame initial images, and the method for extracting the current initial image includes:

以当前初始图像的手掌节点为中心,在宽度为W,高度为H的矩形区域内进行手掌像素点搜索,将满足式(2)的手掌像素点放入手掌像素点集合Sk中,即得到当前手掌区域图像;Taking the palm node of the current initial image as the center, search for palm pixels in a rectangular area with a width of W and a height of H, and put the palm pixels that satisfy the formula (2) into the palm pixel set S k , that is, we get Current palm area image;

Figure GDA0001399495770000111
Figure GDA0001399495770000111

式(2)中,k=1,2,...,W×H,k表示搜寻的次数,dp表示当当前初始图像的手掌节点到RGBD传感器之间的距离,gi表示第i个像素点,di表示矩形区域内第i个像素点到RGBD传感器之间的距离,abs(dp-di,j)表示当前初始图像的手掌节点与矩形区域内第i个像素点到RGBD传感器之间的距离之差的绝对值,threshold表示阈值,25≤threshold≤35,Sk表示第k次搜索到的手势像素点集合,Sk-1表示第k-1次搜索到的手势像素点集合;In formula (2), k=1,2,...,W×H, k represents the number of searches, d p represents the distance between the palm node of the current initial image and the RGBD sensor, and gi represents the i-th Pixel point, d i represents the distance from the i-th pixel in the rectangular area to the RGBD sensor, abs(d p -d i,j ) represents the palm node of the current initial image and the i-th pixel in the rectangular area to the RGBD sensor The absolute value of the difference between the distances between the sensors, threshold represents the threshold, 25≤threshold≤35, S k represents the set of gesture pixels searched for the kth time, and S k-1 represents the gesture pixels searched for the k-1th time point set;

由于矩形搜索区域的W和H不能设置的过小,防止手势区域大小的变化导致手势检测的不完全,矩形搜寻区域的高、宽为:

Figure GDA0001399495770000121
其中(xw,yw)和(xe,ye)分别表示当前初始图像的手掌节点中手腕节点以及手肘节点的坐标;Since the W and H of the rectangular search area cannot be set too small to prevent the gesture detection from being incomplete due to the change of the size of the gesture area, the height and width of the rectangular search area are:
Figure GDA0001399495770000121
where (x w , y w ) and (x e , y e ) represent the coordinates of the wrist node and the elbow node in the palm node of the current initial image, respectively;

步骤22,重复步骤22,直至提取到T帧初始图像的手掌区域图像;Step 22, repeat step 22, until the palm area image of the T frame initial image is extracted;

步骤3,对T帧初始图像的手掌区域图像中的每一帧手掌区域图像进行膨胀和腐蚀操作,得到T帧去噪后的手掌区域图像;Step 3, performing expansion and erosion operations on each frame of the palm area image in the palm area image of the T frame initial image, to obtain a denoised palm area image of the T frame;

手掌区域图像通常包含一些噪声,这些噪声包括手势边缘的一些毛刺和图像内部的一些孔洞等,为了得到更加精确的手势图像,需要进行膨胀与腐蚀操作,膨胀可以去除二值手势图像边缘的一些毛刺以及分散的噪声点等,而腐蚀则可以填平图像内部的一些孔洞。The palm area image usually contains some noise, including some burrs on the edge of the gesture and some holes inside the image, etc. In order to obtain a more accurate gesture image, dilation and erosion operations are required. Dilation can remove some burrs on the edge of the binary gesture image. And scattered noise points, etc., while erosion can fill in some holes inside the image.

步骤4,通过神经网络识别T帧去噪后的手掌区域图像中的手掌区域的手势,并通过式(1)得到识别出的手掌区域的手势的分数PpalmStep 4: Identify the gestures in the palm area in the denoised palm area image of the T frame through the neural network, and obtain the recognized gesture score P palm in the palm area by formula (1):

Figure GDA0001399495770000122
Figure GDA0001399495770000122

式(1)中,为将T帧去噪后的手掌区域图像输入到神经网络中后神经网络第t帧的输出结果,T为去噪后的手掌区域图像的总帧数,round(·)表示取整数。In formula (1), It is the output result of the t-th frame of the neural network after inputting the denoised palm region image of T frame into the neural network, where T is the total number of frames of the denoised palm region image, and round( ) represents an integer.

实施例2Example 2

一种基于RGBD深度传感器的实时手臂手势分析与评价方法,如图3,包括以下步骤:A real-time arm gesture analysis and evaluation method based on RGBD depth sensor, as shown in Figure 3, includes the following steps:

步骤1,利用RGBD传感器在一段时间内获取T帧初始图像,从T帧初始图像中任选一帧初始图像作为第t帧初始图像,提取第t帧初始图像的手臂骨骼节点运动序列;Step 1, use the RGBD sensor to obtain the initial image of T frames within a period of time, select a frame of initial image from the initial image of the T frame as the initial image of the t frame, and extract the arm bone node motion sequence of the initial image of the t frame;

包括:include:

所述第t帧初始图像包括初始手掌节点P1 t、手腕节点P2 t、手肘节点P3 t、肩节点P4 t和肩中心节点Ps t,通过式(3)得到节点Pn t到肩中心节点Ps t的距离Dsn tThe initial image of the t-th frame includes the initial palm node P 1 t , the wrist node P 2 t , the elbow node P 3 t , the shoulder node P 4 t and the shoulder center node P s t , and the node P n is obtained by formula (3). Distance D sn t from t to shoulder center node P s t :

Figure GDA0001399495770000131
Figure GDA0001399495770000131

式(3)中,n=1,2,3,4,t=1,2,...,T,Dsn t表示第t帧初始图像中节点Pn t到肩中心节点Ps t的距离,T为初始图像的总帧数,xn t,yn t,zn t分别表示第t帧初始图像中手掌节点、手腕节点、手肘节点、肩节点的坐标值;xs t,ys t,zs t表示第t帧初始图像中肩中心节点的坐标;In formula (3), n=1, 2, 3, 4, t=1, 2,..., T, D sn t represents the distance from node P n t to the shoulder center node P s t in the initial image of the t-th frame. distance, T is the total number of frames of the initial image, x n t , y n t , and z n t respectively represent the coordinate values of the palm node, wrist node, elbow node, and shoulder node in the initial image of the t-th frame; x s t , y s t , z s t represent the coordinates of the shoulder center node in the initial image of the t-th frame;

即可得到初始图像中手臂骨骼节点的运动序列Dsn=(Dsn 1,Dsn 2,...,Dsn t,...,Dsn T),所述手臂骨骼节点包括手掌节点、手腕节点、手肘节点、肩节点;The motion sequence D sn =(D sn 1 ,D sn 2 ,...,D sn t ,...,D sn T ) of the arm skeleton nodes in the initial image can be obtained, where the arm skeleton nodes include palm nodes, Wrist node, elbow node, shoulder node;

步骤2,在驾驶员标准动态手势样本的运动序列库中找到与初始图像中手臂骨骼节点的运动序列之间的对应点距离之和最小的驾驶员标准动态手势样本的运动序列;Step 2, find the motion sequence of the driver's standard dynamic gesture sample with the smallest sum of the corresponding point distances between the motion sequences of the arm skeleton nodes in the initial image in the motion sequence library of the driver's standard dynamic gesture samples;

本实施例采用DTW算法来解决两条运动序列长度不同的问题,设驾驶员标准动态手势样本的运动序列Da=(Da 1,Da 2,...,Da T′),初始图像中手臂骨骼节点的运动序列Dsn=(Dsn 1,Dsn 2,...,Dsn T),设两条序列之间的点对关系为φ(k)=(φs(k),φa(k)),其中1≤φs(k)≤T,1≤φa(k)≤T′,max(T,T′)≤k≤T+T′,DTW算法目的是找出两条序列之间的最佳点对关系φ(k),使得对应点之间距离之和DTW(Da,Dsn)最小:In this embodiment, the DTW algorithm is used to solve the problem of different lengths of the two motion sequences. Assume that the motion sequence D a = (D a 1 , D a 2 , . . . , D a T′ ) of the standard dynamic gesture sample of the driver, the initial The motion sequence D sn = (D sn 1 , D sn 2 ,..., D sn T ) of the arm skeleton nodes in the image, and the point-to-point relationship between the two sequences is φ(k)=(φ s (k ), φ a (k)), where 1≤φ s (k)≤T, 1≤φ a (k)≤T′, max(T,T′)≤k≤T+T′, the purpose of the DTW algorithm is Find the best point-to-point relationship φ(k) between the two sequences such that the sum of the distances between the corresponding points DTW(D a , D sn ) is the smallest:

Figure GDA0001399495770000132
Figure GDA0001399495770000132

步骤3,通过式(4)得到手臂动态手势的得分ParmStep 3, the score P arm of the arm dynamic gesture is obtained by formula (4):

Figure GDA0001399495770000141
Figure GDA0001399495770000141

式(4)中,α为标准手势序列样本之间的DTW距离平均值,

Figure GDA0001399495770000142
Da,Db表示驾驶员标准动态手势样本的运动序列库中的任一运动序列,a=1,2,...,N,b=1,2,...,N,a≠b,N为驾驶员标准动态手势样本的运动序列库中运动序列的总数。In formula (4), α is the average DTW distance between standard gesture sequence samples,
Figure GDA0001399495770000142
D a , D b represent any motion sequence in the motion sequence library of driver's standard dynamic gesture samples, a=1,2,...,N, b=1,2,...,N, a≠b , N is the total number of motion sequences in the motion sequence library of standard dynamic gesture samples of drivers.

实施例3Example 3

本实施例在实施例1、2的基础上,提供了一种基于RGBD深度传感器的实时手势分析与评价方法,包括实施例1所提供的实时手掌手势分析与评价方法和实施例2所提供的实时手臂手势分析与评价方法。该实施例能够同时识别驾驶员手掌手势与手臂动态手势,而且能够依据识别算法的输出结果对驾驶员手势进行规范性评价并给出手掌手势和手臂动态手势的得分,不仅可以实时监督驾驶员手势以确保火车行驶安全,还可以避免人为的对火车驾驶员手势监控,减轻人力资源消耗。Based on Embodiments 1 and 2, this embodiment provides a real-time gesture analysis and evaluation method based on an RGBD depth sensor, including the real-time palm gesture analysis and evaluation method provided in Embodiment 1 and the method provided in Embodiment 2. A real-time arm gesture analysis and evaluation method. This embodiment can recognize the driver's palm gesture and arm dynamic gesture at the same time, and can perform normative evaluation on the driver's gesture according to the output result of the recognition algorithm, and give the score of the palm gesture and the arm dynamic gesture, which can not only monitor the driver's gesture in real time In order to ensure the safety of the train, it can also avoid artificial gesture monitoring of the train driver and reduce the consumption of human resources.

实施例4Example 4

本实施例提供了火车驾驶员手掌的静态手势识别与评价系统,如图1所示,包括:This embodiment provides a static gesture recognition and evaluation system for the palm of a train driver, as shown in Figure 1, including:

确定手掌掌心位置模块,用于利用RGBD传感器在一段时间内获取T帧初始图像,所述T帧初始图像中的每一帧初始图像均包括手掌节点、手腕节点和手肘节点,确定T帧初始图像的手掌节点坐标;The module for determining the palm position of the palm is used to obtain an initial T frame image by using the RGBD sensor within a period of time, and each initial image in the T frame initial image includes a palm node, a wrist node and an elbow node, and determines the initial T frame image. The palm node coordinates of the image;

包括:include:

步骤11,从T帧初始图像中任选一帧初始图像作为当前帧初始图像,通过初始手掌 节点得到当前帧初始图像中手掌节点P的坐标

Figure GDA0001399495770000151
Step 11, select a frame of initial image from the initial image of T frames as the initial image of the current frame, and obtain the coordinates of the palm node P in the initial image of the current frame through the initial palm node
Figure GDA0001399495770000151

其中,M表示区域圆内白色像素点的个数,M为大于等于1的自然数,xi表示第i个像素点的横坐标,yi表示第i个像素点的纵坐标,zi表示第i个像素点到RGBD传感器之间的距离;Among them, M represents the number of white pixels in the area circle, M is a natural number greater than or equal to 1, xi represents the abscissa of the ith pixel, yi represents the ordinate of the ith pixel, and zi represents the ith pixel The distance between i pixels and the RGBD sensor;

如图2,所述区域圆为以初始手掌节点P1为圆心,以初始手掌节点到手腕节点之间的距离为半径的圆;As shown in Figure 2, the area circle is a circle with the initial palm node P1 as the center and the distance between the initial palm node and the wrist node as the radius;

如图4,所述手掌节点坐标P是在以RGBD传感器中心为原点,以水平方向为X轴,竖直方向为Y轴,传感器指向驾驶员方向为Z轴的坐标系下;As shown in Figure 4, the palm node coordinate P is in the coordinate system with the center of the RGBD sensor as the origin, the horizontal direction as the X axis, the vertical direction as the Y axis, and the direction of the sensor pointing to the driver as the Z axis;

由于RGBD传感器跟踪人体骨骼节点时容易发生节点漂移现象,导致手掌与传感器之间的距离存在偏差,为了减小偏差,需对初始手掌节点进行校正来得到手掌节点P的准确坐标:Since the RGBD sensor is prone to node drift when tracking human skeleton nodes, resulting in a deviation in the distance between the palm and the sensor, in order to reduce the deviation, it is necessary to correct the initial palm node to obtain the accurate coordinates of the palm node P:

提取手掌区域图像模块,提取手掌区域图像模块,用于根据T帧初始图像的手掌节点坐标,提取T帧初始图像的手掌区域图像:The palm area image module is extracted, and the palm area image module is extracted, which is used to extract the palm area image of the T frame initial image according to the palm node coordinates of the T frame initial image:

包括:include:

步骤21,从T帧初始图像中任选一帧初始图像作为当前初始图像,提取该当前初始图像的方法包括:Step 21, select one frame of initial image as the current initial image from the T frame initial images, and the method for extracting the current initial image includes:

以当前初始图像的手掌节点为中心,在宽度为W,高度为H的矩形区域内进行手掌像素点搜索,将满足式(2)的手掌像素点放入手掌像素点集合Sk中,即得到当前手掌区域图像;Taking the palm node of the current initial image as the center, search for palm pixels in a rectangular area with a width of W and a height of H, and put the palm pixels that satisfy the formula (2) into the palm pixel set S k , that is, we get Current palm area image;

Figure GDA0001399495770000161
Figure GDA0001399495770000161

式(2)中,k=1,2,...,W×H,k表示搜寻的次数,dp表示当当前初始图像的手掌节点到RGBD传感器之间的距离,gi表示第i个像素点,di表示矩形区域内第i个像素点到RGBD传感器之间的距离,abs(dp-di,j)表示当前初始图像的手掌节点与矩形区域内第i个像素点到RGBD传感器之间的距离之差的绝对值,threshold表示阈值,25≤threshold≤35,Sk表示第k次搜索到的手势像素点集合,Sk-1表示第k-1次搜索到的手势像素点集合;In formula (2), k=1,2,...,W×H, k represents the number of searches, d p represents the distance between the palm node of the current initial image and the RGBD sensor, and gi represents the i-th Pixel point, d i represents the distance from the i-th pixel in the rectangular area to the RGBD sensor, abs(d p -d i,j ) represents the palm node of the current initial image and the i-th pixel in the rectangular area to the RGBD sensor The absolute value of the difference between the distances between the sensors, threshold represents the threshold, 25≤threshold≤35, S k represents the set of gesture pixels searched for the kth time, and S k-1 represents the gesture pixels searched for the k-1th time point set;

由于矩形搜索区域的W和H不能设置的过小,防止手势区域大小的变化导致手势检测的不完全,矩形搜寻区域的高、宽为:其中(xw,yw)和(xe,ye)分别表示当前初始图像的手掌节点中手腕节点以及手肘节点的坐标;Since the W and H of the rectangular search area cannot be set too small to prevent the gesture detection from being incomplete due to the change of the size of the gesture area, the height and width of the rectangular search area are: where (x w , y w ) and (x e , y e ) represent the coordinates of the wrist node and the elbow node in the palm node of the current initial image, respectively;

步骤22,重复步骤22,直至提取到T帧初始图像的手掌区域图像;Step 22, repeat step 22, until the palm area image of the T frame initial image is extracted;

去噪模块:用于对T帧初始图像的手掌区域图像中的每一帧手掌区域图像进行膨胀和腐蚀操作,得到T帧去噪后的手掌区域图像;Denoising module: It is used to dilate and corrode each palm area image in the palm area image of the T frame initial image, and obtain the palm area image after T frame denoising;

手掌区域图像通常包含一些噪声,这些噪声包括手势边缘的一些毛刺和图像内部的一些孔洞等,为了得到更加精确的手势图像,需要进行膨胀与腐蚀操作,膨胀可以去除二值手势图像边缘的一些毛刺以及分散的噪声点等,而腐蚀则可以填平图像内部的一些孔洞。The palm area image usually contains some noise, including some burrs on the edge of the gesture and some holes inside the image, etc. In order to obtain a more accurate gesture image, dilation and erosion operations are required. Dilation can remove some burrs on the edge of the binary gesture image. And scattered noise points, etc., while erosion can fill in some holes inside the image.

手势识别及评价模块:用于通过神经网络识别T帧去噪后的手掌区域图像中的手掌区域的手势,并通过式(1)得到识别出的手掌区域的手势的分数PpalmGesture recognition and evaluation module: used to recognize the gesture of the palm area in the palm area image after T frame denoising through the neural network, and obtain the recognized gesture score P palm of the palm area by formula (1):

Figure GDA0001399495770000171
Figure GDA0001399495770000171

式(1)中,

Figure GDA0001399495770000175
为将T帧去噪后的手掌区域图像输入到神经网络中后神经网络第t帧的输出结果,T为去噪后的手掌区域图像的总帧数,round(·)表示取整数。In formula (1),
Figure GDA0001399495770000175
It is the output result of the t-th frame of the neural network after inputting the denoised palm region image of T frame into the neural network, where T is the total number of frames of the denoised palm region image, and round( ) represents an integer.

实施例5Example 5

本实施例提供了火车驾驶员手臂的动态手势识别与评价系统,如图3,包括:This embodiment provides a dynamic gesture recognition and evaluation system for the train driver's arm, as shown in Figure 3, including:

手臂骨骼节点运动序列提取模块,利用RGBD传感器在一段时间内获取T帧初始图像,从T帧初始图像中任选一帧初始图像作为第t帧初始图像,提取第t帧初始图像的手臂骨骼节点运动序列;The arm bone node motion sequence extraction module uses the RGBD sensor to obtain the T frame initial image within a period of time, selects one frame of the initial image from the T frame initial image as the t frame initial image, and extracts the arm bone node of the t frame initial image. motion sequence;

包括:include:

所述第t帧初始图像包括初始手掌节点P1 t、手腕节点P2 t、手肘节点P3 t、肩节点P4 t和肩中心节点Ps t,通过式(3)得到节点Pn t到肩中心节点Ps t的距离Dsn tThe initial image of the t-th frame includes the initial palm node P 1 t , the wrist node P 2 t , the elbow node P 3 t , the shoulder node P 4 t and the shoulder center node P s t , and the node P n is obtained by formula (3). Distance D sn t from t to shoulder center node P s t :

Figure GDA0001399495770000174
Figure GDA0001399495770000174

式(3)中,n=1,2,3,4,t=1,2,...,T,Dsn t表示第t帧初始图像中节点Pn t到肩中心节点Ps t的距离,T为初始图像的总帧数,xn t,yn t,zn t分别表示第t帧初始图像中手掌节点、手腕节点、手肘节点、肩节点的坐标值;xs t,ys t,zs t表示第t帧初始图像中肩中心节点的坐标;In formula (3), n=1, 2, 3, 4, t=1, 2,..., T, D sn t represents the distance from node P n t to the shoulder center node P s t in the initial image of the t-th frame. distance, T is the total number of frames of the initial image, x n t , y n t , and z n t respectively represent the coordinate values of the palm node, wrist node, elbow node, and shoulder node in the initial image of the t-th frame; x s t , y s t , z s t represent the coordinates of the shoulder center node in the initial image of the t-th frame;

即可得到初始图像中手臂骨骼节点的运动序列Dsn=(Dsn 1,Dsn 2,...,Dsn t,...,Dsn T),所述手臂骨骼节点包括手掌节点、手腕节点、手肘节点、肩节点;The motion sequence D sn =(D sn 1 ,D sn 2 ,...,D sn t ,...,D sn T ) of the arm skeleton nodes in the initial image can be obtained, where the arm skeleton nodes include palm nodes, Wrist node, elbow node, shoulder node;

动态手势最优匹配模块,用于在驾驶员标准动态手势样本的运动序列库中找到与初始图像中手臂骨骼节点的运动序列之间的对应点距离之和最小的驾驶员标准动态手势样本的运动序列;The dynamic gesture optimal matching module is used to find the motion of the driver's standard dynamic gesture sample with the smallest sum of the corresponding point distances between the motion sequences of the arm skeleton nodes in the initial image in the motion sequence library of the driver's standard dynamic gesture samples sequence;

本实施例采用DTW算法来解决两条运动序列长度不同的问题,设驾驶员标准动态手势样本的运动序列Da=(Da 1,Da 2,...,Da T′),初始图像中手臂骨骼节点的运动序列Dsn=(Dsn 1,Dsn 2,...,Dsn T),设两条序列之间的点对关系为φ(k)=(φs(k),φa(k)),其中1≤φs(k)≤T,1≤φa(k)≤T′,max(T,T′)≤k≤T+T′,DTW算法目的是找出两条序列之间的最佳点对关系φ(k),使得对应点之间距离之和DTW(Da,Dsn)最小:In this embodiment, the DTW algorithm is used to solve the problem of different lengths of the two motion sequences. Assume that the motion sequence D a = (D a 1 , D a 2 , . . . , D a T′ ) of the standard dynamic gesture sample of the driver, the initial The motion sequence D sn = (D sn 1 , D sn 2 ,..., D sn T ) of the arm skeleton nodes in the image, and the point-to-point relationship between the two sequences is φ(k)=(φ s (k ), φ a (k)), where 1≤φ s (k)≤T, 1≤φ a (k)≤T′, max(T,T′)≤k≤T+T′, the purpose of the DTW algorithm is Find the best point-to-point relationship φ(k) between the two sequences such that the sum of the distances between the corresponding points DTW(D a , D sn ) is the smallest:

手臂动态手势评价模块,用于通过式(4)得到手臂动态手势的得分ParmThe arm dynamic gesture evaluation module is used to obtain the arm dynamic gesture score P arm by formula (4):

Figure GDA0001399495770000182
Figure GDA0001399495770000182

式(4)中,α为标准手势序列样本之间的DTW距离平均值,

Figure GDA0001399495770000183
Da,Db表示驾驶员标准动态手势样本的运动序列库中的任一运动序列,a=1,2,...,N,b=1,2,...,N,a≠b,N为驾驶员标准动态手势样本的运动序列库中运动序列的总数;Dsn=(Dsn 1,Dsn 2,...,Dsn t,...,Dsn T),Dsn为初始图像中手臂骨骼节点的运动序列。In formula (4), α is the average DTW distance between standard gesture sequence samples,
Figure GDA0001399495770000183
D a , D b represent any motion sequence in the motion sequence library of driver's standard dynamic gesture samples, a=1,2,...,N, b=1,2,...,N, a≠b , N is the total number of motion sequences in the motion sequence library of the driver's standard dynamic gesture samples; D sn =(D sn 1 ,D sn 2 ,...,D sn t ,...,D sn T ), D sn is the motion sequence of the arm bone node in the initial image.

实施例6Example 6

本实施例在实施例4、5的基础上,提供了一种基于RGBD深度传感器的实时手势分析与评价系统,包括实施例4所提供的实时手掌手势分析与评价系统和实施例5所提供的实时手臂手势分析与评价系统。该实施例能够同时识别驾驶员手掌手势与手臂动态手势,而且能够依据识别算法的输出结果对驾驶员手势进行规范性评价并给出手掌手势和手臂动态手势的得分,不仅可以实时监督驾驶员手势以确保火车行驶安全,还可以避免人为的对火车驾驶员手势监控,减轻人力资源消耗。Based on Embodiments 4 and 5, this embodiment provides a real-time gesture analysis and evaluation system based on an RGBD depth sensor, including the real-time palm gesture analysis and evaluation system provided in Embodiment 4 and the system provided in Embodiment 5. Real-time arm gesture analysis and evaluation system. This embodiment can recognize the driver's palm gesture and arm dynamic gesture at the same time, and can perform normative evaluation on the driver's gesture according to the output result of the recognition algorithm, and give the score of the palm gesture and the arm dynamic gesture, which can not only monitor the driver's gesture in real time In order to ensure the safety of the train, it can also avoid artificial gesture monitoring of the train driver and reduce the consumption of human resources.

Claims (4)

1. A real-time palm gesture analysis and evaluation method based on an RGBD depth sensor is characterized by comprising the following steps:
step 1, acquiring T frames of initial images within a period of time by using an RGBD sensor, wherein each frame of initial image in the T frames of initial images comprises a palm node, a wrist node and an elbow node, and determining the coordinates of the palm node of the T frames of initial images;
the method comprises the following steps:
step 11, selecting one frame of initial image from the T frame of initial image as the current frame of initial image, and obtaining the coordinates of the palm node P in the current frame of initial image through the initial palm node
Figure FDA0002221045120000011
Wherein M represents the number of white pixel points in the region circle, M is a natural number greater than or equal to 1, and xiAbscissa, y, representing the ith pixeliThe ordinate, z, representing the ith pixeliRepresenting the distance from the ith pixel point to the RGBD sensor;
the region circle is an initial palm node P1A circle with the distance between the initial palm node and the wrist node as the radius and the center of the circle;
the palm node coordinate P is in a coordinate system which takes the center of the RGBD sensor as an origin, takes the horizontal direction as an X axis, takes the vertical direction as a Y axis, and takes the direction of the sensor pointing to the driver as a Z axis;
step 12, repeating step 11 until each frame initial image in the T frame initial images is used as a current frame initial image, and obtaining palm node coordinates of the T frame initial images;
step 2, extracting a palm region image of the T frame initial image according to the palm node coordinates of the T frame initial image, wherein the method comprises the following steps:
step 21, selecting one frame of initial image from the T frames of initial images as a current initial image, and the method for extracting the current initial image includes:
taking a palm node of the current initial image as a center, searching for palm pixel points in a rectangular region with the width of W and the height of H, and putting the palm pixel points satisfying the formula (2) into a palm pixel point set SkObtaining a current palm area image;
Figure FDA0002221045120000021
in formula (2), k is 1,2Indicating the number of searches, dpRepresents the distance, g, between the palm node of the current initial image and the RGBD sensoriRepresents the ith pixel point, diRepresenting the distance between the ith pixel point in the rectangular area and the RGBD sensor, abs (d)p-di) Representing the absolute value of the difference between the distance from the palm node of the current initial image and the ith pixel point in the rectangular region to the RGBD sensor, wherein threshold represents a threshold value, is more than or equal to 25 and less than or equal to 35, and SkRepresenting the gesture pixel point set searched for the k time, Sk-1Representing the gesture pixel point set searched at the k-1 st time;
Figure FDA0002221045120000022
wherein (x)w,yw) And (x)e,ye) Respectively representing the coordinates of a wrist node and an elbow node in a palm node of the current initial image;
step 22, repeating the step 22 until a palm area image of the T-frame initial image is extracted;
step 3, performing expansion and corrosion operations on each frame of palm area image in the palm area image of the T frame initial image to obtain a T frame denoised palm area image;
step 4, recognizing the gesture of the palm area in the T-frame de-noised palm area image through a neural network, and obtaining the fraction P of the recognized gesture of the palm area through the formula (1)palm
Figure FDA0002221045120000023
In the formula (1), the reaction mixture is,
Figure FDA0002221045120000031
Figure FDA0002221045120000032
inputting the denoised palm region image of the T frame into the output result of the T frame of the neural network, wherein T is the denoised output result of the T frame of the neural networkThe total frame number of the palm region image of (1), round (·) represents an integer.
2. A real-time gesture analysis and evaluation method based on an RGBD depth sensor is characterized by comprising the real-time palm gesture analysis and evaluation method and the real-time arm gesture analysis and evaluation method of claim 1;
the real-time arm gesture analysis and evaluation method comprises the following steps:
step 1, acquiring a T frame initial image within a period of time by using an RGBD sensor, selecting one frame initial image from the T frame initial image as a T frame initial image, and extracting an arm skeleton node motion sequence of the T frame initial image;
the method comprises the following steps:
the t-th frame initial image comprises an initial palm node P1 tWrist node P2 tElbow joint P3 tShoulder node P4 tAnd shoulder center node Ps tObtaining a node P by the formula (3)n tTo shoulder center node Ps tDistance D ofsn t
Figure FDA0002221045120000033
In formula (3), n is 1,2,3,4, T is 1,2sn tIndicating the node P in the initial image of the t-th framen tTo shoulder center node Ps tT is the total frame number of the initial image, xn t,yn t,zn tRespectively representing coordinate values of a palm node, a wrist node, an elbow node and a shoulder node in the t-th frame initial image; x is the number ofs t,ys t,zs tRepresenting the coordinates of the shoulder center node in the t-th frame initial image;
obtaining the motion sequence D of the arm skeleton node in the initial imagesn=(Dsn 1,Dsn 2,...,Dsn t,...,Dsn T) The arm skeleton nodes comprise palm nodes, wrist nodes, elbow nodes and shoulder nodes;
step 2, finding the motion sequence of the driver standard dynamic gesture sample with the minimum sum of the distances of corresponding points between the motion sequence of the arm skeleton nodes in the initial image and the motion sequence of the driver standard dynamic gesture sample in the motion sequence library of the driver standard dynamic gesture sample;
step 3, obtaining the score P of the dynamic gesture of the arm through the formula (4)arm
Figure FDA0002221045120000041
In equation (4), α is the average of the DTW distances between the standard gesture sequence samples,
Figure FDA0002221045120000042
Da,Dbrepresenting any motion sequence in the motion sequence library of the driver standard dynamic gesture sample, wherein a is 1,2, and N, b is 1, 2.
3. A real-time palm gesture analysis and evaluation system based on an RGBD depth sensor is characterized by comprising:
the device comprises a palm center position determining module, a palm image acquiring module and a palm image acquiring module, wherein the palm center position determining module is used for acquiring T frames of initial images in a period of time by using an RGBD sensor, each frame of initial image in the T frames of initial images comprises a palm node, a wrist node and an elbow node, and the palm node coordinates of the T frames of initial images are determined;
the method comprises the following steps:
step 11, selecting one frame of initial image from the T frame of initial image as the current frame of initial image, and obtaining the coordinates of the palm node P in the current frame of initial image through the initial palm node
Figure FDA0002221045120000043
Wherein M represents the number of white pixel points in the region circle, M is a natural number greater than or equal to 1, and xiAbscissa, y, representing the ith pixeliThe ordinate, z, representing the ith pixeliRepresenting the distance from the ith pixel point to the RGBD sensor;
the region circle is an initial palm node P1A circle with the distance between the initial palm node and the wrist node as the radius and the center of the circle;
the palm node coordinate P is in a coordinate system which takes the center of the RGBD sensor as an origin, takes the horizontal direction as an X axis, takes the vertical direction as a Y axis, and takes the direction of the sensor pointing to the driver as a Z axis;
step 12, repeating step 11 until each frame initial image in the T frame initial images is used as a current frame initial image, and obtaining palm node coordinates of the T frame initial images;
the palm region image extracting module is used for extracting a palm region image of the T frame initial image according to the palm node coordinates of the T frame initial image:
the method comprises the following steps:
step 21, selecting one frame of initial image from the T frames of initial images as a current initial image, and the method for extracting the current initial image includes:
taking a palm node of the current initial image as a center, searching for palm pixel points in a rectangular region with the width of W and the height of H, and putting the palm pixel points satisfying the formula (2) into a palm pixel point set SkObtaining a current palm area image;
Figure FDA0002221045120000051
in formula (2), k is 1,2pRepresents the distance, g, between the palm node of the current initial image and the RGBD sensoriRepresents the ith pixel point, diRepresenting the distance between the ith pixel point in the rectangular area and the RGBD sensor, abs (d)p-di) Representing the palm node of the current initial image and the ith image in the rectangular areaThe absolute value of the difference between the distances from the prime point to the RGBD sensor, where threshold represents a threshold, 25 ≦ threshold ≦ 35, SkRepresenting the gesture pixel point set searched for the k time, Sk-1Representing the gesture pixel point set searched at the k-1 st time;
Figure FDA0002221045120000061
wherein (x)w,yw) And (x)e,ye) Respectively representing the coordinates of a wrist node and an elbow node in a palm node of the current initial image;
step 22, repeating the step 22 until a palm area image of the T-frame initial image is extracted;
a denoising module: the image processing device is used for performing expansion and corrosion operations on each frame of palm area image in the palm area image of the T frame initial image to obtain a T frame denoised palm area image;
gesture recognition and evaluation module: the method is used for recognizing the gesture of the palm region in the T-frame de-noised palm region image through a neural network and obtaining the fraction P of the recognized gesture of the palm region through the formula (1)palm
Figure FDA0002221045120000062
In the formula (1), the reaction mixture is,
Figure FDA0002221045120000063
and (3) inputting the denoised palm region image of the T frame into an output result of the T frame of the neural network, wherein T is the total frame number of the denoised palm region image, and round (·) represents an integer.
4. A real-time gesture analysis and evaluation system based on an RGBD depth sensor, comprising the real-time palm gesture analysis and evaluation system and the real-time arm gesture analysis and evaluation system of claim 3;
the real-time arm gesture analysis and evaluation system comprises:
the arm skeleton node motion sequence extraction module is used for acquiring T frame initial images within a period of time by using an RGBD sensor, selecting one frame initial image from the T frame initial images as a T frame initial image, and extracting an arm skeleton node motion sequence of the T frame initial image;
the method comprises the following steps:
the t-th frame initial image comprises an initial palm node P1 tWrist node P2 tElbow joint P3 tShoulder node P4 tAnd shoulder center node Ps tObtaining a node P by the formula (3)n tTo shoulder center node Ps tDistance D ofsn t
In formula (3), n is 1,2,3,4, T is 1,2sn tIndicating the node P in the initial image of the t-th framen tTo shoulder center node Ps tT is the total frame number of the initial image, xn t,yn t,zn tRespectively representing coordinate values of a palm node, a wrist node, an elbow node and a shoulder node in the t-th frame initial image; x is the number ofs t,ys t,zs tRepresenting the coordinates of the shoulder center node in the t-th frame initial image;
obtaining the motion sequence D of the arm skeleton node in the initial imagesn=(Dsn 1,Dsn 2,...,Dsn t,...,Dsn T) The arm skeleton nodes comprise palm nodes, wrist nodes, elbow nodes and shoulder nodes;
the dynamic gesture optimal matching module is used for finding the motion sequence of the driver standard dynamic gesture sample with the minimum sum of corresponding point distances between the motion sequence and the arm skeleton nodes in the initial image in the motion sequence library of the driver standard dynamic gesture sample;
an arm dynamic gesture evaluation module for obtaining the score P of the arm dynamic gesture through the formula (4)arm
Figure FDA0002221045120000072
In equation (4), α is the average of the DTW distances between the standard gesture sequence samples,
Figure FDA0002221045120000073
Da,Dbrepresenting any motion sequence in the motion sequence library of the driver standard dynamic gesture sample, wherein a is 1,2, and N, b is 1, 2.
CN201710523575.5A 2017-06-30 2017-06-30 A real-time gesture analysis and evaluation method and system based on RGBD depth sensor Expired - Fee Related CN107463873B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710523575.5A CN107463873B (en) 2017-06-30 2017-06-30 A real-time gesture analysis and evaluation method and system based on RGBD depth sensor

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710523575.5A CN107463873B (en) 2017-06-30 2017-06-30 A real-time gesture analysis and evaluation method and system based on RGBD depth sensor

Publications (2)

Publication Number Publication Date
CN107463873A CN107463873A (en) 2017-12-12
CN107463873B true CN107463873B (en) 2020-02-21

Family

ID=60546461

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710523575.5A Expired - Fee Related CN107463873B (en) 2017-06-30 2017-06-30 A real-time gesture analysis and evaluation method and system based on RGBD depth sensor

Country Status (1)

Country Link
CN (1) CN107463873B (en)

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP6962878B2 (en) * 2018-07-24 2021-11-05 本田技研工業株式会社 Operation assistance system and operation assistance method
CN110032957B (en) * 2019-03-27 2023-10-17 长春理工大学 A gesture spatial domain matching method based on skeletal node information
CN110175566B (en) * 2019-05-27 2022-12-23 大连理工大学 Hand posture estimation system and method based on RGBD fusion network
CN110717385A (en) * 2019-08-30 2020-01-21 西安文理学院 Dynamic gesture recognition method
CN113657346A (en) * 2021-08-31 2021-11-16 深圳市比一比网络科技有限公司 Driver action recognition method based on combination of target detection and key point detection

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101923669A (en) * 2008-07-18 2010-12-22 史迪芬·凯斯 Smart Adaptive Design
CN103914132A (en) * 2013-01-07 2014-07-09 富士通株式会社 Method and system for recognizing gestures based on fingers
CN103926999A (en) * 2013-01-16 2014-07-16 株式会社理光 Palm opening and closing gesture recognition method and device and man-machine interaction method and device

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101923669A (en) * 2008-07-18 2010-12-22 史迪芬·凯斯 Smart Adaptive Design
CN103914132A (en) * 2013-01-07 2014-07-09 富士通株式会社 Method and system for recognizing gestures based on fingers
CN103926999A (en) * 2013-01-16 2014-07-16 株式会社理光 Palm opening and closing gesture recognition method and device and man-machine interaction method and device

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
Finger detection and hand posture recognition based on depth information;Stergios Poularakis等;《IEEE International Conference on Acoustics, Speech and Processing》;20140531;1-5页 *
基于Kinect骨架信息的交通警察手势识别;刘阳等;《计算机工程与应用》;20150331(第3期);157-161页 *

Also Published As

Publication number Publication date
CN107463873A (en) 2017-12-12

Similar Documents

Publication Publication Date Title
CN107463873B (en) A real-time gesture analysis and evaluation method and system based on RGBD depth sensor
US11643076B2 (en) Forward collision control method and apparatus, electronic device, program, and medium
WO2019232894A1 (en) Complex scene-based human body key point detection system and method
US12175684B2 (en) Pedestrian tracking method, computing device, pedestrian tracking system and storage medium
WO2020042419A1 (en) Gait-based identity recognition method and apparatus, and electronic device
US8005263B2 (en) Hand sign recognition using label assignment
CN103294996B (en) A kind of 3D gesture identification method
CN108509859A (en) A kind of non-overlapping region pedestrian tracting method based on deep neural network
CN108256421A (en) Dynamic gesture sequence real-time identification method, system and device
CN112766159A (en) Cross-database micro-expression identification method based on multi-feature fusion
JP2016062610A (en) Feature model generation method and feature model generation device
CN102855461B (en) In image, detect the method and apparatus of finger
CN107563286A (en) A kind of dynamic gesture identification method based on Kinect depth information
Ji et al. Integrating visual selective attention model with HOG features for traffic light detection and recognition
Kalsh et al. Sign language recognition system
CN113255779B (en) Multi-source perception data fusion identification method, system and computer readable storage medium
CN111008576B (en) Pedestrian detection and model training method, device and readable storage medium
CN111444764A (en) Gesture recognition method based on depth residual error network
CN115527269B (en) Intelligent human body posture image recognition method and system
CN103793926A (en) Target tracking method based on sample reselecting
CN110796101A (en) Face recognition method and system of embedded platform
CN108121972A (en) A kind of target identification method under the conditions of partial occlusion
CN108898623A (en) Method for tracking target and equipment
CN114663835A (en) A pedestrian tracking method, system, device and storage medium
CN110717385A (en) Dynamic gesture recognition method

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20200221

CF01 Termination of patent right due to non-payment of annual fee