CN108062170A - Multi-class human posture recognition method based on convolutional neural networks and intelligent terminal - Google Patents

Multi-class human posture recognition method based on convolutional neural networks and intelligent terminal Download PDF

Info

Publication number
CN108062170A
CN108062170A CN201711346910.5A CN201711346910A CN108062170A CN 108062170 A CN108062170 A CN 108062170A CN 201711346910 A CN201711346910 A CN 201711346910A CN 108062170 A CN108062170 A CN 108062170A
Authority
CN
China
Prior art keywords
convolutional neural
neural networks
intelligent terminal
data
method based
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201711346910.5A
Other languages
Chinese (zh)
Inventor
张雷
张志浩
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing Normal University
Original Assignee
Nanjing Normal University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing Normal University filed Critical Nanjing Normal University
Priority to CN201711346910.5A priority Critical patent/CN108062170A/en
Publication of CN108062170A publication Critical patent/CN108062170A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/033Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
    • G06F3/0346Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of the device orientation or free movement in a 3D space, e.g. 3D mice, 6-DOF [six degrees of freedom] pointers using gyroscopes, accelerometers or tilt-sensors
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • G06F3/014Hand-worn input/output arrangements, e.g. data gloves
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Human Computer Interaction (AREA)
  • General Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Biophysics (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Computational Linguistics (AREA)
  • Artificial Intelligence (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Image Analysis (AREA)

Abstract

本发明公开一种基于卷积神经网络和智能终端的多类别人体姿态识别方法,包括如下步骤:步骤1,采集移动智能终端设备的三轴加速度传感器数据,并记录对应的动作类别;步骤2,对三轴加速度传感器数据进行预处理后,将数据分为两类,一类是训练样本,一类是测试样本;步骤3,用训练样本训练卷积神经网络,并用测试样本测试其准确率并根据需求不断调整;步骤4,将训练好的卷积神经网络模型移植到移动智能终端上;步骤5,利用移动智能终端采集三轴加速度传感器数据,进行预处理后,输入到训练好的卷积神经网络模型,得到人体姿态识别结果。此种方法识别精度高,能够识别的类型多。

The invention discloses a multi-category human gesture recognition method based on a convolutional neural network and an intelligent terminal, comprising the following steps: Step 1, collecting triaxial acceleration sensor data of a mobile intelligent terminal device, and recording corresponding action categories; Step 2, After preprocessing the triaxial acceleration sensor data, divide the data into two categories, one is training samples and the other is testing samples; step 3, train the convolutional neural network with training samples, and test its accuracy with test samples and Constantly adjust according to demand; step 4, transplant the trained convolutional neural network model to the mobile smart terminal; step 5, use the mobile smart terminal to collect the three-axis acceleration sensor data, preprocess it, and input it to the trained convolutional neural network Neural network model to obtain the result of human gesture recognition. This method has high recognition accuracy and can recognize many types.

Description

基于卷积神经网络和智能终端的多类别人体姿态识别方法Multi-class Human Pose Recognition Method Based on Convolutional Neural Network and Intelligent Terminal

技术领域technical field

本发明属于人工智能研究领域,涉及穿戴式智能监控领域,特别涉及一种利用传感器进行人体姿态识别的方法。The invention belongs to the field of artificial intelligence research, relates to the field of wearable intelligent monitoring, in particular to a method for recognizing human body gestures by using sensors.

背景技术Background technique

人体姿态识别技术在虚拟现实、移动游戏、医疗保健、人机交互、图像识别等领域有着广泛应用。大体上姿态识别技术分为二种:非穿戴式和穿戴式。非穿戴式技术顾名思义,指姿态识别设备不与人体直接接触的人体姿态识别技术,例如图像识别技术。穿戴式人体姿态识别技术相比于非穿戴式,有着空间不受限的优点,在研究和应用上有着更好的发展空间。由于人体姿态的多样性,以及个体动作的差异性,如何建立一种高识别精度的姿态识别模型是目前一直探讨和关注的研究课题。Human gesture recognition technology has been widely used in virtual reality, mobile games, healthcare, human-computer interaction, image recognition and other fields. Generally, gesture recognition technology is divided into two types: non-wearable and wearable. As the name implies, non-wearable technology refers to human gesture recognition technology in which gesture recognition equipment does not come into direct contact with the human body, such as image recognition technology. Compared with non-wearable body posture recognition technology, wearable body posture recognition technology has the advantage of unlimited space, and has a better development space in research and application. Due to the diversity of human body postures and the differences in individual movements, how to establish a posture recognition model with high recognition accuracy is a research topic that has been discussed and paid attention to at present.

通常,为了保持较高的识别精度,会在人体多出关节上安置多个传感器设备。虽然这种方法能够直观地发现各种动作的加速度特征,但实际应用中要求使用者携带多个传感器,很是不方便。如何使用较少甚至只是用一组传感器进行高准确率的人体姿态识别是一个非常实际的研究问题。Usually, in order to maintain a high recognition accuracy, multiple sensor devices are placed on multiple joints of the human body. Although this method can intuitively discover the acceleration characteristics of various actions, it is inconvenient to require the user to carry multiple sensors in practical applications. How to perform high-accuracy human pose recognition with fewer or even a set of sensors is a very practical research problem.

使用智能手机或智能手表的内置传感器进行人体姿态识别,国内外早已有很多研究应用,目前市面上多数智能手环手表和手机均有姿态识别的应用程序APP。此类人体姿态识别方法绝大多数为阈值检测法,即通过判断传感器原始或处理后的数据是否大于或小于预先设定的好阈值来分类动作类型。这种方法计算简单,占用智能移动设备的内存少,但与此同时,其缺点也很明显:不同产品准确率参差不齐,能够识别的动作类别也十分有限。这一方面是各个公司研发人员技术差距的原因,更重要的一方面原因是此类方法的局限。需要识别的动作类别越多,此种算法构建起来越复杂。Using the built-in sensors of smart phones or smart watches for human body posture recognition has already had many research applications at home and abroad. At present, most smart bracelet watches and mobile phones on the market have gesture recognition applications APP. The vast majority of such human gesture recognition methods are threshold detection methods, which classify action types by judging whether the raw or processed data of the sensor is greater than or less than a preset good threshold. This method is simple to calculate and takes up less memory of smart mobile devices, but at the same time, its shortcomings are also obvious: the accuracy rate of different products is uneven, and the types of actions that can be recognized are also very limited. On the one hand, this is the reason for the technical gap of R&D personnel in various companies, and more importantly, on the one hand, it is the limitation of such methods. The more categories of actions that need to be recognized, the more complex the algorithm is to build.

深度学习在模式识别上有着很好的发展前景。深度学习(Deep Learning)起源于人工神经网络(Artificial Neural Network,ANN)的研究。其中卷积神经网络是含有卷积层(Convolutional Layer)的神经网络。卷积神经网络在计算机视觉领域受到极大关注,卷积神经网络不仅可以处理一维数据(例如,文本),它还特别适合处理二维数据(例如,图像)和三维数据(例如,视频以及本专利提及的三维加速度数据)。卷积神经网络属于人工智能范畴,在模式识别分类器的构建上比传统方法效率更高,且易于扩展,能够实现比传统方法动作识别类型更多的识别模型。Deep learning has great prospects for development in pattern recognition. Deep Learning originated from the research of Artificial Neural Network (ANN). The convolutional neural network is a neural network containing a convolutional layer. Convolutional neural networks have received great attention in the field of computer vision. Convolutional neural networks can not only process one-dimensional data (such as text), but are also particularly suitable for processing two-dimensional data (such as images) and three-dimensional data (such as video and The three-dimensional acceleration data mentioned in this patent). The convolutional neural network belongs to the category of artificial intelligence. It is more efficient than the traditional method in the construction of the pattern recognition classifier, and it is easy to expand. It can realize more recognition models than the traditional method of action recognition.

发明内容Contents of the invention

本发明的目的,在于提供一种基于卷积神经网络和智能终端的多类别人体姿态识别方法,其识别精度高,能够识别的类型多。The purpose of the present invention is to provide a multi-category human gesture recognition method based on convolutional neural network and intelligent terminal, which has high recognition accuracy and can recognize many types.

为了达成上述目的,本发明的解决方案是:In order to achieve the above object, the solution of the present invention is:

一种基于卷积神经网络和智能终端的多类别人体姿态识别方法,包括如下步骤:A multi-category human gesture recognition method based on convolutional neural network and intelligent terminal, comprising the following steps:

步骤1,采集移动智能终端设备的三轴加速度传感器数据,并记录对应的动作类别;Step 1, collect the three-axis acceleration sensor data of the mobile smart terminal device, and record the corresponding action category;

步骤2,对三轴加速度传感器数据进行预处理后,将数据分为两类,一类是训练样本,一类是测试样本;Step 2, after preprocessing the triaxial acceleration sensor data, divide the data into two categories, one is training samples and the other is testing samples;

步骤3,用训练样本训练卷积神经网络,并用测试样本测试其准确率并根据需求不断调整;Step 3, use the training samples to train the convolutional neural network, and use the test samples to test its accuracy and adjust it according to the needs;

步骤4,将训练好的卷积神经网络模型移植到移动智能终端上;Step 4, transplanting the trained convolutional neural network model to the mobile smart terminal;

步骤5,利用移动智能终端采集三轴加速度传感器数据,进行预处理后,输入到训练好的卷积神经网络模型,得到人体姿态识别结果。Step 5: Use the mobile smart terminal to collect the data of the three-axis acceleration sensor, and after preprocessing, input it into the trained convolutional neural network model to obtain the human body posture recognition result.

上述步骤1中,采样频率设定为25Hz。In the above step 1, the sampling frequency is set to 25Hz.

上述步骤2中,对数据进行预处理,包括对数据进行滤波、归一化处理,并将数据调整成卷积神经网络的输入格式。In the above step 2, the data is preprocessed, including filtering and normalizing the data, and adjusting the data to the input format of the convolutional neural network.

上述步骤2中,将数据中的75%作为训练样本,把数据中的25%作为测试样本。In the above step 2, 75% of the data are used as training samples, and 25% of the data are used as test samples.

上述步骤3的具体步骤是:The specific steps of the above step 3 are:

a,建立多层的卷积神经网络模型;a, Establish a multi-layer convolutional neural network model;

b,导入训练样本调节卷积神经网络模型参数,得到高准确率的模型。b. Import training samples to adjust the parameters of the convolutional neural network model to obtain a high-accuracy model.

上述步骤b中,调节卷积神经网络模型参数包括各层神经元数量调节,损失函数及卷积核调节。In the above step b, the adjustment of the parameters of the convolutional neural network model includes the adjustment of the number of neurons in each layer, the adjustment of the loss function and the convolution kernel.

上述步骤a中,卷积神经网络模型的结构包括:输入层,两层卷积和最大池化层,一个全连接层和一个输出层。In the above step a, the structure of the convolutional neural network model includes: an input layer, two layers of convolution and maximum pooling layers, a fully connected layer and an output layer.

上述卷积神经网络模型中,卷积核大小为3*3,两个卷积层的神经元个数分别为96和198,第一层卷积核的数据尺寸为(5,5,3),共有96个卷积核;整个实验的池化核都为(2,2),池化步长都为2,都使用最大池化策略;第二层卷积层的卷积核数据尺寸为(3,3,96),共有198个卷积核;全连接层包含1024个隐藏节点;学习率为0.0001;drop-out为1。In the above convolutional neural network model, the size of the convolution kernel is 3*3, the number of neurons in the two convolution layers is 96 and 198 respectively, and the data size of the convolution kernel in the first layer is (5,5,3) , a total of 96 convolution kernels; the pooling kernels of the whole experiment are (2,2), the pooling step size is 2, and the maximum pooling strategy is used; the convolution kernel data size of the second convolutional layer is (3,3,96), a total of 198 convolution kernels; the fully connected layer contains 1024 hidden nodes; the learning rate is 0.0001; the drop-out is 1.

采用上述方案后,由于卷积神经网络的优势,只要样本数量足够,通过调整参数,本发明可以将能够分类的动作类别扩展到更多。本发明在智能监控、人体姿态识别等方面具有重要实际应用意义。After adopting the above solution, due to the advantages of the convolutional neural network, as long as the number of samples is sufficient, the present invention can expand the action categories that can be classified to more by adjusting the parameters. The invention has important practical application significance in aspects such as intelligent monitoring and human body gesture recognition.

本发明具有如下优点:The present invention has the following advantages:

(1)本发明利用人工智能-卷积神经网络识别方法,识别精度高,能够识别的类型多;(1) The present invention utilizes the artificial intelligence-convolutional neural network identification method, which has high identification accuracy and many types that can be identified;

(2)本发明识别方法识别的动作数量具有可扩展性,且扩展操作简单,易于开发人员操作;(2) The number of actions recognized by the recognition method of the present invention is scalable, and the expansion operation is simple and easy for developers to operate;

(3)本发明相比于视频或者图像识别的方法,可以有效的保护用户隐私;(3) Compared with video or image recognition methods, the present invention can effectively protect user privacy;

(4)本发明可应用于人们常用的安卓智能手机和智能手表,有很好的推广性。(4) The present invention can be applied to Android smart phones and smart watches commonly used by people, and has good promotion.

附图说明Description of drawings

图1是本发明的流程图;Fig. 1 is a flow chart of the present invention;

图2是本发明的原理图;Fig. 2 is a schematic diagram of the present invention;

图3是手机三轴加速度传感器方向示意图;Fig. 3 is a schematic diagram of the direction of the three-axis acceleration sensor of the mobile phone;

图4是不同动作对应的部分加速度数据波形示意图;Fig. 4 is a schematic diagram of partial acceleration data waveforms corresponding to different actions;

图5是交叉熵(cross_entropy)随训练次数的变化图。Figure 5 is a diagram of the change of cross entropy (cross_entropy) with the number of training.

具体实施方式Detailed ways

以下将结合附图,对本发明的技术方案及有益效果进行详细说明。The technical solutions and beneficial effects of the present invention will be described in detail below in conjunction with the accompanying drawings.

本发明提供一种基于卷积神经网络和智能终端的多类别人体姿态识别方法,包括如下步骤:The present invention provides a multi-category human posture recognition method based on convolutional neural network and intelligent terminal, comprising the following steps:

步骤1,在第三者监督记录的情况下采集移动智能终端设备的三轴加速度传感器数据,并预先附上动作类别标签,作为人体姿态识别模型进行训练时的样本来使用;Step 1, collect the three-axis acceleration sensor data of the mobile smart terminal device under the supervision and recording of a third party, and pre-attach the action category label, and use it as a sample for human body posture recognition model training;

步骤2,对三轴加速度传感器数据进行预处理,包括对数据进行滤波、归一化处理,并将数据调整成卷积神经网络的输入格式,将数据分为两类,一类是训练样本,一类是测试样本;Step 2, preprocessing the data of the three-axis acceleration sensor, including filtering and normalizing the data, and adjusting the data into the input format of the convolutional neural network, dividing the data into two categories, one is training samples, One is the test sample;

步骤3,用训练样本训练卷积神经网络,并用测试样本测试其准确率并根据需求不断调整;具体包括:Step 3, use the training samples to train the convolutional neural network, and use the test samples to test its accuracy and adjust it according to the needs; specifically include:

a,建立多层的卷积神经网络模型;a, Establish a multi-layer convolutional neural network model;

b,导入训练样本调节卷积神经网络模型参数,得到高准确率的模型;其中,所述的卷积神经网络模型参数调节,包括:各层神经元数量调节,损失函数及卷积核调节。b. Importing training samples to adjust convolutional neural network model parameters to obtain a high-accuracy model; wherein, the convolutional neural network model parameter adjustment includes: adjustment of the number of neurons in each layer, loss function and convolution kernel adjustment.

步骤4,将训练好的卷积神经网络模型(人体姿态识别模型)移植到移动智能终端上,实现实时终端姿态识别处理功能;Step 4, transplanting the trained convolutional neural network model (human gesture recognition model) to the mobile smart terminal to realize the real-time terminal gesture recognition processing function;

步骤5,利用移动智能终端采集三轴加速度传感器数据,进行预处理后,输入到训练好的卷积神经网络模型,得到人体姿态识别结果。Step 5: Use the mobile smart terminal to collect the data of the three-axis acceleration sensor, and after preprocessing, input it into the trained convolutional neural network model to obtain the human body posture recognition result.

本发明基于预设的训练集和卷积神经网络结构训练得到人体姿态识别模型,能对走、跑、上楼、下楼、仰卧起坐、扫地、擦七种动作姿态进行识别。The present invention obtains a human body posture recognition model based on a preset training set and convolutional neural network structure training, and can recognize seven postures of walking, running, going upstairs, going downstairs, sit-ups, sweeping the floor, and wiping.

图1为目标处理的流程图,从智能移动终端采集到人体运动的三维加速度时间序列后,整合处理后输入至初始卷积神经网络进行模型训练,将训练好的符合设计要求的模型输出至移动终端上,使之能在移动智能终端上离线识别人体动作。Figure 1 is the flow chart of target processing. After collecting the three-dimensional acceleration time series of human body movement from the smart mobile terminal, after integration and processing, it is input to the initial convolutional neural network for model training, and the trained model that meets the design requirements is output to the mobile On the terminal, it can recognize human body movements offline on the mobile smart terminal.

图2为卷积神经网络结构图,主要包括:输入层,两层卷积和最大池化层,一个全连接层和一个输出层。输入是预处理后的三轴加速度数据x、y、z,例如智能手机的各个加速度方向如图3所示。Figure 2 is a structural diagram of a convolutional neural network, which mainly includes: an input layer, two convolutional and maximum pooling layers, a fully connected layer and an output layer. The input is the preprocessed three-axis acceleration data x, y, z, for example, each acceleration direction of a smart phone is shown in Figure 3.

可选的,采集智能终端的采样频率可设定为25Hz。以此频率采集到的部分动作加速度数据波形如图4所示。可选的,本发明示例定义每2.56秒为一个动作样本,即每64组数据为一个样本。当然,采样频率可以根据实际需求自行设置合适的值,此处不做限定。Optionally, the sampling frequency of collecting the smart terminal can be set to 25Hz. Part of the motion acceleration data waveform collected at this frequency is shown in Figure 4. Optionally, the example of the present invention defines an action sample every 2.56 seconds, that is, every 64 sets of data is a sample. Of course, the sampling frequency can be set to an appropriate value according to actual needs, which is not limited here.

为了训练卷积神经网络,本发明将采集到的样本分为两类:训练样本和测试样本。训练样本作为卷积神经网络的输入进行模型训练,测试样本作为识别准确率的考量依据。默认的,把数据集的75%作为训练集,把数据集的25%作为测试集。In order to train the convolutional neural network, the present invention divides the collected samples into two categories: training samples and testing samples. The training samples are used as the input of the convolutional neural network for model training, and the test samples are used as the basis for considering the recognition accuracy. By default, 75% of the dataset is used as the training set and 25% of the dataset is used as the test set.

作为卷积神经网络的输入,加速度数据需要进行折叠。本发明示例将三轴加速度数据尺寸设置为(8,8,3),分别代表长、宽和深度。其中每一轴的数据矩阵形式如下:As input to a convolutional neural network, the acceleration data needs to be folded. In the example of the present invention, the three-axis acceleration data size is set to (8,8,3), representing length, width and depth respectively. The data matrix form of each axis is as follows:

这样就可以使每一小段时间内的三轴加速度数据形如像素图片,以适配卷积神经网络的训练。当然,可以根据实际需求自行设置合适的值,此处不做限定。In this way, the three-axis acceleration data in each short period of time can be shaped like a pixel picture to adapt to the training of the convolutional neural network. Of course, you can set an appropriate value according to actual needs, and there is no limitation here.

神经网络基本单元神经元的公式如下:The formula of neuron, the basic unit of neural network, is as follows:

其中,x是神经元输入,n是输入参数个数,b是偏置,hW,b(x)是神经元输出。Among them, x is the neuron input, n is the number of input parameters, b is the bias, h W,b (x) is the neuron output.

卷积神经网络与普通神经网络的区别在于,卷积神经网络包含了一个由卷积层和子采样层构成的特征抽取器。在卷积神经网络的卷积层中,一个神经元只与部分邻层神经元连接。在CNN的一个卷积层中,通常包含若干个特征平面(featureMap),每个特征平面由一些矩形排列的神经元组成,同一特征平面的神经元共享权值,这里共享的权值就是卷积核。卷积核一般以随机小数矩阵的形式初始化,在网络的训练过程中卷积核将学习得到合理的权值。共享权值(卷积核)带来的直接好处是减少网络各层之间的连接,同时又降低了过拟合的风险。The difference between a convolutional neural network and an ordinary neural network is that the convolutional neural network contains a feature extractor composed of a convolutional layer and a subsampling layer. In a convolutional layer of a convolutional neural network, a neuron is only connected to some neurons in neighboring layers. In a convolutional layer of CNN, it usually contains several feature planes (featureMap). Each feature plane is composed of some neurons arranged in a rectangle. The neurons of the same feature plane share weights. The shared weights here are convolutions. nuclear. The convolution kernel is generally initialized in the form of a random decimal matrix, and the convolution kernel will learn to obtain reasonable weights during the training process of the network. The direct benefit of sharing weights (convolution kernels) is to reduce the connections between layers of the network while reducing the risk of overfitting.

本发明此部分只需要设置卷积核的大小及神经元个数即可。卷积核大小和神经元个数的取值为经验值,没有固定的取值方法,本发明示例中卷积核大小为3*3,两个卷积层的神经元个数分别为96和198,此数据仅供参考。In this part of the present invention, only the size of the convolution kernel and the number of neurons need to be set. The value of the convolution kernel size and the number of neurons is an empirical value, and there is no fixed value method. In the example of the present invention, the convolution kernel size is 3*3, and the neurons of the two convolution layers are respectively 96 and 198, this data is for reference only.

子采样也叫做池化(pooling),通常有均值子采样(mean pooling)和最大值子采样(max pooling)两种形式。子采样可以看作一种特殊的卷积过程。卷积和子采样大大简化了模型复杂度,减少了模型的参数。Subsampling is also called pooling, and usually has two forms: mean pooling and max pooling. Subsampling can be seen as a special kind of convolution process. Convolution and subsampling greatly simplify the model complexity and reduce the parameters of the model.

模型最终的具体实验参数列举如下:第一层卷积核的数据尺寸为(5,5,3),共有96个卷积核;整个实验的池化核都为(2,2),池化步长都为2,都使用最大池化策略;第二层卷积层的卷积核数据尺寸为(3,3,96),共有198个卷积核;全连接层包含1024个隐藏节点;学习率为0.0001;drop-out为1。The final specific experimental parameters of the model are listed as follows: the data size of the first layer of convolution kernels is (5,5,3), and there are 96 convolution kernels in total; the pooling kernels of the entire experiment are (2,2), and the pooling The step size is 2, and the maximum pooling strategy is used; the convolution kernel data size of the second convolutional layer is (3,3,96), and there are 198 convolution kernels in total; the fully connected layer contains 1024 hidden nodes; The learning rate is 0.0001; the drop-out is 1.

若训练数据量不够大,则需要对数据进行重复使用。每次将50个数据输入至神经网络进行训练,每50次测量一次识别准确率及交叉熵,其中交叉熵的变化图如图5所示。If the amount of training data is not large enough, the data needs to be reused. Each time, 50 pieces of data are input into the neural network for training, and the recognition accuracy and cross-entropy are measured every 50 times. The change diagram of the cross-entropy is shown in Figure 5.

当训练的卷积神经网络符合设计要求,即可将该模型提取到移动智能终端上使用。若训练的卷积神经网络不符合设计要求,需要修改各隐藏层的神经元个数。神经元个数修改到哪个值为宜,需要反复测试。若上述修改各隐藏层的神经元个数的方法对识别准确率影响甚微,建议添加隐藏层数或增加训练样本数。When the trained convolutional neural network meets the design requirements, the model can be extracted to the mobile smart terminal for use. If the trained convolutional neural network does not meet the design requirements, the number of neurons in each hidden layer needs to be modified. Which value is appropriate to modify the number of neurons needs to be tested repeatedly. If the above method of modifying the number of neurons in each hidden layer has little effect on the recognition accuracy, it is recommended to add the number of hidden layers or increase the number of training samples.

需要说明的是,本发明实施例中的人体姿态识别装置具体可以集成在智能移动终端中,上述智能终端具体可以为智能手机、智能手表等终端,此处不作限定。It should be noted that the human body gesture recognition device in the embodiment of the present invention may be integrated into a smart mobile terminal, and the above-mentioned smart terminal may specifically be a smart phone, a smart watch and other terminals, which are not limited here.

应理解,本发明实施例中的人体姿态识别装置可以实现上述方法实施例中的全部技术方案,其各个功能模块的功能可以根据上述方法实施例中的方法具体实现,其具体实现过程可参照上述实施例中的相关描述,此处不再赘述。It should be understood that the human body posture recognition device in the embodiment of the present invention can realize all the technical solutions in the above-mentioned method embodiments, and the functions of each functional module can be specifically realized according to the methods in the above-mentioned method embodiments, and its specific implementation process can refer to the above-mentioned Relevant descriptions in the embodiments will not be repeated here.

由上可见,本发明实施例中的人体姿态识别装置通过采集智能终端的加速度数据,基于采集到的所述智能终端的加速度数据,并将预处理后的数据输入已训练好的人体姿态识别模型,得到人体姿态识别结果。由于人体姿态识别模型是基于预设的训练集合卷积神经网络训练得到,因此,通过将加速度数据预处理后输入已训练好的人体姿态识别模型,即可实现对人体姿态的识别,从而实现了基于加速度数据的非视觉手段的人体姿态识别。It can be seen from the above that the human body posture recognition device in the embodiment of the present invention collects the acceleration data of the smart terminal, based on the collected acceleration data of the smart terminal, and inputs the preprocessed data into the trained human body posture recognition model , to get the result of human gesture recognition. Since the human body posture recognition model is trained based on the preset training set convolutional neural network, the recognition of human body posture can be realized by preprocessing the acceleration data and inputting it into the trained human body posture recognition model, thereby realizing Human pose recognition based on non-visual means of acceleration data.

在上述实施例中,对各个实施例的描述都各有侧重,某个实施例中没有详述的部分,可以参见其它实施例的相关描述。In the foregoing embodiments, the descriptions of each embodiment have their own emphases, and for parts not described in detail in a certain embodiment, reference may be made to relevant descriptions of other embodiments.

以上实施例仅为说明本发明的技术思想,不能以此限定本发明的保护范围,凡是按照本发明提出的技术思想,在技术方案基础上所做的任何改动,均落入本发明保护范围之内。The above embodiments are only to illustrate the technical ideas of the present invention, and can not limit the protection scope of the present invention with this. All technical ideas proposed in accordance with the present invention, any changes made on the basis of technical solutions, all fall within the protection scope of the present invention. Inside.

Claims (8)

1. a kind of multi-class human posture recognition method based on convolutional neural networks and intelligent terminal, it is characterised in that including such as Lower step:
Step 1, the 3-axis acceleration sensor data of mobile intelligent terminal equipment are gathered, and record corresponding action classification;
Step 2, two classes are splitted data into after being pre-processed to 3-axis acceleration sensor data, one kind is training sample, one Class is test sample;
Step 3, with training sample training convolutional neural networks, and test its accuracy rate with test sample and constantly adjust according to demand It is whole;
It step 4, will be on trained convolutional neural networks model transplantations to mobile intelligent terminal;
Step 5,3-axis acceleration sensor data are gathered using mobile intelligent terminal, after being pre-processed, is input to and trains Convolutional neural networks model, obtain human body attitude recognition result.
2. the multi-class human posture recognition method based on convolutional neural networks and intelligent terminal as described in claim 1, It is characterized in that:In the step 1, sample frequency is set as 25Hz.
3. the multi-class human posture recognition method based on convolutional neural networks and intelligent terminal as described in claim 1, It is characterized in that:In the step 2, data are pre-processed, including being filtered to data, normalized, and by data It is adjusted to the input format of convolutional neural networks.
4. the multi-class human posture recognition method based on convolutional neural networks and intelligent terminal as described in claim 1, It is characterized in that:In the step 2, using 75% in data as training sample, using 25% in data as test sample.
5. the multi-class human posture recognition method based on convolutional neural networks and intelligent terminal as described in claim 1, It is characterized in that:The step 3 comprises the concrete steps that:
A establishes the convolutional neural networks model of multilayer;
B imports training sample and adjusts convolutional neural networks model parameter, obtains the model of high-accuracy.
6. the multi-class human posture recognition method based on convolutional neural networks and intelligent terminal as claimed in claim 5, It is characterized in that:In the step b, adjust convolutional neural networks model parameter and adjusted including each layer neuronal quantity, loss function And convolution kernel is adjusted.
7. the multi-class human posture recognition method based on convolutional neural networks and intelligent terminal as claimed in claim 5, It is characterized in that:In the step a, the structure of convolutional neural networks model includes:Input layer, two layers of convolution and maximum pond layer, One full articulamentum and an output layer.
8. the multi-class human posture recognition method based on convolutional neural networks and intelligent terminal as claimed in claim 7, It is characterized in that:In the convolutional neural networks model, convolution kernel size is 3*3, and the neuron number of two convolutional layers is respectively 96 and 198, the data size of first layer convolution kernel is (5,5,3), shares 96 convolution kernels;The Chi Huahe entirely tested is (2,2), pond step-length are all 2, all using maximum pondization strategy;The convolution kernel data size of second layer convolutional layer for (3,3, 96) 198 convolution kernels, are shared;Full articulamentum includes 1024 concealed nodes;Learning rate is 0.0001;Drop-out is 1.
CN201711346910.5A 2017-12-15 2017-12-15 Multi-class human posture recognition method based on convolutional neural networks and intelligent terminal Pending CN108062170A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201711346910.5A CN108062170A (en) 2017-12-15 2017-12-15 Multi-class human posture recognition method based on convolutional neural networks and intelligent terminal

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201711346910.5A CN108062170A (en) 2017-12-15 2017-12-15 Multi-class human posture recognition method based on convolutional neural networks and intelligent terminal

Publications (1)

Publication Number Publication Date
CN108062170A true CN108062170A (en) 2018-05-22

Family

ID=62139047

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201711346910.5A Pending CN108062170A (en) 2017-12-15 2017-12-15 Multi-class human posture recognition method based on convolutional neural networks and intelligent terminal

Country Status (1)

Country Link
CN (1) CN108062170A (en)

Cited By (24)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109033995A (en) * 2018-06-29 2018-12-18 出门问问信息科技有限公司 Identify the method, apparatus and intelligence wearable device of user behavior
CN109346166A (en) * 2018-11-22 2019-02-15 深圳市云护宝计算机技术有限公司 A kind of inpatient department intelligent medical care bracelet and its deep learning modeling method
CN109670548A (en) * 2018-12-20 2019-04-23 电子科技大学 HAR algorithm is inputted based on the more sizes for improving LSTM-CNN
CN109685148A (en) * 2018-12-28 2019-04-26 南京师范大学 Multi-class human motion recognition method and identifying system
CN109711324A (en) * 2018-12-24 2019-05-03 南京师范大学 Human pose recognition method based on Fourier transform and convolutional neural network
CN109726662A (en) * 2018-12-24 2019-05-07 南京师范大学 Multi-category Human Pose Recognition Method Based on Convolutional and Recurrent Combination Neural Networks
CN109740651A (en) * 2018-12-24 2019-05-10 南京师范大学 Limb recognition method based on 1-norm data processing transformation and convolutional neural network
CN109770912A (en) * 2019-01-23 2019-05-21 复旦大学 A gait anomaly classification method based on deep convolutional neural network
CN109770913A (en) * 2019-01-23 2019-05-21 复旦大学 A method of gait anomaly recognition based on back-propagation neural network
CN110275161A (en) * 2019-06-28 2019-09-24 台州睿联科技有限公司 A kind of wireless human body gesture recognition method applied to Intelligent bathroom
EP3582196A1 (en) * 2018-06-11 2019-12-18 Verisure Sàrl Shock sensor in an alarm system
CN110610158A (en) * 2019-09-16 2019-12-24 南京师范大学 A human body posture recognition method and system based on convolution and gated recurrent neural network
CN111178288A (en) * 2019-12-31 2020-05-19 南京师范大学 Human pose recognition method and device based on local error layer-by-layer training
CN111222459A (en) * 2020-01-06 2020-06-02 上海交通大学 Visual angle-independent video three-dimensional human body posture identification method
CN111700624A (en) * 2020-07-27 2020-09-25 中国科学院合肥物质科学研究院 A pattern recognition method and system for detecting motion posture of smart bracelet
CN111723662A (en) * 2020-05-18 2020-09-29 南京师范大学 A Human Gesture Recognition Method Based on Convolutional Neural Network
CN111753683A (en) * 2020-06-11 2020-10-09 南京师范大学 A Human Gesture Recognition Method Based on Multi-Expert Convolutional Neural Network
CN111860188A (en) * 2020-06-24 2020-10-30 南京师范大学 A Human Gesture Recognition Method Based on Time and Channel Dual Attention
CN111860191A (en) * 2020-06-24 2020-10-30 南京师范大学 A Human Pose Recognition Method Based on Channel Selection Convolutional Neural Network
CN112115964A (en) * 2020-08-04 2020-12-22 深圳市联合视觉创新科技有限公司 Acceleration labeling model generation method, acceleration labeling method, equipment and medium
CN112287810A (en) * 2020-10-27 2021-01-29 南京大学 A device and method for dynamically increasing motion recognition gesture
CN113705507A (en) * 2021-09-02 2021-11-26 上海交通大学 Mixed reality open set human body posture recognition method based on deep learning
CN114167984A (en) * 2021-01-28 2022-03-11 Oppo广东移动通信有限公司 Device control method, device, storage medium and electronic device
CN114916928A (en) * 2022-05-12 2022-08-19 电子科技大学 A Human Pose Multi-Channel Convolutional Neural Network Detection Method

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106022285A (en) * 2016-05-30 2016-10-12 北京智芯原动科技有限公司 Vehicle type identification method and vehicle type identification device based on convolutional neural network
CN106228177A (en) * 2016-06-30 2016-12-14 浙江大学 Daily life subject image recognition methods based on convolutional neural networks
CN106388831A (en) * 2016-11-04 2017-02-15 郑州航空工业管理学院 Method for detecting falling actions based on sample weighting algorithm
CN107153871A (en) * 2017-05-09 2017-09-12 浙江农林大学 Fall detection method based on convolutional neural networks and mobile phone sensor data
CN107180225A (en) * 2017-04-19 2017-09-19 华南理工大学 A kind of recognition methods for cartoon figure's facial expression

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106022285A (en) * 2016-05-30 2016-10-12 北京智芯原动科技有限公司 Vehicle type identification method and vehicle type identification device based on convolutional neural network
CN106228177A (en) * 2016-06-30 2016-12-14 浙江大学 Daily life subject image recognition methods based on convolutional neural networks
CN106388831A (en) * 2016-11-04 2017-02-15 郑州航空工业管理学院 Method for detecting falling actions based on sample weighting algorithm
CN107180225A (en) * 2017-04-19 2017-09-19 华南理工大学 A kind of recognition methods for cartoon figure's facial expression
CN107153871A (en) * 2017-05-09 2017-09-12 浙江农林大学 Fall detection method based on convolutional neural networks and mobile phone sensor data

Cited By (33)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3582196A1 (en) * 2018-06-11 2019-12-18 Verisure Sàrl Shock sensor in an alarm system
WO2019238256A1 (en) * 2018-06-11 2019-12-19 Verisure Sàrl Shock sensor in an alarm system
CN109033995A (en) * 2018-06-29 2018-12-18 出门问问信息科技有限公司 Identify the method, apparatus and intelligence wearable device of user behavior
CN109346166A (en) * 2018-11-22 2019-02-15 深圳市云护宝计算机技术有限公司 A kind of inpatient department intelligent medical care bracelet and its deep learning modeling method
CN109670548A (en) * 2018-12-20 2019-04-23 电子科技大学 HAR algorithm is inputted based on the more sizes for improving LSTM-CNN
CN109670548B (en) * 2018-12-20 2023-01-06 电子科技大学 Multi-scale input HAR algorithm based on improved LSTM-CNN
CN109711324A (en) * 2018-12-24 2019-05-03 南京师范大学 Human pose recognition method based on Fourier transform and convolutional neural network
CN109726662A (en) * 2018-12-24 2019-05-07 南京师范大学 Multi-category Human Pose Recognition Method Based on Convolutional and Recurrent Combination Neural Networks
CN109740651A (en) * 2018-12-24 2019-05-10 南京师范大学 Limb recognition method based on 1-norm data processing transformation and convolutional neural network
CN109685148A (en) * 2018-12-28 2019-04-26 南京师范大学 Multi-class human motion recognition method and identifying system
CN109770913A (en) * 2019-01-23 2019-05-21 复旦大学 A method of gait anomaly recognition based on back-propagation neural network
CN109770912A (en) * 2019-01-23 2019-05-21 复旦大学 A gait anomaly classification method based on deep convolutional neural network
CN110275161A (en) * 2019-06-28 2019-09-24 台州睿联科技有限公司 A kind of wireless human body gesture recognition method applied to Intelligent bathroom
CN110610158A (en) * 2019-09-16 2019-12-24 南京师范大学 A human body posture recognition method and system based on convolution and gated recurrent neural network
CN111178288A (en) * 2019-12-31 2020-05-19 南京师范大学 Human pose recognition method and device based on local error layer-by-layer training
CN111178288B (en) * 2019-12-31 2024-03-01 南京师范大学 Human body posture recognition method and device based on local error layer-by-layer training
CN111222459A (en) * 2020-01-06 2020-06-02 上海交通大学 Visual angle-independent video three-dimensional human body posture identification method
CN111222459B (en) * 2020-01-06 2023-05-12 上海交通大学 Visual angle independent video three-dimensional human body gesture recognition method
CN111723662A (en) * 2020-05-18 2020-09-29 南京师范大学 A Human Gesture Recognition Method Based on Convolutional Neural Network
CN111753683A (en) * 2020-06-11 2020-10-09 南京师范大学 A Human Gesture Recognition Method Based on Multi-Expert Convolutional Neural Network
CN111860191A (en) * 2020-06-24 2020-10-30 南京师范大学 A Human Pose Recognition Method Based on Channel Selection Convolutional Neural Network
CN111860188A (en) * 2020-06-24 2020-10-30 南京师范大学 A Human Gesture Recognition Method Based on Time and Channel Dual Attention
CN111700624A (en) * 2020-07-27 2020-09-25 中国科学院合肥物质科学研究院 A pattern recognition method and system for detecting motion posture of smart bracelet
CN111700624B (en) * 2020-07-27 2024-03-12 中国科学院合肥物质科学研究院 Pattern recognition method and system for detecting motion gesture by intelligent bracelet
CN112115964A (en) * 2020-08-04 2020-12-22 深圳市联合视觉创新科技有限公司 Acceleration labeling model generation method, acceleration labeling method, equipment and medium
CN112115964B (en) * 2020-08-04 2025-05-16 深圳市联合视觉创新科技有限公司 Acceleration annotation model generation method, acceleration annotation method, device and medium
CN112287810A (en) * 2020-10-27 2021-01-29 南京大学 A device and method for dynamically increasing motion recognition gesture
CN114167984A (en) * 2021-01-28 2022-03-11 Oppo广东移动通信有限公司 Device control method, device, storage medium and electronic device
CN114167984B (en) * 2021-01-28 2024-03-12 Oppo广东移动通信有限公司 Equipment control method and device, storage medium and electronic equipment
CN113705507A (en) * 2021-09-02 2021-11-26 上海交通大学 Mixed reality open set human body posture recognition method based on deep learning
CN113705507B (en) * 2021-09-02 2023-09-19 上海交通大学 Mixed reality open set human posture recognition method based on deep learning
CN114916928B (en) * 2022-05-12 2023-08-04 电子科技大学 Human body posture multichannel convolutional neural network detection method
CN114916928A (en) * 2022-05-12 2022-08-19 电子科技大学 A Human Pose Multi-Channel Convolutional Neural Network Detection Method

Similar Documents

Publication Publication Date Title
CN108062170A (en) Multi-class human posture recognition method based on convolutional neural networks and intelligent terminal
Liu et al. A facial expression emotion recognition based human-robot interaction system.
CN110610158A (en) A human body posture recognition method and system based on convolution and gated recurrent neural network
Singh et al. A deeply coupled ConvNet for human activity recognition using dynamic and RGB images
CN107886061B (en) Human Behavior Recognition Method and System Based on Multimodal Deep Boltzmann Machine
CN108345846A (en) A kind of Human bodys' response method and identifying system based on convolutional neural networks
CN111723662B (en) Human body posture recognition method based on convolutional neural network
CN109726662A (en) Multi-category Human Pose Recognition Method Based on Convolutional and Recurrent Combination Neural Networks
CN105740773A (en) Deep learning and multi-scale information based behavior identification method
CN109670548A (en) HAR algorithm is inputted based on the more sizes for improving LSTM-CNN
CN109685148A (en) Multi-class human motion recognition method and identifying system
CN103984416A (en) Gesture recognition method based on acceleration sensor
CN106909220A (en) A kind of sight line exchange method suitable for touch-control
CN111753683A (en) A Human Gesture Recognition Method Based on Multi-Expert Convolutional Neural Network
Mutegeki et al. Feature-representation transfer learning for human activity recognition
Tang et al. A hybrid SAE and CNN classifier for motor imagery EEG classification
CN105426961B (en) The method for catching user's intention using Intelligent bracelet and smart mobile phone
CN109711324A (en) Human pose recognition method based on Fourier transform and convolutional neural network
CN115905819A (en) rPPG signal generation method and device based on generation countermeasure network
CN115438691A (en) Small sample gesture recognition method based on wireless signals
Zhang et al. A lightweight network deployed on ARM devices for hand gesture recognition
CN112101094A (en) A body language-based approach to assessing suicide risk
CN114863572A (en) An EMG gesture recognition method based on multi-channel heterogeneous sensors
Xu et al. A study of falling behavior recognition of the elderly based on deep learning
Jhang et al. Sensor based dynamic hand gesture recognition by pairnet

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20180522

RJ01 Rejection of invention patent application after publication