CN110321833A - Human bodys' response method based on convolutional neural networks and Recognition with Recurrent Neural Network - Google Patents

Human bodys' response method based on convolutional neural networks and Recognition with Recurrent Neural Network Download PDF

Info

Publication number
CN110321833A
CN110321833A CN201910580116.XA CN201910580116A CN110321833A CN 110321833 A CN110321833 A CN 110321833A CN 201910580116 A CN201910580116 A CN 201910580116A CN 110321833 A CN110321833 A CN 110321833A
Authority
CN
China
Prior art keywords
recognition
neural network
neural networks
convolutional neural
recurrent neural
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201910580116.XA
Other languages
Chinese (zh)
Other versions
CN110321833B (en
Inventor
谢子凡
陈志�
岳文静
葛宇轩
王多
崔明浩
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing Post and Telecommunication University
Original Assignee
Nanjing Post and Telecommunication University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing Post and Telecommunication University filed Critical Nanjing Post and Telecommunication University
Priority to CN201910580116.XA priority Critical patent/CN110321833B/en
Publication of CN110321833A publication Critical patent/CN110321833A/en
Application granted granted Critical
Publication of CN110321833B publication Critical patent/CN110321833B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2411Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Evolutionary Computation (AREA)
  • Health & Medical Sciences (AREA)
  • General Engineering & Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Computational Linguistics (AREA)
  • Computing Systems (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Molecular Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Psychiatry (AREA)
  • Social Psychology (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

The Human bodys' response method based on convolutional neural networks and Recognition with Recurrent Neural Network that the invention discloses a kind of tracks human body behavior using sensor, collects the 3 dimension coordinate vector groups and rgb video of human synovial in the period.Then it is trained using 3 dimension coordinates of the Recognition with Recurrent Neural Network RNN to human synovial, obtains temporal characteristics vector.Using convolutional neural networks CNN to rgb video training, obtain space-time characteristic vector, last binding time feature vector and space-time characteristic vector simultaneously normalize, and it is fed to the classifier of Linear SVM, use validation data set, the parameter C for finding linear SVM SVM finally obtains a comprehensive identification model.The present invention is able to solve model during model training to the overfitting problem of the classification of motion, while can also effectively improve Human bodys' response efficiency and accuracy.

Description

Human bodys' response method based on convolutional neural networks and Recognition with Recurrent Neural Network
Technical field
The Human bodys' response method based on convolutional neural networks and Recognition with Recurrent Neural Network that the present invention relates to a kind of, belongs to row For the interleaving techniques such as identification, deep learning, machine learning field.
Background technique
An important subject in Activity recognition and calculating sorting-machine visual field in video, video with Track, motion analysis, virtual reality and artificial intelligence interaction extensive use and become computer vision field research hotspot,
Since identical action scene will be different under the conditions ofs different illumination, visual angle and background etc., and different Identical target and movement can also generate apparent difference in characteristic and posture in action scene, even if in constant movement In scene, human action also has a biggish freedom degree, and every kind of identical movement direction, in terms of have it is very big Otherness.In addition, the problems such as partial occlusion and individual difference is all the embodiment of action recognition complexity spatially.Using The method of manual designs is difficult to obtain the substantive characteristics of object from the violent scene of variation, therefore needs to propose more popular Feature extracting method come the problems such as solution by manual feature extracting method bring one-sidedness and blindness.
Current many action recognition algorithms are shallow-layer online learning methods, are had some limitations, such as in training sample Under this limited situation, indicate that the ability of complicated function is limited, and to the generalization ability of complicated classification problem also can therefore by Certain constraint.
Action recognition class problem is considered a kind of classification problem, much is designed to use actuation in classificatory method Identify specifically have using Logistic regression analysis, decision-tree model, Naive Bayes Classifier and support vector machines.This A little methods have pros and cons in practical applications.
In general TV, video image, often 3D video, and at present for used in the Activity recognition in 3D video Technology is not overripened, and most of Human bodys' response system depends on and processing manually is marked to data, then places data into It is identified in model.There is stronger dependence to data, running efficiency of system is low, is not suitable for industrialization and commercialization Demand.
Summary of the invention
Goal of the invention: in order to overcome the deficiencies in the prior art, the present invention provides a kind of based on convolutional neural networks With the Human bodys' response method of Recognition with Recurrent Neural Network, by a set of algorithm, in conjunction with convolutional neural networks to sample time feature Extraction to sample space-time characteristic of extraction and convolutional neural networks, and further combine Linear SVM, training obtains one The model that figure action in video can be judged.
Technical solution: to achieve the above object, the technical solution adopted by the present invention are as follows:
A kind of Human bodys' response method based on convolutional neural networks and Recognition with Recurrent Neural Network, enables a user exist first Fixed position is persistently waved 5 times, using Microsoft Kinect v2 sensor track human body behavior, with 16 milliseconds for when Between step-length, collect the user main 25 joints period Nei 3 dimension coordinate vector groups and in the period RGB regard Frequently.Then using has gating cycle unit GRU as the basic unit of circulation layer and uses bidirectional circulating neural network to people 3 Wei Zuobiaoshuojuji in body joint are trained, and show that human synovial 3 ties up the temporal characteristics vector of coordinate data.Then will Rgb video is divided into the continuous RGB frame with 16 milliseconds for time step, then one convolutional Neural of training on the data set Network obtains the space-time characteristic vector of RGB video, finally combines the defeated of the two Recognition with Recurrent Neural Network and convolutional neural networks Out as a result, being normalized after connection, and it is fed to the classifier of Linear SVM, using validation data set, finds linear support The parameter C of vector machine SVM finally obtains a comprehensive identification model, specifically includes the following steps:
Step 1), user persistently wave 5 times in fixed position, track human body behavior using sensor, when collection is arranged Time step is 16 milliseconds, collects 3 dimension coordinate the V={ (x in the user main 25 joints period Nei1,y1,z1), (x2,y2,z2),...,(x25,y25,z25)};Wherein, x-axis is that human body vertical direction is vertical and positive axis is directing forwardly, y-axis and people Body vertical direction is parallel and positive axis is directed toward head, and z-axis is that human body vertical direction is vertical and positive axis is directed toward on the left of human body;It adopts simultaneously Collect the rgb video of this scene period Nei;
Step 2), using the 3 dimension coordinates in 25 joints of human body in step 1) as the training set of Recognition with Recurrent Neural Network RNN, with It is followed in training set the training one Recognition with Recurrent Neural Network RNN, Recognition with Recurrent Neural Network RNN of Recognition with Recurrent Neural Network RNN using gate Ring element GRU operates the coordinate of input, and there is batch to normalize, the function of random loss, the circulation nerve net Network RNN includes the output layer of two bidirectional valve controlled circulation layers, hiding a full articulamentum and a softmax model, linearly Function, that is, ReLU function is rectified as activation primitive, and has used dropout mechanism that motion feature is mapped to the movement waved Classification obtains trained Recognition with Recurrent Neural Network RNN;
The collected rgb video of step 1) is divided into continuous RGB frame with 16 milliseconds for time step by step 3), and The resolution ratio for adjusting them, as the training set of convolutional neural networks CNN;Convolutional neural networks CNN uses more convolution kernels Convolution operation carried out to the RGB frame stream of input, the convolutional neural networks CNN include convolutional layer, pond layer, full articulamentum and The output layer of one softmax model, using the pre-training small parameter perturbations model on Sports-1M data set, to reduce Degree fitting and training time, obtain trained convolutional neural networks CNN;
The output result of the trained Recognition with Recurrent Neural Network RNN of step 2) is tieed up number of coordinates by step 4) According to temporal characteristics vector, using the output result of the trained convolutional neural networks CNN of step 3) as the space-time of rgb video Feature vector, the connected binding of the space-time characteristic vector that human synovial 3 is tieed up into the temporal characteristics vector sum rgb video of coordinate data It closes, is normalized after connection, and be fed to the classifier of linear SVM SVM, using validation data set, find line Property support vector machines penalty coefficient C, obtain comprehensive identification model;
When identification, the 3 of 25 joints of human body of human body behavior to be identified are acquired using the method in step 1 for step 5) Coordinate and rgb video are tieed up, the 3 dimension coordinates in 25 joints of collected human body are put into the trained circulation nerve of step 2) Temporal characteristics vector is obtained in network RNN, by collected rgb video into the trained convolutional neural networks CNN of step 3) Space-time characteristic vector is obtained, temporal characteristics vector sum space-time characteristic vector is connected into a feature array, and lead after normalizing Enter comprehensive identification model, utilizes the behavior of comprehensive identification model identification people.
Preferred: sensor described in step 1) is Microsoft Kinect v2 sensor.
Preferred: the method that trained Recognition with Recurrent Neural Network RNN is obtained in the step 2) is as follows:
Step 21) is followed using the 3 dimension coordinates in 25 joints of human body in step 1) as the training set of Recognition with Recurrent Neural Network RNN The dimension of the training set of ring neural network RNN is 16 × (25 × 3);
Step 22), the training Recognition with Recurrent Neural Network on the training set of Recognition with Recurrent Neural Network RNN, first passes around two pairs To gating cycle layer, gating cycle unit GRU is used as the basic unit of the circulation layer and uses forward-backward recutrnce nerve net Network, is respectively trained from t=1 to t=T and the input data of t=T to t=1, wherein t indicates the time in the data set Variable, T indicate the time at the last moment in the data set;
Step 23) carries out scene characteristic classification using GRU type Recognition with Recurrent Neural Network RNN, when controlling previous using update door The status information at quarter is brought into the degree in current state, updates door zt=σ (Wz·[ht-1,xt]), wherein ht-1For t-1 The value of moment candidate's memory unit, xtIndicate the data of t moment input, WzIndicate the corresponding power for updating gated data memory unit value Value, ztIndicate that t moment updates the value of door, σ indicates sigmoid activation primitive;
Step 24) is written to t moment candidate's memory unit h using resetting door control how many information of previous statet, Reset door rt=σ (Wr·[ht-1,xt]), wherein ht-1For the value of t-1 moment candidate memory unit, xtIndicate t moment input Data, WrIndicate the corresponding weight for updating gated data memory unit value, rtIndicate the value of t moment resetting door;
Step 25) calculates candidate memory unit value, candidate memory unitWherein htFor t The value of moment candidate's memory unit, ht-1The value of candidate memory unit, x are carved for t-1tIndicate the data of t moment input, W is indicated The weight of current time GRU unit,Indicate the value of candidate memory unit, tanh indicates trigonometric tangential function;
Step 26) calculates current time memory unit state value, t moment memory unitWherein, ⊙ indicates point-by-point product, and memory unit state, which updates, depends on t-1 quarter candidate note Recall the value h of unitt-1With the value of candidate memory unitAnd by update door and resetting door respectively to this two parts factor into Row is adjusted;
Step 27) finally obtains the output y of Recognition with Recurrent Neural Networkt=σ (Wht), it is transmitted using softmax activation primitive Probability, y are construed to full articulamentum, and by outputtIndicate the probability predicted using Recognition with Recurrent Neural Network.
Preferred: the method that the step 3) obtains trained convolutional neural networks CNN is as follows:
The collected rgb video of step 1) is divided into continuous RGB frame with 16 milliseconds for time step by step 31), and The resolution ratio for adjusting them is inputted referred to as c × t × h using the RGB frame stream as the training set of convolutional neural networks CNN × w size, wherein c is port number, and t is time step, and h and w are the height and width of RGB frame;
Step 32), convolutional neural networks CNN receive the parallelization input of video frame, the convolution window for being h to length, volume Product neural network CNN is by convolution kernel to input matrix x1:nConvolution operation is carried out, convolution kernel output valve is ci=f (w xI:i+h-1+ b), whereinFor convolution kernel weight,Indicate w value in real number field, d indicates xiDimension,For Amount of bias, f are activation primitive, xI:i+h-1For the frame vector matrix of a convolution window, time gradient is the frame stream of n, passes through volume Feature vector c=[c after convolution can be obtained in product operation1, c2..., cn-h+1]。
Step 33) extracts a maximum value from each feature vector, has the window of m convolution kernel to obtain feature VectorWherein,For the feature vector that convolutional neural networks extract,It indicates m-th The feature vector of convolution kernel.
Step 34) obtains following formula by softmax function output category result:
Wherein, y indicates the probability predicted using convolutional neural networks,For the regular terms limit of down-sampling layer output System,It is multiplied for corresponding element,For full articulamentum weight matrix,For amount of bias, by under stochastic gradient It drops optimizer and optimizes convolutional neural networks CNN.
Preferred: the method that comprehensive identification model is obtained in the step 4) is as follows:
The trained Recognition with Recurrent Neural Network RNN of step 2) is tieed up coordinate data from human body 3 and extracted by step 41), defeated Result extracts the trained convolutional neural networks CNN of step 3) from rgb video as temporal characteristics vector out, defeated Temporal characteristics vector sum space-time characteristic vector is connected into a feature array as space-time characteristic vector by result out, and is carried out Normalization;
Step 42), it is corresponding with the feature vector group after each normalization using the feature vector group after normalizing as input Specific movement or behavior label as output, be delivered in linear SVM SVM training, Optimized model is for example following Formula:
yiTxi+b0)≥1-ξi
s.t.ξi≥0
I=1,2 ..., N
Wherein, ω indicates that the vector of input, C indicate penalty coefficient, ξiIndicate the Classification Loss for i-th of sample point,
yiIndicate the corresponding action mark of each sample, b0Indicate intercept, N indicates the feature vector sum of input SVM;
Step 43) is collected by training and verifying, finds the optimum value of penalty coefficient C, and comprehensive identification model is obtained.
It is preferred: initial to learn when optimizing convolutional neural networks CNN by stochastic gradient descent optimizer in step 33) Rate is 0.0001, when observing not training progress, learning rate drop by half.
It is preferred: normalized formula in step 41):
Wherein, xiIndicate the element in feature array, x'iThe element in feature vector after indicating normalization.
It is preferred: resolution ratio being adjusted to 320*240 pixel from 1920*1080 pixel in step 31).
The present invention compared with prior art, has the advantages that
The present invention ties up coordinates and rgb video using the 3 of Microsoft Kinect v2 sensor acquisition human synovial, And the temporal characteristics and space-time characteristic of people's behavioral data in video are obtained using convolutional neural networks and Recognition with Recurrent Neural Network, and It is effectively combined, the model for adapting to complex environment and movement, the phase that future passes through one section of video of input are finally obtained Data are answered, the identification of figure action in video just can be quickly and effectively obtained, there is good accuracy and stability, specifically For:
(1) present invention uses double fluid RNN/CNN neural network algorithm, effectively consideration can act continuity to action recognition Influence, can be identified and prediction action in the short period
(2) scene information is combined consideration with action message by the present invention, by scene information as label in behavior Action sequence is matched in database, it is more acurrate to complete the identification to human body behavior.
(3) the invention proposes a kind of effective, practical system structures, and configure user interface mould group and operation note Module is recorded, the stability of Human bodys' response, the concrete application of the convenient invention industrially are improved.
Detailed description of the invention
Fig. 1 is the 3D video actions identification process figure based on convolutional neural networks and Recognition with Recurrent Neural Network
Fig. 2 is based on GRU-RNN neural network schematic diagram.
Fig. 3 is the three-dimensional data sampled point schematic diagram of human synovial.
Specific embodiment
In the following with reference to the drawings and specific embodiments, the present invention is furture elucidated, it should be understood that these examples are merely to illustrate this It invents rather than limits the scope of the invention, after the present invention has been read, those skilled in the art are to of the invention various The modification of equivalent form falls within the application range as defined in the appended claims.
A kind of Human bodys' response method based on convolutional neural networks and Recognition with Recurrent Neural Network, as shown in Figure 1-3, with packet Include following steps:
Step 1), a user persistently wave 5 times in fixed position, use Microsoft Kinect v2 sensor Human body behavior is tracked, is to the sampled point of human synovial using Microsoft Kinect v2 sensor: the base portion of 1- backbone, The middle part of 2- backbone, 3- neck, the head 4-, 5- left shoulder, 6- left hand elbow, 7- left finesse, 8- hand, 9- right shoulder, 10- are right Side ancon, the right side 11- wrist, the 12- right hand, the left stern of 13-, the left knee of 14-, 15- left ankle, 16- left foot, the right stern of 17-, 18- are right Knee, 19- right ankle, 20- right crus of diaphragm, 21- vertebra, 22- left hand point, 24- right hand point, the right thumb of 25-.Time step is set when collection A length of 16 milliseconds, collect 3 dimension coordinate the V={ (x in the user main 25 joints period Nei1,y1,z1),(x2,y2, z2),...,(x25,y25,z25)};Wherein, x-axis is that human body vertical direction is vertical and positive axis is directing forwardly, and y-axis is vertical with human body Direction is parallel and positive axis is directed toward head, and z-axis is that human body vertical direction is vertical and positive axis is directed toward on the left of human body;When acquiring this simultaneously Between in section this scene rgb video;
Step 2), using the 3 dimension coordinates in 25 joints of human body in step 1) as the training set of Recognition with Recurrent Neural Network RNN, with It is followed in training set the training one Recognition with Recurrent Neural Network RNN, Recognition with Recurrent Neural Network RNN of Recognition with Recurrent Neural Network RNN using gate Ring element GRU operates the coordinate of input, and there is batch to normalize, the function of random loss, the circulation nerve net Network RNN includes the output layer of two bidirectional valve controlled circulation layers, hiding a full articulamentum and a softmax model, linearly Function, that is, ReLU function is rectified as activation primitive, and has used dropout mechanism that motion feature is mapped to the movement waved Classification obtains trained Recognition with Recurrent Neural Network RNN;
Step 21) is followed using the 3 dimension coordinates in 25 joints of human body in step 1) as the training set of Recognition with Recurrent Neural Network RNN The dimension of the training set of ring neural network RNN is 16 (time steps) × (25 × 3) (3 dimension joint coordinates);
Step 22), the training Recognition with Recurrent Neural Network on the training set of Recognition with Recurrent Neural Network RNN, first passes around two pairs To gating cycle layer, gating cycle unit GRU is used as the basic unit of the circulation layer and uses forward-backward recutrnce nerve net Network, is respectively trained from t=1 to t=T and the input data of t=T to t=1, wherein t indicates the time in the data set Variable, T indicate the time at the last moment in the data set;
Step 23) carries out scene characteristic classification using GRU type Recognition with Recurrent Neural Network RNN, when controlling previous using update door The status information at quarter is brought into the degree in current state, updates door zt=σ (Wz·[ht-1,xt]), wherein ht-1For t-1 The value of moment candidate's memory unit, xtIndicate the data of t moment input, WzIndicate the corresponding power for updating gated data memory unit value Value, ztIndicate that t moment updates the value of door, σ indicates sigmoid activation primitive;
Step 24) is written to t moment candidate's memory unit h using resetting door control how many information of previous statet, Reset door rt=σ (Wr·[ht-1,xt]), wherein ht-1For the value of t-1 moment candidate memory unit, xtIndicate t moment input Data, WrIndicate the corresponding weight for updating gated data memory unit value, rtIndicate the value of t moment resetting door;
Step 25) calculates candidate memory unit value, candidate memory unitWherein htFor t The value of moment candidate's memory unit, ht-1The value of candidate memory unit, x are carved for t-1tIndicate the data of t moment input, W is indicated The weight of current time GRU unit,Indicate the value of candidate memory unit, tanh indicates trigonometric tangential function;
Step 26) calculates current time memory unit state value, t moment memory unitWherein, ⊙ indicates point-by-point product, and memory unit state, which updates, depends on t-1 quarter candidate note Recall the value h of unitt-1With the value of candidate memory unitAnd by update door and resetting door respectively to this two parts factor into Row is adjusted;
Step 27) finally obtains the output y of Recognition with Recurrent Neural Networkt=σ (Wht), it is transmitted using softmax activation primitive Probability, y are construed to full articulamentum, and by outputtIndicate the probability predicted using Recognition with Recurrent Neural Network.
The collected rgb video of step 1) is divided into continuous RGB frame with 16 milliseconds for time step by step 3), and The resolution ratio for adjusting them is adjusted to 320*240 pixel from 1920*1080 pixel, as the instruction of convolutional neural networks CNN Practice collection;Convolutional neural networks CNN carries out convolution operation, the convolutional neural networks using the RGB frame stream of multireel product verification input CNN includes the output layer of convolutional layer, pond layer, full articulamentum and a softmax model, using on Sports-1M data set The pre-training small parameter perturbations models obtain trained convolutional neural networks CNN to reduce overfitting and training time;
The method that the step 3) obtains trained convolutional neural networks CNN is as follows:
The collected rgb video of step 1) is divided into continuous RGB frame with 16 milliseconds for time step by step 31), and The resolution ratio for adjusting them is inputted referred to as c × t × h using the RGB frame stream as the training set of convolutional neural networks CNN × w size, wherein c is port number, and t is time step, and h and w are the height and width of RGB frame;
Step 32), convolutional neural networks CNN receive the parallelization input of video frame, the convolution window for being h to length, volume Product neural network CNN is by convolution kernel to input matrix x1:nConvolution operation is carried out, convolution kernel output valve is ci=f (w xI:i+h-1+ b), whereinFor convolution kernel weight,Indicate w value in real number field, d indicates xiDimension,For Amount of bias, f are activation primitive, xI:i+h-1For the frame vector matrix of a convolution window, time gradient is the frame stream of n, passes through volume Feature vector c=[c after convolution can be obtained in product operation1, c2..., cn-h+1]。
Step 33) extracts a maximum value from each feature vector, has the window of m convolution kernel to obtain feature VectorWherein,For the feature vector that convolutional neural networks extract,It indicates m-th The feature vector of convolution kernel.
Step 34) obtains following formula by softmax function output category result:
Wherein, y indicates the probability predicted using convolutional neural networks,For the regular terms limit of down-sampling layer output System,It is multiplied for corresponding element,For full articulamentum weight matrix,For amount of bias, pass through stochastic gradient descent Optimizer optimizes convolutional neural networks CNN.
The output result of the trained Recognition with Recurrent Neural Network RNN of step 2) is tieed up number of coordinates by step 4) According to temporal characteristics vector, using the output result of the trained convolutional neural networks CNN of step 3) as the space-time of rgb video Feature vector, the connected binding of the space-time characteristic vector that human synovial 3 is tieed up into the temporal characteristics vector sum rgb video of coordinate data A feature array is synthesized, is normalized after connection, and is fed to the classifier of linear SVM SVM, uses verifying Data set, finds the penalty coefficient C of linear SVM SVM, and penalty coefficient represents the degrees of tolerance to error, obtains comprehensive Close identification model;
The method that comprehensive identification model is obtained in the step 4) is as follows:
The trained Recognition with Recurrent Neural Network RNN of step 2) is tieed up coordinate data from human body 3 and extracted by step 41), defeated Result mentions the trained convolutional neural networks CNN of step 3) from rgb video as temporal characteristics vector (600 dimension) out It takes, temporal characteristics vector sum space-time characteristic vector is connected into one as space-time characteristic vector (4096 dimension) by output result Feature array (4696 dimension), and be normalized;
Normalized formula:
Wherein, xiIndicate the element in feature array, x'iThe element in feature vector after indicating normalization.
Step 42), it is corresponding with the feature vector group after each normalization using the feature vector group after normalizing as input Specific movement or behavior label as output, be delivered in linear SVM SVM training, Optimized model is for example following Formula:
yiTxi+b0)≥1-ξi
s.t.ξi≥0
I=1,2 ..., N
Wherein, ω indicates that the vector of input, C indicate penalty coefficient, ξiIndicate the Classification Loss for i-th of sample point,
yiIndicate the corresponding action mark of each sample, b0Indicate intercept, N indicates the feature vector sum of input SVM;
Step 43) is collected by training and verifying, finds the optimum value of penalty coefficient C, obtains comprehensive identification model, finally With the action classification in model verifying human body.
When identification, the 3 of 25 joints of human body of human body behavior to be identified are acquired using the method in step 1 for step 5) Coordinate and rgb video are tieed up, the 3 dimension coordinates in 25 joints of collected human body are put into the trained circulation nerve of step 2) Temporal characteristics vector is obtained in network RNN, by collected rgb video into the trained convolutional neural networks CNN of step 3) Space-time characteristic vector is obtained, temporal characteristics vector sum space-time characteristic vector is connected into a feature array, and lead after normalizing Enter comprehensive identification model, utilizes the behavior of comprehensive identification model identification people.
Emulation:
One user persistently waves 5 times in fixed position, tracks human body using Microsoft Kinect v2 sensor Behavior, setting time step is 16 milliseconds when collection, and coordinates are tieed up in collect the user main 25 joints period Nei 3 Vector Groups V={ (x1,y1,z1),(x2,y2,z2),...,(x25,y25,z25)};There are 25 3D coordinates in each joint.It is dynamic to handle this Making sample dimension is 16 (time steps) × (25 × 3) (3 dimension joint coordinates), using the sample action as Recognition with Recurrent Neural Network Training set.
Using multiple GPU training Recognition with Recurrent Neural Network, 0.001 is set by learning rate, rate of decay 0.9, so Network is trained to be used for single-layer model from the beginning using 1000 sequences of small lot afterwards;It is used for using 650 sequences of small lot Bilayer model.For all RNN networks, each unidirectional ply uses 300 neurons, and bidirectional layer neuronal quantity doubles, makes It is lost with 75% holding probability, eventually passes through and be fully connected layer, obtain the temporal characteristics of human body 3D joint coordinates data Vector (600 dimension).
Rgb video is subsequently introduced for convolutional neural networks training, 3D-CNN model is used in Caffe, is regarded from RGB Frame is extracted in frequency, and cuts them from 1920*1080 pixel to 320*240 pixel, and time step is 16 milliseconds, we are by CNN The input of model is known as the size of c*t*h*w, and wherein c is the quantity in channel, and t is time step, and h and w are RGB frame respectively Height and width.Network is using video clipping as input, and the movement label waved for data markers.RGB frame will then be inputted Size be adjusted to 128*171 pixel resolution, the size of input becomes 3 × 16 × 128 × 171, uses stochastic gradient descent Optimizer carries out network verification to network, and initial learning rate is 0.0001.Then, when training progress is not observed, study Rate reduces half.Eventually pass through the space-time characteristic vector (4096 dimension) that rgb video is obtained after being fully connected layer.
For Fusion Features, the temporal characteristics vector that we will extract from RNN, with the space-time characteristic vector extracted from CNN It connects into feature array (4696 dimension) and carries out L2 normalization.Finally, we pass through training, verifying and test segmentation pair RNN/CNN function is standardized.We find the optimal of the parameter C of Linear SVM model using training and verifying division Value.This collective model is used to verify the accuracy of action recognition.
The present invention is able to solve model during model training to the overfitting problem of the classification of motion, while can also be effective Improve Human bodys' response efficiency and accuracy.
The above is only a preferred embodiment of the present invention, it should be pointed out that: for the ordinary skill people of the art For member, various improvements and modifications may be made without departing from the principle of the present invention, these improvements and modifications It should be regarded as protection scope of the present invention.

Claims (8)

1. a kind of Human bodys' response method based on convolutional neural networks and Recognition with Recurrent Neural Network, which is characterized in that with including Following steps:
Step 1), user persistently wave 5 times in fixed position, track human body behavior using sensor, the time is arranged in when collection Step-length is 16 milliseconds, collects 3 dimension coordinate the V={ (x in the user main 25 joints period Nei1,y1,z1),(x2,y2, z2),...,(x25,y25,z25)};Wherein, x-axis is that human body vertical direction is vertical and positive axis is directing forwardly, and y-axis and human body are square vertically It is directed toward head to parallel and positive axis, z-axis is that human body vertical direction is vertical and positive axis is directed toward on the left of human body;The period is acquired simultaneously The rgb video of this interior scene;
Step 2) is followed using the 3 dimension coordinates in 25 joints of human body in step 1) as the training set of Recognition with Recurrent Neural Network RNN with this Gating cycle list is used in training set the training one Recognition with Recurrent Neural Network RNN, Recognition with Recurrent Neural Network RNN of ring neural network RNN First GRU operates the coordinate of input, and there is batch to normalize, the function of random loss, the Recognition with Recurrent Neural Network RNN Output layer including two bidirectional valve controlled circulation layers, hiding a full articulamentum and a softmax model, line rectification letter Number is ReLU function as activation primitive, and has used dropout mechanism that motion feature is mapped to the action classification waved, and is obtained To trained Recognition with Recurrent Neural Network RNN;
The collected rgb video of step 1) is divided into continuous RGB frame with 16 milliseconds for time step, and adjusted by step 3) Their resolution ratio, as the training set of convolutional neural networks CNN;Convolutional neural networks CNN is defeated using multireel product verification The RGB frame stream that enters carries out convolution operation, and the convolutional neural networks CNN includes convolutional layer, pond layer, full articulamentum and one The output layer of softmax model, it is excessively quasi- to reduce using the pre-training small parameter perturbations model on Sports-1M data set Conjunction and training time, obtain trained convolutional neural networks CNN;
The output result of the trained Recognition with Recurrent Neural Network RNN of step 2) is tieed up coordinate data by step 4) Temporal characteristics vector, using the output result of the trained convolutional neural networks CNN of step 3) as the space-time characteristic of rgb video to The space-time characteristic vector that human synovial 3 ties up the temporal characteristics vector sum rgb video of coordinate data is connected combinations by amount, connects Normalized afterwards, and be fed to the classifier of linear SVM SVM, using validation data set, find linearly support to The penalty coefficient C of amount machine SVM obtains comprehensive identification model;
Step 5) when identification, is sat using 3 dimensions that the method in step 1 acquires 25 joints of human body of human body behavior to be identified The 3 dimension coordinates in 25 joints of collected human body are put into the trained Recognition with Recurrent Neural Network of step 2) by mark and rgb video Temporal characteristics vector is obtained in RNN, and collected rgb video is obtained into the trained convolutional neural networks CNN of step 3) Temporal characteristics vector sum space-time characteristic vector is connected into a feature array, and imported after normalizing comprehensive by space-time characteristic vector Identification model is closed, the behavior of comprehensive identification model identification people is utilized.
2. the Human bodys' response method based on convolutional neural networks and Recognition with Recurrent Neural Network according to claim 1, special Sign is: sensor described in step 1) is Microsoft Kinect v2 sensor.
3. the Human bodys' response method based on convolutional neural networks and Recognition with Recurrent Neural Network according to claim 2, special Sign is: the method that trained Recognition with Recurrent Neural Network RNN is obtained in the step 2) is as follows:
Step 21), using the 3 dimension coordinates in 25 joints of human body in step 1) as the training set of Recognition with Recurrent Neural Network RNN, circulation mind The dimension of training set through network RNN is 16 × (25 × 3);
Step 22), training Recognition with Recurrent Neural Network, first passes around two bidirectional gates on the training set of Recognition with Recurrent Neural Network RNN Circulation layer is controlled, gating cycle unit GRU is used as the basic unit of the circulation layer and uses forward-backward recutrnce neural network, point It Xun Lian not be from t=1 to t=T and the input data of t=T to t=1, wherein t indicates the time variable in the data set, T table Show the time at the last moment in the data set;
Step 23) carries out scene characteristic classification using GRU type Recognition with Recurrent Neural Network RNN, controls previous moment using door is updated Status information is brought into the degree in current state, updates door zt=σ (Wz·[ht-1,xt]), wherein ht-1It is waited for the t-1 moment Select the value of memory unit, xtIndicate the data of t moment input, WzIndicate the corresponding weight for updating gated data memory unit value, ztTable Show that t moment updates the value of door, σ indicates sigmoid activation primitive;
Step 24) is written to t moment candidate's memory unit h using resetting door control how many information of previous statet, reset door rt=σ (Wr·[ht-1,xt]), wherein ht-1For the value of t-1 moment candidate memory unit, xtIndicate the data of t moment input, Wr Indicate the corresponding weight for updating gated data memory unit value, rtIndicate the value of t moment resetting door;
Step 25) calculates candidate memory unit value, candidate memory unitWherein htFor t moment The value of candidate memory unit, ht-1The value of candidate memory unit, x are carved for t-1tThe data of t moment input are indicated, when W indicates current The weight of GRU unit is carved,Indicate the value of candidate memory unit, tanh indicates trigonometric tangential function;
Step 26) calculates current time memory unit state value, t moment memory unitIts s In, ⊙ indicates point-by-point product, and memory unit state, which updates, depends on the value h that t-1 carves candidate memory unitt-1It is single with candidate's memory The value of memberAnd this two parts factor is adjusted respectively by updating door and resetting door;
Step 27) finally obtains the output y of Recognition with Recurrent Neural Networkt=σ (Wht), it is transmitted to entirely using softmax activation primitive Articulamentum, and output is construed to probability, ytIndicate the probability predicted using Recognition with Recurrent Neural Network.
4. the Human bodys' response method based on convolutional neural networks and Recognition with Recurrent Neural Network according to claim 3, special Sign is: the method that the step 3) obtains trained convolutional neural networks CNN is as follows:
The collected rgb video of step 1) is divided into continuous RGB frame with 16 milliseconds for time step, and adjusted by step 31) It is big to be inputted referred to as c × t × h × w using the RGB frame stream as the training set of convolutional neural networks CNN for their resolution ratio It is small, wherein c is port number, and t is time step, and h and w are the height and width of RGB frame;
Step 32), convolutional neural networks CNN receive the parallelization input of video frame, the convolution window for being h to length, convolution mind Through network C NN by convolution kernel to input matrix x1:nConvolution operation is carried out, convolution kernel output valve is ci=f (wxi:i+h-1+ b), WhereinFor convolution kernel weight,Indicate w value in real number field, d indicates xiDimension,For amount of bias, f is sharp Function living, xi:i+h-1For the frame vector matrix of a convolution window, time gradient is the frame stream of n, available by convolution operation Feature vector c=[c after convolution1,c2,…,cn-h+1];
Step 33) extracts a maximum value from each feature vector, has the window of m convolution kernel to obtain feature vectorWherein,For the feature vector that convolutional neural networks extract,Indicate m-th of convolution kernel Feature vector;
Step 34) obtains following formula by softmax function output category result:
Wherein, y indicates the probability predicted using convolutional neural networks,Regular terms for the output of down-sampling layer limits,For Corresponding element is multiplied,For full articulamentum weight matrix,For amount of bias, pass through stochastic gradient descent optimizer Optimize convolutional neural networks CNN.
5. the Human bodys' response method based on convolutional neural networks and Recognition with Recurrent Neural Network according to claim 4, special Sign is: the method that comprehensive identification model is obtained in the step 4) is as follows:
The trained Recognition with Recurrent Neural Network RNN of step 2) is tieed up coordinate data from human body 3 and extracted by step 41), output knot Fruit extracts as temporal characteristics vector, by the trained convolutional neural networks CNN of step 3) from rgb video, output knot Fruit connects into a feature array as space-time characteristic vector, by temporal characteristics vector sum space-time characteristic vector, and carries out normalizing Change;
Step 42), using the feature vector group after normalizing as input, with the corresponding tool of feature vector group after each normalization Body movement or behavior label are delivered to training in linear SVM SVM, Optimized model such as following formula as output:
yiTxi+b0)≥1-ξi
s.t.ξi≥0
I=1,2 ..., N
Wherein, ω indicates that the vector of input, C indicate penalty coefficient, ξiIndicate the Classification Loss for i-th of sample point, yiIt indicates The corresponding action mark of each sample, b0Indicate intercept, N indicates the feature vector sum of input SVM;
Step 43) is collected by training and verifying, finds the optimum value of penalty coefficient C, and comprehensive identification model is obtained.
6. the Human bodys' response method based on convolutional neural networks and Recognition with Recurrent Neural Network according to claim 5, special Sign is: when optimizing convolutional neural networks CNN by stochastic gradient descent optimizer in step 33), initial learning rate is 0.0001, when observing not training progress, learning rate drop by half.
7. the Human bodys' response method based on convolutional neural networks and Recognition with Recurrent Neural Network according to claim 6, special Sign is: normalized formula in step 41):
Wherein, xiIndicate the element in feature array, x'iThe element in feature vector after indicating normalization.
8. the Human bodys' response method based on convolutional neural networks and Recognition with Recurrent Neural Network according to claim 7, special Sign is: resolution ratio being adjusted to 320*240 pixel from 1920*1080 pixel in step 31).
CN201910580116.XA 2019-06-28 2019-06-28 Human body behavior identification method based on convolutional neural network and cyclic neural network Active CN110321833B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910580116.XA CN110321833B (en) 2019-06-28 2019-06-28 Human body behavior identification method based on convolutional neural network and cyclic neural network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910580116.XA CN110321833B (en) 2019-06-28 2019-06-28 Human body behavior identification method based on convolutional neural network and cyclic neural network

Publications (2)

Publication Number Publication Date
CN110321833A true CN110321833A (en) 2019-10-11
CN110321833B CN110321833B (en) 2022-05-20

Family

ID=68121381

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910580116.XA Active CN110321833B (en) 2019-06-28 2019-06-28 Human body behavior identification method based on convolutional neural network and cyclic neural network

Country Status (1)

Country Link
CN (1) CN110321833B (en)

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110880172A (en) * 2019-11-12 2020-03-13 中山大学 Video face tampering detection method and system based on cyclic convolution neural network
CN111079928A (en) * 2019-12-14 2020-04-28 大连大学 Method for predicting human motion by using recurrent neural network based on antagonistic learning
CN111459283A (en) * 2020-04-07 2020-07-28 电子科技大学 Man-machine interaction implementation method integrating artificial intelligence and Web3D
CN111597881A (en) * 2020-04-03 2020-08-28 浙江工业大学 Human body complex behavior identification method based on data separation multi-scale feature combination
CN111681321A (en) * 2020-06-05 2020-09-18 大连大学 Method for synthesizing three-dimensional human motion by using recurrent neural network based on layered learning
CN111860269A (en) * 2020-07-13 2020-10-30 南京航空航天大学 Multi-feature fusion tandem RNN structure and pedestrian prediction method
CN111914638A (en) * 2020-06-29 2020-11-10 南京邮电大学 Character action recognition method based on improved long-term recursive deep convolution model
CN112232489A (en) * 2020-10-26 2021-01-15 南京明德产业互联网研究院有限公司 Method and device for gating cycle network and method and device for link prediction
CN112488014A (en) * 2020-12-04 2021-03-12 重庆邮电大学 Video prediction method based on gated cyclic unit
CN112906509A (en) * 2021-01-28 2021-06-04 浙江省隧道工程集团有限公司 Method and system for identifying operation state of water delivery tunnel excavator
CN112906383A (en) * 2021-02-05 2021-06-04 成都信息工程大学 Integrated adaptive water army identification method based on incremental learning
CN113111756A (en) * 2021-04-02 2021-07-13 浙江工业大学 Human body tumble identification method based on human body skeleton key points and long-term and short-term memory artificial neural network
CN113378638A (en) * 2021-05-11 2021-09-10 大连海事大学 Human body joint point detection and D-GRU network-based abnormal behavior identification method for wheelers
CN113568819A (en) * 2021-01-31 2021-10-29 腾讯科技(深圳)有限公司 Abnormal data detection method and device, computer readable medium and electronic equipment
CN116911955A (en) * 2023-09-12 2023-10-20 深圳须弥云图空间科技有限公司 Training method and device for target recommendation model

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107943967A (en) * 2017-11-28 2018-04-20 华南理工大学 Algorithm of documents categorization based on multi-angle convolutional neural networks and Recognition with Recurrent Neural Network
CN108280406A (en) * 2017-12-30 2018-07-13 广州海昇计算机科技有限公司 A kind of Activity recognition method, system and device based on segmentation double-stream digestion
AU2018101512A4 (en) * 2018-10-11 2018-11-15 Dong, Xun Miss A comprehensive stock trend predicting method based on neural networks
CN109117701A (en) * 2018-06-05 2019-01-01 东南大学 Pedestrian's intension recognizing method based on picture scroll product

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107943967A (en) * 2017-11-28 2018-04-20 华南理工大学 Algorithm of documents categorization based on multi-angle convolutional neural networks and Recognition with Recurrent Neural Network
CN108280406A (en) * 2017-12-30 2018-07-13 广州海昇计算机科技有限公司 A kind of Activity recognition method, system and device based on segmentation double-stream digestion
CN109117701A (en) * 2018-06-05 2019-01-01 东南大学 Pedestrian's intension recognizing method based on picture scroll product
AU2018101512A4 (en) * 2018-10-11 2018-11-15 Dong, Xun Miss A comprehensive stock trend predicting method based on neural networks

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
BHARGAVI SUVARNAM等: "Combination of CNN-GRU Model to Recognize Characters of a License Plate number without Segmentation", 《2019 5TH INTERNATIONAL CONFERENCE ON ADVANCED COMPUTING & COMMUNICATION SYSTEMS (ICACCS)》 *
尹良亮等: "基于注意力机制和BGRU网络的文本情感分析方法研究", 《无线互联科技》 *
杨天明等: "基于视频深度学习的时空双流人物动作识别模型", 《计算机应用》 *
马静: "基于姿态和骨架信息的行为识别方法研究与实现", 《中国优秀硕士学位论文全文数据库信息科技辑》 *

Cited By (24)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110880172A (en) * 2019-11-12 2020-03-13 中山大学 Video face tampering detection method and system based on cyclic convolution neural network
CN111079928A (en) * 2019-12-14 2020-04-28 大连大学 Method for predicting human motion by using recurrent neural network based on antagonistic learning
CN111079928B (en) * 2019-12-14 2023-07-07 大连大学 Method for predicting human body movement by using circulating neural network based on countermeasure learning
CN111597881A (en) * 2020-04-03 2020-08-28 浙江工业大学 Human body complex behavior identification method based on data separation multi-scale feature combination
CN111597881B (en) * 2020-04-03 2022-04-05 浙江工业大学 Human body complex behavior identification method based on data separation multi-scale feature combination
CN111459283A (en) * 2020-04-07 2020-07-28 电子科技大学 Man-machine interaction implementation method integrating artificial intelligence and Web3D
CN111681321A (en) * 2020-06-05 2020-09-18 大连大学 Method for synthesizing three-dimensional human motion by using recurrent neural network based on layered learning
CN111681321B (en) * 2020-06-05 2023-07-04 大连大学 Method for synthesizing three-dimensional human motion by using cyclic neural network based on layered learning
CN111914638A (en) * 2020-06-29 2020-11-10 南京邮电大学 Character action recognition method based on improved long-term recursive deep convolution model
CN111914638B (en) * 2020-06-29 2022-08-12 南京邮电大学 Character action recognition method based on improved long-term recursive deep convolution model
CN111860269A (en) * 2020-07-13 2020-10-30 南京航空航天大学 Multi-feature fusion tandem RNN structure and pedestrian prediction method
CN111860269B (en) * 2020-07-13 2024-04-16 南京航空航天大学 Multi-feature fusion series RNN structure and pedestrian prediction method
CN112232489A (en) * 2020-10-26 2021-01-15 南京明德产业互联网研究院有限公司 Method and device for gating cycle network and method and device for link prediction
CN112488014A (en) * 2020-12-04 2021-03-12 重庆邮电大学 Video prediction method based on gated cyclic unit
CN112906509A (en) * 2021-01-28 2021-06-04 浙江省隧道工程集团有限公司 Method and system for identifying operation state of water delivery tunnel excavator
CN113568819A (en) * 2021-01-31 2021-10-29 腾讯科技(深圳)有限公司 Abnormal data detection method and device, computer readable medium and electronic equipment
CN113568819B (en) * 2021-01-31 2024-04-16 腾讯科技(深圳)有限公司 Abnormal data detection method, device, computer readable medium and electronic equipment
CN112906383A (en) * 2021-02-05 2021-06-04 成都信息工程大学 Integrated adaptive water army identification method based on incremental learning
CN113111756A (en) * 2021-04-02 2021-07-13 浙江工业大学 Human body tumble identification method based on human body skeleton key points and long-term and short-term memory artificial neural network
CN113111756B (en) * 2021-04-02 2024-05-03 浙江工业大学 Human body fall recognition method based on human body skeleton key points and long-short-term memory artificial neural network
CN113378638A (en) * 2021-05-11 2021-09-10 大连海事大学 Human body joint point detection and D-GRU network-based abnormal behavior identification method for wheelers
CN113378638B (en) * 2021-05-11 2023-12-22 大连海事大学 Method for identifying abnormal behavior of turbine operator based on human body joint point detection and D-GRU network
CN116911955B (en) * 2023-09-12 2024-01-05 深圳须弥云图空间科技有限公司 Training method and device for target recommendation model
CN116911955A (en) * 2023-09-12 2023-10-20 深圳须弥云图空间科技有限公司 Training method and device for target recommendation model

Also Published As

Publication number Publication date
CN110321833B (en) 2022-05-20

Similar Documents

Publication Publication Date Title
CN110321833A (en) Human bodys' response method based on convolutional neural networks and Recognition with Recurrent Neural Network
Zhang et al. Dynamic hand gesture recognition based on short-term sampling neural networks
CN103268495B (en) Human body behavior modeling recognition methods based on priori knowledge cluster in computer system
CN101777116B (en) Method for analyzing facial expressions on basis of motion tracking
CN108664947A (en) A kind of fatigue driving method for early warning based on Expression Recognition
CN106096535A (en) A kind of face verification method based on bilinearity associating CNN
CN110232412B (en) Human gait prediction method based on multi-mode deep learning
CN108681725A (en) A kind of weighting sparse representation face identification method
CN109389045A (en) Micro- expression recognition method and device based on mixing space-time convolution model
Vakanski et al. Mathematical modeling and evaluation of human motions in physical therapy using mixture density neural networks
CN106909938A (en) Viewing angle independence Activity recognition method based on deep learning network
Li et al. Meta auxiliary learning for facial action unit detection
CN111382699A (en) Dynamic gesture recognition method based on particle swarm optimization LSTM algorithm
CN110096976A (en) Human behavior micro-Doppler classification method based on sparse migration network
Wu et al. Improving NeuCube spiking neural network for EEG-based pattern recognition using transfer learning
CN111062245A (en) Locomotive driver fatigue state monitoring method based on upper body posture
Wu et al. An unsupervised real-time framework of human pose tracking from range image sequences
Khan Human activity analysis in visual surveillance and healthcare
Li et al. A shortcut enhanced LSTM-GCN network for multi-sensor based human motion tracking
CN105046193B (en) A kind of human motion recognition method based on fusion rarefaction representation matrix
CN106096598A (en) A kind of method and device utilizing degree of depth related neural network model to identify human face expression
CN114191797B (en) Free skiing intelligent training system
AU2021100829A4 (en) A Face Recognition Method Based On Bidirectional 2DPCA And Cascaded Feedforward Neural Network
Wang et al. Spatial-temporal feature representation learning for facial fatigue detection
Mahajan et al. Classification of emotions using a 2-channel convolution neural network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant