CN110956141B - Human body continuous action rapid analysis method based on local recognition - Google Patents

Human body continuous action rapid analysis method based on local recognition Download PDF

Info

Publication number
CN110956141B
CN110956141B CN201911216130.8A CN201911216130A CN110956141B CN 110956141 B CN110956141 B CN 110956141B CN 201911216130 A CN201911216130 A CN 201911216130A CN 110956141 B CN110956141 B CN 110956141B
Authority
CN
China
Prior art keywords
coordinate
layer
score
human body
video
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201911216130.8A
Other languages
Chinese (zh)
Other versions
CN110956141A (en
Inventor
赵红领
李润知
崔莉亚
刘皓东
王菁
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhengzhou University
Original Assignee
Zhengzhou University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhengzhou University filed Critical Zhengzhou University
Priority to CN201911216130.8A priority Critical patent/CN110956141B/en
Publication of CN110956141A publication Critical patent/CN110956141A/en
Application granted granted Critical
Publication of CN110956141B publication Critical patent/CN110956141B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components

Abstract

The invention discloses a method for quickly analyzing human body continuous actions based on local identification, which mainly comprises rope skipping video acquisition, video data preprocessing, coordinate point acquisition, coordinate point windowing repeated identification, feature sequence construction, regression prediction model establishment and model calling, wherein video stream data in the process of double-foot jumping during the front face of a human body is acquired through a mobile phone, the basic information of a target object, including name, sex, age, height, weight and made action name, is recorded and stored, and a regression prediction model is constructed by utilizing a deep neural network.

Description

Human body continuous action rapid analysis method based on local recognition
Technical Field
The invention relates to the technical field of human motion action analysis, in particular to a human body continuous action rapid analysis method based on local recognition.
Background
With the development and application of computer technology and machine learning in recent years, video-based time series human motion analysis technology is rapidly emerging and has gained wide attention. At present, human motion analysis based on video time sequences is still a very challenging subject in computer vision, relates to multiple subjects such as image processing, pattern recognition and machine learning, and has wide application prospects in the fields of intelligent monitoring, man-machine interaction, rehabilitation exercise, physical training evaluation and the like.
The traditional neural network is easily misled by irrelevant information in the image when processing an input image with fuzzy characteristics and large ambiguity, and is difficult to ensure higher recognition rate.
The existing human behavior recognition algorithm based on videos has the problems of high complexity, poor robustness, low precision and the like. In addition, research work on the aspect of regression prediction analysis of human body actions is less, so the invention provides the human body action analysis method based on time series regression prediction, which has high robustness and stable time overhead, and has important significance for human body action analysis modeling, action quality evaluation and potential mining of sporters.
Disclosure of Invention
In view of the above situation, in order to overcome the defects of the prior art, the present invention aims to provide a method for rapidly analyzing human body continuous actions based on local recognition, which solves the problem of low accuracy of long-time regression prediction of actions in the exercise training process.
The technical scheme for solving the problem is that the method for rapidly analyzing the continuous actions of the human body based on local recognition mainly comprises rope skipping video acquisition, video data preprocessing, coordinate point acquisition, coordinate point windowing re-recognition, feature sequence construction, regression prediction model establishment and model calling, and specifically comprises the following steps:
step 1: collecting video stream data of the human body in the process of shaking the front face of the human body and jumping with two feet through a mobile phone, and recording and storing basic information of a target object, wherein the basic information comprises name, sex, age, height, weight and name of action;
step 2: preprocessing video data, and performing posture estimation on a human body of each frame in a video to obtain a key point coordinate, wherein the method comprises the following steps of:
a1, converting video data shot by different mobile phones into a uniform scale, setting the height of a video to be 530cm and the width to be 460cm;
step A2: obtaining the coordinate positions of 14 joint points of the nose, the neck, the right shoulder, the right elbow, the right wrist, the left shoulder, the left elbow, the left wrist, the right hip, the right knee, the right ankle, the left hip, the left knee and the left ankle of each frame of human body in the video by utilizing an Open-position method, wherein the coordinate is represented as CP i =(x i ,y i ) Wherein i is from 1 to 14, wherein (x) i ,y i ) Coordinates of key points of the human body;
step A3: open-pos uses the Gaussian distance between predicted keypoints and true values to define confidence in keypoints and normalizes the confidence to [0,1]Is defined herein as the score of the keypoint score and results in output results inputs i =(x i ,y i ,score i );
Step A4: the scores of the 14 key points were averaged and standard deviations, and the sum of the average and standard deviation was used as the score result for the overall key point,
Figure BDA0002299567150000021
Figure BDA0002299567150000022
score general (1) =score Average out +score Standard deviation of
And 3, step 3: windowing the low predicted positions of the key points by using an image windowing technology and re-identifying the low predicted positions of the key points, and improving the prediction accuracy of the key points by using global information and local parts, wherein the method comprises the following steps of:
step B1, obtaining inputs by utilizing Open-pos i =(x i ,y i ,score i ) Setting a threshold th for the score, and finding out key points smaller than the threshold th;
b2, windowing the key points smaller than the threshold around the key points, and putting the image frames in the windows into the Open-pos network with the input modified;
b3, updating the coordinates of the key points obtained in the local frame by using the global information;
and 4, step 4: in order to further improve the robustness of the algorithm to shooting angles, target distance and recording process jitter factors, the key points are respectively set as the left hip C 1 Right hip C 11 Neck C 8 Performing coordinate conversion by taking the gravity center of the three coordinate points as an origin to obtain data
Figure BDA0002299567150000031
Wherein
Figure BDA0002299567150000032
Defining the origin of coordinates for the transformed relative coordinates, i.e., C 1 、C 11 、C 8 Center of gravity of three points C 0 =(cx 0 ,cy 0 ) Wherein
Figure BDA0002299567150000033
Figure BDA0002299567150000034
Updating all coordinate points by taking the original point coordinates as reference;
and 5: accumulating the coordinate matrix inputs obtained by each frame to obtain an accumulated coordinate matrix of each video segment, and performing window segmentation on the accumulated coordinate matrix by using a sliding window, namely setting the length of the sliding window to be the accumulated coordinate and the score obtained by k frames, and setting the step length to be 1;
step 6: a regression prediction model is constructed by utilizing a deep neural network, and the method comprises the following steps:
step C1, constructing a network model by fusing RNN (neural network) with CNN, namely fusing a bidirectional LSTM (BilsTM) model and a model with two convolution layers and a global pooling layer to construct a deep neural network model;
and step C2: dividing data into a training set and a test set, training a training network model by using the test set, and storing a pre-training model;
and C3: and inputting the test data into the trained model to obtain a predicted result.
Due to the adoption of the technical scheme, compared with the prior art, the invention has the following advantages;
1. through rope skipping video acquisition, video data preprocessing, coordinate point acquisition, coordinate point windowing re-identification, feature sequence construction, regression prediction model establishment and model calling, the problems of low key point prediction in the alignment posture estimation process and long-time action analysis and motion potential discovery in the motion training process are solved, and a reference basis is provided for accurate action analysis in the motion process.
Drawings
FIG. 1 is a schematic diagram of the single swing foot jumping analysis of the preferred embodiment of the present invention;
FIG. 2 is a diagram of human body pose estimates at 14 points during rope skipping;
FIG. 3 is a partial position fenestration view of a human body;
FIG. 4 presents a graph of coordinate transformation of human key points in a rectangular coordinate system;
FIG. 5 is a general network architecture diagram;
FIG. 6 is a visualization result diagram of each layer of the network structure.
Detailed Description
The foregoing and other aspects, features and advantages of the invention will be apparent from the following more particular description of embodiments of the invention, as illustrated in the accompanying drawings in which reference is made to figures 1 to 6. The structural contents mentioned in the following embodiments are all referred to the attached drawings of the specification.
A human body continuous action rapid analysis method based on local identification mainly comprises rope skipping video acquisition, video data preprocessing, coordinate point acquisition, coordinate point windowing re-identification, feature sequence construction, regression prediction model establishment and model calling, and specifically comprises the following steps:
step 1: collecting video stream data of the human body in the process of shaking the front face of the human body and jumping with two feet through a mobile phone, and recording and storing basic information of a target object, wherein the basic information comprises name, sex, age, height, weight and name of action;
and 2, step: preprocessing video data, and performing posture estimation on a human body of each frame in a video to obtain a key point coordinate, wherein the method comprises the following steps of:
a1, converting video data shot by different mobile phones into a uniform scale, setting the height of a video to be 530cm and the width to be 460cm;
step A2: coordinate positions of 14 joint points of the nose, the neck, the right shoulder, the right elbow, the right wrist, the left shoulder, the left elbow, the left wrist, the right hip, the right knee, the right ankle, the left hip, the left knee and the left ankle of each frame of human body in the obtained video are represented by CP coordinates by utilizing an Open-position method i =(x i ,y i ) Wherein i is from 1 to 14, wherein (x) i ,y i ) Coordinates of key points of the human body;
step A3: open-pos uses the Gaussian distance between predicted keypoints and true values to define confidence in keypoints and normalizes the confidence to [0,1]Is defined herein as the score of the keypoint score and results in output results inputs i =(x i ,y i ,score i );
Step A4: the scores of the 14 key points were averaged and standard deviations, and the sum of the average and standard deviation was used as the score result for the overall key point,
Figure BDA0002299567150000051
Figure BDA0002299567150000052
score general (1) =score Average out +score Standard deviation of
And 3, step 3: windowing and re-identifying the low predicted position of the key point by utilizing an image windowing technology, and improving the prediction accuracy of the key point by utilizing global information and local parts, wherein the method comprises the following steps:
step B1, obtaining inputs by utilizing Open-pos i =(x i ,y i ,score i ) Setting a threshold th for the score, and finding out key points smaller than the threshold th;
b2, windowing the key points smaller than the threshold around the key points, and putting the image frames in the windows into the Open-pos network with the input modified;
b3, updating the coordinates of the key points obtained in the local frame by using the global information;
and 4, step 4: in order to further improve the robustness of the algorithm to shooting angles, target distance and recording process jitter factors, the key points are respectively set as the left hip C 1 Right hip C 11 Neck C 8 Performing coordinate conversion by taking the gravity center of the three coordinate points as an origin to obtain data
Figure BDA0002299567150000053
Wherein
Figure BDA0002299567150000054
I.e. defining the origin of coordinates, i.e. C, for the transformed relative coordinates 1 、C 11 、C 8 Center of gravity of three points C 0 =(cx 0 ,cy 0 ) Wherein
Figure BDA0002299567150000055
Figure BDA0002299567150000056
Updating all coordinate points by taking the original point coordinates as reference;
and 5: accumulating the coordinate matrix inputs obtained by each frame to obtain an accumulated coordinate matrix of each video segment, and performing window segmentation on the accumulated coordinate matrix by using a sliding window, namely setting the length of the sliding window to be the accumulated coordinate and the score obtained by k frames, and setting the step length to be 1;
and 6: a regression prediction model is constructed by utilizing a deep neural network, and the method comprises the following steps:
step C1, constructing a network model by fusing a CNN with an RNN, namely fusing a bidirectional LSTM (BilSTM) model and a model with two convolution layers and a global pooling layer to construct a deep neural network model;
and step C2: dividing data into a training set and a test set, training a training network model by using the test set, and storing a pre-training model;
and C3: and inputting the test data into the trained model to obtain a predicted result.
In step 6, the BilSTM is a bidirectional LSTM, the bidirectional LSTM is formed by two LSTMs which are superposed up and down, the output is determined by the shape of the two LSTMs together, and a recursive network calculates the hidden vector from front to back
Figure BDA0002299567150000061
Another recurrent neural network calculates the hidden vector from back to front
Figure BDA0002299567150000062
Final output
Figure BDA0002299567150000063
The first layer of the convolution layer is a one-dimensional convolution with convolution kernel of 5 multiplied by 5 and added with regularization, the second layer of the convolution layer is a convolution kernel of 3 multiplied by 3 and added with regularization, the third layer is a global pooling layer, and the number of filters in the convolution process is respectively 64 and 32;
adding a Dropout layer in the BilSTM layer, setting the last layer of activation function as a linear activation function in order to realize linear regression, and selecting a regression loss function Mean Square Error (MSE) loss function as a loss function; in order to accelerate the convergence of the network model, control overfitting to accelerate the convergence and control overfitting to add a batch normalization layer into each convolution layer, the batch normalization layer comprises the following calculation processes:
Figure BDA0002299567150000064
Figure BDA0002299567150000065
Figure BDA0002299567150000066
Figure BDA0002299567150000067
wherein B = { z = 1,...,m Is the input of the batch, m is the size of the batch, μ B Is the average of the batch data and,
Figure BDA0002299567150000068
is the variance of the batch process and,
Figure BDA0002299567150000069
to normalize the result,. Epsilon.is a minimum value, h i Gamma and beta are parameters input into the network model for learning, the results after scale change and offset;
on the basis of the scheme: the convolutional layer is represented in the form of a layer when step 6 is performed
Figure BDA0002299567150000071
Wherein l is the number of layers,
Figure BDA0002299567150000072
is the output of the jth neuron at layer l,
Figure BDA0002299567150000073
is the ith input of the l-th layer, an
Figure BDA0002299567150000074
* Represents the convolution, w ij Is the convolution kernel, bias is the bias term, M j For a set of input feature maps, f (-) represents an activation function;
giving k continuous subsections in the regression prediction output to predict the key point coordinates and scores of the next frame, obtaining the total scores according to the step 2, and adding the obtained frames into the current sequence by using recursion to predict the data of the next frame;
the image windowing technology is that required target information is highlighted in an original image, and the purpose is to detect the coordinate position of a key point in the window;
when step 3 is executed, the area of the window is S, and the coordinates of the four points of the window are ld (kx) 1 ,ky 1 )、lu(kx 2 ,ky 2 )、rd(kx 3 ,ky 3 )、ru(kx 4 ,ky 4 );
The maximum area of the image windowing is determined by the processing speed d of a CPU and the total operation amount m of a software fusion algorithm, wherein the larger the d, the smaller the m, the larger the windowing area, otherwise, the smaller the area, and the background image area is S 1 The maximum area of the window is S 2 ,S 2 =dS 1 /25;
The modified Open-position method adjusts the size format of input data into the length and width of the windowing, and outputs the coordinates of the target information converted according to the coordinates of the origin.
While the invention has been described in further detail with reference to specific embodiments thereof, it is not intended that the invention be limited to the specific embodiments thereof; for those skilled in the art to which the present invention pertains and related technologies, the extension, operation method and data replacement should fall within the protection scope of the present invention based on the technical solution of the present invention.

Claims (4)

1. A human body continuous action rapid analysis method based on local identification is characterized by mainly comprising rope skipping video acquisition, video data preprocessing, coordinate point acquisition, coordinate point windowing re-identification, feature sequence construction, regression prediction model establishment and model calling, and specifically comprises the following steps:
step 1: collecting video stream data in the process of shaking the front face of the human body and jumping with two feet through a mobile phone, and recording and storing basic information of a target object, wherein the basic information comprises name, sex, age, height, weight and name of an action to be made;
step 2: preprocessing video data, and performing posture estimation on a human body of each frame in a video to obtain a key point coordinate, wherein the method comprises the following steps of:
a1, converting video data shot by different mobile phones into a uniform scale, setting the height of a video to be 530cm, and setting the width to be 460cm;
step A2: obtaining the coordinate positions of 14 joint points of the nose, the neck, the right shoulder, the right elbow, the right wrist, the left shoulder, the left elbow, the left wrist, the right hip, the right knee, the right ankle, the left hip, the left knee and the left ankle of each frame of human body in the video by utilizing an Open-position method, wherein the coordinate is represented as CP i =(x i ,y i ) Wherein i is from 1 to 14, wherein (x) i ,y i ) Coordinates of key points of the human body;
step A3: open-pose uses the Gaussian distance between the predicted keypoint and the true value to define the confidence of the keypoint and normalizes the confidence to [0,1]Is defined herein as the score of the keypoint score and results in output results inputs i =(x i ,y i ,score i );
Step A4: the scores of the 14 key points are averaged and standard deviations are calculated, and the sum of the average and standard deviation is used as the score result of the whole key points,
Figure FDA0002299567140000011
Figure FDA0002299567140000012
score general (1) =score Average out +score Standard deviation of
And step 3: windowing and re-identifying the low predicted position of the key point by utilizing an image windowing technology, and improving the prediction accuracy of the key point by utilizing global information and local parts, wherein the method comprises the following steps:
step B1, obtaining inputs by utilizing Open-pos i =(x i ,y i ,score i ) Setting a threshold th for the score, and finding out key points smaller than the threshold th;
b2, windowing the key points smaller than the threshold around the key points, and putting the image frames in the windows into the Open-pos network with the input modified;
b3, updating the coordinates of the key points obtained in the local frame by using the global information;
and 4, step 4: in order to further improve the robustness of the algorithm to shooting angles, target distance and recording process jitter factors, the key points are respectively set as the left hip C 1 Right hip C 11 Neck C 8 Performing coordinate conversion by taking the gravity center of the three coordinate points as an origin to obtain data
Figure FDA0002299567140000021
Wherein
Figure FDA0002299567140000022
I.e. the converted relative coordinates
Defining origin of coordinates, i.e. C 1 、C 11 、C 8 Center of gravity of three points C 0 =(cx 0 ,cy 0 ) Wherein
Figure FDA0002299567140000023
Updating all coordinate points by taking the original point coordinates as reference;
and 5: accumulating the coordinate matrix inputs obtained by each frame to obtain an accumulated coordinate matrix of each video segment, and performing window segmentation on the accumulated coordinate matrix by using a sliding window, namely setting the length of the sliding window to be the accumulated coordinate and the score obtained by k frames, and setting the step length to be 1;
step 6: a regression prediction model is constructed by utilizing a deep neural network, and the method comprises the following steps:
step C1, constructing a network model by fusing RNN (neural network) with CNN, namely fusing a bidirectional LSTM (BilsTM) model and a model with two convolution layers and a global pooling layer to construct a deep neural network model;
and step C2: dividing data into a training set and a test set, training a training network model by using the test set, and storing a pre-training model;
and C3: and inputting the test data into the trained model to obtain a predicted result.
2. The method as claimed in claim 1, wherein the BilSTM is a bidirectional LSTM in step 6, the bidirectional LSTM has two LSTMs superimposed one on top of the other, and the output has a common shape decision of the two LSTMs, wherein a recursive network calculates the hidden vector from front to back
Figure FDA0002299567140000031
Another recurrent neural network calculates the hidden vector from back to front
Figure FDA0002299567140000032
Final output
Figure FDA0002299567140000033
The convolution layer first layer is a one-dimensional convolution with convolution kernel of 5 x 5 and added with regularization, the convolution kernel of the second layer is 3 x 3 and added with regularization, the third layer is a global pooling layer, and the number of filters in the convolution process is 64 and 32 respectively.
3. The method for rapidly analyzing human body continuous actions based on local recognition as claimed in claim 1, wherein a Dropout layer is added in a BilSTM layer, in order to realize linear regression, the last layer of activation function is set as a linear activation function, and the loss function selects a regression loss function Mean Square Error (MSE) loss function;
in order to accelerate the convergence of the network model, control overfitting to accelerate the convergence and control overfitting to add a batch normalization layer into each convolution layer, the batch normalization layer comprises the following calculation processes:
Figure FDA0002299567140000034
Figure FDA0002299567140000035
Figure FDA0002299567140000036
Figure FDA0002299567140000037
wherein B = { z = 1,...,m "is the input to the batch, m is the batch size, μ B Is the average of the batch data and,
Figure FDA0002299567140000038
is the variance of the batch process and,
Figure FDA0002299567140000039
for normalized results, ε is a minimum value, h i Gamma and beta are parameters input into the network model for learning as a result of the scale change and the offset;
the convolutional layer is represented in the form of a layer when step 6 is performed
Figure FDA00022995671400000310
Wherein l is the number of layers,
Figure FDA00022995671400000311
is the output of the jth neuron at level l,
Figure FDA00022995671400000312
is the ith input of the l-th layer, an
Figure FDA00022995671400000313
Represents the convolution, w ij Is the convolution kernel, bias is the bias term, M j For the set of input feature maps, f (-) denotes the stimulusA live function;
and (3) giving k continuous subsegments in the regression prediction output to predict the key point coordinates and the scores of the next frame, obtaining the total scores according to the step (2), and adding the obtained frames into the current sequence by utilizing recursion to predict the data of the next frame.
4. The method as claimed in claim 1, wherein the image windowing technique is to highlight the required target information in the original image in order to detect the coordinate position of the key point in the window;
when step 3 is executed, the area of the window is S, and the coordinates of four points of the window are ld (kx) at the upper left, the lower left, the upper right and the lower right 1 ,ky 1 )、lu(kx 2 ,ky 2 )、rd(kx 3 ,ky 3 )、ru(kx 4 ,ky 4 );
The maximum area of the image windowing is determined by the processing speed d of a CPU and the total operation amount m of a software fusion algorithm, wherein the larger the d, the smaller the m, the larger the windowing area, otherwise, the smaller the area, and the background image area is S 1 The maximum area of the window is S 2 ,S 2 =dS 1 /25;
The modified Open-position method adjusts the size format of input data into the length and width of the windowing, and outputs the coordinates of the target information converted according to the coordinates of the origin.
CN201911216130.8A 2019-12-02 2019-12-02 Human body continuous action rapid analysis method based on local recognition Active CN110956141B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911216130.8A CN110956141B (en) 2019-12-02 2019-12-02 Human body continuous action rapid analysis method based on local recognition

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911216130.8A CN110956141B (en) 2019-12-02 2019-12-02 Human body continuous action rapid analysis method based on local recognition

Publications (2)

Publication Number Publication Date
CN110956141A CN110956141A (en) 2020-04-03
CN110956141B true CN110956141B (en) 2023-02-28

Family

ID=69979435

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911216130.8A Active CN110956141B (en) 2019-12-02 2019-12-02 Human body continuous action rapid analysis method based on local recognition

Country Status (1)

Country Link
CN (1) CN110956141B (en)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112037310A (en) * 2020-08-27 2020-12-04 成都先知者科技有限公司 Game character action recognition generation method based on neural network
CN112569564A (en) * 2020-11-20 2021-03-30 融梦科技发展(北京)有限公司 Rope skipping detection method, device and system and storage medium
CN113095248B (en) * 2021-04-19 2022-10-25 中国石油大学(华东) Technical action correcting method for badminton
CN114596451B (en) * 2022-04-01 2022-11-11 此刻启动(北京)智能科技有限公司 Body fitness testing method and device based on AI vision and storage medium

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017133009A1 (en) * 2016-02-04 2017-08-10 广州新节奏智能科技有限公司 Method for positioning human joint using depth image of convolutional neural network
CN107643759A (en) * 2016-07-22 2018-01-30 鹦鹉无人机股份有限公司 From the autonomous system with target following and positioning of unmanned plane shooting mobile image
CN109708627A (en) * 2018-11-15 2019-05-03 上海航天控制技术研究所 A kind of moving platform down space dynamic point target rapid detection method
CN110503077A (en) * 2019-08-29 2019-11-26 郑州大学 A kind of real-time body's action-analysing method of view-based access control model

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017133009A1 (en) * 2016-02-04 2017-08-10 广州新节奏智能科技有限公司 Method for positioning human joint using depth image of convolutional neural network
CN107643759A (en) * 2016-07-22 2018-01-30 鹦鹉无人机股份有限公司 From the autonomous system with target following and positioning of unmanned plane shooting mobile image
CN109708627A (en) * 2018-11-15 2019-05-03 上海航天控制技术研究所 A kind of moving platform down space dynamic point target rapid detection method
CN110503077A (en) * 2019-08-29 2019-11-26 郑州大学 A kind of real-time body's action-analysing method of view-based access control model

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
基于骨架模型的人体行为分析;朱凌飞等;《电子测量技术》;20190423(第08期);全文 *

Also Published As

Publication number Publication date
CN110956141A (en) 2020-04-03

Similar Documents

Publication Publication Date Title
CN110956141B (en) Human body continuous action rapid analysis method based on local recognition
CN110956139B (en) Human motion analysis method based on time sequence regression prediction
CN109460702B (en) Passenger abnormal behavior identification method based on human body skeleton sequence
WO2018120964A1 (en) Posture correction method based on depth information and skeleton information
CN110991340B (en) Human body action analysis method based on image compression
CN110503077B (en) Real-time human body action analysis method based on vision
Chaudhari et al. Yog-guru: Real-time yoga pose correction system using deep learning methods
CN109753891A (en) Football player's orientation calibration method and system based on human body critical point detection
CN110575663B (en) Physical education auxiliary training method based on artificial intelligence
CN111753747B (en) Violent motion detection method based on monocular camera and three-dimensional attitude estimation
CN108399435A (en) A kind of video classification methods based on sound feature
CN111476161A (en) Somatosensory dynamic gesture recognition method fusing image and physiological signal dual channels
CN110991268B (en) Depth image-based Parkinson hand motion quantization analysis method and system
CN106548194B (en) The construction method and localization method of two dimensional image human joint points location model
CN108830170A (en) A kind of end-to-end method for tracking target indicated based on layered characteristic
CN114783611B (en) Neural recovered action detecting system based on artificial intelligence
CN112200074A (en) Attitude comparison method and terminal
Tsai et al. Enhancing accuracy of human action Recognition System using Skeleton Point correction method
CN109993116B (en) Pedestrian re-identification method based on mutual learning of human bones
CN115346272A (en) Real-time tumble detection method based on depth image sequence
CN111046715A (en) Human body action comparison analysis method based on image retrieval
CN117137435A (en) Rehabilitation action recognition method and system based on multi-mode information fusion
Nandyal et al. Cricket event recognition and classification from umpire action gestures using convolutional neural network
Zhang et al. Automatic construction and extraction of sports moment feature variables using artificial intelligence
Zhang et al. Multi-STMT: Multi-level Network for Human Activity Recognition Based on Wearable Sensors

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant