CN110956141A - Human body continuous action rapid analysis method based on local recognition - Google Patents

Human body continuous action rapid analysis method based on local recognition Download PDF

Info

Publication number
CN110956141A
CN110956141A CN201911216130.8A CN201911216130A CN110956141A CN 110956141 A CN110956141 A CN 110956141A CN 201911216130 A CN201911216130 A CN 201911216130A CN 110956141 A CN110956141 A CN 110956141A
Authority
CN
China
Prior art keywords
coordinate
layer
human body
score
video
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201911216130.8A
Other languages
Chinese (zh)
Other versions
CN110956141B (en
Inventor
赵红领
李润知
崔莉亚
刘皓东
王菁
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhengzhou University
Original Assignee
Zhengzhou University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhengzhou University filed Critical Zhengzhou University
Priority to CN201911216130.8A priority Critical patent/CN110956141B/en
Publication of CN110956141A publication Critical patent/CN110956141A/en
Application granted granted Critical
Publication of CN110956141B publication Critical patent/CN110956141B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Biophysics (AREA)
  • Molecular Biology (AREA)
  • Biomedical Technology (AREA)
  • General Engineering & Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Computational Linguistics (AREA)
  • Software Systems (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Multimedia (AREA)
  • Psychiatry (AREA)
  • Social Psychology (AREA)
  • Human Computer Interaction (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a human body continuous action rapid analysis method based on local identification, which mainly comprises rope skipping video acquisition, video data preprocessing, coordinate point acquisition, coordinate point windowing repeated identification, feature sequence construction, regression prediction model establishment and model calling, wherein video stream data in the process of double-foot jumping of the front side of a human body are acquired through a mobile phone, basic information of a target object, including name, gender, age, height, weight and made action name, is recorded and stored, and a regression prediction model is constructed by utilizing a deep neural network.

Description

Human body continuous action rapid analysis method based on local recognition
Technical Field
The invention relates to the technical field of human motion and motion analysis, in particular to a human continuous motion rapid analysis method based on local recognition.
Background
With the development and application of computer technology and machine learning in recent years, video-based time series human motion analysis technology is rapidly emerging and has gained wide attention. At present, human body action analysis based on a video time sequence is still a very challenging subject in computer vision, relates to a plurality of subjects such as image processing, pattern recognition and machine learning, and has wide application prospects in the fields of intelligent monitoring, man-machine interaction, rehabilitation exercise, physical training evaluation and the like.
When the traditional neural network is used for processing input images with fuzzy features and large ambiguity, the traditional neural network is easily misled by irrelevant information in the images, and is difficult to ensure higher recognition rate.
The existing human behavior recognition algorithm based on videos has the problems of high complexity, poor robustness, low precision and the like. In addition, research work on the aspect of regression prediction analysis of human body actions is less, so the invention provides the human body action analysis method based on time series regression prediction, which has high robustness and stable time overhead, and has important significance for human body action analysis modeling, action quality evaluation and potential mining of sporters.
Disclosure of Invention
In view of the above situation, in order to overcome the defects of the prior art, the present invention aims to provide a method for rapidly analyzing human body continuous actions based on local recognition, which solves the problem of low accuracy of long-time regression prediction of actions in the exercise training process.
The technical scheme for solving the problem is that the method for rapidly analyzing the continuous actions of the human body based on local recognition mainly comprises rope skipping video acquisition, video data preprocessing, coordinate point acquisition, coordinate point windowing re-recognition, feature sequence construction, regression prediction model establishment and model calling, and specifically comprises the following steps:
step 1: collecting video stream data of the human body in the process of shaking the front face of the human body and jumping with two feet through a mobile phone, and recording and storing basic information of a target object, wherein the basic information comprises name, sex, age, height, weight and name of action;
step 2: preprocessing video data, and performing posture estimation on a human body of each frame in a video to obtain a key point coordinate, wherein the method comprises the following steps of:
a1, converting video data shot by different mobile phones into a uniform scale, setting the height of a video to be 530cm and the width to be 460 cm;
step A2: obtaining the coordinate positions of 14 joint points of the nose, the neck, the right shoulder, the right elbow, the right wrist, the left shoulder, the left elbow, the left wrist, the right hip, the right knee, the right ankle, the left hip, the left knee and the left ankle of each frame of human body in the video by utilizing an Open-position method, wherein the coordinate is represented as CPi=(xi,yi) Wherein i is from 1 to 14, wherein (x)i,yi) Coordinates of key points of the human body;
step A3: open-pos uses the Gaussian distance between predicted keypoints and true values to define confidence in keypoints and normalizes the confidence to [0, 1%]Is defined herein as the score of the keypoint and results in output results inputsi=(xi,yi,scorei);
Step A4: the scores of the 14 key points were averaged and standard deviations, and the sum of the average and standard deviation was used as the score result for the overall key point,
Figure BDA0002299567150000021
Figure BDA0002299567150000022
scoregeneral assembly=scoreAverage+scoreStandard deviation of
And step 3: windowing and re-identifying the low predicted position of the key point by utilizing an image windowing technology, and improving the prediction accuracy of the key point by utilizing global information and local parts, wherein the method comprises the following steps:
step B1, obtaining inputs by using Open-posi=(xi,yi,scorei) Setting a threshold th for the score, and finding out key points smaller than the threshold th;
b2, windowing the key points smaller than the threshold around the key points, and putting the image frames in the windows into the Open-pos network with modified input;
step B3, updating the key point coordinates obtained in the local frame by using the global information;
and 4, step 4: in order to further improve the robustness of the algorithm to shooting angles, target distance and recording process jitter factors, the key points are respectively set as the left hip C1Right hip C11Neck C8Performing coordinate conversion by taking the gravity center of the three coordinate points as an origin to obtain data
Figure BDA0002299567150000031
Wherein
Figure BDA0002299567150000032
I.e. defining the origin of coordinates, i.e. C, for the transformed relative coordinates1、C11、C8Center of gravity of three points C0=(cx0,cy0) Wherein
Figure BDA0002299567150000033
Figure BDA0002299567150000034
Updating all coordinate points by taking the original point coordinates as reference;
and 5: accumulating the coordinate matrix inputs obtained by each frame to obtain an accumulated coordinate matrix of each video segment, and performing window segmentation on the accumulated coordinate matrix by using a sliding window, namely setting the length of the sliding window to be the accumulated coordinate and the score obtained by k frames, and setting the step length to be 1;
step 6: a regression prediction model is constructed by utilizing a deep neural network, and the method comprises the following steps:
step C1, constructing a network model by fusing RNN through CNN, namely, fusing a bidirectional LSTM (BilsTM) model and a model with two convolution layers and a global pooling layer to construct a deep neural network model;
step C2: dividing data into a training set and a test set, training a training network model by using the test set, and storing a pre-training model;
step C3: and inputting the test data into the trained model to obtain a predicted result.
Due to the adoption of the technical scheme, compared with the prior art, the invention has the following advantages;
1. through rope skipping video acquisition, video data preprocessing, coordinate point acquisition, coordinate point windowing re-identification, feature sequence construction, regression prediction model establishment and model calling, the problems of low key point prediction in the alignment posture estimation process and long-time action analysis and movement potential discovery in the movement training process are solved, and a reference basis is provided for accurate action analysis in the movement process.
Drawings
FIG. 1 is a schematic diagram of the single-swing double-foot jumping analysis in the preferred embodiment of the present invention;
FIG. 2 is a diagram of human body pose estimates at 14 points during rope skipping;
FIG. 3 is a partial position fenestration view of a human body;
FIG. 4 is a diagram of coordinate transformation of key points of a human body in a rectangular coordinate system;
FIG. 5 is a diagram of an overall network architecture;
FIG. 6 is a visualization result diagram of each layer of the network structure.
Detailed Description
The foregoing and other aspects, features and advantages of the invention will be apparent from the following more particular description of embodiments of the invention, as illustrated in the accompanying drawings in which reference is made to figures 1 to 6. The structural contents mentioned in the following embodiments are all referred to the attached drawings of the specification.
A human body continuous action rapid analysis method based on local identification mainly comprises rope skipping video acquisition, video data preprocessing, coordinate point acquisition, coordinate point windowing re-identification, feature sequence construction, regression prediction model establishment and model calling, and specifically comprises the following steps:
step 1: collecting video stream data of the human body in the process of shaking the front face of the human body and jumping with two feet through a mobile phone, and recording and storing basic information of a target object, wherein the basic information comprises name, sex, age, height, weight and name of action;
step 2: preprocessing video data, and performing posture estimation on a human body of each frame in a video to obtain a key point coordinate, wherein the method comprises the following steps of:
a1, converting video data shot by different mobile phones into a uniform scale, setting the height of a video to be 530cm and the width to be 460 cm;
step A2: obtaining the coordinate positions of 14 joint points of the nose, the neck, the right shoulder, the right elbow, the right wrist, the left shoulder, the left elbow, the left wrist, the right hip, the right knee, the right ankle, the left hip, the left knee and the left ankle of each frame of human body in the video by utilizing an Open-position method, wherein the coordinate is represented as CPi=(xi,yi) Wherein i is from 1 to 14, wherein (x)i,yi) Coordinates of key points of the human body;
step A3: open-pos uses the Gaussian distance between predicted keypoints and true values to define confidence in keypoints and normalizes the confidence to [0, 1%]Is defined herein as the score of the keypoint and results in output results inputsi=(xi,yi,scorei);
Step A4: the scores of the 14 key points were averaged and standard deviations, and the sum of the average and standard deviation was used as the score result for the overall key point,
Figure BDA0002299567150000051
Figure BDA0002299567150000052
scoregeneral assembly=scoreAverage+scoreStandard deviation of
And step 3: windowing and re-identifying the low predicted position of the key point by utilizing an image windowing technology, and improving the prediction accuracy of the key point by utilizing global information and local parts, wherein the method comprises the following steps:
step B1, obtaining inputs by using Open-posi=(xi,yi,scorei) Setting a threshold th for the score, and finding out key points smaller than the threshold th;
b2, windowing the key points smaller than the threshold around the key points, and putting the image frames in the windows into the Open-pos network with modified input;
step B3, updating the key point coordinates obtained in the local frame by using the global information;
and 4, step 4: in order to further improve the robustness of the algorithm to shooting angles, target distance and recording process jitter factors, the key points are respectively set as the left hip C1Right hip C11Neck C8Performing coordinate conversion by taking the gravity center of the three coordinate points as an origin to obtain data
Figure BDA0002299567150000053
Wherein
Figure BDA0002299567150000054
I.e. defining the origin of coordinates, i.e. C, for the transformed relative coordinates1、C11、C8Center of gravity of three points C0=(cx0,cy0) Wherein
Figure BDA0002299567150000055
Figure BDA0002299567150000056
Updating all coordinate points by taking the original point coordinates as reference;
and 5: accumulating the coordinate matrix inputs obtained by each frame to obtain an accumulated coordinate matrix of each video segment, and performing window segmentation on the accumulated coordinate matrix by using a sliding window, namely setting the length of the sliding window to be the accumulated coordinate and the score obtained by k frames, and setting the step length to be 1;
step 6: a regression prediction model is constructed by utilizing a deep neural network, and the method comprises the following steps:
step C1, constructing a network model by fusing RNN through CNN, namely, fusing a bidirectional LSTM (BilsTM) model and a model with two convolution layers and a global pooling layer to construct a deep neural network model;
step C2: dividing data into a training set and a test set, training a training network model by using the test set, and storing a pre-training model;
step C3: and inputting the test data into the trained model to obtain a predicted result.
In step 6, the BilSTM is a bidirectional LSTM, the bidirectional LSTM is formed by two LSTMs which are superposed up and down, the output is determined by the shape of the two LSTMs together, and a recursive network calculates the hidden vector from front to back
Figure BDA0002299567150000061
Another recurrent neural network calculates the hidden vector from back to front
Figure BDA0002299567150000062
Final output
Figure BDA0002299567150000063
The first layer of the convolution layer is a one-dimensional convolution with convolution kernel of 5 multiplied by 5 and added with regularization, the second layer of the convolution layer is a convolution kernel of 3 multiplied by 3 and added with regularization, the third layer is a global pooling layer, and the number of filters in the convolution process is respectively 64 and 32;
adding a Dropout layer in the BilSTM layer, setting the last layer of activation function as a linear activation function in order to realize linear regression, and selecting a regression loss function Mean Square Error (MSE) loss function as a loss function; in order to accelerate the convergence of the network model, control overfitting to accelerate the convergence and control overfitting to add a batch normalization layer into each convolution layer, the batch normalization layer comprises the following calculation processes:
Figure BDA0002299567150000064
Figure BDA0002299567150000065
Figure BDA0002299567150000066
Figure BDA0002299567150000067
wherein B ═ { z ═ z1,...,m"is the input to the batch, m is the batch size, μBIs the average of the batch data and,
Figure BDA0002299567150000068
is the variance of the batch process and,
Figure BDA0002299567150000069
for normalized results, ε is a minimum value, hiAs a result of the scaling and shifting, γ and β are parameters learned in the input network model;
on the basis of the scheme: the convolutional layer is represented in the form of a layer when step 6 is performed
Figure BDA0002299567150000071
Wherein l is the number of layers,
Figure BDA0002299567150000072
is the output of the jth neuron at level l,
Figure BDA0002299567150000073
is the ith input of the l-th layer, an
Figure BDA0002299567150000074
Represents a convolution, wijIs the convolution kernel, bias is the bias term, MjFor a set of input feature maps, f (-) represents an activation function;
giving k continuous subsegments in the output of regression prediction to predict the key point coordinates and scores of the next frame, obtaining total scores according to the step 2, and adding the obtained frames into the current sequence by utilizing recursion to predict the data of the next frame;
the image windowing technology is that required target information is highlighted in an original image, and the purpose is to detect the coordinate position of a key point in the window;
when step 3 is executed, the area of the window is S, and the coordinates of the four points of the window are ld (kx)1,ky1)、lu(kx2,ky2)、rd(kx3,ky3)、ru(kx4,ky4);
The maximum area of the image windowing is determined by the processing speed d of a CPU and the total operation amount m of a software fusion algorithm, wherein the larger the d, the smaller the m, the larger the windowing area, otherwise, the smaller the area, and the background image area is S1The maximum area of the window is S2,S2=dS1/25;
The modified Open-position method adjusts the size format of input data into the length and width of a window, and outputs the coordinates of target information converted according to the origin coordinates.
While the invention has been described in further detail with reference to specific embodiments thereof, it is not intended that the invention be limited to the specific embodiments thereof; for those skilled in the art to which the present invention pertains and related technologies, the extension, operation method and data replacement should fall within the protection scope of the present invention based on the technical solution of the present invention.

Claims (4)

1. A human body continuous action rapid analysis method based on local identification is characterized by mainly comprising rope skipping video acquisition, video data preprocessing, coordinate point acquisition, coordinate point windowing re-identification, feature sequence construction, regression prediction model establishment and model calling, and specifically comprises the following steps:
step 1: collecting video stream data of the human body in the process of shaking the front face of the human body and jumping with two feet through a mobile phone, and recording and storing basic information of a target object, wherein the basic information comprises name, sex, age, height, weight and name of action;
step 2: preprocessing video data, and performing posture estimation on a human body of each frame in a video to obtain a key point coordinate, wherein the method comprises the following steps of:
a1, converting video data shot by different mobile phones into a uniform scale, setting the height of a video to be 530cm and the width to be 460 cm;
step A2: obtaining the coordinate positions of 14 joint points of the nose, the neck, the right shoulder, the right elbow, the right wrist, the left shoulder, the left elbow, the left wrist, the right hip, the right knee, the right ankle, the left hip, the left knee and the left ankle of each frame of human body in the video by utilizing an Open-position method, wherein the coordinate is represented as CPi=(xi,yi) Wherein i is from 1 to 14, wherein (x)i,yi) Coordinates of key points of the human body;
step A3: open-pos uses the Gaussian distance between predicted keypoints and true values to define confidence in keypoints and normalizes the confidence to [0, 1%]Is defined herein as the score of the keypoint and results in output results inputsi=(xi,yi,scorei);
Step A4: the scores of the 14 key points were averaged and standard deviations, and the sum of the average and standard deviation was used as the score result for the overall key point,
Figure FDA0002299567140000011
Figure FDA0002299567140000012
scoregeneral assembly=scoreAverage+scoreStandard deviation of
And step 3: windowing and re-identifying the low predicted position of the key point by utilizing an image windowing technology, and improving the prediction accuracy of the key point by utilizing global information and local parts, wherein the method comprises the following steps:
step B1, obtaining inputs by using Open-posi=(xi,yi,scorei) Setting a threshold th for the score, and finding out key points smaller than the threshold th;
b2, windowing the key points smaller than the threshold around the key points, and putting the image frames in the windows into the Open-pos network with modified input;
step B3, updating the key point coordinates obtained in the local frame by using the global information;
and 4, step 4: in order to further improve the robustness of the algorithm to shooting angles, target distance and recording process jitter factors, the key points are respectively set as the left hip C1Right hip C11Neck C8Performing coordinate conversion by taking the gravity center of the three coordinate points as an origin to obtain data
Figure FDA0002299567140000021
Wherein
Figure FDA0002299567140000022
I.e. the relative coordinates after conversion
Defining origin of coordinates, i.e. C1、C11、C8Center of gravity of three points C0=(cx0,cy0) Wherein
Figure FDA0002299567140000023
Updating all coordinate points by taking the original point coordinates as reference;
and 5: accumulating the coordinate matrix inputs obtained by each frame to obtain an accumulated coordinate matrix of each video segment, and performing window segmentation on the accumulated coordinate matrix by using a sliding window, namely setting the length of the sliding window to be the accumulated coordinate and the score obtained by k frames, and setting the step length to be 1;
step 6: a regression prediction model is constructed by utilizing a deep neural network, and the method comprises the following steps:
step C1, constructing a network model by fusing RNN through CNN, namely, fusing a bidirectional LSTM (BilsTM) model and a model with two convolution layers and a global pooling layer to construct a deep neural network model;
step C2: dividing data into a training set and a test set, training a training network model by using the test set, and storing a pre-training model;
step C3: and inputting the test data into the trained model to obtain a predicted result.
2. The method as claimed in claim 1, wherein the BilSTM is a bidirectional LSTM in step 6, the bidirectional LSTM has two LSTMs superimposed one on top of the other, and the output has a common shape decision of the two LSTMs, wherein a recursive network calculates the hidden vector from front to back
Figure FDA0002299567140000031
Another recurrent neural network calculates the hidden vector from back to front
Figure FDA0002299567140000032
Final output
Figure FDA0002299567140000033
The convolution layer first layer is a one-dimensional convolution with convolution kernel of 5 x 5 and added with regularization, the convolution kernel of the second layer is 3 x 3 and added with regularization, the third layer is a global pooling layer, and the number of filters in the convolution process is 64 and 32 respectively.
3. The method for rapidly analyzing human body continuous actions based on local recognition as claimed in claim 1, wherein a Dropout layer is added in a BilSTM layer, in order to realize linear regression, the last layer of activation function is set as a linear activation function, and the loss function selects a regression loss function Mean Square Error (MSE) loss function;
in order to accelerate the convergence of the network model, control overfitting to accelerate the convergence and control overfitting to add a batch normalization layer into each convolution layer, the batch normalization layer comprises the following calculation processes:
Figure FDA0002299567140000034
Figure FDA0002299567140000035
Figure FDA0002299567140000036
Figure FDA0002299567140000037
wherein B ═ { z ═ z1,...,m"is the input to the batch, m is the batch size, μBIs the average of the batch data and,
Figure FDA0002299567140000038
is the variance of the batch process and,
Figure FDA0002299567140000039
for normalized results, ε is a minimum value, hiAs a result of the scaling and shifting, γ and β are parameters learned in the input network model;
the convolutional layer is represented in the form of a layer when step 6 is performed
Figure FDA00022995671400000310
Wherein l is the number of layers,
Figure FDA00022995671400000311
is the output of the jth neuron at level l,
Figure FDA00022995671400000312
is the l-th layerThe ith input, and
Figure FDA00022995671400000313
represents a convolution, wijIs the convolution kernel, bias is the bias term, MjFor a set of input feature maps, f (-) represents an activation function;
and (3) giving k continuous subsegments in the regression prediction output to predict the key point coordinates and the scores of the next frame, obtaining the total scores according to the step (2), and adding the obtained frames into the current sequence by utilizing recursion to predict the data of the next frame.
4. The method as claimed in claim 1, wherein the image windowing technique is used to highlight the required target information in the original image in order to detect the coordinate position of the key point in the window;
when step 3 is executed, the area of the window is S, and the coordinates of the four points of the window are ld (kx)1,ky1)、lu(kx2,ky2)、rd(kx3,ky3)、ru(kx4,ky4);
The maximum area of the image windowing is determined by the processing speed d of a CPU and the total operation amount m of a software fusion algorithm, wherein the larger the d, the smaller the m, the larger the windowing area, otherwise, the smaller the area, and the background image area is S1The maximum area of the window is S2,S2=dS1/25;
The modified Open-position method adjusts the size format of input data into the length and width of a window, and outputs the coordinates of target information converted according to the origin coordinates.
CN201911216130.8A 2019-12-02 2019-12-02 Human body continuous action rapid analysis method based on local recognition Active CN110956141B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911216130.8A CN110956141B (en) 2019-12-02 2019-12-02 Human body continuous action rapid analysis method based on local recognition

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911216130.8A CN110956141B (en) 2019-12-02 2019-12-02 Human body continuous action rapid analysis method based on local recognition

Publications (2)

Publication Number Publication Date
CN110956141A true CN110956141A (en) 2020-04-03
CN110956141B CN110956141B (en) 2023-02-28

Family

ID=69979435

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911216130.8A Active CN110956141B (en) 2019-12-02 2019-12-02 Human body continuous action rapid analysis method based on local recognition

Country Status (1)

Country Link
CN (1) CN110956141B (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112037310A (en) * 2020-08-27 2020-12-04 成都先知者科技有限公司 Game character action recognition generation method based on neural network
CN112569564A (en) * 2020-11-20 2021-03-30 融梦科技发展(北京)有限公司 Rope skipping detection method, device and system and storage medium
CN113095248A (en) * 2021-04-19 2021-07-09 中国石油大学(华东) Technical action correction method for badminton
CN114596451A (en) * 2022-04-01 2022-06-07 此刻启动(北京)智能科技有限公司 Body fitness testing method and device based on AI vision and storage medium

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017133009A1 (en) * 2016-02-04 2017-08-10 广州新节奏智能科技有限公司 Method for positioning human joint using depth image of convolutional neural network
CN107643759A (en) * 2016-07-22 2018-01-30 鹦鹉无人机股份有限公司 From the autonomous system with target following and positioning of unmanned plane shooting mobile image
CN109708627A (en) * 2018-11-15 2019-05-03 上海航天控制技术研究所 A kind of moving platform down space dynamic point target rapid detection method
CN110503077A (en) * 2019-08-29 2019-11-26 郑州大学 A kind of real-time body's action-analysing method of view-based access control model

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017133009A1 (en) * 2016-02-04 2017-08-10 广州新节奏智能科技有限公司 Method for positioning human joint using depth image of convolutional neural network
CN107643759A (en) * 2016-07-22 2018-01-30 鹦鹉无人机股份有限公司 From the autonomous system with target following and positioning of unmanned plane shooting mobile image
CN109708627A (en) * 2018-11-15 2019-05-03 上海航天控制技术研究所 A kind of moving platform down space dynamic point target rapid detection method
CN110503077A (en) * 2019-08-29 2019-11-26 郑州大学 A kind of real-time body's action-analysing method of view-based access control model

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
朱凌飞等: "基于骨架模型的人体行为分析", 《电子测量技术》 *

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112037310A (en) * 2020-08-27 2020-12-04 成都先知者科技有限公司 Game character action recognition generation method based on neural network
CN112569564A (en) * 2020-11-20 2021-03-30 融梦科技发展(北京)有限公司 Rope skipping detection method, device and system and storage medium
CN113095248A (en) * 2021-04-19 2021-07-09 中国石油大学(华东) Technical action correction method for badminton
CN114596451A (en) * 2022-04-01 2022-06-07 此刻启动(北京)智能科技有限公司 Body fitness testing method and device based on AI vision and storage medium

Also Published As

Publication number Publication date
CN110956141B (en) 2023-02-28

Similar Documents

Publication Publication Date Title
CN110956141B (en) Human body continuous action rapid analysis method based on local recognition
CN106650687B (en) Posture correction method based on depth information and skeleton information
CN110991340B (en) Human body action analysis method based on image compression
CN109460702B (en) Passenger abnormal behavior identification method based on human body skeleton sequence
CN110956139B (en) Human motion analysis method based on time sequence regression prediction
CN110503077B (en) Real-time human body action analysis method based on vision
Chaudhari et al. Yog-guru: Real-time yoga pose correction system using deep learning methods
CN110575663B (en) Physical education auxiliary training method based on artificial intelligence
Zhou et al. Learning to estimate 3d human pose from point cloud
CN111753747B (en) Violent motion detection method based on monocular camera and three-dimensional attitude estimation
CN111582349B (en) Improved target tracking algorithm based on YOLOv3 and kernel correlation filtering
CN106548194B (en) The construction method and localization method of two dimensional image human joint points location model
Jia et al. Two-stream temporal convolutional networks for skeleton-based human action recognition
Xu et al. Robust hand gesture recognition based on RGB-D Data for natural human–computer interaction
CN109325408A (en) A kind of gesture judging method and storage medium
CN108830170A (en) A kind of end-to-end method for tracking target indicated based on layered characteristic
CN117541994A (en) Abnormal behavior detection model and detection method in dense multi-person scene
CN111626152B (en) Space-time line-of-sight direction estimation prototype design method based on Few-shot
CN109993116B (en) Pedestrian re-identification method based on mutual learning of human bones
CN113408435B (en) Security monitoring method, device, equipment and storage medium
CN111046715A (en) Human body action comparison analysis method based on image retrieval
Hoang et al. Advances in skeleton-based fall detection in RGB videos: From handcrafted to deep learning approaches
CN117137435B (en) Rehabilitation action recognition method and system based on multi-mode information fusion
CN115205744A (en) Intelligent exercise assisting method and device for figure skating
CN113065504A (en) Behavior identification method and device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant