CN110956141A - Human body continuous action rapid analysis method based on local recognition - Google Patents
Human body continuous action rapid analysis method based on local recognition Download PDFInfo
- Publication number
- CN110956141A CN110956141A CN201911216130.8A CN201911216130A CN110956141A CN 110956141 A CN110956141 A CN 110956141A CN 201911216130 A CN201911216130 A CN 201911216130A CN 110956141 A CN110956141 A CN 110956141A
- Authority
- CN
- China
- Prior art keywords
- coordinate
- layer
- human body
- score
- video
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/20—Movements or behaviour, e.g. gesture recognition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/044—Recurrent networks, e.g. Hopfield networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/44—Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Biophysics (AREA)
- Molecular Biology (AREA)
- Biomedical Technology (AREA)
- General Engineering & Computer Science (AREA)
- Artificial Intelligence (AREA)
- Computational Linguistics (AREA)
- Software Systems (AREA)
- Life Sciences & Earth Sciences (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Multimedia (AREA)
- Psychiatry (AREA)
- Social Psychology (AREA)
- Human Computer Interaction (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses a human body continuous action rapid analysis method based on local identification, which mainly comprises rope skipping video acquisition, video data preprocessing, coordinate point acquisition, coordinate point windowing repeated identification, feature sequence construction, regression prediction model establishment and model calling, wherein video stream data in the process of double-foot jumping of the front side of a human body are acquired through a mobile phone, basic information of a target object, including name, gender, age, height, weight and made action name, is recorded and stored, and a regression prediction model is constructed by utilizing a deep neural network.
Description
Technical Field
The invention relates to the technical field of human motion and motion analysis, in particular to a human continuous motion rapid analysis method based on local recognition.
Background
With the development and application of computer technology and machine learning in recent years, video-based time series human motion analysis technology is rapidly emerging and has gained wide attention. At present, human body action analysis based on a video time sequence is still a very challenging subject in computer vision, relates to a plurality of subjects such as image processing, pattern recognition and machine learning, and has wide application prospects in the fields of intelligent monitoring, man-machine interaction, rehabilitation exercise, physical training evaluation and the like.
When the traditional neural network is used for processing input images with fuzzy features and large ambiguity, the traditional neural network is easily misled by irrelevant information in the images, and is difficult to ensure higher recognition rate.
The existing human behavior recognition algorithm based on videos has the problems of high complexity, poor robustness, low precision and the like. In addition, research work on the aspect of regression prediction analysis of human body actions is less, so the invention provides the human body action analysis method based on time series regression prediction, which has high robustness and stable time overhead, and has important significance for human body action analysis modeling, action quality evaluation and potential mining of sporters.
Disclosure of Invention
In view of the above situation, in order to overcome the defects of the prior art, the present invention aims to provide a method for rapidly analyzing human body continuous actions based on local recognition, which solves the problem of low accuracy of long-time regression prediction of actions in the exercise training process.
The technical scheme for solving the problem is that the method for rapidly analyzing the continuous actions of the human body based on local recognition mainly comprises rope skipping video acquisition, video data preprocessing, coordinate point acquisition, coordinate point windowing re-recognition, feature sequence construction, regression prediction model establishment and model calling, and specifically comprises the following steps:
step 1: collecting video stream data of the human body in the process of shaking the front face of the human body and jumping with two feet through a mobile phone, and recording and storing basic information of a target object, wherein the basic information comprises name, sex, age, height, weight and name of action;
step 2: preprocessing video data, and performing posture estimation on a human body of each frame in a video to obtain a key point coordinate, wherein the method comprises the following steps of:
a1, converting video data shot by different mobile phones into a uniform scale, setting the height of a video to be 530cm and the width to be 460 cm;
step A2: obtaining the coordinate positions of 14 joint points of the nose, the neck, the right shoulder, the right elbow, the right wrist, the left shoulder, the left elbow, the left wrist, the right hip, the right knee, the right ankle, the left hip, the left knee and the left ankle of each frame of human body in the video by utilizing an Open-position method, wherein the coordinate is represented as CPi=(xi,yi) Wherein i is from 1 to 14, wherein (x)i,yi) Coordinates of key points of the human body;
step A3: open-pos uses the Gaussian distance between predicted keypoints and true values to define confidence in keypoints and normalizes the confidence to [0, 1%]Is defined herein as the score of the keypoint and results in output results inputsi=(xi,yi,scorei);
Step A4: the scores of the 14 key points were averaged and standard deviations, and the sum of the average and standard deviation was used as the score result for the overall key point,
scoregeneral assembly=scoreAverage+scoreStandard deviation of;
And step 3: windowing and re-identifying the low predicted position of the key point by utilizing an image windowing technology, and improving the prediction accuracy of the key point by utilizing global information and local parts, wherein the method comprises the following steps:
step B1, obtaining inputs by using Open-posi=(xi,yi,scorei) Setting a threshold th for the score, and finding out key points smaller than the threshold th;
b2, windowing the key points smaller than the threshold around the key points, and putting the image frames in the windows into the Open-pos network with modified input;
step B3, updating the key point coordinates obtained in the local frame by using the global information;
and 4, step 4: in order to further improve the robustness of the algorithm to shooting angles, target distance and recording process jitter factors, the key points are respectively set as the left hip C1Right hip C11Neck C8Performing coordinate conversion by taking the gravity center of the three coordinate points as an origin to obtain dataWhereinI.e. defining the origin of coordinates, i.e. C, for the transformed relative coordinates1、C11、C8Center of gravity of three points C0=(cx0,cy0) Wherein Updating all coordinate points by taking the original point coordinates as reference;
and 5: accumulating the coordinate matrix inputs obtained by each frame to obtain an accumulated coordinate matrix of each video segment, and performing window segmentation on the accumulated coordinate matrix by using a sliding window, namely setting the length of the sliding window to be the accumulated coordinate and the score obtained by k frames, and setting the step length to be 1;
step 6: a regression prediction model is constructed by utilizing a deep neural network, and the method comprises the following steps:
step C1, constructing a network model by fusing RNN through CNN, namely, fusing a bidirectional LSTM (BilsTM) model and a model with two convolution layers and a global pooling layer to construct a deep neural network model;
step C2: dividing data into a training set and a test set, training a training network model by using the test set, and storing a pre-training model;
step C3: and inputting the test data into the trained model to obtain a predicted result.
Due to the adoption of the technical scheme, compared with the prior art, the invention has the following advantages;
1. through rope skipping video acquisition, video data preprocessing, coordinate point acquisition, coordinate point windowing re-identification, feature sequence construction, regression prediction model establishment and model calling, the problems of low key point prediction in the alignment posture estimation process and long-time action analysis and movement potential discovery in the movement training process are solved, and a reference basis is provided for accurate action analysis in the movement process.
Drawings
FIG. 1 is a schematic diagram of the single-swing double-foot jumping analysis in the preferred embodiment of the present invention;
FIG. 2 is a diagram of human body pose estimates at 14 points during rope skipping;
FIG. 3 is a partial position fenestration view of a human body;
FIG. 4 is a diagram of coordinate transformation of key points of a human body in a rectangular coordinate system;
FIG. 5 is a diagram of an overall network architecture;
FIG. 6 is a visualization result diagram of each layer of the network structure.
Detailed Description
The foregoing and other aspects, features and advantages of the invention will be apparent from the following more particular description of embodiments of the invention, as illustrated in the accompanying drawings in which reference is made to figures 1 to 6. The structural contents mentioned in the following embodiments are all referred to the attached drawings of the specification.
A human body continuous action rapid analysis method based on local identification mainly comprises rope skipping video acquisition, video data preprocessing, coordinate point acquisition, coordinate point windowing re-identification, feature sequence construction, regression prediction model establishment and model calling, and specifically comprises the following steps:
step 1: collecting video stream data of the human body in the process of shaking the front face of the human body and jumping with two feet through a mobile phone, and recording and storing basic information of a target object, wherein the basic information comprises name, sex, age, height, weight and name of action;
step 2: preprocessing video data, and performing posture estimation on a human body of each frame in a video to obtain a key point coordinate, wherein the method comprises the following steps of:
a1, converting video data shot by different mobile phones into a uniform scale, setting the height of a video to be 530cm and the width to be 460 cm;
step A2: obtaining the coordinate positions of 14 joint points of the nose, the neck, the right shoulder, the right elbow, the right wrist, the left shoulder, the left elbow, the left wrist, the right hip, the right knee, the right ankle, the left hip, the left knee and the left ankle of each frame of human body in the video by utilizing an Open-position method, wherein the coordinate is represented as CPi=(xi,yi) Wherein i is from 1 to 14, wherein (x)i,yi) Coordinates of key points of the human body;
step A3: open-pos uses the Gaussian distance between predicted keypoints and true values to define confidence in keypoints and normalizes the confidence to [0, 1%]Is defined herein as the score of the keypoint and results in output results inputsi=(xi,yi,scorei);
Step A4: the scores of the 14 key points were averaged and standard deviations, and the sum of the average and standard deviation was used as the score result for the overall key point,
scoregeneral assembly=scoreAverage+scoreStandard deviation of;
And step 3: windowing and re-identifying the low predicted position of the key point by utilizing an image windowing technology, and improving the prediction accuracy of the key point by utilizing global information and local parts, wherein the method comprises the following steps:
step B1, obtaining inputs by using Open-posi=(xi,yi,scorei) Setting a threshold th for the score, and finding out key points smaller than the threshold th;
b2, windowing the key points smaller than the threshold around the key points, and putting the image frames in the windows into the Open-pos network with modified input;
step B3, updating the key point coordinates obtained in the local frame by using the global information;
and 4, step 4: in order to further improve the robustness of the algorithm to shooting angles, target distance and recording process jitter factors, the key points are respectively set as the left hip C1Right hip C11Neck C8Performing coordinate conversion by taking the gravity center of the three coordinate points as an origin to obtain dataWhereinI.e. defining the origin of coordinates, i.e. C, for the transformed relative coordinates1、C11、C8Center of gravity of three points C0=(cx0,cy0) Wherein Updating all coordinate points by taking the original point coordinates as reference;
and 5: accumulating the coordinate matrix inputs obtained by each frame to obtain an accumulated coordinate matrix of each video segment, and performing window segmentation on the accumulated coordinate matrix by using a sliding window, namely setting the length of the sliding window to be the accumulated coordinate and the score obtained by k frames, and setting the step length to be 1;
step 6: a regression prediction model is constructed by utilizing a deep neural network, and the method comprises the following steps:
step C1, constructing a network model by fusing RNN through CNN, namely, fusing a bidirectional LSTM (BilsTM) model and a model with two convolution layers and a global pooling layer to construct a deep neural network model;
step C2: dividing data into a training set and a test set, training a training network model by using the test set, and storing a pre-training model;
step C3: and inputting the test data into the trained model to obtain a predicted result.
In step 6, the BilSTM is a bidirectional LSTM, the bidirectional LSTM is formed by two LSTMs which are superposed up and down, the output is determined by the shape of the two LSTMs together, and a recursive network calculates the hidden vector from front to backAnother recurrent neural network calculates the hidden vector from back to frontFinal outputThe first layer of the convolution layer is a one-dimensional convolution with convolution kernel of 5 multiplied by 5 and added with regularization, the second layer of the convolution layer is a convolution kernel of 3 multiplied by 3 and added with regularization, the third layer is a global pooling layer, and the number of filters in the convolution process is respectively 64 and 32;
adding a Dropout layer in the BilSTM layer, setting the last layer of activation function as a linear activation function in order to realize linear regression, and selecting a regression loss function Mean Square Error (MSE) loss function as a loss function; in order to accelerate the convergence of the network model, control overfitting to accelerate the convergence and control overfitting to add a batch normalization layer into each convolution layer, the batch normalization layer comprises the following calculation processes:
wherein B ═ { z ═ z1,...,m"is the input to the batch, m is the batch size, μBIs the average of the batch data and,is the variance of the batch process and,for normalized results, ε is a minimum value, hiAs a result of the scaling and shifting, γ and β are parameters learned in the input network model;
on the basis of the scheme: the convolutional layer is represented in the form of a layer when step 6 is performedWherein l is the number of layers,is the output of the jth neuron at level l,is the ith input of the l-th layer, anRepresents a convolution, wijIs the convolution kernel, bias is the bias term, MjFor a set of input feature maps, f (-) represents an activation function;
giving k continuous subsegments in the output of regression prediction to predict the key point coordinates and scores of the next frame, obtaining total scores according to the step 2, and adding the obtained frames into the current sequence by utilizing recursion to predict the data of the next frame;
the image windowing technology is that required target information is highlighted in an original image, and the purpose is to detect the coordinate position of a key point in the window;
when step 3 is executed, the area of the window is S, and the coordinates of the four points of the window are ld (kx)1,ky1)、lu(kx2,ky2)、rd(kx3,ky3)、ru(kx4,ky4);
The maximum area of the image windowing is determined by the processing speed d of a CPU and the total operation amount m of a software fusion algorithm, wherein the larger the d, the smaller the m, the larger the windowing area, otherwise, the smaller the area, and the background image area is S1The maximum area of the window is S2,S2=dS1/25;
The modified Open-position method adjusts the size format of input data into the length and width of a window, and outputs the coordinates of target information converted according to the origin coordinates.
While the invention has been described in further detail with reference to specific embodiments thereof, it is not intended that the invention be limited to the specific embodiments thereof; for those skilled in the art to which the present invention pertains and related technologies, the extension, operation method and data replacement should fall within the protection scope of the present invention based on the technical solution of the present invention.
Claims (4)
1. A human body continuous action rapid analysis method based on local identification is characterized by mainly comprising rope skipping video acquisition, video data preprocessing, coordinate point acquisition, coordinate point windowing re-identification, feature sequence construction, regression prediction model establishment and model calling, and specifically comprises the following steps:
step 1: collecting video stream data of the human body in the process of shaking the front face of the human body and jumping with two feet through a mobile phone, and recording and storing basic information of a target object, wherein the basic information comprises name, sex, age, height, weight and name of action;
step 2: preprocessing video data, and performing posture estimation on a human body of each frame in a video to obtain a key point coordinate, wherein the method comprises the following steps of:
a1, converting video data shot by different mobile phones into a uniform scale, setting the height of a video to be 530cm and the width to be 460 cm;
step A2: obtaining the coordinate positions of 14 joint points of the nose, the neck, the right shoulder, the right elbow, the right wrist, the left shoulder, the left elbow, the left wrist, the right hip, the right knee, the right ankle, the left hip, the left knee and the left ankle of each frame of human body in the video by utilizing an Open-position method, wherein the coordinate is represented as CPi=(xi,yi) Wherein i is from 1 to 14, wherein (x)i,yi) Coordinates of key points of the human body;
step A3: open-pos uses the Gaussian distance between predicted keypoints and true values to define confidence in keypoints and normalizes the confidence to [0, 1%]Is defined herein as the score of the keypoint and results in output results inputsi=(xi,yi,scorei);
Step A4: the scores of the 14 key points were averaged and standard deviations, and the sum of the average and standard deviation was used as the score result for the overall key point,
scoregeneral assembly=scoreAverage+scoreStandard deviation of;
And step 3: windowing and re-identifying the low predicted position of the key point by utilizing an image windowing technology, and improving the prediction accuracy of the key point by utilizing global information and local parts, wherein the method comprises the following steps:
step B1, obtaining inputs by using Open-posi=(xi,yi,scorei) Setting a threshold th for the score, and finding out key points smaller than the threshold th;
b2, windowing the key points smaller than the threshold around the key points, and putting the image frames in the windows into the Open-pos network with modified input;
step B3, updating the key point coordinates obtained in the local frame by using the global information;
and 4, step 4: in order to further improve the robustness of the algorithm to shooting angles, target distance and recording process jitter factors, the key points are respectively set as the left hip C1Right hip C11Neck C8Performing coordinate conversion by taking the gravity center of the three coordinate points as an origin to obtain dataWhereinI.e. the relative coordinates after conversion
Defining origin of coordinates, i.e. C1、C11、C8Center of gravity of three points C0=(cx0,cy0) WhereinUpdating all coordinate points by taking the original point coordinates as reference;
and 5: accumulating the coordinate matrix inputs obtained by each frame to obtain an accumulated coordinate matrix of each video segment, and performing window segmentation on the accumulated coordinate matrix by using a sliding window, namely setting the length of the sliding window to be the accumulated coordinate and the score obtained by k frames, and setting the step length to be 1;
step 6: a regression prediction model is constructed by utilizing a deep neural network, and the method comprises the following steps:
step C1, constructing a network model by fusing RNN through CNN, namely, fusing a bidirectional LSTM (BilsTM) model and a model with two convolution layers and a global pooling layer to construct a deep neural network model;
step C2: dividing data into a training set and a test set, training a training network model by using the test set, and storing a pre-training model;
step C3: and inputting the test data into the trained model to obtain a predicted result.
2. The method as claimed in claim 1, wherein the BilSTM is a bidirectional LSTM in step 6, the bidirectional LSTM has two LSTMs superimposed one on top of the other, and the output has a common shape decision of the two LSTMs, wherein a recursive network calculates the hidden vector from front to backAnother recurrent neural network calculates the hidden vector from back to frontFinal outputThe convolution layer first layer is a one-dimensional convolution with convolution kernel of 5 x 5 and added with regularization, the convolution kernel of the second layer is 3 x 3 and added with regularization, the third layer is a global pooling layer, and the number of filters in the convolution process is 64 and 32 respectively.
3. The method for rapidly analyzing human body continuous actions based on local recognition as claimed in claim 1, wherein a Dropout layer is added in a BilSTM layer, in order to realize linear regression, the last layer of activation function is set as a linear activation function, and the loss function selects a regression loss function Mean Square Error (MSE) loss function;
in order to accelerate the convergence of the network model, control overfitting to accelerate the convergence and control overfitting to add a batch normalization layer into each convolution layer, the batch normalization layer comprises the following calculation processes:
wherein B ═ { z ═ z1,...,m"is the input to the batch, m is the batch size, μBIs the average of the batch data and,is the variance of the batch process and,for normalized results, ε is a minimum value, hiAs a result of the scaling and shifting, γ and β are parameters learned in the input network model;
the convolutional layer is represented in the form of a layer when step 6 is performedWherein l is the number of layers,is the output of the jth neuron at level l,is the l-th layerThe ith input, andrepresents a convolution, wijIs the convolution kernel, bias is the bias term, MjFor a set of input feature maps, f (-) represents an activation function;
and (3) giving k continuous subsegments in the regression prediction output to predict the key point coordinates and the scores of the next frame, obtaining the total scores according to the step (2), and adding the obtained frames into the current sequence by utilizing recursion to predict the data of the next frame.
4. The method as claimed in claim 1, wherein the image windowing technique is used to highlight the required target information in the original image in order to detect the coordinate position of the key point in the window;
when step 3 is executed, the area of the window is S, and the coordinates of the four points of the window are ld (kx)1,ky1)、lu(kx2,ky2)、rd(kx3,ky3)、ru(kx4,ky4);
The maximum area of the image windowing is determined by the processing speed d of a CPU and the total operation amount m of a software fusion algorithm, wherein the larger the d, the smaller the m, the larger the windowing area, otherwise, the smaller the area, and the background image area is S1The maximum area of the window is S2,S2=dS1/25;
The modified Open-position method adjusts the size format of input data into the length and width of a window, and outputs the coordinates of target information converted according to the origin coordinates.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201911216130.8A CN110956141B (en) | 2019-12-02 | 2019-12-02 | Human body continuous action rapid analysis method based on local recognition |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201911216130.8A CN110956141B (en) | 2019-12-02 | 2019-12-02 | Human body continuous action rapid analysis method based on local recognition |
Publications (2)
Publication Number | Publication Date |
---|---|
CN110956141A true CN110956141A (en) | 2020-04-03 |
CN110956141B CN110956141B (en) | 2023-02-28 |
Family
ID=69979435
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201911216130.8A Active CN110956141B (en) | 2019-12-02 | 2019-12-02 | Human body continuous action rapid analysis method based on local recognition |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN110956141B (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112037310A (en) * | 2020-08-27 | 2020-12-04 | 成都先知者科技有限公司 | Game character action recognition generation method based on neural network |
CN112569564A (en) * | 2020-11-20 | 2021-03-30 | 融梦科技发展(北京)有限公司 | Rope skipping detection method, device and system and storage medium |
CN113095248A (en) * | 2021-04-19 | 2021-07-09 | 中国石油大学(华东) | Technical action correction method for badminton |
CN114596451A (en) * | 2022-04-01 | 2022-06-07 | 此刻启动(北京)智能科技有限公司 | Body fitness testing method and device based on AI vision and storage medium |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2017133009A1 (en) * | 2016-02-04 | 2017-08-10 | 广州新节奏智能科技有限公司 | Method for positioning human joint using depth image of convolutional neural network |
CN107643759A (en) * | 2016-07-22 | 2018-01-30 | 鹦鹉无人机股份有限公司 | From the autonomous system with target following and positioning of unmanned plane shooting mobile image |
CN109708627A (en) * | 2018-11-15 | 2019-05-03 | 上海航天控制技术研究所 | A kind of moving platform down space dynamic point target rapid detection method |
CN110503077A (en) * | 2019-08-29 | 2019-11-26 | 郑州大学 | A kind of real-time body's action-analysing method of view-based access control model |
-
2019
- 2019-12-02 CN CN201911216130.8A patent/CN110956141B/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2017133009A1 (en) * | 2016-02-04 | 2017-08-10 | 广州新节奏智能科技有限公司 | Method for positioning human joint using depth image of convolutional neural network |
CN107643759A (en) * | 2016-07-22 | 2018-01-30 | 鹦鹉无人机股份有限公司 | From the autonomous system with target following and positioning of unmanned plane shooting mobile image |
CN109708627A (en) * | 2018-11-15 | 2019-05-03 | 上海航天控制技术研究所 | A kind of moving platform down space dynamic point target rapid detection method |
CN110503077A (en) * | 2019-08-29 | 2019-11-26 | 郑州大学 | A kind of real-time body's action-analysing method of view-based access control model |
Non-Patent Citations (1)
Title |
---|
朱凌飞等: "基于骨架模型的人体行为分析", 《电子测量技术》 * |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112037310A (en) * | 2020-08-27 | 2020-12-04 | 成都先知者科技有限公司 | Game character action recognition generation method based on neural network |
CN112569564A (en) * | 2020-11-20 | 2021-03-30 | 融梦科技发展(北京)有限公司 | Rope skipping detection method, device and system and storage medium |
CN113095248A (en) * | 2021-04-19 | 2021-07-09 | 中国石油大学(华东) | Technical action correction method for badminton |
CN114596451A (en) * | 2022-04-01 | 2022-06-07 | 此刻启动(北京)智能科技有限公司 | Body fitness testing method and device based on AI vision and storage medium |
Also Published As
Publication number | Publication date |
---|---|
CN110956141B (en) | 2023-02-28 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110956141B (en) | Human body continuous action rapid analysis method based on local recognition | |
CN106650687B (en) | Posture correction method based on depth information and skeleton information | |
CN110991340B (en) | Human body action analysis method based on image compression | |
CN109460702B (en) | Passenger abnormal behavior identification method based on human body skeleton sequence | |
CN110956139B (en) | Human motion analysis method based on time sequence regression prediction | |
CN110503077B (en) | Real-time human body action analysis method based on vision | |
Chaudhari et al. | Yog-guru: Real-time yoga pose correction system using deep learning methods | |
CN110575663B (en) | Physical education auxiliary training method based on artificial intelligence | |
Zhou et al. | Learning to estimate 3d human pose from point cloud | |
CN111753747B (en) | Violent motion detection method based on monocular camera and three-dimensional attitude estimation | |
CN111582349B (en) | Improved target tracking algorithm based on YOLOv3 and kernel correlation filtering | |
CN106548194B (en) | The construction method and localization method of two dimensional image human joint points location model | |
Jia et al. | Two-stream temporal convolutional networks for skeleton-based human action recognition | |
Xu et al. | Robust hand gesture recognition based on RGB-D Data for natural human–computer interaction | |
CN109325408A (en) | A kind of gesture judging method and storage medium | |
CN108830170A (en) | A kind of end-to-end method for tracking target indicated based on layered characteristic | |
CN117541994A (en) | Abnormal behavior detection model and detection method in dense multi-person scene | |
CN111626152B (en) | Space-time line-of-sight direction estimation prototype design method based on Few-shot | |
CN109993116B (en) | Pedestrian re-identification method based on mutual learning of human bones | |
CN113408435B (en) | Security monitoring method, device, equipment and storage medium | |
CN111046715A (en) | Human body action comparison analysis method based on image retrieval | |
Hoang et al. | Advances in skeleton-based fall detection in RGB videos: From handcrafted to deep learning approaches | |
CN117137435B (en) | Rehabilitation action recognition method and system based on multi-mode information fusion | |
CN115205744A (en) | Intelligent exercise assisting method and device for figure skating | |
CN113065504A (en) | Behavior identification method and device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |