CN110555412A - End-to-end human body posture identification method based on combination of RGB and point cloud - Google Patents

End-to-end human body posture identification method based on combination of RGB and point cloud Download PDF

Info

Publication number
CN110555412A
CN110555412A CN201910836867.3A CN201910836867A CN110555412A CN 110555412 A CN110555412 A CN 110555412A CN 201910836867 A CN201910836867 A CN 201910836867A CN 110555412 A CN110555412 A CN 110555412A
Authority
CN
China
Prior art keywords
human body
point cloud
information
rgb
dimensional
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201910836867.3A
Other languages
Chinese (zh)
Other versions
CN110555412B (en
Inventor
张世雄
李楠楠
赵翼飞
李若尘
李革
安欣赏
张伟民
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Longgang Intelligent Audiovisual Research Institute
Original Assignee
Shenzhen Longgang Intelligent Audiovisual Research Institute
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Longgang Intelligent Audiovisual Research Institute filed Critical Shenzhen Longgang Intelligent Audiovisual Research Institute
Priority to CN201910836867.3A priority Critical patent/CN110555412B/en
Publication of CN110555412A publication Critical patent/CN110555412A/en
Application granted granted Critical
Publication of CN110555412B publication Critical patent/CN110555412B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/26Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
    • G06V10/267Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion by performing operations on regions, e.g. growing, shrinking or watersheds
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/60Type of objects
    • G06V20/64Three-dimensional objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/103Static body considered as a whole, e.g. static pedestrian or occupant recognition

Abstract

An end-to-end human body posture identification method based on RGB and point cloud combination comprises the following steps: 1): preprocessing the RGB information and the point cloud information; 2): extracting human two-dimensional (2D) skeleton information using a pre-network; and 3): three-dimensional (3D) skeleton information of a human body is extracted using a three-dimensional (3D) network. The end-to-end human body posture identification method based on the combination of RGB and point cloud provided by the invention can effectively extract an accurate 3D model of a human body from human body data acquired by RGB-D equipment. The method solves a series of problems in gesture recognition, such as gesture appearance change, gesture freedom diversity, similar gestures, self-shielding, great ambiguity of 2D gestures, insufficient 3D gesture precision, lack of 3D data sets and the like.

Description

end-to-end human body posture identification method based on combination of RGB and point cloud
Technical Field
The invention relates to a method for recognizing the gesture of key points of a human body through an RGB-D camera, in particular to an end-to-end human body gesture recognition method based on the combination of RGB and point cloud.
Background
The key point detection of human body posture is an important field of computer vision research, and research results are mainly used for a series of intelligent applications such as new generation human-computer interaction, Virtual Reality (VR) and Augmented Reality (AR) human-computer interaction, behavior recognition and analysis and the like. The traditional gesture recognition algorithm generally adopts a wearable acceleration sensor to recognize and detect the human gesture, so that the cost is high, and the wearing is complicated and active cooperation is required. The early detection of the human body posture based on the video is mainly based on template matching of hand-drawn features, the extraction mode of the hand-drawn features is complex in design, low in reliability, easy to be interfered by the outside, and poor in complex action recognition effect. Meanwhile, interference factors such as camera view angle, illumination, shielding and the like generally exist in a real scene, and the traditional method often has the problem that the identification accuracy is not high or cannot be identified in the real scene. With the continuous maturity of the application of deep learning in the field of computer vision, the method of deep learning is adopted for human posture recognition more and more. On the other hand, more and more three-dimensional (3D) acquisition devices are continuously developed, so that defects Of two-dimensional (2D) projection can be well compensated, including rotation, shielding, similarity and the like existing in human body postures, and for an RGB-D camera, three mainstream schemes are provided in the market at present, namely a 3D structured light scheme (structured light), a ToF 3D scheme (Time Of Flight, Time difference distance measurement technology) and a binocular stereo imaging scheme (StereoSystem). The three acquisition schemes can acquire a point cloud image with depth information, and according to different acquired data, the human body posture recognition can be divided into posture recognition with depth data, namely posture recognition based on three-dimensional (3D) point cloud, and human body posture recognition based on common image data (namely RGB data), namely posture recognition based on 2D images.
because the three-dimensional (3D) point cloud has low precision and contains more noise, the data volume is greatly increased, the dimensionality is larger, one dimension is more than that of a two-dimensional image, the calculation is complex, and the calculation amount is large. Meanwhile, the sparsity of the three-dimensional point cloud is also considered, the calculation efficiency is improved in the voxel-based reconstruction, excessive memory waste in an unoccupied part is avoided, the reconstruction resolution is improved, the network structure is improved to improve the reconstruction effect, and more details can be recovered.
The 2D image contains rich color information, is clear and contains more detailed information. The noise is small, the acquisition equipment is mature, but the spatial depth information is insufficient, so that ambiguity is easily caused. The human body posture of the 2D image generally has a certain unfixed problem. For a conventional three-dimensional mapping of 2D image body poses, one image body pose may often correspond to multiple different three-dimensional body poses. From a statistical point of view, a reasonable prediction of the input image forms a distribution. Reflected in the training set, the two human poses that appear similar in the image may be quite different in actual pose.
Disclosure of Invention
The invention provides an end-to-end human body posture identification method based on RGB and point cloud combination, which can effectively extract an accurate 3D model of a human body from human body data acquired by RGB-D equipment. The method solves a series of problems in gesture recognition, such as gesture appearance change, gesture freedom diversity, similar gestures, self-shielding, great ambiguity of 2D gestures, insufficient 3D gesture precision, lack of 3D data sets and the like.
The technical scheme provided by the invention is as follows:
the invention discloses an end-to-end human body posture identification method based on RGB and point cloud combination, which comprises the following steps: step 1): preprocessing the RGB information and the point cloud information; step 2): extracting human two-dimensional (2D) skeleton information using a pre-network; and step 3): three-dimensional (3D) skeleton information of a human body is extracted using a three-dimensional (3D) network.
In the method for recognizing the end-to-end human body posture based on the combination of RGB and point cloud, before preprocessing, an RGB-D camera is used as the acquisition input of signals, and the acquired signals are divided into RGB information and point cloud information.
In the method for recognizing the end-to-end human body posture based on the combination of RGB and point cloud, in the step 1), filtering and denoising preprocessing are respectively carried out on RGB information and point cloud information, and alignment processing is carried out.
In the method for recognizing the end-to-end human body posture based on the combination of RGB and point cloud, in step 1), a contour feature mapping method is used, a point cloud graph is taken as a coordinate reference, each edge salient feature is extracted, the feature points are mapped one by one, the offset { p1, p2, p3, · · · · · · · · · · · · · · · · · · · · } of each feature point is calculated, finally, the average offset p of all the feature points is calculated, and then RGB is projected to an affine space for conversion alignment.
In the method for recognizing the end-to-end human body posture based on the combination of RGB and point cloud, in the step 2), the preprocessed RGB information is input into a pre-network trained in advance to extract human body two-dimensional (2D) skeleton information, the extracted human body two-dimensional (2D) skeleton information (2D posture) and the point cloud information are input into a point cloud cutting module together, the extracted human body two-dimensional (2D) skeleton information is used for cutting a point cloud picture, and useless background information is removed.
In the method for recognizing the end-to-end human body posture based on the combination of RGB and point cloud, in the step 2), a single human body detection model from bottom to top is adopted by a front network, and the network pre-trained by mass data firstly detects two-dimensional (2D) key nodes of a human body, namely, the coordinates of all the joint nodes of the human body in one image are firstly detected, and then coordinate clustering is carried out to form key point coordinates corresponding to the human body.
In the method for recognizing the end-to-end human body posture based on the combination of RGB and point cloud, in the step 3), the point cloud information and the human body two-dimensional (2D) skeleton information are fused and then are simultaneously input into a three-dimensional (3D) network, and the trained three-dimensional (3D) network can extract accurate human body three-dimensional (3D) skeleton information from the point cloud information.
in the method for recognizing the end-to-end human body posture based on the combination of RGB and point cloud, in the step 3), a convolutional neural network is adopted in a three-dimensional (3D) network, the convolutional neural network is divided into three layers in total, the first two layers of networks are both connected to a pooling layer, and finally the three layers of networks are output through a full link layer.
Compared with the prior art, the invention has the beneficial effects that:
1. A unique mode of fusing two networks is designed, data of RGB and RGB-D types are fused at the same time, a strategy of fusing data in the middle is adopted, the problems of early noise and excessive early fused data of the data are effectively solved, and a novel method for detecting the human posture of the RGB-D camera is provided.
2. The model in the identification method can provide effective real 3D posture information of a human body, and different from the previous models which can only output 2D skeleton information, the model can output a 3D skeleton model with real world coordinate values, detailed data can be provided for the height and each trunk size of the human body, and the accuracy of centimeter level can be realized under the condition of higher accuracy of an RGB-D camera.
3. The 3D attitude can be estimated only according to the current frame, and the method is greatly improved compared with the conventional process of estimating the 3D attitude only by a plurality of continuous image frames or video frames.
drawings
The invention is further illustrated by way of example in the following with reference to the accompanying drawings:
FIG. 1 is a flow chart of an end-to-end human body gesture recognition method based on RGB and point cloud combination.
The specific implementation mode is as follows:
The invention effectively combines the gesture recognition of the 3D point cloud and the gesture recognition based on RGB image data, provides a deep learning method combining front and back networks, namely an end-to-end human body gesture recognition method based on the combination of RGB and point cloud, and simultaneously integrates the respective advantages of the point cloud image and the RGB image.
The end-to-end human body posture identification method based on the combination of RGB and point cloud comprises the following main steps:
1. And preprocessing the RGB information and the point cloud information. Before preprocessing, an RGB-D device (RGB-D camera) is used as a signal acquisition input, and the acquired signal is divided into RGB information and point cloud information. And then, respectively carrying out filtering and denoising pretreatment on the RGB information and the point cloud information, and carrying out alignment treatment. Specifically, since the RGB information and the point cloud information are acquired by different acquisition devices, the image information acquired by the two devices do not completely coincide, and there is a positional offset p between them, the present invention provides a method for contour feature mapping, which corrects the offset. The contour feature mapping method comprises the steps of firstly, taking a point cloud image as a coordinate reference, extracting each edge salient feature, mapping feature points one by one, calculating the offset { p1, p2, p3, · · · · · · · · · · · · · · · } of each feature point, finally calculating the average offset p of all feature points, and projecting RGB to an affine space for conversion alignment. The method has the advantages that the aligned point cloud information and the RGB information are matched in the plane space information, and the cutting of the point cloud information and the mutual fusion of the point cloud information and the RGB information are facilitated.
2. Human 2D skeleton information is extracted using a pre-network. In the step, the preprocessed RGB information is input into a pre-network trained in advance to extract human body 2D skeleton information, the extracted human body 2D skeleton information is used for cutting the point cloud picture, and useless background information is removed. The invention provides a convolutional neural network structure-based front-end network for 2D posture extraction, wherein activation functions used by all network layers are ReLu functions, and output posture information comprises 25 skeleton key point information, such as 25 key points of nose, head, shoulders and the like. The preposed network provided by the invention adopts a bottom-up single human body detection model, and the network pre-trained by a large amount of data firstly detects 2D key nodes of a human body, namely, coordinates of all joint nodes of the human body in an image are detected, and then coordinate clustering is carried out to form key point coordinates corresponding to the human body. The network for feature extraction adopts a convolutional neural network VGG-19 network to extract features, 3 networks with the size of 3 multiplied by 3 are accessed to predict confidence intervals of 25 joint points, and human body 2D skeleton information, namely a human body 2D skeleton diagram, is obtained according to the confidence intervals. The advantage here is that RGB information contains abundant color and context correlation information, and the skeleton precision of extracting is higher, and RGB information data acquisition is relatively easy moreover for training data is more, and the model that trains out is more accurate.
3. Human 3D skeletal information is extracted using a 3D network. Specifically, the point cloud information and the human body 2D skeleton information are fused and then simultaneously input into the 3D network, and the trained 3D network can extract accurate human body 3D skeleton information from the point cloud information. The invention provides a 3D posture estimation network of a 3D network, which adopts a 3-dimensional kernel convolution mode to construct a convolution neural network, wherein the convolution neural network is totally divided into three layers, the first two layers of networks are both connected into a pooling layer, and finally, the networks are output through a full connection layer. In the input process, as the input is two kinds of data, as shown in fig. 1, one kind of data is human body 2D skeleton information output by a front network, namely 2D posture, and the other kind of data is cut human body point cloud information, and normalization processing is performed on the point cloud and the 2D posture information at the same time, so that the numerical values of the two kinds of data are unified to a (-1, 1) interval, and then the 2D posture information (X, Y) is gradually merged into the 3D point cloud information (X, Y, Z), wherein the weight setting ratio of (X, Y) in the (X, Y) and the (X, Y) in the point cloud information of the 2D skeleton information is 10: 1, extracting a confidence region of the 3D skeleton by using 3 multiplied by 3, and finally outputting the 3D skeleton information through a full connection layer. The point cloud information and the RGB information are complementary in advantages, the point cloud has good spatial position information but is sparse, skeleton information cannot be extracted accurately, the RGB information contains rich information but lacks the spatial position information, and the point cloud information and the RGB information can be effectively fused by using a trained network.
The invention provides a network capable of accurately extracting 3D posture information of a human skeleton from end to end, which is characterized in that training is firstly carried out on a human3.6 data set in the early work, and then fine adjustment is carried out on the data set of the human body in the real world acquired by the network in combination with the self. After training is finished, the invention only needs to carry out reasoning of a forward end in the application process.
The flow chart of the end-to-end human body posture identification method based on the combination of RGB and point cloud is shown in figure 1, and the specific implementation flow is as follows:
1. Firstly, an RGB-D camera is used as the acquisition input of signals to acquire RGB-D data;
2. dividing the collected signals into RGB information and point cloud information;
3. Respectively inputting the RGB information and the point cloud information into a preprocessing module for filtering and denoising preprocessing, and carrying out alignment processing;
4. Inputting the preprocessed RGB information into a pre-network (Pose-net) trained in advance to extract human body 2D skeleton information, namely 2D posture;
5. Inputting the extracted human body 2D skeleton information (2D posture) and point cloud information into a point cloud cutting module, cutting the point cloud by using the extracted human body 2D skeleton information, and removing useless background information;
6. then fusing the point cloud information and human body 2D skeleton information (2D posture) and inputting the fused point cloud information and the human body 2D skeleton information into a 3D network at the same time, wherein the trained 3D network can extract accurate human body 3D skeleton information from the point cloud information, so that on one hand, the characteristics of high extraction accuracy and accurate positioning of key points of the 2D skeleton are effectively utilized, and on the other hand, the point cloud information is also utilized to carry out effective geometric constraint on the finally generated 3D human body posture;
And 7.3D network output human body 3D skeleton accurate model.
An RGB-D camera is an acquisition device that can simultaneously acquire a point cloud image and an RGB color image. The method adopts an end-to-end deep neural network method, and simultaneously adopts a scheme of mutually fusing the RGB image and the point cloud image, so that the limitation that the prior gesture recognition singly depends on the RGB image or the point cloud image is overcome, the human body gesture extraction scheme which considers the common 2D space image characteristic and the deep 3D space characteristic is adopted, the recognition precision is improved, and the angle ambiguity of the gesture recognition of a single image is eliminated.
to sum up, the invention provides an effective fully supervised deep learning network model, which provides two network level extraction: one is a front-end network (pos-net) for extracting a skeletal pose of a human body, and the other is a 3D network for extracting 3D pose information by combining skeletal information with point cloud. The deep learning network model provided by the invention can effectively extract an accurate 3D model of a human body from human body data acquired by RGB-D equipment. Unlike the conventional 3D conversion model, the 3D information here is 3D data information including real human body, and the 3D information in the conventional 2D converted 3D human body is often matched by the model, so the obtained 3D data is not real, and there is ambiguity due to the angle of the camera and the distance to the camera. Aiming at the problem, the accurate 3D human body skeleton is obtained by combining 2D-3D model conversion and depth point cloud information.
Although the present invention has been described in detail with reference to the foregoing embodiments, those skilled in the art will understand that: any person skilled in the art can modify or easily conceive the technical solutions described in the foregoing embodiments or equivalent substitutes for some technical features within the technical scope of the present disclosure; such modifications, changes or substitutions do not depart from the spirit and scope of the embodiments of the present invention, and they should be construed as being included therein. Therefore, the protection scope of the present invention shall be subject to the protection scope of the claims.

Claims (8)

1. An end-to-end human body posture identification method based on RGB and point cloud combination is characterized by comprising the following steps:
Step 1): preprocessing the RGB information and the point cloud information;
Step 2): extracting human two-dimensional (2D) skeleton information using a pre-network; and
Step 3): three-dimensional (3D) skeleton information of a human body is extracted using a three-dimensional (3D) network.
2. The method for recognizing the human body posture end-to-end based on the combination of RGB and point cloud as claimed in claim 1, wherein in step 1), before the preprocessing, an RGB-D camera is firstly used as the acquisition input of the signal, and the acquired signal is divided into RGB information and point cloud information.
3. the end-to-end human body posture identifying method based on RGB and point cloud combination as claimed in claim 1, in step 1), filtering and denoising preprocessing are respectively carried out on the RGB information and the point cloud information, and alignment processing is carried out.
4. the end-to-end human body posture identification method based on RGB and point cloud combination as claimed in claim 1, in step 1), using a contour feature mapping method, taking a point cloud image as a coordinate reference, extracting each edge salient feature, performing one-to-one mapping on feature points, calculating the offset { p1, p2, p3, … … } of each feature point, finally calculating the average offset p of all feature points, and projecting RGB to an affine space for conversion and alignment.
5. The end-to-end human body posture identification method based on the combination of RGB and point cloud as claimed in claim 1, in step 2), inputting the preprocessed RGB information into a pre-network trained in advance to extract the human body two-dimensional (2D) skeleton information, inputting the extracted human body two-dimensional (2D) skeleton information (2D posture) and point cloud information into a point cloud cutting module together, and cutting a point cloud picture by using the extracted human body two-dimensional (2D) skeleton information to remove useless background information.
6. The end-to-end human body posture identification method based on the combination of RGB and point cloud as claimed in claim 1, in step 2), the front network adopts a bottom-up single human body detection model, and the network pre-trained by mass data first detects two-dimensional (2D) key nodes of the human body, that is, detects the coordinates of all human body joint points in an image, and then carries out coordinate clustering to form key point coordinates corresponding to the human body.
7. The end-to-end human body posture recognition method based on the combination of RGB and point cloud as claimed in claim 1, in step 3), the point cloud information and the human body two-dimensional (2D) skeleton information are fused and then input into a three-dimensional (3D) network, and the trained three-dimensional (3D) network extracts accurate human body three-dimensional (3D) skeleton information from the point cloud information.
8. The method for recognizing the human body posture end-to-end based on the combination of the RGB and the point cloud as claimed in claim 1, in the step 3), the three-dimensional (3D) network adopts a convolutional neural network, the convolutional neural network is divided into three layers in total, wherein the first two layers of networks are both connected to a layer pooling layer and are finally output through a full connection layer.
CN201910836867.3A 2019-09-05 2019-09-05 End-to-end human body gesture recognition method based on combination of RGB and point cloud Active CN110555412B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910836867.3A CN110555412B (en) 2019-09-05 2019-09-05 End-to-end human body gesture recognition method based on combination of RGB and point cloud

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910836867.3A CN110555412B (en) 2019-09-05 2019-09-05 End-to-end human body gesture recognition method based on combination of RGB and point cloud

Publications (2)

Publication Number Publication Date
CN110555412A true CN110555412A (en) 2019-12-10
CN110555412B CN110555412B (en) 2023-05-16

Family

ID=68739207

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910836867.3A Active CN110555412B (en) 2019-09-05 2019-09-05 End-to-end human body gesture recognition method based on combination of RGB and point cloud

Country Status (1)

Country Link
CN (1) CN110555412B (en)

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111597974A (en) * 2020-05-14 2020-08-28 哈工大机器人(合肥)国际创新研究院 Monitoring method and system based on TOF camera for personnel activities in carriage
CN111723687A (en) * 2020-06-02 2020-09-29 北京的卢深视科技有限公司 Human body action recognition method and device based on neural network
CN111723688A (en) * 2020-06-02 2020-09-29 北京的卢深视科技有限公司 Human body action recognition result evaluation method and device and electronic equipment
CN112070835A (en) * 2020-08-21 2020-12-11 达闼机器人有限公司 Mechanical arm pose prediction method and device, storage medium and electronic equipment
CN112907672A (en) * 2021-05-07 2021-06-04 上海擎朗智能科技有限公司 Robot avoidance method and device, electronic equipment and storage medium
CN114091601A (en) * 2021-11-18 2022-02-25 业成科技(成都)有限公司 Sensor fusion method for detecting personnel condition
CN114694263A (en) * 2022-05-30 2022-07-01 深圳智华科技发展有限公司 Action recognition method, device, equipment and storage medium
WO2022217828A1 (en) * 2021-04-15 2022-10-20 青岛小鸟看看科技有限公司 Gesture recognition and control method and apparatus, and virtual reality device
CN115471561A (en) * 2022-11-14 2022-12-13 科大讯飞股份有限公司 Object key point positioning method, cleaning robot control method and related equipment
TWI789267B (en) * 2022-03-10 2023-01-01 國立臺中科技大學 Method of using two-dimensional image to automatically create ground truth data required for training three-dimensional pointnet

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140226854A1 (en) * 2013-02-13 2014-08-14 Lsi Corporation Three-Dimensional Region of Interest Tracking Based on Key Frame Matching
CN104715493A (en) * 2015-03-23 2015-06-17 北京工业大学 Moving body posture estimating method
CN107180226A (en) * 2017-04-28 2017-09-19 华南理工大学 A kind of dynamic gesture identification method based on combination neural net
CN108830150A (en) * 2018-05-07 2018-11-16 山东师范大学 One kind being based on 3 D human body Attitude estimation method and device
CN109086683A (en) * 2018-07-11 2018-12-25 清华大学 A kind of manpower posture homing method and system based on cloud semantically enhancement

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140226854A1 (en) * 2013-02-13 2014-08-14 Lsi Corporation Three-Dimensional Region of Interest Tracking Based on Key Frame Matching
CN104715493A (en) * 2015-03-23 2015-06-17 北京工业大学 Moving body posture estimating method
CN107180226A (en) * 2017-04-28 2017-09-19 华南理工大学 A kind of dynamic gesture identification method based on combination neural net
CN108830150A (en) * 2018-05-07 2018-11-16 山东师范大学 One kind being based on 3 D human body Attitude estimation method and device
CN109086683A (en) * 2018-07-11 2018-12-25 清华大学 A kind of manpower posture homing method and system based on cloud semantically enhancement

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111597974B (en) * 2020-05-14 2023-05-12 哈工大机器人(合肥)国际创新研究院 Monitoring method and system for personnel activities in carriage based on TOF camera
CN111597974A (en) * 2020-05-14 2020-08-28 哈工大机器人(合肥)国际创新研究院 Monitoring method and system based on TOF camera for personnel activities in carriage
CN111723687A (en) * 2020-06-02 2020-09-29 北京的卢深视科技有限公司 Human body action recognition method and device based on neural network
CN111723688A (en) * 2020-06-02 2020-09-29 北京的卢深视科技有限公司 Human body action recognition result evaluation method and device and electronic equipment
CN111723688B (en) * 2020-06-02 2024-03-12 合肥的卢深视科技有限公司 Human body action recognition result evaluation method and device and electronic equipment
CN112070835A (en) * 2020-08-21 2020-12-11 达闼机器人有限公司 Mechanical arm pose prediction method and device, storage medium and electronic equipment
US11947729B2 (en) 2021-04-15 2024-04-02 Qingdao Pico Technology Co., Ltd. Gesture recognition method and device, gesture control method and device and virtual reality apparatus
WO2022217828A1 (en) * 2021-04-15 2022-10-20 青岛小鸟看看科技有限公司 Gesture recognition and control method and apparatus, and virtual reality device
CN112907672A (en) * 2021-05-07 2021-06-04 上海擎朗智能科技有限公司 Robot avoidance method and device, electronic equipment and storage medium
CN114091601A (en) * 2021-11-18 2022-02-25 业成科技(成都)有限公司 Sensor fusion method for detecting personnel condition
CN114091601B (en) * 2021-11-18 2023-05-05 业成科技(成都)有限公司 Sensor fusion method for detecting personnel condition
TWI789267B (en) * 2022-03-10 2023-01-01 國立臺中科技大學 Method of using two-dimensional image to automatically create ground truth data required for training three-dimensional pointnet
CN114694263A (en) * 2022-05-30 2022-07-01 深圳智华科技发展有限公司 Action recognition method, device, equipment and storage medium
CN114694263B (en) * 2022-05-30 2022-09-02 深圳智华科技发展有限公司 Action recognition method, device, equipment and storage medium
CN115471561A (en) * 2022-11-14 2022-12-13 科大讯飞股份有限公司 Object key point positioning method, cleaning robot control method and related equipment

Also Published As

Publication number Publication date
CN110555412B (en) 2023-05-16

Similar Documents

Publication Publication Date Title
CN110555412B (en) End-to-end human body gesture recognition method based on combination of RGB and point cloud
US11703951B1 (en) Gesture recognition systems
US10109055B2 (en) Multiple hypotheses segmentation-guided 3D object detection and pose estimation
CN106648103B (en) A kind of the gesture tracking method and VR helmet of VR helmet
CN109934848B (en) Method for accurately positioning moving object based on deep learning
Ye et al. Accurate 3d pose estimation from a single depth image
CN108734194B (en) Virtual reality-oriented single-depth-map-based human body joint point identification method
Boisvert et al. Three-dimensional human shape inference from silhouettes: reconstruction and validation
CN108256504A (en) A kind of Three-Dimensional Dynamic gesture identification method based on deep learning
CN105759967B (en) A kind of hand overall situation attitude detecting method based on depth data
CN110852182A (en) Depth video human body behavior recognition method based on three-dimensional space time sequence modeling
CN103839277A (en) Mobile augmented reality registration method of outdoor wide-range natural scene
CN111160291A (en) Human eye detection method based on depth information and CNN
CN111476089B (en) Pedestrian detection method, system and terminal for multi-mode information fusion in image
CN106815855A (en) Based on the human body motion tracking method that production and discriminate combine
CN112668550B (en) Double interaction behavior recognition method based on joint point-depth joint attention RGB modal data
CN110751097A (en) Semi-supervised three-dimensional point cloud gesture key point detection method
CN109670401A (en) A kind of action identification method based on skeleton motion figure
Li et al. Deep learning based monocular depth prediction: Datasets, methods and applications
Chai et al. Human gait recognition: approaches, datasets and challenges
Xu et al. 3D joints estimation of the human body in single-frame point cloud
Swadzba et al. Tracking objects in 6D for reconstructing static scenes
Chang et al. Multi-view 3d human pose estimation with self-supervised learning
CN116129051A (en) Three-dimensional human body posture estimation method and system based on graph and attention interleaving
CN108694348B (en) Tracking registration method and device based on natural features

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant