CN112036324A - A method and system for human posture determination for complex multi-person scenes - Google Patents
A method and system for human posture determination for complex multi-person scenes Download PDFInfo
- Publication number
- CN112036324A CN112036324A CN202010905107.6A CN202010905107A CN112036324A CN 112036324 A CN112036324 A CN 112036324A CN 202010905107 A CN202010905107 A CN 202010905107A CN 112036324 A CN112036324 A CN 112036324A
- Authority
- CN
- China
- Prior art keywords
- key point
- point set
- posture
- included angle
- human body
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 51
- 238000001514 detection method Methods 0.000 claims abstract description 5
- 230000036544 posture Effects 0.000 claims description 125
- 210000002414 leg Anatomy 0.000 claims description 42
- 210000000689 upper leg Anatomy 0.000 claims description 32
- 238000012549 training Methods 0.000 claims description 17
- 238000000605 extraction Methods 0.000 claims description 16
- 238000004364 calculation method Methods 0.000 claims description 10
- 238000013527 convolutional neural network Methods 0.000 claims description 9
- 230000008569 process Effects 0.000 claims description 8
- 238000005452 bending Methods 0.000 claims description 7
- 238000013528 artificial neural network Methods 0.000 claims description 3
- 238000012216 screening Methods 0.000 claims description 2
- 230000011218 segmentation Effects 0.000 claims 1
- 206010000117 Abnormal behaviour Diseases 0.000 abstract 1
- 210000004394 hip joint Anatomy 0.000 description 10
- 210000000629 knee joint Anatomy 0.000 description 8
- 210000000544 articulatio talocruralis Anatomy 0.000 description 6
- 238000010586 diagram Methods 0.000 description 6
- 238000012544 monitoring process Methods 0.000 description 4
- 210000004247 hand Anatomy 0.000 description 3
- 230000033001 locomotion Effects 0.000 description 3
- 238000003909 pattern recognition Methods 0.000 description 3
- 238000013473 artificial intelligence Methods 0.000 description 2
- 238000004422 calculation algorithm Methods 0.000 description 2
- 210000002310 elbow joint Anatomy 0.000 description 2
- 238000003709 image segmentation Methods 0.000 description 2
- 238000005070 sampling Methods 0.000 description 2
- 210000000323 shoulder joint Anatomy 0.000 description 2
- 238000012706 support-vector machine Methods 0.000 description 2
- 210000003857 wrist joint Anatomy 0.000 description 2
- 230000009471 action Effects 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 230000008094 contradictory effect Effects 0.000 description 1
- 238000012217 deletion Methods 0.000 description 1
- 230000037430 deletion Effects 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000018109 developmental process Effects 0.000 description 1
- 238000006073 displacement reaction Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 239000004744 fabric Substances 0.000 description 1
- 230000004807 localization Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000000877 morphologic effect Effects 0.000 description 1
- 230000000474 nursing effect Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000011897 real-time detection Methods 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/103—Static body considered as a whole, e.g. static pedestrian or occupant recognition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- General Physics & Mathematics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- General Engineering & Computer Science (AREA)
- Evolutionary Computation (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Software Systems (AREA)
- Mathematical Physics (AREA)
- Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- General Health & Medical Sciences (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Computational Biology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Human Computer Interaction (AREA)
- Multimedia (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
Abstract
Description
技术领域technical field
本发明属于模式识别和人工智能技术领域,具体涉及一种鲁棒的人体姿态判定方法,特别涉及一种用于复杂多人场景的人体姿态判定方法及系统。The invention belongs to the technical field of pattern recognition and artificial intelligence, in particular to a robust human body posture determination method, in particular to a human body posture determination method and system for complex multi-person scenes.
背景技术Background technique
视频监控设备广泛部署在了各种公共场所,如:幼儿园、养老院、车站、商场等。这些设备在日常运行中会产生海量的监控数据,仅靠人工监视视频场景的方式显然不切实际。监控场景中人体姿态的语义描述,即判定不同的人体姿态,可以帮助人们快速了解场景中个体的状态以及发生的事件,对实时检测意外和突发状况意义重大。Video surveillance equipment is widely deployed in various public places, such as: kindergartens, nursing homes, stations, shopping malls, etc. These devices generate massive amounts of surveillance data in their daily operations, and it is obviously impractical to monitor video scenes manually. The semantic description of human posture in monitoring scenes, that is, determining different human postures, can help people quickly understand the state of individuals in the scene and the events that occur, which is of great significance for real-time detection of accidents and emergencies.
传统的人工监控方式不仅效率低下,而且难以及时对于一些突发状况做出迅速的反应,随着人工智能以及模式识别技术的发展,出现了一些人体姿态的识别方法。根据是否将人体视为“连杆”模型,人体姿态识别方法大致可以分为两类:基于视频特征提取的方法和基于人体关键点提取的方法。下面将对这两类方法进行展开讨论。The traditional manual monitoring method is not only inefficient, but also difficult to respond quickly to some emergencies in time. With the development of artificial intelligence and pattern recognition technology, some methods of human posture recognition have emerged. According to whether the human body is regarded as a "link" model, human pose recognition methods can be roughly divided into two categories: methods based on video feature extraction and methods based on human body key point extraction. Both types of methods are discussed below.
在基于视频特征提取的方法中,文献“Action recognition by densetrajectories[C].IEEE Conference on Computer Vision and Pattern Recognition”提出利用光流场跟踪密集的采样点来获得人体运动的轨迹,进而完成人体姿态的识别。该方法为获取人体运动轨迹需要实时跟踪采样点,而在出现行人交错的场景,这种跟踪很容易出错。文献“Human posture recognition based on projection histogram and SupportVector Machine”利用椭圆拟合人体轮廓,然后沿着椭圆长轴和短轴方向构建直方图用于描述人体形状,最后采用支持向量机识别人体姿态。对于无法准确提取边界框的情况,比如:多人相互咬合或者遮挡,该方法提取的人体形状会发生严重变化,导致姿态识别性能急剧下降。文献“A bio-inspired event-based size and position invariant humanposture recognition algorithm”提出了一种基于简化线段Hausdorff距离的人体姿态识别方法。该方法把视频序列中两个连续的图像帧作为输入,通过比较两帧之间的差异得到活动对象,并将活动对象的轮廓分解为矢量线段。该方法简化了计算,提高了效率,但是无法检测到前后帧之间不动身体部位导致识别的准确率下降。文献“Posture recognitioninvariant to background,cloth textures,body size,and camera distance usingmorphological geometry”利用提取的人体轮廓的长度和宽度识别人体姿态,避免了穿着背景等细节信息的影响。由于人体轮廓长度和宽度只能对粗略的描述人体姿态,因此该方法识别姿态的准确率不高。基于视频特征提取的方法在判断人体姿态之前需要输入视频是对目标进行跟踪拍摄的或者首先通过跟踪算法提取目标的运动过程,导致此类方法很难适用于目标跟踪困难人数较多的复杂场景。In the method based on video feature extraction, the document "Action recognition by densetrajectories[C].IEEE Conference on Computer Vision and Pattern Recognition" proposes to use optical flow field to track dense sampling points to obtain the trajectory of human motion, and then complete the detection of human posture. identify. This method requires real-time tracking of sampling points to obtain human motion trajectories, and this tracking is prone to errors in scenes where pedestrians intersect. The document "Human posture recognition based on projection histogram and SupportVector Machine" uses an ellipse to fit the outline of the human body, then constructs a histogram along the long and short axes of the ellipse to describe the shape of the human body, and finally uses a support vector machine to recognize the human posture. For situations where the bounding box cannot be accurately extracted, such as multiple people occluding or occluding each other, the shape of the human body extracted by this method will change severely, resulting in a sharp drop in the performance of gesture recognition. The document "A bio-inspired event-based size and position invariant humanposture recognition algorithm" proposes a method for human posture recognition based on simplified line segment Hausdorff distance. The method takes two consecutive image frames in a video sequence as input, obtains the moving object by comparing the difference between the two frames, and decomposes the outline of the moving object into vector line segments. This method simplifies the calculation and improves the efficiency, but the inability to detect the moving body parts between the frames leads to a decrease in the accuracy of the recognition. The paper "Posture recognition invariant to background, cloth textures, body size, and camera distance using morphological geometry" uses the length and width of the extracted human silhouette to recognize human posture, avoiding the influence of detailed information such as wearing background. Since the length and width of the human body outline can only roughly describe the human posture, the accuracy of the method in identifying the posture is not high. The method based on video feature extraction needs to input the video to track the target before judging the human posture, or first extract the movement process of the target through the tracking algorithm, which makes it difficult for such methods to be suitable for complex scenes with a large number of people who have difficulty in target tracking.
在基于人体关键点提取的方法中,文献“Realtime Multi-person 2D PoseEstimation Using Part Affinity Fields”提出了一种同时预测身体部位位置和各部位位置关系的模型来提取人体关键点;文献“A Coarse-Fine Network for KeypointLocalization”提出了一种由粗到精多级监督的网络CFN(Coarse-Fine Network)用于提取人体关键点。这类方法无需跟踪目标,因此能够适用于多人场景中的人体姿态识别。文献“Neural Network Approach for 2-Dimension Person Pose Estimation With EncodedMask and Keypoint Detection”利用卷积深度神经网络从图像分割蒙版中提取关键点,学习关键点之间的相互连接关系,实现了将图像分割与自底而上策略相结合来提取人体关键点。基于人体关键点提取的方法大多只是提取特征点,却没做出人体姿态的判定。然而在实际应用中,往往只得到人体的特征点是不够的,给出姿态的判定结果可以帮助人们快速地了解场景中个体的状态以及发生的事件,同时也是智能监控系统自动地对事态做出进一步分析和判断的重要依据。因此,出现了一些根据关键点信息判定人体姿态的方法。文献“Human posture classification using skeleton information”利用人体关键点之间的距离以及关键点连线的夹角判定人体姿态。但是在实际应用中经常会出现提取的关键点存在误差以及由于遮挡造成的关键点缺失等问题,这会使得现有姿态判定方法性能的急剧下降甚至失效。In the method based on human body key point extraction, the document "Realtime Multi-person 2D PoseEstimation Using Part Affinity Fields" proposes a model that simultaneously predicts the position of body parts and the positional relationship of each part to extract human key points; the document "A Coarse- Fine Network for Keypoint Localization" proposes a coarse-to-fine multi-level supervised network CFN (Coarse-Fine Network) for extracting human key points. This kind of method does not need to track the target, so it can be suitable for human gesture recognition in multi-person scenes. The document "Neural Network Approach for 2-Dimension Person Pose Estimation With EncodedMask and Keypoint Detection" uses convolutional deep neural network to extract key points from image segmentation masks, learn the mutual connection between key points, and realizes the image segmentation and A bottom-up strategy is combined to extract human keypoints. Most of the methods based on human body key point extraction only extract feature points, but do not make a judgment of human body posture. However, in practical applications, it is often not enough to only obtain the feature points of the human body. Giving the judgment result of the posture can help people quickly understand the state of the individual in the scene and the events that have occurred, and it is also the intelligent monitoring system that automatically makes decisions about the situation. important basis for further analysis and judgment. Therefore, there are some methods to determine the human pose based on the key point information. The document "Human posture classification using skeleton information" uses the distance between the key points of the human body and the angle between the key points to determine the posture of the human body. However, in practical applications, there are often problems such as errors in the extracted key points and missing key points due to occlusion, which will make the performance of the existing attitude determination methods drop sharply or even fail.
发明内容SUMMARY OF THE INVENTION
本发明的目的在于针对上述现有技术存在的问题,提供一种用于复杂多人场景的人体姿态判定方法及系统,从而快速的了解场景中每个个体的状态以及发生的事件。The purpose of the present invention is to provide a method and system for judging human body posture in a complex multi-person scene, so as to quickly understand the state of each individual in the scene and the events that occur.
实现本发明目的的技术解决方案为:一种用于复杂多人场景的人体姿态判定方法,所述方法包括以下步骤:The technical solution to achieve the purpose of the present invention is: a method for determining human body posture for complex multi-person scenes, the method comprises the following steps:
步骤1,检测人体关键点,并划分关键点集合;Step 1, detect human key points, and divide the key point set;
步骤2,提取各集合中人体关键点几何分布的统计特征;Step 2, extracting the statistical features of the geometric distribution of the human body key points in each set;
步骤3,根据统计特征判定人体姿态。Step 3: Determine the posture of the human body according to the statistical features.
进一步地,步骤1所述检测人体关键点,并划分关键点集合,具体过程包括:Further, the human body key points are detected in step 1, and the key point set is divided, and the specific process includes:
步骤1-1,构建样本训练集,该集合包括标注好人体关键点的若干人体图像;Step 1-1, construct a sample training set, the set includes several human body images marked with key points of the human body;
步骤1-2,利用所述样本训练集训练深度卷积神经网络;Step 1-2, using the sample training set to train a deep convolutional neural network;
步骤1-3,利用训练好的深度卷积神经网络检测待检测人体图像中的关键点;Steps 1-3, using the trained deep convolutional neural network to detect key points in the human image to be detected;
步骤1-4,划分关键点集合,具体为:Steps 1-4, divide the set of key points, specifically:
剔除非人体躯干的关键点,包括手部,形成全身关键点集合;Eliminate key points of non-human torso, including hands, to form a set of whole body key points;
将人体躯干关键点划分为上半身关键点集合、下半身关键点集合、左臂关键点集合和右臂关键点集合;Divide the key points of the human torso into a set of upper body key points, a set of lower body key points, a set of left arm key points and a set of right arm key points;
将下半身关键点继续划分为左腿关键点集合、右腿关键点集合、左侧大腿关键点集合以及右侧大腿关键点集合。The key points of the lower body are further divided into a set of key points of the left leg, a set of key points of the right leg, a set of key points of the left thigh, and a set of key points of the right thigh.
进一步地,步骤2所述提取各集合中人体关键点几何分布的统计特征,具体过程包括:Further, in step 2, the statistical features of the geometric distribution of the key points of the human body in each set are extracted, and the specific process includes:
步骤2-1-1,计算各关键点集合的凸包;Step 2-1-1, calculate the convex hull of each key point set;
步骤2-1-2,从构成凸包的关键点中找出距离最大的两个关键点;Step 2-1-2, find the two key points with the largest distance from the key points forming the convex hull;
步骤2-1-3,计算所述两个关键点连线沿顺时针方向旋转与水平方向的夹角,作为该凸包的水平角度。Step 2-1-3: Calculate the angle between the clockwise rotation and the horizontal direction of the connecting line of the two key points as the horizontal angle of the convex hull.
进一步地,步骤3所述根据统计特征判定人体姿态,具体为:根据各关键点集合凸包的水平角度,判定手臂、腿部以及全身的姿态:Further, in step 3, determining the posture of the human body according to the statistical features is specifically: according to the horizontal angle of the convex hull of each key point set, determining the posture of the arms, legs and the whole body:
假设全身关键点集合凸包的水平夹角为α0;上半身关键点集合凸包水平夹角为α1;下半身关键点集合凸包的水平夹角为α2;左臂关键点集合凸包的水平夹角为α3;右臂关键点集合凸包的水平夹角为α4;左腿关键点集合凸包的水平夹角为α5;右腿关键点集合凸包的水平夹角为α6;左侧大腿关键点集合凸包的水平夹角为α7;右侧大腿关键点集合凸包的水平夹角为α8;Assume that the horizontal included angle of the convex hull of the whole body key point set is α 0 ; the horizontal included angle of the upper body key point set convex hull is α 1 ; the horizontal included angle of the lower body key point set convex hull is α 2 ; the left arm key point set convex hull is α 2 ; The horizontal included angle is α 3 ; the horizontal included angle of the convex hull of the right arm key point set is α 4 ; the horizontal included angle of the left leg key point set convex hull is α 5 ; the horizontal included angle of the right leg key point set convex hull is α 6 ; the horizontal included angle of the convex hull of the left thigh key point set is α 7 ; the horizontal included angle of the convex hull of the right thigh key point set is α 8 ;
若|tanα3|∈(a3,b3)或|tanα4|∈(a3,b3),则判定为抬臂姿态;若|tanα5|∈(a5,b5)或|tanα6|∈(a5,b5),则判定为踢腿姿态;若|tanα0|∈(a0,b0)、|tanα1|∈(a1,b1)且|tanα2|∈(a2,b2),则判定为站立姿态;若|tanα7|∈[a7,b7]或|tanα8|∈[a7,b7],则判定为蹲坐姿态;若|tanα0|∈[c0,d0]、|tanα1|∈[c1,d1]且|tanα2|∈[c2,d2],则判定为平躺姿态。If |tanα 3 |∈(a 3 ,b 3 ) or |tanα 4 |∈(a 3 ,b 3 ), it is judged to be a raised arm posture; if |tanα 5 |∈(a 5 ,b 5 ) or |tanα 6 |∈(a 5 ,b 5 ), it is determined as a kicking posture; if |tanα 0 |∈(a 0 ,b 0 ), |tanα 1 |∈(a 1 ,b 1 ) and |tanα 2 |∈ (a 2 , b 2 ), it is judged as standing posture; if |tanα 7 |∈[a 7 ,b 7 ] or |tanα 8 |∈[a 7 ,b 7 ], it is judged as squatting posture; if | tanα 0 |∈[c 0 ,d 0 ], |tanα 1 |∈[c 1 ,d 1 ] and |tanα 2 |∈[c 2 ,d 2 ], then it is determined to be a lying posture.
进一步地,步骤2所述提取各集合中人体关键点几何分布的统计特征,具体过程包括:Further, in step 2, the statistical features of the geometric distribution of the key points of the human body in each set are extracted, and the specific process includes:
步骤2-2-1,计算各关键点集合的最小外接矩形;Step 2-2-1, calculate the minimum circumscribed rectangle of each key point set;
步骤2-2-2,分别计算最小外接矩形的长边和短边沿顺时针方向旋转与水平方向构成的夹角;Step 2-2-2, respectively calculate the angle formed by the clockwise rotation of the long side and the short side of the minimum circumscribed rectangle and the horizontal direction;
步骤2-2-3,取两个夹角中的较小值作为最小外接矩形的旋转角度。Step 2-2-3, take the smaller of the two included angles as the rotation angle of the minimum circumscribed rectangle.
进一步地,步骤3所述根据统计特征判定人体姿态,具体为:根据各关键点集合最小外接矩形的旋转角度,判定手臂、腿部以及全身的姿态:Further, in step 3, determining the posture of the human body according to the statistical features is specifically: determining the postures of the arms, legs and the whole body according to the rotation angle of the minimum circumscribed rectangle of each key point set:
假设全身关键点集合最小外接矩形的水平夹角为β0;上半身关键点集合最小外接矩形水平夹角为β1;下半身关键点集合最小外接矩形的水平夹角为β2;左臂关键点集合最小外接矩形的水平夹角为β3;右臂关键点集合最小外接矩形的水平夹角为β4;左腿关键点集合最小外接矩形的水平夹角为β5;右腿关键点集合最小外接矩形的水平夹角为β6;左侧大腿关键点集合最小外接矩形的水平夹角为β7;右侧大腿关键点集合最小外接矩形的水平夹角为β8;Assume that the horizontal included angle of the minimum circumscribed rectangle of the whole body key point set is β 0 ; the horizontal included angle of the minimum circumscribed rectangle of the upper body key point set is β 1 ; the horizontal included angle of the minimum circumscribed rectangle of the lower body key point set is β 2 ; the left arm key point set The horizontal included angle of the minimum circumscribed rectangle is β 3 ; the horizontal included angle of the minimum circumscribed rectangle of the right arm key point set is β 4 ; the horizontal included angle of the minimum circumscribed rectangle of the left leg key point set is β 5 ; The horizontal included angle of the rectangle is β 6 ; the horizontal included angle of the minimum circumscribed rectangle of the left thigh key point set is β 7 ; the horizontal included angle of the minimum circumscribed rectangle of the right thigh key point set is β 8 ;
若或则判定为抬臂姿态;若或则判定为踢腿姿态;若且则判定为站立姿态;若β1|∈[λ1,γ1],则判定为弯腰姿态;若或则判定为蹲坐姿态;若|β0|∈[θ0,Θ0]、β1|∈[θ1,Θ1]且β2|∈[θ2,Θ2],则判定为平躺姿态。like or It is judged to be the arm raising posture; if or It is judged as kicking posture; if and It is judged as standing posture; if β 1 |∈[λ 1 ,γ 1 ], it is judged as bending posture; or If |β 0 |∈[θ 0 ,Θ 0 ], β 1 |∈[θ 1 ,Θ 1 ] and β 2 |∈[θ 2 ,Θ 2 ], then it is determined as lying down attitude.
一种用于复杂多人场景的人体姿态判定系统,所述系统包括:A human body posture determination system for complex multi-person scenes, the system includes:
关键点划分模块,用于检测人体关键点,并划分关键点集合;The key point division module is used to detect the key points of the human body and divide the set of key points;
统计特征提取模块,用于提取各集合中人体关键点几何分布的统计特征;The statistical feature extraction module is used to extract the statistical features of the geometric distribution of the key points of the human body in each set;
姿态判定模块,用于根据统计特征判定人体姿态。The posture determination module is used to determine the posture of the human body according to the statistical features.
本发明与现有技术相比,其显著优点为:1)本发明采用了划分人体关键点集合的思想,相对于基于视频特征提取的方法,避免了目标提取的难题,可以适用于多人场景下的人体姿态判定;2)相对于现有基于人体关键点提取的方法,将人体姿态判定问题转换成了关键点集合的统计关系,使得本发明对于部分关键点的缺失和移位不敏感,即使在部分身体被遮挡导致关键点不准确的情况下,依然能够根据集合的统计几个特征得到正确的姿态判定结果,从而大大提高了姿态判定的鲁棒性;3)本发明的判定速度较快,可以实现监控范围内人体姿态的实时判定,帮助人们快速了解当前场景中的个体状态以及发生的事件。Compared with the prior art, the present invention has the following significant advantages: 1) The present invention adopts the idea of dividing the human body key point set, compared with the method based on video feature extraction, it avoids the difficult problem of target extraction, and can be applied to multi-person scenarios 2) Compared with the existing method based on human body key point extraction, the problem of human body posture determination is converted into the statistical relationship of the key point set, so that the present invention is insensitive to the deletion and displacement of some key points, Even if some of the body is occluded and the key points are inaccurate, the correct attitude determination result can still be obtained according to the statistical features of the set, thereby greatly improving the robustness of attitude determination; 3) The determination speed of the present invention is relatively high. Fast, it can realize real-time determination of human body posture within the monitoring range, and help people quickly understand the individual status and events in the current scene.
下面结合附图对本发明作进一步详细描述。The present invention will be described in further detail below with reference to the accompanying drawings.
附图说明Description of drawings
图1为一个实施例中用于复杂多人场景的人体姿态判定方法的流程图。FIG. 1 is a flowchart of a method for determining human posture in a complex multi-person scene in one embodiment.
图2为一个实施例中人体关键点集合的划分示意图。FIG. 2 is a schematic diagram of division of a human body key point set in one embodiment.
图3为一个实施例中计算关键点集合凸包的水平角度示意图。FIG. 3 is a schematic diagram of a horizontal angle of calculating a convex hull of a set of key points in one embodiment.
图4为一个实施例中计算关键点集合最小外接矩形的旋转角度示意图。FIG. 4 is a schematic diagram of calculating the rotation angle of the minimum circumscribed rectangle of the key point set in one embodiment.
图5为一个实施例中单个人基本姿态的判定结果图;其中,人体关键点用圆点标注;姿态的判定结果标注在了人的头部,图(a)至(c)分别为站立姿态、蹲坐姿态和弯腰姿态。Fig. 5 is the judgment result diagram of the basic posture of a single person in one embodiment; wherein, the key points of the human body are marked with dots; the judgment result of the posture is marked on the head of the person, and Figures (a) to (c) are standing postures respectively. , squatting posture and bending posture.
图6为一个实施例中单个人复杂姿态的判定结果图,其中,人体关键点用圆点标注,姿态的判定结果标注在人的头部,图(a)至(c)分别为站立抬臂姿态、站立踢腿姿态和站立抬臂踢腿姿态。Fig. 6 is the judgment result diagram of the complex posture of a single person in one embodiment, wherein, the key points of the human body are marked with dots, and the judgment result of the posture is marked on the head of the person, and Figures (a) to (c) are standing and raising arms respectively. Stance, Standing Kick Stance, and Standing Arm Kick Stance.
图7为一个实施例中多人复杂姿态的判定结果图,其中,人体关键点用圆点进行了标注,姿态的判定结果标注在人的头部。FIG. 7 is a diagram showing the judgment result of complex postures of multiple people in one embodiment, wherein the key points of the human body are marked with dots, and the judgment result of the posture is marked on the head of the person.
具体实施方式Detailed ways
为了使本申请的目的、技术方案及优点更加清楚明白,以下结合附图及实施例,对本申请进行进一步详细说明。应当理解,此处描述的具体实施例仅仅用以解释本申请,并不用于限定本申请。In order to make the purpose, technical solutions and advantages of the present application more clearly understood, the present application will be described in further detail below with reference to the accompanying drawings and embodiments. It should be understood that the specific embodiments described herein are only used to explain the present application, but not to limit the present application.
需要说明,若本发明实施例中有涉及“第一”、“第二”等的描述,则该“第一”、“第二”等的描述仅用于描述目的,而不能理解为指示或暗示其相对重要性或者隐含指明所指示的技术特征的数量。由此,限定有“第一”、“第二”的特征可以明示或者隐含地包括至少一个该特征。另外,各个实施例之间的技术方案可以相互结合,但是必须是以本领域普通技术人员能够实现为基础,当技术方案的结合出现相互矛盾或无法实现时应当认为这种技术方案的结合不存在,也不在本发明要求的保护范围之内。It should be noted that if there are descriptions involving "first", "second", etc. in the embodiments of the present invention, the descriptions of "first", "second", etc. are only used for description purposes, and should not be understood as instructions or Implicit their relative importance or implicitly indicate the number of technical features indicated. Thus, a feature delimited with "first", "second" may expressly or implicitly include at least one of that feature. In addition, the technical solutions between the various embodiments can be combined with each other, but must be based on the realization by those of ordinary skill in the art. When the combination of technical solutions is contradictory or cannot be realized, it should be considered that the combination of such technical solutions does not exist. , is not within the scope of protection required by the present invention.
在一个实施例中,结合图1,提供了一种用于复杂多人场景的人体姿态判定方法,所述方法包括以下步骤:In one embodiment, with reference to FIG. 1 , a method for determining human posture in a complex multi-person scene is provided, and the method includes the following steps:
步骤1,检测人体关键点,并划分关键点集合;Step 1, detect human key points, and divide the key point set;
步骤2,提取各集合中人体关键点几何分布的统计特征;Step 2, extracting the statistical features of the geometric distribution of the human body key points in each set;
步骤3,根据统计特征判定人体姿态。Step 3: Determine the posture of the human body according to the statistical features.
进一步地,在其中一个实施例中,步骤1所述检测人体关键点,并划分关键点集合,具体过程包括:Further, in one of the embodiments, the human body key points are detected in step 1, and the key point set is divided, and the specific process includes:
步骤1-1,构建样本训练集,该集合包括标注好人体关键点的若干人体图像;Step 1-1, construct a sample training set, the set includes several human body images marked with key points of the human body;
步骤1-2,利用所述样本训练集训练深度卷积神经网络;Step 1-2, using the sample training set to train a deep convolutional neural network;
步骤1-3,利用训练好的深度卷积神经网络检测待检测人体图像中的关键点;Steps 1-3, using the trained deep convolutional neural network to detect key points in the human image to be detected;
步骤1-4,结合图2,划分关键点集合,具体为:Steps 1-4, combined with Figure 2, divide the set of key points, specifically:
剔除非人体躯干的关键点,包括手部,形成全身关键点集合;Eliminate key points of non-human torso, including hands, to form a set of whole body key points;
将人体躯干关键点划分为上半身关键点集合、下半身关键点集合、左臂关键点集合和右臂关键点集合;Divide the key points of the human torso into a set of upper body key points, a set of lower body key points, a set of left arm key points and a set of right arm key points;
将下半身关键点继续划分为左腿关键点集合、右腿关键点集合、左侧大腿关键点集合以及右侧大腿关键点集合。The key points of the lower body are further divided into a set of key points of the left leg, a set of key points of the right leg, a set of key points of the left thigh, and a set of key points of the right thigh.
这里,采用文献“Realtime Multi-person 2D Pose Estimation Using PartAffinity Fields”中的方法提取人体关键点。Here, the method in the document "Realtime Multi-person 2D Pose Estimation Using PartAffinity Fields" is used to extract human key points.
这里,人体躯干关键点集合包括左眼、右眼、嘴部、颈部、左髋关节、右髋关节、左膝关节、右膝关节、左踝关节、右踝关节,共十个关键点;Here, the set of key points of the human torso includes ten key points in total: left eye, right eye, mouth, neck, left hip joint, right hip joint, left knee joint, right knee joint, left ankle joint, and right ankle joint;
上半身关键点集合包括左眼、右眼、嘴部、颈部、左髋关节、右髋关节,共六个关键点;The upper body key point collection includes six key points: left eye, right eye, mouth, neck, left hip joint, and right hip joint;
下半身关键点集合包括左髋关节、右髋关节、左膝关节、右膝关节、左踝关节、右踝关节,共六个关键点;The lower body key point set includes six key points in total: left hip joint, right hip joint, left knee joint, right knee joint, left ankle joint, and right ankle joint;
左臂关键点集合包括左腕关节、左肘关节、左肩关节,共三个关键点;The set of key points of the left arm includes the left wrist joint, the left elbow joint, and the left shoulder joint, a total of three key points;
右臂关键点集合包括右腕关节、右肘关节、右肩关节,共三个关键点;The right arm key point set includes the right wrist joint, right elbow joint, right shoulder joint, a total of three key points;
左腿关键点集合包括左踝关节、左膝关节、左髋关节,共三个关键点;The set of key points of the left leg includes the left ankle joint, the left knee joint, and the left hip joint, a total of three key points;
右腿关键点集合包括右踝关节、右膝关节、右髋关节,共三个关键点;The set of key points of the right leg includes the right ankle joint, the right knee joint, and the right hip joint, a total of three key points;
左侧大腿关键点集合包括左膝关节、左髋关节,共两个关键点;The set of key points of the left thigh includes the left knee joint and the left hip joint, a total of two key points;
右侧大腿关键点集合包括右膝关节、右髋关节,共两个关键点。The right thigh key point set includes the right knee joint and the right hip joint, a total of two key points.
进一步地,在其中一个实施例中,结合图3,步骤2所述提取各集合中人体关键点几何分布的统计特征,具体过程包括:Further, in one of the embodiments, with reference to FIG. 3 , in step 2, the statistical features of the geometric distribution of the human body key points in each set are extracted, and the specific process includes:
步骤2-1-1,计算各关键点集合的凸包;Step 2-1-1, calculate the convex hull of each key point set;
步骤2-1-2,从构成凸包的关键点中找出距离最大的两个关键点;Step 2-1-2, find the two key points with the largest distance from the key points forming the convex hull;
步骤2-1-3,计算所述两个关键点连线沿顺时针方向旋转与水平方向的夹角,作为该凸包的水平角度。Step 2-1-3: Calculate the angle between the clockwise rotation and the horizontal direction of the connecting line of the two key points as the horizontal angle of the convex hull.
进一步地,在其中一个实施例中,步骤3所述根据统计特征判定人体姿态,具体为:根据各关键点集合凸包的水平角度,判定手臂、腿部以及全身的姿态:Further, in one of the embodiments, the step 3 determines the posture of the human body according to the statistical features, specifically: according to the horizontal angle of the convex hull of each key point set, the posture of the arm, the leg and the whole body is determined:
假设全身关键点集合凸包的水平夹角为α0;上半身关键点集合凸包水平夹角为α1;下半身关键点集合凸包的水平夹角为α2;左臂关键点集合凸包的水平夹角为α3;右臂关键点集合凸包的水平夹角为α4;左腿关键点集合凸包的水平夹角为α5;右腿关键点集合凸包的水平夹角为α6;左侧大腿关键点集合凸包的水平夹角为α7;右侧大腿关键点集合凸包的水平夹角为α8;Assume that the horizontal included angle of the convex hull of the whole body key point set is α 0 ; the horizontal included angle of the upper body key point set convex hull is α 1 ; the horizontal included angle of the lower body key point set convex hull is α 2 ; the left arm key point set convex hull is α 2 ; The horizontal included angle is α 3 ; the horizontal included angle of the convex hull of the right arm key point set is α 4 ; the horizontal included angle of the left leg key point set convex hull is α 5 ; the horizontal included angle of the right leg key point set convex hull is α 6 ; the horizontal included angle of the convex hull of the left thigh key point set is α 7 ; the horizontal included angle of the convex hull of the right thigh key point set is α 8 ;
若|tanα3|∈(a3,b3)或|tanα4|∈(a3,b3),则判定为抬臂姿态;若|tanα5|∈(a5,b5)或|tanα6|∈(a5,b5),则判定为踢腿姿态;若|tanα0|∈(a0,b0)、|tanα1|∈(a1,b1)且|tanα2|∈(a2,b2),则判定为站立姿态;若|tanα7|∈[a7,b7]或|tanα8|∈[a7,b7],则判定为蹲坐姿态;若|tanα0|∈[c0,d0]、|tanα1|∈[c1,d1]且|tanα2|∈[c2,d2],则判定为平躺姿态。If |tanα 3 |∈(a 3 ,b 3 ) or |tanα 4 |∈(a 3 ,b 3 ), it is judged to be a raised arm posture; if |tanα 5 |∈(a 5 ,b 5 ) or |tanα 6 |∈(a 5 ,b 5 ), it is determined as a kicking posture; if |tanα 0 |∈(a 0 ,b 0 ), |tanα 1 |∈(a 1 ,b 1 ) and |tanα 2 |∈ (a 2 , b 2 ), it is judged as standing posture; if |tanα 7 |∈[a 7 ,b 7 ] or |tanα 8 |∈[a 7 ,b 7 ], it is judged as squatting posture; if | tanα 0 |∈[c 0 ,d 0 ], |tanα 1 |∈[c 1 ,d 1 ] and |tanα 2 |∈[c 2 ,d 2 ], then it is determined to be a lying posture.
这里优选地,(a3,b3)=(0.25,+∞),(a5,b5)=(0.25,5),(a0,b0)=(a1,b1)=(a2,b2)=(4,+∞),[a7,b7]=[0,0.25],[c0,d0]=[c1,d1]=[c2,d2]=[0,0.25]。Here preferably, (a 3 ,b 3 )=(0.25,+∞), (a 5 ,b 5 )=(0.25,5), (a 0 ,b 0 )=(a 1 ,b 1 )=( a 2 ,b 2 )=(4,+∞), [a 7 ,b 7 ]=[0,0.25], [c 0 ,d 0 ]=[c 1 ,d 1 ]=[c 2 ,d 2 ]=[0,0.25].
进一步地,在其中一个实施例中,结合图4,步骤2所述提取各集合中人体关键点几何分布的统计特征,具体过程包括:Further, in one of the embodiments, with reference to FIG. 4 , in step 2, the statistical features of the geometric distribution of the human body key points in each set are extracted, and the specific process includes:
步骤2-2-1,计算各关键点集合的最小外接矩形;Step 2-2-1, calculate the minimum circumscribed rectangle of each key point set;
步骤2-2-2,分别计算最小外接矩形的长边和短边沿顺时针方向旋转与水平方向构成的夹角;Step 2-2-2, respectively calculate the angle formed by the clockwise rotation of the long side and the short side of the minimum circumscribed rectangle and the horizontal direction;
步骤2-2-3,取两个夹角中的较小值作为最小外接矩形的旋转角度。Step 2-2-3, take the smaller of the two included angles as the rotation angle of the minimum circumscribed rectangle.
进一步地,在其中一个实施例中,步骤3所述根据统计特征判定人体姿态,具体为:根据各关键点集合最小外接矩形的旋转角度,判定手臂、腿部以及全身的姿态:Further, in one of the embodiments, the step 3 determines the posture of the human body according to the statistical features, specifically: according to the rotation angle of the minimum circumscribed rectangle of each key point set, the posture of the arm, the leg and the whole body is determined:
假设全身关键点集合最小外接矩形的水平夹角为β0;上半身关键点集合最小外接矩形水平夹角为β1;下半身关键点集合最小外接矩形的水平夹角为β2;左臂关键点集合最小外接矩形的水平夹角为β3;右臂关键点集合最小外接矩形的水平夹角为β4;左腿关键点集合最小外接矩形的水平夹角为β5;右腿关键点集合最小外接矩形的水平夹角为β6;左侧大腿关键点集合最小外接矩形的水平夹角为β7;右侧大腿关键点集合最小外接矩形的水平夹角为β8;Assume that the horizontal included angle of the minimum circumscribed rectangle of the whole body key point set is β 0 ; the horizontal included angle of the minimum circumscribed rectangle of the upper body key point set is β 1 ; the horizontal included angle of the minimum circumscribed rectangle of the lower body key point set is β 2 ; the left arm key point set The horizontal included angle of the minimum circumscribed rectangle is β 3 ; the horizontal included angle of the minimum circumscribed rectangle of the right arm key point set is β 4 ; the horizontal included angle of the minimum circumscribed rectangle of the left leg key point set is β 5 ; The horizontal included angle of the rectangle is β 6 ; the horizontal included angle of the minimum circumscribed rectangle of the left thigh key point set is β 7 ; the horizontal included angle of the minimum circumscribed rectangle of the right thigh key point set is β 8 ;
若或则判定为抬臂姿态;若或则判定为踢腿姿态;若且则判定为站立姿态;若β1|∈[λ1,γ1],则判定为弯腰姿态;若或则判定为蹲坐姿态;若|β0|∈[θ0,Θ0]、β1|∈[θ1,Θ1]且β2|∈[θ2,Θ2],则判定为平躺姿态。like or It is judged to be the arm raising posture; if or It is judged as kicking posture; if and It is judged as standing posture; if β 1 |∈[λ 1 ,γ 1 ], it is judged as bending posture; or If |β 0 |∈[θ 0 ,Θ 0 ], β 1 |∈[θ 1 ,Θ 1 ] and β 2 |∈[θ 2 ,Θ 2 ], then it is determined as lying down attitude.
这里优选地,[λ1,γ1]=[0°,25°],[θ0,Θ0]=[θ1,Θ1]=[θ2,Θ2]=[0°,25°]。Here preferably, [λ 1 ,γ 1 ]=[0°, 25°], [θ 0 , Θ 0 ]=[θ 1 , Θ 1 ]=[θ 2 , Θ 2 ]=[0°, 25°].
在一个实施例中,提供了一种用于复杂多人场景的人体姿态判定系统,所述系统包括:In one embodiment, a system for determining human posture in a complex multi-person scene is provided, the system comprising:
关键点划分模块,用于检测人体关键点,并划分关键点集合;The key point division module is used to detect the key points of the human body and divide the set of key points;
统计特征提取模块,用于提取各集合中人体关键点几何分布的统计特征;The statistical feature extraction module is used to extract the statistical features of the geometric distribution of the key points of the human body in each set;
姿态判定模块,用于根据统计特征判定人体姿态。The posture determination module is used to determine the posture of the human body according to the statistical features.
进一步地,在其中一个实施例中,所述关键点划分模块包括:Further, in one of the embodiments, the key point division module includes:
训练集构建单元,用于构建样本训练集,该集合包括标注好人体关键点的若干人体图像;The training set construction unit is used to construct a sample training set, the set includes several human body images marked with key points of the human body;
训练单元,用于利用所述样本训练集训练深度卷积神经网络;a training unit for training a deep convolutional neural network using the sample training set;
关键点检测单元,利用训练好的深度卷积神经网络检测待检测人体图像中的关键点;The key point detection unit uses the trained deep convolutional neural network to detect the key points in the human image to be detected;
划分单元,用于划分关键点集合,具体包括:The division unit is used to divide the set of key points, including:
第一划分子单元,用于剔除非人体躯干的关键点,包括手部,形成全身关键点集合;The first division subunit is used to eliminate key points of non-human torso, including hands, to form a set of key points of the whole body;
第二划分子单元,用于将人体躯干关键点划分为上半身关键点集合、下半身关键点集合、左臂关键点集合和右臂关键点集合;The second division subunit is used to divide the key points of the human torso into a set of key points of the upper body, a set of key points of the lower body, a set of key points of the left arm and a set of key points of the right arm;
第三划分子单元,用于将下半身关键点继续划分为左腿关键点集合、右腿关键点集合、左侧大腿关键点集合以及右侧大腿关键点集合。The third division subunit is used to further divide the key points of the lower body into a set of key points of the left leg, a set of key points of the right leg, a set of key points of the left thigh, and a set of key points of the right thigh.
进一步地,在其中一个实施例中,所述统计特征提取模块包括:Further, in one of the embodiments, the statistical feature extraction module includes:
第一计算单元,用于计算各关键点集合的凸包;a first calculation unit, used to calculate the convex hull of each key point set;
关键点筛选单元,用于从构成凸包的关键点中找出距离最大的两个关键点;The key point screening unit is used to find the two key points with the largest distance from the key points that constitute the convex hull;
第二计算单元,用于计算所述两个关键点连线沿顺时针方向旋转与水平方向的夹角,作为该凸包的水平角度;The second calculation unit is used to calculate the angle between the clockwise rotation and the horizontal direction of the connecting line of the two key points, as the horizontal angle of the convex hull;
所述姿态判定模块,用于根据各关键点集合凸包的水平角度,判定手臂、腿部以及全身的姿态:The attitude determination module is used to determine the attitude of the arms, legs and the whole body according to the horizontal angle of the convex hull of each key point set:
假设全身关键点集合凸包的水平夹角为α0;上半身关键点集合凸包水平夹角为α1;下半身关键点集合凸包的水平夹角为α2;左臂关键点集合凸包的水平夹角为α3;右臂关键点集合凸包的水平夹角为α4;左腿关键点集合凸包的水平夹角为α5;右腿关键点集合凸包的水平夹角为α6;左侧大腿关键点集合凸包的水平夹角为α7;右侧大腿关键点集合凸包的水平夹角为α8;Assume that the horizontal included angle of the convex hull of the whole body key point set is α 0 ; the horizontal included angle of the upper body key point set convex hull is α 1 ; the horizontal included angle of the lower body key point set convex hull is α 2 ; the left arm key point set convex hull is α 2 ; The horizontal included angle is α 3 ; the horizontal included angle of the convex hull of the right arm key point set is α 4 ; the horizontal included angle of the left leg key point set convex hull is α 5 ; the horizontal included angle of the right leg key point set convex hull is α 6 ; the horizontal included angle of the convex hull of the left thigh key point set is α 7 ; the horizontal included angle of the convex hull of the right thigh key point set is α 8 ;
若|tanα3|∈(a3,b3)或|tanα4|∈(a3,b3),则判定为抬臂姿态;若|tanα5|∈(a5,b5)或|tanα6|∈(a5,b5),则判定为踢腿姿态;若|tanα0|∈(a0,b0)、|tanα1|∈(a1,b1)且|tanα2|∈(a2,b2),则判定为站立姿态;若|tanα7|∈[a7,b7]或|tanα8|∈[a7,b7],则判定为蹲坐姿态;若|tanα0|∈[c0,d0]、|tanα1|∈[c1,d1]且|tanα2|∈[c2,d2],则判定为平躺姿态。If |tanα 3 |∈(a 3 ,b 3 ) or |tanα 4 |∈(a 3 ,b 3 ), it is judged to be a raised arm posture; if |tanα 5 |∈(a 5 ,b 5 ) or |tanα 6 |∈(a 5 ,b 5 ), it is determined as a kicking posture; if |tanα 0 |∈(a 0 ,b 0 ), |tanα 1 |∈(a 1 ,b 1 ) and |tanα 2 |∈ (a 2 , b 2 ), it is judged as standing posture; if |tanα 7 |∈[a 7 ,b 7 ] or |tanα 8 |∈[a 7 ,b 7 ], it is judged as squatting posture; if | tanα 0 |∈[c 0 ,d 0 ], |tanα 1 |∈[c 1 ,d 1 ] and |tanα 2 |∈[c 2 ,d 2 ], then it is determined to be a lying posture.
进一步地,在其中一个实施例中,所述统计特征提取模块包括:Further, in one of the embodiments, the statistical feature extraction module includes:
第三计算单元,用于计算各关键点集合的最小外接矩形;The third calculation unit is used to calculate the minimum circumscribed rectangle of each key point set;
第四计算单元,用于分别计算最小外接矩形的长边和短边沿顺时针方向旋转与水平方向构成的夹角;The fourth calculation unit is used to calculate the angle formed by the clockwise rotation of the long side and the short side of the minimum circumscribed rectangle and the horizontal direction respectively;
第五计算单元,用于取两个夹角中的较小值作为最小外接矩形的旋转角度;The fifth calculation unit is used to take the smaller value of the two included angles as the rotation angle of the minimum circumscribed rectangle;
所述姿态判定模块,用于根据各关键点集合最小外接矩形的旋转角度,判定手臂、腿部以及全身的姿态:The attitude determination module is used to determine the attitude of the arm, the leg and the whole body according to the rotation angle of the minimum circumscribed rectangle of each key point set:
假设全身关键点集合最小外接矩形的水平夹角为β0;上半身关键点集合最小外接矩形水平夹角为β1;下半身关键点集合最小外接矩形的水平夹角为β2;左臂关键点集合最小外接矩形的水平夹角为β3;右臂关键点集合最小外接矩形的水平夹角为β4;左腿关键点集合最小外接矩形的水平夹角为β5;右腿关键点集合最小外接矩形的水平夹角为β6;左侧大腿关键点集合最小外接矩形的水平夹角为β7;右侧大腿关键点集合最小外接矩形的水平夹角为β8;Assume that the horizontal included angle of the minimum circumscribed rectangle of the whole body key point set is β 0 ; the horizontal included angle of the minimum circumscribed rectangle of the upper body key point set is β 1 ; the horizontal included angle of the minimum circumscribed rectangle of the lower body key point set is β 2 ; the left arm key point set The horizontal included angle of the minimum circumscribed rectangle is β 3 ; the horizontal included angle of the minimum circumscribed rectangle of the right arm key point set is β 4 ; the horizontal included angle of the minimum circumscribed rectangle of the left leg key point set is β 5 ; The horizontal included angle of the rectangle is β 6 ; the horizontal included angle of the minimum circumscribed rectangle of the left thigh key point set is β 7 ; the horizontal included angle of the minimum circumscribed rectangle of the right thigh key point set is β 8 ;
若或则判定为抬臂姿态;若或则判定为踢腿姿态;若且则判定为站立姿态;若β1|∈[λ1,γ1],则判定为弯腰姿态;若或则判定为蹲坐姿态;若|β0|∈[θ0,Θ0]、β1|∈[θ1,Θ1]且β2|∈[θ2,Θ2],则判定为平躺姿态。like or It is judged to be the arm raising posture; if or It is judged as kicking posture; if and It is judged as standing posture; if β 1 |∈[λ 1 ,γ 1 ], it is judged as bending posture; or If |β 0 |∈[θ 0 ,Θ 0 ], β 1 |∈[θ 1 ,Θ 1 ] and β 2 |∈[θ 2 ,Θ 2 ], then it is determined as lying down attitude.
关于用于复杂多人场景的人体姿态判定系统的具体限定可以参见上文中对于用于复杂多人场景的人体姿态判定方法的限定,在此不再赘述。上述用于复杂多人场景的人体姿态判定系统中的各个模块可全部或部分通过软件、硬件及其组合来实现。上述各模块可以硬件形式内嵌于或独立于计算机设备中的处理器中,也可以以软件形式存储于计算机设备中的存储器中,以便于处理器调用执行以上各个模块对应的操作。For the specific limitation of the human body posture determination system used in the complex multi-person scene, please refer to the definition of the human body posture determination method used in the complex multi-person scene above, which will not be repeated here. Each module in the above-mentioned human posture determination system for complex multi-person scenarios can be implemented in whole or in part by software, hardware and combinations thereof. The above modules can be embedded in or independent of the processor in the computer device in the form of hardware, or stored in the memory in the computer device in the form of software, so that the processor can call and execute the operations corresponding to the above modules.
作为一种具体示例,在其中一个实施例中,对本发明进一步验证说明,利用上述最小外界矩形的方法判定监控视频中的人体姿态,测试结果如图5、图6、图7和表1所示。As a specific example, in one of the embodiments, the present invention is further verified and explained, the above-mentioned method of minimum external rectangle is used to determine the human body posture in the surveillance video, and the test results are shown in Figure 5, Figure 6, Figure 7 and Table 1 .
表1 人体姿态判定结果统计Table 1 Statistics of human posture determination results
由图5单个人多种基本姿态的判定结果和图6单人多种复杂姿态的判定结果可以看出,本发明能够较为准确的判定单个人的姿态。由图7存在相互遮挡的多人复杂姿态的判定结果可以看出,本发明能够同时判定多人的姿态,同时对遮挡等影响不敏感。如表1所示,本发明判定人体姿态的平均正确率为92.833%,其中,对于站立、蹲坐、抬臂、平躺等区分度较明显的姿态判别的正确率较高,而对于弯腰、踢腿等易混淆的姿态判别的正确率稍低。It can be seen from the judgment results of multiple basic postures of a single person in FIG. 5 and the judgment results of multiple complex postures of a single person in FIG. 6 that the present invention can more accurately determine the posture of a single person. It can be seen from the determination result of complex postures of multiple people with mutual occlusion in FIG. 7 that the present invention can simultaneously determine the postures of multiple people, and is insensitive to influences such as occlusion. As shown in Table 1, the present invention has an average correct rate of 92.833% in determining the human body posture. Among them, the correct rate for judging postures with obvious distinctions such as standing, squatting, raising arms, and lying down is higher, while for bending over , kicking and other confusing gestures, the accuracy rate is slightly lower.
以上显示和描述了本发明的基本原理、主要特征及优点。本行业的技术人员应该了解,本发明不受上述实施例的限制,上述实施例和说明书中描述的只是说明本发明的原理,在不脱离本发明精神和范围的前提下,本发明还会有各种变化和改进,这些变化和改进都落入要求保护的本发明范围内。本发明要求保护范围由所附的权利要求书及其等效物界定。The foregoing has shown and described the basic principles, main features and advantages of the present invention. Those skilled in the art should understand that the present invention is not limited by the above-mentioned embodiments, and the descriptions in the above-mentioned embodiments and the description are only to illustrate the principle of the present invention. Without departing from the spirit and scope of the present invention, the present invention will have Various changes and modifications fall within the scope of the claimed invention. The claimed scope of the present invention is defined by the appended claims and their equivalents.
Claims (10)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010905107.6A CN112036324B (en) | 2020-09-01 | 2020-09-01 | A human body posture determination method and system for complex multi-person scenes |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010905107.6A CN112036324B (en) | 2020-09-01 | 2020-09-01 | A human body posture determination method and system for complex multi-person scenes |
Publications (2)
Publication Number | Publication Date |
---|---|
CN112036324A true CN112036324A (en) | 2020-12-04 |
CN112036324B CN112036324B (en) | 2025-02-14 |
Family
ID=73590874
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010905107.6A Active CN112036324B (en) | 2020-09-01 | 2020-09-01 | A human body posture determination method and system for complex multi-person scenes |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112036324B (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113537234A (en) * | 2021-06-10 | 2021-10-22 | 浙江大华技术股份有限公司 | Quantity counting method and device, electronic device and computer equipment |
CN114155518A (en) * | 2021-11-08 | 2022-03-08 | 西安西光产业发展有限公司 | Expressway visor inclination identification method based on deep semantic segmentation network and image correction |
Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR20140114741A (en) * | 2013-03-19 | 2014-09-29 | 삼성전자주식회사 | Apparatus and method for human pose estimation |
US20180268256A1 (en) * | 2017-03-16 | 2018-09-20 | Aquifi, Inc. | Systems and methods for keypoint detection with convolutional neural networks |
CN109165552A (en) * | 2018-07-14 | 2019-01-08 | 深圳神目信息技术有限公司 | A kind of gesture recognition method based on human body key point, system and memory |
CN109558865A (en) * | 2019-01-22 | 2019-04-02 | 郭道宁 | A kind of abnormal state detection method to the special caregiver of need based on human body key point |
CN109871804A (en) * | 2019-02-19 | 2019-06-11 | 上海宝尊电子商务有限公司 | A kind of method and system of shop stream of people discriminance analysis |
CN109934111A (en) * | 2019-02-12 | 2019-06-25 | 清华大学深圳研究生院 | A kind of body-building Attitude estimation method and system based on key point |
CN110287923A (en) * | 2019-06-29 | 2019-09-27 | 腾讯科技(深圳)有限公司 | Human body attitude acquisition methods, device, computer equipment and storage medium |
WO2019232894A1 (en) * | 2018-06-05 | 2019-12-12 | 中国石油大学(华东) | Complex scene-based human body key point detection system and method |
CN110598554A (en) * | 2019-08-09 | 2019-12-20 | 中国地质大学(武汉) | Multi-person posture estimation method based on counterstudy |
CN111597879A (en) * | 2020-04-03 | 2020-08-28 | 成都云盯科技有限公司 | Gesture detection method, device and system based on monitoring video |
-
2020
- 2020-09-01 CN CN202010905107.6A patent/CN112036324B/en active Active
Patent Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR20140114741A (en) * | 2013-03-19 | 2014-09-29 | 삼성전자주식회사 | Apparatus and method for human pose estimation |
US20180268256A1 (en) * | 2017-03-16 | 2018-09-20 | Aquifi, Inc. | Systems and methods for keypoint detection with convolutional neural networks |
WO2019232894A1 (en) * | 2018-06-05 | 2019-12-12 | 中国石油大学(华东) | Complex scene-based human body key point detection system and method |
CN109165552A (en) * | 2018-07-14 | 2019-01-08 | 深圳神目信息技术有限公司 | A kind of gesture recognition method based on human body key point, system and memory |
CN109558865A (en) * | 2019-01-22 | 2019-04-02 | 郭道宁 | A kind of abnormal state detection method to the special caregiver of need based on human body key point |
CN109934111A (en) * | 2019-02-12 | 2019-06-25 | 清华大学深圳研究生院 | A kind of body-building Attitude estimation method and system based on key point |
CN109871804A (en) * | 2019-02-19 | 2019-06-11 | 上海宝尊电子商务有限公司 | A kind of method and system of shop stream of people discriminance analysis |
CN110287923A (en) * | 2019-06-29 | 2019-09-27 | 腾讯科技(深圳)有限公司 | Human body attitude acquisition methods, device, computer equipment and storage medium |
CN110598554A (en) * | 2019-08-09 | 2019-12-20 | 中国地质大学(武汉) | Multi-person posture estimation method based on counterstudy |
CN111597879A (en) * | 2020-04-03 | 2020-08-28 | 成都云盯科技有限公司 | Gesture detection method, device and system based on monitoring video |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113537234A (en) * | 2021-06-10 | 2021-10-22 | 浙江大华技术股份有限公司 | Quantity counting method and device, electronic device and computer equipment |
CN114155518A (en) * | 2021-11-08 | 2022-03-08 | 西安西光产业发展有限公司 | Expressway visor inclination identification method based on deep semantic segmentation network and image correction |
CN114155518B (en) * | 2021-11-08 | 2024-09-27 | 西安西光产业发展有限公司 | Highway light shield inclination recognition method based on depth semantic segmentation network and image correction |
Also Published As
Publication number | Publication date |
---|---|
CN112036324B (en) | 2025-02-14 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN112560741B (en) | A safety wear detection method based on key points of the human body | |
CN110490080B (en) | An image-based human fall detection method | |
CN103210421B (en) | Article detection device and object detecting method | |
CN110490109B (en) | An online human rehabilitation action recognition method based on monocular vision | |
CN111753747B (en) | Violent motion detection method based on monocular camera and three-dimensional attitude estimation | |
CN109766838B (en) | Gait cycle detection method based on convolutional neural network | |
CN102609683A (en) | Automatic labeling method for human joint based on monocular video | |
CN113408435B (en) | A security monitoring method, device, equipment and storage medium | |
JP2012123667A (en) | Attitude estimation device and attitude estimation method | |
CN111709365A (en) | An automatic detection method of human motion pose based on convolutional neural network | |
CN112379773B (en) | Multi-person three-dimensional motion capturing method, storage medium and electronic equipment | |
CN115116127A (en) | A fall detection method based on computer vision and artificial intelligence | |
CN117037272B (en) | Method and system for monitoring fall of old people | |
Alaoui et al. | Fall detection of elderly people using the manifold of positive semidefinite matrices | |
CN112036324B (en) | A human body posture determination method and system for complex multi-person scenes | |
CN114170686A (en) | Elbow bending behavior detection method based on human body key points | |
CN109993116B (en) | Pedestrian re-identification method based on mutual learning of human bones | |
CN116311497A (en) | Tunnel worker abnormal behavior detection method and system based on machine vision | |
CN111178201A (en) | Human body sectional type tracking method based on OpenPose posture detection | |
Das et al. | A fusion of appearance based CNNs and temporal evolution of skeleton with LSTM for daily living action recognition | |
CN113449609A (en) | Subway violation early warning method based on improved HigherHRNet model and DNN (deep neural network) | |
CN115909499A (en) | Multi-person human body posture estimation method based on deep learning | |
CN113408433B (en) | Intelligent monitoring gesture recognition method, device, equipment and storage medium | |
JP2017097549A (en) | Image processing apparatus, method, and program | |
Lv et al. | Cheating behavior detection based-on pictorial structure model |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |