CN105631861A - Method of restoring three-dimensional human body posture from unmarked monocular image in combination with height map - Google Patents
Method of restoring three-dimensional human body posture from unmarked monocular image in combination with height map Download PDFInfo
- Publication number
- CN105631861A CN105631861A CN201510970682.3A CN201510970682A CN105631861A CN 105631861 A CN105631861 A CN 105631861A CN 201510970682 A CN201510970682 A CN 201510970682A CN 105631861 A CN105631861 A CN 105631861A
- Authority
- CN
- China
- Prior art keywords
- image
- human body
- height
- height map
- articulare
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10004—Still image; Photographic image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10024—Color image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30196—Human being; Person
Landscapes
- Image Analysis (AREA)
Abstract
The invention discloses a method of restoring a three-dimensional human body posture from an unmarked monocular image in combination with a height map. The method comprises the following steps: 1) a color image and a height image are used for training to obtain a deep convolutional network-based two-dimensional joint point recognition model; 2) a video frame image sequence and a camera parameter are inputted, and a height map corresponding to each frame of image is calculated; 3) the video frame image and the height map obtained in the second step are inputted, and the two-dimensional joint point recognition model obtained through training the first step is used to obtain two-dimensional joint point coordinates of a human body in each frame of image; and 4) the two-dimensional joint point coordinates obtained in the third step are inputted, and the human body three-dimensional posture is restored according to an optimization model. During the two-dimensional joint point recognition process, the color image and the height image are used integrally, and the two-dimensional joint point recognition accuracy is improved; and time sequence consistency constraints are added to the optimization model which can restore the three-dimensional human body posture from the two-dimensional joint point, and thus the restored three-dimensional human body posture is closer to the real human body posture.
Description
Technical field
The present invention relates to 3 D human body pose recovery method, particularly relate to a kind of method recovering 3 D human body attitude in conjunction with height map from unmarked monocular image sequence.
Background technology
Human body three-dimensional Attitude estimation, because it is widely applied prospect, receives the concern of a lot of researcher. The method of existing 3 D human body Attitude estimation can be largely classified into the method based on monocular camera and based on the multi views method as sequence. At present, the method for monocular is just being subject to industry and is more and more paying close attention to. Because while many vision methods provide more viewdata, and then more rich information can be provided for Attitude estimation, but these data not always can obtain in reality, particularly in video monitoring and sanatorium etc. apply.
Recovering 3 D human body attitude from monocular image sequence is an intrinsic ill-conditioning problem, because when 3 d pose derived by our view from the single-view image that monocular camera obtains time, the projection for observing on same two dimensional image can obtain the combination of multiple 3 d pose and camera position. Add under real conditions due to environmental factors or the impact blocked etc., feature (the such as human body contour outline of image, extremity or two dimension articulare) can not be accurately detected, so that recovering 3 D human body attitude based on monocular image to become more challenging. But, human observer but can pass through eyes and accurately estimate the attitude of human body. In most cases, the mankind can also organize the anatomical landmarks in three dimensions and then the relative position of prediction camera easily. Why the mankind can accomplish this point, it is likely that be because in the brain of people storing the configuration information of substantial amounts of 3 D human body attitude, and then can eliminate the ambiguousness of 2 d-to-3 d. Therefore, it can obtain a rational agency of such ability by substantial amounts of three-dimensional motion data in study motion capture database.
Summary of the invention
Present invention aims to the deficiencies in the prior art, improve a kind of method recovering 3 D human body attitude in conjunction with height map from unmarked monocular image.
It is an object of the invention to be achieved through the following technical solutions: a kind of method recovering 3 D human body attitude in conjunction with height map from unmarked monocular image, the step of the method is as follows:
1) color image data collection and the training of height image data set is used to obtain the two-dimentional articulare model of cognition based on degree of depth convolutional network;
2) reading in color video frame image sequence and camera parameter, estimation obtains the height image that every color image frame is corresponding;
3) by step 2) the coloured image input step 1 of the height map that obtains and correspondence thereof) the two-dimentional articulare model of cognition based on degree of depth convolutional network that training obtains, identify and obtain the two-dimentional body joint point coordinate of human body in every two field picture;
4) input step 3) the two-dimentional body joint point coordinate of multiple image that obtains, recover the 3 d pose of human body in every two field picture according to Optimized model.
Further, described height map calculates in the following ways and obtains:
Height map is the middle mark of a kind of new human body that the present invention proposes. In anatomical structure, there is an empirical scalar in the height of each position of major joint point of human body and the length of skeleton and human body. Therefore elevation information contains the spatial relationship between each articulare of framing structure. For each pixel in human body contour outline, using the Height Estimation method that Park and partners thereof mention in the paper being entitled as " RobustEstimationofHeightofMovingPeopleUsingaSingleCamera " delivered for 2012 to carry out Height Estimation, the method is by the height calculating human body in two dimensional character (i.e. head point and the foot point) back projection on the plane of delineation to three-dimensional scenic. What the value of each pixel in height map represented is the height (namely relative to the height on ground in world coordinate system) of this point. But owing to the height of different people is different, in order to unrelated with human height, the value H of each pixel in height map, (x y) is normalized to relative altitudeThat is:
Wherein x and y is pixel coordinate, hiRepresent the height of i-th people. K be one for relative altitude figure is mapped to required interval convergent-divergent constant, be empirically set to 255.
Further, step 1) described in the two-dimentional articulare model of cognition mentioned in the paper being entitled as " ArticulatedPoseEstimationbyaGraphicalModelwithImageDepen dentPairwiseRelations " delivered for 2014 based on Chen and partners thereof of the two-dimentional articulare model of cognition based on degree of depth convolutional network, paper is X.ChenandA.L.Yuille, " ArticulatedPoseEstimationbyaGraphicalModelwithImageDepen dentPairwiseRelations ", InNIPS, pages1736 1744, 2014. here this model has been improved, increase a new data stream (i.e. height map), (structure originally only uses RGB image to make its degree of depth convolutional network become double-flow design, use now RGB image and the height image estimated) simultaneously, and before last graph model, add a fused layer, merge the output of double fluid convolutional network.
Further, described step 1) in based on the training process of the two-dimentional articulare model of cognition of degree of depth convolutional network be:
1.1) use public data collection " LeedsSportsPoses " (LSP) (coloured image) and synthesis height image data set the two-dimentional articulare model of cognition based on degree of depth convolutional network is trained, the height image data set of synthesis by exercise data driving anthropometric dummy obtain.
1.2) model of cognition obtained is finely tuned by the height map using the RGB image in real video and correspondence thereof.
Further, described step 2) in height map generalization process corresponding to colored video frame images as follows:
2.1) the color video two field picture of input is carried out foreground extraction, obtain the prospect bianry image of every two field picture.
2.2) read in the prospect bianry image of camera internal reference, outer ginseng and every two field picture, generate the height map that every two field picture is corresponding.
Further, described step 3) in identify that the two-dimentional articulare obtained has 14, respectively: left/right ankle, left/right knee, left/right buttocks, left/right wrist, left and right/ancon, left/right shoulder, cervical region and head. The coordinate of two dimension articulare obtains by optimizing scoring function F based on the graph model at position (l, t | I):
Wherein l={li| i �� V} is the set of articulare position, t={tij| (i, j) �� �� } it is paired relationship type, ��0It it is a bias term. V and �� is the set on the summit of graph model and limit respectively. U and R is obtained by the Joint Distribution of the convolutional network that marginalisation training obtains, and comprises position type and the mixing of paired relationship type. The input of convolutional network is an image block, and exporting is a position (i.e. articulare) probability that drops on this image block.
Further, described step 4) particularly as follows:
It is represented by t 3 D human body state:
Wherein PtRepresent t 3 D human body attitude, VtRepresent the articulare speed of t.
3 D human body state can be expressed as a series of main constituent B={b1..., bkAnd the linear combination of an average vector ��:
Wherein ��tIt is the coefficient of main constituent,It is a majorized subset of B. B obtains by different types of exercise data carries out principal component analysis (PCA).
The two-dimentional body joint point coordinate of a given image sequenceCorresponding 3 d poseCan be obtained by the following object function of optimization:
WhereinM and n represents the number of frame number that the image sequence of input comprises and three-dimensional articulare respectively; Weak perspective camera modelFor camera projection matrix; �� is used for balancing back projection's error and temporal consistency,
The invention has the beneficial effects as follows: the present invention is comprehensive in the identification process of two dimension articulare uses coloured image and height image, makes the recognition accuracy of two dimension articulare be improved. On the other hand, in the Optimized model recovering 3 D human body attitude from two dimension articulare, add temporal consistency constraint, make the 3 D human body attitude recovered be more nearly real human body attitude.
Accompanying drawing explanation
Fig. 1 is the overview flow chart of the present invention;
Fig. 2 is based on the skeleton of height and dissects exploded view;
Fig. 3 (a)-(d) is four different synthesis height image of two dimension articulare model of cognition training in embodiment;
Fig. 4 is the true altitude map generalization process that the present invention uses;
Fig. 5 (a)-(c) is three real video images that embodiment uses;
Fig. 6 (a)-(c) is the height map that real video images that embodiment estimates is corresponding;
Fig. 7 (a)-(c) is the two-dimension human body skeleton that embodiment only uses coloured image to estimate;
Fig. 8 (a)-(c) is that embodiment uses colour to add the two-dimension human body skeleton that height image estimates;
Fig. 9 (a)-(c) is the 3 D human body attitude that embodiment recovers.
Detailed description of the invention
The core of the present invention is to increase elevation information (namely using height map) in the process of two dimension articulare identification, then adds temporal consistency constraint in the process recovering 3 d pose according to two dimension articulare.
Following with an embodiment, the embodiment of idiographic flow is described, step following (see Fig. 1):
1) color image data collection and the training of height image data set is used to obtain the two-dimentional articulare model of cognition based on degree of depth convolutional network; A given image sequenceWherein w and h is width and the height of image respectively, and d is number of active lanes. The target of two dimension articulare identification is exactly utilize RGB image with the height image estimated to calculate the position of two dimension articulare on every two field picture. Two dimension articulare identify the two-dimentional articulare model of cognition that the model used is mentioned in the paper being entitled as " ArticulatedPoseEstimationbyaGraphicalModelwithImageDepen dentPairwiseRelations " delivered for 2014 based on Chen and partners thereof, paper is X.ChenandA.L.Yuille, " ArticulatedPoseEstimationbyaGraphicalModelwithImageDepen dentPairwiseRelations ", InNIPS, pages1736 1744,2014. Here carried out improving (see Fig. 1) to this model, increase a new data stream (i.e. height map), its degree of depth convolutional network is made to become double-flow design (structure originally only uses RGB image, uses now RGB image and the height image estimated simultaneously). And before last graph model, add a fused layer, merge the output of double fluid convolutional network. The training process of two dimension articulare model of cognition is:
1.1) use public data collection " LeedsSportsPoses " (LSP) (coloured image) and synthesize height image data set the two-dimentional articulare model of cognition based on degree of depth convolutional network is trained. The data set that LSP data collection is Johnson and partners thereof to be announced in the paper being entitled as " ClusteredPoseandNonlinearAppearanceModelsforHumanPoseEst imation " delivered for 2010, paper is S.JohnsonandM.Everingham, " ClusteredPoseandNonlinearAppearanceModelsforHumanPoseEst imation ", InBMVC, pages1 11,2010. Synthesis altitude information, by using CMU and the action data that gathers voluntarily to drive 9 models to obtain, including the position of height map (see Fig. 3) and the two and three dimensions of predefined 14 articulares, synthesizes nearly 180,000 of height image.
1.2) convolutional network obtained is finely tuned by the height map simultaneously using the RGB image in real video and correspondence thereof. Here pick totally 3811 frame data walked and run of 10 people in real video, the position of its two dimension articulare has been carried out manual mark, and has estimated the height image that every two field picture is corresponding.
2) reading in real video frame image sequence (see Fig. 5) and camera parameter, estimation obtains height image corresponding to every two field picture (see Fig. 6). Detailed process is:
2.1) video frame images of input is carried out foreground extraction, obtain the prospect bianry image of every two field picture.
2.2) read in the prospect bianry image of camera internal reference, outer ginseng and every two field picture, according to above-mentioned height map generalization method, calculate and obtain the height map that every two field picture is corresponding. Height map calculates in the following ways and obtains (see Fig. 4):
Height map is the middle mark of a kind of new human body that the present invention proposes. In anatomical structure, there is an empirical scalar (as shown in Figure 2) in the height of each position of major joint point of human body and the length of skeleton and human body. Therefore elevation information contains the spatial relationship between each articulare of framing structure. For each pixel in human body contour outline, using the Height Estimation method that Park and partners thereof mention in the paper being entitled as " RobustEstimationofHeightofMovingPeopleUsingaSingleCamera " delivered for 2012 to carry out Height Estimation, the method is by the height calculating human body in two dimensional character (i.e. head point and the foot point) back projection on the plane of delineation to three-dimensional scenic. What the value of each pixel in height map represented is the height (namely relative to the height on ground in world coordinate system) of this point. But owing to the height of different people is different, in order to unrelated with human height, the value H of each pixel in height map, (x y) is normalized to relative altitudeThat is:
Wherein x and y is pixel coordinate, hiRepresent the height of i-th people. K be one for relative altitude figure is mapped to required interval convergent-divergent constant, be empirically set to 255.
3) height map that input real video images and estimation obtain, the two-dimentional articulare model of cognition that training obtains is used to be identified, obtain 14 two-dimentional body joint point coordinate of human body in every two field picture, the two-dimension human body skeleton (see Fig. 8) that the two-dimentional articulare of correspondence that is linked in sequence obtains. Here the real video images inputted does not appear in training data.
Also the two-dimentional articulare model of cognition only using coloured image is tested on real video images (the two-dimentional articulare model of cognition that namely Chen and partners thereof mention in the paper being entitled as " ArticulatedPoseEstimationbyaGraphicalModelwithImageDepen dentPairwiseRelations " delivered for 2014), partial test result such as Fig. 7 in an embodiment). After can be seen that addition elevation information according to comparing result, the recognition accuracy of two dimension articulare is improved.
4) input step 3) the two-dimentional body joint point coordinate of multiple image that obtains, the 3 d pose (see Fig. 9) of human body in every two field picture is recovered according to Optimized model. The Optimized model that the Optimized model of 3 D human body attitude proposes in the paper being entitled as " Reconstructing3DHumanPosefrom2DImageLandmarks " delivered for 2012 is recovered based on Ramakrishna and partners thereof according to two dimension articulare. Here this model is improved, adds joint angle speed and temporal consistency constraint, be described in detail below:
It is represented by t 3 D human body state:
Wherein PtRepresent t 3 D human body attitude, VtRepresent the articulare speed of t.
3 D human body state can be expressed as a series of main constituent B={b1..., bkAnd the linear combination of an average vector ��:
Wherein ��tIt is the coefficient of main constituent,It is a majorized subset of B. B obtains by different types of motion carries out principal component analysis (PCA).
The two-dimentional body joint point coordinate of a given image sequenceCorresponding 3 d poseCan be obtained by the following object function of optimization:
WhereinM and n represents the number of frame number that the image sequence of input comprises and three-dimensional articulare respectively; Weak perspective camera modelI is unit matrix,For Kronecker product,For camera projection matrix; �� is used for balancing back projection's error and temporal consistency,
Claims (7)
1. the method recovering 3 D human body attitude in conjunction with height map from unmarked monocular image, it is characterised in that the method comprises the steps:
1) color image data collection and the training of height image data set is used to obtain the two-dimentional articulare model of cognition based on degree of depth convolutional network;
2) reading in color video frame image sequence and camera parameter, estimation obtains the height image that every color image frame is corresponding;
3) by step 2) the coloured image input step 1 of the height map that obtains and correspondence thereof) the two-dimentional articulare model of cognition based on degree of depth convolutional network that training obtains, identify and obtain the two-dimentional body joint point coordinate of human body in every two field picture;
4) input step 3) the two-dimentional body joint point coordinate of multiple image that obtains, recover the 3 d pose of human body in every two field picture according to Optimized model.
2. the method recovering 3 D human body attitude as claimed in claim 1 in conjunction with height map from unmarked monocular image, it is characterised in that described height map calculates in the following ways and obtains:
For each pixel in human body contour outline, using Height Estimation method to carry out Height Estimation, the method is by the height calculating human body in two dimensional character (i.e. head point and the foot point) back projection on the plane of delineation to three-dimensional scenic. What the value of each pixel in height map represented is the height (namely relative to the height on ground in world coordinate system) of this point. The value H of each pixel in height map, (x y) is normalized to relative altitudeThat is:
Wherein x and y is pixel coordinate, hiRepresent the height of i-th people. K is a convergent-divergent constant for relative altitude figure being mapped to required interval.
3. the method recovering 3 D human body attitude as claimed in claim 1 in conjunction with height map from unmarked monocular image, it is characterized in that, described step 1) in based on the two-dimentional articulare model of cognition of degree of depth convolutional network traditional based on the model of cognition of RGB image increases height map, its degree of depth convolutional network is made to become double-flow design, and before last graph model, add a fused layer, merge the output of double fluid convolutional network.
4. the method recovering 3 D human body attitude as claimed in claim 1 in conjunction with height map from unmarked monocular image, it is characterised in that step 1) in based on the training process of the two-dimentional articulare model of cognition of degree of depth convolutional network be:
1.1) use public data collection " LeedsSportsPoses " (LSP) (coloured image) and synthesis height image data set the two-dimentional articulare model of cognition based on degree of depth convolutional network is trained, the height image data set of synthesis by exercise data driving anthropometric dummy obtain.
1.2) model of cognition obtained is finely tuned by the height map using the RGB image in real video and correspondence thereof.
5. the method recovering 3 D human body attitude as claimed in claim 1 in conjunction with height map from unmarked monocular image, it is characterised in that described step 2) in height map generalization process corresponding to colored video frame images as follows:
2.1) the color video two field picture of input is carried out foreground extraction, obtain the prospect bianry image of every two field picture.
2.2) read in the prospect bianry image of camera internal reference, outer ginseng and every two field picture, generate the height map that every two field picture is corresponding.
6. the method recovering 3 D human body attitude as claimed in claim 1 in conjunction with height map from unmarked monocular image, it is characterized in that, described step 3) in identify that the two-dimentional articulare that obtains has 14, respectively: left/right ankle, left/right knee, left/right buttocks, left/right wrist, left and right/ancon, left/right shoulder, cervical region and head. The coordinate of two dimension articulare obtains by optimizing scoring function F based on the graph model at position (l, t | I):
Wherein l={li| i �� V} is the set of articulare position, t={tij| (i, j) �� �� } it is paired relationship type, ��0It it is a bias term. V and �� is the set on the summit of graph model and limit respectively. U and R is obtained by the Joint Distribution of the convolutional network that marginalisation training obtains, and comprises position type and the mixing of paired relationship type. The input of convolutional network is an image block, and exporting is a position (i.e. articulare) probability that drops on this image block.
7. the method recovering 3 D human body attitude as claimed in claim 1 in conjunction with height map from unmarked monocular image, it is characterised in that described step 4) particularly as follows:
It is represented by t 3 D human body state:
Wherein PtRepresent t 3 D human body attitude, VtRepresent the articulare speed of t.
3 D human body state can be expressed as a series of main constituent B={b1,��,bkAnd the linear combination of an average vector ��:
Wherein ��tIt is the coefficient of main constituent,It is a majorized subset of B. B obtains by different types of exercise data carries out principal component analysis (PCA).
The two-dimentional body joint point coordinate of a given image sequenceCorresponding 3 d poseCan be obtained by the following object function of optimization:
WhereinM and n represents the number of frame number that the image sequence of input comprises and three-dimensional articulare respectively; Weak perspective camera modelI is unit matrix,For Kronecker product,For camera projection matrix; 0nFor the null matrix of n �� n,�� is used for balancing back projection's error and temporal consistency,
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201510970682.3A CN105631861B (en) | 2015-12-21 | 2015-12-21 | Restore the method for 3 D human body posture from unmarked monocular image in conjunction with height map |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201510970682.3A CN105631861B (en) | 2015-12-21 | 2015-12-21 | Restore the method for 3 D human body posture from unmarked monocular image in conjunction with height map |
Publications (2)
Publication Number | Publication Date |
---|---|
CN105631861A true CN105631861A (en) | 2016-06-01 |
CN105631861B CN105631861B (en) | 2019-10-01 |
Family
ID=56046747
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201510970682.3A Active CN105631861B (en) | 2015-12-21 | 2015-12-21 | Restore the method for 3 D human body posture from unmarked monocular image in conjunction with height map |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN105631861B (en) |
Cited By (22)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106529399A (en) * | 2016-09-26 | 2017-03-22 | 深圳奥比中光科技有限公司 | Human body information acquisition method, device and system |
CN107392097A (en) * | 2017-06-15 | 2017-11-24 | 中山大学 | A kind of 3 D human body intra-articular irrigation method of monocular color video |
CN109145788A (en) * | 2018-08-08 | 2019-01-04 | 北京云舶在线科技有限公司 | Attitude data method for catching and system based on video |
CN109176512A (en) * | 2018-08-31 | 2019-01-11 | 南昌与德通讯技术有限公司 | A kind of method, robot and the control device of motion sensing control robot |
CN109821239A (en) * | 2019-02-20 | 2019-05-31 | 网易(杭州)网络有限公司 | Implementation method, device, equipment and the storage medium of somatic sensation television game |
CN109960962A (en) * | 2017-12-14 | 2019-07-02 | 腾讯科技(深圳)有限公司 | Image-recognizing method, device, electronic equipment and readable storage medium storing program for executing |
CN110633005A (en) * | 2019-04-02 | 2019-12-31 | 北京理工大学 | Optical unmarked three-dimensional human body motion capture method |
CN110647991A (en) * | 2019-09-19 | 2020-01-03 | 浙江大学 | Three-dimensional human body posture estimation method based on unsupervised field self-adaption |
CN110895830A (en) * | 2018-09-12 | 2020-03-20 | 珠海格力电器股份有限公司 | Method and device for acquiring 3D image |
CN111222459A (en) * | 2020-01-06 | 2020-06-02 | 上海交通大学 | Visual angle-independent video three-dimensional human body posture identification method |
WO2020156143A1 (en) * | 2019-01-31 | 2020-08-06 | 深圳市商汤科技有限公司 | Three-dimensional human pose information detection method and apparatus, electronic device and storage medium |
CN111557022A (en) * | 2018-12-10 | 2020-08-18 | 三星电子株式会社 | Two-dimensional image processing method and apparatus for performing the same |
CN112183316A (en) * | 2020-09-27 | 2021-01-05 | 中山大学 | Method for measuring human body posture of athlete |
CN112419388A (en) * | 2020-11-24 | 2021-02-26 | 深圳市商汤科技有限公司 | Depth detection method and device, electronic equipment and computer readable storage medium |
CN112465890A (en) * | 2020-11-24 | 2021-03-09 | 深圳市商汤科技有限公司 | Depth detection method and device, electronic equipment and computer readable storage medium |
CN113158766A (en) * | 2021-02-24 | 2021-07-23 | 北京科技大学 | Pedestrian behavior recognition method facing unmanned driving and based on attitude estimation |
CN113192186A (en) * | 2021-05-19 | 2021-07-30 | 华中科技大学 | 3D human body posture estimation model establishing method based on single-frame image and application thereof |
CN113272818A (en) * | 2019-01-10 | 2021-08-17 | 微软技术许可有限责任公司 | Camera environment rendering |
CN113421286A (en) * | 2021-07-12 | 2021-09-21 | 北京未来天远科技开发有限公司 | Motion capture system and method |
CN113989928A (en) * | 2021-10-27 | 2022-01-28 | 南京硅基智能科技有限公司 | Motion capturing and redirecting method |
CN115984972A (en) * | 2023-03-20 | 2023-04-18 | 乐歌人体工学科技股份有限公司 | Human body posture identification method based on motion video drive |
US11893681B2 (en) | 2018-12-10 | 2024-02-06 | Samsung Electronics Co., Ltd. | Method for processing two-dimensional image and device for executing method |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20080137956A1 (en) * | 2006-12-06 | 2008-06-12 | Honda Motor Co., Ltd. | Fast Human Pose Estimation Using Appearance And Motion Via Multi-Dimensional Boosting Regression |
CN101789126A (en) * | 2010-01-26 | 2010-07-28 | 北京航空航天大学 | Three-dimensional human body motion tracking method based on volume pixels |
CN101989326A (en) * | 2009-07-31 | 2011-03-23 | 三星电子株式会社 | Human posture recognition method and device |
CN104636729A (en) * | 2015-02-10 | 2015-05-20 | 浙江工业大学 | Three-dimensional face recognition method based on Bayesian multivariate distribution characteristic extraction |
CN105069423A (en) * | 2015-07-29 | 2015-11-18 | 北京格灵深瞳信息技术有限公司 | Human body posture detection method and device |
-
2015
- 2015-12-21 CN CN201510970682.3A patent/CN105631861B/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20080137956A1 (en) * | 2006-12-06 | 2008-06-12 | Honda Motor Co., Ltd. | Fast Human Pose Estimation Using Appearance And Motion Via Multi-Dimensional Boosting Regression |
CN101989326A (en) * | 2009-07-31 | 2011-03-23 | 三星电子株式会社 | Human posture recognition method and device |
CN101789126A (en) * | 2010-01-26 | 2010-07-28 | 北京航空航天大学 | Three-dimensional human body motion tracking method based on volume pixels |
CN104636729A (en) * | 2015-02-10 | 2015-05-20 | 浙江工业大学 | Three-dimensional face recognition method based on Bayesian multivariate distribution characteristic extraction |
CN105069423A (en) * | 2015-07-29 | 2015-11-18 | 北京格灵深瞳信息技术有限公司 | Human body posture detection method and device |
Non-Patent Citations (2)
Title |
---|
CHEN XIANJIE,ALAN YUILLE: "Articulated Pose Estimation by a Graphical Model with Image Dependent Pairwise Relations", 《NEURAL INFORMATION PROCEEDING SYSTEM》 * |
SANG-WOOK PARK 等: "Robust Estimation of Heights of Moving People Using a Single Camera", 《LNEE》 * |
Cited By (33)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106529399A (en) * | 2016-09-26 | 2017-03-22 | 深圳奥比中光科技有限公司 | Human body information acquisition method, device and system |
CN107392097B (en) * | 2017-06-15 | 2020-07-07 | 中山大学 | Three-dimensional human body joint point positioning method of monocular color video |
CN107392097A (en) * | 2017-06-15 | 2017-11-24 | 中山大学 | A kind of 3 D human body intra-articular irrigation method of monocular color video |
CN109960962A (en) * | 2017-12-14 | 2019-07-02 | 腾讯科技(深圳)有限公司 | Image-recognizing method, device, electronic equipment and readable storage medium storing program for executing |
CN109960962B (en) * | 2017-12-14 | 2022-10-21 | 腾讯科技(深圳)有限公司 | Image recognition method and device, electronic equipment and readable storage medium |
CN109145788A (en) * | 2018-08-08 | 2019-01-04 | 北京云舶在线科技有限公司 | Attitude data method for catching and system based on video |
CN109145788B (en) * | 2018-08-08 | 2020-07-07 | 北京云舶在线科技有限公司 | Video-based attitude data capturing method and system |
CN109176512A (en) * | 2018-08-31 | 2019-01-11 | 南昌与德通讯技术有限公司 | A kind of method, robot and the control device of motion sensing control robot |
CN110895830A (en) * | 2018-09-12 | 2020-03-20 | 珠海格力电器股份有限公司 | Method and device for acquiring 3D image |
CN111557022A (en) * | 2018-12-10 | 2020-08-18 | 三星电子株式会社 | Two-dimensional image processing method and apparatus for performing the same |
US11893681B2 (en) | 2018-12-10 | 2024-02-06 | Samsung Electronics Co., Ltd. | Method for processing two-dimensional image and device for executing method |
CN111557022B (en) * | 2018-12-10 | 2024-05-14 | 三星电子株式会社 | Two-dimensional image processing method and device for executing the method |
CN113272818A (en) * | 2019-01-10 | 2021-08-17 | 微软技术许可有限责任公司 | Camera environment rendering |
WO2020156143A1 (en) * | 2019-01-31 | 2020-08-06 | 深圳市商汤科技有限公司 | Three-dimensional human pose information detection method and apparatus, electronic device and storage medium |
CN109821239B (en) * | 2019-02-20 | 2024-05-28 | 网易(杭州)网络有限公司 | Method, device, equipment and storage medium for realizing somatosensory game |
CN109821239A (en) * | 2019-02-20 | 2019-05-31 | 网易(杭州)网络有限公司 | Implementation method, device, equipment and the storage medium of somatic sensation television game |
CN110633005A (en) * | 2019-04-02 | 2019-12-31 | 北京理工大学 | Optical unmarked three-dimensional human body motion capture method |
CN110647991A (en) * | 2019-09-19 | 2020-01-03 | 浙江大学 | Three-dimensional human body posture estimation method based on unsupervised field self-adaption |
CN110647991B (en) * | 2019-09-19 | 2022-04-05 | 浙江大学 | Three-dimensional human body posture estimation method based on unsupervised field self-adaption |
CN111222459A (en) * | 2020-01-06 | 2020-06-02 | 上海交通大学 | Visual angle-independent video three-dimensional human body posture identification method |
CN111222459B (en) * | 2020-01-06 | 2023-05-12 | 上海交通大学 | Visual angle independent video three-dimensional human body gesture recognition method |
CN112183316A (en) * | 2020-09-27 | 2021-01-05 | 中山大学 | Method for measuring human body posture of athlete |
CN112183316B (en) * | 2020-09-27 | 2023-06-30 | 中山大学 | Athlete human body posture measuring method |
CN112465890A (en) * | 2020-11-24 | 2021-03-09 | 深圳市商汤科技有限公司 | Depth detection method and device, electronic equipment and computer readable storage medium |
CN112419388A (en) * | 2020-11-24 | 2021-02-26 | 深圳市商汤科技有限公司 | Depth detection method and device, electronic equipment and computer readable storage medium |
CN113158766A (en) * | 2021-02-24 | 2021-07-23 | 北京科技大学 | Pedestrian behavior recognition method facing unmanned driving and based on attitude estimation |
CN113192186A (en) * | 2021-05-19 | 2021-07-30 | 华中科技大学 | 3D human body posture estimation model establishing method based on single-frame image and application thereof |
CN113421286A (en) * | 2021-07-12 | 2021-09-21 | 北京未来天远科技开发有限公司 | Motion capture system and method |
CN113421286B (en) * | 2021-07-12 | 2024-01-02 | 北京未来天远科技开发有限公司 | Motion capturing system and method |
CN113989928A (en) * | 2021-10-27 | 2022-01-28 | 南京硅基智能科技有限公司 | Motion capturing and redirecting method |
CN113989928B (en) * | 2021-10-27 | 2023-09-05 | 南京硅基智能科技有限公司 | Motion capturing and redirecting method |
CN115984972A (en) * | 2023-03-20 | 2023-04-18 | 乐歌人体工学科技股份有限公司 | Human body posture identification method based on motion video drive |
CN115984972B (en) * | 2023-03-20 | 2023-08-11 | 乐歌人体工学科技股份有限公司 | Human body posture recognition method based on motion video driving |
Also Published As
Publication number | Publication date |
---|---|
CN105631861B (en) | 2019-10-01 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN105631861A (en) | Method of restoring three-dimensional human body posture from unmarked monocular image in combination with height map | |
CN109166149B (en) | Positioning and three-dimensional line frame structure reconstruction method and system integrating binocular camera and IMU | |
CN107204010B (en) | A kind of monocular image depth estimation method and system | |
US11030525B2 (en) | Systems and methods for deep localization and segmentation with a 3D semantic map | |
CN102999942B (en) | Three-dimensional face reconstruction method | |
WO2017128934A1 (en) | Method, server, terminal and system for implementing augmented reality | |
CN107103613B (en) | A kind of three-dimension gesture Attitude estimation method | |
CN102880866B (en) | Method for extracting face features | |
CN108256504A (en) | A kind of Three-Dimensional Dynamic gesture identification method based on deep learning | |
CN108537191A (en) | A kind of three-dimensional face identification method based on structure light video camera head | |
CN106548519A (en) | Augmented reality method based on ORB SLAM and the sense of reality of depth camera | |
CN104318569A (en) | Space salient region extraction method based on depth variation model | |
CN104794722A (en) | Dressed human body three-dimensional bare body model calculation method through single Kinect | |
CN110399809A (en) | The face critical point detection method and device of multiple features fusion | |
CN102800126A (en) | Method for recovering real-time three-dimensional body posture based on multimodal fusion | |
CN109341703A (en) | A kind of complete period uses the vision SLAM algorithm of CNNs feature detection | |
CN103747240B (en) | The vision significance filtering method of Fusion of Color and movable information | |
CN103839277A (en) | Mobile augmented reality registration method of outdoor wide-range natural scene | |
CN102360504A (en) | Self-adaptation virtual and actual three-dimensional registration method based on multiple natural characteristics | |
CN105096311A (en) | Technology for restoring depth image and combining virtual and real scenes based on GPU (Graphic Processing Unit) | |
CN117456136A (en) | Digital twin scene intelligent generation method based on multi-mode visual recognition | |
CN104463962B (en) | Three-dimensional scene reconstruction method based on GPS information video | |
Zou et al. | Automatic reconstruction of 3D human motion pose from uncalibrated monocular video sequences based on markerless human motion tracking | |
CN117711066A (en) | Three-dimensional human body posture estimation method, device, equipment and medium | |
CN116403275B (en) | Method and system for detecting personnel advancing posture in closed space based on multi-vision |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |