CN105631861A - Method of restoring three-dimensional human body posture from unmarked monocular image in combination with height map - Google Patents

Method of restoring three-dimensional human body posture from unmarked monocular image in combination with height map Download PDF

Info

Publication number
CN105631861A
CN105631861A CN201510970682.3A CN201510970682A CN105631861A CN 105631861 A CN105631861 A CN 105631861A CN 201510970682 A CN201510970682 A CN 201510970682A CN 105631861 A CN105631861 A CN 105631861A
Authority
CN
China
Prior art keywords
image
human body
height
height map
articulare
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201510970682.3A
Other languages
Chinese (zh)
Other versions
CN105631861B (en
Inventor
耿卫东
杜宇
刘永豪
韩菲琳
桂义林
王镇
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang University ZJU
Original Assignee
Zhejiang University ZJU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang University ZJU filed Critical Zhejiang University ZJU
Priority to CN201510970682.3A priority Critical patent/CN105631861B/en
Publication of CN105631861A publication Critical patent/CN105631861A/en
Application granted granted Critical
Publication of CN105631861B publication Critical patent/CN105631861B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10004Still image; Photographic image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10024Color image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30196Human being; Person

Landscapes

  • Image Analysis (AREA)

Abstract

The invention discloses a method of restoring a three-dimensional human body posture from an unmarked monocular image in combination with a height map. The method comprises the following steps: 1) a color image and a height image are used for training to obtain a deep convolutional network-based two-dimensional joint point recognition model; 2) a video frame image sequence and a camera parameter are inputted, and a height map corresponding to each frame of image is calculated; 3) the video frame image and the height map obtained in the second step are inputted, and the two-dimensional joint point recognition model obtained through training the first step is used to obtain two-dimensional joint point coordinates of a human body in each frame of image; and 4) the two-dimensional joint point coordinates obtained in the third step are inputted, and the human body three-dimensional posture is restored according to an optimization model. During the two-dimensional joint point recognition process, the color image and the height image are used integrally, and the two-dimensional joint point recognition accuracy is improved; and time sequence consistency constraints are added to the optimization model which can restore the three-dimensional human body posture from the two-dimensional joint point, and thus the restored three-dimensional human body posture is closer to the real human body posture.

Description

The method recovering 3 D human body attitude in conjunction with height map from unmarked monocular image
Technical field
The present invention relates to 3 D human body pose recovery method, particularly relate to a kind of method recovering 3 D human body attitude in conjunction with height map from unmarked monocular image sequence.
Background technology
Human body three-dimensional Attitude estimation, because it is widely applied prospect, receives the concern of a lot of researcher. The method of existing 3 D human body Attitude estimation can be largely classified into the method based on monocular camera and based on the multi views method as sequence. At present, the method for monocular is just being subject to industry and is more and more paying close attention to. Because while many vision methods provide more viewdata, and then more rich information can be provided for Attitude estimation, but these data not always can obtain in reality, particularly in video monitoring and sanatorium etc. apply.
Recovering 3 D human body attitude from monocular image sequence is an intrinsic ill-conditioning problem, because when 3 d pose derived by our view from the single-view image that monocular camera obtains time, the projection for observing on same two dimensional image can obtain the combination of multiple 3 d pose and camera position. Add under real conditions due to environmental factors or the impact blocked etc., feature (the such as human body contour outline of image, extremity or two dimension articulare) can not be accurately detected, so that recovering 3 D human body attitude based on monocular image to become more challenging. But, human observer but can pass through eyes and accurately estimate the attitude of human body. In most cases, the mankind can also organize the anatomical landmarks in three dimensions and then the relative position of prediction camera easily. Why the mankind can accomplish this point, it is likely that be because in the brain of people storing the configuration information of substantial amounts of 3 D human body attitude, and then can eliminate the ambiguousness of 2 d-to-3 d. Therefore, it can obtain a rational agency of such ability by substantial amounts of three-dimensional motion data in study motion capture database.
Summary of the invention
Present invention aims to the deficiencies in the prior art, improve a kind of method recovering 3 D human body attitude in conjunction with height map from unmarked monocular image.
It is an object of the invention to be achieved through the following technical solutions: a kind of method recovering 3 D human body attitude in conjunction with height map from unmarked monocular image, the step of the method is as follows:
1) color image data collection and the training of height image data set is used to obtain the two-dimentional articulare model of cognition based on degree of depth convolutional network;
2) reading in color video frame image sequence and camera parameter, estimation obtains the height image that every color image frame is corresponding;
3) by step 2) the coloured image input step 1 of the height map that obtains and correspondence thereof) the two-dimentional articulare model of cognition based on degree of depth convolutional network that training obtains, identify and obtain the two-dimentional body joint point coordinate of human body in every two field picture;
4) input step 3) the two-dimentional body joint point coordinate of multiple image that obtains, recover the 3 d pose of human body in every two field picture according to Optimized model.
Further, described height map calculates in the following ways and obtains:
Height map is the middle mark of a kind of new human body that the present invention proposes. In anatomical structure, there is an empirical scalar in the height of each position of major joint point of human body and the length of skeleton and human body. Therefore elevation information contains the spatial relationship between each articulare of framing structure. For each pixel in human body contour outline, using the Height Estimation method that Park and partners thereof mention in the paper being entitled as " RobustEstimationofHeightofMovingPeopleUsingaSingleCamera " delivered for 2012 to carry out Height Estimation, the method is by the height calculating human body in two dimensional character (i.e. head point and the foot point) back projection on the plane of delineation to three-dimensional scenic. What the value of each pixel in height map represented is the height (namely relative to the height on ground in world coordinate system) of this point. But owing to the height of different people is different, in order to unrelated with human height, the value H of each pixel in height map, (x y) is normalized to relative altitudeThat is:
H ^ ( x , y ) = k · H ( x , y ) h i
Wherein x and y is pixel coordinate, hiRepresent the height of i-th people. K be one for relative altitude figure is mapped to required interval convergent-divergent constant, be empirically set to 255.
Further, step 1) described in the two-dimentional articulare model of cognition mentioned in the paper being entitled as " ArticulatedPoseEstimationbyaGraphicalModelwithImageDepen dentPairwiseRelations " delivered for 2014 based on Chen and partners thereof of the two-dimentional articulare model of cognition based on degree of depth convolutional network, paper is X.ChenandA.L.Yuille, " ArticulatedPoseEstimationbyaGraphicalModelwithImageDepen dentPairwiseRelations ", InNIPS, pages1736 1744, 2014. here this model has been improved, increase a new data stream (i.e. height map), (structure originally only uses RGB image to make its degree of depth convolutional network become double-flow design, use now RGB image and the height image estimated) simultaneously, and before last graph model, add a fused layer, merge the output of double fluid convolutional network.
Further, described step 1) in based on the training process of the two-dimentional articulare model of cognition of degree of depth convolutional network be:
1.1) use public data collection " LeedsSportsPoses " (LSP) (coloured image) and synthesis height image data set the two-dimentional articulare model of cognition based on degree of depth convolutional network is trained, the height image data set of synthesis by exercise data driving anthropometric dummy obtain.
1.2) model of cognition obtained is finely tuned by the height map using the RGB image in real video and correspondence thereof.
Further, described step 2) in height map generalization process corresponding to colored video frame images as follows:
2.1) the color video two field picture of input is carried out foreground extraction, obtain the prospect bianry image of every two field picture.
2.2) read in the prospect bianry image of camera internal reference, outer ginseng and every two field picture, generate the height map that every two field picture is corresponding.
Further, described step 3) in identify that the two-dimentional articulare obtained has 14, respectively: left/right ankle, left/right knee, left/right buttocks, left/right wrist, left and right/ancon, left/right shoulder, cervical region and head. The coordinate of two dimension articulare obtains by optimizing scoring function F based on the graph model at position (l, t | I):
F ( l , t | I ) = Σ i ∈ V U ( l i | I ) + Σ ( i , j ) ∈ ϵ R ( l i , l j , t i j , t j i | I ) + ω 0
Wherein l={li| i �� V} is the set of articulare position, t={tij| (i, j) �� �� } it is paired relationship type, ��0It it is a bias term. V and �� is the set on the summit of graph model and limit respectively. U and R is obtained by the Joint Distribution of the convolutional network that marginalisation training obtains, and comprises position type and the mixing of paired relationship type. The input of convolutional network is an image block, and exporting is a position (i.e. articulare) probability that drops on this image block.
Further, described step 4) particularly as follows:
It is represented by t 3 D human body state:
X t = [ P t T , V t T ] T
Wherein PtRepresent t 3 D human body attitude, VtRepresent the articulare speed of t.
3 D human body state can be expressed as a series of main constituent B={b1..., bkAnd the linear combination of an average vector ��:
X t = μ + B t * ω t
{ b i } i ∈ I B t * ∈ B t * ⋐ B
Wherein ��tIt is the coefficient of main constituent,It is a majorized subset of B. B obtains by different types of exercise data carries out principal component analysis (PCA).
The two-dimentional body joint point coordinate of a given image sequenceCorresponding 3 d poseCan be obtained by the following object function of optimization:
m i n Ω | | p - CA 0 ( M + B * Ω ) | | 2 + α | | D ( M + B * Ω ) | | 2
WhereinM and n represents the number of frame number that the image sequence of input comprises and three-dimensional articulare respectively; Weak perspective camera modelFor camera projection matrix; M = I m ⊗ μ , Ω = [ ω 1 T ... ω m T ] T , A 0 = I m ⊗ [ I n 0 n ] , B * = Σ t = 1 m B t * ; �� is used for balancing back projection's error and temporal consistency,
The invention has the beneficial effects as follows: the present invention is comprehensive in the identification process of two dimension articulare uses coloured image and height image, makes the recognition accuracy of two dimension articulare be improved. On the other hand, in the Optimized model recovering 3 D human body attitude from two dimension articulare, add temporal consistency constraint, make the 3 D human body attitude recovered be more nearly real human body attitude.
Accompanying drawing explanation
Fig. 1 is the overview flow chart of the present invention;
Fig. 2 is based on the skeleton of height and dissects exploded view;
Fig. 3 (a)-(d) is four different synthesis height image of two dimension articulare model of cognition training in embodiment;
Fig. 4 is the true altitude map generalization process that the present invention uses;
Fig. 5 (a)-(c) is three real video images that embodiment uses;
Fig. 6 (a)-(c) is the height map that real video images that embodiment estimates is corresponding;
Fig. 7 (a)-(c) is the two-dimension human body skeleton that embodiment only uses coloured image to estimate;
Fig. 8 (a)-(c) is that embodiment uses colour to add the two-dimension human body skeleton that height image estimates;
Fig. 9 (a)-(c) is the 3 D human body attitude that embodiment recovers.
Detailed description of the invention
The core of the present invention is to increase elevation information (namely using height map) in the process of two dimension articulare identification, then adds temporal consistency constraint in the process recovering 3 d pose according to two dimension articulare.
Following with an embodiment, the embodiment of idiographic flow is described, step following (see Fig. 1):
1) color image data collection and the training of height image data set is used to obtain the two-dimentional articulare model of cognition based on degree of depth convolutional network; A given image sequenceWherein w and h is width and the height of image respectively, and d is number of active lanes. The target of two dimension articulare identification is exactly utilize RGB image with the height image estimated to calculate the position of two dimension articulare on every two field picture. Two dimension articulare identify the two-dimentional articulare model of cognition that the model used is mentioned in the paper being entitled as " ArticulatedPoseEstimationbyaGraphicalModelwithImageDepen dentPairwiseRelations " delivered for 2014 based on Chen and partners thereof, paper is X.ChenandA.L.Yuille, " ArticulatedPoseEstimationbyaGraphicalModelwithImageDepen dentPairwiseRelations ", InNIPS, pages1736 1744,2014. Here carried out improving (see Fig. 1) to this model, increase a new data stream (i.e. height map), its degree of depth convolutional network is made to become double-flow design (structure originally only uses RGB image, uses now RGB image and the height image estimated simultaneously). And before last graph model, add a fused layer, merge the output of double fluid convolutional network. The training process of two dimension articulare model of cognition is:
1.1) use public data collection " LeedsSportsPoses " (LSP) (coloured image) and synthesize height image data set the two-dimentional articulare model of cognition based on degree of depth convolutional network is trained. The data set that LSP data collection is Johnson and partners thereof to be announced in the paper being entitled as " ClusteredPoseandNonlinearAppearanceModelsforHumanPoseEst imation " delivered for 2010, paper is S.JohnsonandM.Everingham, " ClusteredPoseandNonlinearAppearanceModelsforHumanPoseEst imation ", InBMVC, pages1 11,2010. Synthesis altitude information, by using CMU and the action data that gathers voluntarily to drive 9 models to obtain, including the position of height map (see Fig. 3) and the two and three dimensions of predefined 14 articulares, synthesizes nearly 180,000 of height image.
1.2) convolutional network obtained is finely tuned by the height map simultaneously using the RGB image in real video and correspondence thereof. Here pick totally 3811 frame data walked and run of 10 people in real video, the position of its two dimension articulare has been carried out manual mark, and has estimated the height image that every two field picture is corresponding.
2) reading in real video frame image sequence (see Fig. 5) and camera parameter, estimation obtains height image corresponding to every two field picture (see Fig. 6). Detailed process is:
2.1) video frame images of input is carried out foreground extraction, obtain the prospect bianry image of every two field picture.
2.2) read in the prospect bianry image of camera internal reference, outer ginseng and every two field picture, according to above-mentioned height map generalization method, calculate and obtain the height map that every two field picture is corresponding. Height map calculates in the following ways and obtains (see Fig. 4):
Height map is the middle mark of a kind of new human body that the present invention proposes. In anatomical structure, there is an empirical scalar (as shown in Figure 2) in the height of each position of major joint point of human body and the length of skeleton and human body. Therefore elevation information contains the spatial relationship between each articulare of framing structure. For each pixel in human body contour outline, using the Height Estimation method that Park and partners thereof mention in the paper being entitled as " RobustEstimationofHeightofMovingPeopleUsingaSingleCamera " delivered for 2012 to carry out Height Estimation, the method is by the height calculating human body in two dimensional character (i.e. head point and the foot point) back projection on the plane of delineation to three-dimensional scenic. What the value of each pixel in height map represented is the height (namely relative to the height on ground in world coordinate system) of this point. But owing to the height of different people is different, in order to unrelated with human height, the value H of each pixel in height map, (x y) is normalized to relative altitudeThat is:
H ^ ( x , y ) = k · H ( x , y ) h i
Wherein x and y is pixel coordinate, hiRepresent the height of i-th people. K be one for relative altitude figure is mapped to required interval convergent-divergent constant, be empirically set to 255.
3) height map that input real video images and estimation obtain, the two-dimentional articulare model of cognition that training obtains is used to be identified, obtain 14 two-dimentional body joint point coordinate of human body in every two field picture, the two-dimension human body skeleton (see Fig. 8) that the two-dimentional articulare of correspondence that is linked in sequence obtains. Here the real video images inputted does not appear in training data.
Also the two-dimentional articulare model of cognition only using coloured image is tested on real video images (the two-dimentional articulare model of cognition that namely Chen and partners thereof mention in the paper being entitled as " ArticulatedPoseEstimationbyaGraphicalModelwithImageDepen dentPairwiseRelations " delivered for 2014), partial test result such as Fig. 7 in an embodiment). After can be seen that addition elevation information according to comparing result, the recognition accuracy of two dimension articulare is improved.
4) input step 3) the two-dimentional body joint point coordinate of multiple image that obtains, the 3 d pose (see Fig. 9) of human body in every two field picture is recovered according to Optimized model. The Optimized model that the Optimized model of 3 D human body attitude proposes in the paper being entitled as " Reconstructing3DHumanPosefrom2DImageLandmarks " delivered for 2012 is recovered based on Ramakrishna and partners thereof according to two dimension articulare. Here this model is improved, adds joint angle speed and temporal consistency constraint, be described in detail below:
It is represented by t 3 D human body state:
X t = [ P t T , V t T ] T
Wherein PtRepresent t 3 D human body attitude, VtRepresent the articulare speed of t.
3 D human body state can be expressed as a series of main constituent B={b1..., bkAnd the linear combination of an average vector ��:
X t = μ + B t * ω t
{ b i } i ∈ I B t * ∈ B t * ⋐ B
Wherein ��tIt is the coefficient of main constituent,It is a majorized subset of B. B obtains by different types of motion carries out principal component analysis (PCA).
The two-dimentional body joint point coordinate of a given image sequenceCorresponding 3 d poseCan be obtained by the following object function of optimization:
m i n Ω | | p - CA 0 ( M + B * Ω ) | | 2 + α | | D ( M + B * Ω ) | | 2
WhereinM and n represents the number of frame number that the image sequence of input comprises and three-dimensional articulare respectively; Weak perspective camera modelI is unit matrix,For Kronecker product,For camera projection matrix; M = I m ⊗ μ , Ω = [ ω 1 T ... ω m T ] T , A 0 = I m ⊗ [ I n 0 n ] , B * = Σ t = 1 m B t * ; �� is used for balancing back projection's error and temporal consistency,

Claims (7)

1. the method recovering 3 D human body attitude in conjunction with height map from unmarked monocular image, it is characterised in that the method comprises the steps:
1) color image data collection and the training of height image data set is used to obtain the two-dimentional articulare model of cognition based on degree of depth convolutional network;
2) reading in color video frame image sequence and camera parameter, estimation obtains the height image that every color image frame is corresponding;
3) by step 2) the coloured image input step 1 of the height map that obtains and correspondence thereof) the two-dimentional articulare model of cognition based on degree of depth convolutional network that training obtains, identify and obtain the two-dimentional body joint point coordinate of human body in every two field picture;
4) input step 3) the two-dimentional body joint point coordinate of multiple image that obtains, recover the 3 d pose of human body in every two field picture according to Optimized model.
2. the method recovering 3 D human body attitude as claimed in claim 1 in conjunction with height map from unmarked monocular image, it is characterised in that described height map calculates in the following ways and obtains:
For each pixel in human body contour outline, using Height Estimation method to carry out Height Estimation, the method is by the height calculating human body in two dimensional character (i.e. head point and the foot point) back projection on the plane of delineation to three-dimensional scenic. What the value of each pixel in height map represented is the height (namely relative to the height on ground in world coordinate system) of this point. The value H of each pixel in height map, (x y) is normalized to relative altitudeThat is:
H ^ ( x , y ) = k · H ( x , y ) h i
Wherein x and y is pixel coordinate, hiRepresent the height of i-th people. K is a convergent-divergent constant for relative altitude figure being mapped to required interval.
3. the method recovering 3 D human body attitude as claimed in claim 1 in conjunction with height map from unmarked monocular image, it is characterized in that, described step 1) in based on the two-dimentional articulare model of cognition of degree of depth convolutional network traditional based on the model of cognition of RGB image increases height map, its degree of depth convolutional network is made to become double-flow design, and before last graph model, add a fused layer, merge the output of double fluid convolutional network.
4. the method recovering 3 D human body attitude as claimed in claim 1 in conjunction with height map from unmarked monocular image, it is characterised in that step 1) in based on the training process of the two-dimentional articulare model of cognition of degree of depth convolutional network be:
1.1) use public data collection " LeedsSportsPoses " (LSP) (coloured image) and synthesis height image data set the two-dimentional articulare model of cognition based on degree of depth convolutional network is trained, the height image data set of synthesis by exercise data driving anthropometric dummy obtain.
1.2) model of cognition obtained is finely tuned by the height map using the RGB image in real video and correspondence thereof.
5. the method recovering 3 D human body attitude as claimed in claim 1 in conjunction with height map from unmarked monocular image, it is characterised in that described step 2) in height map generalization process corresponding to colored video frame images as follows:
2.1) the color video two field picture of input is carried out foreground extraction, obtain the prospect bianry image of every two field picture.
2.2) read in the prospect bianry image of camera internal reference, outer ginseng and every two field picture, generate the height map that every two field picture is corresponding.
6. the method recovering 3 D human body attitude as claimed in claim 1 in conjunction with height map from unmarked monocular image, it is characterized in that, described step 3) in identify that the two-dimentional articulare that obtains has 14, respectively: left/right ankle, left/right knee, left/right buttocks, left/right wrist, left and right/ancon, left/right shoulder, cervical region and head. The coordinate of two dimension articulare obtains by optimizing scoring function F based on the graph model at position (l, t | I):
F ( l , t | I ) = Σ i ∈ V U ( l i | I ) + Σ ( i , j ) ∈ ϵ R ( l i , l j , t i j , t j i | I ) + ω 0
Wherein l={li| i �� V} is the set of articulare position, t={tij| (i, j) �� �� } it is paired relationship type, ��0It it is a bias term. V and �� is the set on the summit of graph model and limit respectively. U and R is obtained by the Joint Distribution of the convolutional network that marginalisation training obtains, and comprises position type and the mixing of paired relationship type. The input of convolutional network is an image block, and exporting is a position (i.e. articulare) probability that drops on this image block.
7. the method recovering 3 D human body attitude as claimed in claim 1 in conjunction with height map from unmarked monocular image, it is characterised in that described step 4) particularly as follows:
It is represented by t 3 D human body state:
X t = [ P t T , V t T ] T
Wherein PtRepresent t 3 D human body attitude, VtRepresent the articulare speed of t.
3 D human body state can be expressed as a series of main constituent B={b1,��,bkAnd the linear combination of an average vector ��:
X t = μ + B t * ω t
{ b i } i ∈ I B t * ∈ B t * ⋐ B
Wherein ��tIt is the coefficient of main constituent,It is a majorized subset of B. B obtains by different types of exercise data carries out principal component analysis (PCA).
The two-dimentional body joint point coordinate of a given image sequenceCorresponding 3 d poseCan be obtained by the following object function of optimization:
min Ω | | p - CA 0 ( M + B * Ω ) | | 2 + α | | D ( M + B * Ω ) | | 2
WhereinM and n represents the number of frame number that the image sequence of input comprises and three-dimensional articulare respectively; Weak perspective camera modelI is unit matrix,For Kronecker product,For camera projection matrix; M = I m ⊗ μ , Ω = ω 1 T ... ω m T T , 0nFor the null matrix of n �� n,�� is used for balancing back projection's error and temporal consistency,
CN201510970682.3A 2015-12-21 2015-12-21 Restore the method for 3 D human body posture from unmarked monocular image in conjunction with height map Active CN105631861B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510970682.3A CN105631861B (en) 2015-12-21 2015-12-21 Restore the method for 3 D human body posture from unmarked monocular image in conjunction with height map

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510970682.3A CN105631861B (en) 2015-12-21 2015-12-21 Restore the method for 3 D human body posture from unmarked monocular image in conjunction with height map

Publications (2)

Publication Number Publication Date
CN105631861A true CN105631861A (en) 2016-06-01
CN105631861B CN105631861B (en) 2019-10-01

Family

ID=56046747

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510970682.3A Active CN105631861B (en) 2015-12-21 2015-12-21 Restore the method for 3 D human body posture from unmarked monocular image in conjunction with height map

Country Status (1)

Country Link
CN (1) CN105631861B (en)

Cited By (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106529399A (en) * 2016-09-26 2017-03-22 深圳奥比中光科技有限公司 Human body information acquisition method, device and system
CN107392097A (en) * 2017-06-15 2017-11-24 中山大学 A kind of 3 D human body intra-articular irrigation method of monocular color video
CN109145788A (en) * 2018-08-08 2019-01-04 北京云舶在线科技有限公司 Attitude data method for catching and system based on video
CN109176512A (en) * 2018-08-31 2019-01-11 南昌与德通讯技术有限公司 A kind of method, robot and the control device of motion sensing control robot
CN109821239A (en) * 2019-02-20 2019-05-31 网易(杭州)网络有限公司 Implementation method, device, equipment and the storage medium of somatic sensation television game
CN109960962A (en) * 2017-12-14 2019-07-02 腾讯科技(深圳)有限公司 Image-recognizing method, device, electronic equipment and readable storage medium storing program for executing
CN110633005A (en) * 2019-04-02 2019-12-31 北京理工大学 Optical unmarked three-dimensional human body motion capture method
CN110647991A (en) * 2019-09-19 2020-01-03 浙江大学 Three-dimensional human body posture estimation method based on unsupervised field self-adaption
CN110895830A (en) * 2018-09-12 2020-03-20 珠海格力电器股份有限公司 Method and device for acquiring 3D image
CN111222459A (en) * 2020-01-06 2020-06-02 上海交通大学 Visual angle-independent video three-dimensional human body posture identification method
WO2020156143A1 (en) * 2019-01-31 2020-08-06 深圳市商汤科技有限公司 Three-dimensional human pose information detection method and apparatus, electronic device and storage medium
CN111557022A (en) * 2018-12-10 2020-08-18 三星电子株式会社 Two-dimensional image processing method and apparatus for performing the same
CN112183316A (en) * 2020-09-27 2021-01-05 中山大学 Method for measuring human body posture of athlete
CN112419388A (en) * 2020-11-24 2021-02-26 深圳市商汤科技有限公司 Depth detection method and device, electronic equipment and computer readable storage medium
CN112465890A (en) * 2020-11-24 2021-03-09 深圳市商汤科技有限公司 Depth detection method and device, electronic equipment and computer readable storage medium
CN113158766A (en) * 2021-02-24 2021-07-23 北京科技大学 Pedestrian behavior recognition method facing unmanned driving and based on attitude estimation
CN113192186A (en) * 2021-05-19 2021-07-30 华中科技大学 3D human body posture estimation model establishing method based on single-frame image and application thereof
CN113272818A (en) * 2019-01-10 2021-08-17 微软技术许可有限责任公司 Camera environment rendering
CN113421286A (en) * 2021-07-12 2021-09-21 北京未来天远科技开发有限公司 Motion capture system and method
CN113989928A (en) * 2021-10-27 2022-01-28 南京硅基智能科技有限公司 Motion capturing and redirecting method
CN115984972A (en) * 2023-03-20 2023-04-18 乐歌人体工学科技股份有限公司 Human body posture identification method based on motion video drive
US11893681B2 (en) 2018-12-10 2024-02-06 Samsung Electronics Co., Ltd. Method for processing two-dimensional image and device for executing method

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080137956A1 (en) * 2006-12-06 2008-06-12 Honda Motor Co., Ltd. Fast Human Pose Estimation Using Appearance And Motion Via Multi-Dimensional Boosting Regression
CN101789126A (en) * 2010-01-26 2010-07-28 北京航空航天大学 Three-dimensional human body motion tracking method based on volume pixels
CN101989326A (en) * 2009-07-31 2011-03-23 三星电子株式会社 Human posture recognition method and device
CN104636729A (en) * 2015-02-10 2015-05-20 浙江工业大学 Three-dimensional face recognition method based on Bayesian multivariate distribution characteristic extraction
CN105069423A (en) * 2015-07-29 2015-11-18 北京格灵深瞳信息技术有限公司 Human body posture detection method and device

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080137956A1 (en) * 2006-12-06 2008-06-12 Honda Motor Co., Ltd. Fast Human Pose Estimation Using Appearance And Motion Via Multi-Dimensional Boosting Regression
CN101989326A (en) * 2009-07-31 2011-03-23 三星电子株式会社 Human posture recognition method and device
CN101789126A (en) * 2010-01-26 2010-07-28 北京航空航天大学 Three-dimensional human body motion tracking method based on volume pixels
CN104636729A (en) * 2015-02-10 2015-05-20 浙江工业大学 Three-dimensional face recognition method based on Bayesian multivariate distribution characteristic extraction
CN105069423A (en) * 2015-07-29 2015-11-18 北京格灵深瞳信息技术有限公司 Human body posture detection method and device

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
CHEN XIANJIE,ALAN YUILLE: "Articulated Pose Estimation by a Graphical Model with Image Dependent Pairwise Relations", 《NEURAL INFORMATION PROCEEDING SYSTEM》 *
SANG-WOOK PARK 等: "Robust Estimation of Heights of Moving People Using a Single Camera", 《LNEE》 *

Cited By (33)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106529399A (en) * 2016-09-26 2017-03-22 深圳奥比中光科技有限公司 Human body information acquisition method, device and system
CN107392097B (en) * 2017-06-15 2020-07-07 中山大学 Three-dimensional human body joint point positioning method of monocular color video
CN107392097A (en) * 2017-06-15 2017-11-24 中山大学 A kind of 3 D human body intra-articular irrigation method of monocular color video
CN109960962A (en) * 2017-12-14 2019-07-02 腾讯科技(深圳)有限公司 Image-recognizing method, device, electronic equipment and readable storage medium storing program for executing
CN109960962B (en) * 2017-12-14 2022-10-21 腾讯科技(深圳)有限公司 Image recognition method and device, electronic equipment and readable storage medium
CN109145788A (en) * 2018-08-08 2019-01-04 北京云舶在线科技有限公司 Attitude data method for catching and system based on video
CN109145788B (en) * 2018-08-08 2020-07-07 北京云舶在线科技有限公司 Video-based attitude data capturing method and system
CN109176512A (en) * 2018-08-31 2019-01-11 南昌与德通讯技术有限公司 A kind of method, robot and the control device of motion sensing control robot
CN110895830A (en) * 2018-09-12 2020-03-20 珠海格力电器股份有限公司 Method and device for acquiring 3D image
CN111557022A (en) * 2018-12-10 2020-08-18 三星电子株式会社 Two-dimensional image processing method and apparatus for performing the same
US11893681B2 (en) 2018-12-10 2024-02-06 Samsung Electronics Co., Ltd. Method for processing two-dimensional image and device for executing method
CN111557022B (en) * 2018-12-10 2024-05-14 三星电子株式会社 Two-dimensional image processing method and device for executing the method
CN113272818A (en) * 2019-01-10 2021-08-17 微软技术许可有限责任公司 Camera environment rendering
WO2020156143A1 (en) * 2019-01-31 2020-08-06 深圳市商汤科技有限公司 Three-dimensional human pose information detection method and apparatus, electronic device and storage medium
CN109821239B (en) * 2019-02-20 2024-05-28 网易(杭州)网络有限公司 Method, device, equipment and storage medium for realizing somatosensory game
CN109821239A (en) * 2019-02-20 2019-05-31 网易(杭州)网络有限公司 Implementation method, device, equipment and the storage medium of somatic sensation television game
CN110633005A (en) * 2019-04-02 2019-12-31 北京理工大学 Optical unmarked three-dimensional human body motion capture method
CN110647991A (en) * 2019-09-19 2020-01-03 浙江大学 Three-dimensional human body posture estimation method based on unsupervised field self-adaption
CN110647991B (en) * 2019-09-19 2022-04-05 浙江大学 Three-dimensional human body posture estimation method based on unsupervised field self-adaption
CN111222459A (en) * 2020-01-06 2020-06-02 上海交通大学 Visual angle-independent video three-dimensional human body posture identification method
CN111222459B (en) * 2020-01-06 2023-05-12 上海交通大学 Visual angle independent video three-dimensional human body gesture recognition method
CN112183316A (en) * 2020-09-27 2021-01-05 中山大学 Method for measuring human body posture of athlete
CN112183316B (en) * 2020-09-27 2023-06-30 中山大学 Athlete human body posture measuring method
CN112465890A (en) * 2020-11-24 2021-03-09 深圳市商汤科技有限公司 Depth detection method and device, electronic equipment and computer readable storage medium
CN112419388A (en) * 2020-11-24 2021-02-26 深圳市商汤科技有限公司 Depth detection method and device, electronic equipment and computer readable storage medium
CN113158766A (en) * 2021-02-24 2021-07-23 北京科技大学 Pedestrian behavior recognition method facing unmanned driving and based on attitude estimation
CN113192186A (en) * 2021-05-19 2021-07-30 华中科技大学 3D human body posture estimation model establishing method based on single-frame image and application thereof
CN113421286A (en) * 2021-07-12 2021-09-21 北京未来天远科技开发有限公司 Motion capture system and method
CN113421286B (en) * 2021-07-12 2024-01-02 北京未来天远科技开发有限公司 Motion capturing system and method
CN113989928A (en) * 2021-10-27 2022-01-28 南京硅基智能科技有限公司 Motion capturing and redirecting method
CN113989928B (en) * 2021-10-27 2023-09-05 南京硅基智能科技有限公司 Motion capturing and redirecting method
CN115984972A (en) * 2023-03-20 2023-04-18 乐歌人体工学科技股份有限公司 Human body posture identification method based on motion video drive
CN115984972B (en) * 2023-03-20 2023-08-11 乐歌人体工学科技股份有限公司 Human body posture recognition method based on motion video driving

Also Published As

Publication number Publication date
CN105631861B (en) 2019-10-01

Similar Documents

Publication Publication Date Title
CN105631861A (en) Method of restoring three-dimensional human body posture from unmarked monocular image in combination with height map
CN109166149B (en) Positioning and three-dimensional line frame structure reconstruction method and system integrating binocular camera and IMU
CN107204010B (en) A kind of monocular image depth estimation method and system
US11030525B2 (en) Systems and methods for deep localization and segmentation with a 3D semantic map
CN102999942B (en) Three-dimensional face reconstruction method
WO2017128934A1 (en) Method, server, terminal and system for implementing augmented reality
CN107103613B (en) A kind of three-dimension gesture Attitude estimation method
CN102880866B (en) Method for extracting face features
CN108256504A (en) A kind of Three-Dimensional Dynamic gesture identification method based on deep learning
CN108537191A (en) A kind of three-dimensional face identification method based on structure light video camera head
CN106548519A (en) Augmented reality method based on ORB SLAM and the sense of reality of depth camera
CN104318569A (en) Space salient region extraction method based on depth variation model
CN104794722A (en) Dressed human body three-dimensional bare body model calculation method through single Kinect
CN110399809A (en) The face critical point detection method and device of multiple features fusion
CN102800126A (en) Method for recovering real-time three-dimensional body posture based on multimodal fusion
CN109341703A (en) A kind of complete period uses the vision SLAM algorithm of CNNs feature detection
CN103747240B (en) The vision significance filtering method of Fusion of Color and movable information
CN103839277A (en) Mobile augmented reality registration method of outdoor wide-range natural scene
CN102360504A (en) Self-adaptation virtual and actual three-dimensional registration method based on multiple natural characteristics
CN105096311A (en) Technology for restoring depth image and combining virtual and real scenes based on GPU (Graphic Processing Unit)
CN117456136A (en) Digital twin scene intelligent generation method based on multi-mode visual recognition
CN104463962B (en) Three-dimensional scene reconstruction method based on GPS information video
Zou et al. Automatic reconstruction of 3D human motion pose from uncalibrated monocular video sequences based on markerless human motion tracking
CN117711066A (en) Three-dimensional human body posture estimation method, device, equipment and medium
CN116403275B (en) Method and system for detecting personnel advancing posture in closed space based on multi-vision

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant