CN113077519A - Multi-phase external parameter automatic calibration method based on human skeleton extraction - Google Patents

Multi-phase external parameter automatic calibration method based on human skeleton extraction Download PDF

Info

Publication number
CN113077519A
CN113077519A CN202110289301.0A CN202110289301A CN113077519A CN 113077519 A CN113077519 A CN 113077519A CN 202110289301 A CN202110289301 A CN 202110289301A CN 113077519 A CN113077519 A CN 113077519A
Authority
CN
China
Prior art keywords
camera
image
points
cameras
joint points
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110289301.0A
Other languages
Chinese (zh)
Other versions
CN113077519B (en
Inventor
关俊志
耿虎军
高峰
柴兴华
陈彦桥
张泽勇
李晨阳
王雅涵
彭会湘
陈韬亦
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
CETC 54 Research Institute
Original Assignee
CETC 54 Research Institute
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by CETC 54 Research Institute filed Critical CETC 54 Research Institute
Priority to CN202110289301.0A priority Critical patent/CN113077519B/en
Publication of CN113077519A publication Critical patent/CN113077519A/en
Application granted granted Critical
Publication of CN113077519B publication Critical patent/CN113077519B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/80Analysis of captured images to determine intrinsic or extrinsic camera parameters, i.e. camera calibration
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F17/00Digital computing or data processing equipment or methods, specially adapted for specific functions
    • G06F17/10Complex mathematical operations
    • G06F17/16Matrix or vector computation, e.g. matrix-matrix or matrix-vector multiplication, matrix factorization
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10028Range image; Depth image; 3D point clouds
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30196Human being; Person

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Data Mining & Analysis (AREA)
  • Computing Systems (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Software Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Pure & Applied Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Evolutionary Computation (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computational Mathematics (AREA)
  • Mathematical Analysis (AREA)
  • Mathematical Optimization (AREA)
  • Multimedia (AREA)
  • Algebra (AREA)
  • Databases & Information Systems (AREA)
  • Image Analysis (AREA)
  • Studio Devices (AREA)

Abstract

The invention discloses a multi-phase external parameter automatic calibration method based on human skeleton extraction, and belongs to the technical field of computer vision. Processing each frame of image, and extracting the positions of human skeletal joint points in the image by using a deep learning method; selecting any one camera coordinate system as a world coordinate system, and calculating external parameters of other cameras through an essential matrix; the translation vector scale is calculated by using the human body size information. The method takes human skeleton joint points as characteristic points, takes point cloud formed by motion tracks of the human skeleton joint points as a virtual calibration object, then calculates an essential matrix between cameras, obtains the relative pose between the cameras through essential matrix decomposition, and completes real-time online accurate external reference calibration of a multi-camera system.

Description

Multi-phase external parameter automatic calibration method based on human skeleton extraction
Technical Field
The invention belongs to the field of computer vision, and particularly relates to a method for calibrating online external parameters of a multi-camera system through pedestrian skeleton extraction.
Background
In the related fields of computer vision technology and artificial intelligence, the application of a multi-camera system in the fields of scene reconstruction, smart city safety monitoring, airport monitoring, motion capture, sports video analysis, industrial measurement and the like needs to carry out accurate and quick external parameter calibration on the multi-camera system. The camera external parameters are a set of parameters representing attributes such as the position, the rotation direction and the like of the camera in a world coordinate system, so that calibration needs to be performed after the camera is installed. The calibration of the external parameters of the multi-phase machine is the process of obtaining the external parameters of the multi-phase machine
In the conventional calibration method, known scene structure information is used for calibration, and the conventional calibration method usually involves the manufacture of an accurate calibration object, a complex calibration process and high-precision known calibration information and requires complex operation of a professional. Moreover, each time the position of the camera set is changed, a recalibration operation is required.
Disclosure of Invention
In order to solve the technical problems, the invention provides a multi-phase external parameter automatic calibration method, which takes pedestrians frequently existing in a scene as a calibration object, can realize online real-time calibration of a camera system, and provides a basis for later application such as scene understanding monitoring and the like.
In order to achieve the purpose, the technical scheme adopted by the invention is as follows:
a multi-phase external parameter automatic calibration method based on human skeleton extraction comprises the following steps:
(1) enabling a single pedestrian to walk in a camera monitoring area, and simultaneously recording videos by a plurality of cameras to obtain synchronized videos;
(2) intercepting images of pedestrians with the same frame number at different positions from each video;
(3) processing each frame of image, and extracting pedestrian bone joint points in the image by using a deep learning algorithm to obtain image pixel coordinates of each bone joint point;
(4) calculating the image physical dimension coordinates of each bone joint point by using the image pixel coordinates of each bone joint point according to the known camera internal parameters;
(5) and selecting any one camera coordinate system as a world coordinate system, and calculating the external parameters of other cameras by using the image physical dimension coordinates of the skeletal joint points and the essential matrix.
Wherein, the specific mode of the step (3) is as follows:
(301) performing neural network prediction on each frame of image to obtain a thermodynamic diagram and a partial affinity field of each skeletal joint;
(302) extracting specific image positions and confidence degrees of the joints from the thermodynamic diagram by applying a non-maximum suppression algorithm;
(303) finding limb links by using the extracted joint information and part of the affinity fields to obtain all connections, wherein each connection is regarded as a limb;
(304) the limbs with the same joint are regarded as the limbs of the same person, the limbs are assembled to form the person, and image pixel coordinates of all the skeletal joint points are obtained.
Wherein, the specific mode of the step (5) is as follows:
(501) recording discrete three-dimensional point cloud of three-dimensional positions of skeleton joint points of human body under different cameras
Figure BDA0002981799460000021
k is a camera mark, i is a skeletal joint mark, and t represents different moments;
(502) the camera coordinate system of one camera is arbitrarily selected as a first world coordinate system, and the discrete three-dimensional point cloud of the camera is
Figure BDA0002981799460000022
Then
Figure BDA0002981799460000023
Wherein R iskAnd ckThe external parameters of the camera k are a rotation matrix and a translation vector;
(503) selecting matching points of at least eight pairs of skeletal joint points to calculate an essential matrix EkThen by decomposing the essential matrix EkTo obtain ckAnd Rk(ii) a Wherein the essential matrix EkThe calculation method is as follows:
the imaging bone joint points and the center points of the two cameras form a plane, i.e.
Figure BDA0002981799460000024
And ckThree vectors are in the same plane, and can be obtained:
Figure BDA0002981799460000025
will be provided with
Figure BDA0002981799460000026
And
Figure BDA0002981799460000027
brought into the above formula, XiaoFalling off device
Figure BDA0002981799460000028
And
Figure BDA0002981799460000029
the following can be obtained:
Figure BDA00029817994600000210
wherein E isk=|ck|×RkIs the essential matrix, | ck|×Is a vector ckThe anti-symmetric matrix of (a) is,
Figure BDA00029817994600000211
and
Figure BDA00029817994600000212
respectively the physical size coordinates of the images of the bone joint point i at the moment t in the selection camera and the camera marked as k,
Figure BDA00029817994600000213
and
Figure BDA00029817994600000214
respectively the vertical coordinates of the bone joint point i at the time t in the selected camera and the camera marked as k;
(504) c obtained by triangulation and calculation in step (503)kAnd RkCalculating the coordinates of two different skeletal joint points in the camera coordinate system denoted by k
Figure BDA00029817994600000215
And
Figure BDA00029817994600000216
the distance between two different skeletal joint points is
Figure BDA00029817994600000217
And using the actual physical length between two known skeletal joint points
Figure BDA00029817994600000218
Calculating to obtain scale information lambdak
Figure BDA00029817994600000219
(505) Adding the scale information to the translation vector to obtain the actual translation vector of each camera as lambdakck
Compared with the prior art, the invention has the beneficial effects that:
1. the invention provides an effective multi-camera system method, which can obtain good calibration effect without additional calibration objects and complicated calibration processes.
2. The method is simple and easy to implement, and can carry out automatic online calibration under the condition that a multi-camera system does not shut down, thereby greatly improving the calibration efficiency.
3. Feature point matching, scale calculation and online calibration of a multi-camera system have been research hotspots in the field, and at present, common methods are roughly divided into two types: one type is a calibration method based on the traditional calibration object, and although the method can obtain good effect, the method has high requirement on the manufacturing precision of the calibration object, the calibration flow is complicated, and online calibration cannot be realized; the other type is a self-calibration method, a specially-made calibration object is not needed in the method, the corresponding relation between cameras is established by depending on feature points in an image, but the method cannot establish the corresponding relation of the feature points under the condition that the visual angle between the cameras is large, so that the application difficulty in a real scene is high, and the translation vector has no actual scale information. In view of the above, the invention firstly uses human skeleton joint points as feature points, uses point cloud formed by motion tracks of the human skeleton joint points as a virtual calibration object, calculates the relative pose between cameras by an essential matrix principle, and provides a scale calculation method based on human physical dimensions to solve the problem of scale uncertainty in camera calibration. This approach is an important innovation over the prior art.
Drawings
Fig. 1 is a flowchart of a calibration method of a multi-camera system according to an embodiment of the present invention.
Fig. 2 is a schematic diagram of a human skeleton extracted by a deep learning algorithm in the embodiment of the present invention.
Fig. 3 is a schematic diagram of an essential matrix adopted in the embodiment of the present invention.
Detailed description of the invention
In order to facilitate the understanding and implementation of the present invention for those of ordinary skill in the art, the present invention is further described in detail with reference to the accompanying drawings and examples, it is to be understood that the embodiments described herein are merely illustrative and explanatory of the present invention and are not restrictive thereof.
A multi-phase external parameter automatic calibration method based on human skeleton extraction comprises the following steps:
step 1, after a multi-camera system is installed, enabling a single pedestrian to walk in a camera monitoring area, and simultaneously recording videos by multiple cameras to obtain synchronized videos;
step 2, intercepting images of pedestrians with the same frame number at different positions from each video;
step 3, processing each frame of image, extracting pedestrian skeleton joint points in the image by using a convolutional neural network, and obtaining the pixel coordinate of each joint point:
step 3.1, carrying out neural network prediction on the image to obtain a thermodynamic diagram (Heatmap) and a partial Affinity Field (Part Affinity Field) of each skeletal joint point;
3.2, extracting the specific image position and the confidence coefficient of the joint from the thermodynamic diagram by applying a non-maximum suppression (NMS) algorithm;
3.3, finding the limb links by utilizing the joint information and part of the affinity fields to obtain all connections, wherein each connection can be regarded as a limb;
and 3.4, after all the limbs are obtained, regarding the limbs with the same joint as the limbs of the same person, assembling the limbs to form a person, and obtaining the image pixel coordinates of the skeletal joint points of the person.
Step 4, calculating the image physical dimension coordinates of each bone joint point according to the known camera internal parameters, and recording the coordinates as
Figure BDA0002981799460000041
k is a camera mark, i is a skeletal joint mark, and t represents different moments;
and 5, selecting any one camera coordinate system as a world coordinate system, and calculating external parameters of other cameras through the essential matrix:
step 5.1, recording discrete three-dimensional point clouds of three-dimensional positions of human skeleton joint points under different cameras
Figure BDA0002981799460000042
k is a camera mark, i is a skeletal joint mark, and t represents different moments;
step 5.2, selecting a camera coordinate system of the camera marked as 1 as a first world coordinate system, wherein the dispersed three-dimensional point cloud of the camera is
Figure BDA0002981799460000043
Figure BDA0002981799460000044
Wherein R iskAnd ckThe external parameters of the camera k are the rotation matrix and the translation vector;
step 5.3, selecting matching points of a plurality of pairs of skeletal joint points to calculate an essential matrix EkThen by decomposing the essential matrix EkTo obtain ckAnd Rk(ii) a Wherein the essential matrix EkThe calculation method is as follows:
the imaging bone joint point and the central points of the two cameras form a plane
Figure BDA0002981799460000045
ckThree vectors are in the same plane, so that the following can be obtained:
Figure BDA0002981799460000046
will be provided with
Figure BDA0002981799460000047
And
Figure BDA0002981799460000048
brought into the above formula and eliminated
Figure BDA0002981799460000049
And
Figure BDA00029817994600000410
the following can be obtained:
Figure BDA00029817994600000411
wherein E isk=|ck|×RkIs the essential matrix, | ck|×Is a vector ckAn antisymmetric matrix of (a);
step 5.4, c obtained in step (503)kAll lengths of (A) are normalized to 1, i.e. | | ck1, in practical cases, c is generally | | |k||≠||cmI.e. the distances of the different cameras to the camera 1 are unequal, so that the scale information needs to be calculated. C is obtained by the above calculation assuming two different skeletal joint pointskAnd RkAnd triangulation can calculate its coordinates in the k-coordinate system of the camera as
Figure BDA00029817994600000412
Figure BDA0002981799460000051
The distance between them is
Figure BDA0002981799460000052
If the actual physical length between two skeletal joint points is known to be
Figure BDA0002981799460000053
For example, the average length of a human arm, the following can be calculated:
Figure BDA0002981799460000054
step 5.5, finally, calculating and adding the scale information to the translation vector to obtain the translation vector lambda of each camerakck
The following is a more specific example:
referring to fig. 1, a method for calibrating a multi-camera system based on pedestrian head recognition includes the following steps:
a multi-camera system calibration method based on pedestrian head recognition comprises the following steps:
step 1, after a multi-camera system is installed, enabling a single pedestrian to walk in a camera monitoring area, and simultaneously recording videos by multiple cameras to obtain synchronized videos;
step 2, intercepting images of pedestrians with the same frame number at different positions from each video;
step 3, processing each frame of image, extracting pedestrian skeleton joint points in the image by using a convolutional neural network, and obtaining pixel coordinates of each joint point, as shown in fig. 2, the method comprises the following substeps:
step 3.1, carrying out neural network prediction on the image to obtain a thermodynamic diagram (Heatmap) and a partial Affinity Field (Part Affinity Field) of each joint point;
3.2, extracting the specific image position and the confidence coefficient of the joint from the thermodynamic diagram by applying a non-maximum suppression (NMS) algorithm;
3.3, finding the limb links by utilizing the joint information and part of the affinity fields to obtain all connections, wherein each connection can be regarded as a limb;
step 3.4, after all limbs are obtained, regarding the limbs with the same joint as the limbs of the same person, assembling the limbs to form a person, and obtaining image pixel coordinates of human skeletal joint points, wherein a specific skeleton extraction algorithm is shown in documents [1], [1] Z.Cao and G.Hidalgo Martinez and T.Simon and S.Wei and Y.A.sheikh.Openpos: real Multi-person 2D Point Estimation using Part Affinity fields InIEEE Transactions on Pattern Recognition and Machine understanding, doi: 10.1109/TPAMI.2019.29257.
Step 4, calculating the image physical dimension coordinates of each bone joint point according to the known camera internal parameters, and recording the coordinates as
Figure BDA0002981799460000055
k is a camera mark, i is a skeletal joint mark, and t represents different moments;
step 5, selecting any one camera coordinate system as a world coordinate system, and calculating external parameters of other cameras through the intrinsic matrix, wherein the method comprises the following substeps:
step 5.1, recording discrete three-dimensional point clouds of three-dimensional positions of human skeleton joint points under different cameras
Figure BDA0002981799460000056
k is a camera mark, i is a skeletal joint mark, and t represents different moments;
step 5.2, selecting a camera coordinate system of the camera marked as 1 as a first world coordinate system, wherein the dispersed three-dimensional point cloud of the camera is
Figure BDA0002981799460000061
Figure BDA0002981799460000062
Wherein R iskAnd ckThe external parameters of the camera k are the rotation matrix and the translation vector;
step 5.3, obtaining the essence matrix E by calculating at least eight pairs of matching points of the skeletal joint pointskThen by decomposing the essential matrix EkTo obtain ckAnd RkThe specific algorithm is shown in the literature [2]:
[2]H.C.Longuet-Higgins.A computer algorithm for reconstructinga scene from two projections.Nature,vol.293,pages 133-135,September 1981.
Essence matrix EkThe calculation method is as follows:
the imaging bone joint point and the central points of the two cameras form a plane
Figure BDA0002981799460000063
ckThree vectors are in the same plane, as shown in fig. 3, and thus:
Figure BDA0002981799460000064
will be provided with
Figure BDA0002981799460000065
And
Figure BDA0002981799460000066
brought into the above formula and eliminated
Figure BDA0002981799460000067
And
Figure BDA0002981799460000068
the following can be obtained:
Figure BDA0002981799460000069
wherein E isk=|ck|×RkIs the essential matrix, | ck|×Is a vector ckAn antisymmetric matrix of (a);
step 5.4, c obtained in step (503)kAll lengths of (A) are normalized to 1, i.e. | | ck1, in practical cases, c is generally | | |k||≠||cmI, k and m are two different camera labels, i.e. the distances from the different cameras to the camera 1 are not equal, so that the scale information needs to be calculated. Assuming two different skeletal joint points, pass throughStep (c) is calculatedkAnd RkAnd triangulation can calculate its coordinates in the k-coordinate system of the camera as
Figure BDA00029817994600000610
Figure BDA00029817994600000611
The distance between them is
Figure BDA00029817994600000612
If the actual physical length between two skeletal joint points is known to be
Figure BDA00029817994600000613
For example, the average length of a human arm, the following can be calculated:
Figure BDA00029817994600000614
step 5.5, finally, calculating and adding the scale information to the translation vector to obtain the translation vector lambda of each camerakck. The calibration error projection error of the method is 1.4 pixels, the attitude error is 0.5 degrees, the offset error is 1.0 percent, and the calibration result is accurate.
In a word, the method processes each frame of image, and extracts the positions of human skeletal joint points in the image by using a deep learning method; selecting any one camera coordinate system as a world coordinate system, and calculating external parameters of other cameras through an essential matrix; the scale of the translation vector is calculated by using the human body size information. The invention takes human skeleton joint points as characteristic points, takes point cloud formed by motion tracks of the human joint points as a virtual calibration object, solves a camera rotation matrix and a translation vector by using an essential matrix, and provides a translation vector scale calculation method based on human body dimension information to complete real-time online accurate external parameter calibration of a multi-camera system.
The above description is only one embodiment of the present invention, and is not intended to limit the present invention. Any modification, improvement or the like made within the spirit and principle of the present invention should be included in the protection scope of the present invention.

Claims (3)

1. A multi-phase external parameter automatic calibration method based on human skeleton extraction is characterized by comprising the following steps:
(1) enabling a single pedestrian to walk in a camera monitoring area, and simultaneously recording videos by a plurality of cameras to obtain synchronized videos;
(2) intercepting images of pedestrians with the same frame number at different positions from each video;
(3) processing each frame of image, and extracting pedestrian bone joint points in the image by using a deep learning algorithm to obtain image pixel coordinates of each bone joint point;
(4) calculating the image physical dimension coordinates of each bone joint point by using the image pixel coordinates of each bone joint point according to the known camera internal parameters;
(5) and selecting any one camera coordinate system as a world coordinate system, and calculating the external parameters of other cameras by using the image physical dimension coordinates of the skeletal joint points and the essential matrix.
2. The method for automatically calibrating the external parameters of the multiple phases based on the human body skeleton extraction as claimed in claim 1, wherein the specific manner of the step (3) is as follows:
(301) performing neural network prediction on each frame of image to obtain a thermodynamic diagram and a partial affinity field of each skeletal joint;
(302) extracting specific image positions and confidence degrees of the joints from the thermodynamic diagram by applying a non-maximum suppression algorithm;
(303) finding limb links by using the extracted joint information and part of the affinity fields to obtain all connections, wherein each connection is regarded as a limb;
(304) the limbs with the same joint are regarded as the limbs of the same person, the limbs are assembled to form the person, and image pixel coordinates of all the skeletal joint points are obtained.
3. The method for automatically calibrating the external parameters of the multiple phases based on the human skeleton extraction of claim 1, wherein the specific manner of the step (5) is as follows:
(501) recording discrete three-dimensional point cloud of three-dimensional positions of skeleton joint points of human body under different cameras
Figure FDA0002981799450000011
k is a camera mark, i is a skeletal joint mark, and the incense represents different moments;
(502) the camera coordinate system of one camera is arbitrarily selected as a first world coordinate system, and the discrete three-dimensional point cloud of the camera is
Figure FDA0002981799450000012
Then
Figure FDA0002981799450000013
Wherein R iskAnd ckThe external parameters of the camera k are rotation matrix and translation vector;
(503) selecting matching points of a plurality of pairs of skeletal joint points to calculate an essential matrix EkThen by decomposing the essential matrix EkTo obtain ckAnd Rk(ii) a Wherein the essential matrix EkThe calculation method is as follows:
the imaging bone joint points and the center points of the two cameras form a plane, i.e.
Figure FDA0002981799450000014
Figure FDA0002981799450000015
And ckThree vectors are in the same plane, and can be obtained:
Figure FDA0002981799450000021
will be provided with
Figure FDA0002981799450000022
And
Figure FDA0002981799450000023
brought into the above formula and eliminated
Figure FDA0002981799450000024
And
Figure FDA0002981799450000025
the following can be obtained:
Figure FDA0002981799450000026
wherein E isk=|ck|×RkIs the essential matrix, | ck|×Is a vector ckThe anti-symmetric matrix of (a) is,
Figure FDA0002981799450000027
and
Figure FDA0002981799450000028
respectively the physical size coordinates of the images of the bone joint point i at the moment t in the selection camera and the camera marked as k,
Figure FDA0002981799450000029
and
Figure FDA00029817994500000210
respectively the vertical coordinates of the bone joint point i at the time t in the selected camera and the camera marked as k;
(504) c obtained by triangulation and calculation in step (503)kAnd RkCalculating the coordinates of two different skeletal joint points in the camera coordinate system denoted by k
Figure FDA00029817994500000211
And
Figure FDA00029817994500000212
the distance between two different skeletal joint points is
Figure FDA00029817994500000213
And using the actual physical length between two known skeletal joint points
Figure FDA00029817994500000214
Calculating to obtain scale information lambdak
Figure FDA00029817994500000215
(505) Adding the scale information to the translation vector to obtain the actual translation vector of each camera as lambdakck
CN202110289301.0A 2021-03-18 2021-03-18 Multi-phase external parameter automatic calibration method based on human skeleton extraction Active CN113077519B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110289301.0A CN113077519B (en) 2021-03-18 2021-03-18 Multi-phase external parameter automatic calibration method based on human skeleton extraction

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110289301.0A CN113077519B (en) 2021-03-18 2021-03-18 Multi-phase external parameter automatic calibration method based on human skeleton extraction

Publications (2)

Publication Number Publication Date
CN113077519A true CN113077519A (en) 2021-07-06
CN113077519B CN113077519B (en) 2022-12-09

Family

ID=76612748

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110289301.0A Active CN113077519B (en) 2021-03-18 2021-03-18 Multi-phase external parameter automatic calibration method based on human skeleton extraction

Country Status (1)

Country Link
CN (1) CN113077519B (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113925497A (en) * 2021-10-22 2022-01-14 吉林大学 Automobile passenger riding posture extraction method based on binocular vision measurement system
CN114241602A (en) * 2021-12-16 2022-03-25 北京理工大学 Multi-purpose rotational inertia measuring and calculating method based on deep learning
CN114758016A (en) * 2022-06-15 2022-07-15 超节点创新科技(深圳)有限公司 Camera equipment calibration method, electronic equipment and storage medium
WO2023071939A1 (en) * 2021-10-27 2023-05-04 华为技术有限公司 Parameter determination method and related device
CN117934572A (en) * 2024-03-21 2024-04-26 南京起源跃动科技有限公司 Method for aligning 3D skeleton coordinate system with VR head display coordinate system

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102034238A (en) * 2010-12-13 2011-04-27 西安交通大学 Multi-camera system calibrating method based on optical imaging test head and visual graph structure
CN103456016A (en) * 2013-09-06 2013-12-18 同济大学 Body-feeling camera network calibration method unrelated to visual angles
CN108288291A (en) * 2018-06-07 2018-07-17 北京轻威科技有限责任公司 Polyphaser calibration based on single-point calibration object
CN110458897A (en) * 2019-08-13 2019-11-15 北京积加科技有限公司 Multi-cam automatic calibration method and system, monitoring method and system
CN110969668A (en) * 2019-11-22 2020-04-07 大连理工大学 Stereoscopic calibration algorithm of long-focus binocular camera
CN111028271A (en) * 2019-12-06 2020-04-17 浩云科技股份有限公司 Multi-camera personnel three-dimensional positioning and tracking system based on human skeleton detection
CN111667540A (en) * 2020-06-09 2020-09-15 中国电子科技集团公司第五十四研究所 Multi-camera system calibration method based on pedestrian head recognition
CN111739103A (en) * 2020-06-18 2020-10-02 苏州炫感信息科技有限公司 Multi-camera calibration system based on single-point calibration object
CN112001926A (en) * 2020-07-04 2020-11-27 西安电子科技大学 RGBD multi-camera calibration method and system based on multi-dimensional semantic mapping and application

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102034238A (en) * 2010-12-13 2011-04-27 西安交通大学 Multi-camera system calibrating method based on optical imaging test head and visual graph structure
CN103456016A (en) * 2013-09-06 2013-12-18 同济大学 Body-feeling camera network calibration method unrelated to visual angles
CN108288291A (en) * 2018-06-07 2018-07-17 北京轻威科技有限责任公司 Polyphaser calibration based on single-point calibration object
CN110458897A (en) * 2019-08-13 2019-11-15 北京积加科技有限公司 Multi-cam automatic calibration method and system, monitoring method and system
CN110969668A (en) * 2019-11-22 2020-04-07 大连理工大学 Stereoscopic calibration algorithm of long-focus binocular camera
CN111028271A (en) * 2019-12-06 2020-04-17 浩云科技股份有限公司 Multi-camera personnel three-dimensional positioning and tracking system based on human skeleton detection
CN111667540A (en) * 2020-06-09 2020-09-15 中国电子科技集团公司第五十四研究所 Multi-camera system calibration method based on pedestrian head recognition
CN111739103A (en) * 2020-06-18 2020-10-02 苏州炫感信息科技有限公司 Multi-camera calibration system based on single-point calibration object
CN112001926A (en) * 2020-07-04 2020-11-27 西安电子科技大学 RGBD multi-camera calibration method and system based on multi-dimensional semantic mapping and application

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
ANH MINH TRUONG ET AL: "Automatic Multi-Camera Extrinsic Parameter Calibration Based on Pedestrian Torsors", 《SENSORS》 *

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113925497A (en) * 2021-10-22 2022-01-14 吉林大学 Automobile passenger riding posture extraction method based on binocular vision measurement system
CN113925497B (en) * 2021-10-22 2023-09-15 吉林大学 Binocular vision measurement system-based automobile passenger riding posture extraction method
WO2023071939A1 (en) * 2021-10-27 2023-05-04 华为技术有限公司 Parameter determination method and related device
CN114241602A (en) * 2021-12-16 2022-03-25 北京理工大学 Multi-purpose rotational inertia measuring and calculating method based on deep learning
CN114241602B (en) * 2021-12-16 2024-05-28 北京理工大学 Deep learning-based multi-objective moment of inertia measurement and calculation method
CN114758016A (en) * 2022-06-15 2022-07-15 超节点创新科技(深圳)有限公司 Camera equipment calibration method, electronic equipment and storage medium
CN114758016B (en) * 2022-06-15 2022-09-13 超节点创新科技(深圳)有限公司 Camera equipment calibration method, electronic equipment and storage medium
CN117934572A (en) * 2024-03-21 2024-04-26 南京起源跃动科技有限公司 Method for aligning 3D skeleton coordinate system with VR head display coordinate system
CN117934572B (en) * 2024-03-21 2024-05-28 南京起源跃动科技有限公司 Method for aligning 3D skeleton coordinate system with VR head display coordinate system

Also Published As

Publication number Publication date
CN113077519B (en) 2022-12-09

Similar Documents

Publication Publication Date Title
CN113077519B (en) Multi-phase external parameter automatic calibration method based on human skeleton extraction
Xiang et al. Monocular total capture: Posing face, body, and hands in the wild
CN107392964B (en) The indoor SLAM method combined based on indoor characteristic point and structure lines
CN108154550B (en) RGBD camera-based real-time three-dimensional face reconstruction method
US9235928B2 (en) 3D body modeling, from a single or multiple 3D cameras, in the presence of motion
CN111126304A (en) Augmented reality navigation method based on indoor natural scene image deep learning
CN103099623B (en) Extraction method of kinesiology parameters
CN109919141A (en) A kind of recognition methods again of the pedestrian based on skeleton pose
CN107194991B (en) Three-dimensional global visual monitoring system construction method based on skeleton point local dynamic update
US20150243035A1 (en) Method and device for determining a transformation between an image coordinate system and an object coordinate system associated with an object of interest
CN104794722A (en) Dressed human body three-dimensional bare body model calculation method through single Kinect
CN111862299A (en) Human body three-dimensional model construction method and device, robot and storage medium
Wang et al. Outdoor markerless motion capture with sparse handheld video cameras
CN114529605A (en) Human body three-dimensional attitude estimation method based on multi-view fusion
CN112401369A (en) Body parameter measuring method, system, equipment, chip and medium based on human body reconstruction
CN111998862A (en) Dense binocular SLAM method based on BNN
CN117671738B (en) Human body posture recognition system based on artificial intelligence
CN108073855A (en) A kind of recognition methods of human face expression and system
CN113255487A (en) Three-dimensional real-time human body posture recognition method
CN115376034A (en) Motion video acquisition and editing method and device based on human body three-dimensional posture space-time correlation action recognition
Cao et al. Camera calibration using symmetric objects
CN113284249B (en) Multi-view three-dimensional human body reconstruction method and system based on graph neural network
CN114882106A (en) Pose determination method and device, equipment and medium
CN112365589B (en) Virtual three-dimensional scene display method, device and system
CN111667540B (en) Multi-camera system calibration method based on pedestrian head recognition

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant