WO2021063128A1 - Procédé de détermination de pose d'un corps rigide actif dans un environnement à caméra unique, et appareil associé - Google Patents

Procédé de détermination de pose d'un corps rigide actif dans un environnement à caméra unique, et appareil associé Download PDF

Info

Publication number
WO2021063128A1
WO2021063128A1 PCT/CN2020/110254 CN2020110254W WO2021063128A1 WO 2021063128 A1 WO2021063128 A1 WO 2021063128A1 CN 2020110254 W CN2020110254 W CN 2020110254W WO 2021063128 A1 WO2021063128 A1 WO 2021063128A1
Authority
WO
WIPO (PCT)
Prior art keywords
matrix
dimensional space
rigid body
space point
coordinates
Prior art date
Application number
PCT/CN2020/110254
Other languages
English (en)
Chinese (zh)
Inventor
王越
许秋子
Original Assignee
深圳市瑞立视多媒体科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 深圳市瑞立视多媒体科技有限公司 filed Critical 深圳市瑞立视多媒体科技有限公司
Publication of WO2021063128A1 publication Critical patent/WO2021063128A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods

Definitions

  • the invention relates to the technical field of computer vision, in particular to a method for positioning an active rigid body in a single-camera environment and related equipment.
  • the traditional optical motion capture method uses the ultra-high-power near-infrared light source in the motion capture camera to emit infrared light and irradiate it on the passive marking point; the marking point coated with high-reflective material reflects the irradiated infrared light, and this part of the infrared light
  • the ambient light with background information will pass through the low-distortion lens and reach the infrared narrow band pass filter unit of the camera. Since the light band of the infrared narrow band pass filter unit is the same as that of the infrared light source, the ambient light with redundant background information will be filtered out, leaving only the infrared light with the marking point information to pass through and be taken by the camera. Photosensitive element recording.
  • the photosensitive element then converts the light signal into an image signal and outputs it to the control circuit, and the image processing unit in the control circuit uses a Field Programmable Gate Array (FPGA) to preprocess the image signal in the form of hardware, and finally to The tracking software flows out the 2D coordinate information of the marked points.
  • FPGA Field Programmable Gate Array
  • the system server needs to receive the 2D data of each camera in the multi-camera system, and then adopt the principle of multi-eye vision, according to the matching relationship between 2D point clouds And calibrate the calculated pose relationship between the cameras in advance, calculate the 3D coordinates in the three-dimensional space, and use this as the basis to calculate the motion information of the rigid body in the space.
  • This method relies on the collaborative work between multiple cameras, so that it can be used in a relatively large space to realize the identification and tracking of rigid bodies, which leads to the high cost and difficult maintenance of the motion capture system.
  • the main purpose of the present invention is to provide an active rigid body pose positioning method and related equipment in a single-camera environment, aiming to solve the high cost and difficult-to-maintain technology caused by the use of multi-camera systems in current passive or active motion capture methods problem.
  • the present invention provides a method for positioning an active rigid body in a single-camera environment.
  • the method includes the following steps:
  • the two-dimensional space point code corresponding to the two-dimensional space point coordinates, and the camera parameters of the camera.
  • the phase Matching the two-dimensional spatial point coordinates of two adjacent frames to obtain multiple sets of two-dimensional spatial feature pairs, constructing a linear equation system from the multiple sets of two-dimensional spatial feature pairs and the camera parameters, and solving the essential matrix;
  • the two-dimensional space feature pair multiple sets of the rotation matrix and the translation matrix, the three-dimensional space point coordinates are estimated, the depth value of the three-dimensional space point coordinates is detected, and the set of the rotation matrix whose depth value is a positive number
  • the translation matrix is defined as a target rotation matrix and a target translation matrix, and the rigid body pose is determined according to the target rotation matrix and the target translation matrix.
  • the determining the rigid body pose according to the target rotation matrix and the target translation matrix includes:
  • the target translation matrix is optimized by an optimization formula to obtain an optimized target translation matrix, and the rigid body pose is determined according to the target rotation matrix and the optimized target translation matrix;
  • the optimization formula is:
  • L1 is the three-dimensional average distance
  • L2 is the average distance of the rigid body
  • T is the target translation matrix before optimization
  • T′ is the target translation matrix after optimization.
  • the acquiring rigid body coordinates, summing the distances between all rigid body mark points in the rigid body coordinates and then taking the average value, before obtaining the average distance of the rigid body includes:
  • the coordinates of all three-dimensional space points in the same frame are converted into rigid body coordinates in the rigid body coordinate system, and the rigid body coordinates of each of the marked points in each frame are obtained.
  • the multiple cameras are matched in pairs, and each of the marked points is obtained according to the spatial position data of the two cameras and the coordinates of the multiple two-dimensional spatial points in the same frame.
  • the three-dimensional space point coordinates of the frame including:
  • the two two-dimensional space point coordinates captured by the two matching cameras in the same frame are matched by pairwise matching of all the cameras that have captured the same marked point, and the least squares method is solved by singular value decomposition. Calculate a set of three-dimensional space point coordinates;
  • the conversion of all three-dimensional space point coordinates in the same frame into rigid body coordinates in a rigid body coordinate system to obtain the rigid body coordinates of each of the marked points in each frame includes:
  • the difference between the origin and the coordinates of the three-dimensional space points corresponding to each of the marking points in the same frame is respectively calculated to obtain the rigid body coordinates of each of the marking points in each frame.
  • the estimating three-dimensional spatial point coordinates through the two-dimensional spatial feature pair, multiple sets of the rotation matrix and the translation matrix includes:
  • the two cameras are camera 1 and camera 2
  • the two two-dimensional space point coordinates captured in the same frame are A(a1, a2), B(b1, b2)
  • the rotation matrix of camera 1 is R1( R11, R12, R13)
  • R1 is a 3*3 matrix
  • the translation matrix is T1 (T11, T12, T13)
  • T1 is a 3*1 matrix
  • the rotation matrix of camera 2 is R2 (R21, R22, R23)
  • the translation matrix is T2 (T21, T22, T23).
  • R2 is a 3*3 matrix
  • T2 is a 3*1 matrix.
  • the three-dimensional space point coordinates can be obtained by the following method:
  • the detecting the depth value of the three-dimensional space point coordinates, and defining the group of the rotation matrix and the translation matrix whose depth value is a positive number as the target rotation matrix and the target translation matrix includes:
  • the estimated three-dimensional space point coordinates detect whether the depth value corresponding to the three-dimensional space point coordinates is a positive number, and if so, define the corresponding set of the rotation matrix and translation matrix as the target rotation matrix and target translation matrix.
  • the present invention also provides an active rigid body pose positioning device in a single-camera environment, including:
  • the calculation essential matrix module is used to obtain the two-dimensional space point coordinates of two adjacent frames captured by the monocular camera, the two-dimensional space point code corresponding to the two-dimensional space point coordinates, and the camera parameters of the camera, according to the two Two-dimensional space point coding, matching the two-dimensional space point coordinates of two adjacent frames to obtain multiple sets of two-dimensional space feature pairs, and constructing a linear equation system from multiple sets of the two-dimensional space feature pairs and the camera parameters, Solve the essential matrix;
  • Calculating rotation matrix and translation matrix module which is used to decompose the essential matrix through a singular value decomposition algorithm to obtain multiple sets of rotation matrix and translation matrix;
  • the rigid body pose determination module is used to estimate the three-dimensional space point coordinates through the two-dimensional space feature pair, multiple sets of the rotation matrix and the translation matrix, detect the depth value of the three-dimensional space point coordinate, and set the depth value to be positive
  • the number of rotation matrices and translation matrices are defined as a target rotation matrix and a target translation matrix, and the rigid body pose is determined according to the target rotation matrix and the target translation matrix.
  • the present invention also provides a device for positioning an active rigid body in a single-camera environment.
  • the device includes a memory, a processor, and a device that is stored in the memory and can run on the processor.
  • a pose positioning program for an active rigid body in a single-camera environment where the pose positioning program for an active rigid body in a single-camera environment is executed by the processor to realize the pose positioning of an active rigid body in a single-camera environment as described above Method steps.
  • the present invention also provides a computer-readable storage medium that stores a program for positioning the pose of an active rigid body in a single-camera environment.
  • the pose positioning program is executed by the processor, the steps of the active rigid body pose positioning method in the single-camera environment as described above are realized.
  • the essence matrix is solved by matching the characteristic points in the coordinates of two adjacent frames; and the essence is graded by the singular value decomposition algorithm Matrix, multiple groups of rotation matrix and translation matrix are obtained; by detecting the depth value of the feature point, the final target rotation matrix and translation matrix are determined.
  • the whole process does not depend on the rigid body structure, and the required matching data can be obtained according to the code and coordinates to calculate the rigid body pose information.
  • the present invention can realize the tracking and positioning of an active optical rigid body at a lower cost, and has obvious advantages compared with a complex multi-camera environment.
  • the present invention matches the feature points of two adjacent frames each time, so that the active light rigid body can be tracked and positioned every time the current frame is compared to the initial frame's motion posture, thereby avoiding the common features of monocular camera tracking.
  • the cumulative error problem further improves the tracking accuracy.
  • FIG. 1 is a schematic structural diagram of the operating environment of an active rigid body pose positioning device in a single-camera environment related to a solution of an embodiment of the present invention
  • FIG. 2 is a flowchart of a method for positioning an active rigid body in a single-camera environment in an embodiment of the present invention
  • FIG. 3 is a detailed flowchart of step S3 in an embodiment of the present invention.
  • step S302 is a detailed flowchart of step S302 in an embodiment of the present invention.
  • Fig. 5 is a structural diagram of an active rigid body pose positioning device in a single-camera environment in an embodiment of the present invention.
  • the active rigid body pose positioning device in a single-camera environment includes: a processor 1001, such as a CPU, a communication bus 1002, a user interface 1003, a network interface 1004, and a memory 1005.
  • the communication bus 1002 is used to implement connection and communication between these components.
  • the user interface 1003 may include a display screen (Display) and an input unit such as a keyboard (Keyboard), and the network interface 1004 may optionally include a standard wired interface and a wireless interface (such as a WI-FI interface).
  • the memory 1005 may be a high-speed RAM memory, or a non-volatile memory (non-volatile memory), such as a magnetic disk memory.
  • the memory 1005 may also be a storage device independent of the aforementioned processor 1001.
  • the hardware structure of the active rigid body pose positioning device in the single-camera environment shown in FIG. 1 does not constitute a limitation on the active rigid body pose positioning device in the single-camera environment, and may include ratios More or fewer parts are shown, or some parts are combined, or different parts are arranged.
  • the memory 1005 which is a computer-readable storage medium, may include an operating system, a network communication module, a user interface module, and a pose positioning program for an active rigid body in a single-camera environment.
  • the operating system is a program that manages and controls the pose positioning equipment and software resources of the active rigid body in the single-camera environment, and supports the operation of the pose positioning program of the active rigid body in the single-camera environment and other software and/or programs.
  • the network interface 1004 is mainly used to access the network; the user interface 1003 is mainly used to detect confirmation commands and edit commands, and process
  • the device 1001 can be used to call the pose positioning program of the active rigid body in the single-camera environment stored in the memory 1005, and execute the operations of the following embodiments of the pose positioning method of the active rigid body in the single-camera environment.
  • FIG. 2 is a flowchart of a method for positioning an active rigid body in a single-camera environment in an embodiment of the present invention. As shown in FIG. 2, a method for positioning an active rigid body in a single-camera environment includes The following steps:
  • Step S1 solving the essential matrix: Obtain the two-dimensional point coordinates of the two adjacent frames captured by the monocular camera, the two-dimensional point code corresponding to the two-dimensional point coordinates, and the camera parameters of the camera, according to the two-dimensional point code, The two-dimensional space point coordinates of two adjacent frames are matched to obtain multiple sets of two-dimensional spatial feature pairs. The multiple sets of two-dimensional spatial feature pairs and camera parameters are constructed to form a linear equation set to solve the essential matrix.
  • the marker points in this step are generally set at different positions of the rigid body.
  • the two-dimensional space coordinate information of the marker point is captured by the monocular camera to determine the spatial point data.
  • the spatial point data includes two-dimensional space. Point coordinates and corresponding two-dimensional space point codes.
  • marking points there are eight marking points on the rigid body, and the marking points can be eight luminous LED lights. Therefore, a rigid body usually contains eight spatial point data, and each frame of monocular camera data contains eight mark points.
  • the coding of the same mark point in different frames is the same, and different mark points are in the same frame. The encoding is different.
  • the two-dimensional spatial feature pair is the projection of the same marker point in two adjacent frames on the monocular camera.
  • a rigid body contains eight marker points, it has eight sets of two-dimensional spatial feature pairs.
  • the camera parameters of the monocular camera need to be calibrated, that is, the camera's optical center, focal length and distortion parameters, etc. These camera parameters are used as a matrix, recorded as matrix M, and used in the essential matrix calculation.
  • the principle of epipolar geometric constraint is adopted. The linear equations are constructed for multiple sets of two-dimensional spatial feature pairs and camera parameters in the following way to solve the essential matrix:
  • Step S2 Decompose the essential matrix: Decompose the essential matrix through a singular value decomposition algorithm to obtain multiple sets of rotation matrices and translation matrices.
  • the motion information of the rigid body is recovered according to the essential matrix: rotation matrix R and translation matrix T.
  • This process is obtained by singular value decomposition (SVD) in this step.
  • SVD singular value decomposition
  • a total of four possible solutions R, T
  • R, T four sets of rotation matrices and translation matrices, of which only one correct solution is available in the monocular camera Positive depth (the depth value is a positive number). Therefore, the next step of detecting depth information is required.
  • Step S3 Determine the pose of the rigid body: Estimate the three-dimensional space point coordinates through the two-dimensional space feature pairs, multiple sets of rotation matrices and translation matrices, detect the depth value of the three-dimensional space point coordinates, and set the group of rotation matrices with a positive depth value
  • the translation matrix is defined as the target rotation matrix and the target translation matrix, and the rigid body pose is determined according to the target rotation matrix and the target translation matrix.
  • step S2 After decomposing the essential matrix using singular value decomposition in step S2, four possible solutions are obtained. Therefore, in this step, it is necessary to finally determine the correct solution among the four possible solutions. First, it is necessary to estimate the coordinates of the three-dimensional space point, and detect the depth value of the feature point according to the three-dimensional space point coordinate. Only the set of solutions (R, T) with a positive depth value is the final target (R, T).
  • step S3 estimating three-dimensional space point coordinates through two-dimensional spatial feature pairs, multiple sets of rotation matrices and translation matrices, further includes:
  • the two cameras are camera 1 and camera 2
  • the two two-dimensional space point coordinates captured in the same frame are A(a1, a2), B(b1, b2)
  • the rotation matrix of camera 1 is R1( R11, R12, R13)
  • R1 is a 3*3 matrix
  • the translation matrix is T1 (T11, T12, T13)
  • T1 is a 3*1 matrix
  • the rotation matrix of camera 2 is R2 (R21, R22, R23)
  • the translation matrix is T2 (T21, T22, T23).
  • R2 is a 3*3 matrix
  • T2 is a 3*1 matrix.
  • the coordinates of a three-dimensional space point C(c1, c2) in the same frame are obtained by the following method , C3):
  • multiple different rotation matrices and translation matrices such as multiple sets of R1, T1, R2, and T2 rotation matrix and translation matrix data pairs, multiple different three-dimensional space point coordinates are obtained.
  • step S2 For example, if four sets of rotation matrices and translation matrices are obtained in step S2, four different three-dimensional space point coordinates can be estimated through this step, but there is only one three-dimensional space point coordinate C whose coordinate value c3 is greater than 0, then The R and T corresponding to the coordinate C of the three-dimensional space point are the final target data.
  • the matched sets of two-dimensional space feature pairs are combined with four possible solutions (R, T), and the corresponding three-dimensional space coordinate data (x, y, z) is estimated by the above method according to the principle of triangulation. , Provide accurate data for the subsequent detection depth value z.
  • step S3 the depth value of the three-dimensional space point coordinates is detected, and the group of rotation matrices and translation matrices with a positive depth value are defined as the target rotation matrix and the target translation matrix, including:
  • the estimated three-dimensional space point coordinates it is detected whether the depth value corresponding to the three-dimensional space point coordinates is a positive number, and if so, the corresponding group of rotation matrix and translation matrix is defined as the target rotation matrix and the target translation matrix.
  • multiple depth values z are obtained by solving in the above-mentioned manner, and the corresponding solution (R, T) when the depth value z is zero or negative is eliminated, and the corresponding solution (R, T) when the depth value z is positive is retained. And as the final target data, the rigid body pose is determined with this target data.
  • step S3 after defining the set of rotation matrices and translation matrices with a positive depth value as the target rotation matrix and the target translation matrix, before determining the rigid body pose according to the target rotation matrix and the target translation matrix, such as As shown in Figure 3, it includes:
  • Step S301 Calculate the three-dimensional average distance: sum the distances between all three-dimensional space points in the three-dimensional space point coordinates and take the average value to obtain the three-dimensional average distance.
  • D is the distance between two three-dimensional space points
  • (a 1 , a 2 , a 3 ) is the three-dimensional space point coordinates of the three-dimensional space point 1
  • (b 1 , b 2 , b 3 ) is the three-dimensional space point 2
  • Step S302 Calculate the average distance of the rigid body: obtain the coordinates of the rigid body, sum the distances between all rigid body mark points in the coordinates of the rigid body and take the average value to obtain the average distance of the rigid body.
  • a distance calculation formula similar to that in step S301 can be used to calculate the distance between two rigid body mark points in the rigid body coordinate system after calculating the distance between the two coordinates of the rigid body coordinate system, and then the sum is performed and the average value is taken.
  • the rigid body coordinates in this step can be obtained by actually measuring the rigid body coordinates of the marked points, as shown in Figure 4, or can be obtained by a multi-camera system, that is, in the following way, accurate rigid body coordinates can be obtained with only one initialization. Multiple calculations:
  • Step S30201 acquiring data: acquiring the two-dimensional point coordinates of two adjacent frames captured by multiple cameras, the two-dimensional point codes corresponding to the two-dimensional point coordinates, and the spatial position data of multiple cameras, and encode the two-dimensional points
  • the coordinates of the same multiple two-dimensional space points are classified into the same kind, and are marked under the same marking point.
  • the marker points in this step are generally set at different positions of the rigid body.
  • the two-dimensional space coordinate information of the marker points is captured by multiple cameras, and the spatial point data is determined through the preset rigid body encoding technology.
  • the spatial point data includes two-dimensional spatial points. Coordinates and corresponding two-dimensional space point codes.
  • the spatial position data is obtained by calibrating and calculating the spatial position relationship of each camera.
  • there are eight marking points on the rigid body and the marking points can be eight luminous LED lights. Therefore, a rigid body usually contains eight spatial point data.
  • each frame of data for a single camera contains the spatial point data of eight marker points.
  • the encoding of the same marker point in different frames is the same.
  • the spatial point data with the same spatial point code in all cameras can be divided together as the same type, and these spatial point data are considered to be projections of the same marker point in space on different cameras.
  • Step S30202 Calculate the three-dimensional space data: match multiple cameras in pairs, and obtain the three-dimensional space point coordinates of each marker point per frame according to the spatial position data of the two cameras and the multiple two-dimensional space point coordinates of the same frame .
  • the processing of this step is performed on each frame of data of each marker point.
  • multiple cameras that capture the marker point are matched in pairs.
  • singular value decomposition singular Value Decomposition
  • Decomposition, SVD solve the least square method to obtain a set of three-dimensional space point data.
  • the rigid body includes eight marker points
  • eight three-dimensional space point codes and three-dimensional space point coordinates of the eight marker points are obtained through this step.
  • This step further includes:
  • the two cameras are camera 1 and camera 2
  • the coordinates of two two-dimensional space points captured in the same frame are A(a1, a2), B(b1, b2)
  • the rotation matrix of camera 1 is R1( R11, R12, R13)
  • R1 is a 3*3 matrix
  • the translation matrix is T1 (T11, T12, T13)
  • T1 is a 3*1 matrix
  • the rotation matrix of camera 2 is R2 (R21, R22, R23)
  • the translation matrix is T2 (T21, T22, T23).
  • R2 is a 3*3 matrix
  • the translation matrix is T2
  • T2 is a 3*1 matrix.
  • the coordinates of a three-dimensional space point in the same frame are obtained by the following method C(c1, c2, c3):
  • the two two-dimensional space point coordinates captured by all pairwise matching cameras are finally calculated to obtain a set of three-dimensional space point coordinates.
  • this threshold range is a coordinate parameter preset in advance. If the three-dimensional space point coordinates are found to deviate from the threshold range, the three-dimensional space point coordinates are considered to be wrong data and eliminated.
  • Step S30203 Calculate rigid body coordinates: convert all three-dimensional space point codes and three-dimensional space point coordinates in the same frame into rigid body coordinates in the rigid body coordinate system, and obtain the rigid body coordinates of each marker point and each frame.
  • the three-dimensional space point data corresponding to each marking point can be obtained, and the multiple three-dimensional space point data corresponding to multiple marking points can be formed into a rigid body. If the rigid body currently in use has eight luminous LED lights, the rigid body Contains eight three-dimensional space point data. Through multiple three-dimensional space point data, such as three-dimensional space point coordinates in eight three-dimensional space point data, it can be transformed into rigid body coordinates in a rigid body coordinate system.
  • This step further includes:
  • the average value is calculated for each dimension of the three-dimensional space point coordinates corresponding to all the mark points in the same frame to obtain the coordinate average value, and this coordinate average value is recorded as the origin in the rigid coordinate system as all the mark points
  • the reference data of the corresponding three-dimensional space point coordinates is obtained.
  • step S2 obtains eight three-dimensional space point coordinate data, and calculates the average value of each dimension of the eight three-dimensional space point coordinate data to obtain the coordinate average value.
  • the average value of the coordinates is taken as the origin in the rigid body coordinate system, and the difference between the coordinates of each three-dimensional space point and the origin is calculated, and the difference obtained is the rigid body coordinate of each marked point.
  • the three-dimensional space point coordinates corresponding to the eight mark points are calculated as the difference from the origin.
  • the difference between the coordinates of each dimension and the dimension coordinates corresponding to the origin is calculated, and finally Get eight rigid body coordinates.
  • multiple cameras are used to capture multiple two-dimensional space point coordinates, a set of three-dimensional space point data is analyzed through a specific solution algorithm, and after operations such as integration, averaging, and optimization of multiple three-dimensional space point data are performed, Finally, more accurate three-dimensional space point data is obtained, and the accurate three-dimensional space point data is converted into rigid body coordinate data in the rigid body coordinate system, which provides definite and accurate data for the subsequent calculation of the average distance of the rigid body.
  • Step S303 optimization: the target translation matrix is optimized by the optimization formula to obtain the optimized target translation matrix, and the rigid body pose is determined according to the target rotation matrix and the optimized target translation matrix.
  • the optimization formula is:
  • L1 is the three-dimensional average distance
  • L2 is the average distance of the rigid body
  • T is the target translation matrix before optimization
  • T′ is the target translation matrix after optimization.
  • the translation amount may have various situations, so there is no guarantee that the translation matrix T is accurate and true data.
  • the three-dimensional space point coordinates of the rigid body are estimated by the triangulation principle, the three-dimensional space point coordinates and the rigid body coordinates in the rigid body coordinate system are estimated according to the estimated three-dimensional space point coordinates and the rigid body coordinates.
  • the target translation matrix is optimized.
  • the optimized target translation matrix is obtained through the above optimization formula, so that the final rigid body pose is more accurate and more authentic.
  • the pose positioning method of the active rigid body in the single-camera environment of this embodiment The active optical rigid body has coding information so that the motion capture tracking and positioning no longer depends on the rigid body structure, but can directly obtain matching two-dimensional spatial features based on the coding information. Yes, to solve the rigid body pose.
  • the invention can realize the tracking and positioning of a rigid body at a lower cost, which has obvious advantages compared with a complex multi-camera environment.
  • the encoding information of the active optical rigid body is used to match the adjacent two frames, each time the active optical rigid body is tracked and positioned, the motion posture of the current frame compared to the initial frame can be calculated, thereby avoiding monocular camera tracking
  • the common cumulative error problem further improves the tracking accuracy.
  • a device for positioning an active rigid body in a single-camera environment is proposed. As shown in FIG. 5, the device includes:
  • the calculation essential matrix module is used to obtain the two-dimensional space point coordinates of two adjacent frames captured by the monocular camera, the two-dimensional space point code corresponding to the two-dimensional point coordinates, and the camera parameters of the camera. According to the two-dimensional space point code, Match the two-dimensional space point coordinates of two adjacent frames to obtain multiple sets of two-dimensional spatial feature pairs, construct a linear equation system from multiple sets of two-dimensional spatial feature pairs and camera parameters, and solve the essential matrix;
  • Calculate rotation matrix and translation matrix module which is used to decompose the essential matrix through the singular value decomposition algorithm to obtain multiple sets of rotation matrix and translation matrix;
  • the translation matrix is defined as the target rotation matrix and the target translation matrix, and the rigid body pose is determined according to the target rotation matrix and the target translation matrix.
  • this embodiment does not describe the contents of the embodiment of the active rigid body pose positioning device in the single-camera environment. Too much repeat.
  • a device for positioning an active rigid body in a single-camera environment includes a memory, a processor, and an active rigid body in a single-camera environment that is stored in the memory and can run on the processor.
  • the pose positioning program of the active rigid body in the single-camera environment is executed by the processor to implement the steps in the pose positioning method of the active rigid body in the single-camera environment of the foregoing embodiments.
  • a computer-readable storage medium stores a pose positioning program for an active rigid body in a single-camera environment, and the pose positioning program for an active rigid body in a single-camera environment is processed by the processor The steps in the active rigid body pose positioning method in the single-camera environment of the foregoing embodiments are implemented during execution.
  • the storage medium may be a non-volatile storage medium.
  • the program can be stored in a computer-readable storage medium, and the storage medium can include: Read only memory (ROM, Read Only Memory), random access memory (RAM, Random Access Memory), magnetic disk or optical disk, etc.

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Length Measuring Devices By Optical Means (AREA)

Abstract

La présente invention concerne le domaine technique de la vision artificielle, et concerne en particulier un procédé de détermination de pose d'un corps rigide actif dans un environnement à caméra unique, et un appareil associé. Le procédé comprend : l'acquisition de coordonnées de point spatial bidimensionnel et de codes de point spatial bidimensionnel de deux cadres adjacents, et un paramètre de caméra, la mise en correspondance, en fonction des codes de point spatial bidimensionnel, des coordonnées de point spatial bidimensionnel de façon à obtenir de multiples paires de caractéristiques spatiales bidimensionnelles, la construction d'un ensemble d'équations linéaires à partir des multiples paires de caractéristiques spatiales bidimensionnelles et du paramètre de caméra, et la résolution pour obtenir une matrice intrinsèque ; la décomposition de la matrice intrinsèque au moyen d'un algorithme de décomposition de valeurs singulières de manière à obtenir de multiples matrices de rotation et de translation ; et l'estimation des coordonnées de point spatial tridimensionnel, la détection d'une valeur de profondeur, la détermination d'une matrice de rotation cible et d'une matrice de translation cible, et la détermination d'une pose d'un corps rigide selon la matrice de rotation cible et la matrice de translation cible. L'invention permet de suivre et de positionner un corps rigide électroluminescent actif dans un environnement à caméra unique à faible coût, et est donc plus avantageuse que l'utilisation d'un réglage compliqué de caméras multiples.
PCT/CN2020/110254 2019-09-30 2020-08-20 Procédé de détermination de pose d'un corps rigide actif dans un environnement à caméra unique, et appareil associé WO2021063128A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201910938118.1A CN110689577B (zh) 2019-09-30 2019-09-30 单相机环境中主动式刚体的位姿定位方法及相关设备
CN201910938118.1 2019-09-30

Publications (1)

Publication Number Publication Date
WO2021063128A1 true WO2021063128A1 (fr) 2021-04-08

Family

ID=69111063

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/110254 WO2021063128A1 (fr) 2019-09-30 2020-08-20 Procédé de détermination de pose d'un corps rigide actif dans un environnement à caméra unique, et appareil associé

Country Status (2)

Country Link
CN (2) CN110689577B (fr)
WO (1) WO2021063128A1 (fr)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113610979A (zh) * 2021-07-12 2021-11-05 深圳市瑞立视多媒体科技有限公司 一种预警刚体之间相似度的方法、设备及光学动作捕捉系统
CN113850873A (zh) * 2021-09-24 2021-12-28 成都圭目机器人有限公司 一种线阵相机在搭载平台定位坐标系下的偏移位置标定方法
CN114742904A (zh) * 2022-05-23 2022-07-12 轻威科技(绍兴)有限公司 一种剔除干扰点后的商用立体相机组的标定方法及装置
CN115100287A (zh) * 2022-04-14 2022-09-23 美的集团(上海)有限公司 外参标定方法及机器人
CN117523678A (zh) * 2024-01-04 2024-02-06 广东茉莉数字科技集团股份有限公司 一种基于光学动作数据的虚拟主播区分方法及系统

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110689577B (zh) * 2019-09-30 2022-04-01 深圳市瑞立视多媒体科技有限公司 单相机环境中主动式刚体的位姿定位方法及相关设备
CN111566701B (zh) * 2020-04-02 2021-10-15 深圳市瑞立视多媒体科技有限公司 大空间环境下边扫场边标定方法、装置、设备及存储介质
CN113392909B (zh) * 2021-06-17 2022-12-27 深圳市睿联技术股份有限公司 数据处理方法、数据处理装置、终端及可读存储介质
CN113473210A (zh) * 2021-07-15 2021-10-01 北京京东方光电科技有限公司 显示方法、设备和存储介质
CN118298113B (zh) * 2024-06-05 2024-10-01 知行汽车科技(苏州)股份有限公司 一种三维重建方法、装置、设备及介质

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080174594A1 (en) * 2007-01-22 2008-07-24 Sharp Laboratories Of America, Inc. Method for supporting intuitive view specification in the free-viewpoint television application
CN103759670A (zh) * 2014-01-06 2014-04-30 四川虹微技术有限公司 一种基于数字近景摄影的物体三维信息获取方法
CN107341814A (zh) * 2017-06-14 2017-11-10 宁波大学 基于稀疏直接法的四旋翼无人机单目视觉测程方法
CN108648270A (zh) * 2018-05-12 2018-10-12 西北工业大学 基于eg-slam的无人机实时三维场景重建方法
CN110689577A (zh) * 2019-09-30 2020-01-14 深圳市瑞立视多媒体科技有限公司 单相机环境中主动式刚体的位姿定位方法及相关设备
CN110689584A (zh) * 2019-09-30 2020-01-14 深圳市瑞立视多媒体科技有限公司 多相机环境中主动式刚体的位姿定位方法及相关设备

Family Cites Families (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2007034964A (ja) * 2005-07-29 2007-02-08 Nippon Telegr & Teleph Corp <Ntt> カメラ視点運動並びに3次元情報の復元及びレンズ歪パラメータの推定方法、装置、カメラ視点運動並びに3次元情報の復元及びレンズ歪パラメータの推定プログラム
CN102564350A (zh) * 2012-02-10 2012-07-11 华中科技大学 基于面结构光和光笔的复杂零部件精密三维测量方法
CN102768767B (zh) * 2012-08-06 2014-10-22 中国科学院自动化研究所 刚体在线三维重建与定位的方法
CN103759716B (zh) * 2014-01-14 2016-08-17 清华大学 基于机械臂末端单目视觉的动态目标位置和姿态测量方法
CN104180818B (zh) * 2014-08-12 2017-08-11 北京理工大学 一种单目视觉里程计算装置
CN108151713A (zh) * 2017-12-13 2018-06-12 南京航空航天大学 一种单目vo快速位姿估计方法
CN109141396B (zh) * 2018-07-16 2022-04-26 南京航空航天大学 辅助信息与随机抽样一致算法融合的无人机位姿估计方法
CN110285827B (zh) * 2019-04-28 2023-04-07 武汉大学 一种距离约束的摄影测量高精度目标定位方法

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080174594A1 (en) * 2007-01-22 2008-07-24 Sharp Laboratories Of America, Inc. Method for supporting intuitive view specification in the free-viewpoint television application
CN103759670A (zh) * 2014-01-06 2014-04-30 四川虹微技术有限公司 一种基于数字近景摄影的物体三维信息获取方法
CN107341814A (zh) * 2017-06-14 2017-11-10 宁波大学 基于稀疏直接法的四旋翼无人机单目视觉测程方法
CN108648270A (zh) * 2018-05-12 2018-10-12 西北工业大学 基于eg-slam的无人机实时三维场景重建方法
CN110689577A (zh) * 2019-09-30 2020-01-14 深圳市瑞立视多媒体科技有限公司 单相机环境中主动式刚体的位姿定位方法及相关设备
CN110689584A (zh) * 2019-09-30 2020-01-14 深圳市瑞立视多媒体科技有限公司 多相机环境中主动式刚体的位姿定位方法及相关设备

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113610979A (zh) * 2021-07-12 2021-11-05 深圳市瑞立视多媒体科技有限公司 一种预警刚体之间相似度的方法、设备及光学动作捕捉系统
CN113610979B (zh) * 2021-07-12 2023-12-01 深圳市瑞立视多媒体科技有限公司 一种预警刚体之间相似度的方法、设备及光学动作捕捉系统
CN113850873A (zh) * 2021-09-24 2021-12-28 成都圭目机器人有限公司 一种线阵相机在搭载平台定位坐标系下的偏移位置标定方法
CN113850873B (zh) * 2021-09-24 2024-06-07 成都圭目机器人有限公司 一种线阵相机在搭载平台定位坐标系下的偏移位置标定方法
CN115100287A (zh) * 2022-04-14 2022-09-23 美的集团(上海)有限公司 外参标定方法及机器人
CN114742904A (zh) * 2022-05-23 2022-07-12 轻威科技(绍兴)有限公司 一种剔除干扰点后的商用立体相机组的标定方法及装置
CN114742904B (zh) * 2022-05-23 2024-07-02 轻威科技(绍兴)有限公司 一种剔除干扰点后的商用立体相机组的标定方法及装置
CN117523678A (zh) * 2024-01-04 2024-02-06 广东茉莉数字科技集团股份有限公司 一种基于光学动作数据的虚拟主播区分方法及系统
CN117523678B (zh) * 2024-01-04 2024-04-05 广东茉莉数字科技集团股份有限公司 一种基于光学动作数据的虚拟主播区分方法及系统

Also Published As

Publication number Publication date
CN110689577B (zh) 2022-04-01
CN114170307A (zh) 2022-03-11
CN110689577A (zh) 2020-01-14

Similar Documents

Publication Publication Date Title
WO2021063128A1 (fr) Procédé de détermination de pose d&#39;un corps rigide actif dans un environnement à caméra unique, et appareil associé
WO2021063127A1 (fr) Procédé de positionnement de pose et équipement associé de corps rigide actif dans un environnement multi-caméras
CN110136208B (zh) 一种机器人视觉伺服系统的联合自动标定方法及装置
CN106780601B (zh) 一种空间位置追踪方法、装置及智能设备
JP6855587B2 (ja) 視点から距離情報を取得するための装置及び方法
TWI624170B (zh) 影像掃描系統及其方法
CN106875435B (zh) 获取深度图像的方法及系统
CN112150528A (zh) 一种深度图像获取方法及终端、计算机可读存储介质
CN107808398B (zh) 摄像头参数算出装置以及算出方法、程序、记录介质
TWI393980B (zh) The method of calculating the depth of field and its method and the method of calculating the blurred state of the image
US20210256299A1 (en) System and method for correspondence map determination
JP2004340840A (ja) 距離測定装置、距離測定方法、及び距離測定プログラム
CN109640066B (zh) 高精度稠密深度图像的生成方法和装置
JP7489253B2 (ja) デプスマップ生成装置及びそのプログラム、並びに、デプスマップ生成システム
CN111160233B (zh) 基于三维成像辅助的人脸活体检测方法、介质及系统
US10049454B2 (en) Active triangulation calibration
JP6288770B2 (ja) 顔検出方法、顔検出システム、および顔検出プログラム
CN105427302B (zh) 一种基于移动稀疏相机采集阵列的三维采集及重建系统
US11166005B2 (en) Three-dimensional information acquisition system using pitching practice, and method for calculating camera parameters
US11195290B2 (en) Apparatus and method for encoding in structured depth camera system
KR101866107B1 (ko) 평면 모델링을 통한 깊이 정보 보정 방법과 보정 장치 및 부호화 장치
WO2019116518A1 (fr) Dispositif de détection d&#39;objet et procédé de détection d&#39;objet
CN115375772B (zh) 相机标定方法、装置、设备及存储介质
US11882262B2 (en) System and method for stereoscopic image analysis
WO2024009528A1 (fr) Dispositif de calcul de paramètre de caméra, procédé de calcul de paramètre de caméra et programme de calcul de paramètre de caméra

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20871576

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20871576

Country of ref document: EP

Kind code of ref document: A1