WO2020038386A1 - Determination of scale factor in monocular vision-based reconstruction - Google Patents
Determination of scale factor in monocular vision-based reconstruction Download PDFInfo
- Publication number
- WO2020038386A1 WO2020038386A1 PCT/CN2019/101704 CN2019101704W WO2020038386A1 WO 2020038386 A1 WO2020038386 A1 WO 2020038386A1 CN 2019101704 W CN2019101704 W CN 2019101704W WO 2020038386 A1 WO2020038386 A1 WO 2020038386A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- specified
- monocular camera
- moment
- designated
- pose
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/50—Depth or shape recovery
- G06T7/55—Depth or shape recovery from multiple images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10016—Video; Image sequence
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30248—Vehicle exterior or interior
- G06T2207/30252—Vehicle exterior; Vicinity of vehicle
Definitions
- the present application relates to the field of mobile robot technology, and in particular, to a method for determining a mesoscale factor for monocular vision reconstruction and a mobile robot.
- the simultaneous positioning and map construction algorithms based on monocular vision have become the focus of current mobile robot research.
- the traditional simultaneous positioning and map construction methods based on monocular vision can only achieve 3D reconstruction at projective scale or affine-scaling, that is, there is a scale factor between the reconstructed scene and the real-world scene.
- the scale factor is the ratio of the real world map scale to the constructed map scale. Therefore, if the scale factor can be determined when the mobile robot is initialized, the actual rotation and translation of the monocular camera in the real world can be calculated based on the projection model, and a map with the same scale as the real world can be constructed.
- the present application provides a method for determining a mesoscale factor for monocular vision reconstruction and a mobile robot.
- a first aspect of the present application provides a method for determining a mesoscale factor in monocular vision reconstruction.
- the method is applied to a mobile robot, and the method includes:
- the ratio of the modulus of the actual translation vector to the modulus of the normalized translation vector is determined as a scale factor in the monocular vision reconstruction of the device.
- a second aspect of the present application provides a mobile robot, which includes a monocular camera and a processor; wherein,
- the monocular camera is configured to acquire a first image of a designated object at a first moment and a second image of the designated object at a second moment;
- the processor is configured to:
- the ratio of the modulus of the actual translation vector to the modulus of the normalized translation vector is determined as a scale factor in the monocular vision reconstruction of the device.
- a third aspect of the present application provides a computer-readable storage medium on which a computer program is stored, and the program is executed by a processor to implement the steps of any of the methods provided in the first aspect of the present application.
- the method for determining the meso-scale factor of the monocular vision reconstruction and the mobile robot provided in this application, because the position of the designated object is fixed, therefore, in the case of the mobile robot slipping, jamming, etc., by calculating the designated object at the first moment The first pose relative to the monocular camera and the second pose of the designated object relative to the monocular camera at the second moment, so that the real-time The actual translation vector. Therefore, the method provided in this application does not have the problem that the determined scale factor is inaccurate due to slipping, jamming, and the like of the mobile robot.
- FIG. 1 is a flowchart of Embodiment 1 of a method for determining a scale factor in monocular vision reconstruction provided by the present application.
- Fig. 2 is a schematic diagram of a monocular camera acquiring an image of a specified object according to an exemplary embodiment of the present application.
- Fig. 3 is a flowchart of calculating a pose of a specified object relative to a monocular camera according to an exemplary embodiment of the present application.
- FIG. 4 is a hardware structural diagram of a first embodiment of a mobile robot provided in this application.
- first, second, third, etc. may be used in this application to describe various information, such information should not be limited to these terms. These terms are only used to distinguish the same type of information from each other.
- first information may also be referred to as the second information, and similarly, the second information may also be referred to as the first information.
- word “if” as used herein can be interpreted as “at” or "when” or "in response to determination”.
- a related method for determining the meso-scale factor in monocular vision reconstruction uses two adjacent frames of images collected by a monocular camera, and uses epipolar geometry to calculate the normalized translation vector of the monocular camera between the two frames of images; and uses code disk data and IMU (Inertial measurement unit (inertial measurement unit) data calculates the actual translation vector of the monocular camera in the real world between the two frames of images, and then uses the normalized translation vector and the actual translation vector to obtain the scale factor in monocular vision reconstruction.
- IMU Inertial measurement unit (inertial measurement unit) data
- the code disk count is inconsistent with the actual situation due to the mobile robot's slipping, jamming, etc., resulting in the actual calculation of the code disk data combined with the IMU data in this case.
- the translation vector is also inaccurate, and the scale factor calculated based on the actual translation vector is also inaccurate.
- the present application provides a method for determining a scale factor in a monocular vision reconstruction and a mobile robot, so as to solve the problem that the determined scale factor is inaccurate due to the slipping, jamming, etc. of the mobile robot in the existing method.
- the method provided by this embodiment can be applied to a mobile robot.
- it can be applied to a cleaning robot.
- FIG. 1 is a flowchart of Embodiment 1 of a method for determining a scale factor in monocular vision reconstruction provided by the present application.
- the method provided in this embodiment may include:
- the mobile robot is provided with a monocular camera, and images can be collected by the monocular camera.
- the designated object may be a charging device for charging the mobile robot.
- the mobile robot may obtain a first image of the designated object at the first moment through a monocular camera.
- the second image of the specified object at the second moment For example, there are neighboring sampling times: the first time t1 and the second time t2.
- the mobile robot can obtain the first image F1 of the specified object at the first time t1 and the second image of the specified object at the second time t2 through the monocular camera.
- Image F2 Image F2.
- the mobile robot is at different positions at the first time t1 and the second time t2, that is, the monocular camera is at different shooting positions at the first time t1 and the second time t2.
- Fig. 2 is a schematic diagram of a monocular camera acquiring an image of a specified object according to an exemplary embodiment of the present application. Please refer to FIG. 2.
- the designated object is a charging device for charging the mobile robot.
- the monocular camera 110 is at different shooting positions at a first time t1 and a second time t2.
- the mobile robot may turn to the charging device after detecting that the device is disconnected from the charging device 200, and then photograph the charging device 200 with a monocular camera at a position different from the previous one. .
- a first image of the charging device 200 at a first time t1 can be obtained, the first image corresponding to the first shooting position, and a second image of the charging device 200 at a second time t2, the second image corresponds to the second Shooting position.
- S102 Perform feature point extraction and matching on the first image and the second image, and calculate a normalized translation vector of the monocular camera from the first time to the second time according to the paired feature points.
- the pixel coordinates of the matched feature points in the first image and the second image may be used to calculate the monocular from the first time to the second time based on the epipolar constraint.
- the normalized translation vector of the camera between the first shooting position and the second shooting position may be calculated using eight pairs of paired feature points.
- the epipolar constraint can be expressed by the following formula:
- K is the internal parameter matrix of the monocular camera
- p 1 and p 2 are the pixel homogeneous coordinates of the paired feature points on the first image and the second image, respectively
- Rep is the monocular camera from the first time t1 to The rotation change amount at the second time t2
- t ep is the normalized translation vector of the monocular camera from the first time t1 to the second time t2.
- FIG. 3 is a flowchart of calculating a pose of a specified object relative to a monocular camera according to an exemplary embodiment of the present application.
- calculating the pose of the specified object relative to the monocular camera may include:
- the specified object may be identified from the image based on the attribute information of the specified object, and then based on the identified specified object, the pixel coordinates of the specified point on the specified object may be obtained from the image.
- attribute information of the designated object may include material attributes, color attributes, shape attributes, and the like. In this embodiment, this is not limited.
- the designated object may be a charging device for charging the mobile robot.
- the charging device is provided with a marker.
- the marker may be a marker composed of several marker blocks of a specific material, a specific color, a specific shape, a specified number, and / or a specified content.
- the marker may be a designated shape marker made of a specific material.
- the marker when the monocular camera is an infrared camera, the marker may be composed of a specified number of highly reflective material; for another example, when the monocular camera is an RGB camera, the marker may be a specified number of black and white printed Consisting of checkered checkered blocks.
- the specific setting form of the marker is not limited.
- the marker on the charging device can reflect the attribute information of the charging device, and the charging device in the image can be identified based on the marker of the charging device.
- the specific implementation principle and implementation process of identifying the specified object in the image based on the attribute information of the specified object refer to the description in the related technology, and details are not described herein again.
- the designated point on the designated object may be set according to actual needs, for example, the designated point may be a corner point, a center point, etc. of the marker.
- the specific position of the designated point is not limited. It should be noted that the number of the designated points is greater than or equal to four.
- the marker 210 on the charging device 200 is composed of four marker blocks 1, 2, 3, and 4, and designated points on the charging device 200 are designated as the marker blocks.
- the center point At this time, the four marker blocks 1, 2, 3, and 4 can be identified from the image based on attribute information such as the material, color, shape, and the distance between the marker blocks. 4. Furthermore, the pixel coordinates of the center point of each marker block are obtained. In this way, the pixel coordinates of the specified point on the specified object can be obtained.
- the center point of each marked block is sequentially recorded as Bi, where i is equal to 1 to 4.
- the pixel coordinates of the center point Bi of the i-th labeled block are labeled (u i , v i ).
- a distortion correction algorithm is used to calculate the first coordinates of each of the specified points after distortion correction.
- the distortion correction algorithm is expressed by the following formula:
- K is the internal parameter matrix of the monocular camera
- k 1 , k 2 , k 3 , p 1 , p 2 are distortion parameters of the monocular camera
- (x i , y i ) is the first coordinate after distortion correction of the i-th designated point.
- the designated coordinate system is an absolute coordinate system.
- the designated coordinate system is a coordinate system marked on the charging device. That is, in the example shown in FIG. 2, the origin of the designated coordinate system is the center point of the charging device, the X-axis is horizontal to the right, and the Y-axis is perpendicular to the X-axis downward.
- the specific implementation process of this step may include:
- the first formula is:
- the designated point includes a reference designated point and a target designated point
- A is the second coordinate of each target designated point in the designated coordinate system and the first coordinate of the reference designated point in the designated coordinate system.
- a matrix formed by the difference between two coordinates; the X is the X coordinate in the first coordinate after the distortion correction of the designated point of each target and the X coordinate in the first coordinate after the distortion of the reference designated point are corrected
- the Y is between the Y coordinate in the first coordinate after the distortion correction of the designated point of each target and the Y coordinate in the first coordinate after the distortion of the reference designated point are corrected A vector of differences.
- reference designated point may be any designated point.
- the reference designated point is used as an example for description.
- the second sitting mark of the i-th designated point in the designated coordinate system is (a i , b i , 0).
- each of the first vector i, the second vector j ′, the third vector k ′, and the first coefficient z includes three elements.
- the rotation matrix of the specified object relative to the monocular camera is denoted as R, and at this time, there are:
- the second formula is:
- the first rotation matrix R t1 and the first translation vector t t1 of the specified object with respect to the monocular camera at the first time t1 can be calculated.
- a second rotation matrix R t2 and a second translation vector t t2 of the designated object relative to the monocular camera at the second time t2 can be calculated.
- the first pose of the designated object relative to the monocular camera at the first time t1 is recorded as T t1 .
- T t2 the second pose of the designated object relative to the monocular camera at the second time t2 is recorded as T t2 .
- S104 Calculate an actual translation vector of the monocular camera from the first moment to the second moment according to the first pose and the second pose.
- the specific implementation process of this step may include: calculating the change in pose of the monocular camera from the first moment to the second moment according to the first pose and the second pose. Obtaining the true translation vector from the pose change.
- the pose change of the monocular camera from the first time t1 to the second time t2 can be calculated according to the following formula
- the pose change of the monocular camera from the first time t1 to the second time t2 includes an actual rotation matrix and an actual translation vector, and has:
- the actual rotation matrix and actual translation vector of the monocular camera from the first time t1 to the second time t2 can be obtained.
- the actual translation vector is a vector composed of the first three elements of the last column vector in the pose change.
- S105 Determine a ratio between a mode of the actual translation vector and a mode of the normalized translation vector as a scale factor in the monocular vision reconstruction of the device.
- the normalized translation vector of the monocular camera from the first time t1 to the second time t2 is calculated through step S102, and the monocular camera is calculated from the first time t1 to the second time t2 in the real world through step S104.
- the ratio between the modulus of the actual translation vector and the modulus of the normalized translation vector is determined as the scale factor in the monocular vision reconstruction of the device. which is:
- the map corresponding to the change in the pose and feature points of the monocular camera in the real world at two moments can be calculated, and then positioned at the same time in the subsequent
- the existing vision-based simultaneous positioning and map reconstruction algorithms can be used to calculate the change in pose and position of map points of subsequent monocular cameras in the real world by minimizing reprojection errors.
- a map at a real scale can be located and constructed.
- the calculated value of the designated object relative to the first time is calculated by The first pose and designated object of the monocular camera are relative to the second pose of the monocular camera at the second moment, and the monocular camera is calculated from the first moment to the second moment according to the first pose and the second pose.
- the actual amount of translation in the real world Therefore, the method provided by the present application can effectively avoid the problem that the determined scale factor is inaccurate due to slipping, jamming, and the like of the mobile robot.
- FIG. 4 is a hardware structural diagram of a first embodiment of a mobile robot provided in this application.
- the mobile robot 100 provided in this embodiment may include a monocular camera 410 and a processor 420. Among them,
- the monocular camera 410 is configured to acquire a first image of a designated object at a first moment and a second image of the designated object at a second moment;
- the processor 420 is configured to:
- the ratio of the modulus of the actual translation vector to the modulus of the normalized translation vector is determined as a scale factor in the monocular vision reconstruction of the device.
- the mobile robot of this embodiment may be used to execute the technical solution of the method embodiment shown in FIG. 1, and the implementation principles and technical effects thereof are similar, and details are not described herein again.
- processor 420 is specifically configured to:
- a posture of the designated object with respect to the monocular camera is obtained.
- processor 420 is specifically configured to:
- the actual translation vector is obtained from the pose change.
- the processor 420 is configured to identify the specified object from the frame image based on the attribute information of the specified object, and obtain a designation on the specified object based on the identified specified object.
- the pixel coordinates of the point is configured to identify the specified object from the frame image based on the attribute information of the specified object, and obtain a designation on the specified object based on the identified specified object.
- processor 420 is specifically configured to:
- the first formula is:
- the second formula is:
- the designated point includes a reference designated point and a target designated point, and A is a second coordinate of each target designated point in the designated coordinate system and a second coordinate of the reference designated point in the designated coordinate system.
- a matrix formed by the differences; the X is a vector formed by the difference between the X coordinate in the first coordinate after the distortion of the designated point of the target is corrected and the X coordinate in the first coordinate after the distortion of the reference designated point is corrected.
- the Y is a vector formed by the difference between the Y coordinate in the first coordinate after the distortion of the specified point of the target is corrected and the Y coordinate in the first coordinate after the distortion of the reference point is corrected;
- (a 1 , b 1 ) are the second coordinates of the reference designated point under the designated coordinates;
- (x 1 , y 1 ) are the first coordinates of the reference designated point after distortion correction; and
- i 1 and i 2 are the first and second elements in i, respectively;
- the j ' 1 and j' 2 are the first and second elements in j ', respectively Elements;
- the k ′ 1 and the k ′ 2 are the first element and the second element in the k ′, respectively;
- t is a translation vector of the specified object with respect to the monocular camera.
- the designated object is a charging device for charging the device; and the processor 420 is configured to obtain the designated object through a monocular camera after detecting that the device is disconnected from the designated object. A first image at a first moment and a second image of the designated object at a second moment.
- the present application also provides a computer-readable storage medium on which a computer program is stored, and when the program is executed by a processor, the steps of the method according to any one of the first aspect of the application are implemented.
- a computer-readable storage medium suitable for storing computer program instructions includes all forms of non-volatile memory, media, and memory devices, such as semiconductor memory devices (such as EPROM, EEPROM, and flash memory devices), magnetic disks (such as internal hard disks) Or removable disks), magneto-optical disks, and CD ROM and DVD-ROM disks.
- semiconductor memory devices such as EPROM, EEPROM, and flash memory devices
- magnetic disks such as internal hard disks) Or removable disks
- magneto-optical disks and CD ROM and DVD-ROM disks.
Abstract
Description
Claims (10)
- 一种单目视觉重建中尺度因子的确定方法,应用于移动机器人,所述方法包括:A method for determining a mesoscale factor in monocular vision reconstruction, which is applied to a mobile robot, the method includes:通过单目相机获取指定物体在第一时刻的第一图像以及所述指定物体在第二时刻的第二图像;Acquiring a first image of a designated object at a first moment and a second image of the designated object at a second moment through a monocular camera;对所述第一图像和所述第二图像进行特征点提取和匹配,并依据配对好的特征点计算所述单目相机从所述第一时刻到所述第二时刻的归一化平移向量;Performing feature point extraction and matching on the first image and the second image, and calculating a normalized translation vector of the monocular camera from the first moment to the second moment according to the paired feature points ;计算所述指定物体在所述第一时刻相对于所述单目相机的第一位姿,以及所述指定物体在所述第二时刻相对于所述单目相机的第二位姿;Calculating a first pose of the designated object relative to the monocular camera at the first moment, and a second pose of the designated object relative to the monocular camera at the second moment;依据所述第一位姿和所述第二位姿,计算所述单目相机从所述第一时刻到所述第二时刻的实际平移向量;Calculating an actual translation vector of the monocular camera from the first moment to the second moment according to the first pose and the second pose;将所述实际平移向量的模与所述归一化平移向量的模之间的比值确定为单目视觉重建中的尺度因子。A ratio between a mode of the actual translation vector and a mode of the normalized translation vector is determined as a scale factor in monocular vision reconstruction.
- 根据权利要求1所述的方法,其特征在于,计算所述指定物体相对于所述单目相机的位姿,包括:The method according to claim 1, wherein calculating the pose of the designated object relative to the monocular camera comprises:针对所述第一图像和所述第二图像中的每帧图像,从该帧图像中获取所述指定物体上的指定点的像素坐标;所述指定点的数量大于或者等于4;For each frame of the first image and the second image, obtain pixel coordinates of a specified point on the specified object from the frame image; the number of the specified points is greater than or equal to 4;依据各个所述指定点的像素坐标,采用畸变校正算法计算各个所述指定点畸变校正后的第一坐标;Calculating, according to the pixel coordinates of each of the specified points, a first coordinate after distortion correction of each of the specified points by using a distortion correction algorithm;依据各个所述指定点畸变校正后的第一坐标和预存的各个所述指定点在指定坐标系下的第二坐标,计算所述指定物体相对于所述单目相机的旋转矩阵和平移向量;Calculating a rotation matrix and a translation vector of the specified object relative to the monocular camera according to the first coordinates of each of the specified point distortion corrections and the pre-stored second coordinates of each of the specified points in the specified coordinate system;依据所述指定物体相对于所述单目相机的所述旋转矩阵和所述平移向量,得到所述指定物体相对于所述单目相机的位姿。Obtaining the pose of the specified object relative to the monocular camera according to the rotation matrix and the translation vector of the specified object relative to the monocular camera.
- 根据权利要求1所述的方法,其特征在于,依据所述第一位姿和所述第二位姿,计算所述单目相机从所述第一时刻到所述第二时刻的实际平移向量,包括:The method according to claim 1, wherein an actual translation vector of the monocular camera from the first moment to the second moment is calculated according to the first pose and the second pose. ,include:依据所述第一位姿和所述第二位姿,计算所述单目相机从所述第一时刻到所述第二时刻的位姿变化;Calculating a change in pose of the monocular camera from the first moment to the second moment according to the first pose and the second pose;从所述位姿变化中获取所述实际平移向量。The actual translation vector is obtained from the pose change.
- 根据权利要求2所述的方法,其特征在于,从该帧图像中获取所述指定物体上的指定点的像素坐标,包括:The method according to claim 2, wherein obtaining pixel coordinates of a specified point on the specified object from the frame image comprises:基于所述指定物体的属性信息,从该帧图像中识别出所述指定物体;Identifying the designated object from the frame image based on the attribute information of the designated object;基于识别出的所述指定物体,获取所述指定物体上的指定点的像素坐标。Based on the identified specified object, pixel coordinates of a specified point on the specified object are obtained.
- 根据权利要求2所述的方法,其特征在于,依据各个所述指定点畸变校正后的第一坐标和预存的各个所述指定点在指定坐标系下的第二坐标,计算所述指定物体相对于所述单目相机的旋转矩阵和平移向量,包括:The method according to claim 2, characterized in that the relative of the specified object is calculated according to the first coordinates of each of the specified point distortion corrections and the second coordinates of each of the specified points stored in the specified coordinate system in advance. The rotation matrix and translation vector for the monocular camera include:依据各个所述指定点畸变校正后的第一坐标和预存的各个所述指定点在指定坐标系下的第二坐标,按照第一公式计算第一向量i、第二向量j′、第三向量k'和第一系数z;Calculate the first vector i, the second vector j ′, and the third vector according to the first formula according to the first coordinates of each of the specified point distortion corrections and the second coordinates of each of the specified points stored in the specified coordinate system. k 'and the first coefficient z;将所述i、所述j′和所述k'按矩阵行方向依序排列,得到所述指定物体相对于所述单目相机的旋转矩阵;Arranging i, j ′ and k ′ in the row direction of the matrix in order to obtain a rotation matrix of the specified object relative to the monocular camera;依据所述i、所述j′、所述k'和所述z,按照第二公式计算所述指定物体相对于所述单目相机的平移向量;Calculating a translation vector of the specified object with respect to the monocular camera according to the i, the j ′, the k ′, and the z according to a second formula;其中,所述第一公式为:The first formula is:所述第二公式为:The second formula is:其中,所述指定点包括参考指定点和目标指定点,所述A为各个所述目标指定点在所述指定坐标系下的第二坐标与所述参考指定点在所述指定坐标系下的第二坐标的差值构成的矩阵;所述X为各个所述目标指定点畸变校正后的第一坐标中的X坐标与所述参考指定点畸变校正后的第一坐标中的X坐标的差值构成的向量;所述Y为各个所述目标指定点畸变校正后的第一坐标中的Y坐标与所述参考指定点畸变校正后的第一坐标中的Y坐标的差值构成的向量;Wherein, the designated point includes a reference designated point and a target designated point, and the A is a second coordinate of each of the target designated points in the designated coordinate system and a reference coordinate of the reference designated point in the designated coordinate system. A matrix formed by the difference of the second coordinates; the X is the difference between the X coordinate in the first coordinate after the distortion correction of the designated point of the target and the X coordinate in the first coordinate after the distortion of the reference designated point are corrected A vector formed by values; the Y is a vector formed by a difference between the Y coordinate in the first coordinate after the distortion correction of the designated point of the target and the Y coordinate in the first coordinate after the distortion of the reference designated point is corrected;其中,(a 1,b 1)为所述参考指定点在所述指定坐标下的第二坐标;(x 1,y 1)为所述参考指定点畸变校正后的第一坐标;所述i 1和所述i 2分别为所述i中的第一个元素和第 二个元素;所述j′ 1和所述j′ 2分别为所述j'中的第一个元素和第二个元素;所述k′ 1和所述k′ 2分别为所述k'中的第一个元素和第二个元素;所述t为所述指定物体相对于所述单目相机的平移向量。 Where (a 1 , b 1 ) are the second coordinates of the reference designated point under the designated coordinates; (x 1 , y 1 ) are the first coordinates of the reference designated point after distortion correction; and i 1 and i 2 are the first and second elements in i, respectively; the j ′ 1 and j ′ 2 are the first and second elements in j ′, respectively Element k ′ 1 and k ′ 2 are the first element and the second element of k ′, respectively; and t is a translation vector of the specified object with respect to the monocular camera.
- 根据权利要求1所述的方法,其特征在于,所述指定物体为用于给本设备充电的充电设备;通过单目相机获取指定物体在第一时刻的第一图像以及所述指定物体在第二时刻的第二图像,包括:The method according to claim 1, wherein the designated object is a charging device used to charge the device; a first image of the designated object at a first moment is obtained through a monocular camera, and the designated object is at the first time. The second image at two moments includes:在检测到本设备与所述指定物体断开连接后,通过所述单目相机获取所述指定物体在第一时刻的第一图像以及所述指定物体在第二时刻的第二图像。After detecting that the device is disconnected from the designated object, the first image of the designated object at the first moment and the second image of the designated object at the second moment are obtained through the monocular camera.
- 一种移动机器人,包括:A mobile robot includes:单目相机,用于获取指定物体在第一时刻的第一图像以及所述指定物体在第二时刻的第二图像;A monocular camera, configured to obtain a first image of a specified object at a first moment and a second image of the specified object at a second moment;处理器,用于:Processors for:对所述第一图像和所述第二图像进行特征点提取和匹配,并依据配对好的特征点计算所述单目相机从所述第一时刻到所述第二时刻的归一化平移向量;Performing feature point extraction and matching on the first image and the second image, and calculating a normalized translation vector of the monocular camera from the first moment to the second moment according to the paired feature points ;计算所述指定物体在所述第一时刻相对于所述单目相机的第一位姿,以及所述指定物体在所述第二时刻相对于所述单目相机的第二位姿;Calculating a first pose of the designated object relative to the monocular camera at the first moment, and a second pose of the designated object relative to the monocular camera at the second moment;依据所述第一位姿和所述第二位姿,计算所述单目相机从所述第一时刻到所述第二时刻的实际平移向量;Calculating an actual translation vector of the monocular camera from the first moment to the second moment according to the first pose and the second pose;将所述实际平移向量的模与所述归一化平移向量的模的比值确定为单目视觉重建中的尺度因子。A ratio of a modulus of the actual translation vector to a modulus of the normalized translation vector is determined as a scale factor in monocular vision reconstruction.
- 根据权利要求7所述的移动机器人,其特征在于,所述处理器,用于:The mobile robot according to claim 7, wherein the processor is configured to:针对所述第一图像和所述第二图像中的每帧图像,从该帧图像中获取所述指定物体上的指定点的像素坐标;所述指定点的数量大于或者等于4;For each frame of the first image and the second image, obtain pixel coordinates of a specified point on the specified object from the frame image; the number of the specified points is greater than or equal to 4;依据各个所述指定点的像素坐标,采用畸变校正算法计算各个所述指定点畸变校正后的第一坐标;Calculating, according to the pixel coordinates of each of the specified points, a first coordinate after distortion correction of each of the specified points by using a distortion correction algorithm;依据各个所述指定点畸变校正后的第一坐标和预存的各个所述指定点在指定坐标系下的第二坐标,计算所述指定物体相对于所述单目相机的旋转矩阵和平移向量;Calculating a rotation matrix and a translation vector of the specified object relative to the monocular camera according to the first coordinates of each of the specified point distortion corrections and the pre-stored second coordinates of each of the specified points in the specified coordinate system;依据所述指定物体相对于所述单目相机的所述旋转矩阵和所述平移向量,得到所述指定物体相对于所述单目相机的位姿。Obtaining the pose of the specified object relative to the monocular camera according to the rotation matrix and the translation vector of the specified object relative to the monocular camera.
- 根据权利要求7所述的移动机器人,其特征在于,所述处理器,具体用于:The mobile robot according to claim 7, wherein the processor is specifically configured to:依据所述第一位姿和所述第二位姿,计算所述单目相机从所述第一时刻到所述第二时刻的位姿变化;Calculating a change in pose of the monocular camera from the first moment to the second moment according to the first pose and the second pose;从所述位姿变化中获取所述实际平移向量。The actual translation vector is obtained from the pose change.
- 一种计算机可读存储介质,其上存储有计算机程序,其特征在于,所述程序被处理器执行实现权利要求1-6任一项所述方法的步骤。A computer-readable storage medium having stored thereon a computer program, characterized in that the program is executed by a processor to implement the steps of the method according to any one of claims 1-6.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810961434.6 | 2018-08-22 | ||
CN201810961434.6A CN110858403B (en) | 2018-08-22 | 2018-08-22 | Method for determining scale factor in monocular vision reconstruction and mobile robot |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2020038386A1 true WO2020038386A1 (en) | 2020-02-27 |
Family
ID=69593088
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2019/101704 WO2020038386A1 (en) | 2018-08-22 | 2019-08-21 | Determination of scale factor in monocular vision-based reconstruction |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN110858403B (en) |
WO (1) | WO2020038386A1 (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111260538A (en) * | 2018-12-03 | 2020-06-09 | 北京初速度科技有限公司 | Positioning and vehicle-mounted terminal based on long-baseline binocular fisheye camera |
CN112102406A (en) * | 2020-09-09 | 2020-12-18 | 东软睿驰汽车技术(沈阳)有限公司 | Monocular vision scale correction method and device and delivery vehicle |
CN112686950A (en) * | 2020-12-04 | 2021-04-20 | 深圳市优必选科技股份有限公司 | Pose estimation method and device, terminal equipment and computer readable storage medium |
CN114406985A (en) * | 2021-10-18 | 2022-04-29 | 苏州迪凯尔医疗科技有限公司 | Target tracking mechanical arm method, system, equipment and storage medium |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113554703B (en) * | 2020-04-23 | 2024-03-01 | 北京京东乾石科技有限公司 | Robot positioning method, apparatus, system and computer readable storage medium |
CN111671360B (en) * | 2020-05-26 | 2021-11-16 | 深圳拓邦股份有限公司 | Sweeping robot position calculating method and device and sweeping robot |
CN112798812B (en) * | 2020-12-30 | 2023-09-26 | 中山联合汽车技术有限公司 | Target speed measuring method based on monocular vision |
CN113126117B (en) * | 2021-04-15 | 2021-08-27 | 湖北亿咖通科技有限公司 | Method for determining absolute scale of SFM map and electronic equipment |
CN116704047B (en) * | 2023-08-01 | 2023-10-27 | 安徽云森物联网科技有限公司 | Pedestrian ReID-based calibration method for monitoring camera equipment position |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101706957A (en) * | 2009-10-30 | 2010-05-12 | 无锡景象数字技术有限公司 | Self-calibration method for binocular stereo vision device |
WO2017114507A1 (en) * | 2015-12-31 | 2017-07-06 | 清华大学 | Method and device for image positioning based on ray model three-dimensional reconstruction |
CN108010125A (en) * | 2017-12-28 | 2018-05-08 | 中国科学院西安光学精密机械研究所 | True scale three-dimensional reconstruction system and method based on line-structured light and image information |
CN108090435A (en) * | 2017-12-13 | 2018-05-29 | 深圳市航盛电子股份有限公司 | One kind can parking area recognition methods, system and medium |
Family Cites Families (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103234454B (en) * | 2013-04-23 | 2016-03-30 | 合肥米克光电技术有限公司 | A kind of self-calibrating method of image measurer |
CN103278138B (en) * | 2013-05-03 | 2015-05-06 | 中国科学院自动化研究所 | Method for measuring three-dimensional position and posture of thin component with complex structure |
CN104346829A (en) * | 2013-07-29 | 2015-02-11 | 中国农业机械化科学研究院 | Three-dimensional color reconstruction system and method based on PMD (photonic mixer device) cameras and photographing head |
CN104732518B (en) * | 2015-01-19 | 2017-09-01 | 北京工业大学 | A kind of PTAM improved methods based on intelligent robot terrain surface specifications |
CN105118055B (en) * | 2015-08-11 | 2017-12-15 | 北京电影学院 | Camera position amendment scaling method and system |
CN105931222B (en) * | 2016-04-13 | 2018-11-02 | 成都信息工程大学 | The method for realizing high-precision camera calibration with low precision two dimensional surface target |
CN106529538A (en) * | 2016-11-24 | 2017-03-22 | 腾讯科技(深圳)有限公司 | Method and device for positioning aircraft |
CN106920259B (en) * | 2017-02-28 | 2019-12-06 | 武汉工程大学 | positioning method and system |
-
2018
- 2018-08-22 CN CN201810961434.6A patent/CN110858403B/en active Active
-
2019
- 2019-08-21 WO PCT/CN2019/101704 patent/WO2020038386A1/en active Application Filing
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101706957A (en) * | 2009-10-30 | 2010-05-12 | 无锡景象数字技术有限公司 | Self-calibration method for binocular stereo vision device |
WO2017114507A1 (en) * | 2015-12-31 | 2017-07-06 | 清华大学 | Method and device for image positioning based on ray model three-dimensional reconstruction |
CN108090435A (en) * | 2017-12-13 | 2018-05-29 | 深圳市航盛电子股份有限公司 | One kind can parking area recognition methods, system and medium |
CN108010125A (en) * | 2017-12-28 | 2018-05-08 | 中国科学院西安光学精密机械研究所 | True scale three-dimensional reconstruction system and method based on line-structured light and image information |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111260538A (en) * | 2018-12-03 | 2020-06-09 | 北京初速度科技有限公司 | Positioning and vehicle-mounted terminal based on long-baseline binocular fisheye camera |
CN111260538B (en) * | 2018-12-03 | 2023-10-03 | 北京魔门塔科技有限公司 | Positioning and vehicle-mounted terminal based on long-baseline binocular fisheye camera |
CN112102406A (en) * | 2020-09-09 | 2020-12-18 | 东软睿驰汽车技术(沈阳)有限公司 | Monocular vision scale correction method and device and delivery vehicle |
CN112686950A (en) * | 2020-12-04 | 2021-04-20 | 深圳市优必选科技股份有限公司 | Pose estimation method and device, terminal equipment and computer readable storage medium |
CN112686950B (en) * | 2020-12-04 | 2023-12-15 | 深圳市优必选科技股份有限公司 | Pose estimation method, pose estimation device, terminal equipment and computer readable storage medium |
CN114406985A (en) * | 2021-10-18 | 2022-04-29 | 苏州迪凯尔医疗科技有限公司 | Target tracking mechanical arm method, system, equipment and storage medium |
CN114406985B (en) * | 2021-10-18 | 2024-04-12 | 苏州迪凯尔医疗科技有限公司 | Mechanical arm method, system, equipment and storage medium for target tracking |
Also Published As
Publication number | Publication date |
---|---|
CN110858403B (en) | 2022-09-27 |
CN110858403A (en) | 2020-03-03 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2020038386A1 (en) | Determination of scale factor in monocular vision-based reconstruction | |
US9420265B2 (en) | Tracking poses of 3D camera using points and planes | |
JP5832341B2 (en) | Movie processing apparatus, movie processing method, and movie processing program | |
US10636151B2 (en) | Method for estimating the speed of movement of a camera | |
WO2018076154A1 (en) | Spatial positioning calibration of fisheye camera-based panoramic video generating method | |
WO2020063708A1 (en) | Method, device and system for calibrating intrinsic parameters of fisheye camera, calibration device controller and calibration tool | |
CN112767542A (en) | Three-dimensional reconstruction method of multi-view camera, VR camera and panoramic camera | |
WO2021004416A1 (en) | Method and apparatus for establishing beacon map on basis of visual beacons | |
US11082633B2 (en) | Method of estimating the speed of displacement of a camera | |
CN108776976B (en) | Method, system and storage medium for simultaneously positioning and establishing image | |
CN106530358A (en) | Method for calibrating PTZ camera by using only two scene images | |
CN106228538A (en) | Binocular vision indoor orientation method based on logo | |
CN110490943B (en) | Rapid and accurate calibration method and system of 4D holographic capture system and storage medium | |
CN111062966B (en) | Method for optimizing camera tracking based on L-M algorithm and polynomial interpolation | |
WO2020024684A1 (en) | Method and device for modeling three-dimensional scene, electronic device, readable storage medium, and computer apparatus | |
JP4109075B2 (en) | Method for measuring the rotation characteristics and flight characteristics of a sphere, and a device for measuring the rotation characteristics and flight characteristics of a sphere | |
CN110567441A (en) | Particle filter-based positioning method, positioning device, mapping and positioning method | |
JP6922348B2 (en) | Information processing equipment, methods, and programs | |
KR20190130407A (en) | Apparatus and method for omni-directional camera calibration | |
JP4109076B2 (en) | Measuring method of rotation amount and rotation axis direction of curved surface body, and measurement device of rotation amount and rotation axis direction of curved surface body | |
CN113034347B (en) | Oblique photography image processing method, device, processing equipment and storage medium | |
JP2002109518A (en) | Three-dimensional shape restoring method and system therefor | |
JP4886661B2 (en) | Camera parameter estimation apparatus and camera parameter estimation program | |
CN116152121A (en) | Curved surface screen generating method and correcting method based on distortion parameters | |
WO2018072087A1 (en) | Method for realizing effect of photo being taken by others through selfie, and photographing device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 19852178 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19852178 Country of ref document: EP Kind code of ref document: A1 |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19852178 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 15.09.2021) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19852178 Country of ref document: EP Kind code of ref document: A1 |