WO2020038386A1 - Determination of scale factor in monocular vision-based reconstruction - Google Patents

Determination of scale factor in monocular vision-based reconstruction Download PDF

Info

Publication number
WO2020038386A1
WO2020038386A1 PCT/CN2019/101704 CN2019101704W WO2020038386A1 WO 2020038386 A1 WO2020038386 A1 WO 2020038386A1 CN 2019101704 W CN2019101704 W CN 2019101704W WO 2020038386 A1 WO2020038386 A1 WO 2020038386A1
Authority
WO
WIPO (PCT)
Prior art keywords
specified
monocular camera
moment
designated
pose
Prior art date
Application number
PCT/CN2019/101704
Other languages
French (fr)
Chinese (zh)
Inventor
沈冰伟
朱建华
蒋腻聪
郭斌
Original Assignee
杭州萤石软件有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 杭州萤石软件有限公司 filed Critical 杭州萤石软件有限公司
Publication of WO2020038386A1 publication Critical patent/WO2020038386A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/50Depth or shape recovery
    • G06T7/55Depth or shape recovery from multiple images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30248Vehicle exterior or interior
    • G06T2207/30252Vehicle exterior; Vicinity of vehicle

Definitions

  • the present application relates to the field of mobile robot technology, and in particular, to a method for determining a mesoscale factor for monocular vision reconstruction and a mobile robot.
  • the simultaneous positioning and map construction algorithms based on monocular vision have become the focus of current mobile robot research.
  • the traditional simultaneous positioning and map construction methods based on monocular vision can only achieve 3D reconstruction at projective scale or affine-scaling, that is, there is a scale factor between the reconstructed scene and the real-world scene.
  • the scale factor is the ratio of the real world map scale to the constructed map scale. Therefore, if the scale factor can be determined when the mobile robot is initialized, the actual rotation and translation of the monocular camera in the real world can be calculated based on the projection model, and a map with the same scale as the real world can be constructed.
  • the present application provides a method for determining a mesoscale factor for monocular vision reconstruction and a mobile robot.
  • a first aspect of the present application provides a method for determining a mesoscale factor in monocular vision reconstruction.
  • the method is applied to a mobile robot, and the method includes:
  • the ratio of the modulus of the actual translation vector to the modulus of the normalized translation vector is determined as a scale factor in the monocular vision reconstruction of the device.
  • a second aspect of the present application provides a mobile robot, which includes a monocular camera and a processor; wherein,
  • the monocular camera is configured to acquire a first image of a designated object at a first moment and a second image of the designated object at a second moment;
  • the processor is configured to:
  • the ratio of the modulus of the actual translation vector to the modulus of the normalized translation vector is determined as a scale factor in the monocular vision reconstruction of the device.
  • a third aspect of the present application provides a computer-readable storage medium on which a computer program is stored, and the program is executed by a processor to implement the steps of any of the methods provided in the first aspect of the present application.
  • the method for determining the meso-scale factor of the monocular vision reconstruction and the mobile robot provided in this application, because the position of the designated object is fixed, therefore, in the case of the mobile robot slipping, jamming, etc., by calculating the designated object at the first moment The first pose relative to the monocular camera and the second pose of the designated object relative to the monocular camera at the second moment, so that the real-time The actual translation vector. Therefore, the method provided in this application does not have the problem that the determined scale factor is inaccurate due to slipping, jamming, and the like of the mobile robot.
  • FIG. 1 is a flowchart of Embodiment 1 of a method for determining a scale factor in monocular vision reconstruction provided by the present application.
  • Fig. 2 is a schematic diagram of a monocular camera acquiring an image of a specified object according to an exemplary embodiment of the present application.
  • Fig. 3 is a flowchart of calculating a pose of a specified object relative to a monocular camera according to an exemplary embodiment of the present application.
  • FIG. 4 is a hardware structural diagram of a first embodiment of a mobile robot provided in this application.
  • first, second, third, etc. may be used in this application to describe various information, such information should not be limited to these terms. These terms are only used to distinguish the same type of information from each other.
  • first information may also be referred to as the second information, and similarly, the second information may also be referred to as the first information.
  • word “if” as used herein can be interpreted as “at” or "when” or "in response to determination”.
  • a related method for determining the meso-scale factor in monocular vision reconstruction uses two adjacent frames of images collected by a monocular camera, and uses epipolar geometry to calculate the normalized translation vector of the monocular camera between the two frames of images; and uses code disk data and IMU (Inertial measurement unit (inertial measurement unit) data calculates the actual translation vector of the monocular camera in the real world between the two frames of images, and then uses the normalized translation vector and the actual translation vector to obtain the scale factor in monocular vision reconstruction.
  • IMU Inertial measurement unit (inertial measurement unit) data
  • the code disk count is inconsistent with the actual situation due to the mobile robot's slipping, jamming, etc., resulting in the actual calculation of the code disk data combined with the IMU data in this case.
  • the translation vector is also inaccurate, and the scale factor calculated based on the actual translation vector is also inaccurate.
  • the present application provides a method for determining a scale factor in a monocular vision reconstruction and a mobile robot, so as to solve the problem that the determined scale factor is inaccurate due to the slipping, jamming, etc. of the mobile robot in the existing method.
  • the method provided by this embodiment can be applied to a mobile robot.
  • it can be applied to a cleaning robot.
  • FIG. 1 is a flowchart of Embodiment 1 of a method for determining a scale factor in monocular vision reconstruction provided by the present application.
  • the method provided in this embodiment may include:
  • the mobile robot is provided with a monocular camera, and images can be collected by the monocular camera.
  • the designated object may be a charging device for charging the mobile robot.
  • the mobile robot may obtain a first image of the designated object at the first moment through a monocular camera.
  • the second image of the specified object at the second moment For example, there are neighboring sampling times: the first time t1 and the second time t2.
  • the mobile robot can obtain the first image F1 of the specified object at the first time t1 and the second image of the specified object at the second time t2 through the monocular camera.
  • Image F2 Image F2.
  • the mobile robot is at different positions at the first time t1 and the second time t2, that is, the monocular camera is at different shooting positions at the first time t1 and the second time t2.
  • Fig. 2 is a schematic diagram of a monocular camera acquiring an image of a specified object according to an exemplary embodiment of the present application. Please refer to FIG. 2.
  • the designated object is a charging device for charging the mobile robot.
  • the monocular camera 110 is at different shooting positions at a first time t1 and a second time t2.
  • the mobile robot may turn to the charging device after detecting that the device is disconnected from the charging device 200, and then photograph the charging device 200 with a monocular camera at a position different from the previous one. .
  • a first image of the charging device 200 at a first time t1 can be obtained, the first image corresponding to the first shooting position, and a second image of the charging device 200 at a second time t2, the second image corresponds to the second Shooting position.
  • S102 Perform feature point extraction and matching on the first image and the second image, and calculate a normalized translation vector of the monocular camera from the first time to the second time according to the paired feature points.
  • the pixel coordinates of the matched feature points in the first image and the second image may be used to calculate the monocular from the first time to the second time based on the epipolar constraint.
  • the normalized translation vector of the camera between the first shooting position and the second shooting position may be calculated using eight pairs of paired feature points.
  • the epipolar constraint can be expressed by the following formula:
  • K is the internal parameter matrix of the monocular camera
  • p 1 and p 2 are the pixel homogeneous coordinates of the paired feature points on the first image and the second image, respectively
  • Rep is the monocular camera from the first time t1 to The rotation change amount at the second time t2
  • t ep is the normalized translation vector of the monocular camera from the first time t1 to the second time t2.
  • FIG. 3 is a flowchart of calculating a pose of a specified object relative to a monocular camera according to an exemplary embodiment of the present application.
  • calculating the pose of the specified object relative to the monocular camera may include:
  • the specified object may be identified from the image based on the attribute information of the specified object, and then based on the identified specified object, the pixel coordinates of the specified point on the specified object may be obtained from the image.
  • attribute information of the designated object may include material attributes, color attributes, shape attributes, and the like. In this embodiment, this is not limited.
  • the designated object may be a charging device for charging the mobile robot.
  • the charging device is provided with a marker.
  • the marker may be a marker composed of several marker blocks of a specific material, a specific color, a specific shape, a specified number, and / or a specified content.
  • the marker may be a designated shape marker made of a specific material.
  • the marker when the monocular camera is an infrared camera, the marker may be composed of a specified number of highly reflective material; for another example, when the monocular camera is an RGB camera, the marker may be a specified number of black and white printed Consisting of checkered checkered blocks.
  • the specific setting form of the marker is not limited.
  • the marker on the charging device can reflect the attribute information of the charging device, and the charging device in the image can be identified based on the marker of the charging device.
  • the specific implementation principle and implementation process of identifying the specified object in the image based on the attribute information of the specified object refer to the description in the related technology, and details are not described herein again.
  • the designated point on the designated object may be set according to actual needs, for example, the designated point may be a corner point, a center point, etc. of the marker.
  • the specific position of the designated point is not limited. It should be noted that the number of the designated points is greater than or equal to four.
  • the marker 210 on the charging device 200 is composed of four marker blocks 1, 2, 3, and 4, and designated points on the charging device 200 are designated as the marker blocks.
  • the center point At this time, the four marker blocks 1, 2, 3, and 4 can be identified from the image based on attribute information such as the material, color, shape, and the distance between the marker blocks. 4. Furthermore, the pixel coordinates of the center point of each marker block are obtained. In this way, the pixel coordinates of the specified point on the specified object can be obtained.
  • the center point of each marked block is sequentially recorded as Bi, where i is equal to 1 to 4.
  • the pixel coordinates of the center point Bi of the i-th labeled block are labeled (u i , v i ).
  • a distortion correction algorithm is used to calculate the first coordinates of each of the specified points after distortion correction.
  • the distortion correction algorithm is expressed by the following formula:
  • K is the internal parameter matrix of the monocular camera
  • k 1 , k 2 , k 3 , p 1 , p 2 are distortion parameters of the monocular camera
  • (x i , y i ) is the first coordinate after distortion correction of the i-th designated point.
  • the designated coordinate system is an absolute coordinate system.
  • the designated coordinate system is a coordinate system marked on the charging device. That is, in the example shown in FIG. 2, the origin of the designated coordinate system is the center point of the charging device, the X-axis is horizontal to the right, and the Y-axis is perpendicular to the X-axis downward.
  • the specific implementation process of this step may include:
  • the first formula is:
  • the designated point includes a reference designated point and a target designated point
  • A is the second coordinate of each target designated point in the designated coordinate system and the first coordinate of the reference designated point in the designated coordinate system.
  • a matrix formed by the difference between two coordinates; the X is the X coordinate in the first coordinate after the distortion correction of the designated point of each target and the X coordinate in the first coordinate after the distortion of the reference designated point are corrected
  • the Y is between the Y coordinate in the first coordinate after the distortion correction of the designated point of each target and the Y coordinate in the first coordinate after the distortion of the reference designated point are corrected A vector of differences.
  • reference designated point may be any designated point.
  • the reference designated point is used as an example for description.
  • the second sitting mark of the i-th designated point in the designated coordinate system is (a i , b i , 0).
  • each of the first vector i, the second vector j ′, the third vector k ′, and the first coefficient z includes three elements.
  • the rotation matrix of the specified object relative to the monocular camera is denoted as R, and at this time, there are:
  • the second formula is:
  • the first rotation matrix R t1 and the first translation vector t t1 of the specified object with respect to the monocular camera at the first time t1 can be calculated.
  • a second rotation matrix R t2 and a second translation vector t t2 of the designated object relative to the monocular camera at the second time t2 can be calculated.
  • the first pose of the designated object relative to the monocular camera at the first time t1 is recorded as T t1 .
  • T t2 the second pose of the designated object relative to the monocular camera at the second time t2 is recorded as T t2 .
  • S104 Calculate an actual translation vector of the monocular camera from the first moment to the second moment according to the first pose and the second pose.
  • the specific implementation process of this step may include: calculating the change in pose of the monocular camera from the first moment to the second moment according to the first pose and the second pose. Obtaining the true translation vector from the pose change.
  • the pose change of the monocular camera from the first time t1 to the second time t2 can be calculated according to the following formula
  • the pose change of the monocular camera from the first time t1 to the second time t2 includes an actual rotation matrix and an actual translation vector, and has:
  • the actual rotation matrix and actual translation vector of the monocular camera from the first time t1 to the second time t2 can be obtained.
  • the actual translation vector is a vector composed of the first three elements of the last column vector in the pose change.
  • S105 Determine a ratio between a mode of the actual translation vector and a mode of the normalized translation vector as a scale factor in the monocular vision reconstruction of the device.
  • the normalized translation vector of the monocular camera from the first time t1 to the second time t2 is calculated through step S102, and the monocular camera is calculated from the first time t1 to the second time t2 in the real world through step S104.
  • the ratio between the modulus of the actual translation vector and the modulus of the normalized translation vector is determined as the scale factor in the monocular vision reconstruction of the device. which is:
  • the map corresponding to the change in the pose and feature points of the monocular camera in the real world at two moments can be calculated, and then positioned at the same time in the subsequent
  • the existing vision-based simultaneous positioning and map reconstruction algorithms can be used to calculate the change in pose and position of map points of subsequent monocular cameras in the real world by minimizing reprojection errors.
  • a map at a real scale can be located and constructed.
  • the calculated value of the designated object relative to the first time is calculated by The first pose and designated object of the monocular camera are relative to the second pose of the monocular camera at the second moment, and the monocular camera is calculated from the first moment to the second moment according to the first pose and the second pose.
  • the actual amount of translation in the real world Therefore, the method provided by the present application can effectively avoid the problem that the determined scale factor is inaccurate due to slipping, jamming, and the like of the mobile robot.
  • FIG. 4 is a hardware structural diagram of a first embodiment of a mobile robot provided in this application.
  • the mobile robot 100 provided in this embodiment may include a monocular camera 410 and a processor 420. Among them,
  • the monocular camera 410 is configured to acquire a first image of a designated object at a first moment and a second image of the designated object at a second moment;
  • the processor 420 is configured to:
  • the ratio of the modulus of the actual translation vector to the modulus of the normalized translation vector is determined as a scale factor in the monocular vision reconstruction of the device.
  • the mobile robot of this embodiment may be used to execute the technical solution of the method embodiment shown in FIG. 1, and the implementation principles and technical effects thereof are similar, and details are not described herein again.
  • processor 420 is specifically configured to:
  • a posture of the designated object with respect to the monocular camera is obtained.
  • processor 420 is specifically configured to:
  • the actual translation vector is obtained from the pose change.
  • the processor 420 is configured to identify the specified object from the frame image based on the attribute information of the specified object, and obtain a designation on the specified object based on the identified specified object.
  • the pixel coordinates of the point is configured to identify the specified object from the frame image based on the attribute information of the specified object, and obtain a designation on the specified object based on the identified specified object.
  • processor 420 is specifically configured to:
  • the first formula is:
  • the second formula is:
  • the designated point includes a reference designated point and a target designated point, and A is a second coordinate of each target designated point in the designated coordinate system and a second coordinate of the reference designated point in the designated coordinate system.
  • a matrix formed by the differences; the X is a vector formed by the difference between the X coordinate in the first coordinate after the distortion of the designated point of the target is corrected and the X coordinate in the first coordinate after the distortion of the reference designated point is corrected.
  • the Y is a vector formed by the difference between the Y coordinate in the first coordinate after the distortion of the specified point of the target is corrected and the Y coordinate in the first coordinate after the distortion of the reference point is corrected;
  • (a 1 , b 1 ) are the second coordinates of the reference designated point under the designated coordinates;
  • (x 1 , y 1 ) are the first coordinates of the reference designated point after distortion correction; and
  • i 1 and i 2 are the first and second elements in i, respectively;
  • the j ' 1 and j' 2 are the first and second elements in j ', respectively Elements;
  • the k ′ 1 and the k ′ 2 are the first element and the second element in the k ′, respectively;
  • t is a translation vector of the specified object with respect to the monocular camera.
  • the designated object is a charging device for charging the device; and the processor 420 is configured to obtain the designated object through a monocular camera after detecting that the device is disconnected from the designated object. A first image at a first moment and a second image of the designated object at a second moment.
  • the present application also provides a computer-readable storage medium on which a computer program is stored, and when the program is executed by a processor, the steps of the method according to any one of the first aspect of the application are implemented.
  • a computer-readable storage medium suitable for storing computer program instructions includes all forms of non-volatile memory, media, and memory devices, such as semiconductor memory devices (such as EPROM, EEPROM, and flash memory devices), magnetic disks (such as internal hard disks) Or removable disks), magneto-optical disks, and CD ROM and DVD-ROM disks.
  • semiconductor memory devices such as EPROM, EEPROM, and flash memory devices
  • magnetic disks such as internal hard disks) Or removable disks
  • magneto-optical disks and CD ROM and DVD-ROM disks.

Abstract

The present application provides a method for determining a scale factor in monocular vision-based reconstruction, the method comprising: acquiring, by means of a monocular camera, a first image of a specified object at a first time point and a second image thereof at a second time point; performing feature point extraction and matching on the first image and the second image, and calculating, according to paired feature points, a normalized translation vector of the monocular camera from the first time point to the second time point; calculating a first pose of the specified object relative to the monocular camera at the first time point and a second pose thereof relative to the monocular camera at the second time point; calculating, according to the first pose and the second pose, an actual translation vector of the monocular camera in the physical world from the first time point to the second time point; and determining a ratio of a norm of the actual translation vector to a norm of the normalized translation vector to be a scale factor in monocular vision-based reconstruction.

Description

确定单目视觉重建中的尺度因子Determining scale factors in monocular vision reconstruction
相关申请的交叉引用Cross-reference to related applications
本专利申请要求于2018年8月22日提交的、申请号为2018109614346、发明名称为“一种单目视觉重建中尺度因子的确定方法和移动机器人”的中国专利申请的优先权,该申请的全文以引用的方式并入本文中。This patent application claims priority from a Chinese patent application filed on August 22, 2018, with an application number of 2018109614346, and the invention name is "A Method for Determining Mesoscale Factors in Monocular Visual Reconstruction and a Mobile Robot" The entire text is incorporated herein by reference.
技术领域Technical field
本申请涉及移动机器人技术领域,尤其涉及一种单目视觉重建中尺度因子的确定方法和移动机器人。The present application relates to the field of mobile robot technology, and in particular, to a method for determining a mesoscale factor for monocular vision reconstruction and a mobile robot.
背景技术Background technique
近年来,随着计算机视觉技术的发展,基于单目视觉的同时定位和地图构建算法成为当前移动机器人研究的热点。然而,传统的基于单目视觉的同时定位与地图构建方法大多只能实现射影尺度或仿射尺度(affine-scaling)下的三维重建,即重建后的场景与现实世界场景存在一个尺度因子,该尺度因子即为现实世界地图尺度与构建出的地图尺度的比例。因此,若能够在移动机器人初始化时,确定该尺度因子,即可基于投影模型计算出单目相机在现实世界中的实际旋转量和实际平移量,构建与现实世界尺度相同的地图。In recent years, with the development of computer vision technology, the simultaneous positioning and map construction algorithms based on monocular vision have become the focus of current mobile robot research. However, the traditional simultaneous positioning and map construction methods based on monocular vision can only achieve 3D reconstruction at projective scale or affine-scaling, that is, there is a scale factor between the reconstructed scene and the real-world scene. The scale factor is the ratio of the real world map scale to the constructed map scale. Therefore, if the scale factor can be determined when the mobile robot is initialized, the actual rotation and translation of the monocular camera in the real world can be calculated based on the projection model, and a map with the same scale as the real world can be constructed.
发明内容Summary of the Invention
有鉴于此,本申请提供一种单目视觉重建中尺度因子的确定方法和移动机器人。In view of this, the present application provides a method for determining a mesoscale factor for monocular vision reconstruction and a mobile robot.
本申请第一方面提供一种单目视觉重建中尺度因子的确定方法,所述方法应用于移动机器人,所述方法包括:A first aspect of the present application provides a method for determining a mesoscale factor in monocular vision reconstruction. The method is applied to a mobile robot, and the method includes:
通过单目相机获取指定物体在第一时刻的第一图像以及所述指定物体在第二时刻的第二图像;Acquiring a first image of a designated object at a first moment and a second image of the designated object at a second moment through a monocular camera;
对所述第一图像和所述第二图像进行特征点提取和匹配,并依据配对好的特征点计算所述单目相机从所述第一时刻到所述第二时刻的归一化平移向量;Performing feature point extraction and matching on the first image and the second image, and calculating a normalized translation vector of the monocular camera from the first moment to the second moment according to the paired feature points ;
计算所述指定物体在所述第一时刻相对于所述单目相机的第一位姿,以及所述指定 物体在所述第二时刻相对于所述单目相机的第二位姿;Calculating a first pose of the designated object relative to the monocular camera at the first moment, and a second pose of the designated object relative to the monocular camera at the second moment;
依据所述第一位姿和所述第二位姿,计算所述单目相机从所述第一时刻到所述第二时刻的实际平移向量;Calculating an actual translation vector of the monocular camera from the first moment to the second moment according to the first pose and the second pose;
将所述实际平移向量的模与所述归一化平移向量的模的比值确定为本设备的单目视觉重建中的尺度因子。The ratio of the modulus of the actual translation vector to the modulus of the normalized translation vector is determined as a scale factor in the monocular vision reconstruction of the device.
本申请第二方面提供一种移动机器人,所述移动机器人包括单目相机和处理器;其中,A second aspect of the present application provides a mobile robot, which includes a monocular camera and a processor; wherein,
所述单目相机,用于获取指定物体在第一时刻的第一图像以及所述指定物体在第二时刻的第二图像;The monocular camera is configured to acquire a first image of a designated object at a first moment and a second image of the designated object at a second moment;
所述处理器,用于:The processor is configured to:
对所述第一图像和所述第二图像进行特征点提取和匹配,并依据匹配的特征点计算所述单目相机从所述第一时刻到所述第二时刻的归一化平移向量;Performing feature point extraction and matching on the first image and the second image, and calculating a normalized translation vector of the monocular camera from the first moment to the second moment according to the matched feature points;
计算所述指定物体在所述第一时刻相对于所述单目相机的第一位姿,以及所述指定物体在所述第二时刻相对于所述单目相机的第二位姿;Calculating a first pose of the designated object relative to the monocular camera at the first moment, and a second pose of the designated object relative to the monocular camera at the second moment;
依据所述第一位姿和所述第二位姿,计算所述单目相机从所述第一时刻到所述第二时刻的实际平移向量;Calculating an actual translation vector of the monocular camera from the first moment to the second moment according to the first pose and the second pose;
将所述实际平移向量的模与所述归一化平移向量的模的比值确定为本设备的单目视觉重建中的尺度因子。The ratio of the modulus of the actual translation vector to the modulus of the normalized translation vector is determined as a scale factor in the monocular vision reconstruction of the device.
本申请第三方面提供一种计算机可读存储介质,其上存储有计算机程序,所述程序被处理器执行实现本申请第一方面提供的任一项所述方法的步骤。A third aspect of the present application provides a computer-readable storage medium on which a computer program is stored, and the program is executed by a processor to implement the steps of any of the methods provided in the first aspect of the present application.
本申请提供的单目视觉重建中尺度因子的确定方法和移动机器人,由于指定物体的位置是固定不变的,因此,在移动机器人打滑、卡住等情况下,通过计算指定物体在第一时刻相对于单目相机的第一位姿和指定物体在第二时刻相对于单目相机的第二位姿,从而能够相对准确的计算得到单目相机从第一时刻到第二时刻在现实世界的实际平移向量。因此,本申请提供的方法,不存在因移动机器人打滑、卡住等情况导致确定出的尺度因子不准确的问题。The method for determining the meso-scale factor of the monocular vision reconstruction and the mobile robot provided in this application, because the position of the designated object is fixed, therefore, in the case of the mobile robot slipping, jamming, etc., by calculating the designated object at the first moment The first pose relative to the monocular camera and the second pose of the designated object relative to the monocular camera at the second moment, so that the real-time The actual translation vector. Therefore, the method provided in this application does not have the problem that the determined scale factor is inaccurate due to slipping, jamming, and the like of the mobile robot.
附图说明BRIEF DESCRIPTION OF THE DRAWINGS
图1为本申请提供的单目视觉重建中尺度因子的确定方法实施例一的流程图。FIG. 1 is a flowchart of Embodiment 1 of a method for determining a scale factor in monocular vision reconstruction provided by the present application.
图2为本申请一示例性实施例示出的单目相机采集指定物体的图像的示意图。Fig. 2 is a schematic diagram of a monocular camera acquiring an image of a specified object according to an exemplary embodiment of the present application.
图3为本申请一示例性实施例示出的计算指定物体相对于单目相机的位姿的流程图。Fig. 3 is a flowchart of calculating a pose of a specified object relative to a monocular camera according to an exemplary embodiment of the present application.
图4为本申请提供的移动机器人实施例一的硬件结构图。FIG. 4 is a hardware structural diagram of a first embodiment of a mobile robot provided in this application.
具体实施方式detailed description
这里将详细地对示例性实施例进行说明,其示例表示在附图中。下面的描述涉及附图时,除非另有表示,不同附图中的相同数字表示相同或相似的要素。以下示例性实施例中所描述的实施方式并不代表与本申请相一致的所有实施方式。相反,它们仅是与如所附权利要求书中所详述的、本申请的一些方面相一致的装置和方法的例子。Exemplary embodiments will be described in detail herein, examples of which are illustrated in the accompanying drawings. When the following description refers to the accompanying drawings, the same numbers in different drawings represent the same or similar elements unless otherwise indicated. The implementations described in the following exemplary embodiments do not represent all implementations consistent with this application. Rather, they are merely examples of devices and methods consistent with certain aspects of the application as detailed in the appended claims.
在本申请使用的术语是仅仅出于描述特定实施例的目的,而非旨在限制本申请。在本申请和所附权利要求书中所使用的单数形式的“一种”、“所述”和“该”也旨在包括多数形式,除非上下文清楚地表示其他含义。还应当理解,本文中使用的术语“和/或”是指并包含一个或多个相关联的列出项目的任何或所有可能组合。The terminology used in this application is for the purpose of describing particular embodiments only and is not intended to limit the application. As used in this application and the appended claims, the singular forms "a", "the" and "the" are intended to include the plural forms as well, unless the context clearly indicates otherwise. It should also be understood that the term "and / or" as used herein refers to and includes any or all possible combinations of one or more of the associated listed items.
应当理解,尽管在本申请可能采用术语第一、第二、第三等来描述各种信息,但这些信息不应限于这些术语。这些术语仅用来将同一类型的信息彼此区分开。例如,在不脱离本申请范围的情况下,第一信息也可以被称为第二信息,类似地,第二信息也可以被称为第一信息。取决于语境,如在此所使用的词语“如果”可以被解释成为“在……时”或“当……时”或“响应于确定”。It should be understood that although the terms first, second, third, etc. may be used in this application to describe various information, such information should not be limited to these terms. These terms are only used to distinguish the same type of information from each other. For example, without departing from the scope of the present application, the first information may also be referred to as the second information, and similarly, the second information may also be referred to as the first information. Depending on the context, the word "if" as used herein can be interpreted as "at" or "when" or "in response to determination".
相关技术中提出了一种单目视觉重建中尺度因子的确定方法。该方法利用单目相机采集到的相邻的两帧图像,运用对极几何(Epipolar Geometry)计算出这两帧图像之间单目相机的归一化平移向量;并利用码盘数据和IMU(Inertial measurement unit,惯性测量单元)数据计算这两帧图像之间单目相机在现实世界的实际平移向量,进而利用归一化平移向量和实际平移向量得到单目视觉重建中的尺度因子。A related method for determining the meso-scale factor in monocular vision reconstruction is proposed in the related art. This method uses two adjacent frames of images collected by a monocular camera, and uses epipolar geometry to calculate the normalized translation vector of the monocular camera between the two frames of images; and uses code disk data and IMU ( Inertial measurement unit (inertial measurement unit) data calculates the actual translation vector of the monocular camera in the real world between the two frames of images, and then uses the normalized translation vector and the actual translation vector to obtain the scale factor in monocular vision reconstruction.
但是,当采用上述方法确定单目视觉重建中的尺度因子时,由于移动机器人存在打滑、卡住等情况使得码盘计数与实际不符,导致该情况下采用码盘数据结合IMU数据 计算出的实际平移向量也不准确,进而依据该实际平移向量计算出的尺度因子也不准确。However, when the above-mentioned method is used to determine the scale factor in the monocular vision reconstruction, the code disk count is inconsistent with the actual situation due to the mobile robot's slipping, jamming, etc., resulting in the actual calculation of the code disk data combined with the IMU data in this case. The translation vector is also inaccurate, and the scale factor calculated based on the actual translation vector is also inaccurate.
本申请提供一种单目视觉重建中尺度因子的确定方法和移动机器人,以解决现有的方法存在的因移动机器人打滑、卡住等情况导致确定出的尺度因子不准确的问题。The present application provides a method for determining a scale factor in a monocular vision reconstruction and a mobile robot, so as to solve the problem that the determined scale factor is inaccurate due to the slipping, jamming, etc. of the mobile robot in the existing method.
本实施例提供的方法,可应用于移动机器人。例如,可应用于扫地机器人。The method provided by this embodiment can be applied to a mobile robot. For example, it can be applied to a cleaning robot.
下面给出几个具体的实施例,用以详细介绍本申请的技术方案。下面这几个具体的实施例可以相互结合,对于相同或相似的概念或过程可能在某些实施例中不再赘述。Several specific embodiments are given below to introduce the technical solution of the present application in detail. The following specific embodiments may be combined with each other, and the same or similar concepts or processes may not be repeated in some embodiments.
图1为本申请提供的单目视觉重建中尺度因子的确定方法实施例一的流程图。请参照图1,本实施例提供的方法,可以包括:FIG. 1 is a flowchart of Embodiment 1 of a method for determining a scale factor in monocular vision reconstruction provided by the present application. Referring to FIG. 1, the method provided in this embodiment may include:
S101、通过单目相机获取指定物体在第一时刻的第一图像以及上述指定物体在第二时刻的第二图像。S101. Obtain a first image of a specified object at a first moment and a second image of the specified object at a second moment through a monocular camera.
具体的,移动机器人上设置有单目相机,可通过该单目相机采集图像。可选的,指定物体可以是为该移动机器人充电的充电设备,移动机器人可在检测到本设备与上述指定物体断开连接后,通过单目相机获取上述指定物体在第一时刻的第一图像以及上述指定物体在第二时刻的第二图像。例如,存在邻近的采样时刻:第一时刻t1和第二时刻t2,移动机器人可以通过单目相机获取指定物体在第一时刻t1的第一图像F1、以及指定物体在第二时刻t2的第二图像F2。Specifically, the mobile robot is provided with a monocular camera, and images can be collected by the monocular camera. Optionally, the designated object may be a charging device for charging the mobile robot. After detecting that the device is disconnected from the designated object, the mobile robot may obtain a first image of the designated object at the first moment through a monocular camera. And the second image of the specified object at the second moment. For example, there are neighboring sampling times: the first time t1 and the second time t2. The mobile robot can obtain the first image F1 of the specified object at the first time t1 and the second image of the specified object at the second time t2 through the monocular camera. Image F2.
需要说明的是,移动机器人在第一时刻t1和第二时刻t2处于不同的位置,即单目相机在第一时刻t1和第二时刻t2处于不同的拍摄位置。It should be noted that the mobile robot is at different positions at the first time t1 and the second time t2, that is, the monocular camera is at different shooting positions at the first time t1 and the second time t2.
图2为本申请一示例性实施例示出的单目相机采集指定物体的图像的示意图。请参照图2,在图2所示实例中,指定物体是为该移动机器人充电的充电设备。Fig. 2 is a schematic diagram of a monocular camera acquiring an image of a specified object according to an exemplary embodiment of the present application. Please refer to FIG. 2. In the example shown in FIG. 2, the designated object is a charging device for charging the mobile robot.
参照图2,单目相机110在第一时刻t1和第二时刻t2处于不同的拍摄位置。结合前面的介绍,例如,一实施例中,移动机器人可在检测到本设备与充电设备200断开连接后,转向充电设备,进而在不同于之前的位置通过单目相机对充电设备200进行拍摄。这样,可得到充电设备200在第一时刻t1的第一图像,该第一图像对应于第一拍摄位置,以及充电设备200在第二时刻t2的第二图像,该第二图像对应于第二拍摄位置。Referring to FIG. 2, the monocular camera 110 is at different shooting positions at a first time t1 and a second time t2. In conjunction with the previous introduction, for example, in one embodiment, the mobile robot may turn to the charging device after detecting that the device is disconnected from the charging device 200, and then photograph the charging device 200 with a monocular camera at a position different from the previous one. . In this way, a first image of the charging device 200 at a first time t1 can be obtained, the first image corresponding to the first shooting position, and a second image of the charging device 200 at a second time t2, the second image corresponds to the second Shooting position.
S102、对上述第一图像和上述第二图像进行特征点提取和匹配,并依据配对好的特征点计算上述单目相机从上述第一时刻到上述第二时刻的归一化平移向量。S102. Perform feature point extraction and matching on the first image and the second image, and calculate a normalized translation vector of the monocular camera from the first time to the second time according to the paired feature points.
具体的,有关对第一图像和第二图像进行特征点提取和匹配的具体实现原理和实现过程可以参见相关技术中的描述,此处不再赘述。Specifically, for specific implementation principles and implementation processes of performing feature point extraction and matching on the first image and the second image, refer to descriptions in related technologies, and details are not described herein again.
进一步地,在匹配结束后,可利用匹配好的特征点分别在第一图像和第二图像中的像素坐标,基于对极约束,计算出从上述第一时刻到上述第二时刻,上述单目相机在第一拍摄位置与第二拍摄位置间的归一化平移向量。例如,可利用8对配对好的特征点,计算单目相机从第一时刻到第二时刻的归一化平移向量。Further, after the matching is completed, the pixel coordinates of the matched feature points in the first image and the second image may be used to calculate the monocular from the first time to the second time based on the epipolar constraint. The normalized translation vector of the camera between the first shooting position and the second shooting position. For example, the normalized translation vector of the monocular camera from the first moment to the second moment may be calculated using eight pairs of paired feature points.
具体的,对极约束可以用下述公式表示:Specifically, the epipolar constraint can be expressed by the following formula:
Figure PCTCN2019101704-appb-000001
Figure PCTCN2019101704-appb-000001
其中,K为单目相机的内参矩阵,p 1和p 2为配对好的特征点分别在第一图像和第二图像上的像素齐次坐标,R ep为单目相机从第一时刻t1到第二时刻t2的旋转变化量,t ep为单目相机从第一时刻t1到第二时刻t2的归一化平移向量。 Among them, K is the internal parameter matrix of the monocular camera, p 1 and p 2 are the pixel homogeneous coordinates of the paired feature points on the first image and the second image, respectively, and Rep is the monocular camera from the first time t1 to The rotation change amount at the second time t2, t ep is the normalized translation vector of the monocular camera from the first time t1 to the second time t2.
需要说明的是,有关基于对极约束,依据配对好的特征点计算单目相机从第一时刻t1到第二时刻t2的归一化平移向量的具体实现过程可以参见相关技术中的介绍,此处不再赘述。It should be noted that the specific implementation process of calculating the normalized translation vector of the monocular camera from the first time t1 to the second time t2 based on the epipolar constraint and the paired feature points can be referred to the introduction in the related technology. I will not repeat them here.
S103、计算上述指定物体在上述第一时刻相对于上述单目相机的第一位姿,以及上述指定物体在上述第二时刻相对于上述单目相机的第二位姿。S103. Calculate a first pose of the designated object relative to the monocular camera at the first moment, and a second pose of the designated object relative to the monocular camera at the second moment.
具体的,图3为本申请一示例性实施例示出的计算指定物体相对于单目相机的位姿的流程图。请参照图3,计算指定物体相对于单目相机的位姿,可以包括:Specifically, FIG. 3 is a flowchart of calculating a pose of a specified object relative to a monocular camera according to an exemplary embodiment of the present application. Referring to FIG. 3, calculating the pose of the specified object relative to the monocular camera may include:
S301、针对第一图像和第二图像中的每帧图像,从该帧图像中获取上述指定物体上的指定点的像素坐标;上述指定点的数量大于或者等于4。S301. For each frame of the first image and the second image, obtain pixel coordinates of a specified point on the specified object from the frame image; the number of the specified points is greater than or equal to 4.
具体的,本步骤中,可基于指定物体的属性信息,从图像中识别出该指定物体,进而基于识别出的指定物体,从图像中获取该指定物体上的指定点的像素坐标。Specifically, in this step, the specified object may be identified from the image based on the attribute information of the specified object, and then based on the identified specified object, the pixel coordinates of the specified point on the specified object may be obtained from the image.
需要说明的是,指定物体的属性信息可以包括材料属性、颜色属性、形状属性等。本实施例中,不对此作出限定。It should be noted that the attribute information of the designated object may include material attributes, color attributes, shape attributes, and the like. In this embodiment, this is not limited.
例如,一实施例中,指定物体可以是为该移动机器人充电的充电设备。该充电设备上设置有标记物。例如,该标记物可以是由特定材料、特定颜色、特定形状、指定数量和/或指定内容的若干标记块组成的标记物。再例如,该标记物可以是由特定材料制成的 一指定形状的标记物。例如,当单目相机为红外相机时,标记物可以由指定数量的高反射材料制成的标记块组成;再例如,当单目相机为RGB相机时,标记物可以由指定数量的印有黑白相间棋盘格的标记块组成。本实施例中,不对标记物的具体设置形式进行限定。For example, in one embodiment, the designated object may be a charging device for charging the mobile robot. The charging device is provided with a marker. For example, the marker may be a marker composed of several marker blocks of a specific material, a specific color, a specific shape, a specified number, and / or a specified content. As another example, the marker may be a designated shape marker made of a specific material. For example, when the monocular camera is an infrared camera, the marker may be composed of a specified number of highly reflective material; for another example, when the monocular camera is an RGB camera, the marker may be a specified number of black and white printed Consisting of checkered checkered blocks. In this embodiment, the specific setting form of the marker is not limited.
需要说明的是,充电设备上的标记物可反映充电设备的属性信息,可基于充电设备的标记物来识别出图像中的充电设备。有关基于指定物体的属性信息,识别出图像中的指定物体的具体实现原理和实现过程可以参见相关技术中描述,此处不再赘述。It should be noted that the marker on the charging device can reflect the attribute information of the charging device, and the charging device in the image can be identified based on the marker of the charging device. For the specific implementation principle and implementation process of identifying the specified object in the image based on the attribute information of the specified object, refer to the description in the related technology, and details are not described herein again.
进一步地,指定物体上的指定点可以是根据实际需要设定的,例如,指定点可以是标记物的角点、中心点等。本实施例中,不对指定点的具体位置进行限定。需要说明的是,该指定点的数量大于或者等于4。Further, the designated point on the designated object may be set according to actual needs, for example, the designated point may be a corner point, a center point, etc. of the marker. In this embodiment, the specific position of the designated point is not limited. It should be noted that the number of the designated points is greater than or equal to four.
下面以图2所示示例,详细说明本步骤的具体实现过程:The following uses the example shown in Figure 2 to explain the detailed implementation of this step in detail:
具体的,参照图2,在图2所示示例中,充电设备200上的标记物210由4个标记块1、2、3、4组成,指定该充电设备200上的指定点为各标记块的中心点。此时,可基于这4个标记块1、2、3、4的材料、颜色、形状以及各标记块的间隔距离等属性信息,从图像中识别出这4个标记块1、2、3、4,进而得到各个标记块的中心点的像素坐标,这样,即可得到指定物体上的指定点的像素坐标。Specifically, referring to FIG. 2, in the example shown in FIG. 2, the marker 210 on the charging device 200 is composed of four marker blocks 1, 2, 3, and 4, and designated points on the charging device 200 are designated as the marker blocks. The center point. At this time, the four marker blocks 1, 2, 3, and 4 can be identified from the image based on attribute information such as the material, color, shape, and the distance between the marker blocks. 4. Furthermore, the pixel coordinates of the center point of each marker block are obtained. In this way, the pixel coordinates of the specified point on the specified object can be obtained.
为方便说明,将各个标记块的中心点依序记为Bi,其中i等于1到4。将第i个标记块的中心点Bi的像素坐标记为(u i,v i)。 For the convenience of description, the center point of each marked block is sequentially recorded as Bi, where i is equal to 1 to 4. The pixel coordinates of the center point Bi of the i-th labeled block are labeled (u i , v i ).
S302、依据各个上述指定点的像素坐标,采用畸变校正算法计算各个上述指定点畸变校正后的第一坐标。S302. According to the pixel coordinates of each of the specified points, a distortion correction algorithm is used to calculate the first coordinates of each of the specified points after distortion correction.
具体,畸变校正算法采用如下公式表示:Specifically, the distortion correction algorithm is expressed by the following formula:
Figure PCTCN2019101704-appb-000002
Figure PCTCN2019101704-appb-000002
其中,K为单目相机的内参矩阵;Among them, K is the internal parameter matrix of the monocular camera;
k 1,k 2,k 3,p 1,p 2为单目相机的畸变参数; k 1 , k 2 , k 3 , p 1 , p 2 are distortion parameters of the monocular camera;
(u i,v i)为第i个指定点的像素坐标; (u i , v i ) are the pixel coordinates of the i-th specified point;
(x i,y i)为第i个指定点畸变校正后的第一坐标。 (x i , y i ) is the first coordinate after distortion correction of the i-th designated point.
S303、依据各个上述指定点畸变校正后的第一坐标和预存的各个上述指定点在指定坐标系下的第二坐标,计算上述指定物体相对于上述单目相机的旋转矩阵和平移向量。S303. Calculate a rotation matrix and a translation vector of the specified object relative to the monocular camera according to the first coordinates of each of the specified point distortion corrections and the pre-stored second coordinates of each of the specified points in the specified coordinate system.
具体的,指定坐标系为一绝对坐标系。具体的,在图2所示示例中,指定坐标系为充电设备上标示的坐标系。即在图2所示实例中,指定坐标系的原点为充电设备的中心点,X轴为水平向右,Y轴为垂直于X轴向下。Specifically, the designated coordinate system is an absolute coordinate system. Specifically, in the example shown in FIG. 2, the designated coordinate system is a coordinate system marked on the charging device. That is, in the example shown in FIG. 2, the origin of the designated coordinate system is the center point of the charging device, the X-axis is horizontal to the right, and the Y-axis is perpendicular to the X-axis downward.
需要说明的是,本步骤中,可基于视觉伺服中的比例正交投影迭代变换算法(Pose from Orthography and Scaling with Iterations,POSIT),根据指定物体上的多个指定点在指定坐标系下的第二坐标以及这多个指定点畸变校正后的第一坐标,进行正交投影迭代,计算出指定物体相对于单目相机的旋转矩阵和平移向量。It should be noted that, in this step, based on the proportional orthogonal projection iterative transformation algorithm (Position from Orthography and Scaling with Iterations, POSIT) in the visual servoing, according to a plurality of specified points on the specified object in the specified coordinate system. The two coordinates and the first coordinates after the distortion correction of the plurality of designated points are subjected to orthogonal projection iteration to calculate a rotation matrix and a translation vector of the designated object relative to the monocular camera.
具体的,该步骤的具体实现过程,可以包括:Specifically, the specific implementation process of this step may include:
依据各个上述指定点畸变校正后的第一坐标和预存的各个上述指定点在指定坐标系下的第二坐标,按照第一公式计算第一向量i、第二向量j′、第三向量k'和第一系数z。Calculate the first vector i, the second vector j ′, and the third vector k ′ according to the first formula according to the first coordinates after the distortion correction of each of the specified points and the second coordinates of the predetermined points in the specified coordinate system that are stored in advance. And the first coefficient z.
具体的,第一公式为:Specifically, the first formula is:
Figure PCTCN2019101704-appb-000003
Figure PCTCN2019101704-appb-000003
其中,所述指定点包括参考指定点和目标指定点,所述A为由各个目标指定点在所述指定坐标系下的第二坐标与所述参考指定点在所述指定坐标系下的第二坐标之间的差值构成的矩阵;所述X为由各个所述目标指定点畸变校正后的第一坐标中的X坐标与所述参考指定点畸变校正后的第一坐标中的X坐标之间的差值构成的向量;所述Y为由各个所述目标指定点畸变校正后的第一坐标中的Y坐标与所述参考指定点畸变校正后的第一坐标中的Y坐标之间的差值构成的向量。Wherein, the designated point includes a reference designated point and a target designated point, and A is the second coordinate of each target designated point in the designated coordinate system and the first coordinate of the reference designated point in the designated coordinate system. A matrix formed by the difference between two coordinates; the X is the X coordinate in the first coordinate after the distortion correction of the designated point of each target and the X coordinate in the first coordinate after the distortion of the reference designated point are corrected A vector formed by the difference between them; the Y is between the Y coordinate in the first coordinate after the distortion correction of the designated point of each target and the Y coordinate in the first coordinate after the distortion of the reference designated point are corrected A vector of differences.
需要说明的是,参考指定点可以为任意一个指定点。本实施例中,以参考指定点为 第1个指定点为例进行说明。It should be noted that the reference designated point may be any designated point. In this embodiment, the reference designated point is used as an example for description.
进一步地,为方便说明,将第i个指定点在指定坐标系下的第二坐标记为(a i,b i,0)。 Further, for convenience of explanation, the second sitting mark of the i-th designated point in the designated coordinate system is (a i , b i , 0).
结合上面的例子,参照图2,此时有:With reference to the above example, referring to FIG. 2, at this time:
Figure PCTCN2019101704-appb-000004
Figure PCTCN2019101704-appb-000004
需要说明的是,第一向量i、第二向量j′、第三向量k'和第一系数z均包括三个元素。It should be noted that each of the first vector i, the second vector j ′, the third vector k ′, and the first coefficient z includes three elements.
将所述i、所述j′和所述k'按矩阵行方向依序排列,得到所述指定物体相对于所述单目相机的旋转矩阵。Arranging the i, the j ′, and the k ′ in a row direction of the matrix in order to obtain a rotation matrix of the specified object relative to the monocular camera.
具体的,将指定物体相对于单目相机的旋转矩阵记为R,此时有:Specifically, the rotation matrix of the specified object relative to the monocular camera is denoted as R, and at this time, there are:
Figure PCTCN2019101704-appb-000005
Figure PCTCN2019101704-appb-000005
依据所述i、所述j′、所述k'和所述z按照第二公式计算指定物体相对于所述单目相机的平移向量。Calculate a translation vector of a specified object relative to the monocular camera according to the i, j ′, k ′, and z according to a second formula.
具体的,第二公式为:Specifically, the second formula is:
Figure PCTCN2019101704-appb-000006
Figure PCTCN2019101704-appb-000006
其中,(a 1,b 1)为所述参考指定点在所述指定坐标下的第二坐标;(x 1,y 1)为所述参考指定点畸变校正后的第一坐标;i 1、i 2分别为所述第一向量i中的第一个元素和第二个元素;j' 1、j' 2分别为所述第二向量j'中的第一个元素和第二个元素;k' 1、k' 2分别为所述第三向量k'中的第一个元素和第二个元素;所述t为所述指定物体相对于所述单目相机的平移向量。 Where (a 1 , b 1 ) are the second coordinates of the reference designated point under the designated coordinates; (x 1 , y 1 ) are the first coordinates of the reference designated point after distortion correction; i 1 , i 2 is a first element and a second element in the first vector i; j ' 1 and j' 2 are a first element and a second element in the second vector j '; k ′ 1 and k ′ 2 are the first element and the second element in the third vector k ′, respectively; and t is a translation vector of the specified object relative to the monocular camera.
这样,通过以上步骤,基于第一图像,便可以计算得到指定物体在第一时刻t1相对于单目相机的第一旋转矩阵R t1和第一平移向量t t1。基于第二图像,便可以计算得到指定物体在第二时刻t2相对于单目相机的第二旋转矩阵R t2和第二平移向量t t2In this way, through the above steps, based on the first image, the first rotation matrix R t1 and the first translation vector t t1 of the specified object with respect to the monocular camera at the first time t1 can be calculated. Based on the second image, a second rotation matrix R t2 and a second translation vector t t2 of the designated object relative to the monocular camera at the second time t2 can be calculated.
S304、依据上述指定物体相对于上述单目相机的旋转矩阵和平移向量,得到上述指定物体相对于上述单目相机的位姿。S304. Obtain the pose of the specified object relative to the monocular camera according to the rotation matrix and translation vector of the specified object relative to the monocular camera.
在一实施例中,将指定物体在第一时刻t1相对于单目相机的第一位姿记为T t1,结合前面的介绍,可知: In an embodiment, the first pose of the designated object relative to the monocular camera at the first time t1 is recorded as T t1 . With the foregoing introduction, it can be known that:
Figure PCTCN2019101704-appb-000007
Figure PCTCN2019101704-appb-000007
进一步地,将指定物体在第二时刻t2相对于单目相机的第二位姿记为T t2,结合前面的介绍,可知: Further, the second pose of the designated object relative to the monocular camera at the second time t2 is recorded as T t2 . In combination with the previous introduction, it can be known that:
Figure PCTCN2019101704-appb-000008
Figure PCTCN2019101704-appb-000008
S104、依据上述第一位姿和上述第二位姿,计算上述单目相机从上述第一时刻到上述第二时刻的实际平移向量。S104. Calculate an actual translation vector of the monocular camera from the first moment to the second moment according to the first pose and the second pose.
具体的,本步骤的具体实现过程,可以包括:依据所述第一位姿和所述第二位姿,计算所述单目相机从所述第一时刻到所述第二时刻的位姿变化;从所述位姿变化中获取所述真实平移向量。Specifically, the specific implementation process of this step may include: calculating the change in pose of the monocular camera from the first moment to the second moment according to the first pose and the second pose. Obtaining the true translation vector from the pose change.
具体的,可按照如下公式计算所述单目相机从所述第一时刻t1到所述第二时刻t2的位姿变化
Figure PCTCN2019101704-appb-000009
Specifically, the pose change of the monocular camera from the first time t1 to the second time t2 can be calculated according to the following formula
Figure PCTCN2019101704-appb-000009
Figure PCTCN2019101704-appb-000010
Figure PCTCN2019101704-appb-000010
进一步地,单目相机从第一时刻t1到第二时刻t2的位姿变化包括实际旋转矩阵和实际平移向量,且有:Further, the pose change of the monocular camera from the first time t1 to the second time t2 includes an actual rotation matrix and an actual translation vector, and has:
Figure PCTCN2019101704-appb-000011
Figure PCTCN2019101704-appb-000011
因此,基于前面计算得到的位姿变化,便可以获取单目相机从第一时刻t1到第二时 刻t2的实际旋转矩阵和实际平移向量。参见上式,可知,该实际平移向量即为位姿变化中最后一个列向量的前三个元素构成的向量。Therefore, based on the previously calculated pose changes, the actual rotation matrix and actual translation vector of the monocular camera from the first time t1 to the second time t2 can be obtained. Seeing the above formula, it can be known that the actual translation vector is a vector composed of the first three elements of the last column vector in the pose change.
S105、将上述实际平移向量的模与上述归一化平移向量的模之间的比值确定为本设备单目视觉重建中的尺度因子。S105. Determine a ratio between a mode of the actual translation vector and a mode of the normalized translation vector as a scale factor in the monocular vision reconstruction of the device.
具体的,经步骤S102计算得到单目相机从第一时刻t1到第二时刻t2的归一化平移向量,并经步骤S104计算得到单目相机从第一时刻t1到第二时刻t2在现实世界的实际平移向量后,将实际平移向量的模与归一化平移向量的模之间的比值确定为本设备单目视觉重建中的尺度因子。即:Specifically, the normalized translation vector of the monocular camera from the first time t1 to the second time t2 is calculated through step S102, and the monocular camera is calculated from the first time t1 to the second time t2 in the real world through step S104. After the actual translation vector, the ratio between the modulus of the actual translation vector and the modulus of the normalized translation vector is determined as the scale factor in the monocular vision reconstruction of the device. which is:
Figure PCTCN2019101704-appb-000012
Figure PCTCN2019101704-appb-000012
需要说明的是,当计算出本设备单目视觉重建中的尺度因子后,即可计算出两时刻单目相机在现实世界的位姿变化量和特征点对应的地图,进而在后续的同时定位与地图重建中,可基于现有的基于视觉的同时定位与地图重建算法,利用最小化重投影误差计算后续单目相机在现实世界中的位姿变化量和地图点位置。这样,结合回环检测纠正单目相机位姿和地图点位置漂移,即可定位和构建真实尺度下的地图。It should be noted that after the scale factor in the monocular vision reconstruction of this device is calculated, the map corresponding to the change in the pose and feature points of the monocular camera in the real world at two moments can be calculated, and then positioned at the same time in the subsequent In map reconstruction, the existing vision-based simultaneous positioning and map reconstruction algorithms can be used to calculate the change in pose and position of map points of subsequent monocular cameras in the real world by minimizing reprojection errors. In this way, combined with loop detection to correct the monocular camera pose and map point position drift, a map at a real scale can be located and constructed.
本实施例提供的单目视觉重建中尺度因子的确定方法,由于指定物体的位置是固定不变的,因此,在移动机器人打滑、卡住等情况下,通过计算指定物体在第一时刻相对于单目相机的第一位姿和指定物体在第二时刻相对于单目相机的第二位姿,进而依据第一位姿和第二位姿计算得到单目相机从第一时刻到第二时刻在现实世界的实际平移量。因此,本申请提供的方法,可以有效避免因移动机器人打滑、卡住等情况导致确定出的尺度因子不准确的问题。In the method for determining the scale factor in the monocular vision reconstruction provided by this embodiment, since the position of the designated object is fixed, when the mobile robot is slipping, stuck, etc., the calculated value of the designated object relative to the first time is calculated by The first pose and designated object of the monocular camera are relative to the second pose of the monocular camera at the second moment, and the monocular camera is calculated from the first moment to the second moment according to the first pose and the second pose. The actual amount of translation in the real world. Therefore, the method provided by the present application can effectively avoid the problem that the determined scale factor is inaccurate due to slipping, jamming, and the like of the mobile robot.
以上对本申请提供的单目视觉重建中尺度因子的确定方法进行了介绍,下面对本申请提供移动机器人进行介绍:The method for determining the meso-scale factor of the monocular vision reconstruction provided in the present application is described above, and the mobile robot provided in the present application is introduced below:
图4为本申请提供的移动机器人实施例一的硬件结构图。请参照图4,本实施例提供的移动机器人100,可以包括单目相机410和处理器420;其中,FIG. 4 is a hardware structural diagram of a first embodiment of a mobile robot provided in this application. Referring to FIG. 4, the mobile robot 100 provided in this embodiment may include a monocular camera 410 and a processor 420. Among them,
所述单目相机410,用于获取指定物体在第一时刻的第一图像以及所述指定物体在第二时刻的第二图像;The monocular camera 410 is configured to acquire a first image of a designated object at a first moment and a second image of the designated object at a second moment;
所述处理器420,用于:The processor 420 is configured to:
对所述第一图像和所述第二图像进行特征点提取和匹配,并依据配对好的特征点计算所述单目相机从所述第一时刻到所述第二时刻的归一化平移向量;Performing feature point extraction and matching on the first image and the second image, and calculating a normalized translation vector of the monocular camera from the first moment to the second moment according to the paired feature points ;
计算所述指定物体在所述第一时刻相对于所述单目相机的第一位姿,以及所述指定物体在所述第二时刻相对于所述单目相机的第二位姿;Calculating a first pose of the designated object relative to the monocular camera at the first moment, and a second pose of the designated object relative to the monocular camera at the second moment;
依据所述第一位姿和所述第二位姿,计算所述单目相机从所述第一时刻到所述第二时刻的实际平移向量;Calculating an actual translation vector of the monocular camera from the first moment to the second moment according to the first pose and the second pose;
将所述实际平移向量的模与所述归一化平移向量的模的比值确定为本设备单目视觉重建中的尺度因子。The ratio of the modulus of the actual translation vector to the modulus of the normalized translation vector is determined as a scale factor in the monocular vision reconstruction of the device.
本实施例的移动机器人,可用于执行图1所示方法实施例的技术方案,其实现原理和技术效果类似,此处不再赘述。The mobile robot of this embodiment may be used to execute the technical solution of the method embodiment shown in FIG. 1, and the implementation principles and technical effects thereof are similar, and details are not described herein again.
进一步地,所述处理器420,具体用于:Further, the processor 420 is specifically configured to:
针对第一图像和第二图像中的每帧图像,从该帧图像中获取所述指定物体上的指定点的像素坐标;所述指定点的数量大于或者等于4;For each frame of the first image and the second image, obtain pixel coordinates of a specified point on the specified object from the frame image; the number of the specified points is greater than or equal to 4;
依据各个所述指定点的像素坐标,采用畸变校正算法,得到各个所述指定点畸变校正后的第一坐标;Using a distortion correction algorithm according to the pixel coordinates of each of the specified points to obtain the first coordinate after the distortion correction of each of the specified points;
依据各个所述指定点畸变校正后的第一坐标和预存的各个所述指定点在指定坐标系下的第二坐标,计算所述指定物体相对于所述单目相机的旋转矩阵和平移向量;Calculating a rotation matrix and a translation vector of the specified object relative to the monocular camera according to the first coordinates of each of the specified point distortion corrections and the pre-stored second coordinates of each of the specified points in the specified coordinate system;
依据所述指定物体相对于所述单目相机的旋转矩阵和平移向量,得到所述指定物体相对于所述单目相机的位姿。According to a rotation matrix and a translation vector of the designated object with respect to the monocular camera, a posture of the designated object with respect to the monocular camera is obtained.
进一步地,所述处理器420,具体用于:Further, the processor 420 is specifically configured to:
依据所述第一位姿和所述第二位姿,计算从所述第一时刻到所述第二时刻所述单目相机的位姿变化;Calculating a change in pose of the monocular camera from the first moment to the second moment according to the first pose and the second pose;
从所述位姿变化中获取所述实际平移向量。The actual translation vector is obtained from the pose change.
进一步地,所述处理器420,用于基于所述指定物体的属性信息,从该帧图像中识别出所述指定物体,并基于识别出的所述指定物体,获取所述指定物体上的指定点的像素坐标。Further, the processor 420 is configured to identify the specified object from the frame image based on the attribute information of the specified object, and obtain a designation on the specified object based on the identified specified object. The pixel coordinates of the point.
进一步地,所述处理器420,具体用于:Further, the processor 420 is specifically configured to:
依据各个所述指定点畸变校正后的第一坐标和预存的各个所述指定点在指定坐标系下的第二坐标,按照第一公式计算第一向量i、第二向量j′、第三向量k'和第一系数z;Calculate the first vector i, the second vector j ′, and the third vector according to the first formula according to the first coordinates of each of the specified point distortion corrections and the second coordinates of each of the specified points stored in the specified coordinate system. k 'and the first coefficient z;
将所述i、所述j′和所述k'按矩阵行方向依序排列,得到所述指定物体相对于所述单目相机的旋转矩阵;Arranging i, j ′ and k ′ in the row direction of the matrix in order to obtain a rotation matrix of the specified object relative to the monocular camera;
依据所述i、所述j′、所述k'和所述z,按照第二公式计算所述指定物体相对于所述单目相机的平移向量;Calculating a translation vector of the specified object with respect to the monocular camera according to the i, the j ′, the k ′, and the z according to a second formula;
其中,所述第一公式为:The first formula is:
Figure PCTCN2019101704-appb-000013
Figure PCTCN2019101704-appb-000013
所述第二公式为:The second formula is:
Figure PCTCN2019101704-appb-000014
Figure PCTCN2019101704-appb-000014
所述指定点包括参考指定点和目标指定点,所述A为各个目标指定点在所述指定坐标系下的第二坐标与所述参考指定点在所述指定坐标系下的第二坐标的差值构成的矩阵;所述X为各个所述目标指定点畸变校正后的第一坐标中的X坐标与所述参考指定点畸变校正后的第一坐标中的X坐标的差值构成的向量;所述Y为各个所述目标指定点畸变校正后的第一坐标中的Y坐标与所述参考指定点畸变校正后的第一坐标中的Y坐标的差值构成的向量;The designated point includes a reference designated point and a target designated point, and A is a second coordinate of each target designated point in the designated coordinate system and a second coordinate of the reference designated point in the designated coordinate system. A matrix formed by the differences; the X is a vector formed by the difference between the X coordinate in the first coordinate after the distortion of the designated point of the target is corrected and the X coordinate in the first coordinate after the distortion of the reference designated point is corrected The Y is a vector formed by the difference between the Y coordinate in the first coordinate after the distortion of the specified point of the target is corrected and the Y coordinate in the first coordinate after the distortion of the reference point is corrected;
其中,(a 1,b 1)为所述参考指定点在所述指定坐标下的第二坐标;(x 1,y 1)为所述参考指定点畸变校正后的第一坐标;所述i 1和所述i 2分别为所述i中的第一个元素和第二个元素;所述j' 1和所述j' 2分别为所述j'中的第一个元素和第二个元素;所述k' 1和所述 k' 2分别为所述k'中的第一个元素和第二个元素;所述t为所述指定物体相对于所述单目相机的平移向量。 Where (a 1 , b 1 ) are the second coordinates of the reference designated point under the designated coordinates; (x 1 , y 1 ) are the first coordinates of the reference designated point after distortion correction; and i 1 and i 2 are the first and second elements in i, respectively; the j ' 1 and j' 2 are the first and second elements in j ', respectively Elements; the k ′ 1 and the k ′ 2 are the first element and the second element in the k ′, respectively; and t is a translation vector of the specified object with respect to the monocular camera.
进一步地,所述指定物体为用于给本设备充电的充电设备;所述处理器420,用于在检测到本设备与所述指定物体断开连接后,通过单目相机获取所述指定物体在第一时刻的第一图像以及所述指定物体在第二时刻的第二图像。Further, the designated object is a charging device for charging the device; and the processor 420 is configured to obtain the designated object through a monocular camera after detecting that the device is disconnected from the designated object. A first image at a first moment and a second image of the designated object at a second moment.
本申请还提供一种计算机可读存储介质,其上存储有计算机程序,所述程序被处理器执行时实现本申请第一方面提供的任一项所述方法的步骤。The present application also provides a computer-readable storage medium on which a computer program is stored, and when the program is executed by a processor, the steps of the method according to any one of the first aspect of the application are implemented.
具体的,适合于存储计算机程序指令的计算机可读存储介质包括所有形式的非易失性存储器、媒介和存储器设备,例如包括半导体存储器设备(例如EPROM、EEPROM和闪存设备)、磁盘(例如内部硬盘或可移动盘)、磁光盘以及CD ROM和DVD-ROM盘。Specifically, a computer-readable storage medium suitable for storing computer program instructions includes all forms of non-volatile memory, media, and memory devices, such as semiconductor memory devices (such as EPROM, EEPROM, and flash memory devices), magnetic disks (such as internal hard disks) Or removable disks), magneto-optical disks, and CD ROM and DVD-ROM disks.
以上所述仅为本申请的较佳实施例而已,并不用以限制本申请,凡在本申请的精神和原则之内,所做的任何修改、等同替换、改进等,均应包含在本申请保护的范围之内。The above are only preferred embodiments of this application, and are not intended to limit this application. Any modification, equivalent replacement, or improvement made within the spirit and principles of this application shall be included in this application Within the scope of protection.

Claims (10)

  1. 一种单目视觉重建中尺度因子的确定方法,应用于移动机器人,所述方法包括:A method for determining a mesoscale factor in monocular vision reconstruction, which is applied to a mobile robot, the method includes:
    通过单目相机获取指定物体在第一时刻的第一图像以及所述指定物体在第二时刻的第二图像;Acquiring a first image of a designated object at a first moment and a second image of the designated object at a second moment through a monocular camera;
    对所述第一图像和所述第二图像进行特征点提取和匹配,并依据配对好的特征点计算所述单目相机从所述第一时刻到所述第二时刻的归一化平移向量;Performing feature point extraction and matching on the first image and the second image, and calculating a normalized translation vector of the monocular camera from the first moment to the second moment according to the paired feature points ;
    计算所述指定物体在所述第一时刻相对于所述单目相机的第一位姿,以及所述指定物体在所述第二时刻相对于所述单目相机的第二位姿;Calculating a first pose of the designated object relative to the monocular camera at the first moment, and a second pose of the designated object relative to the monocular camera at the second moment;
    依据所述第一位姿和所述第二位姿,计算所述单目相机从所述第一时刻到所述第二时刻的实际平移向量;Calculating an actual translation vector of the monocular camera from the first moment to the second moment according to the first pose and the second pose;
    将所述实际平移向量的模与所述归一化平移向量的模之间的比值确定为单目视觉重建中的尺度因子。A ratio between a mode of the actual translation vector and a mode of the normalized translation vector is determined as a scale factor in monocular vision reconstruction.
  2. 根据权利要求1所述的方法,其特征在于,计算所述指定物体相对于所述单目相机的位姿,包括:The method according to claim 1, wherein calculating the pose of the designated object relative to the monocular camera comprises:
    针对所述第一图像和所述第二图像中的每帧图像,从该帧图像中获取所述指定物体上的指定点的像素坐标;所述指定点的数量大于或者等于4;For each frame of the first image and the second image, obtain pixel coordinates of a specified point on the specified object from the frame image; the number of the specified points is greater than or equal to 4;
    依据各个所述指定点的像素坐标,采用畸变校正算法计算各个所述指定点畸变校正后的第一坐标;Calculating, according to the pixel coordinates of each of the specified points, a first coordinate after distortion correction of each of the specified points by using a distortion correction algorithm;
    依据各个所述指定点畸变校正后的第一坐标和预存的各个所述指定点在指定坐标系下的第二坐标,计算所述指定物体相对于所述单目相机的旋转矩阵和平移向量;Calculating a rotation matrix and a translation vector of the specified object relative to the monocular camera according to the first coordinates of each of the specified point distortion corrections and the pre-stored second coordinates of each of the specified points in the specified coordinate system;
    依据所述指定物体相对于所述单目相机的所述旋转矩阵和所述平移向量,得到所述指定物体相对于所述单目相机的位姿。Obtaining the pose of the specified object relative to the monocular camera according to the rotation matrix and the translation vector of the specified object relative to the monocular camera.
  3. 根据权利要求1所述的方法,其特征在于,依据所述第一位姿和所述第二位姿,计算所述单目相机从所述第一时刻到所述第二时刻的实际平移向量,包括:The method according to claim 1, wherein an actual translation vector of the monocular camera from the first moment to the second moment is calculated according to the first pose and the second pose. ,include:
    依据所述第一位姿和所述第二位姿,计算所述单目相机从所述第一时刻到所述第二时刻的位姿变化;Calculating a change in pose of the monocular camera from the first moment to the second moment according to the first pose and the second pose;
    从所述位姿变化中获取所述实际平移向量。The actual translation vector is obtained from the pose change.
  4. 根据权利要求2所述的方法,其特征在于,从该帧图像中获取所述指定物体上的指定点的像素坐标,包括:The method according to claim 2, wherein obtaining pixel coordinates of a specified point on the specified object from the frame image comprises:
    基于所述指定物体的属性信息,从该帧图像中识别出所述指定物体;Identifying the designated object from the frame image based on the attribute information of the designated object;
    基于识别出的所述指定物体,获取所述指定物体上的指定点的像素坐标。Based on the identified specified object, pixel coordinates of a specified point on the specified object are obtained.
  5. 根据权利要求2所述的方法,其特征在于,依据各个所述指定点畸变校正后的第一坐标和预存的各个所述指定点在指定坐标系下的第二坐标,计算所述指定物体相对于所述单目相机的旋转矩阵和平移向量,包括:The method according to claim 2, characterized in that the relative of the specified object is calculated according to the first coordinates of each of the specified point distortion corrections and the second coordinates of each of the specified points stored in the specified coordinate system in advance. The rotation matrix and translation vector for the monocular camera include:
    依据各个所述指定点畸变校正后的第一坐标和预存的各个所述指定点在指定坐标系下的第二坐标,按照第一公式计算第一向量i、第二向量j′、第三向量k'和第一系数z;Calculate the first vector i, the second vector j ′, and the third vector according to the first formula according to the first coordinates of each of the specified point distortion corrections and the second coordinates of each of the specified points stored in the specified coordinate system. k 'and the first coefficient z;
    将所述i、所述j′和所述k'按矩阵行方向依序排列,得到所述指定物体相对于所述单目相机的旋转矩阵;Arranging i, j ′ and k ′ in the row direction of the matrix in order to obtain a rotation matrix of the specified object relative to the monocular camera;
    依据所述i、所述j′、所述k'和所述z,按照第二公式计算所述指定物体相对于所述单目相机的平移向量;Calculating a translation vector of the specified object with respect to the monocular camera according to the i, the j ′, the k ′, and the z according to a second formula;
    其中,所述第一公式为:The first formula is:
    Figure PCTCN2019101704-appb-100001
    Figure PCTCN2019101704-appb-100001
    所述第二公式为:The second formula is:
    Figure PCTCN2019101704-appb-100002
    Figure PCTCN2019101704-appb-100002
    其中,所述指定点包括参考指定点和目标指定点,所述A为各个所述目标指定点在所述指定坐标系下的第二坐标与所述参考指定点在所述指定坐标系下的第二坐标的差值构成的矩阵;所述X为各个所述目标指定点畸变校正后的第一坐标中的X坐标与所述参考指定点畸变校正后的第一坐标中的X坐标的差值构成的向量;所述Y为各个所述目标指定点畸变校正后的第一坐标中的Y坐标与所述参考指定点畸变校正后的第一坐标中的Y坐标的差值构成的向量;Wherein, the designated point includes a reference designated point and a target designated point, and the A is a second coordinate of each of the target designated points in the designated coordinate system and a reference coordinate of the reference designated point in the designated coordinate system. A matrix formed by the difference of the second coordinates; the X is the difference between the X coordinate in the first coordinate after the distortion correction of the designated point of the target and the X coordinate in the first coordinate after the distortion of the reference designated point are corrected A vector formed by values; the Y is a vector formed by a difference between the Y coordinate in the first coordinate after the distortion correction of the designated point of the target and the Y coordinate in the first coordinate after the distortion of the reference designated point is corrected;
    其中,(a 1,b 1)为所述参考指定点在所述指定坐标下的第二坐标;(x 1,y 1)为所述参考指定点畸变校正后的第一坐标;所述i 1和所述i 2分别为所述i中的第一个元素和第 二个元素;所述j′ 1和所述j′ 2分别为所述j'中的第一个元素和第二个元素;所述k′ 1和所述k′ 2分别为所述k'中的第一个元素和第二个元素;所述t为所述指定物体相对于所述单目相机的平移向量。 Where (a 1 , b 1 ) are the second coordinates of the reference designated point under the designated coordinates; (x 1 , y 1 ) are the first coordinates of the reference designated point after distortion correction; and i 1 and i 2 are the first and second elements in i, respectively; the j ′ 1 and j ′ 2 are the first and second elements in j ′, respectively Element k ′ 1 and k ′ 2 are the first element and the second element of k ′, respectively; and t is a translation vector of the specified object with respect to the monocular camera.
  6. 根据权利要求1所述的方法,其特征在于,所述指定物体为用于给本设备充电的充电设备;通过单目相机获取指定物体在第一时刻的第一图像以及所述指定物体在第二时刻的第二图像,包括:The method according to claim 1, wherein the designated object is a charging device used to charge the device; a first image of the designated object at a first moment is obtained through a monocular camera, and the designated object is at the first time. The second image at two moments includes:
    在检测到本设备与所述指定物体断开连接后,通过所述单目相机获取所述指定物体在第一时刻的第一图像以及所述指定物体在第二时刻的第二图像。After detecting that the device is disconnected from the designated object, the first image of the designated object at the first moment and the second image of the designated object at the second moment are obtained through the monocular camera.
  7. 一种移动机器人,包括:A mobile robot includes:
    单目相机,用于获取指定物体在第一时刻的第一图像以及所述指定物体在第二时刻的第二图像;A monocular camera, configured to obtain a first image of a specified object at a first moment and a second image of the specified object at a second moment;
    处理器,用于:Processors for:
    对所述第一图像和所述第二图像进行特征点提取和匹配,并依据配对好的特征点计算所述单目相机从所述第一时刻到所述第二时刻的归一化平移向量;Performing feature point extraction and matching on the first image and the second image, and calculating a normalized translation vector of the monocular camera from the first moment to the second moment according to the paired feature points ;
    计算所述指定物体在所述第一时刻相对于所述单目相机的第一位姿,以及所述指定物体在所述第二时刻相对于所述单目相机的第二位姿;Calculating a first pose of the designated object relative to the monocular camera at the first moment, and a second pose of the designated object relative to the monocular camera at the second moment;
    依据所述第一位姿和所述第二位姿,计算所述单目相机从所述第一时刻到所述第二时刻的实际平移向量;Calculating an actual translation vector of the monocular camera from the first moment to the second moment according to the first pose and the second pose;
    将所述实际平移向量的模与所述归一化平移向量的模的比值确定为单目视觉重建中的尺度因子。A ratio of a modulus of the actual translation vector to a modulus of the normalized translation vector is determined as a scale factor in monocular vision reconstruction.
  8. 根据权利要求7所述的移动机器人,其特征在于,所述处理器,用于:The mobile robot according to claim 7, wherein the processor is configured to:
    针对所述第一图像和所述第二图像中的每帧图像,从该帧图像中获取所述指定物体上的指定点的像素坐标;所述指定点的数量大于或者等于4;For each frame of the first image and the second image, obtain pixel coordinates of a specified point on the specified object from the frame image; the number of the specified points is greater than or equal to 4;
    依据各个所述指定点的像素坐标,采用畸变校正算法计算各个所述指定点畸变校正后的第一坐标;Calculating, according to the pixel coordinates of each of the specified points, a first coordinate after distortion correction of each of the specified points by using a distortion correction algorithm;
    依据各个所述指定点畸变校正后的第一坐标和预存的各个所述指定点在指定坐标系下的第二坐标,计算所述指定物体相对于所述单目相机的旋转矩阵和平移向量;Calculating a rotation matrix and a translation vector of the specified object relative to the monocular camera according to the first coordinates of each of the specified point distortion corrections and the pre-stored second coordinates of each of the specified points in the specified coordinate system;
    依据所述指定物体相对于所述单目相机的所述旋转矩阵和所述平移向量,得到所述指定物体相对于所述单目相机的位姿。Obtaining the pose of the specified object relative to the monocular camera according to the rotation matrix and the translation vector of the specified object relative to the monocular camera.
  9. 根据权利要求7所述的移动机器人,其特征在于,所述处理器,具体用于:The mobile robot according to claim 7, wherein the processor is specifically configured to:
    依据所述第一位姿和所述第二位姿,计算所述单目相机从所述第一时刻到所述第二时刻的位姿变化;Calculating a change in pose of the monocular camera from the first moment to the second moment according to the first pose and the second pose;
    从所述位姿变化中获取所述实际平移向量。The actual translation vector is obtained from the pose change.
  10. 一种计算机可读存储介质,其上存储有计算机程序,其特征在于,所述程序被处理器执行实现权利要求1-6任一项所述方法的步骤。A computer-readable storage medium having stored thereon a computer program, characterized in that the program is executed by a processor to implement the steps of the method according to any one of claims 1-6.
PCT/CN2019/101704 2018-08-22 2019-08-21 Determination of scale factor in monocular vision-based reconstruction WO2020038386A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201810961434.6 2018-08-22
CN201810961434.6A CN110858403B (en) 2018-08-22 2018-08-22 Method for determining scale factor in monocular vision reconstruction and mobile robot

Publications (1)

Publication Number Publication Date
WO2020038386A1 true WO2020038386A1 (en) 2020-02-27

Family

ID=69593088

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2019/101704 WO2020038386A1 (en) 2018-08-22 2019-08-21 Determination of scale factor in monocular vision-based reconstruction

Country Status (2)

Country Link
CN (1) CN110858403B (en)
WO (1) WO2020038386A1 (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111260538A (en) * 2018-12-03 2020-06-09 北京初速度科技有限公司 Positioning and vehicle-mounted terminal based on long-baseline binocular fisheye camera
CN112102406A (en) * 2020-09-09 2020-12-18 东软睿驰汽车技术(沈阳)有限公司 Monocular vision scale correction method and device and delivery vehicle
CN112686950A (en) * 2020-12-04 2021-04-20 深圳市优必选科技股份有限公司 Pose estimation method and device, terminal equipment and computer readable storage medium
CN114406985A (en) * 2021-10-18 2022-04-29 苏州迪凯尔医疗科技有限公司 Target tracking mechanical arm method, system, equipment and storage medium

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113554703B (en) * 2020-04-23 2024-03-01 北京京东乾石科技有限公司 Robot positioning method, apparatus, system and computer readable storage medium
CN111671360B (en) * 2020-05-26 2021-11-16 深圳拓邦股份有限公司 Sweeping robot position calculating method and device and sweeping robot
CN112798812B (en) * 2020-12-30 2023-09-26 中山联合汽车技术有限公司 Target speed measuring method based on monocular vision
CN113126117B (en) * 2021-04-15 2021-08-27 湖北亿咖通科技有限公司 Method for determining absolute scale of SFM map and electronic equipment
CN116704047B (en) * 2023-08-01 2023-10-27 安徽云森物联网科技有限公司 Pedestrian ReID-based calibration method for monitoring camera equipment position

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101706957A (en) * 2009-10-30 2010-05-12 无锡景象数字技术有限公司 Self-calibration method for binocular stereo vision device
WO2017114507A1 (en) * 2015-12-31 2017-07-06 清华大学 Method and device for image positioning based on ray model three-dimensional reconstruction
CN108010125A (en) * 2017-12-28 2018-05-08 中国科学院西安光学精密机械研究所 True scale three-dimensional reconstruction system and method based on line-structured light and image information
CN108090435A (en) * 2017-12-13 2018-05-29 深圳市航盛电子股份有限公司 One kind can parking area recognition methods, system and medium

Family Cites Families (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103234454B (en) * 2013-04-23 2016-03-30 合肥米克光电技术有限公司 A kind of self-calibrating method of image measurer
CN103278138B (en) * 2013-05-03 2015-05-06 中国科学院自动化研究所 Method for measuring three-dimensional position and posture of thin component with complex structure
CN104346829A (en) * 2013-07-29 2015-02-11 中国农业机械化科学研究院 Three-dimensional color reconstruction system and method based on PMD (photonic mixer device) cameras and photographing head
CN104732518B (en) * 2015-01-19 2017-09-01 北京工业大学 A kind of PTAM improved methods based on intelligent robot terrain surface specifications
CN105118055B (en) * 2015-08-11 2017-12-15 北京电影学院 Camera position amendment scaling method and system
CN105931222B (en) * 2016-04-13 2018-11-02 成都信息工程大学 The method for realizing high-precision camera calibration with low precision two dimensional surface target
CN106529538A (en) * 2016-11-24 2017-03-22 腾讯科技(深圳)有限公司 Method and device for positioning aircraft
CN106920259B (en) * 2017-02-28 2019-12-06 武汉工程大学 positioning method and system

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101706957A (en) * 2009-10-30 2010-05-12 无锡景象数字技术有限公司 Self-calibration method for binocular stereo vision device
WO2017114507A1 (en) * 2015-12-31 2017-07-06 清华大学 Method and device for image positioning based on ray model three-dimensional reconstruction
CN108090435A (en) * 2017-12-13 2018-05-29 深圳市航盛电子股份有限公司 One kind can parking area recognition methods, system and medium
CN108010125A (en) * 2017-12-28 2018-05-08 中国科学院西安光学精密机械研究所 True scale three-dimensional reconstruction system and method based on line-structured light and image information

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111260538A (en) * 2018-12-03 2020-06-09 北京初速度科技有限公司 Positioning and vehicle-mounted terminal based on long-baseline binocular fisheye camera
CN111260538B (en) * 2018-12-03 2023-10-03 北京魔门塔科技有限公司 Positioning and vehicle-mounted terminal based on long-baseline binocular fisheye camera
CN112102406A (en) * 2020-09-09 2020-12-18 东软睿驰汽车技术(沈阳)有限公司 Monocular vision scale correction method and device and delivery vehicle
CN112686950A (en) * 2020-12-04 2021-04-20 深圳市优必选科技股份有限公司 Pose estimation method and device, terminal equipment and computer readable storage medium
CN112686950B (en) * 2020-12-04 2023-12-15 深圳市优必选科技股份有限公司 Pose estimation method, pose estimation device, terminal equipment and computer readable storage medium
CN114406985A (en) * 2021-10-18 2022-04-29 苏州迪凯尔医疗科技有限公司 Target tracking mechanical arm method, system, equipment and storage medium
CN114406985B (en) * 2021-10-18 2024-04-12 苏州迪凯尔医疗科技有限公司 Mechanical arm method, system, equipment and storage medium for target tracking

Also Published As

Publication number Publication date
CN110858403B (en) 2022-09-27
CN110858403A (en) 2020-03-03

Similar Documents

Publication Publication Date Title
WO2020038386A1 (en) Determination of scale factor in monocular vision-based reconstruction
US9420265B2 (en) Tracking poses of 3D camera using points and planes
JP5832341B2 (en) Movie processing apparatus, movie processing method, and movie processing program
US10636151B2 (en) Method for estimating the speed of movement of a camera
WO2018076154A1 (en) Spatial positioning calibration of fisheye camera-based panoramic video generating method
WO2020063708A1 (en) Method, device and system for calibrating intrinsic parameters of fisheye camera, calibration device controller and calibration tool
CN112767542A (en) Three-dimensional reconstruction method of multi-view camera, VR camera and panoramic camera
WO2021004416A1 (en) Method and apparatus for establishing beacon map on basis of visual beacons
US11082633B2 (en) Method of estimating the speed of displacement of a camera
CN108776976B (en) Method, system and storage medium for simultaneously positioning and establishing image
CN106530358A (en) Method for calibrating PTZ camera by using only two scene images
CN106228538A (en) Binocular vision indoor orientation method based on logo
CN110490943B (en) Rapid and accurate calibration method and system of 4D holographic capture system and storage medium
CN111062966B (en) Method for optimizing camera tracking based on L-M algorithm and polynomial interpolation
WO2020024684A1 (en) Method and device for modeling three-dimensional scene, electronic device, readable storage medium, and computer apparatus
JP4109075B2 (en) Method for measuring the rotation characteristics and flight characteristics of a sphere, and a device for measuring the rotation characteristics and flight characteristics of a sphere
CN110567441A (en) Particle filter-based positioning method, positioning device, mapping and positioning method
JP6922348B2 (en) Information processing equipment, methods, and programs
KR20190130407A (en) Apparatus and method for omni-directional camera calibration
JP4109076B2 (en) Measuring method of rotation amount and rotation axis direction of curved surface body, and measurement device of rotation amount and rotation axis direction of curved surface body
CN113034347B (en) Oblique photography image processing method, device, processing equipment and storage medium
JP2002109518A (en) Three-dimensional shape restoring method and system therefor
JP4886661B2 (en) Camera parameter estimation apparatus and camera parameter estimation program
CN116152121A (en) Curved surface screen generating method and correcting method based on distortion parameters
WO2018072087A1 (en) Method for realizing effect of photo being taken by others through selfie, and photographing device

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19852178

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 19852178

Country of ref document: EP

Kind code of ref document: A1

122 Ep: pct application non-entry in european phase

Ref document number: 19852178

Country of ref document: EP

Kind code of ref document: A1

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 15.09.2021)

122 Ep: pct application non-entry in european phase

Ref document number: 19852178

Country of ref document: EP

Kind code of ref document: A1