WO2022077296A1 - Three-dimensional reconstruction method, gimbal load, removable platform and computer-readable storage medium - Google Patents

Three-dimensional reconstruction method, gimbal load, removable platform and computer-readable storage medium Download PDF

Info

Publication number
WO2022077296A1
WO2022077296A1 PCT/CN2020/120978 CN2020120978W WO2022077296A1 WO 2022077296 A1 WO2022077296 A1 WO 2022077296A1 CN 2020120978 W CN2020120978 W CN 2020120978W WO 2022077296 A1 WO2022077296 A1 WO 2022077296A1
Authority
WO
WIPO (PCT)
Prior art keywords
pose
dimensional point
image
acquisition device
point clouds
Prior art date
Application number
PCT/CN2020/120978
Other languages
French (fr)
Chinese (zh)
Inventor
徐骥飞
杜劼熹
上官政和
Original Assignee
深圳市大疆创新科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 深圳市大疆创新科技有限公司 filed Critical 深圳市大疆创新科技有限公司
Priority to PCT/CN2020/120978 priority Critical patent/WO2022077296A1/en
Publication of WO2022077296A1 publication Critical patent/WO2022077296A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • G06T17/05Geographic models
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/30Determination of transform parameters for the alignment of images, i.e. image registration
    • G06T7/38Registration of image sequences

Definitions

  • the present application relates to the technical field of three-dimensional reconstruction, and in particular, to a three-dimensional reconstruction method, a PTZ load, a movable platform, and a computer-readable storage medium.
  • Simultaneous Localization and Mapping is a technology for robots to estimate their own motion and build a map of the surrounding environment in an unknown environment. It has a wide range of applications in drones, autonomous driving, mobile robot navigation, virtual reality and augmented reality.
  • the positioning under the global map can be achieved through high-precision GPS signals and prior maps.
  • the SLAM method is usually implemented based on the camera. This method uses images for 3D reconstruction, and uses the 3D reconstructed map to perceive its own motion and the surrounding environment, but from a single image Therefore, this method has the problems of low positioning accuracy and low quality of 3D model construction.
  • one of the objectives of the present application is to provide a three-dimensional reconstruction method, a pan-tilt load, a movable platform, and a computer-readable storage medium.
  • an embodiment of the present application provides a three-dimensional reconstruction method, which is applied to a pan-tilt load, and the pan-tilt payload is provided with a first image acquisition device and a second image acquisition device, and the method includes:
  • the motion data includes: during the acquisition of the three-dimensional point cloud and image sequence of the target scene, the pose information of the load on the PTZ;
  • the three-dimensional point cloud is registered according to the motion data to obtain the relative pose between the three-dimensional point clouds; and the image sequence is registered to obtain the relative pose between the images;
  • the relative pose between the three-dimensional point clouds, and the relative pose between the images obtain the pose and/or the first image acquisition device when acquiring the three-dimensional point cloud. 2. the pose when the image acquisition device acquires the image;
  • a 3D model of the target scene is generated according to the 3D point cloud, the sequence of images, and the pose when the first image acquisition device acquires the 3D point cloud and/or the pose when the second image acquisition device acquires the image .
  • an embodiment of the present application provides a pan-tilt load, including a first image acquisition device, a second image acquisition device, an inertial measurement unit, a memory for storing executable instructions, and a processor;
  • the first image acquisition device is used to acquire a three-dimensional point cloud of the target scene
  • the second image acquisition device is used for acquiring the image sequence of the target scene
  • the inertial measurement unit is used to obtain motion data of the gimbal load;
  • the motion data includes the pose information of the gimbal load;
  • the processor executes the executable instructions, it is configured to:
  • the three-dimensional point cloud is registered according to the motion data to obtain the relative pose between the three-dimensional point clouds; and the image sequence is registered to obtain the relative pose between the images;
  • the relative pose between the three-dimensional point clouds, and the relative pose between the images obtain the pose and/or the first image acquisition device when acquiring the three-dimensional point cloud. 2. the pose when the image acquisition device acquires the image;
  • a 3D model of the target scene is generated according to the 3D point cloud, the sequence of images, and the pose when the first image acquisition device acquires the 3D point cloud and/or the pose when the second image acquisition device acquires the image .
  • an embodiment of the present application provides a movable platform, including the PTZ load described in the second aspect.
  • an embodiment of the present application provides a computer-readable storage medium, where the computer-readable storage medium stores computer-executable instructions, and when the computer-executable instructions are executed by a processor, the implementation is as described in the first aspect Methods.
  • a three-dimensional reconstruction method, a pan-tilt load, a movable platform, and a computer-readable storage medium acquire a three-dimensional point cloud and an image sequence of a target scene by adopting a multi-sensor method;
  • the motion data of the gimbal load is obtained, and the motion data corresponds to the pose information of the gimbal load;
  • the three-dimensional point cloud is registered according to the motion data , obtain the relative pose between the three-dimensional point clouds; and, register the image sequence to obtain the relative pose between the images; then play according to the motion data, the relative position between the three-dimensional point clouds pose and the relative pose between the images, obtain the pose when the first image acquisition device acquires the three-dimensional point cloud and/or the pose when the second image acquisition device acquires the image; finally, according to the three-dimensional point cloud
  • the three-dimensional point cloud can indicate the three-dimensional information of the target scene
  • the The motion data includes the pose information of the pan-tilt load
  • the first image acquisition device and the second image acquisition device are arranged on the pan-tilt payload, so the motion data can indirectly indicate the first image acquisition
  • the pose information of the device and the second image acquisition device is obtained, thereby obtaining high-precision pose information, so that the three-dimensional model generated based on the high-precision pose information has higher positioning accuracy and robustness.
  • FIG. 1 is a schematic structural diagram of a pan-tilt load provided by an embodiment of the present application.
  • FIG. 2 is an application scenario diagram provided by an embodiment of the present application
  • FIG. 3 is a schematic flowchart of a three-dimensional reconstruction method provided by an embodiment of the present application.
  • FIGS. 4 , 5 , 6 and 7 are schematic diagrams of different structures of a pose graph model provided by an embodiment of the present application.
  • FIG. 8 is a schematic structural diagram of another pan-tilt load provided by an embodiment of the present application.
  • Simultaneous Localization and Mapping is a technology for robots to estimate their own motion and build a map of the surrounding environment in an unknown environment. It has a wide range of applications in drones, autonomous driving, mobile robot navigation, virtual reality and augmented reality.
  • the positioning under the global map can be achieved through high-precision GPS signals and prior maps.
  • the SLAM method is usually implemented based on the camera. This method uses images for 3D reconstruction, and uses the 3D reconstructed map to perceive its own motion and the surrounding environment, but from a single image The information obtained is limited, therefore, this method has the problems of low positioning accuracy and low quality of 3D model construction.
  • the embodiments of the present application provide a three-dimensional reconstruction method and a gimbal load.
  • the gimbal load integrates a plurality of different sensors, and obtains a three-dimensional point cloud and an image sequence of a target scene by adopting a multi-sensor method; and , during the acquisition of the three-dimensional point cloud and the image sequence, acquire motion data of the gimbal load, where the motion data includes the pose information of the gimbal load;
  • the point clouds are registered to obtain the relative pose between the three-dimensional point clouds; and the image sequence is registered to obtain the relative pose between the images; then according to the motion data, the three-dimensional point cloud and the relative pose between the images, obtain the pose when the first image acquisition device acquires the three-dimensional point cloud and/or the second image acquisition device when the image is acquired; finally
  • a 3D model of the target scene is generated according to the 3D point cloud, the sequence of images, and the pose when the first image acquisition device acquires the 3D point cloud and/or the pose when the second image
  • the three-dimensional reconstruction method can be applied to a pan-tilt load, where a plurality of different sensors are installed.
  • FIG. 1 which provides a schematic structural diagram of a pan-tilt load.
  • the first image acquisition device 21 is used to acquire the three-dimensional point cloud of the target scene in real time
  • the second image acquisition device 22 is used for to acquire the image sequence of the target scene in real time
  • the inertial measurement unit 23 is used to acquire the motion data of the PTZ load in real time
  • the PTZ load can be based on the real-time acquisition of the three-dimensional point cloud, the image
  • the sequence and the motion data generate a three-dimensional model in real time, so that the real-time requirements can be met while the accuracy requirements are met, so that subsequent positioning or measurement can be performed directly based on the real-time generated three-dimensional model.
  • the inertial measurement unit is used to measure the three-axis attitude angle (or angular rate) and acceleration of the gimbal load.
  • an inertial measurement unit includes three single-axis accelerometers and three single-axis gyroscopes.
  • the accelerometer detects the acceleration signal of the gimbal load
  • the gyro detects the angular velocity signal of the gimbal load, and measures the angular velocity signal of the gimbal load.
  • the angular velocity and acceleration of the gimbal load in the three-dimensional space are calculated, and the attitude of the gimbal load is calculated.
  • the first image acquisition device may include at least one or more of the following: lidar, binocular vision sensor, and structured light depth camera.
  • the lidar is used to transmit a laser pulse sequence to the target scene, then receive the laser pulse sequence reflected from the target, and generate a three-dimensional point cloud according to the reflected laser pulse sequence.
  • the lidar can determine the reception time of the reflected laser pulse sequence, eg, by detecting the rising edge time and/or the falling edge time of the electrical signal pulse to determine the reception time of the laser pulse sequence.
  • the laser radar can calculate TOF (Time of flight, time of flight) by using the receiving time information and the transmitting time of the laser pulse sequence, so as to determine the distance from the detected object to the laser radar.
  • the lidar is a self-illuminating sensor, does not depend on light source illumination, is less disturbed by ambient light, and can work normally even in a closed environment without light, so that a high-precision three-dimensional model can be generated later, and it has a wide range of applicability.
  • the binocular vision sensor obtains two images of the target scene from different positions based on the principle of parallax, and obtains three-dimensional geometric information by calculating the positional deviation between corresponding points of the two images, thereby generating a three-dimensional point cloud.
  • the binocular vision sensor has low hardware requirements, and correspondingly, it can also reduce the cost. It only needs to be an ordinary CMOS (Complementary Metal Oxide Semiconductor) camera. As long as the light is suitable, both indoor and outdoor environments can be used. use, so it also has certain applicability.
  • CMOS Complementary Metal Oxide Semiconductor
  • the structured light depth camera projects light with certain structural characteristics into the target scene and then collects it. This kind of light with certain structure will collect different image phase information due to different depth regions of the subject, and then It is converted into depth information to obtain a 3D point cloud.
  • Structured light depth camera is also a self-illuminating sensor, does not depend on light source illumination, and is less disturbed by ambient light, and can work normally even in a closed environment without light, so as to generate high-precision 3D models later, which has a wide range of applicability .
  • the second image acquisition device can acquire color images, grayscale images, infrared images, and the like.
  • the second image acquisition device includes at least one or more of the following: a visible light camera, a grayscale camera, and an infrared camera.
  • the second image acquisition device may capture a sequence of images at a specified frame rate.
  • the sequence of images may be captured at a standard frame rate such as about 24p, 25p, 30p, 48p, 50p, 60p, 72p, 90p, 100p, 120p, 300p, 50i or 60i.
  • Capture image sequences at frame rates of one image per second, 2 seconds, 5 seconds, or 10 seconds.
  • the second image capture device may have adjustable capture parameters. Under different capture parameters, the second image capture device may capture different images despite being subjected to exactly the same external conditions (eg location, lighting). Capture parameters may include exposure (eg, exposure time, shutter speed, aperture, film speed), gain, gamma, region of interest, binning/subsampling, pixel clock, offset, trigger, ISO, and the like. Exposure-related parameters may control the amount of light reaching the image sensor in the second image capture device. For example, shutter speed can control the amount of time light reaches the image sensor and aperture can control the amount of light that reaches the image sensor in a given time. A gain-related parameter can control the amplification of the signal from the optical sensor. ISO controls the level of sensitivity of the camera to the available light.
  • Exposure-related parameters may control the amount of light reaching the image sensor in the second image capture device. For example, shutter speed can control the amount of time light reaches the image sensor and aperture can control the amount of light that reaches the image sensor in a given time.
  • a lidar, a visible light camera and an inertial measurement unit are installed on the pan/tilt load.
  • the lidar, the visible light camera, and the inertial measurement unit may operate at the same frame rate.
  • the lidar, the visible light camera and the inertial measurement unit may also work at different frame rates, the frame rates of the lidar and the visible light camera and the inertial measurement unit.
  • the pan-tilt payload can be mounted on a movable platform, and the pan-tilt payload can be carried by the movable platform to move in a target scene, so as to perform three-dimensional reconstruction for the target scene, which is convenient for follow-up.
  • the three-dimensional model obtained by the three-dimensional reconstruction can be used for positioning tasks, mapping tasks, or navigation tasks, and the like.
  • the movable platform includes, but is not limited to, an unmanned aerial vehicle, an unmanned vehicle, or a mobile robot.
  • the PTZ load can be connected to the remote control terminal in communication, and the user can control the PTZ load through the remote control terminal to collect data in real time and perform 3D reconstruction in real time, and transmit the 3D model obtained by the 3D reconstruction to the remote control terminal in real time.
  • the remote control terminal is displayed on the remote control terminal, so that the user can know the progress of the three-dimensional reconstruction in real time, which is convenient for the user to use; further, the user can also input feedback information for the three-dimensional model displayed on the remote control terminal.
  • the feedback information is used to indicate the position where the first image acquisition device missed the acquisition, and the remote control terminal transmits the feedback information to the PTZ load, so that the PTZ load can control the first image acquisition device
  • the 3D point cloud at the position indicated by the feedback information in the target scene is collected to realize the online monitoring process, to ensure accurate and effective acquisition of 3D point cloud data, and to improve the efficiency of 3D reconstruction.
  • the gimbal load 20 can be mounted on the unmanned aerial vehicle 10, and the gimbal load 20 and the unmanned aerial vehicle 10 are connected in communication with the remote control terminal 30, The user can control the UAV 10 and the gimbal load 20 through the remote control terminal 30.
  • the remote control terminal 30 can respond to the user's trigger operation and send a three-dimensional reconstruction instruction to the gimbal load 20, so the
  • the pan-tilt load 20 is installed with a first image acquisition device, a second image acquisition device and an inertial measurement unit; the first image acquisition device acquires the three-dimensional point cloud of the target scene in real time in response to the three-dimensional reconstruction instruction;
  • the second image acquisition device acquires the image sequence of the target scene in real time in response to the three-dimensional reconstruction instruction;
  • the inertial measurement unit acquires the motion data of the pan-tilt load 20 in real time in response to the three-dimensional reconstruction instruction, so the
  • the motion data includes: during the acquisition of the three-dimensional point cloud and image sequence of the target scene, the pose information of the gimbal load; then the gimbal load 20 can be based on the real-time acquisition of the three-dimensional point cloud, the image
  • the sequence and the motion data generate a three-dimensional model in real time, so as to meet the real-time
  • the three-dimensional model is described so that the user can know the progress of the three-dimensional reconstruction in real time, which is convenient for the user to use; the user can also control the flight of the unmanned aerial vehicle 10 through the remote control terminal 30, so that the unmanned aerial vehicle 10 can carry the gimbal load 20 and fly to the desired location.
  • different positions of the target scene so that the pan-tilt load 20 can perform three-dimensional reconstruction on different positions of the target scene, and obtain a complete three-dimensional map in the target scene, so that the unmanned aerial vehicle can be based on the 3D map for positioning tasks, measurement tasks or cruise tasks, etc.
  • FIG. 3 is a schematic flowchart of a three-dimensional reconstruction method provided by an embodiment of the present application.
  • the method is applied to a PTZ load, and the method includes:
  • step S101 a 3D point cloud of a target scene, an image sequence, and motion data loaded by the gimbal are acquired; wherein, the 3D point cloud is acquired by the first image acquisition device, and the image sequence is acquired by the first image acquisition device.
  • the motion data includes: during the acquisition of a three-dimensional point cloud and an image sequence of a target scene, the pose information of the PTZ load.
  • step S102 the three-dimensional point cloud is registered according to the motion data to obtain the relative pose between the three-dimensional point clouds; and the image sequence is registered to obtain the relative pose between the images .
  • step S103 according to the motion data, the relative pose between the three-dimensional point clouds, and the relative pose between the images, the pose and the pose when the first image acquisition device acquires the three-dimensional point cloud is acquired /or the pose when the second image acquisition device acquires an image.
  • step S104 according to the three-dimensional point cloud, the image sequence, and the pose when the first image acquisition device acquires the three-dimensional point cloud and/or the pose when the second image acquisition device acquires the image, generate 3D model of the target scene.
  • the target scene may be a scene where the satellite positioning signal is lower than a predetermined intensity
  • the target scene may be an indoor scene or an outdoor non-open scene (such as underwater, a mine, etc.), in this kind of scene, you can use
  • the three-dimensional reconstruction method of this embodiment generates a high-precision three-dimensional model.
  • the satellite positioning signals include, but are not limited to, GPS signals, signals from the Galileo Satellite Navigation System (GALILEO), or signals from the Beidou Satellite Navigation System (BDS).
  • GALILEO Galileo Satellite Navigation System
  • BDS Beidou Satellite Navigation System
  • the strength of the satellite positioning signal may be determined by at least one of the following parameters: the signal-to-noise ratio of the satellite positioning signal, the cold start time or the warm start time of the satellite positioning signal; in an exemplary embodiment , if the signal-to-noise ratio of the satellite positioning signal is lower than the preset threshold, it indicates that the strength of the satellite positioning signal is lower than the preset strength; in another exemplary embodiment, if the signal-to-noise ratio of the satellite positioning signal If the ratio is lower than the preset threshold and the cold start time of the satellite positioning signal does not meet the preset time condition, it indicates that the strength of the satellite positioning signal is lower than the preset strength.
  • the pan-tilt load can use a depth camera to acquire a 3D point cloud of the target scene in real time, and use a second image acquisition device to acquire an image sequence of the target scene in real time; and during acquisition of the 3D point cloud and the image sequence, Using an inertial measurement unit to collect motion data of the gimbal load in real time, the motion data corresponds to the pose information of the gimbal load.
  • This embodiment uses the three-dimensional point cloud, the image sequence, and the motion data acquired in real time to perform real-time three-dimensional reconstruction.
  • the motion data can indicate The pose information of the pan-tilt load, the first image acquisition device and the second image acquisition device are set on the pan-tilt payload, so the motion data can indirectly indicate the first image acquisition device and the The pose information of the second image acquisition device is obtained to obtain high-precision pose information, so that the three-dimensional model generated based on the high-precision pose information has higher positioning accuracy and robustness; and real-time acquisition and real-time three-dimensional reconstruction The process is also enough to meet the needs of the scene with real-time requirements, so that the subsequent positioning or measurement can be directly based on the real-time generated 3D model.
  • the pan-tilt load can be connected to the remote control terminal in communication, and the pan-tilt load can generate a collection trajectory according to the accuracy requirements and/or density requirements of the three-dimensional model input by the user on the remote control terminal. and acquisition time, and then control the first image acquisition device, the second image acquisition device and the inertial measurement unit to acquire data according to the acquisition track and the acquisition time.
  • This embodiment implements automatic planning of the acquisition scheme based on the accuracy requirements and/or density requirements of the three-dimensional model, and controls the first image acquisition device, the second image acquisition device, and the inertial measurement unit to acquire data. There is no need to participate in the process, which can effectively prevent human errors introduced in the acquisition process and improve the accuracy of the generated 3D model.
  • the PTZ load After acquiring the 3D point cloud, the image sequence and the motion data, the PTZ load registers the 3D point cloud according to the motion data, and obtains the relative pose between the 3D point clouds, The relative poses between the three-dimensional point clouds are used to calibrate the three-dimensional point clouds, thereby improving the accuracy of subsequent three-dimensional reconstruction results; The relative pose between the images is used to determine the pose of the second image acquisition device, thereby improving the accuracy of subsequent three-dimensional reconstruction results.
  • each point in the 3D point cloud has an independent coordinate system, so as to ensure the accuracy of the 3D coordinates of the point in the corresponding coordinate system.
  • the 3D point cloud when used for 3D reconstruction in the future, the points in the 3D point cloud need to be in the same coordinate system, so the 3D point cloud needs to be reprojected, the motion of the lidar during the acquisition process is calculated, and the This amount of movement is compensated on the corresponding laser point.
  • the three-dimensional point cloud is transformed by the motion data obtained by the inertial measurement unit installed on the same pan/tilt load as the first image acquisition device, and the motion data obtained by the inertial measurement unit corresponds to the cloud
  • the motion data acquired by the inertial measurement unit also indicates the motion process of the first image acquisition device, so the motion data measured by the inertial measurement unit can be used to accurately determine the transformation between points in different coordinate systems
  • the transformation relationship between the points can be obtained by using the motion data and the external parameter transformation relationship between the inertial measurement unit and the first image acquisition device.
  • the PTZ load obtains the transformation relationship between points according to the motion data corresponding to each point in the group of 3D point clouds, and then according to the relationship between the points The transformation relationship reprojects the points in the set of 3D point clouds to the same coordinate system.
  • a group of three-dimensional point clouds collected in a continuous time series are transformed into the same coordinate system, so as to realize the integration of three-dimensional point cloud data, which is beneficial to improve the efficiency of three-dimensional reconstruction.
  • the PTZ load acquires the motion data corresponding to the point according to the acquisition time of each point;
  • the first image acquisition device is a laser radar, and the acquisition time of the point is based on the reflected laser pulses Determined by the receiving time of the sequence, the acquisition time of the point is the acquisition time of the corresponding motion data, and the gimbal load acquires the motion data from the inertial measurement unit at the same moment when the point is acquired.
  • the first image acquisition device will acquire each point at a preset time interval, and when acquiring the transformation relationship between points, the corresponding motion of each point in the set of point clouds can be obtained.
  • the data determines the relative pose of the inertial measurement unit within the preset time interval, wherein the preset time interval is the time interval for acquiring adjacent points; then the gimbal load is based on the inertial measurement unit at the time interval.
  • the relative pose within a preset time interval, and the extrinsic parameter transformation relationship between the first image acquisition device and the inertial measurement unit obtain the transformation relationship between the points.
  • the motion data acquired by the inertial measurement unit also indicates the movement process of the first image acquisition device, so it can be Using the motion data measured by the inertial measurement unit, the transformation relationship between points in different coordinate systems can be accurately determined, so that a set of three-dimensional point clouds collected in a continuous time series can be transformed based on the transformation relationship between the points. Under the same coordinate system, the integration of 3D point cloud data is realized, which is beneficial to improve the efficiency of 3D reconstruction.
  • the gimbal load can obtain the relative poses between the adjacent two groups of 3D point clouds, and the relative poses between the adjacent two groups of 3D point clouds are used to
  • the two sets of 3D point clouds are reprojected to the same coordinate system to realize the matching and integration of 3D point cloud data, thus facilitating the subsequent 3D reconstruction process.
  • the adjacent two sets of three-dimensional point clouds include points collected for the same object, so as to ensure the smooth matching of the two adjacent three-dimensional point clouds.
  • the PTZ load acquires the surface features and/or corner features of each group of 3D point clouds. Then, according to the surface features and/or corner features of the adjacent two groups of three-dimensional point clouds, the relative poses between the two adjacent groups of three-dimensional point clouds are obtained; the relative poses between the two adjacent groups of three-dimensional point clouds are used for Reproject two adjacent sets of 3D point clouds to the same coordinate system.
  • This embodiment uses the surface features and/or corner features in the 3D point cloud acquisition space to perform high-speed matching, simplifies the matching process, helps improve the efficiency of acquiring the relative poses between two adjacent groups of 3D point clouds, and shortens the matching time. .
  • ICP Iterative Closest Point/Plane
  • the surface features and/or corner features of the three-dimensional point clouds of the group may be determined according to the curvature information between the midpoints of the three-dimensional point clouds of each group.
  • the surface features of each group of 3D point clouds are determined according to curvature information whose curvature is greater than a preset threshold; the corner features of each group of 3D point clouds are determined according to curvature information whose curvature is less than or equal to a preset threshold determined.
  • the matching process between points is converted into a matching process between surface features and/or corner features of a three-dimensional point cloud in space, which is beneficial to simplify the matching process and realize the two adjacent sets of three-dimensional point clouds.
  • the high-speed matching of features and/or corner features is further conducive to improving the efficiency of obtaining relative poses between two adjacent sets of 3D point clouds, shortening the matching time, and meeting the needs of real-time generation of 3D models.
  • the surface feature includes at least one three-dimensional coordinate information indicating a plane and a normal vector of the plane; and/or the corner feature includes at least one three-dimensional coordinate information indicating an edge and a vector indicating the edge.
  • the pan-tilt load may be based on the adjacent two groups of three-dimensional point clouds.
  • the more similar the features are, and/or the more similar the corner features of the adjacent two sets of 3D point clouds are, indicating that the points in the adjacent two sets of 3D point clouds indicated by the surface feature and/or the corner feature may be matched.
  • the relative poses between the adjacent two groups of three-dimensional point clouds are obtained through the similarity between the surface features and/or the similarity between the corner features, so as to ensure the accuracy of the obtained results.
  • the similarity between the surface features of the adjacent two groups of three-dimensional point clouds is determined based on the distance between the surface features of the two adjacent groups of three-dimensional point clouds; and/or, the The similarity between the corner features of the adjacent two groups of three-dimensional point clouds is determined based on the distance between the corner features of the two adjacent groups of three-dimensional point clouds. Wherein, and/or represent both or one of the two.
  • the optimization algorithm can be used to solve: in, represents the distance difference between the surface features of the adjacent two groups of three-dimensional point clouds, Represents the distance difference between the corner features of the adjacent two groups of three-dimensional point clouds.
  • the relative poses between two adjacent groups of 3D point clouds can be obtained through the transformation and expression of LieAlgebra algorithm.
  • This embodiment uses the surface feature and the corner feature to perform high-speed matching, which is beneficial to improve the matching efficiency and meet the requirement of generating a three-dimensional model in real time.
  • the image sequence includes a plurality of images acquired in a continuous time sequence.
  • the PTZ payload can extract feature points of each of the images, and then perform registration based on the feature points of the images to obtain the relative poses between the images.
  • the PTZ load may use a bundle adjustment (Bundle Adjustment, BA) algorithm to register the feature points of the images, and obtain the relative poses between the images.
  • BA Bunle Adjustment
  • the depth map obtained by the back-projection of the 3D point cloud can be used to assist the registration of the image process to improve image registration accuracy.
  • the PTZ load can reproject the 3D point cloud by using the motion data, and generate a depth map corresponding to the image according to the reprojected 3D point cloud and the motion data; It is mentioned that each point in the 3D point cloud has independent coordinates.
  • each point in the 3D point cloud needs to be in the same coordinate system, so the PTZ load uses The motion data reprojects the 3D point cloud, so that each point in the reprojected 3D point cloud is in the same coordinate system; further, the motion data obtained by the inertial measurement unit corresponds to the PTZ
  • the pose information of the load since the inertial measurement unit, the second image acquisition device and the first image acquisition device are all installed on the PTZ load, the second image acquisition device and the first image acquisition device are The motion process is the same as or similar to the motion process of the PTZ load, so the motion data acquired by the inertial measurement unit can also indicate the motion process of the second image acquisition device and the second image acquisition device, Then, a depth map corresponding to the image may be generated using the reprojected 3D point cloud and the motion data.
  • a polyhedron may be generated based on the reprojected 3D point cloud, and then projected onto the camera plane according to the multi-facet map and the motion data to obtain the image corresponding to the depth map.
  • the PTZ load also extracts the feature points in the image, and obtains the depth information of the feature points from the depth map corresponding to the image; and then according to the feature points of the image and the The depth information is registered to obtain the relative pose between the images.
  • the depth map obtained by the back-projection of the three-dimensional point cloud is used to assist the registration process of the image, which is beneficial to improve the accuracy of image registration, thereby improving the accuracy of the subsequently generated three-dimensional model.
  • the pan/tilt load in the embodiment of the present application uses the motion data to reproject the three-dimensional point cloud , so that each point in the reprojected three-dimensional point cloud is in the same coordinate system, and considering that the inertial measurement unit, the second image acquisition device and the first image acquisition device are all installed on the PTZ load,
  • the motion data obtained by the inertial measurement unit can also indicate the motion process of the second image acquisition device and the second image acquisition device, and then can generate and match the motion data according to the re-projected three-dimensional point cloud and the motion data.
  • the depth map corresponding to the images in the image sequence is obtained; finally, registration is performed according to the feature points extracted from the depth map to obtain the relative pose between the images.
  • the depth map obtained by the back-projection of the three-dimensional point cloud is used to perform registration instead of the image, so as to ensure that an accurate image registration result can be obtained even in the absence of light, thereby improving the accuracy of the subsequently generated three-dimensional model.
  • a distributed accelerator in order to meet the requirement of generating a 3D model online in real time, can be used to obtain the relative posture between the 3D point clouds and the relative posture between the images, so as to speed up the matching process and improve data acquisition. efficiency.
  • the PTZ load can be based on the three-dimensional point cloud, the image sequence, the motion data, the three-dimensional point
  • the relative poses between the clouds and the relative poses between the images generate a three-dimensional model of the target scene in real time, so as to meet the requirement of generating a three-dimensional model online in real time.
  • the PTZ load can transmit the 3D model to the remote control terminal in real time after generating the 3D model, so that the 3D model can be displayed on the remote control terminal in real time, so that the user can know the progress of the 3D reconstruction in real time, which is convenient.
  • the remote control terminal uses the user; if the user finds a missing location after seeing the 3D model, he can also input feedback information for the 3D model displayed on the remote control terminal, where the feedback information is used to instruct the first image acquisition device
  • the remote control terminal transmits the feedback information to the PTZ load, so that the PTZ load can control the first image acquisition device to collect the information indicated by the feedback information in the target scene.
  • the 3D point cloud at the location realizes the online monitoring process, ensures the accurate and effective acquisition of 3D point cloud data, and is conducive to improving the efficiency of 3D reconstruction.
  • the PTZ load is obtained according to the motion data, the relative pose between the 3D point clouds and the relative pose between the images.
  • the pose when the acquisition device acquires the three-dimensional point cloud and/or the pose when the second image acquisition device acquires the image generates a three-dimensional model in real time.
  • This embodiment realizes the generation of a high-precision three-dimensional model based on the fusion of various data.
  • the gimbal load first establishes a pose graph model, and the pose graph model includes but is not limited to a sliding window pose graph model. Then, the pose graph model is optimized by using the motion data, the relative poses between the three-dimensional point clouds, and the relative poses between the images, and the first image acquisition device acquires the three-dimensional point cloud. The pose and/or the pose when the second image acquisition device acquires the image.
  • the motion data obtained by the inertial measurement unit corresponds to the pose information of the gimbal load, since the inertial measurement unit, the second image acquisition device and the first image acquisition device are all installed on the gimbal load,
  • the motion process of the second image acquisition device and the first image acquisition device is the same or similar to the motion process of the gimbal load, so the motion data acquired by the inertial measurement unit can also indicate the first image acquisition device.
  • the motion process of the second image acquisition device and the second image acquisition device therefore, the motion data can be used as one of the optimization factors for optimizing the pose graph model.
  • motion data includes acceleration, which is integrated to obtain velocity, and velocity is integrated to obtain relative displacement. The relative displacement can constrain the displacement between the two positions of the gimbal load.
  • the motion data includes the angular velocity, and the angle is obtained after the angular velocity is integrated, and then the relative attitude information is obtained.
  • the relative attitude can constrain the rotation between the two attitudes of the gimbal load.
  • each vertex is configured to indicate the pose when the first image acquisition device acquires the three-dimensional point cloud and/or the pose when the second image acquisition device acquires an image; and , the edges between the vertices are configured to indicate the relative poses between the three-dimensional point clouds, the relative poses between the images, or the motion data, and each edge constitutes the constraint relationship of each vertex, through The relative pose between the three-dimensional point clouds indicated by each edge, the relative pose between the images, or the motion data are used to optimize the position of the first image acquisition device in each vertex when the three-dimensional point cloud is collected. pose and/or the pose when the second image acquisition device acquires the image.
  • the pose graph model includes a sliding window pose graph model, and in each iterative optimization process of the window, the pose indicated by the ith (i is an integer greater than 1) vertex is selected and the pose indicated by at least one reference vertex (for example, the reference vertex can be the i-jth vertex, and j is an integer greater than 0) as the pose to be iteratively optimized in the window, using the three-dimensional point indicated by the edge between the vertices
  • the relative pose between clouds, the relative pose between the images, or the motion data is used for pose optimization.
  • the window is slid to select the pose indicated by the next vertex for optimization.
  • each vertex instructs the first image acquisition device to acquire a three-dimensional point cloud n is an integer greater than 0, and each edge indicates the relative pose between the three-dimensional point clouds and the relative pose between the images or the motion data; in another example, please refer to FIG.
  • n is an integer greater than 0
  • each edge indicates the relative pose between the three-dimensional point clouds, all The relative pose between the images or the motion data.
  • the PTZ load may also be equipped with a locator, which is used to obtain positioning information such as GPS information and/or RTK information, the locator collects positioning information during the process of collecting the three-dimensional point cloud by the first image collecting device and acquiring the image sequence by the second image collecting device, please refer to FIG.
  • the edges between the vertices can also be configured to indicate GPS information and/or RTK information, so as to obtain high-precision pose information and improve the positioning accuracy of the generated three-dimensional model.
  • the pan-tilt load can utilize the The pose of the three-dimensional point cloud reprojects the three-dimensional point cloud to the target coordinate system, for example, the target coordinate system can be the world coordinate system, and then generates an initial three-dimensional model according to the re-projected three-dimensional point cloud, and then according to the The pose when the second image acquisition device acquires the image and the multiple images in the image sequence perform texture mapping on the initial three-dimensional model to obtain the final three-dimensional model.
  • the PTZ load may add a tag index to the 3D model, and then store the 3D model in the storage location pointed to by the tag index, so that the tag index can be used in subsequent use based on the tag index.
  • the three-dimensional model is found by index; and/or, the three-dimensional model is transmitted to the remote control terminal in real time, so that the three-dimensional model is displayed on the remote control terminal in real time, so that the progress of the three-dimensional reconstruction can be displayed to the user in real time, which is convenient for the user to use.
  • the PTZ load can also perform loop closure detection according to the 3D point cloud or the image sequence, and optimize the pose graph model according to the results of the loop closure detection, thereby optimizing all
  • the 3D model is described to ensure the accuracy and global uniformity of the optimized 3D model.
  • a bag-of-words model may be constructed according to the image sequence or the depth map obtained by re-projection of the three-dimensional point cloud, so that the bag-of-words model can measure the similarity or all of the images in the image sequence.
  • the similarity between the multiple depth images, the pan-tilt load can perform loop closure detection according to the similarity between the multiple images in the image sequence, and determine the image forming a loop; or, according to the three-dimensional point cloud
  • the similarity between the multiple depth maps obtained by reprojection is subjected to loop closure detection to determine the depth map forming a loop closure, and then to determine the image corresponding to the depth map;
  • the pose graph model is globally optimized, that is, the relative poses between the looped images are formed to globally optimize the pose graph model, and the optimized pose information is obtained, that is, the optimized first image acquisition device
  • the pose when collecting the three-dimensional point cloud and/or the pose when the second image acquisition device acquires the image, and then update the three-dimensional model according to the globally optimized pose graph model, that is, update according to the optimized pose information
  • the three-dimensional model ensures the accuracy and global uniformity of the optimized three-dimensional model.
  • the down-sampled global model and the three-dimensional model stored in the first image acquisition device may be used to perform loop closure detection, to determine the relative poses between the three-dimensional point clouds forming loop closures, and then to perform loop closure detection.
  • the gimbal load can globally optimize the pose graph model by using the result of the loopback detection, that is, the relative poses between the three-dimensional point clouds forming the loopback can be used to globally optimize the pose graph model, and the optimized pose graph model can be obtained.
  • the pose information that is, the optimized pose when the first image acquisition device collects the 3D point cloud and/or the pose when the second image acquisition device acquires the image, and then updated according to the globally optimized pose graph model
  • the three-dimensional model is to update the three-dimensional model according to the optimized pose information, so as to ensure the accuracy and global unity of the optimized three-dimensional model.
  • the edges in the pose graph model are also configured to indicate the relative poses between the images forming the loops or the relative poses between the three-dimensional point clouds forming the loops, thereby
  • the pose graph model is optimized, and the three-dimensional model is updated according to the globally optimized pose graph model.
  • the PTZ load can store the updated 3D model in the storage location pointed to by the corresponding label index; and/or, transmit the updated 3D model to the remote control in real time terminal to display the 3D model in real time on the remote control terminal, so that the user can know the progress of the 3D reconstruction in real time, which is convenient for the user to use; if the user finds a missing location after seeing the 3D model, he can also target the display on the remote control terminal.
  • Feedback information is input to the 3D model on the PTZ, and the feedback information is used to indicate the position where the first image acquisition device is missing, and the remote control terminal transmits the feedback information to the PTZ load, so that the PTZ The load can control the first image acquisition device to collect the three-dimensional point cloud at the position indicated by the feedback information in the target scene, realize the online monitoring process, ensure the accurate and effective acquisition of the three-dimensional point cloud data, and help improve the three-dimensional reconstruction. efficiency.
  • an embodiment of the present application further provides a pan-tilt load 20 , including a first image acquisition device 21 , a second image acquisition device 22 , an inertial measurement unit 23 , and a memory for storing executable instructions 24 and processor 25.
  • the first image acquisition device 21 is used to acquire a three-dimensional point cloud of the target scene.
  • the second image acquisition device 22 is configured to acquire an image sequence of the target scene.
  • the inertial measurement unit 23 is configured to acquire motion data of the gimbal load; the motion data includes the pose information of the gimbal load.
  • processor 25 executes the executable instructions, it is configured to:
  • the three-dimensional point cloud is registered according to the motion data to obtain the relative pose between the three-dimensional point clouds; and the image sequence is registered to obtain the relative pose between the images;
  • the relative pose between the three-dimensional point clouds, and the relative pose between the images obtain the pose and/or the first image acquisition device when acquiring the three-dimensional point cloud. 2. the pose when the image acquisition device acquires the image;
  • a 3D model of the target scene is generated according to the 3D point cloud, the sequence of images, and the pose when the first image acquisition device acquires the 3D point cloud and/or the pose when the second image acquisition device acquires the image .
  • the processor 25 executes the executable instructions included in the memory 24, and the processor 25 may be a central processing unit (Central Processing Unit, CPU), or other general-purpose processors, digital signal processors (Digital Signal Processors) Processor, DSP), Application Specific Integrated Circuit (ASIC), Field-Programmable Gate Array (FPGA) or other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components, etc.
  • CPU Central Processing Unit
  • DSP Digital Signal Processors
  • ASIC Application Specific Integrated Circuit
  • FPGA Field-Programmable Gate Array
  • a general purpose processor may be a microprocessor or the processor may be any conventional processor or the like.
  • the memory 24 stores executable instructions for the three-dimensional reconstruction method, and the memory 24 may include at least one type of storage medium, including flash memory, hard disk, multimedia card, card-type memory (eg, SD or DX memory, etc.) , Random Access Memory (RAM), Static Random Access Memory (SRAM), Read Only Memory (ROM), Electrically Erasable Programmable Read Only Memory (EEPROM), Programmable Read Only Memory (PROM), Magnetic Memory, Disk, CD and so on. Also, the pan-tilt load 20 may cooperate with a network storage device that performs the storage function of the memory through a network connection.
  • the memory 24 may be an internal storage unit of the gimbal load 20 , such as a hard disk or a memory of the gimbal load 20 .
  • the memory 24 can also be an external storage device of the gimbal load 20, such as a pluggable hard disk, a smart memory card (Smart Media Card, SMC), a secure digital (Secure Digital, SD) card, and a flash memory card equipped on the gimbal load 20. (Flash Card) etc. Further, the memory 24 may also include both an internal storage unit of the pan-tilt load 20 and an external storage device. Memory 24 is used to store executable instructions and other programs and data required by the device. The memory 24 may also be used to temporarily store data that has been or will be output.
  • an external storage device of the gimbal load 20 such as a pluggable hard disk, a smart memory card (Smart Media Card, SMC), a secure digital (Secure Digital, SD) card, and a flash memory card equipped on the gimbal load 20. (Flash Card) etc.
  • the memory 24 may also include both an internal storage unit of the pan-tilt load 20 and an external storage device. Memory 24 is used to store
  • the various embodiments described herein can be implemented using computer readable media such as computer software, hardware, or any combination thereof.
  • the embodiments described herein can be implemented using application specific integrated circuits (ASICs), digital signal processors (DSPs), digital signal processing devices (DSPDs), programmable logic devices (PLDs), field programmable gate arrays ( FPGA), processors, controllers, microcontrollers, microprocessors, electronic units designed to perform the functions described herein are implemented.
  • ASICs application specific integrated circuits
  • DSPs digital signal processors
  • DSPDs digital signal processing devices
  • PLDs programmable logic devices
  • FPGA field programmable gate arrays
  • processors controllers, microcontrollers, microprocessors, electronic units designed to perform the functions described herein are implemented.
  • embodiments such as procedures or functions may be implemented with separate software modules that allow the performance of at least one function or operation.
  • the software codes may be implemented by a software application (or program) written in any suitable programming language, which may be stored in
  • FIG. 8 is only an example of the pan-tilt load 20, and does not constitute a limitation to the pan-tilt payload 20. It may include more or less components than the one shown, or combine some components, or different Components, such as devices, may also include input and output devices, network access devices, buses, and the like.
  • the first image acquisition device 21 is specifically configured to acquire the 3D point cloud of the target scene in real time; the second image acquisition device 22 is specifically configured to acquire the image sequence of the target scene in real time; and the The inertial measurement unit 23 is specifically configured to acquire motion data of the PTZ load in real time.
  • the target scene includes an indoor scene or an outdoor non-empty scene, and/or a scene where the satellite positioning signal is lower than a predetermined intensity.
  • the first image acquisition device 21 includes at least one or more of the following: lidar, binocular vision sensor, and structured light depth camera;
  • the second image acquisition device 22 includes at least one or more of the following: a visible light camera, a grayscale camera, and an infrared camera.
  • the processor 25 is specifically configured to: acquire the pose and/or the second image according to the three-dimensional point cloud, the image sequence, and the first image acquisition device when acquiring the three-dimensional point cloud
  • the acquisition device acquires the pose of the image, and generates a three-dimensional model of the target scene in real time.
  • the processor 25 when acquiring the relative pose between the three-dimensional point clouds, is further configured to: for a group of three-dimensional point clouds collected in a continuous time series, according to the The transformation relationship between the motion data acquisition points corresponding to each point;
  • the points in the set of three-dimensional point clouds are reprojected to the same coordinate system according to the transformation relationship between the points.
  • the processor 25 is further configured to: determine the relative pose of the inertial measurement unit 23 within a preset time interval according to the motion data corresponding to each point in the set of three-dimensional point clouds; Let the time interval be the time interval for obtaining adjacent points;
  • the motion data corresponding to each point is determined according to the acquisition time of the point.
  • the processor 25 when acquiring the relative pose between the three-dimensional point clouds, the processor 25 is further configured to:
  • For two adjacent groups of 3D point clouds obtain the surface features and/or corner features of each group of 3D point clouds; wherein, the adjacent two groups of 3D point clouds both include points collected for the same object;
  • the relative poses between the two adjacent groups of three-dimensional point clouds are obtained; the relative poses between the two adjacent groups of three-dimensional point clouds are used to The adjacent two sets of 3D point clouds are reprojected to the same coordinate system.
  • the surface features and/or corner features of each group of three-dimensional point clouds are determined according to curvature information between three-dimensional points in the group of three-dimensional point clouds.
  • the surface features of each group of three-dimensional point clouds are determined according to curvature information whose curvature is greater than a preset threshold;
  • the corner features of each group of three-dimensional point clouds are determined according to curvature information whose curvature is less than or equal to a preset threshold.
  • the processor 25 is further configured to: according to the similarity between the surface features of the adjacent two groups of three-dimensional point clouds, and/or the similarity between the corner features of the adjacent two groups of three-dimensional point clouds degrees to obtain the relative pose between 3D point clouds.
  • the similarity between the surface features of the adjacent two groups of three-dimensional point clouds is determined based on the distance between the surface features of the two adjacent groups of three-dimensional point clouds; and/or,
  • the similarity between the corner features of the adjacent two groups of three-dimensional point clouds is determined based on the distance between the corner features of the two adjacent groups of three-dimensional point clouds.
  • the surface feature includes at least one three-dimensional coordinate information indicating a plane and a normal vector of the plane; and/or,
  • the corner feature includes at least one three-dimensional coordinate information indicating an edge and a vector indicating the edge.
  • the sequence of images includes a plurality of images acquired in a continuous time sequence
  • the processor 25 is further configured to extract feature points in the images, and perform registration based on the feature points of the images to obtain the relative poses between the images .
  • the sequence of images includes a plurality of images acquired in a continuous time sequence
  • the processor 25 is further configured to:
  • the registration is performed according to the feature points of the images and the depth information of the feature points, and the relative poses between the images are acquired.
  • the processor 25 when acquiring the relative pose between the images, is further configured to: in the absence of light, use the motion data to reproject the three-dimensional point cloud, and re-project the three-dimensional point cloud according to the re-projection.
  • the projected three-dimensional point cloud and the motion data generate a depth map corresponding to the images in the image sequence; perform registration according to the feature points extracted from the depth map to obtain the relative poses between the images.
  • the processor 25 is further configured to:
  • the pose graph model is optimized by using the motion data, the relative poses between the three-dimensional point clouds, and the relative poses between the images, and obtains the data obtained when the first image acquisition device 21 obtains the three-dimensional point cloud.
  • each vertex is configured to instruct the first image acquisition device 21 to acquire a pose when acquiring a three-dimensional point cloud and/or the second image acquisition device 22 to acquire an image. posture when
  • edges between the vertices are configured to indicate relative poses between the three-dimensional point clouds, relative poses between the images, or the motion data.
  • the edges between the various vertices are further configured to indicate GPS information and/or RTK information.
  • the pose graph model includes a sliding window pose graph model.
  • the processor 25 is further configured to: after adding a label index to the 3D model, store the 3D model in the storage location pointed to by the label index; and/or, store the 3D model Real-time transmission to the remote control terminal to display the three-dimensional model in real time on the remote control terminal.
  • the processor 25 is further configured to:
  • Loop closure detection is performed according to the similarity between multiple images in the image sequence; or, loop closure detection is performed according to the similarity between multiple depth maps obtained by re-projection of the 3D point cloud;
  • the pose graph model is globally optimized using the result of loop closure detection; and the three-dimensional model is updated according to the globally optimized pose graph model.
  • the processor 25 is further configured to: perform global optimization on the pose graph model by using the relative pose between the images forming the loop or the relative pose between the points forming the loop.
  • a bag-of-words model is used to measure the similarity between multiple images in the image sequence or the similarity between the multiple depth images.
  • the processor 25 is further configured to:
  • the first image acquisition device 21 , the second image acquisition device 22 and the inertial measurement unit 23 are controlled to acquire data according to the acquisition track and acquisition time.
  • the processor 25 is further configured to:
  • Receive feedback information transmitted by the remote control terminal the feedback information is used to indicate the position of the missed acquisition by the first image acquisition device 21, and the feedback information is input by the user with respect to the three-dimensional model displayed on the remote control terminal;
  • the first image acquisition device 21 is controlled to acquire the three-dimensional point cloud at the position indicated by the feedback information in the target scene.
  • an embodiment of the present application also provides a movable platform, including the above-mentioned PTZ load.
  • the movable platform includes at least an unmanned aerial vehicle, an unmanned vehicle or a mobile robot.
  • non-transitory computer-readable storage medium such as a memory including instructions, executable by a processor of an apparatus to perform the above-described method.
  • the non-transitory computer-readable storage medium may be ROM, random access memory (RAM), CD-ROM, magnetic tape, floppy disk, optical data storage device, and the like.
  • a non-transitory computer-readable storage medium when the instructions in the storage medium are executed by the processor of the terminal, enable the terminal to execute the above method.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Geometry (AREA)
  • Software Systems (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Remote Sensing (AREA)
  • Computer Graphics (AREA)
  • Length Measuring Devices By Optical Means (AREA)
  • Processing Or Creating Images (AREA)

Abstract

A three-dimensional reconstruction method, a gimbal load (20), a removable platform and a computer-readable storage medium. The method comprises: acquiring three-dimensional point clouds of a target scene, an image sequence, and movement data of a gimbal load (20), wherein the three-dimensional point clouds are acquired by a first image collection apparatus (21), the image sequence is acquired by a second image collection apparatus (22), and the movement data comprises pose information of the gimbal load (20) during the acquisition of the three-dimensional point clouds of the target scene and the image sequence; registering the three-dimensional point clouds according to the movement data, so as to acquire a relative pose between the three-dimensional point clouds, and registering the image sequence, so as to acquire a relative pose between images; and generating a three-dimensional model of the target scene according to the three-dimensional point clouds, the image sequence, the movement data, the relative pose between the three-dimensional point clouds and the relative pose between the images. A high-precision three-dimensional model is acquired.

Description

三维重建方法、云台负载、可移动平台以及计算机可读存储介质Three-dimensional reconstruction method, pan-tilt load, movable platform, and computer-readable storage medium 技术领域technical field
本申请涉及三维重建技术领域,具体而言,涉及一种三维重建方法、云台负载、可移动平台以及计算机可读存储介质。The present application relates to the technical field of three-dimensional reconstruction, and in particular, to a three-dimensional reconstruction method, a PTZ load, a movable platform, and a computer-readable storage medium.
背景技术Background technique
同步定位与建图(Simultaneously Localization and Mapping,SLAM)是一种机器人在未知环境中估计自身运动并且建立周围环境地图的技术。它在无人机、自动驾驶、移动机器人导航、虚拟现实和增强现实等领域有着广泛的应用。Simultaneous Localization and Mapping (SLAM) is a technology for robots to estimate their own motion and build a map of the surrounding environment in an unknown environment. It has a wide range of applications in drones, autonomous driving, mobile robot navigation, virtual reality and augmented reality.
当机器人处于室外环境时,可通过高精度的GPS信号和先验地图实现全局地图下的定位。但当机器人处于GPS信号不可达的环境时,通常基于摄像头来实现SLAM方法,该方式利用图像进行三维重建,并利用三维重建后的地图来感知自身的运动和周围的环境,但从单一的图像中获取的信息有限,因此,该种方式存在定位精度低,三维模型构建质量不高的问题。When the robot is in an outdoor environment, the positioning under the global map can be achieved through high-precision GPS signals and prior maps. However, when the robot is in an environment where GPS signals cannot be reached, the SLAM method is usually implemented based on the camera. This method uses images for 3D reconstruction, and uses the 3D reconstructed map to perceive its own motion and the surrounding environment, but from a single image Therefore, this method has the problems of low positioning accuracy and low quality of 3D model construction.
发明内容SUMMARY OF THE INVENTION
有鉴于此,本申请的目的之一是提供一种三维重建方法、云台负载、可移动平台以及计算机可读存储介质。In view of this, one of the objectives of the present application is to provide a three-dimensional reconstruction method, a pan-tilt load, a movable platform, and a computer-readable storage medium.
第一方面,本申请实施例提供了一种三维重建方法,应用于云台负载上,所述云台负载设有第一图像采集装置、第二图像采集装置,所述方法包括:In a first aspect, an embodiment of the present application provides a three-dimensional reconstruction method, which is applied to a pan-tilt load, and the pan-tilt payload is provided with a first image acquisition device and a second image acquisition device, and the method includes:
获取目标场景的三维点云、图像序列以及所述云台负载的运动数据;其中,所述三维点云由所述第一图像采集装置获取,所述图像序列由所述第二图像采集装置获取;所述运动数据包括:在获取目标场景的三维点云、图像序列的期间,所述云台负载的位姿信息;;Acquiring a 3D point cloud, an image sequence, and motion data of the pan/tilt load of the target scene; wherein, the 3D point cloud is acquired by the first image acquisition device, and the image sequence is acquired by the second image acquisition device ; The motion data includes: during the acquisition of the three-dimensional point cloud and image sequence of the target scene, the pose information of the load on the PTZ;
根据所述运动数据对所述三维点云进行配准,获取三维点云之间的相对位姿;以及,对所述图像序列进行配准,获取图像之间的相对位姿;The three-dimensional point cloud is registered according to the motion data to obtain the relative pose between the three-dimensional point clouds; and the image sequence is registered to obtain the relative pose between the images;
根据所述运动数据、所述三维点云之间的相对位姿以及所述图像之间的相对位姿,获取所述第一图像采集装置获取三维点云时的位姿和/或所述第二图像采集装置获取图像时的位姿;According to the motion data, the relative pose between the three-dimensional point clouds, and the relative pose between the images, obtain the pose and/or the first image acquisition device when acquiring the three-dimensional point cloud. 2. the pose when the image acquisition device acquires the image;
根据所述三维点云、所述图像序列以及所述第一图像采集装置获取三维点云时的位姿和/或所述第二图像采集装置获取图像时的位姿,生成目标场景的三维模型。A 3D model of the target scene is generated according to the 3D point cloud, the sequence of images, and the pose when the first image acquisition device acquires the 3D point cloud and/or the pose when the second image acquisition device acquires the image .
第二方面,本申请实施例提供了一种云台负载,包括第一图像采集装置、第二图像采集装置、惯性测量单元、用于存储可执行指令的存储器以及处理器;In a second aspect, an embodiment of the present application provides a pan-tilt load, including a first image acquisition device, a second image acquisition device, an inertial measurement unit, a memory for storing executable instructions, and a processor;
所述第一图像采集装置用于获取目标场景的三维点云;The first image acquisition device is used to acquire a three-dimensional point cloud of the target scene;
所述第二图像采集装置用于获取所述目标场景的图像序列;The second image acquisition device is used for acquiring the image sequence of the target scene;
在采集所述三维点云和所述图像序列期间,所述惯性测量单元用于获取所述云台负载的运动数据;所述运动数据包括所述云台负载的位姿信息;During the collection of the three-dimensional point cloud and the image sequence, the inertial measurement unit is used to obtain motion data of the gimbal load; the motion data includes the pose information of the gimbal load;
当所述处理器执行所述可执行指令时,被配置为:When the processor executes the executable instructions, it is configured to:
根据所述运动数据对所述三维点云进行配准,获取三维点云之间的相对位姿;以及,对所述图像序列进行配准,获取图像之间的相对位姿;The three-dimensional point cloud is registered according to the motion data to obtain the relative pose between the three-dimensional point clouds; and the image sequence is registered to obtain the relative pose between the images;
根据所述运动数据、所述三维点云之间的相对位姿以及所述图像之间的相对位姿,获取所述第一图像采集装置获取三维点云时的位姿和/或所述第二图像采集装置获取图像时的位姿;According to the motion data, the relative pose between the three-dimensional point clouds, and the relative pose between the images, obtain the pose and/or the first image acquisition device when acquiring the three-dimensional point cloud. 2. the pose when the image acquisition device acquires the image;
根据所述三维点云、所述图像序列以及所述第一图像采集装置获取三维点云时的位姿和/或所述第二图像采集装置获取图像时的位姿,生成目标场景的三维模型。A 3D model of the target scene is generated according to the 3D point cloud, the sequence of images, and the pose when the first image acquisition device acquires the 3D point cloud and/or the pose when the second image acquisition device acquires the image .
第三方面,本申请实施例提供了一种可移动平台,包括第二方面所述的云台负载。In a third aspect, an embodiment of the present application provides a movable platform, including the PTZ load described in the second aspect.
第四方面,本申请实施例提供了一种计算机可读存储介质,所述计算机可读存储介质存储有计算机可执行指令,所述计算机可执行指令被处理器执行时实现如第一方面所述的方法。In a fourth aspect, an embodiment of the present application provides a computer-readable storage medium, where the computer-readable storage medium stores computer-executable instructions, and when the computer-executable instructions are executed by a processor, the implementation is as described in the first aspect Methods.
本申请实施例所提供的一种三维重建方法、云台负载、可移动平台以及计算机可读存储介质,通过采用多传感器方式,获取目标场景的三维点云和图像序列;以及,在获取所述三维点云和所述图像序列期间,获取所述云台负载的运动数据,所述运动数据对应所述云台负载的位姿信息;然后根据所述运动数据对所述三维点云进行配准,获取三维点云之间的相对位姿;以及,对所述图像序列进行配准,获取图像之间的相对位姿;接着玩根据所述运动数据、所述三维点云之间的相对位姿以及所述图像之间的相对位姿,获取所述第一图像采集装置获取三维点云时的位姿和/或所述第二图像采 集装置获取图像时的位姿;最后根据所述三维点云、所述图像序列以及所述第一图像采集装置获取三维点云时的位姿和/或所述第二图像采集装置获取图像时的位姿,生成目标场景的三维模型。。本实施例实现通过多传感器方式获取多种数据,使用所述三维点云、所述图像序列以及所述运动数据进行三维重建,由于所述三维点云可以指示所述目标场景的三维信息,所述运动数据包括所述云台负载的位姿信息,所述第一图像采集装置、第二图像采集装置设置于所述云台负载上,因此所述运动数据可以间接指示所述第一图像采集装置和所述第二图像采集装置的位姿信息,从而获得具备高精度的位姿信息,进而使得基于高精度位姿信息生成的三维模型具备更高的定位精度和鲁棒性。A three-dimensional reconstruction method, a pan-tilt load, a movable platform, and a computer-readable storage medium provided by the embodiments of the present application acquire a three-dimensional point cloud and an image sequence of a target scene by adopting a multi-sensor method; During the three-dimensional point cloud and the image sequence, the motion data of the gimbal load is obtained, and the motion data corresponds to the pose information of the gimbal load; then the three-dimensional point cloud is registered according to the motion data , obtain the relative pose between the three-dimensional point clouds; and, register the image sequence to obtain the relative pose between the images; then play according to the motion data, the relative position between the three-dimensional point clouds pose and the relative pose between the images, obtain the pose when the first image acquisition device acquires the three-dimensional point cloud and/or the pose when the second image acquisition device acquires the image; finally, according to the three-dimensional point cloud The point cloud, the image sequence, and the pose when the first image acquisition device acquires the three-dimensional point cloud and/or the pose when the second image acquisition device acquires the image generates a three-dimensional model of the target scene. . In this embodiment, multiple types of data are acquired through multiple sensors, and the three-dimensional point cloud, the image sequence, and the motion data are used for three-dimensional reconstruction. Since the three-dimensional point cloud can indicate the three-dimensional information of the target scene, the The motion data includes the pose information of the pan-tilt load, and the first image acquisition device and the second image acquisition device are arranged on the pan-tilt payload, so the motion data can indirectly indicate the first image acquisition The pose information of the device and the second image acquisition device is obtained, thereby obtaining high-precision pose information, so that the three-dimensional model generated based on the high-precision pose information has higher positioning accuracy and robustness.
附图说明Description of drawings
为了更清楚地说明本申请实施例中的技术方案,下面将对实施例描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本申请的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动性的前提下,还可以根据这些附图获得其他的附图。In order to illustrate the technical solutions in the embodiments of the present application more clearly, the following briefly introduces the drawings that are used in the description of the embodiments. Obviously, the drawings in the following description are only some embodiments of the present application. For those of ordinary skill in the art, other drawings can also be obtained from these drawings without creative labor.
图1是本申请一个实施例提供的一种云台负载的结构示意图;1 is a schematic structural diagram of a pan-tilt load provided by an embodiment of the present application;
图2是本申请一个实施例提供的一种应用场景图;FIG. 2 is an application scenario diagram provided by an embodiment of the present application;
图3是本申请一个实施例提供的一种三维重建方法的流程示意图;3 is a schematic flowchart of a three-dimensional reconstruction method provided by an embodiment of the present application;
图4、图5、图6以及图7是本申请一个实施例提供的位姿图模型的不同结构示意图;4 , 5 , 6 and 7 are schematic diagrams of different structures of a pose graph model provided by an embodiment of the present application;
图8是本申请一个实施例提供的另一种云台负载的结构示意图。FIG. 8 is a schematic structural diagram of another pan-tilt load provided by an embodiment of the present application.
具体实施方式Detailed ways
下面将结合本申请实施例中的附图,对本申请实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例仅仅是本申请一部分实施例,而不是全部的实施例。基于本申请中的实施例,本领域普通技术人员在没有做出创造性劳动前提下所获得的所有其他实施例,都属于本申请保护的范围。The technical solutions in the embodiments of the present application will be clearly and completely described below with reference to the drawings in the embodiments of the present application. Obviously, the described embodiments are only a part of the embodiments of the present application, but not all of the embodiments. Based on the embodiments in the present application, all other embodiments obtained by those of ordinary skill in the art without creative efforts shall fall within the protection scope of the present application.
同步定位与建图(Simultaneously Localization and Mapping,SLAM)是一种机器人在未知环境中估计自身运动并且建立周围环境地图的技术。它在无人机、自动驾驶、移动机器人导航、虚拟现实和增强现实等领域有着广泛的应用。Simultaneous Localization and Mapping (SLAM) is a technology for robots to estimate their own motion and build a map of the surrounding environment in an unknown environment. It has a wide range of applications in drones, autonomous driving, mobile robot navigation, virtual reality and augmented reality.
当机器人处于室外环境时,可通过高精度的GPS信号和先验地图实现全局地图下的定位。但当机器人处于GPS信号不可达的环境时,通常基于摄像头来实现SLAM方法,该方式利用图像进行三维重建,并利用三维重建后的地图来感知自身的运动和周围的环境,但从单一的图像获取的信息有限,因此,该种方式存在定位精度低,三维模型构建质量不高的问题。When the robot is in an outdoor environment, the positioning under the global map can be achieved through high-precision GPS signals and prior maps. However, when the robot is in an environment where GPS signals cannot be reached, the SLAM method is usually implemented based on the camera. This method uses images for 3D reconstruction, and uses the 3D reconstructed map to perceive its own motion and the surrounding environment, but from a single image The information obtained is limited, therefore, this method has the problems of low positioning accuracy and low quality of 3D model construction.
基于此,本申请实施例提供了一种三维重建方法及云台负载,所述云台负载上集成多个不同的传感器,通过采用多传感器方式,获取目标场景的三维点云、图像序列;以及,在获取所述三维点云和所述图像序列期间,获取所述云台负载的运动数据,所述运动数据包括所述云台负载的位姿信息;然后根据所述运动数据对所述三维点云进行配准,获取三维点云之间的相对位姿;以及,对所述图像序列进行配准,获取图像之间的相对位姿;接着根据所述运动数据、所述三维点云之间的相对位姿以及所述图像之间的相对位姿,获取所述第一图像采集装置获取三维点云时的位姿和/或所述第二图像采集装置获取图像时的位姿;最后根据所述三维点云、所述图像序列以及所述第一图像采集装置获取三维点云时的位姿和/或所述第二图像采集装置获取图像时的位姿,生成目标场景的三维模型。本实施例实现通过多传感器方式获取多种数据,基于多种数据的组合来获取高精度的位姿信息,使得基于高精度的位姿信息生成的三维模型具备更高的定位精度和鲁棒性。Based on this, the embodiments of the present application provide a three-dimensional reconstruction method and a gimbal load. The gimbal load integrates a plurality of different sensors, and obtains a three-dimensional point cloud and an image sequence of a target scene by adopting a multi-sensor method; and , during the acquisition of the three-dimensional point cloud and the image sequence, acquire motion data of the gimbal load, where the motion data includes the pose information of the gimbal load; The point clouds are registered to obtain the relative pose between the three-dimensional point clouds; and the image sequence is registered to obtain the relative pose between the images; then according to the motion data, the three-dimensional point cloud and the relative pose between the images, obtain the pose when the first image acquisition device acquires the three-dimensional point cloud and/or the second image acquisition device when the image is acquired; finally A 3D model of the target scene is generated according to the 3D point cloud, the sequence of images, and the pose when the first image acquisition device acquires the 3D point cloud and/or the pose when the second image acquisition device acquires the image . In this embodiment, a variety of data are acquired through multi-sensors, and high-precision pose information is acquired based on the combination of various data, so that the three-dimensional model generated based on the high-precision pose information has higher positioning accuracy and robustness .
所述三维重建方法可应用于云台负载上,所述云台负载安装有多个不同的传感器,比如,请参阅图1,提供了一种云台负载的结构示意图,所述云台负载安装有第一图像采集装置21,第二图像采集装置22以及惯性测量单元23;所述第一图像采集装置21用于实时获取所述目标场景的三维点云;所述第二图像采集装置22用于实时获取所述目标场景的图像序列;所述惯性测量单元23用于实时获取所述云台负载的运动数据;然后所述云台负载可以基于实时获取的所述三维点云、所述图像序列以及所述运动数据实时生成三维模型,实现在满足精度要求的同时也能够满足实时性的需求,以便后续可以直接基于实时生成的三维模型进行定位或者测量等。The three-dimensional reconstruction method can be applied to a pan-tilt load, where a plurality of different sensors are installed. For example, please refer to FIG. 1 , which provides a schematic structural diagram of a pan-tilt load. There are a first image acquisition device 21, a second image acquisition device 22 and an inertial measurement unit 23; the first image acquisition device 21 is used to acquire the three-dimensional point cloud of the target scene in real time; the second image acquisition device 22 is used for to acquire the image sequence of the target scene in real time; the inertial measurement unit 23 is used to acquire the motion data of the PTZ load in real time; then the PTZ load can be based on the real-time acquisition of the three-dimensional point cloud, the image The sequence and the motion data generate a three-dimensional model in real time, so that the real-time requirements can be met while the accuracy requirements are met, so that subsequent positioning or measurement can be performed directly based on the real-time generated three-dimensional model.
所述惯性测量单元用于测量所述云台负载的三轴姿态角(或角速率)以及加速度。一般的,一个惯性测量单元包含了三个单轴的加速度计和三个单轴的陀螺,加速度计检测所述云台负载的加速度信号,而陀螺检测所述云台负载的角速度信号,测量所述云台负载在三维空间中的角速度和加速度,并以此解算出所述云台负载的姿态。The inertial measurement unit is used to measure the three-axis attitude angle (or angular rate) and acceleration of the gimbal load. Generally, an inertial measurement unit includes three single-axis accelerometers and three single-axis gyroscopes. The accelerometer detects the acceleration signal of the gimbal load, and the gyro detects the angular velocity signal of the gimbal load, and measures the angular velocity signal of the gimbal load. The angular velocity and acceleration of the gimbal load in the three-dimensional space are calculated, and the attitude of the gimbal load is calculated.
所述第一图像采集装置可以包括以下至少一种或多种:激光雷达、双目视觉传感器以及结构光深度相机。The first image acquisition device may include at least one or more of the following: lidar, binocular vision sensor, and structured light depth camera.
所述激光雷达用于向目标场景发射激光脉冲序列,然后接收从目标反射回来的激光脉冲序列,并根据反射回来的激光脉冲序列生成三维点云。在一个例子中,所述激光雷达可以确定反射回来的激光脉冲序列的接收时间,例如,通过探测电信号脉冲的上升沿时间和/或下降沿时间确定激光脉冲序列的接收时间。如此,所述激光雷达可以利用激光脉冲序列的接收时间信息和发射时间计算TOF(Time of flight,飞行时间),从而确定探测物到所述激光雷达的距离。所述激光雷达属于自主发光的传感器,不依赖于光源光照,受环境光干扰比较小,即使在无光封闭环境内也可以正常工作,以便后续生成高精度的三维模型,具有广泛的适用性。The lidar is used to transmit a laser pulse sequence to the target scene, then receive the laser pulse sequence reflected from the target, and generate a three-dimensional point cloud according to the reflected laser pulse sequence. In one example, the lidar can determine the reception time of the reflected laser pulse sequence, eg, by detecting the rising edge time and/or the falling edge time of the electrical signal pulse to determine the reception time of the laser pulse sequence. In this way, the laser radar can calculate TOF (Time of flight, time of flight) by using the receiving time information and the transmitting time of the laser pulse sequence, so as to determine the distance from the detected object to the laser radar. The lidar is a self-illuminating sensor, does not depend on light source illumination, is less disturbed by ambient light, and can work normally even in a closed environment without light, so that a high-precision three-dimensional model can be generated later, and it has a wide range of applicability.
所述双目视觉传感器是基于视差原理从不同的位置获取目标场景的两幅图像,通过计算两幅图像对应点间的位置偏差,来获取三维几何信息,以此生成三维点云。双目视觉传感器对于硬件要求低,相应的,也可以降低成本,只需是普通的CMOS(Complementary Metal Oxide Semiconductor,互补金属氧化物半导体)相机即可,只要光线合适,室内环境和室外环境均可使用,因此也具有一定的适用性。The binocular vision sensor obtains two images of the target scene from different positions based on the principle of parallax, and obtains three-dimensional geometric information by calculating the positional deviation between corresponding points of the two images, thereby generating a three-dimensional point cloud. The binocular vision sensor has low hardware requirements, and correspondingly, it can also reduce the cost. It only needs to be an ordinary CMOS (Complementary Metal Oxide Semiconductor) camera. As long as the light is suitable, both indoor and outdoor environments can be used. use, so it also has certain applicability.
所述结构光深度相机是将具有一定结构特征的光线投射到目标场景中再进行采集,这种具备一定结构的光线,会因被摄物体的不同深度区域而采集不同的图像相位信息,然后将其换算成深度信息,以此来获得三维点云。结构光深度相机也是属于自主发光的传感器,不依赖于光源光照,受环境光干扰比较小,即使在无光封闭环境内也可以正常工作,以便后续生成高精度的三维模型,具有广泛的适用性。The structured light depth camera projects light with certain structural characteristics into the target scene and then collects it. This kind of light with certain structure will collect different image phase information due to different depth regions of the subject, and then It is converted into depth information to obtain a 3D point cloud. Structured light depth camera is also a self-illuminating sensor, does not depend on light source illumination, and is less disturbed by ambient light, and can work normally even in a closed environment without light, so as to generate high-precision 3D models later, which has a wide range of applicability .
所述第二图像采集装置可以获取彩色图像、灰度图像、红外图像等等。所述第二图像采集装置包括以下至少一种或多种:可见光相机、灰度相机以及红外相机。The second image acquisition device can acquire color images, grayscale images, infrared images, and the like. The second image acquisition device includes at least one or more of the following: a visible light camera, a grayscale camera, and an infrared camera.
所述第二图像采集装置可以以指定帧率捕捉图像序列。在一些实施方式中,可以以如约24p、25p、30p、48p、50p、60p、72p、90p、100p、120p、300p、50i或60i的标准帧率拍摄图像序列。在一些实施方式中,可以以小于或等于约每0.0001秒、0.0002秒、0.0005秒、0.001秒、0.002秒、0.005秒、0.01秒、0.02秒、0.05秒、0.1秒、0.2秒、0.5秒、1秒、2秒、5秒或10秒一张图像的帧率拍摄图像序列。The second image acquisition device may capture a sequence of images at a specified frame rate. In some embodiments, the sequence of images may be captured at a standard frame rate such as about 24p, 25p, 30p, 48p, 50p, 60p, 72p, 90p, 100p, 120p, 300p, 50i or 60i. In some embodiments, less than or equal to about every 0.0001 second, 0.0002 second, 0.0005 second, 0.001 second, 0.002 second, 0.005 second, 0.01 second, 0.02 second, 0.05 second, 0.1 second, 0.2 second, 0.5 second, 1 Capture image sequences at frame rates of one image per second, 2 seconds, 5 seconds, or 10 seconds.
所述第二图像采集装置可以具有可调的拍摄参数。在不同拍摄参数下,尽管经受完全相同的外部条件(例如,位置、光照),所述第二图像采集装置可以拍摄不同的图像。拍摄参数可以包括曝光(例如,曝光时间、快门速度、孔径、胶片速度)、增益、伽玛、兴趣区、像素合并(binning)/子采样、像素时钟、偏移、触发、ISO等。与曝光相关的参数可以控制到达所述第二图像采集装置中的图像传感器的光量。例如,快门速度可以控制光到达图像传感器的时间量而孔径可以控制在给定时间内到达图像传感 器的光量。与增益相关的参数可以控制对来自光学传感器的信号的放大。ISO可以控制相机对可用光的灵敏度水平。The second image capture device may have adjustable capture parameters. Under different capture parameters, the second image capture device may capture different images despite being subjected to exactly the same external conditions (eg location, lighting). Capture parameters may include exposure (eg, exposure time, shutter speed, aperture, film speed), gain, gamma, region of interest, binning/subsampling, pixel clock, offset, trigger, ISO, and the like. Exposure-related parameters may control the amount of light reaching the image sensor in the second image capture device. For example, shutter speed can control the amount of time light reaches the image sensor and aperture can control the amount of light that reaches the image sensor in a given time. A gain-related parameter can control the amplification of the signal from the optical sensor. ISO controls the level of sensitivity of the camera to the available light.
在一个示例性的实施例中,所述云台负载上安装有激光雷达、可见光相机以及惯性测量单元。在一个例子中,所述激光雷达、所述可见光相机和所述惯性测量单元可以以相同的帧率进行工作。在另一个例子中,所述激光雷达、所述可见光相机和所述惯性测量单元也可以以不同的帧率进行工作,所述激光雷达和、所述可见光相机和所述惯性测量单元的帧率满足在预设的时间周期内能够获取到三维点云、图像序列以及所述云台负载的运动数据。In an exemplary embodiment, a lidar, a visible light camera and an inertial measurement unit are installed on the pan/tilt load. In one example, the lidar, the visible light camera, and the inertial measurement unit may operate at the same frame rate. In another example, the lidar, the visible light camera and the inertial measurement unit may also work at different frame rates, the frame rates of the lidar and the visible light camera and the inertial measurement unit The three-dimensional point cloud, the image sequence and the motion data of the PTZ load can be acquired within a preset time period.
在一实施例中,所述云台负载可以挂载在可移动平台上,由所述可移动平台携带所述云台负载在目标场景中移动,以针对于所述目标场景进行三维重建,便于后续可以利用三维重建得到的三维模型进行定位任务、测绘任务或者导航任务等等。其中,所述可移动平台包括但不限于无人飞行器、无人驾驶车辆或者移动机器人等。In one embodiment, the pan-tilt payload can be mounted on a movable platform, and the pan-tilt payload can be carried by the movable platform to move in a target scene, so as to perform three-dimensional reconstruction for the target scene, which is convenient for follow-up. The three-dimensional model obtained by the three-dimensional reconstruction can be used for positioning tasks, mapping tasks, or navigation tasks, and the like. Wherein, the movable platform includes, but is not limited to, an unmanned aerial vehicle, an unmanned vehicle, or a mobile robot.
在一实施例中,所述云台负载可以与遥控终端通信连接,用户可以通过遥控终端控制所述云台负载实时采集数据并实时进行三维重建,并将三维重建得到的三维模型实时传输给所述遥控终端,以在所述遥控终端上显示,从而用户可以实时了解三维重建进度,方便用户使用;进一步地,用户也可以针对于显示在所述遥控终端上的三维模型输入反馈信息,所述反馈信息用于指示所述第一图像采集装置漏采的位置,由所述遥控终端将所述反馈信息传输给所述云台负载,使得所述云台负载可以控制所述第一图像采集装置采集所述目标场景中所述反馈信息指示的位置处的三维点云,实现在线监控过程,保证三维点云数据准确且有效地获取,有利于提高三维重建效率。In one embodiment, the PTZ load can be connected to the remote control terminal in communication, and the user can control the PTZ load through the remote control terminal to collect data in real time and perform 3D reconstruction in real time, and transmit the 3D model obtained by the 3D reconstruction to the remote control terminal in real time. The remote control terminal is displayed on the remote control terminal, so that the user can know the progress of the three-dimensional reconstruction in real time, which is convenient for the user to use; further, the user can also input feedback information for the three-dimensional model displayed on the remote control terminal. The feedback information is used to indicate the position where the first image acquisition device missed the acquisition, and the remote control terminal transmits the feedback information to the PTZ load, so that the PTZ load can control the first image acquisition device The 3D point cloud at the position indicated by the feedback information in the target scene is collected to realize the online monitoring process, to ensure accurate and effective acquisition of 3D point cloud data, and to improve the efficiency of 3D reconstruction.
在一个示例性的应用场景中,请参阅图2,所述云台负载20可以挂载在无人飞行器10上,所述云台负载20和所述无人飞行器10与遥控终端30通信连接,用户可以通过遥控终端30对所述无人飞行器10以及所述云台负载20进行控制,比如所述遥控终端30可以响应于用户的触发操作,向所述云台负载20发送三维重建指令,所述云台负载20安装有第一图像采集装置,第二图像采集装置以及惯性测量单元;所述第一图像采集装置响应于所述三维重建指令,实时获取所述目标场景的三维点云;所述第二图像采集装置响应于所述三维重建指令,实时获取所述目标场景的图像序列;所述惯性测量单元响应于所述三维重建指令,实时获取所述云台负载20的运动数据,所述运动数据包括:在获取目标场景的三维点云、图像序列的期间,所述云台负载的位姿信息;然后所述云台负载20可以基于实时获取的所述三维点云、所述图像序列以及所述运动数据实时生成三维模型,实现在满足精度要求的同时也能够满足实时性的需求, 并将实时生成的三维模型发送给所述遥控终端30,使得所述遥控终端30可以显示所述三维模型,以便用户可以实时了解三维重建进度,方便用户使用;用户还可以通过所述遥控终端30控制所述无人飞行器10飞行,使得无人飞行器10携带所述云台负载20飞行至所述目标场景的不同位置,从而所述云台负载20可以对所述目标场景的不同位置进行三维重建,获取所述目标场景内的完整的三维地图,以便后续所述无人飞行器可以基于所述三维地图进行定位任务、测量任务或者巡航任务等。In an exemplary application scenario, please refer to FIG. 2 , the gimbal load 20 can be mounted on the unmanned aerial vehicle 10, and the gimbal load 20 and the unmanned aerial vehicle 10 are connected in communication with the remote control terminal 30, The user can control the UAV 10 and the gimbal load 20 through the remote control terminal 30. For example, the remote control terminal 30 can respond to the user's trigger operation and send a three-dimensional reconstruction instruction to the gimbal load 20, so the The pan-tilt load 20 is installed with a first image acquisition device, a second image acquisition device and an inertial measurement unit; the first image acquisition device acquires the three-dimensional point cloud of the target scene in real time in response to the three-dimensional reconstruction instruction; The second image acquisition device acquires the image sequence of the target scene in real time in response to the three-dimensional reconstruction instruction; the inertial measurement unit acquires the motion data of the pan-tilt load 20 in real time in response to the three-dimensional reconstruction instruction, so the The motion data includes: during the acquisition of the three-dimensional point cloud and image sequence of the target scene, the pose information of the gimbal load; then the gimbal load 20 can be based on the real-time acquisition of the three-dimensional point cloud, the image The sequence and the motion data generate a three-dimensional model in real time, so as to meet the real-time requirements while meeting the accuracy requirements, and send the three-dimensional model generated in real time to the remote control terminal 30, so that the remote control terminal 30 can display all the data. The three-dimensional model is described so that the user can know the progress of the three-dimensional reconstruction in real time, which is convenient for the user to use; the user can also control the flight of the unmanned aerial vehicle 10 through the remote control terminal 30, so that the unmanned aerial vehicle 10 can carry the gimbal load 20 and fly to the desired location. different positions of the target scene, so that the pan-tilt load 20 can perform three-dimensional reconstruction on different positions of the target scene, and obtain a complete three-dimensional map in the target scene, so that the unmanned aerial vehicle can be based on the 3D map for positioning tasks, measurement tasks or cruise tasks, etc.
接下来对三维重建过程进行说明,请参阅图3,为本申请一个实施例提供的三维重建方法的流程示意图,所述方法应用于云台负载上,所述方法包括:Next, the three-dimensional reconstruction process will be described. Please refer to FIG. 3 , which is a schematic flowchart of a three-dimensional reconstruction method provided by an embodiment of the present application. The method is applied to a PTZ load, and the method includes:
在步骤S101中,获取目标场景的三维点云、图像序列以及所述云台负载的运动数据;其中,所述三维点云由所述第一图像采集装置获取,所述图像序列由所述第二图像采集装置获取;所述运动数据包括:在获取目标场景的三维点云、图像序列的期间,所述云台负载的位姿信息。In step S101, a 3D point cloud of a target scene, an image sequence, and motion data loaded by the gimbal are acquired; wherein, the 3D point cloud is acquired by the first image acquisition device, and the image sequence is acquired by the first image acquisition device. 2. Acquired by an image acquisition device; the motion data includes: during the acquisition of a three-dimensional point cloud and an image sequence of a target scene, the pose information of the PTZ load.
在步骤S102中,根据所述运动数据对所述三维点云进行配准,获取三维点云之间的相对位姿;以及,对所述图像序列进行配准,获取图像之间的相对位姿。In step S102, the three-dimensional point cloud is registered according to the motion data to obtain the relative pose between the three-dimensional point clouds; and the image sequence is registered to obtain the relative pose between the images .
在步骤S103中,根据所述运动数据、所述三维点云之间的相对位姿以及所述图像之间的相对位姿,获取所述第一图像采集装置获取三维点云时的位姿和/或所述第二图像采集装置获取图像时的位姿。In step S103, according to the motion data, the relative pose between the three-dimensional point clouds, and the relative pose between the images, the pose and the pose when the first image acquisition device acquires the three-dimensional point cloud is acquired /or the pose when the second image acquisition device acquires an image.
在步骤S104中,根据所述三维点云、所述图像序列以及所述第一图像采集装置获取三维点云时的位姿和/或所述第二图像采集装置获取图像时的位姿,生成目标场景的三维模型。In step S104, according to the three-dimensional point cloud, the image sequence, and the pose when the first image acquisition device acquires the three-dimensional point cloud and/or the pose when the second image acquisition device acquires the image, generate 3D model of the target scene.
其中,所述目标场景可以是卫星定位信号低于预定强度的场景,比如所述目标场景可以是室内场景或者室外非空旷场景(比如水下、矿洞等),在该种场景下,可以使用本实施例的三维重建方法生成高精度三维模型。Wherein, the target scene may be a scene where the satellite positioning signal is lower than a predetermined intensity, for example, the target scene may be an indoor scene or an outdoor non-open scene (such as underwater, a mine, etc.), in this kind of scene, you can use The three-dimensional reconstruction method of this embodiment generates a high-precision three-dimensional model.
所述卫星定位信号(即GNSS信号)包括但不限于GPS信号、来自伽利略卫星导航系统(GALILEO)的信号或者来自北斗卫星导航系统(BDS)的信号等。The satellite positioning signals (ie, GNSS signals) include, but are not limited to, GPS signals, signals from the Galileo Satellite Navigation System (GALILEO), or signals from the Beidou Satellite Navigation System (BDS).
在一个例子中,所述卫星定位信号的强度可以通过以下至少一种参数来确定:卫星定位信号的信噪比、卫星定位信号的冷启动时间或者热启动时间;在一个示例性的实施例中,若所述卫星定位信号的信噪比低于预设阈值,表明所述卫星定位信号的强度低于预设强度;在另一示例性的实施例中,若所述卫星定位信号的信噪比低于预设阈值且所述卫星定位信号的冷启动时间不符合预设时间条件,表明所述卫星定位信号的强度低于预设强度。In one example, the strength of the satellite positioning signal may be determined by at least one of the following parameters: the signal-to-noise ratio of the satellite positioning signal, the cold start time or the warm start time of the satellite positioning signal; in an exemplary embodiment , if the signal-to-noise ratio of the satellite positioning signal is lower than the preset threshold, it indicates that the strength of the satellite positioning signal is lower than the preset strength; in another exemplary embodiment, if the signal-to-noise ratio of the satellite positioning signal If the ratio is lower than the preset threshold and the cold start time of the satellite positioning signal does not meet the preset time condition, it indicates that the strength of the satellite positioning signal is lower than the preset strength.
所述云台负载可以使用深处相机实时获取目标场景的三维点云,使用第二图像采集装置实时获取所述目标场景的图像序列;以及在获取所述三维点云和所述图像序列期间,使用惯性测量单元实时采集所述云台负载的运动数据,所述运动数据对应所述云台负载的位姿信息。本实施例使用实时获取的所述三维点云、所述图像序列以及所述运动数据进行实时的三维重建,由于所述三维点云可以指示所述目标场景的三维信息,所述运动数据可以指示所述云台负载的位姿信息,所述第一图像采集装置、第二图像采集装置设置于所述云台负载上,因此所述运动数据可以间接指示所述第一图像采集装置和所述第二图像采集装置的位姿信息,从而获得具备高精度的位姿信息,进而使得基于高精度位姿信息生成的三维模型具备更高的定位精度和鲁棒性;而且实时获取以及实时三维重建的过程也够满足具有实时性要求的场景的需求,以便后续可以直接基于实时生成的三维模型进行定位或者测量等。The pan-tilt load can use a depth camera to acquire a 3D point cloud of the target scene in real time, and use a second image acquisition device to acquire an image sequence of the target scene in real time; and during acquisition of the 3D point cloud and the image sequence, Using an inertial measurement unit to collect motion data of the gimbal load in real time, the motion data corresponds to the pose information of the gimbal load. This embodiment uses the three-dimensional point cloud, the image sequence, and the motion data acquired in real time to perform real-time three-dimensional reconstruction. Since the three-dimensional point cloud can indicate the three-dimensional information of the target scene, the motion data can indicate The pose information of the pan-tilt load, the first image acquisition device and the second image acquisition device are set on the pan-tilt payload, so the motion data can indirectly indicate the first image acquisition device and the The pose information of the second image acquisition device is obtained to obtain high-precision pose information, so that the three-dimensional model generated based on the high-precision pose information has higher positioning accuracy and robustness; and real-time acquisition and real-time three-dimensional reconstruction The process is also enough to meet the needs of the scene with real-time requirements, so that the subsequent positioning or measurement can be directly based on the real-time generated 3D model.
在一实施例中,所述云台负载可以与所述遥控终端通信连接,所述云台负载可以通过用户在遥控终端上输入的所述三维模型的精度要求和/或密度要求,生成采集轨迹和采集时间,然后根据所述采集轨迹和采集时间控制所述第一图像采集装置、所述第二图像采集装置以及所述惯性测量单元采集数据。本实施例实现基于所述三维模型的精度要求和/或密度要求自动规划采集方案并控制所述第一图像采集装置、所述第二图像采集装置以及所述惯性测量单元采集数据,用户在采集过程中无需参与,可以有效防止在采集过程中引入的人为误差,提高生成的三维模型的准确性。In an embodiment, the pan-tilt load can be connected to the remote control terminal in communication, and the pan-tilt load can generate a collection trajectory according to the accuracy requirements and/or density requirements of the three-dimensional model input by the user on the remote control terminal. and acquisition time, and then control the first image acquisition device, the second image acquisition device and the inertial measurement unit to acquire data according to the acquisition track and the acquisition time. This embodiment implements automatic planning of the acquisition scheme based on the accuracy requirements and/or density requirements of the three-dimensional model, and controls the first image acquisition device, the second image acquisition device, and the inertial measurement unit to acquire data. There is no need to participate in the process, which can effectively prevent human errors introduced in the acquisition process and improve the accuracy of the generated 3D model.
在获取所述三维点云、所述图像序列以及所述运动数据之后,所述云台负载根据所述运动数据对所述三维点云进行配准,获取三维点云之间的相对位姿,所述三维点云之间的相对位姿用于对三维点云进行校准,从而提高后续三维重建结果的准确性;以及,所述云台负载对所述图像序列进行配准,获取图像之间的相对位姿,所述图像之间的相对位姿用于确定所述第二图像采集装置的位姿,从而提高后续三维重建结果的准确性。After acquiring the 3D point cloud, the image sequence and the motion data, the PTZ load registers the 3D point cloud according to the motion data, and obtains the relative pose between the 3D point clouds, The relative poses between the three-dimensional point clouds are used to calibrate the three-dimensional point clouds, thereby improving the accuracy of subsequent three-dimensional reconstruction results; The relative pose between the images is used to determine the pose of the second image acquisition device, thereby improving the accuracy of subsequent three-dimensional reconstruction results.
这里对获取三维点云之间的相对位姿的过程进行说明:The process of obtaining the relative pose between 3D point clouds is described here:
考虑到如激光雷达、结构光深度相机等,在同一时刻只能采集到三维点云中的一个点的数据,这个点表示在三维坐标系下的一个三维坐标点,在采集三维点云的过程中会产生类似于果冻效应(rolling shutter effect)的问题,一帧三维点云中的点不是在同一时刻采集的,在采集过程中,激光雷达随着载体在运动,但是激光雷达点测量的是物体和雷达之间的距离,所以不同激光点的坐标系就不一样了。为了解决这个问题,三维点云中的每个点均对应有独立的坐标系,从而保证点在对应的坐标系下的三维坐 标的准确性。但在后续利用三维点云进行三维重建时,需要三维点云中的点在同一坐标系下,因此需要对所述三维点云进行重投影,把采集过程中激光雷达的运动计算出来,并在对应的激光点上补偿这个运动量。本实施例利用与所述第一图像采集装置安装于同一云台负载上的惯性测量单元获得的运动数据来对所述三维点云进行变换,所述惯性测量单元获取的运动数据对应所述云台负载的位姿信息,由于惯性测量单元以及所述第一图像采集装置均安装于云台负载上,所述第一图像采集装置的运动过程与所述云台负载的运动过程是相同或相似的,因此所述惯性测量单元获取的运动数据也指示着所述第一图像采集装置的运动过程,因此可以利用惯性测量单元测得的运动数据可以准确确定处于不同坐标系的点之间的变换关系,比如可以利用所述运动数据以及所述惯性测量单元与所述第一图像采集装置之间的外参转换关系,获取所述点之间的变换关系。Considering that such as lidar, structured light depth camera, etc., only one point in the 3D point cloud can be collected at the same time. This point represents a 3D coordinate point in the 3D coordinate system. In the process of collecting the 3D point cloud There will be a problem similar to the rolling shutter effect. The points in a frame of 3D point cloud are not collected at the same time. During the collection process, the lidar moves with the carrier, but the lidar points measure the The distance between the object and the radar, so the coordinate system of different laser points is different. In order to solve this problem, each point in the 3D point cloud has an independent coordinate system, so as to ensure the accuracy of the 3D coordinates of the point in the corresponding coordinate system. However, when the 3D point cloud is used for 3D reconstruction in the future, the points in the 3D point cloud need to be in the same coordinate system, so the 3D point cloud needs to be reprojected, the motion of the lidar during the acquisition process is calculated, and the This amount of movement is compensated on the corresponding laser point. In this embodiment, the three-dimensional point cloud is transformed by the motion data obtained by the inertial measurement unit installed on the same pan/tilt load as the first image acquisition device, and the motion data obtained by the inertial measurement unit corresponds to the cloud The pose information of the platform load, since the inertial measurement unit and the first image acquisition device are both installed on the pan-tilt load, the motion process of the first image acquisition device and the motion process of the pan-tilt load are the same or similar Therefore, the motion data acquired by the inertial measurement unit also indicates the motion process of the first image acquisition device, so the motion data measured by the inertial measurement unit can be used to accurately determine the transformation between points in different coordinate systems For example, the transformation relationship between the points can be obtained by using the motion data and the external parameter transformation relationship between the inertial measurement unit and the first image acquisition device.
对于在连续时间序列上采集到的一组三维点云,所述云台负载根据该组三维点云中每个点对应的运动数据获取点之间的变换关系,然后根据所述点之间的变换关系将该组三维点云中的点重投影至同一坐标系。本实施例将在连续时间序列上采集到的一组三维点云变换至同一坐标系下,实现三维点云数据的整合,有利于提高三维重建效率。For a group of 3D point clouds collected in a continuous time series, the PTZ load obtains the transformation relationship between points according to the motion data corresponding to each point in the group of 3D point clouds, and then according to the relationship between the points The transformation relationship reprojects the points in the set of 3D point clouds to the same coordinate system. In this embodiment, a group of three-dimensional point clouds collected in a continuous time series are transformed into the same coordinate system, so as to realize the integration of three-dimensional point cloud data, which is beneficial to improve the efficiency of three-dimensional reconstruction.
其中,所述云台负载根据每个点的获取时间获取所述点对应的运动数据;在一个例子,所述第一图像采集装置为激光雷达,所述点的获取时间根据反射回的激光脉冲序列的接收时间所确定,所述点的获取时间即为对应的运动数据的获取时间,则所述云台负载在获取所述点的相同时刻从所述惯性测量单元获取运动数据。Wherein, the PTZ load acquires the motion data corresponding to the point according to the acquisition time of each point; in an example, the first image acquisition device is a laser radar, and the acquisition time of the point is based on the reflected laser pulses Determined by the receiving time of the sequence, the acquisition time of the point is the acquisition time of the corresponding motion data, and the gimbal load acquires the motion data from the inertial measurement unit at the same moment when the point is acquired.
在连续的时间序列上,所述第一图像采集装置会以预设的时间间隔来获取每个点,在获取点之间的变换关系时,可以根据该组点云中每个点对应的运动数据确定所述惯性测量单元在所述预设时间间隔内的相对位姿,其中,所述预设时间间隔为获取相邻点的时间间隔;然后所述云台负载根据所述惯性测量单元在预设时间间隔内的相对位姿、以及所述第一图像采集装置与所述惯性测量单元之间的外参转换关系,获取所述点之间的变换关系。本实施例中,由于惯性测量单元以及所述第一图像采集装置均安装于云台负载上,所述惯性测量单元获取的运动数据也指示着所述第一图像采集装置的运动过程,因此可以利用惯性测量单元测得的运动数据可以准确确定处于不同坐标系的点之间的变换关系,从而可以基于所述点之间的变换关系将在连续时间序列上采集到的一组三维点云变换至同一坐标系下,实现三维点云数据的整合,有利于提高三维重建效率。In a continuous time series, the first image acquisition device will acquire each point at a preset time interval, and when acquiring the transformation relationship between points, the corresponding motion of each point in the set of point clouds can be obtained. The data determines the relative pose of the inertial measurement unit within the preset time interval, wherein the preset time interval is the time interval for acquiring adjacent points; then the gimbal load is based on the inertial measurement unit at the time interval. The relative pose within a preset time interval, and the extrinsic parameter transformation relationship between the first image acquisition device and the inertial measurement unit, obtain the transformation relationship between the points. In this embodiment, since the inertial measurement unit and the first image acquisition device are both installed on the pan-tilt load, the motion data acquired by the inertial measurement unit also indicates the movement process of the first image acquisition device, so it can be Using the motion data measured by the inertial measurement unit, the transformation relationship between points in different coordinate systems can be accurately determined, so that a set of three-dimensional point clouds collected in a continuous time series can be transformed based on the transformation relationship between the points. Under the same coordinate system, the integration of 3D point cloud data is realized, which is beneficial to improve the efficiency of 3D reconstruction.
在一个例子,对每组以{Δt,Δt=t i+1-t i}为时间间隔累计的三维点云
Figure PCTCN2020120978-appb-000001
Figure PCTCN2020120978-appb-000002
n为大于1的整数,对每一个点
Figure PCTCN2020120978-appb-000003
使用对应的运动数据进行积分、滤波、非线性差值后得到的相对应的变换矩阵
Figure PCTCN2020120978-appb-000004
j为小于或等于n的整数,然后可以将所有的点重投影至T i时刻定义的坐标系下,即
Figure PCTCN2020120978-appb-000005
从而得到处于同一坐标系下的三维点云
Figure PCTCN2020120978-appb-000006
In one example, for each set of 3D point clouds accumulated at time intervals of {Δt,Δt=t i+1 -t i }
Figure PCTCN2020120978-appb-000001
Figure PCTCN2020120978-appb-000002
n is an integer greater than 1, for each point
Figure PCTCN2020120978-appb-000003
The corresponding transformation matrix obtained after integrating, filtering, and nonlinear difference using the corresponding motion data
Figure PCTCN2020120978-appb-000004
j is an integer less than or equal to n, and then all points can be reprojected to the coordinate system defined at time T i , that is
Figure PCTCN2020120978-appb-000005
So as to get the 3D point cloud in the same coordinate system
Figure PCTCN2020120978-appb-000006
将在连续时间序列上采集到的一组三维点云变换至同一坐标系下之后,为了方便后续的三维重建过程,进一步提高三维重建效率,考虑到不同组三维点云之间的坐标系也可能不同,因此,需要进一步对三维点云数据进行整合。对于相邻两组三维点云,所述云台负载可以获取相邻两组三维点云之间的相对位姿,所述相邻两组三维点云之间的相对位姿用于将相邻两组三维点云重投影至同一坐标系,实现三维点云数据的匹配与整合,从而方便后续的三维重建过程。其中,相邻两组三维点云中均包括有针对同一物体采集到的点,从而保证相邻两组三维点云的顺利匹配。After transforming a group of 3D point clouds collected in a continuous time series into the same coordinate system, in order to facilitate the subsequent 3D reconstruction process and further improve the efficiency of 3D reconstruction, considering that the coordinate systems between different groups of 3D point clouds may also be Therefore, further integration of 3D point cloud data is required. For two adjacent groups of 3D point clouds, the gimbal load can obtain the relative poses between the adjacent two groups of 3D point clouds, and the relative poses between the adjacent two groups of 3D point clouds are used to The two sets of 3D point clouds are reprojected to the same coordinate system to realize the matching and integration of 3D point cloud data, thus facilitating the subsequent 3D reconstruction process. Among them, the adjacent two sets of three-dimensional point clouds include points collected for the same object, so as to ensure the smooth matching of the two adjacent three-dimensional point clouds.
进一步地,考虑到相关技术中通常采用Iterative Closest Point/Plane(ICP)算法进行三维点云之间的匹配,但该算法比较耗时,可能无法满足三维模型实时生成的需求。因此,在要求能够实时生成三维模型的情况下,在获取相邻两组三维点云之间的相对位姿时,所述云台负载获取每组三维点云的面特征和/或拐角特征,然后根据相邻两组三维点云的面特征和/或拐角特征,获取相邻两组三维点云之间的相对位姿;所述相邻两组三维点云之间的相对位姿用于将相邻两组三维点云重投影至同一坐标系。本实施例利用三维点云获取空间中的面特征和/或拐角特征来进行高速匹配,简化匹配过程,有利于提高获取相邻两组三维点云之间的相对位姿的效率,缩短匹配时长。Further, considering that the Iterative Closest Point/Plane (ICP) algorithm is usually used in related technologies to match between 3D point clouds, but this algorithm is time-consuming and may not meet the needs of real-time generation of 3D models. Therefore, when it is required to be able to generate a 3D model in real time, when acquiring the relative poses between two adjacent groups of 3D point clouds, the PTZ load acquires the surface features and/or corner features of each group of 3D point clouds, Then, according to the surface features and/or corner features of the adjacent two groups of three-dimensional point clouds, the relative poses between the two adjacent groups of three-dimensional point clouds are obtained; the relative poses between the two adjacent groups of three-dimensional point clouds are used for Reproject two adjacent sets of 3D point clouds to the same coordinate system. This embodiment uses the surface features and/or corner features in the 3D point cloud acquisition space to perform high-speed matching, simplifies the matching process, helps improve the efficiency of acquiring the relative poses between two adjacent groups of 3D point clouds, and shortens the matching time. .
可以理解的是,在没有实时性要求的情况下,也可以使用相关技术中的Iterative Closest Point/Plane(ICP)算法进行三维点云之间的匹配,获取三维点云之间的相对位姿,本实施例对此不足任何限制。It can be understood that in the absence of real-time requirements, the Iterative Closest Point/Plane (ICP) algorithm in related technologies can also be used to match between 3D point clouds to obtain the relative pose between 3D point clouds. This embodiment does not have any limitation on this.
在一种实现方式中,对于相邻两组三维点云,可以根据各组三维点云中点之间的曲度信息来确定该组三维点云的面特征和/或拐角特征。作为例子,所述每组三维点云的面特征根据曲度大于预设阈值的曲度信息所确定;所述每组三维点云的拐角特征根据曲度小于或等于预设阈值的曲度信息所确定。本实施例中,将点之间的匹配过程转换为三维点云在空间中的面特征和/或拐角特征之间的匹配过程,有利于简化匹配过程,实现相邻两组三维点云的面特征和/或拐角特征的高速匹配,进一步有利于提高获取相邻两组三维点云之间的相对位姿的效率,缩短匹配时长,满足实时生成三维模型的需求。In an implementation manner, for two adjacent groups of three-dimensional point clouds, the surface features and/or corner features of the three-dimensional point clouds of the group may be determined according to the curvature information between the midpoints of the three-dimensional point clouds of each group. As an example, the surface features of each group of 3D point clouds are determined according to curvature information whose curvature is greater than a preset threshold; the corner features of each group of 3D point clouds are determined according to curvature information whose curvature is less than or equal to a preset threshold determined. In this embodiment, the matching process between points is converted into a matching process between surface features and/or corner features of a three-dimensional point cloud in space, which is beneficial to simplify the matching process and realize the two adjacent sets of three-dimensional point clouds. The high-speed matching of features and/or corner features is further conducive to improving the efficiency of obtaining relative poses between two adjacent sets of 3D point clouds, shortening the matching time, and meeting the needs of real-time generation of 3D models.
其中,所述面特征包括指示平面的至少一个三维坐标信息以及该平面的法向量;和/或,所述拐角特征包括指示边缘的至少一个三维坐标信息以及指示该边缘的向量。Wherein, the surface feature includes at least one three-dimensional coordinate information indicating a plane and a normal vector of the plane; and/or the corner feature includes at least one three-dimensional coordinate information indicating an edge and a vector indicating the edge.
具体来说,在根据相邻两组三维点云的面特征和/或拐角特征获取相邻两组三维点云之间的相对位姿时,所述云台负载可以根据相邻两组三维点云的面特征之间的相似度,和/或,相邻两组三维点云的拐角特征之间的相似度,获取三维点云之间的相对位姿;相邻两组三维点云的面特征之间越相似,和/或,相邻两组三维点云的拐角特征之间越相似,表明该面特征和/或拐角特征指示的相邻两组三维点云中的点可能是匹配的。本实施例通过面特征之间的相似度和/或拐角特征的相似度来获取相邻两组三维点云之间的相对位姿,保证获取结果的准确性。Specifically, when the relative poses between the adjacent two groups of three-dimensional point clouds are obtained according to the surface features and/or corner features of the two adjacent groups of three-dimensional point clouds, the pan-tilt load may be based on the adjacent two groups of three-dimensional point clouds. The similarity between the surface features of the cloud, and/or the similarity between the corner features of the adjacent two sets of three-dimensional point clouds, to obtain the relative pose between the three-dimensional point clouds; The more similar the features are, and/or the more similar the corner features of the adjacent two sets of 3D point clouds are, indicating that the points in the adjacent two sets of 3D point clouds indicated by the surface feature and/or the corner feature may be matched. . In this embodiment, the relative poses between the adjacent two groups of three-dimensional point clouds are obtained through the similarity between the surface features and/or the similarity between the corner features, so as to ensure the accuracy of the obtained results.
在一种实现方式中,所述相邻两组三维点云的面特征之间的相似度基于所述相邻两组三维点云的面特征之间的距离确定;和/或,所述相邻两组三维点云的拐角特征之间的相似度基于所述相邻两组三维点云的拐角特征之间的距离确定。其中,和/或表示两者或者两者之一。In an implementation manner, the similarity between the surface features of the adjacent two groups of three-dimensional point clouds is determined based on the distance between the surface features of the two adjacent groups of three-dimensional point clouds; and/or, the The similarity between the corner features of the adjacent two groups of three-dimensional point clouds is determined based on the distance between the corner features of the two adjacent groups of three-dimensional point clouds. Wherein, and/or represent both or one of the two.
在一个例子中,根据处于同一坐标系下的三维点云
Figure PCTCN2020120978-appb-000007
生成
Figure PCTCN2020120978-appb-000008
个面(Surface)特征
Figure PCTCN2020120978-appb-000009
Figure PCTCN2020120978-appb-000010
个拐角(Edge)特征
Figure PCTCN2020120978-appb-000011
根据以上信息可利用优化算法解:
Figure PCTCN2020120978-appb-000012
其中,
Figure PCTCN2020120978-appb-000013
表示所述相邻两组三维点云的面特征之间的距离差,
Figure PCTCN2020120978-appb-000014
表示所述相邻两组三维点云的拐角特征之间的距离差。
Figure PCTCN2020120978-appb-000015
可由李代数(LieAlgebra)算法转换表达得到相邻两组三维点云之间的相对位姿。本实施例使用面特征和拐角特征进行高速匹配,有利于提高匹配效率,满足实时生成三维模型的需求。
In one example, according to the three-dimensional point cloud in the same coordinate system
Figure PCTCN2020120978-appb-000007
generate
Figure PCTCN2020120978-appb-000008
Surface features
Figure PCTCN2020120978-appb-000009
and
Figure PCTCN2020120978-appb-000010
Edge features
Figure PCTCN2020120978-appb-000011
According to the above information, the optimization algorithm can be used to solve:
Figure PCTCN2020120978-appb-000012
in,
Figure PCTCN2020120978-appb-000013
represents the distance difference between the surface features of the adjacent two groups of three-dimensional point clouds,
Figure PCTCN2020120978-appb-000014
Represents the distance difference between the corner features of the adjacent two groups of three-dimensional point clouds.
Figure PCTCN2020120978-appb-000015
The relative poses between two adjacent groups of 3D point clouds can be obtained through the transformation and expression of LieAlgebra algorithm. This embodiment uses the surface feature and the corner feature to perform high-speed matching, which is beneficial to improve the matching efficiency and meet the requirement of generating a three-dimensional model in real time.
这里对获取图像之间的相对位姿的过程进行说明:所述图像序列包括在连续的时间序列上获取的多个图像。The process of acquiring the relative pose between images is described here: the image sequence includes a plurality of images acquired in a continuous time sequence.
在一种实现方式中,所述云台负载可以提取各个所述图像的特征点,然后基于所述图像的特征点进行配准,获取所述图像之间的相对位姿。在一个例子中,所述云台负载可以使用光束法平差(Bundle Adjustment,BA)算法来对所述图像的特征点进行配准,获取所述图像之间的相对位姿。In an implementation manner, the PTZ payload can extract feature points of each of the images, and then perform registration based on the feature points of the images to obtain the relative poses between the images. In one example, the PTZ load may use a bundle adjustment (Bundle Adjustment, BA) algorithm to register the feature points of the images, and obtain the relative poses between the images.
在第二种实现方式中,考虑到本申请实施例在获取所述图像序列的同时也获取所述三维点云,则可以利用三维点云反投影得到的深度图来辅助所述图像的配准过程,提高图像配准精度。具体来说,所述云台负载可以使用所述运动数据将所述三维点云 进行重投影,并根据重投影后的三维点云以及所述运动数据生成与所述图像对应的深度图;上述提到,三维点云中的每个点均对应有独立的坐标,为了保证生成的深度图的准确性,需要三维点云中的每个点在同一坐标系下,因此所述云台负载使用所述运动数据将所述三维点云进行重投影,使得重投影后的三维点云中的每个点在同一坐标系下;进一步地,所述惯性测量单元获取的运动数据对应所述云台负载的位姿信息,由于惯性测量单元、所述第二图像采集装置以及所述第一图像采集装置均安装于云台负载上,所述第二图像采集装置和所述第一图像采集装置的运动过程与所述云台负载的运动过程是相同或相似的,因此所述惯性测量单元获取的运动数据也可以指示着所述第二图像采集装置和所述第二图像采集装置的运动过程,则可以使用重投影后的三维点云以及所述运动数据生成与所述图像对应的深度图。In the second implementation manner, considering that the embodiment of the present application also acquires the 3D point cloud while acquiring the image sequence, the depth map obtained by the back-projection of the 3D point cloud can be used to assist the registration of the image process to improve image registration accuracy. Specifically, the PTZ load can reproject the 3D point cloud by using the motion data, and generate a depth map corresponding to the image according to the reprojected 3D point cloud and the motion data; It is mentioned that each point in the 3D point cloud has independent coordinates. In order to ensure the accuracy of the generated depth map, each point in the 3D point cloud needs to be in the same coordinate system, so the PTZ load uses The motion data reprojects the 3D point cloud, so that each point in the reprojected 3D point cloud is in the same coordinate system; further, the motion data obtained by the inertial measurement unit corresponds to the PTZ The pose information of the load, since the inertial measurement unit, the second image acquisition device and the first image acquisition device are all installed on the PTZ load, the second image acquisition device and the first image acquisition device are The motion process is the same as or similar to the motion process of the PTZ load, so the motion data acquired by the inertial measurement unit can also indicate the motion process of the second image acquisition device and the second image acquisition device, Then, a depth map corresponding to the image may be generated using the reprojected 3D point cloud and the motion data.
在一个例子中,考虑到三维点云的密度有限,可以先基于所述重投影后的三维点云生成多面体,再根据所述多面图和所述运动数据投影到相机平面,得到所述图像对应的深度图。In one example, considering the limited density of the 3D point cloud, a polyhedron may be generated based on the reprojected 3D point cloud, and then projected onto the camera plane according to the multi-facet map and the motion data to obtain the image corresponding to the depth map.
接着,所述云台负载还提取所述图像中的特征点,并从所述图像对应的深度图中获取所述特征点的深度信息;然后根据所述图像的特征点以及所述特征点的深度信息进行配准,获取所述图像之间的相对位姿。本实施例利用三维点云反投影得到的深度图来辅助所述图像的配准过程,有利于提高图像配准精度,进而提高后续生成的三维模型的精度。Next, the PTZ load also extracts the feature points in the image, and obtains the depth information of the feature points from the depth map corresponding to the image; and then according to the feature points of the image and the The depth information is registered to obtain the relative pose between the images. In this embodiment, the depth map obtained by the back-projection of the three-dimensional point cloud is used to assist the registration process of the image, which is beneficial to improve the accuracy of image registration, thereby improving the accuracy of the subsequently generated three-dimensional model.
在第三种实现方式中,如果所述第二图像采集装置为可见光相机,考虑到可将光相机对于光照存在依赖性,在无光情况下,由可见光相机获取的图像序列可能是不可用的,基于此,为了保证图像配准的准确性,本申请实施例中所述云台负载为了统一三维点云中处于不同坐标系的点,使用所述运动数据将所述三维点云进行重投影,使得重投影后的三维点云中的每个点在同一坐标系下,而且考虑到惯性测量单元、所述第二图像采集装置以及所述第一图像采集装置均安装于云台负载上,所述惯性测量单元获取的运动数据也可以指示着所述第二图像采集装置和所述第二图像采集装置的运动过程,进而可以根据重投影后的三维点云以及所述运动数据生成与所述图像序列中的图像对应的深度图;最后根据所述深度图中提取的特征点进行配准,获取所述图像之间的相对位姿。本实施例利用三维点云反投影得到的深度图来代替所述图像进行配准,保证在无光情况下也能获取准确的图像配准结果,进而提高后续生成的三维模型的精度。In a third implementation manner, if the second image acquisition device is a visible light camera, considering that the light camera may be dependent on illumination, the image sequence acquired by the visible light camera may not be available in the absence of light , based on this, in order to ensure the accuracy of image registration, in order to unify points in different coordinate systems in the three-dimensional point cloud, the pan/tilt load in the embodiment of the present application uses the motion data to reproject the three-dimensional point cloud , so that each point in the reprojected three-dimensional point cloud is in the same coordinate system, and considering that the inertial measurement unit, the second image acquisition device and the first image acquisition device are all installed on the PTZ load, The motion data obtained by the inertial measurement unit can also indicate the motion process of the second image acquisition device and the second image acquisition device, and then can generate and match the motion data according to the re-projected three-dimensional point cloud and the motion data. The depth map corresponding to the images in the image sequence is obtained; finally, registration is performed according to the feature points extracted from the depth map to obtain the relative pose between the images. In this embodiment, the depth map obtained by the back-projection of the three-dimensional point cloud is used to perform registration instead of the image, so as to ensure that an accurate image registration result can be obtained even in the absence of light, thereby improving the accuracy of the subsequently generated three-dimensional model.
在一实施例中,为了达到实时在线生成三维模型的需求,可以使用分布式加速器 来获取所述三维点云之间的相对姿态以及所述图像之间的相对姿态,加快匹配过程,提高数据获取效率。In one embodiment, in order to meet the requirement of generating a 3D model online in real time, a distributed accelerator can be used to obtain the relative posture between the 3D point clouds and the relative posture between the images, so as to speed up the matching process and improve data acquisition. efficiency.
在获取所述三维点云之间的相对姿态以及所述图像之间的相对姿态之后,所述云台负载可以根据所述三维点云、所述图像序列、所述运动数据、所述三维点云之间的相对位姿以及所述图像之间的相对位姿,实时生成目标场景的三维模型,从而满足实时在线生成三维模型的需求。After acquiring the relative poses between the three-dimensional point clouds and the relative poses between the images, the PTZ load can be based on the three-dimensional point cloud, the image sequence, the motion data, the three-dimensional point The relative poses between the clouds and the relative poses between the images generate a three-dimensional model of the target scene in real time, so as to meet the requirement of generating a three-dimensional model online in real time.
进一步地,所述云台负载可以在生成所述三维模型之后,将所述三维模型实时传输给遥控终端,以在遥控终端上实时显示所述三维模型,从而用户可以实时了解三维重建进度,方便用户使用;用户如果在看到三维模型后发现有漏采的位置,也可以针对于显示在所述遥控终端上的三维模型输入反馈信息,所述反馈信息用于指示所述第一图像采集装置漏采的位置,由所述遥控终端将所述反馈信息传输给所述云台负载,使得所述云台负载可以控制所述第一图像采集装置采集所述目标场景中所述反馈信息指示的位置处的三维点云,实现在线监控过程,保证三维点云数据准确且有效地获取,有利于提高三维重建效率。Further, the PTZ load can transmit the 3D model to the remote control terminal in real time after generating the 3D model, so that the 3D model can be displayed on the remote control terminal in real time, so that the user can know the progress of the 3D reconstruction in real time, which is convenient. Used by the user; if the user finds a missing location after seeing the 3D model, he can also input feedback information for the 3D model displayed on the remote control terminal, where the feedback information is used to instruct the first image acquisition device The position of the missed acquisition, the remote control terminal transmits the feedback information to the PTZ load, so that the PTZ load can control the first image acquisition device to collect the information indicated by the feedback information in the target scene. The 3D point cloud at the location realizes the online monitoring process, ensures the accurate and effective acquisition of 3D point cloud data, and is conducive to improving the efficiency of 3D reconstruction.
在一实施例中,在实时生成目标场景的三维模型时,所述云台负载根据所述运动数据、所述三维点云之间的相对位姿以及所述图像之间的相对位姿,获取所述第一图像采集装置获取三维点云时的位姿和/或所述第二图像采集装置获取图像时的位姿;然后根据所述三维点云、所述图像序列以及所述第一图像采集装置获取三维点云时的位姿和/或所述第二图像采集装置获取图像时的位姿,实时生成三维模型。本实施例实现基于多种数据的融合生成高精度的三维模型。In one embodiment, when the 3D model of the target scene is generated in real time, the PTZ load is obtained according to the motion data, the relative pose between the 3D point clouds and the relative pose between the images. the pose when the first image acquisition device acquires the three-dimensional point cloud and/or the pose when the second image acquisition device acquires the image; and then according to the three-dimensional point cloud, the image sequence and the first image The pose when the acquisition device acquires the three-dimensional point cloud and/or the pose when the second image acquisition device acquires the image, generates a three-dimensional model in real time. This embodiment realizes the generation of a high-precision three-dimensional model based on the fusion of various data.
在一种实现方式中,所述云台负载首先建立位姿图模型,所述位姿图模型包括但不限于滑动窗口位姿图模型。然后利用所述运动数据、所述三维点云之间的相对位姿以及所述图像之间的相对位姿优化所述位姿图模型,获取所述第一图像采集装置获取三维点云时的位姿和/或所述第二图像采集装置获取图像时的位姿。其中,所述惯性测量单元获取的运动数据对应所述云台负载的位姿信息,由于惯性测量单元、所述第二图像采集装置以及所述第一图像采集装置均安装于云台负载上,所述第二图像采集装置和所述第一图像采集装置的运动过程与所述云台负载的运动过程是相同或相似的,因此所述惯性测量单元获取的运动数据也可以指示着所述第二图像采集装置和所述第二图像采集装置的运动过程,因此可以将所述运动数据作为优化所述位姿图模型的优化因素之一。例如,运动数据包含加速度,加速度进行积分得到速度,速度积分得到相对位移。相对位移就可以对云台负载两个位置间的位移进行约束。又如,运动数据 包含角速度,角速度积分后得到角度,进而得到相对姿态信息,相对姿态就可以对对云台负载两个姿态间的旋转进行约束。本实施例基于位姿图模型进行优化,获取高精度的位姿数据,使得基于高精度的位姿数据生成的三维模型具备更高的定位精度和鲁棒性。其中,在所述位姿图模型中,各个顶点被配置为指示所述第一图像采集装置采集三维点云时的位姿和/或所述第二图像采集装置获取图像时的位姿;以及,各个顶点之间的边被配置为指示所述三维点云之间的相对位姿、所述图像之间的相对位姿或者所述运动数据,通过各个边构成了各个顶点的约束关系,通过各个边指示的所述三维点云之间的相对位姿、所述图像之间的相对位姿或者所述运动数据来优化各个顶点中的所述第一图像采集装置采集三维点云时的位姿和/或所述第二图像采集装置获取图像时的位姿。In an implementation manner, the gimbal load first establishes a pose graph model, and the pose graph model includes but is not limited to a sliding window pose graph model. Then, the pose graph model is optimized by using the motion data, the relative poses between the three-dimensional point clouds, and the relative poses between the images, and the first image acquisition device acquires the three-dimensional point cloud. The pose and/or the pose when the second image acquisition device acquires the image. Wherein, the motion data obtained by the inertial measurement unit corresponds to the pose information of the gimbal load, since the inertial measurement unit, the second image acquisition device and the first image acquisition device are all installed on the gimbal load, The motion process of the second image acquisition device and the first image acquisition device is the same or similar to the motion process of the gimbal load, so the motion data acquired by the inertial measurement unit can also indicate the first image acquisition device. The motion process of the second image acquisition device and the second image acquisition device, therefore, the motion data can be used as one of the optimization factors for optimizing the pose graph model. For example, motion data includes acceleration, which is integrated to obtain velocity, and velocity is integrated to obtain relative displacement. The relative displacement can constrain the displacement between the two positions of the gimbal load. For another example, the motion data includes the angular velocity, and the angle is obtained after the angular velocity is integrated, and then the relative attitude information is obtained. The relative attitude can constrain the rotation between the two attitudes of the gimbal load. This embodiment performs optimization based on the pose graph model to obtain high-precision pose data, so that the three-dimensional model generated based on the high-precision pose data has higher positioning accuracy and robustness. Wherein, in the pose graph model, each vertex is configured to indicate the pose when the first image acquisition device acquires the three-dimensional point cloud and/or the pose when the second image acquisition device acquires an image; and , the edges between the vertices are configured to indicate the relative poses between the three-dimensional point clouds, the relative poses between the images, or the motion data, and each edge constitutes the constraint relationship of each vertex, through The relative pose between the three-dimensional point clouds indicated by each edge, the relative pose between the images, or the motion data are used to optimize the position of the first image acquisition device in each vertex when the three-dimensional point cloud is collected. pose and/or the pose when the second image acquisition device acquires the image.
在一个示例性的实施例中,所述位姿图模型包括滑动窗口位姿图模型,在每一次窗口的迭代优化过程中,选择第i(i为大于1的整数)个顶点指示的位姿以及至少一个参考顶点指示的位姿(参考顶点比如可以是第i-j个顶点,j为大于0的整数)作为窗口中要进行迭代优化的位姿,使用顶点之间的边指示的所述三维点云之间的相对位姿、所述图像之间的相对位姿或者所述运动数据来进行位姿优化,在优化完成后,滑动窗口以选择下一个顶点指示的位姿进行优化。In an exemplary embodiment, the pose graph model includes a sliding window pose graph model, and in each iterative optimization process of the window, the pose indicated by the ith (i is an integer greater than 1) vertex is selected and the pose indicated by at least one reference vertex (for example, the reference vertex can be the i-jth vertex, and j is an integer greater than 0) as the pose to be iteratively optimized in the window, using the three-dimensional point indicated by the edge between the vertices The relative pose between clouds, the relative pose between the images, or the motion data is used for pose optimization. After the optimization is completed, the window is slid to select the pose indicated by the next vertex for optimization.
在一个例子中,请参阅图4,假设所述第一图像采集装置和所述第二图像采集装置以相同的帧率获取数据,则每个顶点指示所述第一图像采集装置采集三维点云时的位姿和所述第二图像采集装置获取图像时的位姿,n为大于0的整数,各个边指示所述三维点云之间的相对位姿、所述图像之间的相对位姿或者所述运动数据;在另一个例子中,请参阅图5,假设所述第一图像采集装置和所述第二图像采集装置以不同的帧率获取数据,每个顶点指示所述第一图像采集装置采集三维点云时的位姿和/或所述第二图像采集装置获取图像时的位姿,n为大于0的整数,各个边指示所述三维点云之间的相对位姿、所述图像之间的相对位姿或者所述运动数据。In one example, referring to FIG. 4 , assuming that the first image acquisition device and the second image acquisition device acquire data at the same frame rate, each vertex instructs the first image acquisition device to acquire a three-dimensional point cloud n is an integer greater than 0, and each edge indicates the relative pose between the three-dimensional point clouds and the relative pose between the images or the motion data; in another example, please refer to FIG. 5 , assuming that the first image acquisition device and the second image acquisition device acquire data at different frame rates, and each vertex indicates the first image The pose when the collecting device collects the three-dimensional point cloud and/or the pose when the second image collecting device obtains the image, n is an integer greater than 0, and each edge indicates the relative pose between the three-dimensional point clouds, all The relative pose between the images or the motion data.
在一些场景下,在定位信号高于预定强度的场景下,比如室外场景,为了进一步提高三维模型的精度,所述云台负载还可以安装有定位器,所述定位器用于获取定位信息诸如GPS信息和/或RTK信息,所述定位器在所述第一图像采集装置采集三维点云以及所述第二图像采集装置获取图像序列的过程中采集定位信息,则请参阅图6,所述各个顶点之间的边还可以被配置为指示GPS信息和/或RTK信息,从而获取高精度的位姿信息,提高生成的三维模型的定位精度。In some scenarios, in scenarios where the positioning signal is higher than a predetermined strength, such as outdoor scenarios, in order to further improve the accuracy of the three-dimensional model, the PTZ load may also be equipped with a locator, which is used to obtain positioning information such as GPS information and/or RTK information, the locator collects positioning information during the process of collecting the three-dimensional point cloud by the first image collecting device and acquiring the image sequence by the second image collecting device, please refer to FIG. The edges between the vertices can also be configured to indicate GPS information and/or RTK information, so as to obtain high-precision pose information and improve the positioning accuracy of the generated three-dimensional model.
在一个实施例中,在获取所述第一图像采集装置获取三维点云时的位姿和/或所述 第二图像采集装置获取图像时的位姿之后,所述云台负载可以利用所述三维点云时的位姿将所述三维点云重投影至目标坐标系下,比如所述目标坐标系可以是世界坐标系,然后根据重投影后的三维点云生成初始三维模型,然后根据所述第二图像采集装置获取图像时的位姿以及所述图像序列中的多张图像对所述初始三维模型进行纹理映射,获取最终的三维模型。In one embodiment, after acquiring the pose when the first image acquisition device acquires the three-dimensional point cloud and/or the pose when the second image acquisition device acquires the image, the pan-tilt load can utilize the The pose of the three-dimensional point cloud reprojects the three-dimensional point cloud to the target coordinate system, for example, the target coordinate system can be the world coordinate system, and then generates an initial three-dimensional model according to the re-projected three-dimensional point cloud, and then according to the The pose when the second image acquisition device acquires the image and the multiple images in the image sequence perform texture mapping on the initial three-dimensional model to obtain the final three-dimensional model.
在获取所述三维模型之后,所述云台负载可以为所述三维模型添加标签索引后,将所述三维模型存储至所述标签索引指向的存储位置,以便后续使用过程中可以基于所述标签索引找到所述三维模型;和/或,将所述三维模型实时传输给遥控终端,以在遥控终端上实时显示所述三维模型,从而可以将三维重建进度实时展示给用户,方便用户使用。After acquiring the 3D model, the PTZ load may add a tag index to the 3D model, and then store the 3D model in the storage location pointed to by the tag index, so that the tag index can be used in subsequent use based on the tag index. The three-dimensional model is found by index; and/or, the three-dimensional model is transmitted to the remote control terminal in real time, so that the three-dimensional model is displayed on the remote control terminal in real time, so that the progress of the three-dimensional reconstruction can be displayed to the user in real time, which is convenient for the user to use.
为了进一步提高所述三维模型的精度,所述云台负载还可以根据所述三维点云或者所述图像序列进行回环检测,并根据回环检测的结果来优化所述位姿图模型,进而优化所述三维模型,保证优化后的三维模型的精确度和全局统一性。In order to further improve the accuracy of the 3D model, the PTZ load can also perform loop closure detection according to the 3D point cloud or the image sequence, and optimize the pose graph model according to the results of the loop closure detection, thereby optimizing all The 3D model is described to ensure the accuracy and global uniformity of the optimized 3D model.
可以理解的是,本申请实施例回环检测的具体实现过程不做任何限制,可依据实际应用场景进行具体设置。It can be understood that, the specific implementation process of loopback detection in this embodiment of the present application is not limited, and specific settings may be made according to actual application scenarios.
在一个例子中,可以根据所述图像序列或者所述三维点云重投影得到的深度图构建词袋模型,使得词袋模型来度量所述图像序列中的多张图像之间的相似度或者所述多张深度图像之间的相似度,所述云台负载可以根据所述图像序列中的多张图像之间的相似度进行回环检测,确定形成回环的图像;或者,根据所述三维点云重投影得到的多张深度图之间的相似度进行回环检测,确定形成回环的深度图,进而确定与所述深度图对应的图像;然后所述云台负载可以利用回环检测的结果对所述位姿图模型进行全局优化,即形成回环的图像之间的相对位姿来对所述位姿图模型进行全局优化,获取优化后的位姿信息,即优化后的所述第一图像采集装置采集三维点云时的位姿和/或所述第二图像采集装置获取图像时的位姿,接着根据全局优化后的位姿图模型更新所述三维模型,即根据优化后的位姿信息更新所述三维模型,保证优化后的三维模型的精确度和全局统一性。In one example, a bag-of-words model may be constructed according to the image sequence or the depth map obtained by re-projection of the three-dimensional point cloud, so that the bag-of-words model can measure the similarity or all of the images in the image sequence. The similarity between the multiple depth images, the pan-tilt load can perform loop closure detection according to the similarity between the multiple images in the image sequence, and determine the image forming a loop; or, according to the three-dimensional point cloud The similarity between the multiple depth maps obtained by reprojection is subjected to loop closure detection to determine the depth map forming a loop closure, and then to determine the image corresponding to the depth map; The pose graph model is globally optimized, that is, the relative poses between the looped images are formed to globally optimize the pose graph model, and the optimized pose information is obtained, that is, the optimized first image acquisition device The pose when collecting the three-dimensional point cloud and/or the pose when the second image acquisition device acquires the image, and then update the three-dimensional model according to the globally optimized pose graph model, that is, update according to the optimized pose information The three-dimensional model ensures the accuracy and global uniformity of the optimized three-dimensional model.
在另一个例子中,可以利用所述第一图像采集装置中存储的降采样后的全局模型和所述三维模型进行回环检测,确定形成回环的三维点云之间的相对位姿,然后所述云台负载可以利用回环检测的结果对所述位姿图模型进行全局优化,即形成回环的三维点云之间的相对位姿来对所述位姿图模型进行全局优化,获取优化后的位姿信息,即优化后的所述第一图像采集装置采集三维点云时的位姿和/或所述第二图像采集装 置获取图像时的位姿,接着根据全局优化后的位姿图模型更新所述三维模型,即根据优化后的位姿信息更新所述三维模型,保证优化后的三维模型的精确度和全局统一性。In another example, the down-sampled global model and the three-dimensional model stored in the first image acquisition device may be used to perform loop closure detection, to determine the relative poses between the three-dimensional point clouds forming loop closures, and then to perform loop closure detection. The gimbal load can globally optimize the pose graph model by using the result of the loopback detection, that is, the relative poses between the three-dimensional point clouds forming the loopback can be used to globally optimize the pose graph model, and the optimized pose graph model can be obtained. The pose information, that is, the optimized pose when the first image acquisition device collects the 3D point cloud and/or the pose when the second image acquisition device acquires the image, and then updated according to the globally optimized pose graph model The three-dimensional model is to update the three-dimensional model according to the optimized pose information, so as to ensure the accuracy and global unity of the optimized three-dimensional model.
在全局优化过程中,请参阅图7,所述位姿图模型中的边还被配置为指示形成回环的图像之间的相对位姿或者形成回环的三维点云之间的相对位姿,从而实现对所述位姿图模型进行优化,进而根据全局优化后的位姿图模型更新所述三维模型。During the global optimization process, referring to FIG. 7, the edges in the pose graph model are also configured to indicate the relative poses between the images forming the loops or the relative poses between the three-dimensional point clouds forming the loops, thereby The pose graph model is optimized, and the three-dimensional model is updated according to the globally optimized pose graph model.
在得到更新后的三维模型之后,所述云台负载可以将所述更新后的三维模型存储至对应的标签索引指向的存储位置;和/或,将所述更新后的三维模型实时传输给遥控终端,以在遥控终端上实时显示三维模型,从而用户可以实时了解三维重建进度,方便用户使用;用户如果在看到三维模型后发现有漏采的位置,也可以针对于显示在所述遥控终端上的三维模型输入反馈信息,所述反馈信息用于指示所述第一图像采集装置漏采的位置,由所述遥控终端将所述反馈信息传输给所述云台负载,使得所述云台负载可以控制所述第一图像采集装置采集所述目标场景中所述反馈信息指示的位置处的三维点云,实现在线监控过程,保证三维点云数据准确且有效地获取,有利于提高三维重建效率。After obtaining the updated 3D model, the PTZ load can store the updated 3D model in the storage location pointed to by the corresponding label index; and/or, transmit the updated 3D model to the remote control in real time terminal to display the 3D model in real time on the remote control terminal, so that the user can know the progress of the 3D reconstruction in real time, which is convenient for the user to use; if the user finds a missing location after seeing the 3D model, he can also target the display on the remote control terminal. Feedback information is input to the 3D model on the PTZ, and the feedback information is used to indicate the position where the first image acquisition device is missing, and the remote control terminal transmits the feedback information to the PTZ load, so that the PTZ The load can control the first image acquisition device to collect the three-dimensional point cloud at the position indicated by the feedback information in the target scene, realize the online monitoring process, ensure the accurate and effective acquisition of the three-dimensional point cloud data, and help improve the three-dimensional reconstruction. efficiency.
相应的,请参阅图8,本申请实施例还提供了一种云台负载20,包括第一图像采集装置21、第二图像采集装置22、惯性测量单元23、用于存储可执行指令的存储器24以及处理器25。Correspondingly, referring to FIG. 8 , an embodiment of the present application further provides a pan-tilt load 20 , including a first image acquisition device 21 , a second image acquisition device 22 , an inertial measurement unit 23 , and a memory for storing executable instructions 24 and processor 25.
所述第一图像采集装置21用于获取目标场景的三维点云。The first image acquisition device 21 is used to acquire a three-dimensional point cloud of the target scene.
所述第二图像采集装置22用于获取所述目标场景的图像序列。The second image acquisition device 22 is configured to acquire an image sequence of the target scene.
在采集所述三维点云和所述图像序列期间,所述惯性测量单元23用于获取所述云台负载的运动数据;所述运动数据包括所述云台负载的位姿信息。During the collection of the three-dimensional point cloud and the image sequence, the inertial measurement unit 23 is configured to acquire motion data of the gimbal load; the motion data includes the pose information of the gimbal load.
当所述处理器25执行所述可执行指令时,被配置为:When the processor 25 executes the executable instructions, it is configured to:
根据所述运动数据对所述三维点云进行配准,获取三维点云之间的相对位姿;以及,对所述图像序列进行配准,获取图像之间的相对位姿;The three-dimensional point cloud is registered according to the motion data to obtain the relative pose between the three-dimensional point clouds; and the image sequence is registered to obtain the relative pose between the images;
根据所述运动数据、所述三维点云之间的相对位姿以及所述图像之间的相对位姿,获取所述第一图像采集装置获取三维点云时的位姿和/或所述第二图像采集装置获取图像时的位姿;According to the motion data, the relative pose between the three-dimensional point clouds, and the relative pose between the images, obtain the pose and/or the first image acquisition device when acquiring the three-dimensional point cloud. 2. the pose when the image acquisition device acquires the image;
根据所述三维点云、所述图像序列以及所述第一图像采集装置获取三维点云时的位姿和/或所述第二图像采集装置获取图像时的位姿,生成目标场景的三维模型。A 3D model of the target scene is generated according to the 3D point cloud, the sequence of images, and the pose when the first image acquisition device acquires the 3D point cloud and/or the pose when the second image acquisition device acquires the image .
所述处理器25执行所述存储器24中包括的可执行指令,所述处理器25可以是中 央处理单元(Central Processing Unit,CPU),还可以是其他通用处理器、数字信号处理器(Digital Signal Processor,DSP)、专用集成电路(Application Specific Integrated Circuit,ASIC)、现成可编程门阵列(Field-Programmable Gate Array,FPGA)或者其他可编程逻辑器件、分立门或者晶体管逻辑器件、分立硬件组件等。通用处理器可以是微处理器或者该处理器也可以是任何常规的处理器等。The processor 25 executes the executable instructions included in the memory 24, and the processor 25 may be a central processing unit (Central Processing Unit, CPU), or other general-purpose processors, digital signal processors (Digital Signal Processors) Processor, DSP), Application Specific Integrated Circuit (ASIC), Field-Programmable Gate Array (FPGA) or other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components, etc. A general purpose processor may be a microprocessor or the processor may be any conventional processor or the like.
所述存储器24存储三维重建方法的可执行指令,所述存储器24可以包括至少一种类型的存储介质,存储介质包括闪存、硬盘、多媒体卡、卡型存储器(例如,SD或DX存储器等等)、随机访问存储器(RAM)、静态随机访问存储器(SRAM)、只读存储器(ROM)、电可擦除可编程只读存储器(EEPROM)、可编程只读存储器(PROM)、磁性存储器、磁盘、光盘等等。而且,云台负载20可以与通过网络连接执行存储器的存储功能的网络存储装置协作。存储器24可以是云台负载20的内部存储单元,例如云台负载20的硬盘或内存。存储器24也可以是云台负载20的外部存储设备,例如云台负载20上配备的插接式硬盘,智能存储卡(Smart Media Card,SMC),安全数字(Secure Digital,SD)卡,闪存卡(Flash Card)等。进一步地,存储器24还可以既包括云台负载20的内部存储单元也包括外部存储设备。存储器24用于存储可执行指令以及设备所需的其他程序和数据。存储器24还可以用于暂时地存储已经输出或者将要输出的数据。The memory 24 stores executable instructions for the three-dimensional reconstruction method, and the memory 24 may include at least one type of storage medium, including flash memory, hard disk, multimedia card, card-type memory (eg, SD or DX memory, etc.) , Random Access Memory (RAM), Static Random Access Memory (SRAM), Read Only Memory (ROM), Electrically Erasable Programmable Read Only Memory (EEPROM), Programmable Read Only Memory (PROM), Magnetic Memory, Disk, CD and so on. Also, the pan-tilt load 20 may cooperate with a network storage device that performs the storage function of the memory through a network connection. The memory 24 may be an internal storage unit of the gimbal load 20 , such as a hard disk or a memory of the gimbal load 20 . The memory 24 can also be an external storage device of the gimbal load 20, such as a pluggable hard disk, a smart memory card (Smart Media Card, SMC), a secure digital (Secure Digital, SD) card, and a flash memory card equipped on the gimbal load 20. (Flash Card) etc. Further, the memory 24 may also include both an internal storage unit of the pan-tilt load 20 and an external storage device. Memory 24 is used to store executable instructions and other programs and data required by the device. The memory 24 may also be used to temporarily store data that has been or will be output.
这里描述的各种实施方式可以使用例如计算机软件、硬件或其任何组合的计算机可读介质来实施。对于硬件实施,这里描述的实施方式可以通过使用特定用途集成电路(ASIC)、数字信号处理器(DSP)、数字信号处理装置(DSPD)、可编程逻辑装置(PLD)、现场可编程门阵列(FPGA)、处理器、控制器、微控制器、微处理器、被设计为执行这里描述的功能的电子单元中的至少一种来实施。对于软件实施,诸如过程或功能的实施方式可以与允许执行至少一种功能或操作的单独的软件模块来实施。软件代码可以由以任何适当的编程语言编写的软件应用程序(或程序)来实施,软件代码可以存储在存储器中并且由控制器执行。The various embodiments described herein can be implemented using computer readable media such as computer software, hardware, or any combination thereof. For hardware implementation, the embodiments described herein can be implemented using application specific integrated circuits (ASICs), digital signal processors (DSPs), digital signal processing devices (DSPDs), programmable logic devices (PLDs), field programmable gate arrays ( FPGA), processors, controllers, microcontrollers, microprocessors, electronic units designed to perform the functions described herein are implemented. For software implementation, embodiments such as procedures or functions may be implemented with separate software modules that allow the performance of at least one function or operation. The software codes may be implemented by a software application (or program) written in any suitable programming language, which may be stored in memory and executed by a controller.
本领域技术人员可以理解,图8仅仅是云台负载20的示例,并不构成对云台负载20的限定,可以包括比图示更多或更少的部件,或者组合某些部件,或者不同的部件,例如设备还可以包括输入输出设备、网络接入设备、总线等。Those skilled in the art can understand that FIG. 8 is only an example of the pan-tilt load 20, and does not constitute a limitation to the pan-tilt payload 20. It may include more or less components than the one shown, or combine some components, or different Components, such as devices, may also include input and output devices, network access devices, buses, and the like.
在一实施例中,所述第一图像采集装置21具体用于实时获取目标场景的三维点云;所述第二图像采集装置22具体用于实时获取所述目标场景的图像序列;以及所述惯性测量单元23具体用于实时获取所述云台负载的运动数据。In one embodiment, the first image acquisition device 21 is specifically configured to acquire the 3D point cloud of the target scene in real time; the second image acquisition device 22 is specifically configured to acquire the image sequence of the target scene in real time; and the The inertial measurement unit 23 is specifically configured to acquire motion data of the PTZ load in real time.
在一实施例中,所述目标场景包括室内场景或者室外非空旷场景,和/或,卫星定位信号低于预定强度的场景。In an embodiment, the target scene includes an indoor scene or an outdoor non-empty scene, and/or a scene where the satellite positioning signal is lower than a predetermined intensity.
在一实施例中,所述第一图像采集装置21包括以下至少一种或多种:激光雷达、双目视觉传感器以及结构光深度相机;In one embodiment, the first image acquisition device 21 includes at least one or more of the following: lidar, binocular vision sensor, and structured light depth camera;
所述第二图像采集装置22包括以下至少一种或多种:可见光相机、灰度相机以及红外相机。The second image acquisition device 22 includes at least one or more of the following: a visible light camera, a grayscale camera, and an infrared camera.
在一实施例中,所述处理器25具体用于:根据所述三维点云、所述图像序列以及所述第一图像采集装置获取三维点云时的位姿和/或所述第二图像采集装置获取图像时的位姿,实时生成目标场景的三维模型。In an embodiment, the processor 25 is specifically configured to: acquire the pose and/or the second image according to the three-dimensional point cloud, the image sequence, and the first image acquisition device when acquiring the three-dimensional point cloud The acquisition device acquires the pose of the image, and generates a three-dimensional model of the target scene in real time.
在一实施例中,在获取三维点云之间的相对位姿时,所述处理器25还用于:对于在连续时间序列上采集到的一组三维点云,根据该组三维点云中每个点对应的运动数据获取点之间的变换关系;In one embodiment, when acquiring the relative pose between the three-dimensional point clouds, the processor 25 is further configured to: for a group of three-dimensional point clouds collected in a continuous time series, according to the The transformation relationship between the motion data acquisition points corresponding to each point;
根据所述点之间的变换关系将该组三维点云中的点重投影至同一坐标系。The points in the set of three-dimensional point clouds are reprojected to the same coordinate system according to the transformation relationship between the points.
在一实施例中,所述处理器25还用于:根据该组三维点云中每个点对应的运动数据确定所述惯性测量单元23在预设时间间隔内的相对位姿;所述预设时间间隔为获取相邻点的时间间隔;In one embodiment, the processor 25 is further configured to: determine the relative pose of the inertial measurement unit 23 within a preset time interval according to the motion data corresponding to each point in the set of three-dimensional point clouds; Let the time interval be the time interval for obtaining adjacent points;
根据所述惯性测量单元23在预设时间间隔内的相对位姿、以及所述第一图像采集装置21与所述惯性测量单元23之间的外参转换关系,获取所述点之间的变换关系。Obtain the transformation between the points according to the relative pose of the inertial measurement unit 23 within a preset time interval and the external parameter conversion relationship between the first image acquisition device 21 and the inertial measurement unit 23 relation.
在一实施例中,所述每个点对应的运动数据根据所述点的获取时间所确定。In one embodiment, the motion data corresponding to each point is determined according to the acquisition time of the point.
在一实施例中,在获取三维点云之间的相对位姿时,所述处理器25还用于:In one embodiment, when acquiring the relative pose between the three-dimensional point clouds, the processor 25 is further configured to:
对于相邻两组三维点云,获取每组三维点云的面特征和/或拐角特征;其中,相邻两组三维点云中均包括有针对同一物体采集到的点;For two adjacent groups of 3D point clouds, obtain the surface features and/or corner features of each group of 3D point clouds; wherein, the adjacent two groups of 3D point clouds both include points collected for the same object;
根据相邻两组三维点云的面特征和/或拐角特征,获取相邻两组三维点云之间的相对位姿;所述相邻两组三维点云之间的相对位姿用于将相邻两组三维点云重投影至同一坐标系。According to the surface features and/or corner features of the adjacent two groups of three-dimensional point clouds, the relative poses between the two adjacent groups of three-dimensional point clouds are obtained; the relative poses between the two adjacent groups of three-dimensional point clouds are used to The adjacent two sets of 3D point clouds are reprojected to the same coordinate system.
在一实施例中,所述每组三维点云的面特征和/或拐角特征根据该组三维点云中三维点之间的曲度信息所确定。In one embodiment, the surface features and/or corner features of each group of three-dimensional point clouds are determined according to curvature information between three-dimensional points in the group of three-dimensional point clouds.
在一实施例中,所述每组三维点云的面特征根据曲度大于预设阈值的曲度信息所确定;In one embodiment, the surface features of each group of three-dimensional point clouds are determined according to curvature information whose curvature is greater than a preset threshold;
所述每组三维点云的拐角特征根据曲度小于或等于预设阈值的曲度信息所确定。The corner features of each group of three-dimensional point clouds are determined according to curvature information whose curvature is less than or equal to a preset threshold.
在一实施例中,所述处理器25还用于:根据相邻两组三维点云的面特征之间的相 似度,和/或,相邻两组三维点云的拐角特征之间的相似度,获取三维点云之间的相对位姿。In one embodiment, the processor 25 is further configured to: according to the similarity between the surface features of the adjacent two groups of three-dimensional point clouds, and/or the similarity between the corner features of the adjacent two groups of three-dimensional point clouds degrees to obtain the relative pose between 3D point clouds.
在一实施例中,所述相邻两组三维点云的面特征之间的相似度基于所述相邻两组三维点云的面特征之间的距离确定;和/或,In one embodiment, the similarity between the surface features of the adjacent two groups of three-dimensional point clouds is determined based on the distance between the surface features of the two adjacent groups of three-dimensional point clouds; and/or,
所述相邻两组三维点云的拐角特征之间的相似度基于所述相邻两组三维点云的拐角特征之间的距离确定。The similarity between the corner features of the adjacent two groups of three-dimensional point clouds is determined based on the distance between the corner features of the two adjacent groups of three-dimensional point clouds.
在一实施例中,所述面特征包括指示平面的至少一个三维坐标信息以及该平面的法向量;和/或,In an embodiment, the surface feature includes at least one three-dimensional coordinate information indicating a plane and a normal vector of the plane; and/or,
所述拐角特征包括指示边缘的至少一个三维坐标信息以及指示该边缘的向量。The corner feature includes at least one three-dimensional coordinate information indicating an edge and a vector indicating the edge.
在一实施例中,所述图像序列包括在连续的时间序列上获取的多个图像;In one embodiment, the sequence of images includes a plurality of images acquired in a continuous time sequence;
在获取图像之间的相对位姿时,所述处理器25还用于提取所述图像中的特征点,并基于所述图像的特征点进行配准,获取所述图像之间的相对位姿。When acquiring the relative poses between the images, the processor 25 is further configured to extract feature points in the images, and perform registration based on the feature points of the images to obtain the relative poses between the images .
在一实施例中,所述图像序列包括在连续的时间序列上获取的多个图像;In one embodiment, the sequence of images includes a plurality of images acquired in a continuous time sequence;
在获取图像之间的相对位姿时,所述处理器25还用于:When acquiring the relative pose between the images, the processor 25 is further configured to:
使用所述运动数据将所述三维点云进行重投影,并根据重投影后的三维点云以及所述运动数据生成与所述图像对应的深度图;reprojecting the 3D point cloud using the motion data, and generating a depth map corresponding to the image according to the reprojected 3D point cloud and the motion data;
提取所述图像中的特征点,并从所述图像对应的深度图中获取所述特征点的深度信息;extracting feature points in the image, and obtaining depth information of the feature points from a depth map corresponding to the image;
根据所述图像的特征点以及所述特征点的深度信息进行配准,获取所述图像之间的相对位姿。The registration is performed according to the feature points of the images and the depth information of the feature points, and the relative poses between the images are acquired.
在一实施例中,在获取图像之间的相对位姿时,所述处理器25还用于:在无光情况下,使用所述运动数据将所述三维点云进行重投影,并根据重投影后的三维点云以及所述运动数据生成与所述图像序列中的图像对应的深度图;根据所述深度图中提取的特征点进行配准,获取所述图像之间的相对位姿。In one embodiment, when acquiring the relative pose between the images, the processor 25 is further configured to: in the absence of light, use the motion data to reproject the three-dimensional point cloud, and re-project the three-dimensional point cloud according to the re-projection. The projected three-dimensional point cloud and the motion data generate a depth map corresponding to the images in the image sequence; perform registration according to the feature points extracted from the depth map to obtain the relative poses between the images.
在一实施例中,所述处理器25还用于:In one embodiment, the processor 25 is further configured to:
建立位姿图模型;Build a pose graph model;
利用所述运动数据、所述三维点云之间的相对位姿以及所述图像之间的相对位姿优化所述位姿图模型,获取所述第一图像采集装置21获取三维点云时的位姿和/或所述第二图像采集装置22获取图像时的位姿。The pose graph model is optimized by using the motion data, the relative poses between the three-dimensional point clouds, and the relative poses between the images, and obtains the data obtained when the first image acquisition device 21 obtains the three-dimensional point cloud. The pose and/or the pose when the second image acquisition device 22 acquires the image.
在一实施例中,在所述位姿图模型中,各个顶点被配置为指示所述第一图像采集装置21采集三维点云时的位姿和/或所述第二图像采集装置22获取图像时的位姿;In one embodiment, in the pose graph model, each vertex is configured to instruct the first image acquisition device 21 to acquire a pose when acquiring a three-dimensional point cloud and/or the second image acquisition device 22 to acquire an image. posture when
以及,各个顶点之间的边被配置为指示所述三维点云之间的相对位姿、所述图像之间的相对位姿或者所述运动数据。And, the edges between the vertices are configured to indicate relative poses between the three-dimensional point clouds, relative poses between the images, or the motion data.
在一实施例中,所述各个顶点之间的边还被配置为指示GPS信息和/或RTK信息。In one embodiment, the edges between the various vertices are further configured to indicate GPS information and/or RTK information.
在一实施例中,所述位姿图模型包括滑动窗口位姿图模型。In one embodiment, the pose graph model includes a sliding window pose graph model.
在一实施例中,所述处理器25还用于:为所述三维模型添加标签索引后,将所述三维模型存储至所述标签索引指向的存储位置;和/或,将所述三维模型实时传输给遥控终端,以在遥控终端上实时显示所述三维模型。In one embodiment, the processor 25 is further configured to: after adding a label index to the 3D model, store the 3D model in the storage location pointed to by the label index; and/or, store the 3D model Real-time transmission to the remote control terminal to display the three-dimensional model in real time on the remote control terminal.
在一实施例中,所述处理器25还用于:In one embodiment, the processor 25 is further configured to:
根据所述图像序列中的多张图像之间的相似度进行回环检测;或者,根据所述三维点云重投影得到的多张深度图之间的相似度进行回环检测;或者,根据降采样后的全局模型与所述三维模型进行回环检测;Loop closure detection is performed according to the similarity between multiple images in the image sequence; or, loop closure detection is performed according to the similarity between multiple depth maps obtained by re-projection of the 3D point cloud; The global model and the three-dimensional model for loop closure detection;
利用回环检测的结果对所述位姿图模型进行全局优化;以及,根据全局优化后的位姿图模型更新所述三维模型。The pose graph model is globally optimized using the result of loop closure detection; and the three-dimensional model is updated according to the globally optimized pose graph model.
在一实施例中,所述处理器25还用于:利用形成回环的图像之间的相对位姿或者形成回环的点之间的相对位姿,对所述位姿图模型进行全局优化。In one embodiment, the processor 25 is further configured to: perform global optimization on the pose graph model by using the relative pose between the images forming the loop or the relative pose between the points forming the loop.
在一实施例中,使用词袋模型来度量所述图像序列中的多张图像之间的相似度或者所述多张深度图像之间的相似度。In one embodiment, a bag-of-words model is used to measure the similarity between multiple images in the image sequence or the similarity between the multiple depth images.
在一实施例中,所述处理器25还用于:In one embodiment, the processor 25 is further configured to:
根据所述三维模型的精度要求和/或密度要求,生成采集轨迹和采集时间;generating an acquisition trajectory and acquisition time according to the accuracy requirements and/or density requirements of the three-dimensional model;
根据所述采集轨迹和采集时间控制所述第一图像采集装置21、所述第二图像采集装置22以及所述惯性测量单元23采集数据。The first image acquisition device 21 , the second image acquisition device 22 and the inertial measurement unit 23 are controlled to acquire data according to the acquisition track and acquisition time.
在一实施例中,所述处理器25还用于:In one embodiment, the processor 25 is further configured to:
接收所述遥控终端传输的反馈信息;所述反馈信息用于指示所述第一图像采集装置21漏采的位置,所述反馈信息由用户针对于显示在所述遥控终端上的三维模型输入;Receive feedback information transmitted by the remote control terminal; the feedback information is used to indicate the position of the missed acquisition by the first image acquisition device 21, and the feedback information is input by the user with respect to the three-dimensional model displayed on the remote control terminal;
控制所述第一图像采集装置21采集所述目标场景中所述反馈信息指示的位置处的三维点云。The first image acquisition device 21 is controlled to acquire the three-dimensional point cloud at the position indicated by the feedback information in the target scene.
上述设备中各个单元的功能和作用的实现过程具体详见上述方法中对应步骤的实现过程,在此不再赘述。For details of the implementation process of the functions and functions of the units in the above device, please refer to the implementation process of the corresponding steps in the above method, which will not be repeated here.
相应的,本申请实施例还提供了一种可移动平台,包括上述的云台负载。Correspondingly, an embodiment of the present application also provides a movable platform, including the above-mentioned PTZ load.
其中,所述可移动平台至少包括无人飞行器、无人驾驶车辆或者移动机器人。Wherein, the movable platform includes at least an unmanned aerial vehicle, an unmanned vehicle or a mobile robot.
在示例性实施例中,还提供了一种包括指令的非临时性计算机可读存储介质,例 如包括指令的存储器,上述指令可由装置的处理器执行以完成上述方法。例如,非临时性计算机可读存储介质可以是ROM、随机存取存储器(RAM)、CD-ROM、磁带、软盘和光数据存储设备等。In an exemplary embodiment, there is also provided a non-transitory computer-readable storage medium, such as a memory including instructions, executable by a processor of an apparatus to perform the above-described method. For example, the non-transitory computer-readable storage medium may be ROM, random access memory (RAM), CD-ROM, magnetic tape, floppy disk, optical data storage device, and the like.
一种非临时性计算机可读存储介质,当存储介质中的指令由终端的处理器执行时,使得终端能够执行上述方法。A non-transitory computer-readable storage medium, when the instructions in the storage medium are executed by the processor of the terminal, enable the terminal to execute the above method.
需要说明的是,在本文中,诸如第一和第二等之类的关系术语仅仅用来将一个实体或者操作与另一个实体或操作区分开来,而不一定要求或者暗示这些实体或操作之间存在任何这种实际的关系或者顺序。术语“包括”、“包含”或者其任何其他变体意在涵盖非排他性的包含,从而使得包括一系列要素的过程、方法、物品或者设备不仅包括那些要素,而且还包括没有明确列出的其他要素,或者是还包括为这种过程、方法、物品或者设备所固有的要素。在没有更多限制的情况下,由语句“包括一个……”限定的要素,并不排除在包括所述要素的过程、方法、物品或者设备中还存在另外的相同要素。It should be noted that, in this document, relational terms such as first and second are only used to distinguish one entity or operation from another entity or operation, and do not necessarily require or imply any relationship between these entities or operations. any such actual relationship or sequence exists. The terms "comprising", "comprising" or any other variation thereof are intended to encompass non-exclusive inclusion such that a process, method, article or device comprising a list of elements includes not only those elements, but also other not expressly listed elements, or also include elements inherent to such a process, method, article or apparatus. Without further limitation, an element qualified by the phrase "comprising a..." does not preclude the presence of additional identical elements in a process, method, article or apparatus that includes the element.
以上对本申请实施例所提供的方法和装置进行了详细介绍,本文中应用了具体个例对本申请的原理及实施方式进行了阐述,以上实施例的说明只是用于帮助理解本申请的方法及其核心思想;同时,对于本领域的一般技术人员,依据本申请的思想,在具体实施方式及应用范围上均会有改变之处,综上所述,本说明书内容不应理解为对本申请的限制。The methods and devices provided by the embodiments of the present application have been introduced in detail above, and specific examples are used to illustrate the principles and implementations of the present application. At the same time, for those of ordinary skill in the art, according to the idea of the application, there will be changes in the specific implementation and application scope. In summary, the content of this specification should not be construed as a limitation to the application. .

Claims (58)

  1. 一种三维重建方法,其特征在于,应用于云台负载上,所述云台负载设有第一图像采集装置、第二图像采集装置,所述方法包括:A three-dimensional reconstruction method, characterized in that it is applied to a pan-tilt load, wherein the pan-tilt load is provided with a first image acquisition device and a second image acquisition device, and the method comprises:
    获取目标场景的三维点云、图像序列以及所述云台负载的运动数据;其中,所述三维点云由所述第一图像采集装置获取,所述图像序列由所述第二图像采集装置获取;所述运动数据包括:在获取目标场景的三维点云、图像序列的期间,所述云台负载的位姿信息;Acquiring a 3D point cloud, an image sequence, and motion data of the pan/tilt load of the target scene; wherein, the 3D point cloud is acquired by the first image acquisition device, and the image sequence is acquired by the second image acquisition device ; The motion data includes: during the acquisition of the three-dimensional point cloud and the image sequence of the target scene, the pose information of the load on the PTZ;
    根据所述运动数据对所述三维点云进行配准,获取三维点云之间的相对位姿;以及,对所述图像序列进行配准,获取图像之间的相对位姿;The three-dimensional point cloud is registered according to the motion data to obtain the relative pose between the three-dimensional point clouds; and the image sequence is registered to obtain the relative pose between the images;
    根据所述运动数据、所述三维点云之间的相对位姿以及所述图像之间的相对位姿,获取所述第一图像采集装置获取三维点云时的位姿和/或所述第二图像采集装置获取图像时的位姿;According to the motion data, the relative pose between the three-dimensional point clouds, and the relative pose between the images, obtain the pose and/or the first image acquisition device when acquiring the three-dimensional point cloud. 2. the pose when the image acquisition device acquires the image;
    根据所述三维点云、所述图像序列,以及所述第一图像采集装置获取三维点云时的位姿和/或所述第二图像采集装置获取图像时的位姿,生成目标场景的三维模型。According to the 3D point cloud, the image sequence, and the pose when the first image acquisition device acquires the 3D point cloud and/or the pose when the second image acquisition device acquires the image, a 3D image of the target scene is generated Model.
  2. 根据权利要求1所述的方法,其特征在于,所述获取目标场景的三维点云、图像序列以及所述云台负载的运动数据,包括:实时获取目标场景的三维点云、图像序列以及所述云台负载的运动数据。The method according to claim 1, wherein the acquiring the 3D point cloud of the target scene, the image sequence and the motion data loaded by the PTZ comprises: acquiring the 3D point cloud, the image sequence and all the motion data of the target scene in real time. Describe the motion data of the gimbal load.
  3. 根据权利要求1所述的方法,其特征在于,所述目标场景包括室内场景或者室外非空旷场景;和/或,卫星定位信号低于预定强度的场景。The method according to claim 1, wherein the target scene includes an indoor scene or an outdoor non-empty scene; and/or a scene where the satellite positioning signal is lower than a predetermined intensity.
  4. 根据权利要求1或2所述的方法,其特征在于,所述云台负载还安装有惯性测量单元;The method according to claim 1 or 2, wherein an inertial measurement unit is further installed on the pan-tilt load;
    所述第一图像采集装置用于获取所述目标场景的三维点云;The first image acquisition device is configured to acquire a three-dimensional point cloud of the target scene;
    所述第二图像采集装置用于获取所述目标场景的图像序列;The second image acquisition device is used for acquiring the image sequence of the target scene;
    所述惯性测量单元用于获取所述云台负载的运动数据。The inertial measurement unit is used for acquiring motion data of the PTZ load.
  5. 根据权利要求4所述的方法,其特征在于,所述第一图像采集装置包括以下至少一种或多种:激光雷达、双目视觉传感器以及结构光深度相机;所述第二图像采集装置包括以下至少一种或多种:可见光相机、灰度相机以及红外相机。The method according to claim 4, wherein the first image acquisition device comprises at least one or more of the following: lidar, binocular vision sensor and structured light depth camera; the second image acquisition device comprises At least one or more of the following: visible light cameras, grayscale cameras, and infrared cameras.
  6. 根据权利要求1所述的方法,其特征在于,所述根据所述三维点云、所述图像序列以及所述第一图像采集装置获取三维点云时的位姿和/或所述第二图像采集装置获取图像时的位姿,生成目标场景的三维模型,包括:The method according to claim 1, wherein the pose and/or the second image when acquiring the three-dimensional point cloud according to the three-dimensional point cloud, the image sequence and the first image acquisition device The acquisition device acquires the pose of the image, and generates a three-dimensional model of the target scene, including:
    根据所述三维点云、所述图像序列以及所述第一图像采集装置获取三维点云时的 位姿和/或所述第二图像采集装置获取图像时的位姿,实时生成目标场景的三维模型。According to the 3D point cloud, the image sequence, and the pose when the first image acquisition device acquires the 3D point cloud and/or the pose when the second image acquisition device acquires the image, a 3D image of the target scene is generated in real time Model.
  7. 根据权利要求4所述的方法,其特征在于,所述根据所述运动数据对所述三维点云进行配准,获取三维点云之间的相对位姿,包括:The method according to claim 4, wherein the registering the three-dimensional point cloud according to the motion data, and obtaining the relative pose between the three-dimensional point clouds, comprises:
    对于在连续时间序列上采集到的一组三维点云,根据该组三维点云中每个点对应的运动数据获取点之间的变换关系;For a group of 3D point clouds collected on a continuous time series, obtain the transformation relationship between points according to the motion data corresponding to each point in the group of 3D point clouds;
    根据所述点之间的变换关系将该组三维点云中的点重投影至同一坐标系。The points in the set of three-dimensional point clouds are reprojected to the same coordinate system according to the transformation relationship between the points.
  8. 根据权利要求7所述的方法,其特征在于,所述根据该组三维点云中每个点对应的运动数据获取所述点之间的变换关系,包括:The method according to claim 7, wherein the obtaining the transformation relationship between the points according to the motion data corresponding to each point in the set of three-dimensional point clouds comprises:
    根据该组三维点云中每个点对应的运动数据确定所述惯性测量单元在预设时间间隔内的相对位姿;所述预设时间间隔为获取相邻点的时间间隔;Determine the relative pose of the inertial measurement unit within a preset time interval according to the motion data corresponding to each point in the group of three-dimensional point clouds; the preset time interval is the time interval for obtaining adjacent points;
    根据所述惯性测量单元在预设时间间隔内的相对位姿、以及所述第一图像采集装置与所述惯性测量单元之间的外参转换关系,获取所述点之间的变换关系。The transformation relationship between the points is acquired according to the relative pose of the inertial measurement unit within a preset time interval and the external parameter transformation relationship between the first image acquisition device and the inertial measurement unit.
  9. 根据权利要求7所述的方法,其特征在于,所述每个点对应的运动数据根据所述点的获取时间所确定。The method according to claim 7, wherein the motion data corresponding to each point is determined according to the acquisition time of the point.
  10. 根据权利要求7所述的方法,其特征在于,所述根据所述运动数据对所述三维点云进行配准,获取三维点云之间的相对位姿,还包括:The method according to claim 7, wherein the registering the three-dimensional point cloud according to the motion data to obtain the relative pose between the three-dimensional point clouds, further comprising:
    对于相邻两组三维点云,获取每组三维点云的面特征和/或拐角特征;其中,相邻两组三维点云中均包括有针对同一物体采集到的点;For two adjacent groups of 3D point clouds, obtain the surface features and/or corner features of each group of 3D point clouds; wherein, the adjacent two groups of 3D point clouds both include points collected for the same object;
    根据相邻两组三维点云的面特征和/或拐角特征,获取相邻两组三维点云之间的相对位姿;所述相邻两组三维点云之间的相对位姿用于将相邻两组三维点云重投影至同一坐标系。According to the surface features and/or corner features of the adjacent two groups of three-dimensional point clouds, the relative poses between the two adjacent groups of three-dimensional point clouds are obtained; the relative poses between the two adjacent groups of three-dimensional point clouds are used to The adjacent two sets of 3D point clouds are reprojected to the same coordinate system.
  11. 根据权利要求10所述的方法,其特征在于,所述每组三维点云的面特征和/或拐角特征根据该组三维点云中点之间的曲度信息所确定。The method according to claim 10, wherein the surface features and/or corner features of each group of three-dimensional point clouds are determined according to the curvature information between the middle points of the group of three-dimensional point clouds.
  12. 根据权利要求11所述的方法,其特征在于,所述每组三维点云的面特征根据曲度大于预设阈值的曲度信息所确定;The method according to claim 11, wherein the surface features of each group of three-dimensional point clouds are determined according to curvature information whose curvature is greater than a preset threshold;
    所述每组三维点云的拐角特征根据曲度小于或等于预设阈值的曲度信息所确定。The corner features of each group of three-dimensional point clouds are determined according to curvature information whose curvature is less than or equal to a preset threshold.
  13. 根据权利要求10所述的方法,其特征在于,所述根据相邻两组三维点云的面特征和/或拐角特征,获取三维点云之间的相对位姿,包括:The method according to claim 10, wherein the obtaining the relative pose between the three-dimensional point clouds according to the surface features and/or corner features of two adjacent groups of three-dimensional point clouds, comprising:
    根据相邻两组三维点云的面特征之间的相似度,和/或,相邻两组三维点云的拐角特征之间的相似度,获取三维点云之间的相对位姿。The relative pose between the three-dimensional point clouds is obtained according to the similarity between the surface features of the two adjacent groups of three-dimensional point clouds, and/or the similarity between the corner features of the two adjacent groups of three-dimensional point clouds.
  14. 根据权利要求13所述的方法,其特征在于,所述相邻两组三维点云的面特征 之间的相似度基于所述相邻两组三维点云的面特征之间的距离确定;和/或,The method according to claim 13, wherein the similarity between the surface features of the adjacent two groups of three-dimensional point clouds is determined based on the distance between the surface features of the adjacent two groups of three-dimensional point clouds; and /or,
    所述相邻两组三维点云的拐角特征之间的相似度基于所述相邻两组三维点云的拐角特征之间的距离确定。The similarity between the corner features of the adjacent two groups of three-dimensional point clouds is determined based on the distance between the corner features of the two adjacent groups of three-dimensional point clouds.
  15. 根据权利要求10所述的方法,其特征在于,所述面特征包括指示平面的至少一个三维坐标信息以及该平面的法向量;和/或,The method according to claim 10, wherein the surface feature comprises at least one three-dimensional coordinate information indicating a plane and a normal vector of the plane; and/or,
    所述拐角特征包括指示边缘的至少一个三维坐标信息以及指示该边缘的向量。The corner feature includes at least one three-dimensional coordinate information indicating an edge and a vector indicating the edge.
  16. 根据权利要求1所述的方法,其特征在于,所述图像序列包括在连续的时间序列上获取的多个图像;The method of claim 1, wherein the image sequence comprises a plurality of images acquired in a continuous time series;
    所述对所述图像序列进行配准,获取图像之间的相对位姿,包括:The registering the image sequence to obtain the relative pose between the images includes:
    提取所述图像中的特征点,并基于所述图像的特征点进行配准,获取所述图像之间的相对位姿。Feature points in the images are extracted, and registration is performed based on the feature points of the images to obtain relative poses between the images.
  17. 根据权利要求1所述的方法,其特征在于,所述图像序列包括在连续的时间序列上获取的多个图像;The method of claim 1, wherein the image sequence comprises a plurality of images acquired in a continuous time series;
    所述对所述图像序列进行配准,获取图像之间的相对位姿,包括:The registering the image sequence to obtain the relative pose between the images includes:
    使用所述运动数据将所述三维点云进行重投影,并根据重投影后的三维点云以及所述运动数据生成与所述图像对应的深度图;reprojecting the 3D point cloud using the motion data, and generating a depth map corresponding to the image according to the reprojected 3D point cloud and the motion data;
    提取所述图像中的特征点,并从所述图像对应的深度图中获取所述特征点的深度信息;extracting feature points in the image, and obtaining depth information of the feature points from a depth map corresponding to the image;
    根据所述图像的特征点以及所述特征点的深度信息进行配准,获取所述图像之间的相对位姿。The registration is performed according to the feature points of the images and the depth information of the feature points, and the relative poses between the images are acquired.
  18. 根据权利要求1所述的方法,其特征在于,所述对所述图像序列进行配准,获取图像之间的相对位姿,还包括:The method according to claim 1, wherein the registering the image sequence to obtain the relative pose between the images further comprises:
    在无光情况下,使用所述运动数据将所述三维点云进行重投影,并根据重投影后的三维点云以及所述运动数据生成与所述图像序列中的图像对应的深度图;In the absence of light, the three-dimensional point cloud is re-projected using the motion data, and a depth map corresponding to the images in the image sequence is generated according to the re-projected three-dimensional point cloud and the motion data;
    根据所述深度图中提取的特征点进行配准,获取所述图像之间的相对位姿。The registration is performed according to the feature points extracted from the depth map, and the relative pose between the images is obtained.
  19. 根据权利要求1所述的方法,其特征在于,The method of claim 1, wherein:
    所述根据所述运动数据、所述三维点云之间的相对位姿以及所述图像之间的相对位姿,获取所述第一图像采集装置获取三维点云时的位姿和/或所述第二图像采集装置获取图像时的位姿,包括:According to the motion data, the relative pose between the three-dimensional point clouds, and the relative pose between the images, the pose and/or the pose and/or the pose when the first image acquisition device acquires the three-dimensional point cloud is acquired. Describe the pose when the second image acquisition device acquires the image, including:
    建立位姿图模型;Build a pose graph model;
    利用所述运动数据、所述三维点云之间的相对位姿以及所述图像之间的相对位姿 优化所述位姿图模型,获取所述第一图像采集装置获取三维点云时的位姿和/或所述第二图像采集装置获取图像时的位姿。The pose graph model is optimized by using the motion data, the relative poses between the three-dimensional point clouds, and the relative poses between the images, and obtains the pose when the first image acquisition device acquires the three-dimensional point cloud. pose and/or the pose when the second image acquisition device acquires the image.
  20. 根据权利要求19所述的方法,其特征在于,在所述位姿图模型中,各个顶点被配置为指示所述第一图像采集装置采集三维点云时的位姿和/或所述第二图像采集装置获取图像时的位姿;The method according to claim 19, wherein, in the pose graph model, each vertex is configured to indicate the pose and/or the second pose when the first image capturing device collects the three-dimensional point cloud the pose when the image acquisition device acquires the image;
    以及,各个顶点之间的边被配置为指示所述三维点云之间的相对位姿、所述图像之间的相对位姿或者所述运动数据。And, the edges between the vertices are configured to indicate relative poses between the three-dimensional point clouds, relative poses between the images, or the motion data.
  21. 根据权利要求20所述的方法,其特征在于,所述各个顶点之间的边还被配置为指示定位信息;所述定位信息至少包括GPS信息和/或RTK信息。The method according to claim 20, wherein the edges between the vertices are further configured to indicate positioning information; the positioning information includes at least GPS information and/or RTK information.
  22. 根据权利要求19所述的方法,其特征在于,所述位姿图模型包括滑动窗口位姿图模型。The method of claim 19, wherein the pose graph model comprises a sliding window pose graph model.
  23. 根据权利要求19所述的方法,其特征在于,还包括:The method of claim 19, further comprising:
    为所述三维模型添加标签索引后,将所述三维模型存储至所述标签索引指向的存储位置;和/或,将所述三维模型实时传输给遥控终端,以在遥控终端上实时显示所述三维模型。After adding a label index to the 3D model, store the 3D model in a storage location pointed to by the label index; and/or, transmit the 3D model to a remote control terminal in real time, so as to display the 3D model on the remote control terminal in real time 3D model.
  24. 根据权利要求1所述的方法,其特征在于,在所述根据所述三维点云、所述图像序列以及所述第一图像采集装置获取三维点云时的位姿和/或所述第二图像采集装置获取图像时的位姿,生成目标场景的三维模型之后,还包括:The method according to claim 1, characterized in that, when the three-dimensional point cloud is acquired according to the three-dimensional point cloud, the image sequence and the first image acquisition device, the pose and/or the second After the image acquisition device acquires the pose of the image and generates the three-dimensional model of the target scene, it further includes:
    根据所述图像序列中的多张图像之间的相似度进行回环检测;或者,根据所述三维点云重投影得到的多张深度图之间的相似度进行回环检测;或者,根据降采样后的全局模型与所述三维模型进行回环检测;Loop closure detection is performed according to the similarity between multiple images in the image sequence; or, loop closure detection is performed according to the similarity between multiple depth maps obtained by re-projection of the 3D point cloud; The global model and the three-dimensional model for loop closure detection;
    利用回环检测的结果对所述位姿图模型进行全局优化;以及,using the results of loop closure detection to globally optimize the pose graph model; and,
    根据全局优化后的位姿图模型更新所述三维模型。The three-dimensional model is updated according to the globally optimized pose graph model.
  25. 根据权利要求24所述的方法,其特征在于,所述利用回环检测的结果对所述位姿图模型进行全局优化,包括:The method according to claim 24, wherein the performing global optimization on the pose graph model using a result of loop closure detection comprises:
    利用形成回环的图像之间的相对位姿或者形成回环的三维点云之间的相对位姿,对所述位姿图模型进行全局优化。The pose graph model is globally optimized using the relative poses between the images forming the loops or the relative poses between the three-dimensional point clouds forming the loops.
  26. 根据权利要求24所述的方法,其特征在于,使用词袋模型来度量所述图像序列中的多张图像之间的相似度或者所述多张深度图像之间的相似度。The method according to claim 24, wherein a bag-of-words model is used to measure the similarity between multiple images in the image sequence or the similarity between the multiple depth images.
  27. 根据权利要求4所述的方法,其特征在于,还包括:The method of claim 4, further comprising:
    根据所述三维模型的精度要求和/或密度要求,生成采集轨迹和采集时间;generating an acquisition trajectory and acquisition time according to the accuracy requirements and/or density requirements of the three-dimensional model;
    根据所述采集轨迹和采集时间控制所述第一图像采集装置、所述第二图像采集装置以及所述惯性测量单元采集数据。The first image acquisition device, the second image acquisition device and the inertial measurement unit are controlled to acquire data according to the acquisition trajectory and acquisition time.
  28. 根据权利要求23所述的方法,其特征在于,还包括:The method of claim 23, further comprising:
    接收所述遥控终端传输的反馈信息;所述反馈信息用于指示所述第一图像采集装置漏采的位置,所述反馈信息由用户针对于显示在所述遥控终端上的三维模型输入;Receive feedback information transmitted by the remote control terminal; the feedback information is used to indicate the position of the missed acquisition by the first image acquisition device, and the feedback information is input by the user with respect to the three-dimensional model displayed on the remote control terminal;
    控制所述第一图像采集装置采集所述目标场景中所述反馈信息指示的位置处的三维点云。The first image acquisition device is controlled to acquire the three-dimensional point cloud at the position indicated by the feedback information in the target scene.
  29. 一种云台负载,其特征在于,包括第一图像采集装置、第二图像采集装置、惯性测量单元、用于存储可执行指令的存储器以及处理器;A pan-tilt load, characterized by comprising a first image acquisition device, a second image acquisition device, an inertial measurement unit, a memory for storing executable instructions, and a processor;
    所述第一图像采集装置用于获取目标场景的三维点云;The first image acquisition device is used to acquire a three-dimensional point cloud of the target scene;
    所述第二图像采集装置用于获取所述目标场景的图像序列;The second image acquisition device is used for acquiring the image sequence of the target scene;
    在采集所述三维点云和所述图像序列期间,所述惯性测量单元用于获取所述云台负载的运动数据;所述运动数据包括所述云台负载的位姿信息;During the collection of the three-dimensional point cloud and the image sequence, the inertial measurement unit is used to obtain motion data of the gimbal load; the motion data includes the pose information of the gimbal load;
    当所述处理器执行所述可执行指令时,被配置为:When the processor executes the executable instructions, it is configured to:
    根据所述运动数据对所述三维点云进行配准,获取三维点云之间的相对位姿;以及,对所述图像序列进行配准,获取图像之间的相对位姿;The three-dimensional point cloud is registered according to the motion data to obtain the relative pose between the three-dimensional point clouds; and the image sequence is registered to obtain the relative pose between the images;
    根据所述运动数据、所述三维点云之间的相对位姿以及所述图像之间的相对位姿,获取所述第一图像采集装置获取三维点云时的位姿和/或所述第二图像采集装置获取图像时的位姿;According to the motion data, the relative pose between the three-dimensional point clouds, and the relative pose between the images, obtain the pose and/or the first image acquisition device when acquiring the three-dimensional point cloud. 2. the pose when the image acquisition device acquires the image;
    根据所述三维点云、所述图像序列以及所述第一图像采集装置获取三维点云时的位姿和/或所述第二图像采集装置获取图像时的位姿,生成目标场景的三维模型。A 3D model of the target scene is generated according to the 3D point cloud, the sequence of images, and the pose when the first image acquisition device acquires the 3D point cloud and/or the pose when the second image acquisition device acquires the image .
    根据所述运动数据、所述三维点云之间的相对位姿以及所述图像之间的相对位姿,获取所述第一图像采集装置获取三维点云时的位姿和/或所述第二图像采集装置获取图像时的位姿According to the motion data, the relative pose between the three-dimensional point clouds, and the relative pose between the images, obtain the pose and/or the first image acquisition device when acquiring the three-dimensional point cloud. The pose when the image acquisition device acquires the image
  30. 根据权利要求29所述的云台负载,其特征在于,The PTZ load according to claim 29, wherein,
    所述第一图像采集装置具体用于实时获取目标场景的三维点云;The first image acquisition device is specifically configured to acquire the three-dimensional point cloud of the target scene in real time;
    所述第二图像采集装置具体用于实时获取所述目标场景的图像序列;以及The second image acquisition device is specifically configured to acquire the image sequence of the target scene in real time; and
    所述惯性测量单元具体用于实时获取所述云台负载的运动数据。The inertial measurement unit is specifically configured to acquire motion data of the PTZ load in real time.
  31. 根据权利要求29所述的云台负载,其特征在于,所述目标场景包括室内场景或者室外非空旷场景,和/或,卫星定位信号低于预定强度的场景。The PTZ load according to claim 29, wherein the target scene includes an indoor scene or an outdoor non-open scene, and/or a scene where the satellite positioning signal is lower than a predetermined strength.
  32. 根据权利要求29所述的云台负载,其特征在于,所述第一图像采集装置包括以下至少一种或多种:激光雷达、双目视觉传感器以及结构光第一图像采集装置;The pan-tilt load according to claim 29, wherein the first image acquisition device comprises at least one or more of the following: a lidar, a binocular vision sensor, and a structured light first image acquisition device;
    所述第二图像采集装置包括以下至少一种或多种:可见光相机、灰度相机以及红外相机。The second image acquisition device includes at least one or more of the following: a visible light camera, a grayscale camera, and an infrared camera.
  33. 根据权利要求29所述的云台负载,其特征在于,所述处理器具体用于:根据所述三维点云、所述图像序列以及所述第一图像采集装置获取三维点云时的位姿和/或所述第二图像采集装置获取图像时的位姿,实时生成目标场景的三维模型。The PTZ payload according to claim 29, wherein the processor is specifically configured to: acquire the pose of the 3D point cloud according to the 3D point cloud, the image sequence and the first image acquisition device when acquiring the 3D point cloud And/or the pose when the second image acquisition device acquires the image, and generates a three-dimensional model of the target scene in real time.
  34. 根据权利要求29所述的云台负载,其特征在于,在获取三维点云之间的相对位姿时,所述处理器还用于:对于在连续时间序列上采集到的一组三维点云,根据该组三维点云中每个点对应的运动数据获取点之间的变换关系;The PTZ load according to claim 29, wherein when acquiring the relative pose between the three-dimensional point clouds, the processor is further configured to: for a group of three-dimensional point clouds collected in a continuous time series , obtain the transformation relationship between points according to the motion data corresponding to each point in the set of three-dimensional point clouds;
    根据所述点之间的变换关系将该组三维点云中的点重投影至同一坐标系。The points in the set of three-dimensional point clouds are reprojected to the same coordinate system according to the transformation relationship between the points.
  35. 根据权利要求34所述的云台负载,其特征在于,所述处理器还用于:根据该组三维点云中每个点对应的运动数据确定所述惯性测量单元在预设时间间隔内的相对位姿;所述预设时间间隔为获取相邻点的时间间隔;The pan-tilt load according to claim 34, wherein the processor is further configured to: determine the motion data of the inertial measurement unit within a preset time interval according to the motion data corresponding to each point in the set of three-dimensional point clouds relative pose; the preset time interval is the time interval for obtaining adjacent points;
    根据所述惯性测量单元在预设时间间隔内的相对位姿、以及所述第一图像采集装置与所述惯性测量单元之间的外参转换关系,获取所述点之间的变换关系。The transformation relationship between the points is acquired according to the relative pose of the inertial measurement unit within a preset time interval and the external parameter transformation relationship between the first image acquisition device and the inertial measurement unit.
  36. 根据权利要求34所述的云台负载,其特征在于,所述每个点对应的运动数据根据所述点的获取时间所确定。The PTZ load according to claim 34, wherein the motion data corresponding to each point is determined according to the acquisition time of the point.
  37. 根据权利要求34所述的云台负载,其特征在于,在获取三维点云之间的相对位姿时,所述处理器还用于:The pan-tilt load according to claim 34, wherein when acquiring the relative pose between the three-dimensional point clouds, the processor is further configured to:
    对于相邻两组三维点云,获取每组三维点云的面特征和/或拐角特征;其中,相邻两组三维点云中均包括有针对同一物体采集到的点;For two adjacent groups of 3D point clouds, obtain the surface features and/or corner features of each group of 3D point clouds; wherein, the adjacent two groups of 3D point clouds both include points collected for the same object;
    根据相邻两组三维点云的面特征和/或拐角特征,获取相邻两组三维点云之间的相对位姿;所述相邻两组三维点云之间的相对位姿用于将相邻两组三维点云重投影至同一坐标系。According to the surface features and/or corner features of the adjacent two groups of three-dimensional point clouds, the relative poses between the two adjacent groups of three-dimensional point clouds are obtained; the relative poses between the two adjacent groups of three-dimensional point clouds are used to The adjacent two sets of 3D point clouds are reprojected to the same coordinate system.
  38. 根据权利要求37所述的云台负载,其特征在于,所述每组三维点云的面特征和/或拐角特征根据该组三维点云中点之间的曲度信息所确定。The pan/tilt load according to claim 37, wherein the surface features and/or corner features of each group of three-dimensional point clouds are determined according to the curvature information between the midpoints of the group of three-dimensional point clouds.
  39. 根据权利要求38所述的云台负载,其特征在于,所述每组三维点云的面特征根据曲度大于预设阈值的曲度信息所确定;The PTZ load according to claim 38, wherein the surface features of each group of three-dimensional point clouds are determined according to curvature information whose curvature is greater than a preset threshold;
    所述每组三维点云的拐角特征根据曲度小于或等于预设阈值的曲度信息所确定。The corner features of each group of three-dimensional point clouds are determined according to curvature information whose curvature is less than or equal to a preset threshold.
  40. 根据权利要求37所述的云台负载,其特征在于,所述处理器还用于:根据相 邻两组三维点云的面特征之间的相似度,和/或,相邻两组三维点云的拐角特征之间的相似度,获取三维点云之间的相对位姿。The PTZ load according to claim 37, wherein the processor is further configured to: according to the similarity between the surface features of the adjacent two groups of three-dimensional point clouds, and/or, the adjacent two groups of three-dimensional point clouds The similarity between the corner features of the cloud and the relative pose between the three-dimensional point clouds are obtained.
  41. 根据权利要求40所述的云台负载,其特征在于,所述相邻两组三维点云的面特征之间的相似度基于所述相邻两组三维点云的面特征之间的距离确定;和/或,The pan/tilt load according to claim 40, wherein the similarity between the surface features of the adjacent two groups of three-dimensional point clouds is determined based on the distance between the surface features of the two adjacent groups of three-dimensional point clouds ;and / or,
    所述相邻两组三维点云的拐角特征之间的相似度基于所述相邻两组三维点云的拐角特征之间的距离确定。The similarity between the corner features of the adjacent two groups of three-dimensional point clouds is determined based on the distance between the corner features of the two adjacent groups of three-dimensional point clouds.
  42. 根据权利要求37所述的云台负载,其特征在于,所述面特征包括指示平面的至少一个三维坐标信息以及该平面的法向量;和/或,The PTZ payload according to claim 37, wherein the surface feature includes at least one three-dimensional coordinate information indicating a plane and a normal vector of the plane; and/or,
    所述拐角特征包括指示边缘的至少一个三维坐标信息以及指示该边缘的向量。The corner feature includes at least one three-dimensional coordinate information indicating an edge and a vector indicating the edge.
  43. 根据权利要求29所述的云台负载,其特征在于,所述图像序列包括在连续的时间序列上获取的多个图像;The pan-tilt load of claim 29, wherein the image sequence comprises a plurality of images acquired in a continuous time series;
    在获取图像之间的相对位姿时,所述处理器还用于提取所述图像中的特征点,并基于所述图像的特征点进行配准,获取所述图像之间的相对位姿。When acquiring the relative poses between the images, the processor is further configured to extract feature points in the images, perform registration based on the feature points of the images, and obtain the relative poses between the images.
  44. 根据权利要求29所述的云台负载,其特征在于,所述图像序列包括在连续的时间序列上获取的多个图像;The pan-tilt load of claim 29, wherein the image sequence comprises a plurality of images acquired in a continuous time series;
    在获取图像之间的相对位姿时,所述处理器还用于:When acquiring the relative pose between the images, the processor is further configured to:
    使用所述运动数据将所述三维点云进行重投影,并根据重投影后的三维点云以及所述运动数据生成与所述图像对应的深度图;reprojecting the 3D point cloud using the motion data, and generating a depth map corresponding to the image according to the reprojected 3D point cloud and the motion data;
    提取所述图像中的特征点,并从所述图像对应的深度图中获取所述特征点的深度信息;extracting feature points in the image, and obtaining depth information of the feature points from a depth map corresponding to the image;
    根据所述图像的特征点以及所述特征点的深度信息进行配准,获取所述图像之间的相对位姿。The registration is performed according to the feature points of the images and the depth information of the feature points, and the relative poses between the images are acquired.
  45. 根据权利要求29所述的云台负载,其特征在于,在获取图像之间的相对位姿时,所述处理器还用于:在无光情况下,使用所述运动数据将所述三维点云进行重投影,并根据重投影后的三维点云以及所述运动数据生成与所述图像序列中的图像对应的深度图;根据所述深度图中提取的特征点进行配准,获取所述图像之间的相对位姿。The pan-tilt load according to claim 29, wherein when acquiring the relative pose between images, the processor is further configured to: use the motion data to convert the three-dimensional point in the absence of light Reproject the cloud, and generate a depth map corresponding to the image in the image sequence according to the reprojected 3D point cloud and the motion data; perform registration according to the feature points extracted from the depth map, and obtain the Relative pose between images.
  46. 根据权利要求29所述的云台负载,其特征在于,所述处理器还用于:The PTZ load according to claim 29, wherein the processor is further configured to:
    建立位姿图模型;Build a pose graph model;
    利用所述运动数据、所述三维点云之间的相对位姿以及所述图像之间的相对位姿优化所述位姿图模型,获取所述第一图像采集装置获取三维点云时的位姿和/或所述第二图像采集装置获取图像时的位姿。The pose graph model is optimized by using the motion data, the relative poses between the three-dimensional point clouds, and the relative poses between the images, and obtains the pose when the first image acquisition device acquires the three-dimensional point cloud. pose and/or the pose when the second image acquisition device acquires the image.
  47. 根据权利要求46所述的云台负载,其特征在于,在所述位姿图模型中,各个顶点被配置为指示所述第一图像采集装置采集三维点云时的位姿和/或所述第二图像采集装置获取图像时的位姿;The PTZ payload according to claim 46, wherein in the pose graph model, each vertex is configured to indicate the pose and/or the pose when the first image capturing device collects a three-dimensional point cloud The pose when the second image acquisition device acquires the image;
    以及,各个顶点之间的边被配置为指示所述三维点云之间的相对位姿、所述图像之间的相对位姿或者所述运动数据。And, the edges between the vertices are configured to indicate relative poses between the three-dimensional point clouds, relative poses between the images, or the motion data.
  48. 根据权利要求47所述的云台负载,其特征在于,所述各个顶点之间的边还被配置为指示定位信息;所述定位信息至少包括GPS信息和/或RTK信息。The PTZ load according to claim 47, wherein the edges between the vertices are further configured to indicate positioning information; the positioning information at least includes GPS information and/or RTK information.
  49. 根据权利要求46所述的云台负载,其特征在于,所述位姿图模型包括滑动窗口位姿图模型。The gimbal payload according to claim 46, wherein the pose graph model comprises a sliding window pose graph model.
  50. 根据权利要求46所述的云台负载,其特征在于,所述处理器还用于:为所述三维模型添加标签索引后,将所述三维模型存储至所述标签索引指向的存储位置;和/或,将所述三维模型实时传输给遥控终端,以在遥控终端上实时显示所述三维模型。The PTZ load according to claim 46, wherein the processor is further configured to: after adding a label index to the 3D model, store the 3D model to a storage location pointed to by the label index; and /or, transmitting the three-dimensional model to the remote control terminal in real time, so as to display the three-dimensional model on the remote control terminal in real time.
  51. 根据权利要求29所述的云台负载,其特征在于,所述处理器还用于:The PTZ load according to claim 29, wherein the processor is further configured to:
    根据所述图像序列中的多张图像之间的相似度进行回环检测;或者,根据所述三维点云重投影得到的多张深度图之间的相似度进行回环检测;或者,根据降采样后的全局模型与所述三维模型进行回环检测;Loop closure detection is performed according to the similarity between multiple images in the image sequence; or, loop closure detection is performed according to the similarity between multiple depth maps obtained by re-projection of the 3D point cloud; The global model and the three-dimensional model for loop closure detection;
    利用回环检测的结果对所述位姿图模型进行全局优化;以及,根据全局优化后的位姿图模型更新所述三维模型。The pose graph model is globally optimized using the result of loop closure detection; and the three-dimensional model is updated according to the globally optimized pose graph model.
  52. 根据权利要求51所述的云台负载,其特征在于,所述处理器还用于:利用形成回环的图像之间的相对位姿或者形成回环的三维点之间的相对位姿,对所述位姿图模型进行全局优化。The pan-tilt load according to claim 51, wherein the processor is further configured to: use the relative pose between the images forming the loop or the relative pose between the three-dimensional points forming the loop to The pose graph model is globally optimized.
  53. 根据权利要求51所述的云台负载,其特征在于,使用词袋模型来度量所述图像序列中的多张图像之间的相似度或者所述多张深度图像之间的相似度。The PTZ payload according to claim 51, wherein a bag-of-words model is used to measure the similarity between multiple images in the image sequence or the similarity between the multiple depth images.
  54. 根据权利要求29所述的云台负载,其特征在于,所述处理器还用于:The PTZ load according to claim 29, wherein the processor is further configured to:
    根据所述三维模型的精度要求和/或密度要求,生成采集轨迹和采集时间;generating an acquisition trajectory and acquisition time according to the accuracy requirements and/or density requirements of the three-dimensional model;
    根据所述采集轨迹和采集时间控制所述第一图像采集装置、所述第二图像采集装置以及所述惯性测量单元采集数据。The first image acquisition device, the second image acquisition device and the inertial measurement unit are controlled to acquire data according to the acquisition trajectory and acquisition time.
  55. 根据权利要求50所述的云台负载,其特征在于,所述处理器还用于:The PTZ load according to claim 50, wherein the processor is further configured to:
    接收所述遥控终端传输的反馈信息;所述反馈信息用于指示所述第一图像采集装置漏采的位置,所述反馈信息由用户针对于显示在所述遥控终端上的三维模型输入;Receive feedback information transmitted by the remote control terminal; the feedback information is used to indicate the position of the missed acquisition by the first image acquisition device, and the feedback information is input by the user with respect to the three-dimensional model displayed on the remote control terminal;
    控制所述第一图像采集装置采集所述目标场景中所述反馈信息指示的位置处的三 维点云。The first image acquisition device is controlled to acquire a three-dimensional point cloud at the position indicated by the feedback information in the target scene.
  56. 一种可移动平台,其特征在于,包括如权利要求29至55任意一项所述的云台负载。A movable platform, characterized by comprising the PTZ load according to any one of claims 29 to 55.
  57. 根据权利要求56所述的可移动平台,其特征在于,所述可移动平台至少包括无人飞行器、无人驾驶车辆或者移动机器人。The movable platform of claim 56, wherein the movable platform comprises at least an unmanned aerial vehicle, an unmanned vehicle or a mobile robot.
  58. 一种计算机可读存储介质,其特征在于,所述计算机可读存储介质存储有计算机可执行指令,所述计算机可执行指令被处理器执行时实现如权利要求1至28任一项所述的方法。A computer-readable storage medium, characterized in that the computer-readable storage medium stores computer-executable instructions, and when the computer-executable instructions are executed by a processor, implements the method according to any one of claims 1 to 28 method.
PCT/CN2020/120978 2020-10-14 2020-10-14 Three-dimensional reconstruction method, gimbal load, removable platform and computer-readable storage medium WO2022077296A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/CN2020/120978 WO2022077296A1 (en) 2020-10-14 2020-10-14 Three-dimensional reconstruction method, gimbal load, removable platform and computer-readable storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2020/120978 WO2022077296A1 (en) 2020-10-14 2020-10-14 Three-dimensional reconstruction method, gimbal load, removable platform and computer-readable storage medium

Publications (1)

Publication Number Publication Date
WO2022077296A1 true WO2022077296A1 (en) 2022-04-21

Family

ID=81207575

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/120978 WO2022077296A1 (en) 2020-10-14 2020-10-14 Three-dimensional reconstruction method, gimbal load, removable platform and computer-readable storage medium

Country Status (1)

Country Link
WO (1) WO2022077296A1 (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114529652A (en) * 2022-04-24 2022-05-24 深圳思谋信息科技有限公司 Point cloud compensation method, device, equipment, storage medium and computer program product
CN116844074A (en) * 2023-07-25 2023-10-03 北京爱科农科技有限公司 Panoramic display linkage method for three-dimensional scene and key area of orchard
CN117168441A (en) * 2023-11-02 2023-12-05 西安因诺航空科技有限公司 Multi-sensor fusion SLAM positioning and reconstructing method and system
WO2023237065A1 (en) * 2022-06-08 2023-12-14 先临三维科技股份有限公司 Loop closure detection method and apparatus, and electronic device and medium

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104236548A (en) * 2014-09-12 2014-12-24 清华大学 Indoor autonomous navigation method for micro unmanned aerial vehicle
CN106780729A (en) * 2016-11-10 2017-05-31 中国人民解放军理工大学 A kind of unmanned plane sequential images batch processing three-dimensional rebuilding method
CN110060332A (en) * 2019-04-09 2019-07-26 上海科技大学 High-precision three-dimensional based on airborne acquisition equipment builds figure and modeling
US20190325089A1 (en) * 2018-04-18 2019-10-24 Reconstruct Inc. Computation of point clouds and joint display of point clouds and building information models with project schedules for monitoring construction progress, productivity, and risk for delays

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104236548A (en) * 2014-09-12 2014-12-24 清华大学 Indoor autonomous navigation method for micro unmanned aerial vehicle
CN106780729A (en) * 2016-11-10 2017-05-31 中国人民解放军理工大学 A kind of unmanned plane sequential images batch processing three-dimensional rebuilding method
US20190325089A1 (en) * 2018-04-18 2019-10-24 Reconstruct Inc. Computation of point clouds and joint display of point clouds and building information models with project schedules for monitoring construction progress, productivity, and risk for delays
CN110060332A (en) * 2019-04-09 2019-07-26 上海科技大学 High-precision three-dimensional based on airborne acquisition equipment builds figure and modeling

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114529652A (en) * 2022-04-24 2022-05-24 深圳思谋信息科技有限公司 Point cloud compensation method, device, equipment, storage medium and computer program product
CN114529652B (en) * 2022-04-24 2022-07-19 深圳思谋信息科技有限公司 Point cloud compensation method, device, equipment and storage medium
WO2023237065A1 (en) * 2022-06-08 2023-12-14 先临三维科技股份有限公司 Loop closure detection method and apparatus, and electronic device and medium
CN116844074A (en) * 2023-07-25 2023-10-03 北京爱科农科技有限公司 Panoramic display linkage method for three-dimensional scene and key area of orchard
CN117168441A (en) * 2023-11-02 2023-12-05 西安因诺航空科技有限公司 Multi-sensor fusion SLAM positioning and reconstructing method and system
CN117168441B (en) * 2023-11-02 2024-02-20 西安因诺航空科技有限公司 Multi-sensor fusion SLAM positioning and reconstructing method and system

Similar Documents

Publication Publication Date Title
WO2022077296A1 (en) Three-dimensional reconstruction method, gimbal load, removable platform and computer-readable storage medium
US11263761B2 (en) Systems and methods for visual target tracking
CN112567201B (en) Distance measuring method and device
CN111156998B (en) Mobile robot positioning method based on RGB-D camera and IMU information fusion
CN109887057B (en) Method and device for generating high-precision map
US20210141378A1 (en) Imaging method and device, and unmanned aerial vehicle
KR102191445B1 (en) Need-sensitive image and location capture system and method
WO2022036980A1 (en) Pose determination method and apparatus, electronic device, storage medium, and program
US11906983B2 (en) System and method for tracking targets
CN108810473B (en) Method and system for realizing GPS mapping camera picture coordinate on mobile platform
CN111936821A (en) System and method for positioning
JP2015535980A (en) Image processing method used for vision-based positioning, particularly for apparatus
WO2018193574A1 (en) Flight path generation method, information processing device, flight path generation system, program and recording medium
CN113587934A (en) Robot, indoor positioning method and device and readable storage medium
US11460302B2 (en) Terrestrial observation device having location determination functionality
US20210229810A1 (en) Information processing device, flight control method, and flight control system
WO2020019175A1 (en) Image processing method and apparatus, and photographing device and unmanned aerial vehicle
CA3069813A1 (en) Capturing, connecting and using building interior data from mobile devices
US20210233271A1 (en) Information processing apparatus, server, movable object device, information processing method, and program
KR102402949B1 (en) Acquisition method of image information with improved precision
CN116027351A (en) Hand-held/knapsack type SLAM device and positioning method
US20210256732A1 (en) Image processing method and unmanned aerial vehicle
CN111581322B (en) Method, device and equipment for displaying region of interest in video in map window
TWI738315B (en) Automatic tracking photographic system based on light label
Zhao et al. 2D monocular visual odometry using mobile-phone sensors

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20957076

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20957076

Country of ref document: EP

Kind code of ref document: A1