WO2020248851A1 - 库位的检测方法及装置、存储介质和电子装置 - Google Patents
库位的检测方法及装置、存储介质和电子装置 Download PDFInfo
- Publication number
- WO2020248851A1 WO2020248851A1 PCT/CN2020/093657 CN2020093657W WO2020248851A1 WO 2020248851 A1 WO2020248851 A1 WO 2020248851A1 CN 2020093657 W CN2020093657 W CN 2020093657W WO 2020248851 A1 WO2020248851 A1 WO 2020248851A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- dimensional
- preset
- storage location
- obstacle
- voxel
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/246—Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T17/00—Three dimensional [3D] modelling, e.g. data description of 3D objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10004—Still image; Photographic image
- G06T2207/10012—Stereo images
Definitions
- This application relates to the field of data processing, and in particular to a method and device for detecting storage locations, a storage medium and an electronic device.
- the automatic parking system is a system that finds storage spaces and controls vehicles to automatically plan the route to park in.
- This system can greatly improve the automation level of the current auxiliary parking system, and solve the "last mile" problem of parking.
- User travel comfort provides technical guarantee, so it has become the current research hotspot of unmanned driving.
- the system needs to detect accurate storage location and obstacle perception to provide prerequisites for driving decision planning.
- accurate three-dimensional location detection will greatly improve the automatic parking system Security.
- the location detection in the prior art is often limited to two-dimensional location detection due to the selection or method of sensors, and it is easy to have a limited detection range and low detection accuracy due to sensor distance, angular resolution, field of view, etc. .
- the embodiments of the present application provide a method and device for detecting a warehouse location, a storage medium, and an electronic device.
- a method for detecting storage locations includes: establishing three-dimensional voxel probabilities based on the relative positions of ToF cameras flying at multiple times from different directions and the RGB images of the multiple ToF cameras Map; Determine the semantics of the preset obstacle according to the three-dimensional voxel probability map and the vertical height distribution characteristics of the preset obstacle; According to the preset storage location model and the determined semantic extraction of the preset obstacle, stop can be The two-dimensional storage location of the storage location is combined with the three-dimensional voxel probability map to determine the three-dimensional storage location corresponding to the two-dimensional storage location.
- a storage location detection device including: a establishing module for flying relative positions between ToF cameras at multiple times from different directions and RGB images of the multiple ToF cameras A three-dimensional voxel probability map is established; a determining module is used to determine the semantics of the preset obstacle according to the three-dimensional voxel probability map and the vertical height distribution characteristics of the preset obstacle; a processing module is used to determine the semantics of the preset obstacle according to the preset storage location The model and the determined semantics of the preset obstacle extract a two-dimensional location of a dockable location, and combine the three-dimensional voxel probability map to determine a three-dimensional location corresponding to the two-dimensional location.
- a storage medium in which a computer program is stored, wherein the computer program is configured to execute the steps in any one of the foregoing method embodiments when running.
- an electronic device including a memory and a processor, the memory is stored with a computer program, and the processor is configured to run the computer program to execute any of the above Steps in the method embodiment.
- Fig. 1 is a flowchart of a method for detecting a warehouse location according to an embodiment of the present application
- Fig. 2 is a schematic diagram of the arrangement of multiple ToF devices according to an embodiment of the present application.
- Fig. 3 is a schematic diagram of a single-frame splicing point cloud according to an embodiment of the present application
- FIG. 4 is a schematic diagram of a three-dimensional voxel probability fusion map according to an embodiment of the present application.
- Fig. 5 is a flow chart of a warehouse location detection according to an embodiment of the present application.
- Fig. 6 is a schematic diagram of a detection direction according to an embodiment of the present application.
- Fig. 7 is a schematic diagram of a two-dimensional location detection result according to an embodiment of the present application.
- Fig. 8 is a schematic diagram of a three-dimensional storage location detection result according to an embodiment of the present application.
- Fig. 9 is a structural block diagram of a storage location detection device according to an embodiment of the present application.
- FIG. 1 is a flowchart of a storage location detection method according to an embodiment of the present application. As shown in FIG. 1, the process includes the following steps:
- Step S102 establishing a three-dimensional voxel probability map based on the relative positions of ToF cameras flying at multiple times from different directions and the RGB images of the multiple ToF cameras;
- Step S104 Determine the semantics of the preset obstacle according to the three-dimensional voxel probability map and the vertical height distribution characteristics of the preset obstacle;
- Step S106 Extract the two-dimensional storage location of the parkable storage location according to the preset storage location model and the determined semantics of the preset obstacle, and determine the three-dimensional storage location corresponding to the two-dimensional storage location in combination with the three-dimensional voxel probability map.
- a three-dimensional voxel probability map is established based on the relative positions of multiple time-flight ToF cameras from different directions and the RGB images of the multiple ToF cameras; and then based on the three-dimensional voxel probability map and presets
- the vertical height distribution feature of the obstacle determines the semantics of the preset obstacle; finally, the two-dimensional storage location of the dockable storage location is extracted according to the preset location model and the determined semantics of the preset obstacle, and combined with the three-dimensional voxel probability map
- the three-dimensional location corresponding to the two-dimensional location is determined, thereby solving the problem that the location can only be detected by the two-dimensional location detection method, and improving the efficiency of automatic parking.
- the three-dimensional voxel is established based on the relative positions of multiple time-flight ToF cameras from different directions and the RGB images of the multiple ToF cameras.
- the probability map method can be implemented in the following ways:
- Step S102-11 obtaining a wide-angle point cloud formed by stitching point clouds within a preset range from the relative position calibration results between multiple time-flying ToF cameras in different directions;
- Step S102-12 Determine the camera self-motion between adjacent frames in the RGB image based on the feature point method to determine the motion change of the three-dimensional volume between the previous and next frames;
- Step S102-13 Establish the three-dimensional voxel probability map based on the local point cloud and the global position in the wide-angle point cloud, and update the voxel probability value in the three-dimensional voxel probability map.
- the method of determining the self-motion of the camera between adjacent frames in the RGB image based on the feature point method involved in the above step S102-12 to determine the motion change of the three-dimensional body between the previous and subsequent frames can be implemented in the following manner:
- Step S1 Acquire RGB images of adjacent frames through multiple ToF cameras, and extract feature points in the RGB images of adjacent frames based on the scale-invariant feature transformation SIFT algorithm;
- Step S2 matching the extracted feature points according to the proximity algorithm
- Step S3 Calculate changes in the positions of adjacent cameras according to the matching results of the feature points to obtain a change matrix, and use the change matrix as a visual odometer.
- step S1 to step S3 in a specific application scenario for RGB images of adjacent frames, the SIFT feature points of the images are extracted respectively and the nearest neighbor algorithm is used for matching. Since the feature point matching result is the basis for calculating the camera position change between adjacent frames, in the matching result, matches that are greater than N times the minimum feature distance are deleted to ensure the reliability of the feature point matching result.
- the change matrix T is used to express the position transformation of the camera, which is composed of a rotation matrix R and a translation vector t.
- the rotation matrix R is calculated and determined by the rotation angles ⁇ , ⁇ , ⁇ of the x, y, and z axes, and the translation vector t is composed of the position difference ( ⁇ x, ⁇ y, ⁇ z) of the origin of the two coordinate systems.
- the reprojection error is minimized to calculate the camera external parameter change matrix. Since the position transformation is calculated using a three-dimensional point projection method, the scale of the translation vector can be guaranteed to be the same. Finally obtain the optimal parameter estimate versus Form the transformation matrix T as the result of the visual odometer.
- the method of updating the voxel probability value in the three-dimensional voxel probability map involved in step S102-13 of this embodiment may be: after the voxel has been assigned In this case, the voxel probability value is subsequently updated according to the new point cloud results of each frame, where the update method is:
- L_new is the updated grid probability value
- L_old is the grid probability value before update
- p(n) is the probability value divided by the point cloud of the current frame
- l_min and l_max are the probability limit thresholds.
- Step S104-11 setting three height intervals for detection, where the three height intervals include at least one of the following: ground detection interval, low-level obstacle detection interval, and high-level obstacle detection interval;
- Step S104-12 performing through filtering on the three height intervals to obtain a set of voxels in a range corresponding to the three height intervals;
- Step S104-13 Determine the distribution interval and geometric characteristics of the preset obstacle according to the voxel sets corresponding to the different height intervals, and perform semantic judgment on the obstacle to determine the semantics of the preset obstacle.
- each voxel records the probability of being an obstacle, so the probability threshold ⁇ obstacle is taken as the basis for obstacle classification, and the obstacle perception in the three-dimensional map is extracted result.
- the probability threshold ⁇ obstacle is taken as the basis for obstacle classification, and the obstacle perception in the three-dimensional map is extracted result.
- three height intervals (unit: meters) are set:
- the voxels in the detection interval of high and low obstacles are projected to a two-dimensional horizontal plane.
- There are major obstacles in the garage environment such as pillars, vehicles, walls, etc., but the distribution intervals and geometric characteristics of obstacles with different semantics are obviously different.
- the walls appear on both high and low floors and have obvious linear characteristics.
- the lower floors all appear and are rectangular, and the vehicles only appear in a rectangular shape on the lower floors. Therefore, the common obstacles in the scene can be quickly semantically judged according to the differences in distribution and geometric characteristics, and the two-dimensional distribution of the corresponding semantic obstacles can be obtained.
- Combining the three-dimensional voxel probability map to determine the three-dimensional location corresponding to the two-dimensional location can be achieved in the following ways:
- Step S106-11 searching for a two-dimensional storage location that represents an empty storage location according to the distribution result of the two-dimensional preset obstacles on the horizontal plane in the preset storage location model;
- Step S106-12 based on the planar position of the two-dimensional storage location, use the preset length as the step length to grow toward the top surface of the garage to form a three-dimensional bounding box;
- Step S106-13 When the number of voxels of the obstacle contained in the three-dimensional bounding box exceeds a preset threshold, it is determined that the obstacle is touched and recorded as the maximum vertical height value of the storage location;
- Step S106-14 taking the height value in the ground segmentation result from the maximum height value as the ground height value, and taking the ground height value as the height limit value of the three-dimensional storage location;
- step S106-15 the two-dimensional storage location and the height limit value are formed into a three-dimensional storage location.
- This optional embodiment provides a new solution based on multiple ToF (time of flight) cameras instead of conventional in-vehicle ultrasonic radar and surround view cameras, which can provide automatic parking systems with real-time three-dimensional environment perception in a local area. Location detection to ensure driving safety during parking.
- ToF time of flight
- the invention proposes to deploy multiple ToF cameras with different orientations to jointly observe the scene, and pre-calibrate the results of the relative positions between the devices to combine the multiple ToF cameras
- the point clouds are spliced into a wide-angle point cloud, and then the self-motion of the camera is estimated using the visual odometer, real-time 3D voxel probability mapping of the local scene of the indoor parking lot is performed, and multi-frame point clouds are merged through the update of the probability.
- the semantic results such as walls, pillars, and vehicles are allocated to the obstacles, combined with the assumptions of the location model and the three-dimensional
- the top surface and ground structure extracted from the map information provide a real-time detection method for three-dimensional storage locations, and assist automatic parking by providing localized three-dimensional environmental perception and storage location detection information.
- Step S202 arranging multiple ToF cameras with different orientations, and stitching a small range point cloud into a wide-angle point cloud based on the relative position calibration result between the cameras;
- Step S204 using the RGB image of the ToF camera, use the feature point method to estimate the 6-DOF self-motion of the camera device, and calculate the three-dimensional rigid body motion change between the previous and next frames;
- Step S206 3D voxel probability mapping is performed from the local point cloud and the global position, and multi-frame point cloud data is merged by updating the voxel probability value to eliminate dynamic obstacles and reduce the influence of noise points;
- Step S208 by setting different height intervals, the interval through filtering voxels are obtained, and the vertical height distribution characteristics of common obstacles and structures in the parking lot are used to quickly distinguish the obstacle semantics and extract the ground structure;
- Step S210 based on the location model assumptions and the location standard parameters, extract the two-dimensional location of the dockable vacancy, and combine the three-dimensional grid map to search for a limited height in the vertical direction of the corresponding location to detect the three-dimensional location;
- step S212 a suitable parking location is selected in combination with vehicle parameters to obtain the target location and the passable range perception result.
- the ToF camera is limited by the horizontal field of view, and it is difficult to perceive the wide-angle range around the vehicle body.
- multiple ToF cameras with different orientations are arranged to jointly observe the scene.
- the point clouds of multiple ToF cameras are stitched into a wide-angle point cloud.
- Fig. 2 is a schematic diagram of the arrangement of multiple ToF devices according to the embodiment of the present application. The way the camera is fixed on the platform is shown in Fig. 2. Realize the joint observation of surround view, and there is a certain common field of view between adjacent cameras to ensure the continuity of stitching. When connecting multiple Kinect-V2 cameras, ensure that the bandwidth of the USB3.0 interface and the control bus is sufficient.
- Kinect-V2 contains color and infrared cameras to obtain RGB and depth images respectively (comprising RGB-D data).
- the original depth data contains a lot of noise.
- the bilateral filtering method is used to filter noise points while maintaining clearer clarity.
- the boundary information Due to different camera positions, image registration is first performed according to camera parameters. Then the point cloud recovered from the depth map and camera parameters is converted to the assumed coordinate system according to the pre-calibrated inter-camera change parameters, and stitched into a wide-angle point cloud.
- the stitched point cloud is shown in Figure 3.
- 3 is a schematic diagram of a single-frame splicing point cloud according to an embodiment of the present application.
- step S204 in order to perceive a wider range, combining the characteristics of the ToF camera data, using the RGB images taken at different times, using the feature point method to estimate the camera self-motion between adjacent frames, and build a visual odometer.
- SIFT Scale-Invariant Feature Transform
- feature point matching For RGB images of adjacent frames, the SIFT feature points of the image are extracted and the nearest neighbor algorithm is used for matching. Since the feature point matching result is the basis for calculating the camera position change between adjacent frames, in the matching result, matches that are greater than N times the minimum feature distance are deleted to ensure the reliability of the feature point matching result.
- calculate the transformation matrix use the transformation matrix T to express the position transformation of the camera, which is composed of the rotation matrix R and the translation vector t.
- the rotation matrix R is calculated and determined by the rotation angles ⁇ , ⁇ , ⁇ of the x, y, and z axes, and the translation vector t is composed of the position difference ( ⁇ x, ⁇ y, ⁇ z) of the origin of the two coordinate systems.
- the reprojection error is minimized to calculate the camera external parameter change matrix. Since the position transformation is calculated using a three-dimensional point projection method, the scale of the translation vector can be guaranteed to be the same. Finally obtain the optimal parameter estimate versus Form the transformation matrix T as the result of the visual odometer.
- the three-dimensional voxel map uses probability to describe the possibility of the voxel state as an obstacle point. Within the limited range of probability, the higher the probability value, the greater the probability of the voxel as an obstacle.
- a hierarchical data structure is realized. The smallest voxel boundary size represents the resolution of the map, and the voxel indicates whether the area is occupied and Record the probability value of the obstacle.
- the inserted point cloud after the position transformation is classified into the nearest intersecting voxel, and the ray is constructed from the current position center and each point cloud, and the ray terminal corresponds to In the state of hitting the obstacle, the voxel passing through the middle of the ray corresponds to the state of not hitting.
- the voxels in the scanning range are divided into two sets of hit and miss. Initialize the unobserved voxels, and assign the probability values to p hit and p miss .
- the update method is: if the voxel has been assigned, then the voxel probability value is updated according to the new point cloud results of each frame.
- the update method is:
- L new is the mesh probability value after update
- L old is the mesh probability value before update
- p(n) is the probability value divided by the point cloud of the current frame
- l min and l max are the probability limit thresholds, which guarantees Voxel status change update times.
- the update strategy guarantees from the mathematical model that the voxel map can increase the probability value of the target fixed obstacle, thereby eliminating the ghosting of the dynamic obstacle, and eliminating the influence of spatial noise points on the mapping.
- FIG. 4 is a schematic diagram of the 3D voxel probability fusion map according to an embodiment of the present application.
- each voxel records the probability of being an obstacle, so the probability threshold ⁇ obstacle is taken as the basis for obstacle classification, and the result of obstacle perception in the three-dimensional map is extracted.
- the probability threshold ⁇ obstacle is taken as the basis for obstacle classification, and the result of obstacle perception in the three-dimensional map is extracted.
- three height intervals (unit: meters) are set:
- the voxels in the detection interval of high and low obstacles are projected to a two-dimensional horizontal plane.
- There are major obstacles in the garage environment such as pillars, vehicles, walls, etc., but the distribution intervals and geometric characteristics of obstacles with different semantics are obviously different.
- the walls appear on both high and low floors and have obvious linear characteristics.
- the lower floors all appear and are rectangular, and the vehicles only appear in a rectangular shape on the lower floors. Therefore, the common obstacles in the scene can be quickly semantically judged according to the differences in distribution and geometric characteristics, and the two-dimensional distribution of the corresponding semantic obstacles can be obtained.
- Fig. 5 is a flow chart of the location detection according to an embodiment of the present application, searching for a vacant location area that characterizes the location according to the horizontal two-dimensional obstacle distribution result.
- the design of parking spaces is often vertical or parallel to the wall.
- the direction of the parking lot road is usually vertical or parallel to the storage space, and the columns are often distributed along the road. Therefore, the two-dimensional location detection method is based on the following assumptions:
- FIG. 6 is a schematic diagram of the detection direction according to an embodiment of the present application.
- the two-dimensional inspection image is rotated with the center of the vehicle as the origin, and the longitudinal axis of the image is parallel to the direction line, and then the smallest envelope rectangle of various obstacles is projected onto the detection direction, and the overlapping area rectangles are merged.
- the rectangular space between adjacent obstacles is greater than the minimum standard for parking spaces, for example, the parallel type is 6.0 meters, the vertical type is 2.4 meters, and the empty rectangular area is used as the parking area.
- FIG. 7, is a schematic diagram of the detection result of the two-dimensional storage location according to an embodiment of the present application.
- the sensor data is fully utilized, and the data is merged by establishing a local probability grid map, which expands the sensing range and reduces the impact of various noises; realizes real-time environmental perception And the three-dimensional location detection results provide a basis for driving decision-making for the automatic parking system.
- the method according to the above embodiment can be implemented by means of software plus the necessary general hardware platform, of course, it can also be implemented by hardware, but in many cases the former is Better implementation.
- the technical solution of this application essentially or the part that contributes to the existing technology can be embodied in the form of a software product, and the computer software product is stored in a storage medium (such as ROM/RAM, magnetic disk, The optical disc) includes several instructions to enable a terminal device (which can be a mobile phone, a computer, a server, or a network device, etc.) to execute the method described in each embodiment of the present application.
- a storage location detection device is also provided, and the device is used to implement the above-mentioned embodiments and preferred implementations, and those that have been explained will not be repeated.
- the term "module” can implement a combination of software and/or hardware with predetermined functions.
- the devices described in the following embodiments are preferably implemented by software, hardware or a combination of software and hardware is also possible and conceived.
- FIG. 9 is a structural block diagram of a storage location detection device according to an embodiment of the present application.
- the device includes: a establishing module 902 for flying relative positions between ToF cameras at multiple times from different directions and The RGB images of the multiple ToF cameras establish a three-dimensional voxel probability map; the determining module 904 is coupled to the establishing module 902, and is used to determine the preset obstacle according to the three-dimensional voxel probability map and the vertical height distribution characteristics of the preset obstacle.
- the processing module 906, coupled with the determination module 904 is used to extract the two-dimensional location of the dockable location according to the preset location model and the determined semantics of the preset obstacle, and combine the three-dimensional voxel probability map Determine the three-dimensional location corresponding to the two-dimensional location.
- the establishing module may include: an acquiring unit, configured to acquire a wide-angle point cloud formed by stitching point clouds within a preset range from the relative position calibration results between multiple time-flight ToF cameras with different orientations; and a determining unit , For determining the self-motion of the camera between adjacent frames in the RGB image based on the feature point method to determine the motion change of the three-dimensional body between the previous and subsequent frames; a processing unit for determining the local point cloud and the global position in the wide-angle point cloud The three-dimensional voxel probability map is established, and the voxel probability values in the three-dimensional voxel probability map are updated.
- the determining unit may include: a first processing unit, configured to obtain RGB images of adjacent frames through a plurality of ToF cameras, and respectively extract features in the RGB images of adjacent frames based on the scale-invariant feature transformation SIFT algorithm Point; matching unit, used to match the extracted feature points according to the proximity algorithm; the second processing unit, used to calculate the changes in the positions of adjacent cameras according to the matching results of the feature points to obtain the change matrix, and use the change matrix as the vision Odometer.
- a first processing unit configured to obtain RGB images of adjacent frames through a plurality of ToF cameras, and respectively extract features in the RGB images of adjacent frames based on the scale-invariant feature transformation SIFT algorithm Point
- matching unit used to match the extracted feature points according to the proximity algorithm
- the second processing unit used to calculate the changes in the positions of adjacent cameras according to the matching results of the feature points to obtain the change matrix, and use the change matrix as the vision Odometer.
- the processing unit is further configured to subsequently update the probability value of the voxel according to the new point cloud result of each frame when the voxel has been assigned; wherein the update method is:
- L new is the mesh probability value after update
- L old is the mesh probability value before update
- p(n) is the probability value divided by the point cloud of the current frame
- l min and l max are the probability limit thresholds.
- the determination module 908 includes: a setting unit for setting three height intervals for detection, wherein the three height intervals include at least one of the following: ground detection interval, bottom obstacle detection interval, and high Obstacle detection interval; the third processing unit is used to perform through filtering on the three height intervals to obtain the voxel set within the corresponding range of the three height intervals; the fourth processing unit is used to according to the voxels corresponding to different height intervals
- the collection determines the distribution interval and geometric characteristics of the preset obstacles to make a semantic judgment on the obstacles to determine the semantics of the preset obstacles.
- the processing module 910 includes: a searching unit for searching for a two-dimensional storage location that characterizes a vacant storage location according to the distribution result of a two-dimensional preset obstacle in the horizontal plane in the preset storage location model; Based on the plane position of the two-dimensional storage location, the preset length is used as the step length to grow toward the top surface of the garage to form a three-dimensional bounding box; the sixth processing unit is used for the number of voxels of obstacles contained in the three-dimensional bounding box When the preset threshold is exceeded, the obstacle is determined to be touched and recorded as the maximum height value in the vertical direction of the location; the seventh processing unit is used to take the height value in the ground segmentation result from the maximum height value as the ground The height value, and the ground height value is used as the height limit value of the three-dimensional storage location; the eighth processing unit is used to form a three-dimensional storage location with the two-dimensional storage location and the height limit value.
- each of the above modules can be implemented by software or hardware.
- it can be implemented in the following manner, but not limited to this: the above modules are all located in the same processor; or, the above modules are combined in any combination The forms are located in different processors.
- the embodiment of the present application also provides a storage medium in which a computer program is stored, wherein the computer program is configured to execute the steps in any of the foregoing method embodiments when running.
- the storage medium may be configured to store a computer program for executing the following steps:
- S2 Determine the semantics of the preset obstacle according to the three-dimensional voxel probability map and the vertical height distribution characteristics of the preset obstacle;
- S3 Extract the two-dimensional storage location of the dockable storage location according to the preset storage location model and the determined semantics of the preset obstacle, and determine the three-dimensional storage location corresponding to the two-dimensional storage location in combination with the three-dimensional voxel probability map.
- the storage medium may include, but is not limited to: U disk, Read-Only Memory (Read-Only Memory, ROM for short), Random Access Memory (RAM for short), mobile hard disk, magnetic disk Or various media that can store computer programs, such as an optical disk.
- the embodiment of the present application also provides an electronic device, including a memory and a processor, the memory is stored with a computer program, and the processor is configured to run the computer program to execute the steps in any of the foregoing method embodiments.
- the electronic device may further include a transmission device and an input-output device, wherein the transmission device is connected to the aforementioned processor, and the input-output device is connected to the aforementioned processor.
- the processor may be configured to execute the following steps through a computer program:
- S2 Determine the semantics of the preset obstacle according to the three-dimensional voxel probability map and the vertical height distribution characteristics of the preset obstacle;
- S3 Extract the two-dimensional storage location of the dockable storage location according to the preset storage location model and the determined semantics of the preset obstacle, and determine the three-dimensional storage location corresponding to the two-dimensional storage location in combination with the three-dimensional voxel probability map.
- modules or steps of this application can be implemented by a general computing device, and they can be concentrated on a single computing device or distributed in a network composed of multiple computing devices.
- they can be implemented with program codes executable by the computing device, so that they can be stored in the storage device for execution by the computing device, and in some cases, can be executed in a different order than here.
Abstract
Description
Claims (14)
- 一种库位的检测方法,其中,包括:基于由不同朝向的多个时间飞行ToF相机间的相对位置以及所述多个ToF相机的RGB图像建立三维体素概率地图;根据所述三维体素概率地图和预设障碍物的垂直高度分布特征确定所述预设障碍物的语义;根据预设库位模型和确定的所述预设障碍物的语义提取可停靠的库位的二维库位,并结合所述三维体素概率地图确定与所述二维库位对应的三维库位。
- 根据权利要求1所述的方法,其中,所述基于由不同朝向的多个时间飞行ToF相机间的相对位置以及所述多个ToF相机的RGB图像建立三维体素概率地图,包括:获取由不同朝向的多个时间飞行ToF相机间的相对位置标定结果将预设范围内的点云拼接成的广角点云;基于特征点法确定所述RGB图像中相邻帧间相机自运动以确定前后帧间三维体的运动变化;基于所述广角点云中的局部点云以及全局位置建立所述三维体素概率地图,并对所述三维体素概率地图中的体素概率值进行更新。
- 根据权利要求2所述的方法,其中,所述获取所述多个ToF相机的RGB图像,并使用特征点法估计相邻帧间相机自运动以确定前后帧间三维体的运动变化,包括:通过所述多个ToF相机获取相邻帧的RGB图像,并基于尺度不变特征变换SIFT算法分别提取相邻帧RGB图像中的特征点;根据邻近算法对提取出的特征点进行匹配;根据特征点的匹配结果计算相邻相机位置的变化以得到变化矩阵,并将所述变化矩阵作为视觉里程计。
- 根据权利要求1所述的方法,其中,所述根据所述三维体素概率地图和预设障碍物的垂直高度分布特征确定所述预设障碍物的语义,包括:设置用于进行探测的三个高度区间,其中,所述三个高度区间包括以下至少之一:地面探测区间、底层障碍物探测区间、高层障碍物探测区间;对所述三个高度区间进行直通滤波获取与所述三个高度区间对应范围内的体素集合;根据与所述不同高度区间对应的体素集合确定所述预设障碍物的分布区间及几何特征对障碍物进行语义判断,以确定所述预设障碍物的语义。
- 根据权利要求1所述的方法,其中,所述根据预设库位模型和确定的所述预设障碍物的语义提取可停靠的库位的二维库位,并结合所述三维体素概率地图确定与所述二维库位对应的三维库位,包括:根据预设库位模型中水平面二维预设障碍物的分布结果查找表征空位库位的二维库位;基于所述二维库位的平面位置,以预设长度作为步长向车库顶面方向增长,组成三维包围盒;在所述三维包围盒内包含的障碍物的体素个数超过预设阈值的情况下,确定触碰到障碍物,并记录为库位竖直方向上最大高度值;从所述最大高度值中取地面分割结果中的高度值作为地面高度值,并将所述地面高度值作为三维库位的限高值;将所述二维库位与所述限高值组成所述三维库位。
- 一种库位的检测装置,其中,包括:建立模块,用于基于由不同朝向的多个时间飞行ToF相机间的相对位置以及所述多个ToF相机的RGB图像建立三维体素概率地图;确定模块,用于根据所述三维体素概率地图和预设障碍物的垂直高度分布特征确定所述预设障碍物的语义;处理模块,用于根据预设库位模型和确定的所述预设障碍物的语义提取可停靠的库位的二维库位,并结合所述三维体素概率地图确定与所述二维库位对应的三维库位。
- 根据权利要求7所述的装置,其中,所述建立模块包括:获取单元,用于获取由不同朝向的多个时间飞行ToF相机间的相对位置标定结果将预设范围内的点云拼接成的广角点云;确定单元,用于基于特征点法确定所述RGB图像中相邻帧间相机自运动以确定前后帧间三维体的运动变化;处理单元,用于基于所述广角点云中的局部点云以及全局位置建立所述三维体素概率地图,并对所述三维体素概率地图中的体素概率值进行更新。
- 根据权利要求8所述的装置,其中,所述确定单元包括:第一处理单元,用于通过所述多个ToF相机获取相邻帧的RGB图像,并基于尺度不变特征变换SIFT算法分别提取相邻帧RGB图像中的特征点;匹配单元,用于根据邻近算法对提取出的特征点进行匹配;第二处理单元,用于根据特征点的匹配结果计算相邻相机位置的变化以得到变化矩阵,并将所述变化矩阵作为视觉里程计。
- 根据权利要求7所述的装置,其中,所述确定模块包括:设置单元,用于设置用于进行探测的三个高度区间,其中,所述三个高度区间包括以下至少之一:地面探测区间、底层障碍物探测区间、高层障碍物探测区间;第三处理单元,用于对所述三个高度区间进行直通滤波获取与所述三个高度区间对应范围内的体素集合;第四处理单元,用于根据与所述不同高度区间对应的体素集合确定所述预设障碍物的分布区间及几何特征对障碍物进行语义判断,以确定所述预设障碍物的语义。
- 根据权利要求7所述的装置,其中,所述处理模块包括:查找单元,用于根据预设库位模型中水平面二维预设障碍物的分布结果查找表征空位库位的二维库位;第五处理单元,用于基于所述二维库位的平面位置,以预设长度作为步长向车库顶面方向增长,组成三维包围盒;第六处理单元,用于在所述三维包围盒内包含的障碍物的体素个数超过预设阈值的情况下,确定触碰到障碍物,并记录为库位竖直方 向上最大高度值;第七处理单元,用于从所述最大高度值中取地面分割结果中的高度值作为地面高度值,并将所述地面高度值作为三维库位的限高值;第八处理单元,用于将所述二维库位与所述限高值组成所述三维库位。
- 一种存储介质,其中,所述存储介质中存储有计算机程序,其中,所述计算机程序被设置为运行时执行所述权利要求1至6任一项中所述的方法。
- 一种电子装置,包括存储器和处理器,其中,所述存储器中存储有计算机程序,所述处理器被设置为运行所述计算机程序以执行所述权利要求1至6任一项中所述的方法。
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910502737.6 | 2019-06-11 | ||
CN201910502737.6A CN112070798A (zh) | 2019-06-11 | 2019-06-11 | 库位的检测方法及装置、存储介质和电子装置 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2020248851A1 true WO2020248851A1 (zh) | 2020-12-17 |
Family
ID=73658591
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2020/093657 WO2020248851A1 (zh) | 2019-06-11 | 2020-06-01 | 库位的检测方法及装置、存储介质和电子装置 |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN112070798A (zh) |
WO (1) | WO2020248851A1 (zh) |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140232569A1 (en) * | 2013-02-21 | 2014-08-21 | Apple Inc. | Automatic identification of vehicle location |
CN205230369U (zh) * | 2015-12-24 | 2016-05-11 | 北京万集科技股份有限公司 | 一种基于tof相机的停车位检测系统 |
CN106056643A (zh) * | 2016-04-27 | 2016-10-26 | 武汉大学 | 一种基于点云的室内动态场景slam方法及系统 |
CN107036594A (zh) * | 2017-05-07 | 2017-08-11 | 郑州大学 | 智能电站巡检智能体的定位与多粒度环境感知技术 |
CN108122412A (zh) * | 2016-11-26 | 2018-06-05 | 沈阳新松机器人自动化股份有限公司 | 用于监控机器人检测车辆乱停的方法 |
-
2019
- 2019-06-11 CN CN201910502737.6A patent/CN112070798A/zh active Pending
-
2020
- 2020-06-01 WO PCT/CN2020/093657 patent/WO2020248851A1/zh active Application Filing
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140232569A1 (en) * | 2013-02-21 | 2014-08-21 | Apple Inc. | Automatic identification of vehicle location |
CN205230369U (zh) * | 2015-12-24 | 2016-05-11 | 北京万集科技股份有限公司 | 一种基于tof相机的停车位检测系统 |
CN106056643A (zh) * | 2016-04-27 | 2016-10-26 | 武汉大学 | 一种基于点云的室内动态场景slam方法及系统 |
CN108122412A (zh) * | 2016-11-26 | 2018-06-05 | 沈阳新松机器人自动化股份有限公司 | 用于监控机器人检测车辆乱停的方法 |
CN107036594A (zh) * | 2017-05-07 | 2017-08-11 | 郑州大学 | 智能电站巡检智能体的定位与多粒度环境感知技术 |
Non-Patent Citations (1)
Title |
---|
ZHAO, JUNJIAO ET AL.: "Three-dimensional Parking Slot Detection Using Mapping and Structural Semantics Based on Multiple Time-of-flight Cameras", JOURNAL OF TONGJI UNIVERSITY (NATURAL SCIENCE), vol. 47, no. 4, 6 May 2019 (2019-05-06), pages 1 - 6, XP055764445, ISSN: 0253-374X * |
Also Published As
Publication number | Publication date |
---|---|
CN112070798A (zh) | 2020-12-11 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110163930B (zh) | 车道线生成方法、装置、设备、系统及可读存储介质 | |
CN108763287B (zh) | 大规模可通行区域驾驶地图的构建方法及其无人驾驶应用方法 | |
US9483703B2 (en) | Online coupled camera pose estimation and dense reconstruction from video | |
CN111376895B (zh) | 一种环视泊车感知方法、装置、自动泊车系统和车辆 | |
US9042639B2 (en) | Method for representing surroundings | |
US20190370989A1 (en) | Method and apparatus for 3-dimensional point cloud reconstruction | |
JP7209115B2 (ja) | 複数の相対的に接近する様に動いているリジッドなオブジェクトの検出、3d再現および追跡 | |
WO2020082777A1 (zh) | 车位检测方法及装置 | |
CN111213153A (zh) | 目标物体运动状态检测方法、设备及存储介质 | |
JP2008262333A (ja) | 路面判別装置および路面判別方法 | |
CN114454875A (zh) | 一种基于强化学习的城市道路自动泊车方法及系统 | |
CN111507158A (zh) | 利用语义分割检测停车区域的方法和装置 | |
WO2021017211A1 (zh) | 一种基于视觉的车辆定位方法、装置及车载终端 | |
CN114972758A (zh) | 一种基于点云弱监督的实例分割方法 | |
CN113763438B (zh) | 一种点云配准方法、装置、设备及存储介质 | |
WO2020248851A1 (zh) | 库位的检测方法及装置、存储介质和电子装置 | |
WO2020248118A1 (zh) | 点云处理方法、系统、设备及存储介质 | |
CN115164919B (zh) | 基于双目相机的空间可行驶区域地图构建方法及装置 | |
CN112150538A (zh) | 一种在三维地图构建过程中车辆位姿的确定方法和装置 | |
CN116508071A (zh) | 用于注释汽车雷达数据的系统和方法 | |
JP2000222563A (ja) | 障害物検出装置および障害物検出装置を搭載した移動体 | |
Habib et al. | Integration of lidar and airborne imagery for realistic visualization of 3d urban environments | |
JPWO2020244717A5 (zh) | ||
KR102540634B1 (ko) | 프로젝션 기반 컬러맵 생성 방법 및 이를 실행하기 위하여 기록매체에 기록된 컴퓨터 프로그램 | |
KR102540629B1 (ko) | 교통 시설물에 대한 학습 데이터 생성 방법 및 이를 실행하기 위하여 기록매체에 기록된 컴퓨터 프로그램 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 20822683 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20822683 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 19.05.2022) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20822683 Country of ref document: EP Kind code of ref document: A1 |