WO2022088611A1 - 障碍检测方法、装置、电子设备、存储介质及计算机程序 - Google Patents
障碍检测方法、装置、电子设备、存储介质及计算机程序 Download PDFInfo
- Publication number
- WO2022088611A1 WO2022088611A1 PCT/CN2021/085927 CN2021085927W WO2022088611A1 WO 2022088611 A1 WO2022088611 A1 WO 2022088611A1 CN 2021085927 W CN2021085927 W CN 2021085927W WO 2022088611 A1 WO2022088611 A1 WO 2022088611A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- detected
- position information
- specific object
- area
- Prior art date
Links
- 238000001514 detection method Methods 0.000 title claims abstract description 135
- 238000003860 storage Methods 0.000 title claims abstract description 13
- 238000004590 computer program Methods 0.000 title claims abstract description 7
- 230000011218 segmentation Effects 0.000 claims abstract description 39
- 238000013528 artificial neural network Methods 0.000 claims description 47
- 230000009466 transformation Effects 0.000 claims description 42
- 230000004044 response Effects 0.000 claims description 12
- 238000000034 method Methods 0.000 description 35
- 230000008569 process Effects 0.000 description 13
- 238000010408 sweeping Methods 0.000 description 11
- 238000004458 analytical method Methods 0.000 description 10
- 230000009286 beneficial effect Effects 0.000 description 9
- 238000010586 diagram Methods 0.000 description 7
- 239000011159 matrix material Substances 0.000 description 7
- 230000006870 function Effects 0.000 description 6
- 230000004927 fusion Effects 0.000 description 6
- 238000012545 processing Methods 0.000 description 5
- 241001417527 Pempheridae Species 0.000 description 4
- 230000008878 coupling Effects 0.000 description 3
- 238000010168 coupling process Methods 0.000 description 3
- 238000005859 coupling reaction Methods 0.000 description 3
- 238000005516 engineering process Methods 0.000 description 3
- 238000004364 calculation method Methods 0.000 description 2
- 238000004140 cleaning Methods 0.000 description 2
- 238000004891 communication Methods 0.000 description 2
- 238000011161 development Methods 0.000 description 2
- 230000018109 developmental process Effects 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 238000011895 specific detection Methods 0.000 description 2
- 230000001131 transforming effect Effects 0.000 description 2
- 230000008901 benefit Effects 0.000 description 1
- 230000001413 cellular effect Effects 0.000 description 1
- 238000012937 correction Methods 0.000 description 1
- 230000007613 environmental effect Effects 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 238000013507 mapping Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000007781 pre-processing Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/10—Terrestrial scenes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/56—Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/80—Geometric correction
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/194—Segmentation; Edge detection involving foreground-background segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/80—Analysis of captured images to determine intrinsic or extrinsic camera parameters, i.e. camera calibration
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/25—Determination of region of interest [ROI] or a volume of interest [VOI]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/26—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
- G06V10/267—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion by performing operations on regions, e.g. growing, shrinking or watersheds
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10004—Still image; Photographic image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V2201/00—Indexing scheme relating to image or video recognition or understanding
- G06V2201/07—Target detection
Definitions
- the embodiments of the present disclosure relate to the technical field of path planning, and in particular, to an obstacle detection method, an apparatus, an electronic device, a storage medium, and a computer program.
- Path planning in the field of robot navigation refers to: first inform the robot of the starting point and target point of walking, and then plan a reasonable path for the robot to walk according to the known map information in an indoor environment with obstacles.
- a reasonable path should at least meet the following conditions: the path is a relatively short path from the starting point to the target point; the path can avoid obstacles in the known map to a great extent.
- the embodiments of the present disclosure provide at least an obstacle detection method, an apparatus, an electronic device, a storage medium, and a computer program.
- a first aspect of the embodiments of the present disclosure provides an obstacle detection method, the obstacle detection method includes: acquiring an image to be detected; subjecting the to-be-detected image to object detection to obtain position information of the object in the to-be-detected image;
- the to-be-detected image is semantically segmented, and the location information of the ground area in the to-be-detected image is acquired; the obstacle area is acquired based on the location information of the object and the location information of the ground area.
- the obstacle area is jointly determined based on the position information of the object and the position information of the ground area, and the position information of the ground area in the image to be detected is obtained based on semantic segmentation, it is possible to improve the accuracy of the determined ground area. Therefore, the obstacle area with high accuracy can be obtained, and the obstacle analysis of the environment can be carried out effectively.
- the acquiring the image to be detected includes: acquiring calibration parameters of the camera, wherein the calibration parameters include distortion parameters; acquiring a captured target image, and correcting the target image based on the distortion parameters to obtain the to-be-detected image.
- the calibration parameter further includes a transformation parameter;
- the performing semantic segmentation of the image to be detected to obtain the location information of the ground area in the image to be detected includes: inputting the image to be detected into the first a deep neural network to obtain the semantic labels of all pixels in the image to be detected, wherein the semantic labels include ground labels and background labels; based on the transformation parameters, the ground labels in the image to be detected correspond to The pixel points of are transformed into the ground area in the camera coordinate system, and the position information of the ground in the camera coordinate system is obtained.
- the ground area and the background area in the image to be detected are distinguished by the semantic label of the first deep neural network, and the information of the image to be detected is further projected to the camera coordinate system through the transformation parameters, which is beneficial to reflect the difference between the ground area, the background area and the The distance relationship of the camera.
- the calibration parameter further includes a transformation parameter
- the object in the image to be detected includes a non-specific object
- the object detection is performed on the image to be detected
- the position information of the object in the image to be detected is obtained , further comprising: inputting the image to be detected into a second deep neural network to obtain a contour frame of a non-specific object in the image to be detected; based on the transformation parameters, transforming the contour frame of the non-specific object into The outline frame of the camera coordinate system is obtained, and the position information of the non-specific object in the camera coordinate system is obtained.
- the non-specific objects in the image to be detected are semantically segmented through the second deep neural network to demarcate the contour boxes of the non-specific objects, and the shape of the non-specific objects is represented by the contour boxes, which is conducive to sticking to the non-specific objects in path planning. Get a plan path with high practicability.
- the calibration parameter further includes a transformation parameter
- the object in the image to be detected includes a specific object
- the object detection is performed on the image to be detected
- the position information of the object in the image to be detected is obtained
- the method includes: inputting the image to be detected into a third deep neural network to obtain image position information of a specific object in the image to be detected; based on the transformation parameter, transforming the image position information of the object into The position information of the camera coordinate system.
- object detection is performed on the specific object in the image to be detected through the third deep neural network, so as to obtain the position information of the specific object.
- the inputting the to-be-detected image into a third deep neural network and acquiring the image position information of a specific object in the to-be-detected image includes: inputting the to-be-detected image into the third deep neural network The network is used to obtain the bounding box of the specific object in the to-be-detected image; and the image position information of the specific object is calculated based on the diagonal coordinates of the bounding box.
- the obstacle detection method further includes: forming a current planning map based on the position information of the object and the position information of the ground, wherein the current planning map includes a drivable area and includes the object The obstacle area is obtained; the planned route is obtained based on the drivable area of the current planning map.
- the driving area is obtained through the position information on the ground, and the obstacle area in the driving area is obtained through the position information of the object, thereby generating a current planning map for path planning.
- the obstacle detection method further includes: in response to the fact that the object in the to-be-detected image includes a non-specific object, determining that the obstacle area includes a contour frame of the non-specific object in the camera coordinate system and/or, in response to the situation that the object in the image to be detected includes a specific object, obtain the category information of the specific object; generate the specific object according to the position information and category information of the specific object.
- a method for forming an obstacle area is proposed, which is beneficial to quickly generate an obstacle area in the driving area according to the object frame; and for a specific object, considering the category information of the specific object can enable the obstacle detection device to determine whether it needs to be close to the specific object.
- planning which is conducive to improving the practicability of the planning path.
- a second aspect of an embodiment of the present disclosure provides an obstacle detection device, the obstacle detection device includes: a camera part configured to acquire an image to be detected; an object detection part configured to perform object detection on the image to be detected, Obtain the position information of the object in the to-be-detected image; the semantic segmentation part is configured to perform semantic segmentation on the to-be-detected image to obtain the location information of the ground area in the to-be-detected image; the obstacle detection part is configured to be based on The position information of the object and the position information of the ground area are used to obtain an obstacle area and a planned path to avoid the obstacle area.
- the camera part is configured to acquire calibration parameters of the camera, wherein the calibration parameters include distortion parameters; acquire a captured target image, and correct the target image based on the distortion parameters, Obtain the to-be-detected image.
- the calibration parameters further include transformation parameters;
- the semantic segmentation part is configured to input the image to be detected into a first deep neural network, and acquire semantic labels of all pixels in the image to be detected , wherein the semantic label includes a ground label and a background label; based on the transformation parameters, transform the pixel points corresponding to the ground label in the to-be-detected image into a ground area in the camera coordinate system to obtain the The location information of the ground area in the camera coordinate system.
- the calibration parameter further includes a transformation parameter
- the object in the image to be detected includes a non-specific object
- the object detection part is further configured to input the image to be detected into a second deep neural network, Obtain the outline frame of the non-specific object in the to-be-detected image; based on the transformation parameters, transform the outline frame of the non-specific object into the outline frame of the camera coordinate system, and obtain the non-specific object in the The position information of the camera coordinate system.
- the calibration parameter further includes a transformation parameter
- the object in the image to be detected includes a specific object
- the object detection part is further configured to input the image to be detected into a third deep neural network, and obtain Image position information of a specific object in the to-be-detected image; based on the transformation parameter, transform the image position information of the object into position information of the specific object in the camera coordinate system.
- the object detection part is further configured to input the image to be detected into the third deep neural network, and obtain a bounding frame of a specific object in the image to be detected;
- the diagonal coordinates calculate the image position information of the specific object.
- the apparatus further includes: a path planning part configured to form a current planning map based on the position information of the object and the position information of the ground, wherein the current planning map includes a drivable area, and an obstacle area including the object; obtaining a planned route based on the drivable area of the current planned map.
- a path planning part configured to form a current planning map based on the position information of the object and the position information of the ground, wherein the current planning map includes a drivable area, and an obstacle area including the object; obtaining a planned route based on the drivable area of the current planned map.
- the apparatus further includes: an area determination part configured to determine that the obstacle area includes the non-specific object in the camera in response to the fact that the object in the to-be-detected image includes a non-specific object The area corresponding to the outline frame of the coordinate system; and/or, configured to obtain the category information of the specific object in response to the fact that the object in the to-be-detected image includes a specific object; according to the position information of the specific object and The category information generates an object frame of the specific object in the camera coordinate system; wherein, the obstacle area includes an area corresponding to the specific object in the object frame of the camera coordinate system.
- a third aspect of the embodiments of the present disclosure provides an electronic device, including a mutually coupled memory and a processor, where the processor is configured to execute program instructions stored in the memory to implement the obstacle detection method in the first aspect.
- a fourth aspect of the embodiments of the present disclosure provides a computer-readable storage medium on which program instructions are stored, and when the program instructions are executed by a processor, the obstacle detection method in the first aspect above is implemented.
- a fifth aspect of the embodiments of the present disclosure provides a computer program, including computer-readable code, when the computer-readable code is executed in an electronic device, the processor in the electronic device executes the program to implement the above-mentioned first aspect The obstacle detection method in .
- the obstacle detection device obtains the image to be detected; performs object detection on the image to be detected to obtain the position information of the object in the image to be detected; performs semantic segmentation on the image to be detected to obtain the position information of the ground area in the image to be detected; and the position information of the ground area to obtain the obstacle area.
- the above solution can perform obstacle analysis on the environment.
- FIG. 1 is a schematic flowchart of an embodiment of an obstacle detection method provided by an embodiment of the present disclosure
- FIG. 2 is a schematic flowchart of an embodiment of step S11 in the obstacle detection method shown in FIG. 1;
- FIG. 3 is a schematic diagram of a frame of an image to be detected including an object frame provided by an embodiment of the present disclosure
- FIG. 4 is a schematic flowchart of an embodiment of step S12 in the obstacle detection method shown in FIG. 1;
- FIG. 5 is a schematic flowchart of another embodiment of step S12 in the obstacle detection method shown in FIG. 1;
- FIG. 6 is a schematic flowchart of an embodiment of step S13 in the obstacle detection method shown in FIG. 1;
- FIG. 7 is a schematic frame diagram of an embodiment of an obstacle detection device provided by an embodiment of the present disclosure.
- FIG. 8 is a schematic diagram of a framework of an embodiment of an electronic device provided by an embodiment of the present disclosure.
- FIG. 9 is a schematic diagram of a framework of an embodiment of a computer-readable storage medium provided by an embodiment of the present disclosure.
- Sweeping robots are an important part of home smart terminals. While intelligent sweeping robots need to locate their own position, they also need to sense the information of surrounding objects for path planning and avoid obstacles, such as shoes, chair feet, and socks. etc., to achieve an efficient cleaning process. Therefore, obstacle detection analysis is an important technical component in the path planning of intelligent sweepers.
- sweeping robots There are generally two types of sweeping robots: 1) sweeping robots based on single-line lidar; 2) sweeping robots based on binocular or monocular cameras.
- single-line lidar solution the advantage is that it covers a large angular range, and the general single-line lidar can achieve a real-time 360-degree scanning range.
- single-line lidar can only scan a single horizontal plane, so it cannot completely cover obstacles of different heights. For example, socks cannot be detected when they are close to the ground.
- the use of multi-line lidar can alleviate this problem to a certain extent, but the current cost of multi-line lidar is high, and it is not suitable for use in consumer-grade products such as home sweepers.
- this method has certain positional constraints between the two camera sensors of the binocular camera (for example, the two cameras must be separated by a certain distance and at the same height in the horizontal direction), and this scheme is relatively small for textures, such as white walls The scene adaptability is not good.
- the embodiment of the present disclosure proposes a solution using a monocular camera, which integrates the target detection technology and An obstacle analysis method for semantic segmentation techniques.
- an embodiment of the present disclosure proposes an obstacle detection method that can be applied to the realization of a sweeping robot in a home place.
- the obstacle detection method proposed in the embodiments of the present disclosure can also be applied to other robots, such as cleaning robots, lawn mowing robots, outdoor delivery robots, etc., which will not be repeated here.
- FIG. 1 is a schematic flowchart of an embodiment of an obstacle detection method provided by an embodiment of the present disclosure.
- the execution subject of the obstacle detection method in the embodiment of the present disclosure may be an obstacle detection apparatus.
- the obstacle detection method may be executed by a terminal device, a server, or other processing equipment, wherein the obstacle detection apparatus may be a user equipment (User Equipment, UE ), mobile devices, user terminals, terminals, cellular phones, wireless phones, Personal Digital Assistant (PDA), handheld devices, computing devices, in-vehicle devices, wearable devices, etc.
- the obstacle detection method may be implemented by the processor invoking computer-readable instructions stored in the memory.
- the obstacle detection method according to the embodiment of the present disclosure may include the following steps:
- Step S11 Acquire an image to be detected.
- the obstacle detection device acquires the to-be-detected image of the environment where the robot is located.
- the image to be detected may be acquired by a camera part mounted on the robot, and the camera part may specifically be a monocular red-green-blue (RGB) camera.
- the camera model is fixed at the fixed position of the robot, generally set at the front position of the robot running direction, and can obtain the environmental image in the forward direction of the robot, that is, the camera field of view needs to cover the ground and the possible obstacles that need to be analyzed.
- Step S12 Perform object detection on the image to be detected, and obtain position information of the object in the image to be detected.
- the obstacle detection apparatus performs object detection on the image to be detected, and obtains position information of the object in the image to be detected.
- the object in the image to be detected is the obstacle; through the position information of the object, the obstacle detection device can generate an effective obstacle area.
- the objects of the embodiments of the present disclosure can be classified into specific objects and non-specific objects.
- specific objects are objects with relatively fixed shapes preset by the staff in advance, such as slippers, paper balls, cans, etc.
- non-specific objects are objects with variable shapes, such as table legs, wires, etc.
- Step S13 Semantically segment the image to be detected, and obtain location information of the ground area in the image to be detected.
- the obstacle detection apparatus performs ground detection on the image to be detected through the semantic segmentation part.
- the semantic segmentation part in this step and the above-mentioned semantic segmentation part for detecting non-specific objects may be the same part or different parts.
- the input of the semantic segmentation part is the image to be detected after correction based on the calibrated distortion parameters, and the output is a number of pixels marked as the ground in the image to be detected, and the ground area composed of several pixels.
- Step S14 Obtain the obstacle area based on the position information of the object and the position information of the ground.
- the obstacle detection apparatus obtains the above-mentioned ground information and object information through the fusion part.
- the input information of the fusion part includes: (1) the information obtained by the semantic segmentation part, that is, relative to the camera, the area belonging to the ground in the current area and the position and shape of non-specific objects; (2) the information obtained by the object detection part, that is, relative to the camera.
- the camera runs the position and class of a specific object ahead.
- the obstacle detection device can form a current planning map according to the above ground information and object information, wherein the current planning map can be a two-dimensional map or a three-dimensional map.
- the obstacle area is jointly determined based on the position information of the object and the position information of the ground area, and the position information of the ground area in the image to be detected is obtained based on semantic segmentation, it is possible to improve the accuracy of the determined ground area. Therefore, the obstacle area with high accuracy can be obtained, and the obstacle analysis of the environment can be carried out effectively.
- the obstacle detection device may also preprocess the image to be detected to improve the accuracy of the image to be detected. Please refer to FIG. 2 for the specific preprocessing process.
- FIG. 2 shows the obstacle detection method shown in FIG. 1
- Step S11 is a schematic flowchart of an embodiment. As shown in Figure 2, step S11 specifically includes the following steps:
- Step S21 Obtain calibration parameters of the camera, wherein the calibration parameters include distortion parameters.
- the obstacle detection device needs to calibrate the camera.
- the obstacle detection device can calibrate the internal and external parameters of the camera part by Zhang Zhengyou's calibration method.
- the internal parameters may include, but are not limited to, the focal length and distortion parameters of the camera
- the external parameters may include, but are not limited to, the homography matrix of the camera.
- the distortion parameters include tangential distortion and radial distortion. The radial distortion occurs in the process of converting the camera coordinate system to the image physical coordinate system, and the tangential distortion occurs in the camera manufacturing process, because the plane of the photosensitive element is not parallel to the plane of the lens.
- the homography matrix is the transformation relationship of the projection mapping between the image physical coordinate system and the camera coordinate system.
- the internal and external parameters of the camera part can also be calibrated by other calibration methods, such as Meng Hu's plane calibration method, Wu Yihong's parallel circle calibration method, and the like.
- Step S22 acquiring the captured target image, and correcting the target image based on the distortion parameter to obtain an image to be detected.
- the tangential distortion and the radial distortion may cause partial deformation of the image to be detected partially captured by the camera. Therefore, after the obstacle detection device acquires the target image, it needs to correct the target image based on the distortion parameters, so as to reduce the influence of the image distortion caused by the tangential distortion and radial distortion of the camera itself, which is beneficial to improve the accuracy of subsequent path planning.
- the obstacle detection device can directly detect the object through the object detection part.
- FIG. 4 is a schematic flowchart of an embodiment of step S12 in the obstacle detection method shown in FIG. 1 .
- the calibration parameters also include transformation parameters.
- step S12 specifically includes the following steps:
- Step S31 Input the image to be detected into a third deep neural network, and obtain image position information of a specific object in the image to be detected.
- the object detection part includes a third deep neural network and a geometric projection part.
- the input of the third deep neural network is the image to be detected corrected based on the calibrated distortion parameters, and the output is the image to be detected containing the image position information of the specific object.
- the image position information of a specific object is represented by the object frame in FIG. 3 .
- the particular object shown in Figure 3 includes two balls of paper and a can.
- step S31 can be implemented in the following ways: inputting the image to be detected into a third deep neural network to obtain the bounding box of the specific object in the image to be detected; calculating the image position information of the specific object based on the diagonal coordinates of the bounding box .
- the third deep neural network pre-trains the specific object in the embodiments of the present disclosure, and the third deep neural network can identify the position and category of the specific object, which is embodied in the form of an object frame.
- the object box may be composed of a bounding box surrounding a specific object and a semantic label, and the semantic label indicates the category of the specific object corresponding to the bounding box.
- Step S32 Based on the transformation parameters, transform the image position information of the object into the position information of the specific object in the camera coordinate system.
- the obstacle detection apparatus transforms the image position information of the object into the position information of the specific object in the camera coordinate system based on the transformation parameters, and reflects it in the form of a bounding box.
- the transformation parameters are the external parameters of the calibrated camera part, including but not limited to the homography matrix of the camera.
- the obstacle detection device uses the geometric projection part to take the midpoint position between the upper left corner coordinate and the lower right corner coordinate of the bounding box, or the focus position between the upper right corner coordinate and the lower left corner coordinate as the specific object in the image , and then combined with the pre-calibrated homography matrix to calculate the position of a specific object relative to the camera.
- the ground area and the background area in the image to be detected are distinguished by the semantic label of the first deep neural network, and the information of the image to be detected is further projected to the camera coordinate system through the transformation parameters, which is beneficial to reflect the difference between the ground area, the background area and the The distance relationship of the camera.
- the obstacle detection apparatus marks the category information of the specific object on the bounding box to form the object frame of the specific object as shown in FIG. 3 .
- the category information of specific objects is conducive to planning a planning path that is closer to real life and has higher practicability. For example, when a specific object belongs to shoes and other objects that will not affect the robot, the planned path can be planned closely to the specific object; when the specific object belongs to more dangerous objects such as desk lamps and electric heaters, the planned path can be combined with Specific objects are planned at a certain distance to prevent hazards during robot operation.
- the obstacle detection device can detect them by means of semantic segmentation. It is worth noting that the process of detecting non-specific objects by means of semantic segmentation can be detected by the above-mentioned object detection part, or by another method.
- the semantic segmentation part is used for detection. Please refer to FIG. 5 for the specific detection process.
- FIG. 5 is a schematic flowchart of another embodiment of step S12 in the obstacle detection method shown in FIG. 1 .
- the calibration parameters also include transformation parameters, and the objects in the image to be detected include non-specific objects.
- step S12 specifically includes the following steps:
- Step S41 Input the image to be detected into the second deep neural network, and obtain the outline frame of the non-specific object in the image to be detected.
- the input of the second deep neural network is an image to be detected corrected based on the calibrated distortion parameters
- the output is an image to be detected that includes image position information of a non-specific object.
- the image position information of the non-specific object is embodied by the outline frame surrounding the non-specific object.
- Step S42 Based on the transformation parameters, transform the outline frame of the non-specific object into the outline frame in the camera coordinate system, and obtain the position information of the non-specific object in the camera coordinate system.
- the obstacle detection apparatus transforms the image position information of the non-specific object into the position information of the non-specific object in the camera coordinate system based on the homography matrix, and reflects it in the form of an outline frame.
- the above-mentioned third deep neural network can only output the object frame of a specific object, wherein the object frame is a rectangular frame surrounding the specific object; the recognition principle of the second deep neural network is to identify the edge points of non-specific objects, and then put a number of These edge points are combined and connected to form a closed edge line, that is, a contour box around a non-specific object.
- the outline frame of a non-specific object can better reflect the specific shape information of the object, which is beneficial to the robot's running path to be planned closely to the non-specific object, so as to improve the practicability of the robot's planning path. .
- the non-specific objects in the image to be detected are semantically segmented through the second deep neural network to demarcate the contour boxes of the non-specific objects, and the shape of the non-specific objects is represented by the contour boxes, which is conducive to sticking to the non-specific objects in path planning. Get a plan path with high practicability.
- FIG. 6 is a schematic flowchart of an embodiment of step S13 in the obstacle detection method shown in FIG. 1 .
- the calibration parameters also include transformation parameters.
- step S13 specifically includes the following steps:
- Step S51 Input the image to be detected into the first deep neural network, and acquire semantic labels of all pixels in the image to be detected, wherein the semantic labels include ground labels and background labels.
- the semantic segmentation part includes a fully convolutional first deep neural network and a geometric projection part.
- the input of the first deep neural network is the image to be detected corrected based on the calibrated distortion parameters, and the output is the semantic label of each pixel in the image to be detected.
- semantic segmentation part can also be used to segment the outline frame of non-specific objects, that is, the above-mentioned second deep neural network and the first deep neural network in this step can be the same deep neural network, which will not be repeated here.
- semantic tags specifically include ground tags and background tags. When a certain pixel is identified as a ground pixel, the semantic label is marked as 1; when a certain pixel is identified as a background pixel, the semantic label is marked as 0.
- Step S52 Based on the transformation parameters, transform the pixel points corresponding to the ground label in the image to be detected into the ground area in the camera coordinate system, and obtain the position information of the ground in the camera coordinate system.
- the obstacle detection apparatus projects each ground pixel point with a semantic label of 1 from the image space to the camera space based on the homography matrix, and obtains the position information of the ground pixel point in the camera coordinate system. Then, the obstacle detection device combines the projected ground pixels into the ground area of the camera coordinate system, and the remaining areas are background areas. Among them, the ground area is the driving area of the robot.
- the obstacle detection device obtains the image to be detected; performs object detection on the image to be detected to obtain the position information of the object in the image to be detected; performs semantic segmentation on the image to be detected to obtain the position information of the ground in the image to be detected; The position information and the position information of the ground area obtain the obstacle area.
- the obstacle detection device can automatically identify and mark the ground area and object position in the to-be-detected image by performing object detection and semantic segmentation on the to-be-detected image, wherein the ground area is the driving area of the robot, and the detected object When the position appears in the driving area, it can be analyzed that the object is an obstacle, so that the obstacle can be analyzed effectively in the environment, and the path planning can be further carried out according to the analysis result.
- the writing order of each step does not mean a strict execution order but constitutes any limitation on the implementation process, and the specific execution order of each step should be based on its function and possible Internal logic is determined.
- the ground area and the background area in the image to be detected are distinguished by the semantic label of the first deep neural network, and the information of the image to be detected is further projected to the camera coordinate system through the transformation parameters, which is beneficial to reflect the difference between the ground area, the background area and the The distance relationship of the camera.
- the obstacle detection method may further include: forming a current planning map based on the position information of the object and the position information of the ground, wherein the current planning map includes a drivable area and an obstacle area including the object; The drivable area gets the planned route.
- the driving area is obtained through the position information on the ground, and the obstacle area in the driving area is obtained through the position information of the object, thereby generating a current planning map for path planning.
- the current planning map includes drivable areas and obstacle areas for objects.
- the obstacle area is an area generated based on the position information of the object, and the drivable area is a part of the ground area excluding the obstacle area.
- the area enclosed by the contour frame of the camera coordinate system of the non-specific object can be regarded as the location of the non-specific object, so this area is the obstacle area ( or a part of the obstacle area);
- the category information of the specific object can be obtained through the above-mentioned third deep neural network (or other methods), and then according to the position information and category of the specific object
- the information generates the object frame of the specific object in the camera coordinate system, and the area enclosed by the object frame of the specific object in the camera coordinate system can be regarded as the location of the specific object, so this area is the obstacle area (or a part of the obstacle area).
- the size of the corresponding object frame may be different. This is considering that different specific objects have different effects on the robot, so they will correspond to different sizes of object frames so that the robot can plan more Precise path; for example, when a specific object belongs to shoes and other objects that will not affect the robot, the size of the object frame is small, that is, the obstacle area is small, and the planned path can be planned close to the specific object; In the case of more dangerous objects such as desk lamps and electric heaters, the size of the object frame is large, that is, the obstacle area is large, and the planned path can be planned at a certain distance from the specific object to prevent the robot from causing danger during operation. In the subsequent path planning process, considering the category information of specific objects is conducive to planning a planning path that is closer to real life and has higher practicability.
- the obstacle detection device can obtain the drivable area of the robot in the driving area, and then input the drivable area into the trajectory planning part.
- the obstacle detection device inputs the end point information through the trajectory planning part, and obtains the planned path to avoid the obstacle area.
- the obstacle detection method may further include: in response to the fact that the object in the image to be detected includes a non-specific object, determining that the obstacle area includes an area corresponding to the contour frame of the non-specific object in the camera coordinate system.
- the obstacle detection method may further include: in response to a situation in which the object in the image to be detected includes a specific object, acquiring category information of the specific object; generating the position information and category information of the specific object in the camera coordinate system The object frame; wherein, the obstacle area includes the area corresponding to the object frame of the specific object in the camera coordinate system.
- a method for forming an obstacle area is proposed, which is beneficial to quickly generate an obstacle area in the driving area according to the object frame; and for a specific object, considering the category information of the specific object can enable the obstacle detection device to determine whether it needs to be close to the specific object.
- planning which is conducive to improving the practicability of the planning path.
- FIG. 7 is a schematic diagram of a framework of an embodiment of an obstacle detection apparatus provided by an embodiment of the present disclosure.
- the obstacle detection device 70 includes:
- the camera part 71 is configured to acquire the image to be detected.
- the object detection part 72 is configured to perform object detection on the image to be detected, and obtain position information of the object in the image to be detected.
- the semantic segmentation part 73 is configured to perform semantic segmentation on the image to be detected, and obtain location information of the ground area in the image to be detected.
- the obstacle detection section 74 is configured to acquire the obstacle area based on the position information of the object and the position information of the ground area.
- the camera part 71 is configured to acquire calibration parameters of the camera, wherein the calibration parameters include distortion parameters; acquire a captured target image, and correct the target image based on the distortion parameters to obtain an image to be detected.
- the calibration parameters further include transformation parameters;
- the semantic segmentation part 73 is configured to input the image to be detected into the first deep neural network, and obtain semantic labels of all pixels in the image to be detected, wherein the semantic labels include ground label and background label; based on the transformation parameters, transform the pixels corresponding to the ground label in the image to be detected into the ground area in the camera coordinate system, and obtain the position information of the ground area in the camera coordinate system.
- the calibration parameter further includes a transformation parameter
- the object in the image to be detected includes a non-specific object
- the object detection part 72 is further configured to input the image to be detected into the second deep neural network, and obtain the non-specific object in the image to be detected.
- the outline frame of the object; based on the transformation parameters, the outline frame of the non-specific object is transformed into the outline frame in the camera coordinate system, and the position information of the non-specific object in the camera coordinate system is obtained.
- the calibration parameters further include transformation parameters
- the object in the image to be detected includes a specific object
- the object detection part 72 is further configured to input the image to be detected into a third deep neural network, and obtain the specific object in the image to be detected.
- Image position information based on the transformation parameters, transform the image position information of the object into the position information of a specific object in the camera coordinate system.
- the object detection part 72 is further configured to input the image to be detected into the third deep neural network to obtain the bounding box of the specific object in the image to be detected; calculate the image position of the specific object based on the diagonal coordinates of the bounding box information.
- the apparatus further includes: a path planning part configured to form a current planning map based on the position information of the object and the position information of the ground, wherein the current planning map includes a drivable area and an obstacle area including the object; Get the planned route from the drivable area of the current planning map.
- the apparatus further includes: an area determination part configured to, in response to a situation in which the object in the image to be detected includes a non-specific object, determine that the obstacle area includes an area corresponding to the contour frame of the non-specific object in the camera coordinate system; Or, it is configured to obtain category information of the specific object in response to the fact that the object in the image to be detected includes a specific object; generate an object frame of the specific object in the camera coordinate system according to the position information and category information of the specific object; wherein the obstacle area includes The area corresponding to a specific object in the object frame of the camera coordinate system.
- the system for implementing the solutions of the embodiments of the present disclosure may include the following parts: a camera part, an object detection part, a ground segmentation part (corresponding to the semantic segmentation part in the above-mentioned embodiment), and a fusion part (corresponding to the above-mentioned semantic segmentation part)
- the obstacle detection part in the embodiment may include the following parts: a camera part, an object detection part, a ground segmentation part (corresponding to the semantic segmentation part in the above-mentioned embodiment), and a fusion part (corresponding to the above-mentioned semantic segmentation part)
- the obstacle detection part in the embodiment and the trajectory planning part.
- the camera part can receive the current image and distribute the image to the object detection part and the semantic segmentation part respectively.
- the object detection part detects all the objects in the set A in the picture, and outputs the object category and the position of the object relative to the current robot.
- the ground segmentation part labels the pixels in the image for the input image, and outputs the area that belongs to the flat ground.
- the fusion part outputs a map of the area where the sweeping robot can travel. According to this map, the trajectory planning part plans the trajectory of obstacle avoidance, and finally the system control part completes the execution of the obstacle avoidance trajectory.
- the camera part needs to be fixed at a fixed position of the sweeping robot to determine whether the field of view of the camera part can cover the ground and possible obstacles that need to be analyzed. After fixing the camera part, there needs to be a calibration process.
- the internal parameter I of the camera part (including focal length and distortion parameters) is calibrated
- the external parameter E (here the external parameter refers to the homography matrix between the camera imaging plane and the two planes on the ground) is calibrated.
- the calibration here can use Zhang Zhengyou's calibration method.
- the input of the object detection part can be the corrected image based on the calibrated distortion parameters, and the output is the category and position of a specific type of obstacle.
- a specific type of obstacle is an object with a relatively fixed shape (eg, slippers, paper balls, cans, etc.) that are formulated before development by the system developer.
- the object detection part can include a deep neural network M1 and a geometric projection part G.
- the input of neural network A is the picture, and the output is the bounding box of the detected object.
- the geometric projection part uses the midpoint between the lower left corner and the lower right corner of the bounding box as the position of the object in the image, combined with pre-calibration
- a good external parameter E can calculate the position of the object relative to the camera part.
- the input of the ground segmentation part is the image rectified based on the calibrated distortion parameters, and the output is the semantic label of each pixel in the input image.
- the semantic label is equal to 0 or 1.
- it means that the pixel is the ground, otherwise it is the background.
- the ground segmentation part can be implemented by a fully convolutional deep network M2. After obtaining the semantic label of each pixel, through the geometric part G, each pixel with semantic ground can be projected from the image space to the camera space (ie, the coordinates relative to the camera).
- the fusion part can determine the following information: 1) The information obtained by the ground segmentation part, that is, which areas are currently on the ground relative to the camera part; 2) The information obtained by the object detection part, that is, relative to the camera part, where are the obstacles ahead? object. These two aspects of information are in the camera space system, so the two aspects of information are comprehensively analyzed, and the fusion part obtains which areas are drivable within the range covered by the current camera part.
- the drivable area can be described using a local map. By inputting the local map into the trajectory planning part, the trajectory of obstacle avoidance can be obtained.
- this solution is used to analyze obstacles, so as to know which obstacles are in front of the robot and which are walkable areas, and feed back this information to the planning control system, thereby Achieve obstacle avoidance.
- FIG. 8 is a schematic diagram of a framework of an embodiment of an electronic device provided by an embodiment of the present disclosure.
- the electronic device 80 includes a memory 81 and a processor 82 coupled to each other, and the processor 82 is configured to execute program instructions stored in the memory 81 to implement the steps in any of the above-mentioned embodiments of the obstacle detection method.
- the electronic device 80 may include, but is not limited to, a microcomputer and a server.
- the electronic device 80 may also include mobile devices such as a notebook computer and a tablet computer, which are not limited herein.
- the processor 82 is configured to control itself and the memory 81 to implement the steps in any of the above-mentioned embodiments of the obstacle detection method.
- the processor 82 may also be referred to as a central processing unit (Central Processing Unit, CPU).
- the processor 82 may be an integrated circuit chip with signal processing capability.
- the processor 82 may also be a general-purpose processor, a digital signal processor (Digital Signal Processor, DSP), an application-specific integrated circuit (Application Specific Integrated Circuit, ASIC), a Field-Programmable Gate Array (Field-Programmable Gate Array, FPGA) or other Programmable logic devices, discrete gate or transistor logic devices, discrete hardware components.
- a general purpose processor may be a microprocessor or the processor may be any conventional processor or the like.
- the processor 82 may be jointly implemented by an integrated circuit chip.
- FIG. 9 is a schematic diagram of a framework of an embodiment of a computer-readable storage medium provided by an embodiment of the present disclosure.
- the computer-readable storage medium 90 stores program instructions 901 that can be executed by the processor, and the program instructions 901 are used to implement the steps in any of the above-mentioned embodiments of the obstacle detection method.
- the functions or included parts of the apparatus may be used to execute the methods described in the above method embodiments, and the specific implementation may refer to the descriptions of the above method embodiments. No longer.
- Embodiments of the present disclosure may further provide a computer program, including computer-readable code, when the computer-readable code is executed in an electronic device, the processor in the electronic device executes the program for implementing any of the above The steps in the embodiment of the obstacle detection method.
- the disclosed method and apparatus may be implemented in other manners.
- the device implementations described above are only illustrative.
- the division of parts or units is only a logical function division. In actual implementation, there may be other divisions.
- units or components may be combined or integrated. to another system, or some features can be ignored, or not implemented.
- the shown or discussed mutual coupling or direct coupling or communication connection may be through some interfaces, indirect coupling or communication connection of devices or units, which may be in electrical, mechanical or other forms.
- each functional unit in each embodiment of the embodiments of the present disclosure may be integrated into one processing unit, or each unit may exist physically alone, or two or more units may be integrated into one unit.
- the above-mentioned integrated units may be implemented in the form of hardware, or may be implemented in the form of software functional units.
- the integrated unit if implemented as a software functional unit and sold or used as a stand-alone product, may be stored in a computer-readable storage medium.
- the technical solutions of the embodiments of the present disclosure may be embodied in the form of software products in essence, or the parts that contribute to the related technology, or all or part of the technical solutions, and the computer software products are stored in a storage medium. , including several instructions to cause a computer device (which may be a personal computer, a server, or a network device, etc.) or a processor (processor) to execute all or part of the steps of the methods of the various implementations of the embodiments of the present disclosure.
- the aforementioned storage medium includes: U disk, mobile hard disk, Read-Only Memory (ROM, Read-Only Memory), Random Access Memory (RAM, Random Access Memory), magnetic disk or optical disk and other media that can store program codes .
- the obstacle area is jointly determined based on the position information of the object and the position information of the ground area, and the position information of the ground area in the image to be detected is obtained based on semantic segmentation, it is possible to improve the determination accuracy.
- the accuracy of the ground area, and then the obstacle area with high accuracy can be obtained, so that the obstacle analysis of the environment can be carried out effectively.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Evolutionary Computation (AREA)
- Multimedia (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Computing Systems (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Biophysics (AREA)
- General Engineering & Computer Science (AREA)
- Biomedical Technology (AREA)
- Mathematical Physics (AREA)
- Data Mining & Analysis (AREA)
- Molecular Biology (AREA)
- Life Sciences & Earth Sciences (AREA)
- Computational Linguistics (AREA)
- Databases & Information Systems (AREA)
- Medical Informatics (AREA)
- Image Analysis (AREA)
- Geometry (AREA)
Abstract
Description
Claims (19)
- 一种障碍检测方法,所述障碍检测方法包括:获取待检测图像;将所述待检测图像进行物体检测,获取所述待检测图像中物体的位置信息;将所述待检测图像进行语义分割,获取所述待检测图像中地面区域的位置信息;基于所述物体的位置信息以及所述地面区域的位置信息获取障碍区域。
- 根据权利要求1所述的障碍检测方法,其中,所述获取待检测图像,包括:获取相机的标定参数,其中,所述标定参数包括畸变参数;获取拍摄的目标图像,并基于所述畸变参数对所述目标图像进行矫正,得到所述待检测图像。
- 根据权利要求2所述的障碍检测方法,其中,所述标定参数还包括变换参数;所述将所述待检测图像进行语义分割,获取所述待检测图像中地面区域的位置信息,包括:将所述待检测图像输入第一深度神经网络,获取所述待检测图像中所有像素点的语义标签,其中,所述语义标签包括地面标签和背景标签;基于所述变换参数,将所述待检测图像中所述地面标签对应的像素点变换为在所述相机坐标系的地面区域,得到所述地面区域在所述相机坐标系的位置信息。
- 根据权利要求2或3所述的障碍检测方法,其中,所述标定参数还包括变换参数,所述待检测图像中物体包括非特定物体;所述将所述待检测图像进行物体检测,获取所述待检测图像中物体的位置信息,还包括:将所述待检测图像输入第二深度神经网络,获取所述待检测图像中所述非特定物体的轮廓框;基于所述变换参数,将所述非特定物体的轮廓框变换为在所述相机坐标系的轮廓框,得到所述非特定物体在所述相机坐标系的位置信息。
- 根据权利要求2至4任一项所述的障碍检测方法,其中,所述标定参数还包括变换参数,所述待检测图像中物体包括特定物体;所述将所述待检测图像进行物体检测,获取所述待检测图像中物体的位置信息,包括:将所述待检测图像输入第三深度神经网络,获取所述待检测图像中特 定物体的图像位置信息;基于所述变换参数,将所述物体的图像位置信息变换为所述特定物体在所述相机坐标系的位置信息。
- 根据权利要求5所述的障碍检测方法,其中,所述将所述待检测图像输入第三深度神经网络,获取所述待检测图像中特定物体的图像位置信息,包括:将所述待检测图像输入所述第三深度神经网络,获取所述待检测图像中特定物体的包围框;基于所述包围框的对角坐标计算所述特定物体的图像位置信息。
- 根据权利要求1至6中任一项所述的障碍检测方法,其中,所述障碍检测方法还包括:基于所述物体的位置信息以及所述地面的位置信息形成当前规划地图,其中,所述当前规划地图包括可行驶区域,以及包括所述物体的障碍区域;基于所述当前规划地图的可行驶区域获取规划路径。
- 根据权利要求7所述的障碍检测方法,其中,所述障碍检测方法还包括:响应于所述待检测图像中物体包括非特定物体的情况,确定所述障碍区域包括所述非特定物体在所述相机坐标系的轮廓框所对应的区域;和/或,响应于所述待检测图像中物体包括特定物体的情况,获取所述特定物体的类别信息;根据所述特定物体的位置信息和类别信息生成所述特定物体在相机坐标系的物体框;其中,所述障碍区域包括所述特定物体在所述相机坐标系的物体框所对应的区域。
- 一种障碍检测装置,所述障碍检测装置包括:相机部分,被配置为获取待检测图像;物体检测部分,被配置为将所述待检测图像进行物体检测,获取所述待检测图像中物体的位置信息;语义分割部分,被配置为将所述待检测图像进行语义分割,获取所述待检测图像中地面区域的位置信息;障碍检测部分,被配置为基于所述物体的位置信息以及所述地面区域的位置信息获取障碍区域。
- 根据权利要求9所述的装置,其中,所述相机部分,被配置为获取相机的标定参数,其中,所述标定参数包括畸变参数;获取拍摄的目标图像,并基于所述畸变参数对所述目标图像进行矫正,得到所述待检测图像。
- 根据权利要求10所述的装置,其中,所述标定参数还包括变换参数;所述语义分割部分,被配置为将所述待检测图像输入第一深度神经网络,获取所述待检测图像中所有像素点的语义标签,其中,所述语义标签包括地面标签和背景标签;基于所述变换参数,将所述待检测图像中所述 地面标签对应的像素点变换为在所述相机坐标系的地面区域,得到所述地面区域在所述相机坐标系的位置信息。
- 根据权利要求10或11所述的装置,其中,所述标定参数还包括变换参数,所述待检测图像中物体包括非特定物体;所述物体检测部分,还被配置为将所述待检测图像输入第二深度神经网络,获取所述待检测图像中所述非特定物体的轮廓框;基于所述变换参数,将所述非特定物体的轮廓框变换为在所述相机坐标系的轮廓框,得到所述非特定物体在所述相机坐标系的位置信息。
- 根据权利要求10至12任一项所述的装置,其中,所述标定参数还包括变换参数,所述待检测图像中物体包括特定物体;所述物体检测部分,还被配置为将所述待检测图像输入第三深度神经网络,获取所述待检测图像中特定物体的图像位置信息;基于所述变换参数,将所述物体的图像位置信息变换为所述特定物体在所述相机坐标系的位置信息。
- 根据权利要求13所述的装置,其中,所述物体检测部分,还被配置为将所述待检测图像输入所述第三深度神经网络,获取所述待检测图像中特定物体的包围框;基于所述包围框的对角坐标计算所述特定物体的图像位置信息。
- 根据权利要求9至14任一项所述的装置,其中,所述装置还包括:路径规划部分,被配置为基于所述物体的位置信息以及所述地面的位置信息形成当前规划地图,其中,所述当前规划地图包括可行驶区域,以及包括所述物体的障碍区域;基于所述当前规划地图的可行驶区域获取规划路径。
- 根据权利要求15所述的装置,其中,所述装置还包括:区域确定部分,被配置为响应于所述待检测图像中物体包括非特定物体的情况,确定所述障碍区域包括所述非特定物体在所述相机坐标系的轮廓框所对应的区域;和/或,被配置为响应于所述待检测图像中物体包括特定物体的情况,获取所述特定物体的类别信息;根据所述特定物体的位置信息和类别信息生成所述特定物体在相机坐标系的物体框;其中,所述障碍区域包括所述特定物体在所述相机坐标系的物体框所对应的区域。
- 一种电子设备,包括相互耦接的存储器和处理器,所述处理器被配置为执行所述存储器中存储的程序指令,以实现权利要求1至8任一项所述的障碍检测方法。
- 一种计算机可读存储介质,其上存储有程序指令,所述程序指令被处理器执行时实现权利要求1至8任一项所述的障碍检测方法。
- 一种计算机程序,包括计算机可读代码,当所述计算机可读代码在电子设备中运行时,所述电子设备中的处理器执行用于实现如权利要求1至8任一项所述的障碍检测方法。
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020217034499A KR20220060500A (ko) | 2020-10-28 | 2021-04-08 | 장애 검출 방법, 장치, 전자 기기, 저장 매체 및 컴퓨터 프로그램 |
JP2021562323A JP2023503747A (ja) | 2020-10-28 | 2021-04-08 | 障害検出方法及び装置、電子機器、記憶媒体並びにコンピュータプログラム |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011172403.6A CN112183476B (zh) | 2020-10-28 | 2020-10-28 | 一种障碍检测方法、装置、电子设备以及存储介质 |
CN202011172403.6 | 2020-10-28 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2022088611A1 true WO2022088611A1 (zh) | 2022-05-05 |
Family
ID=73923531
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2021/085927 WO2022088611A1 (zh) | 2020-10-28 | 2021-04-08 | 障碍检测方法、装置、电子设备、存储介质及计算机程序 |
Country Status (4)
Country | Link |
---|---|
JP (1) | JP2023503747A (zh) |
KR (1) | KR20220060500A (zh) |
CN (1) | CN112183476B (zh) |
WO (1) | WO2022088611A1 (zh) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115376109A (zh) * | 2022-10-25 | 2022-11-22 | 杭州华橙软件技术有限公司 | 障碍物检测方法、障碍物检测装置以及存储介质 |
Families Citing this family (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112183476B (zh) * | 2020-10-28 | 2022-12-23 | 深圳市商汤科技有限公司 | 一种障碍检测方法、装置、电子设备以及存储介质 |
CN112711263A (zh) * | 2021-01-19 | 2021-04-27 | 未来机器人(深圳)有限公司 | 仓储自动导引车避障方法、装置、计算机设备和存储介质 |
CN113158779B (zh) * | 2021-03-09 | 2024-06-07 | 美智纵横科技有限责任公司 | 一种行走方法、装置和计算机存储介质 |
CN112966658B (zh) * | 2021-03-30 | 2024-06-18 | 深圳市优必选科技股份有限公司 | 机器人导航方法、装置、终端设备及计算机可读存储介质 |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR101690511B1 (ko) * | 2015-09-14 | 2016-12-28 | 주식회사 서연전자 | 자율 주행 차량의 제어 시스템 및 그 제어 방법 |
CN111126251A (zh) * | 2019-12-20 | 2020-05-08 | 深圳市商汤科技有限公司 | 图像处理方法、装置、设备和存储介质 |
CN112183476A (zh) * | 2020-10-28 | 2021-01-05 | 深圳市商汤科技有限公司 | 一种障碍检测方法、装置、电子设备以及存储介质 |
Family Cites Families (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP5635770B2 (ja) * | 2009-12-24 | 2014-12-03 | 京セラ株式会社 | 障害物検出装置 |
CN105115490A (zh) * | 2015-07-16 | 2015-12-02 | 深圳前海达闼科技有限公司 | 一种确定室内可活动区域的方法及装置 |
CN110246142A (zh) * | 2019-06-14 | 2019-09-17 | 深圳前海达闼云端智能科技有限公司 | 一种检测障碍物的方法、终端和可读存储介质 |
CN110502019A (zh) * | 2019-09-06 | 2019-11-26 | 北京云迹科技有限公司 | 一种室内机器人的避障方法及装置 |
CN110974088B (zh) * | 2019-11-29 | 2021-09-24 | 深圳市杉川机器人有限公司 | 扫地机器人控制方法、扫地机器人及存储介质 |
CN111462135B (zh) * | 2020-03-31 | 2023-04-21 | 华东理工大学 | 基于视觉slam与二维语义分割的语义建图方法 |
-
2020
- 2020-10-28 CN CN202011172403.6A patent/CN112183476B/zh active Active
-
2021
- 2021-04-08 KR KR1020217034499A patent/KR20220060500A/ko unknown
- 2021-04-08 WO PCT/CN2021/085927 patent/WO2022088611A1/zh active Application Filing
- 2021-04-08 JP JP2021562323A patent/JP2023503747A/ja not_active Withdrawn
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR101690511B1 (ko) * | 2015-09-14 | 2016-12-28 | 주식회사 서연전자 | 자율 주행 차량의 제어 시스템 및 그 제어 방법 |
CN111126251A (zh) * | 2019-12-20 | 2020-05-08 | 深圳市商汤科技有限公司 | 图像处理方法、装置、设备和存储介质 |
CN112183476A (zh) * | 2020-10-28 | 2021-01-05 | 深圳市商汤科技有限公司 | 一种障碍检测方法、装置、电子设备以及存储介质 |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115376109A (zh) * | 2022-10-25 | 2022-11-22 | 杭州华橙软件技术有限公司 | 障碍物检测方法、障碍物检测装置以及存储介质 |
Also Published As
Publication number | Publication date |
---|---|
KR20220060500A (ko) | 2022-05-11 |
CN112183476B (zh) | 2022-12-23 |
JP2023503747A (ja) | 2023-02-01 |
CN112183476A (zh) | 2021-01-05 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2022088611A1 (zh) | 障碍检测方法、装置、电子设备、存储介质及计算机程序 | |
US10740920B1 (en) | Method and apparatus for combining data to construct a floor plan | |
Kang et al. | Automatic targetless camera–lidar calibration by aligning edge with gaussian mixture model | |
CN111126251B (zh) | 图像处理方法、装置、设备和存储介质 | |
Menze et al. | Object scene flow for autonomous vehicles | |
WO2020102944A1 (zh) | 点云处理方法、设备及存储介质 | |
WO2019120011A1 (zh) | 一种目标检测方法及装置 | |
Liang et al. | Image based localization in indoor environments | |
Treible et al. | Cats: A color and thermal stereo benchmark | |
Wang et al. | Color-based road detection and its evaluation on the KITTI road benchmark | |
EP4283567A1 (en) | Three-dimensional map construction method and apparatus | |
US9886769B1 (en) | Use of 3D depth map with low and high resolution 2D images for gesture recognition and object tracking systems | |
WO2022078513A1 (zh) | 定位方法、装置、自移动设备和存储介质 | |
CN112700552A (zh) | 三维物体检测方法、装置、电子设备及介质 | |
CN110838144B (zh) | 一种充电设备识别方法、移动机器人和充电设备识别系统 | |
CN113359692B (zh) | 一种障碍物的避让方法、可移动机器人 | |
WO2022078488A1 (zh) | 定位方法、装置、自移动设备和存储介质 | |
WO2021114776A1 (en) | Object detection method, object detection device, terminal device, and medium | |
Gu et al. | Histograms of the normalized inverse depth and line scanning for urban road detection | |
CN112883955A (zh) | 货架布局检测方法、装置及计算机可读存储介质 | |
TW202029134A (zh) | 行車偵測方法、車輛及行車處理裝置 | |
Cociaş et al. | Multiple-superquadrics based object surface estimation for grasping in service robotics | |
Cheng et al. | Structured light-based shape measurement system | |
CN114155349A (zh) | 一种三维建图方法、三维建图装置及机器人 | |
Bodensteiner et al. | Accurate single image multi-modal camera pose estimation |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
ENP | Entry into the national phase |
Ref document number: 2021562323 Country of ref document: JP Kind code of ref document: A |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 21884355 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 17.08.2023) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 21884355 Country of ref document: EP Kind code of ref document: A1 |