WO2021134325A1 - Obstacle detection method and apparatus based on driverless technology and computer device - Google Patents

Obstacle detection method and apparatus based on driverless technology and computer device Download PDF

Info

Publication number
WO2021134325A1
WO2021134325A1 PCT/CN2019/130155 CN2019130155W WO2021134325A1 WO 2021134325 A1 WO2021134325 A1 WO 2021134325A1 CN 2019130155 W CN2019130155 W CN 2019130155W WO 2021134325 A1 WO2021134325 A1 WO 2021134325A1
Authority
WO
WIPO (PCT)
Prior art keywords
feature information
point cloud
current frame
data
extraction
Prior art date
Application number
PCT/CN2019/130155
Other languages
French (fr)
Chinese (zh)
Inventor
邹晓艺
何明
叶茂盛
吴伟
许双杰
许家妙
曹通易
Original Assignee
深圳元戎启行科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 深圳元戎启行科技有限公司 filed Critical 深圳元戎启行科技有限公司
Priority to PCT/CN2019/130155 priority Critical patent/WO2021134325A1/en
Priority to CN201980037716.XA priority patent/CN113678136B/en
Publication of WO2021134325A1 publication Critical patent/WO2021134325A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition

Definitions

  • This application relates to an obstacle detection method, device, computer equipment and storage medium based on unmanned driving technology.
  • the point cloud data is projected into the image data to obtain feature information of multiple channels, and then obstacle detection is performed based on the feature information of multiple channels.
  • some information may be lost, resulting in the extraction of effective feature information of each source data is not comprehensive enough, resulting in low accuracy of obstacle detection.
  • an obstacle detection method, device, computer device, and storage medium based on an unmanned driving technology that can improve the accuracy of obstacle detection in an unmanned driving process are provided.
  • An obstacle detection method based on unmanned driving technology including:
  • the point cloud feature information corresponding to each perspective and the current frame image data are input into the corresponding feature extraction model, and the spatial feature information corresponding to each perspective and the current frame image data are extracted in parallel through the corresponding feature extraction model Image feature information;
  • the fused feature information is input into a trained detection model, and the fused feature information is predicted and calculated through the detection model, and an obstacle detection result is output.
  • An obstacle detection device based on unmanned driving technology including:
  • the acquisition module is used to acquire the current frame point cloud data and the current frame image data within a preset angle range
  • a projection module configured to project the point cloud data of the current frame on multiple viewing angles to obtain two-dimensional planes corresponding to the multiple viewing angles;
  • the first extraction module is used to perform feature extraction on the two-dimensional plane corresponding to each perspective to obtain point cloud feature information corresponding to each perspective;
  • the second extraction module is used to input the point cloud feature information corresponding to each perspective and the current frame image data into the corresponding feature extraction model, and extract the spatial feature information corresponding to each perspective in parallel through the corresponding feature extraction model.
  • the fusion module is used to fuse the spatial feature information corresponding to multiple perspectives with the image feature information to obtain the fused feature information;
  • the prediction module is used to input the fused feature information into the trained detection model, and perform prediction operations on the fused feature information through the detection model, and output obstacle detection results.
  • a computer device including a memory and one or more processors, the memory stores computer readable instructions, and when the computer readable instructions are executed by the processor, the one or more processors execute The following steps:
  • the point cloud feature information corresponding to each perspective and the current frame image data are input into the corresponding feature extraction model, and the spatial feature information corresponding to each perspective and the current frame image data are extracted in parallel through the corresponding feature extraction model Image feature information;
  • the fused feature information is input into a trained detection model, and the fused feature information is predicted and calculated through the detection model, and an obstacle detection result is output.
  • One or more non-volatile computer-readable storage media storing computer-readable instructions.
  • the computer-readable instructions When executed by one or more processors, the one or more processors perform the following steps:
  • the fused feature information is input into a trained detection model, and the fused feature information is predicted and calculated through the detection model, and an obstacle detection result is output.
  • Fig. 1 is an application environment diagram of an obstacle detection method based on unmanned driving technology in one or more embodiments.
  • Fig. 2 is a schematic flowchart of an obstacle detection method based on unmanned driving technology in one or more embodiments.
  • FIG. 3 is a schematic flowchart of the step of fusing spatial feature information and image feature information corresponding to multiple viewing angles to obtain fused feature information in one or more embodiments.
  • Fig. 4 is a block diagram of an obstacle detection device based on unmanned driving technology in one or more embodiments.
  • Figure 5 is a block diagram of a computer device in one or more embodiments.
  • the obstacle detection method based on the unmanned driving technology provided in this application can be applied to the schematic diagram of obstacle detection during the unmanned driving process as shown in FIG. 1.
  • the first vehicle-mounted sensor 102 sends the collected point cloud data of the current frame to the vehicle-mounted computer device 104.
  • the first vehicle-mounted sensor may be a lidar.
  • On-board computer equipment can be referred to as computer equipment.
  • the second vehicle-mounted sensor 106 sends the collected image data of the current frame within the preset angle range to the computer device 104.
  • the second vehicle-mounted sensor may be a vehicle-mounted camera.
  • the computer device 104 projects the point cloud data of the current frame on multiple viewing angles to obtain two-dimensional planes corresponding to the multiple viewing angles.
  • the computer device 104 performs feature extraction on the two-dimensional plane corresponding to each view angle to obtain point cloud feature information corresponding to each view angle.
  • the computer device 104 inputs the point cloud feature information corresponding to each perspective and the current frame image data into the corresponding feature extraction model, and extracts the spatial feature information corresponding to each perspective and the current frame image data in parallel through the corresponding feature extraction model. Image feature information.
  • the computer device 104 fuses the spatial feature information corresponding to the multiple viewing angles with the current frame image data to obtain the fused feature information.
  • the computer device 104 inputs the fused feature information into the trained detection model, and performs a prediction operation on the fused feature information through the detection model, and outputs an obstacle detection result.
  • an obstacle detection method based on unmanned driving technology is provided. Taking the method applied to the computer equipment in FIG. 1 as an example for description, the method includes the following steps:
  • Step 202 Obtain current frame point cloud data and current frame image data within a preset angle range.
  • the collected current frame point cloud data is transmitted to the computer device through the first on-board sensor installed on the vehicle, and the preset angle range collected by the second on-board sensor installed on the vehicle
  • the image data of the current frame within is sent to the computer device.
  • the first vehicle-mounted sensor may be a lidar.
  • the current frame point cloud data is the current frame point cloud data within a 360-degree range collected by the first vehicle-mounted sensor.
  • the second vehicle-mounted sensor may be a vehicle-mounted camera.
  • the current frame image data within the preset angle range may be the current frame image data within a 360-degree range around the vehicle collected by multiple on-board cameras.
  • Step 204 Project the point cloud data of the current frame on multiple viewing angles to obtain two-dimensional planes corresponding to the multiple viewing angles.
  • Step 206 Perform feature extraction on the two-dimensional plane corresponding to each view angle to obtain point cloud feature information corresponding to each view angle.
  • the point cloud data of the current frame is 3D point cloud data.
  • the computer device projects the acquired point cloud data of the current frame to multiple viewing angles, thereby projecting the 3D point cloud data into the two-dimensional planes corresponding to the multiple viewing angles, and realizes the conversion of the 3D point cloud data into the two-dimensional data in the two-dimensional plane .
  • Multiple viewing angles may include a bird's-eye view and a front view.
  • the computer device projects the current point cloud data on the bird's-eye view angle
  • a two-dimensional plane corresponding to the bird's-eye view angle can be obtained.
  • the computer device projects the point cloud data of the current frame on the orthographic perspective
  • a two-dimensional plane corresponding to the orthographic perspective can be obtained.
  • the two-dimensional plane corresponding to each view includes the point cloud data of the current frame after projection.
  • the computer device can extract the point cloud feature information corresponding to each perspective in the two-dimensional plane corresponding to each perspective.
  • the point cloud feature information may be the local feature information of each point in the current frame point cloud data corresponding to each pixel in the two-dimensional plane, and the local feature information may include local depth, point cloud density, and the like.
  • the trained neural network model is pre-stored in the computer equipment.
  • the neural network model can be a pointnet based on the attention layer.
  • the computer device can input the two-dimensional plane corresponding to each perspective into the trained neural network model, and perform prediction operations on the two-dimensional plane corresponding to each perspective through the neural network model to obtain the point cloud feature information corresponding to each perspective .
  • Step 208 Input the point cloud feature information and current frame image data corresponding to each perspective into the corresponding feature extraction model, and extract the spatial feature information corresponding to each perspective and the current frame image data in parallel through the corresponding feature extraction model.
  • Image feature information Input the point cloud feature information and current frame image data corresponding to each perspective into the corresponding feature extraction model, and extract the spatial feature information corresponding to each perspective and the current frame image data in parallel through the corresponding feature extraction model.
  • the computer device converts the point cloud feature information corresponding to each view angle and the current frame image data to obtain the point cloud feature vector corresponding to each view angle and the image matrix corresponding to the current frame image data.
  • a plurality of feature extraction models are pre-stored in the computer equipment.
  • the multiple feature extraction models may be the same type of feature extraction models.
  • the feature extraction model is obtained by training a large amount of sample data.
  • the feature extraction model may be a 2D convolutional neural network model.
  • the computer device inputs the point cloud feature vector corresponding to each perspective and the image matrix corresponding to the current frame data into the corresponding feature extraction model, and performs parallel feature extraction through the feature extraction model to obtain the spatial feature information corresponding to each perspective and the current frame data. Image feature information corresponding to the frame image data.
  • the feature extraction model can include a pooling layer, and the computer device can perform dimensionality reduction processing on the point cloud feature information corresponding to each perspective according to the first resolution through the pooling layer of the corresponding feature extraction model, and then obtain the corresponding point cloud feature information for each perspective.
  • Spatial feature information The pooling layer of the corresponding feature extraction model performs dimensionality reduction processing on the current frame of image data according to the second resolution, and then obtains the image feature information corresponding to the current frame of image data.
  • the spatial feature information may include information such as the shape of the obstacle.
  • the image feature information may include information such as the shape and color of the obstacle.
  • Step 210 Fusion of spatial feature information and image feature information corresponding to multiple viewing angles to obtain fused feature information.
  • Step 212 Input the fused feature information into the trained detection model, and perform a prediction operation on the fused feature information through the detection model, and output an obstacle detection result.
  • the computer device can merge the spatial feature information and image feature information corresponding to multiple viewing angles.
  • the way of fusion may be to first stitch the spatial feature information and image feature information corresponding to multiple viewing angles according to preset parameters, and then align the stitched feature information to the preset viewing angles to obtain the fused feature information.
  • the computer equipment converts the fused feature information to obtain the fused feature vector.
  • the trained detection model is pre-stored in the computer equipment.
  • the detection model is obtained through training with a large amount of sample data.
  • the detection model may be a 2D convolutional neural network.
  • the detection model includes multiple network layers, for example, it may include an input layer, an attention layer, a convolutional layer, a pooling layer, a fully connected layer, and so on.
  • the computer device inputs the fused feature vector into the detection model, calculates the context vector and weight corresponding to the fused feature vector through the attention layer of the detection model, and generates a first extraction result according to the context feature and weight.
  • the convolutional layer extracts the context feature corresponding to the context vector according to the first extraction result to generate the second extraction result.
  • the second extraction result is reduced in dimensionality through the pooling layer of the detection model.
  • the second extraction result after dimensionality reduction is classified by the fully connected layer, and the classification result can be obtained.
  • the classification results are weighted and output through the output layer.
  • the computer equipment obtains the obstacle detection result according to the classification result outputted by the weighting.
  • the computer device obtains the current frame point cloud data and the current frame image data within a preset angle range, and projects the current frame point cloud data on multiple viewing angles to obtain two-dimensional planes corresponding to the multiple viewing angles. It is conducive to the subsequent fusion of the point cloud data of the current frame and the image data of the current frame.
  • the computer device performs feature extraction on the two-dimensional plane corresponding to each perspective, obtains the point cloud feature information corresponding to each perspective, and inputs the point cloud feature information corresponding to each perspective and the current frame image data into the corresponding feature extraction model.
  • the spatial feature information corresponding to each view angle and the image feature information corresponding to the current frame image data are extracted in parallel through the corresponding feature extraction model.
  • the computer equipment By performing multiple feature extraction on the two-dimensional plane corresponding to each viewing angle, it is possible to extract more comprehensive and effective feature information from the point cloud data of the current frame.
  • the computer equipment fuses the spatial feature information and the image feature information corresponding to multiple viewing angles to obtain the fused feature information. Based on the data characteristics of multiple source data, multiple source data can be complemented to obtain more comprehensive obstacle feature information.
  • the computer equipment predicts and calculates the fused feature information through the detection model, and outputs the obstacle detection result. Since the fused feature information is comprehensive, and the detection model is pre-trained, the accuracy of obstacle detection is effectively improved.
  • the steps of fusing the spatial feature information and the image feature information corresponding to multiple viewing angles to obtain the fused feature information include:
  • step 302 the spatial feature information and the image feature information corresponding to the multiple viewing angles are spliced according to preset parameters to obtain spliced feature information.
  • Step 304 align the spliced feature information to a preset viewing angle according to the preset parameters to obtain the aligned feature information, and use the aligned feature information as the fused feature information.
  • the computer device can perform dimensionality reduction processing on the point cloud feature information corresponding to each perspective according to the first resolution through the pooling layer of the corresponding feature extraction model, and obtain the spatial feature information after the dimensionality reduction processing, that is, the space corresponding to multiple perspectives. In the process of feature information.
  • the computer device can perform dimensionality reduction processing on the current frame image data according to the second resolution through the pooling layer of the corresponding feature extraction model to obtain the image feature information after the dimensionality reduction processing, that is, the image feature information corresponding to the current frame image data.
  • the preset parameter may be the coordinate conversion relationship between the point cloud data and the image data.
  • the computer device splices the spatial feature information corresponding to the bird's-eye view angle and the spatial feature information corresponding to the front view angle with the image feature information respectively according to preset parameters. After the computer device obtains the spliced feature information, it can align the spliced feature information to a preset viewing angle according to preset parameters.
  • the preset viewing angle may be a bird's-eye view.
  • the computer device then obtains the aligned feature information on the preset viewing angle, and uses the aligned feature information as the fused feature information.
  • the computer device stitches the spatial feature information and image feature information corresponding to the multiple viewing angles according to preset parameters, and then aligns the stitched feature information to the preset viewing angles according to the preset parameters to obtain the aligned Feature information, using the aligned feature information as the fused feature information.
  • the spatial feature information corresponding to multiple viewing angles can improve the accurate 3D information, the lack of color information, and the image feature information includes higher-resolution color information, lacks 3D information, by splicing and aligning the spatial feature information and the image feature information , To achieve the fusion of complementary data, so as to perform obstacle detection based on the fused feature information, which can further improve the accuracy of obstacle detection.
  • the two-dimensional plane includes the two-dimensional data corresponding to each point in the point cloud data of the current frame
  • performing feature extraction on the two-dimensional plane corresponding to each perspective to obtain point cloud feature information includes: Extract multiple data dimensions from the two-dimensional data corresponding to each point in the point cloud data of the current frame; input multiple data dimensions into the trained neural network model, and perform prediction operations on the feature information of multiple dimensions through the neural network model , Get point cloud feature information.
  • the computer device can extract multiple data dimensions from the two-dimensional data corresponding to each point in the point cloud data of the current frame. Multiple data dimensions may include the coordinates of points, reflectivity and other dimensions.
  • the trained neural network model is pre-stored in the computer equipment. The trained neural network model is obtained by training with a large amount of sample data.
  • the neural network model can be a pointnet based on the attention layer.
  • the neural network model can include multiple network layers.
  • the network layer may include an attention layer, a convolutional layer, and so on.
  • the computer device can input the extracted multiple data dimensions into the trained neural network model, and calculate the context vectors and weights corresponding to the multiple data dimensions through the attention layer of the neural network model.
  • the neural network model takes the context vector and weight as the input of the convolutional layer, and extracts the context features corresponding to the context vector through the convolutional layer.
  • the neural network model takes the context features and weights as the input of the pooling layer, and reduces the dimensionality of the context features through the pooling layer.
  • the output layer of the neural network model outputs the context features and weights after dimensionality reduction, and uses the context features after dimensionality reduction as point cloud feature information.
  • the computer device extracts multiple data dimensions from the two-dimensional data corresponding to each point in the point cloud data of the current frame, and performs prediction operations on the multiple data dimensions through a neural network model to obtain point cloud feature information. Since the neural network is pre-trained, the local feature information of each point in the current frame of point cloud data can be accurately extracted through the neural network model, which is beneficial to the subsequent extraction of spatial feature information of the current frame of point cloud data.
  • projecting the point cloud data of the current frame on multiple viewing angles to obtain a two-dimensional plane corresponding to the multiple viewing angles includes: projecting the point cloud data of the current frame on a bird's-eye view angle to obtain the corresponding bird's-eye view angle.
  • the two-dimensional plane of the current frame project the point cloud data of the current frame on the orthographic perspective to obtain the two-dimensional plane corresponding to the orthographic perspective.
  • the computer device can project the point cloud data of the current frame to multiple perspectives.
  • the coordinates of the point cloud data of the current frame can be expressed as (x, y, z).
  • Computer equipment can project with preset resolutions. Multiple viewing angles may include a bird's-eye view and a front view. For example, in the process of bird's-eye perspective projection, when the preset resolution is 0.1m per grid, then the point cloud data of the current frame within the range of -60 ⁇ x ⁇ 60 and -60 ⁇ y ⁇ 60 can be set. Projected into a two-dimensional plane with a size of 1200x 1200.
  • the point cloud data of the current frame within the range of 0.05m will all fall on the corresponding grid.
  • the computer device can obtain an unobstructed and intuitive obstruction view, avoiding the problem of inaccurate point cloud feature information extracted by obstructing objects.
  • the computer device can describe the shape of a smaller target more intuitively, for example, can describe pedestrians more intuitively. This is beneficial for the computer equipment to extract more comprehensive and accurate effective feature information from the two-dimensional planes corresponding to multiple viewing angles.
  • the detection model includes a plurality of network layers, and performing prediction operations on the fused feature information through the detection model, and outputting obstacle detection results, includes: inputting the fused feature information into the input layer of the detection model;
  • the fused feature information is input to the attention layer of the detection model through the input layer, and the context vector and weight corresponding to the fused feature information are calculated through the attention layer to generate the first extraction result;
  • the first extraction result is input to the convolutional layer , Extract the context feature corresponding to the context vector through the convolutional layer to generate the second extraction result; input the second extraction result into the pooling layer, and perform dimensionality reduction processing on the second extraction result through the pooling layer;
  • Second, the extraction results are input to the fully connected layer, and the second extraction results after the dimensionality reduction process are classified through the fully connected layer to obtain the classification results, and the classification results are weighted through the output layer and output; the weighted classification results are selected from the weighted output
  • the classification result is used as the obstacle detection result.
  • the trained detection model is pre-stored in the computer equipment.
  • the detection model may be a detection model obtained after pre-training with a large amount of sample data.
  • the detection model may be a 2D convolutional neural network based on the attention layer.
  • the detection model can include multiple network layers.
  • the detection model may include multiple network layers such as an input layer, an attention layer, a convolutional layer, a pooling layer, a fully connected layer, and an output layer.
  • the computer device inputs the fused feature information to the input layer of the detection model by calling the trained detection model.
  • the fused feature information is transmitted to the attention layer through the input layer, and the context vector and weight corresponding to the fused feature information are calculated through the attention layer, and the first extraction result is generated according to the context vector and weight.
  • the detection model uses the first extraction result as the input of the convolution layer, extracts the context feature corresponding to the context vector through the convolution layer, and generates the second extraction result according to the context feature and weight. Furthermore, the computer device uses the second extraction result as the input of the pooling layer, and performs dimensionality reduction processing on the second extraction result through the pooling layer to obtain the second extraction result after the dimensionality reduction processing. The computer device uses the second extraction result after the dimensionality reduction processing as the input of the fully connected layer, and classifies the second extraction result after the dimensionality reduction processing to obtain the classification result.
  • the classification result can include multiple categories of obstacles, multiple location information, and so on. Furthermore, the classification results are weighted and output through the output layer. Furthermore, the computer device selects the classification result with the largest weight among the weighted classification results, which is the obstacle detection result. Obstacle detection results may include the location information of the obstacle, the size of the obstacle, the shape of the obstacle, and so on.
  • the computer device calculates the context vector and weight corresponding to the fused feature information through the attention layer of the detection model, and generates the first extraction result. It can filter the interference information in the fused feature information, and realize the feature focus processing on the fused feature information.
  • the context feature corresponding to the context vector is extracted through the convolutional layer to generate the second extraction result, and the second extraction result is reduced by the pooling layer, which can extract the main context features and avoid the influence of redundant features.
  • the computer device classifies the second extraction result after dimensionality reduction to obtain the classification result, and then weights the classification result and outputs it.
  • the classification result with the largest weight among the weighted classification results is selected as the obstacle detection result, which can classify
  • the results are normalized to further improve the accuracy of obstacle detection.
  • the two-dimensional plane includes multiple pixels, and each pixel corresponds to the two-dimensional data of multiple points in the point cloud data of the current frame.
  • each view angle Before feature extraction is performed on the two-dimensional plane corresponding to each view angle, it also includes : Perform average processing on the two-dimensional data of multiple points corresponding to each pixel to obtain the average value; perform normalization processing on the points corresponding to the corresponding pixels according to the average value.
  • the computer device may also perform normalization processing on multiple points in the current frame point cloud data in the two-dimensional plane.
  • the two-dimensional plane includes multiple pixels, and each pixel may be represented by a grid, and each grid includes multiple points in the point cloud data of the current frame.
  • the computer equipment averages the coordinates of multiple points in each grid to obtain the average value. Furthermore, the computer equipment makes the difference between the coordinates of each point in the grid and the average value to realize the normalization of the point cloud data of the current frame in each grid.
  • the feature extraction is performed on the two-dimensional plane corresponding to each perspective to obtain the point cloud feature information corresponding to each perspective
  • the method further includes: invoking multiple threads to concurrently extract each perspective in the two-dimensional plane corresponding to each perspective.
  • Point cloud feature information corresponding to each perspective before inputting the point cloud feature information corresponding to each perspective and the current frame image data into the corresponding feature extraction model, it also includes: using multi-threaded point cloud features corresponding to each perspective
  • the information and the current frame image data are converted in parallel to obtain the point cloud feature vector corresponding to each view angle and the image matrix corresponding to the current frame image data.
  • the computer device uses multiple threads to concurrently extract the point cloud feature information corresponding to each perspective in the two-dimensional plane corresponding to each perspective through multiple threads. Thereby improving the extraction efficiency of point cloud feature information.
  • the computer device can also use the multi-thread to perform the point cloud feature information corresponding to each perspective and the current frame image data before inputting the point cloud feature information corresponding to each perspective and the current frame image data into the corresponding feature extraction model. Parallel conversion effectively reduces the time-consuming feature extraction model for feature extraction.
  • the computer device may also obtain historical trajectory information of multiple obstacles in the current environment according to the obstacle detection result, and at the same time obtain the current position information of the vehicle. Predict the trajectory of multiple obstacles within a preset time period based on historical trajectory information and current position information.
  • the computer device tracks the movement process of the obstacle in the obstacle detection result, predicts the position information at the current time based on the position information of the obstacle at the previous time, and compares the predicted position information at the current time with the actual position. Information is compared to obtain error information. The computer device corrects the position information at the next moment according to the error information, thereby obtaining historical trajectory information of multiple obstacles.
  • the computer equipment can obtain the current position information sent by the vehicle-mounted locator. Therefore, the computer device can render the acquired historical trajectory information of multiple obstacles into a feature map to obtain a trajectory rendering map.
  • the historical trajectory information may be the trajectory of each frame of the history of multiple obstacles.
  • the computer device renders the historical trajectory information of multiple obstacles in the current frame to obtain a trajectory rendering map.
  • the color of obstacles in each frame in the trajectory rendering diagram changes with the distance from the current frame. The farther away from the current frame, the lighter the color of the obstacle.
  • the obstacle itself and the surrounding environment information can be obtained, and the influence factors of the trajectory can be considered from various aspects, which is more conducive to improving the accuracy of trajectory prediction.
  • the computer equipment obtains the current position information collected by the vehicle-mounted locator.
  • the current location information may be the location information of the vehicle on the high-precision map at the current moment.
  • the current location information can be expressed in the form of latitude and longitude.
  • the computer equipment extracts map elements from the current location information. Map elements can include information such as lane lines, center lines, sidewalks, and stop lines.
  • the computer device may render the extracted map elements according to multiple channel dimensions, and render the map elements into a map element rendering map corresponding to the channel dimensions. When the map elements are different, the channel dimensions corresponding to the map elements can also be different.
  • Channel dimensions can include color channels, element channels, and so on.
  • the color channel can include three channels of red, green, and blue.
  • Elemental passages can include lane-line passages, center-line passages, and sidewalk passages. The current position of the obstacle can be rendered intuitively and accurately through the channel dimension corresponding to the map element, which is conducive to subsequent trajectory prediction.
  • the trajectory rendering image and the map element rendering image can be stitched together.
  • the computer device determines the corresponding channel dimensions of the trajectory rendering map and the map element rendering map, and performs image stitching on the trajectory rendering map and the map element rendering map in the corresponding channel dimensions to obtain a spliced image matrix.
  • the spliced image matrix may be a complete image including the trajectory rendering map and the map element rendering map.
  • the computer device has pre-trained a feature extractor before acquiring the historical trajectory information and current position information of multiple obstacles in the current environment.
  • the computer device calls the trained feature extractor, and inputs the spliced image matrix into the trained feature extractor.
  • the computer device extracts the image feature information and context feature information corresponding to the spliced image matrix through the feature extractor, and then outputs the feature extraction result corresponding to the spliced image matrix through the fully connected layer of the feature extractor. It realizes the combination of various influence factors of the obstacle trajectory, and further improves the comprehensiveness of the feature extraction results.
  • the computer equipment can calculate the feature extraction results by means of regression prediction to obtain the trajectories of multiple obstacles within a preset time period. Because the obstacle detection result is more comprehensive and accurate, and the feature extraction result includes the trajectory of multiple obstacles in the history frame, the scope of environmental information is expanded, and the trajectory prediction based on various influencing factors is realized, thereby effectively providing the trajectory The accuracy of the forecast.
  • an obstacle detection device based on unmanned driving technology including: an acquisition module 402, a projection module 404, a first extraction module 406, a second extraction module 408, The fusion module 410 and the prediction module 412, where:
  • the acquiring module 402 is used to acquire the point cloud data of the current frame and the image data of the current frame within a preset angle range.
  • the projection module 404 is configured to project the point cloud data of the current frame on multiple viewing angles to obtain two-dimensional planes corresponding to the multiple viewing angles.
  • the first extraction module 406 is configured to perform feature extraction on the two-dimensional plane corresponding to each view angle to obtain point cloud feature information corresponding to each view angle.
  • the second extraction module 408 is used to input the point cloud feature information and current frame image data corresponding to each perspective into the corresponding feature extraction model, and extract the spatial feature information corresponding to each perspective and the current frame through the corresponding feature extraction model in parallel. Image feature information corresponding to the frame image data.
  • the fusion module 410 is used for fusing spatial feature information and image feature information corresponding to multiple viewing angles to obtain fused feature information.
  • the prediction module 412 is configured to input the fused feature information into the trained detection model, perform prediction operations on the fused feature information through the detection model, and output obstacle detection results.
  • the fusion module 410 is further configured to splice the spatial characteristic information and image characteristic information corresponding to the multiple viewing angles according to preset parameters to obtain the spliced characteristic information; according to the preset parameters, the spliced characteristic information The alignment is performed to the preset viewing angle to obtain the aligned feature information, and the aligned feature information is used as the fused feature information.
  • the first extraction module 406 is also used to extract multiple data dimensions from the two-dimensional data corresponding to each point in the point cloud data of the current frame; input the multiple data dimensions to the trained neural network model
  • the neural network model is used to perform predictive operations on multiple data dimensions to obtain point cloud feature information.
  • the projection module 404 is also used to project the point cloud data of the current frame on the bird's-eye view angle to obtain a two-dimensional plane corresponding to the bird's-eye view angle; Two-dimensional plane corresponding to the viewing angle
  • the prediction module 412 is also used to input the fused feature information to the input layer of the detection model; the fused feature information is input to the attention layer of the detection model through the input layer, and the attention layer is used to calculate The context vector and weight corresponding to the fused feature information are generated to generate the first extraction result; the first extraction result is input to the convolutional layer, and the context feature corresponding to the context vector is extracted through the convolutional layer to generate the second extraction result; the second extraction The result is input to the pooling layer, and the second extraction result is reduced by the pooling layer; the second extraction result after the dimensionality reduction is input into the fully connected layer, and the second extraction result after the dimensionality reduction is processed through the fully connected layer
  • the classification results are obtained by classification, and the classification results are weighted and output through the output layer; the classification result with the largest weight among the weighted classification results is selected as the obstacle detection result.
  • the above-mentioned device further includes: a normalization processing module for averaging the two-dimensional data of multiple points corresponding to each pixel to obtain an average value; according to the average value, the points corresponding to the corresponding pixels are processed Standardized processing.
  • the first extraction module 406 is also used to call multiple threads to concurrently extract the point cloud feature information corresponding to each perspective in the two-dimensional plane corresponding to each perspective; the above-mentioned device further includes: a conversion module for The point cloud feature information corresponding to each view angle and the current frame image data are converted in parallel by using multiple threads to obtain the point cloud feature vector corresponding to each view angle and the image matrix corresponding to the current frame image data.
  • the various modules in the above-mentioned obstacle detection device based on unmanned driving technology can be implemented in whole or in part by software, hardware, and a combination thereof.
  • the above-mentioned modules may be embedded in the form of hardware or independent of the processor in the computer equipment, or may be stored in the memory of the computer equipment in the form of software, so that the processor can call and execute the operations corresponding to the above-mentioned modules.
  • a computer device is provided, and its internal structure diagram may be as shown in FIG. 5.
  • the computer equipment includes a processor, a memory, a communication interface and a database connected through a system bus.
  • the processor of the computer device is used to provide calculation and control capabilities.
  • the memory of the computer device includes a non-volatile storage medium and an internal memory.
  • the non-volatile storage medium stores an operating system, computer readable instructions, and a database.
  • the internal memory provides an environment for the operation of the operating system and computer-readable instructions in the non-volatile storage medium.
  • the database of the computer equipment is used to store obstacle detection results.
  • the communication interface of the computer device is used to connect and communicate with the first vehicle-mounted sensor, the second vehicle-mounted sensor, and the vehicle-mounted positioning sensor.
  • the computer readable instruction is executed by the processor to realize an obstacle detection method.
  • FIG. 5 is only a block diagram of a part of the structure related to the solution of the present application, and does not constitute a limitation on the computer device to which the solution of the present application is applied.
  • the specific computer device may Including more or fewer parts than shown in the figure, or combining some parts, or having a different arrangement of parts.
  • a computer device that includes a memory and one or more processors.
  • the memory stores computer-readable instructions.
  • the one or more processors execute each of the foregoing method implementations. The steps in the example.
  • One or more non-volatile computer-readable storage media storing computer-readable instructions.
  • the computer-readable instructions are executed by one or more processors, the one or more processors execute the steps in each of the foregoing method embodiments. step.
  • Non-volatile memory may include read only memory (ROM), programmable ROM (PROM), electrically programmable ROM (EPROM), electrically erasable programmable ROM (EEPROM), or flash memory.
  • Volatile memory may include random access memory (RAM) or external cache memory.
  • RAM is available in many forms, such as static RAM (SRAM), dynamic RAM (DRAM), synchronous DRAM (SDRAM), double data rate SDRAM (DDRSDRAM), enhanced SDRAM (ESDRAM), synchronous chain Channel (Synchlink) DRAM (SLDRAM), memory bus (Rambus) direct RAM (RDRAM), direct memory bus dynamic RAM (DRDRAM), and memory bus dynamic RAM (RDRAM), etc.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • Image Analysis (AREA)

Abstract

An obstacle detection method based on driverless technology, comprising: obtaining current frame point cloud data and current frame image data within a preset angle range (202); projecting the current frame point cloud data from a plurality of viewing angles to obtain two-dimensional planes corresponding to the plurality of viewing angles (204); performing feature extraction on the two-dimensional plane corresponding to each viewing angle to obtain point cloud feature information corresponding to each viewing angle (206); inputting the point cloud feature information corresponding to each viewing angle and the current frame image data into a corresponding feature extraction model, and extracting spatial feature information corresponding to each viewing angle and image feature information corresponding to the current frame image data in parallel by means of the corresponding feature extraction model (208); fusing the spatial feature information corresponding to the plurality of viewing angles and the image feature information to obtain fused feature information (210); and inputting the fused feature information into a trained detection model, performing prediction calculation on the fused feature information by means of the detection model, and outputting an obstacle detection result (212).

Description

基于无人驾驶技术的障碍物检测方法、装置和计算机设备Obstacle detection method, device and computer equipment based on unmanned driving technology 技术领域Technical field
本申请涉及一种基于无人驾驶技术的障碍物检测方法、装置、计算机设备和存储介质。This application relates to an obstacle detection method, device, computer equipment and storage medium based on unmanned driving technology.
背景技术Background technique
人工智能技术的发展,促进了无人驾驶技术的发展。在无人驾驶过程中,需要实时检测周围环境中的障碍物。例如,行人、车辆等交通参与者。通过对检测到的障碍物进行跟踪预测,得到障碍物轨迹,能够更好地规划合理路线、躲避障碍物以及遵守交通规则。The development of artificial intelligence technology has promoted the development of unmanned driving technology. In the process of unmanned driving, it is necessary to detect obstacles in the surrounding environment in real time. For example, pedestrians, vehicles and other traffic participants. By tracking and predicting the detected obstacles, the obstacle trajectory is obtained, which can better plan a reasonable route, avoid obstacles, and abide by traffic rules.
传统方式中,是将点云数据投影至图像数据中,获取多个通道的特征信息,进而根据多个通道的特征信息进行障碍物检测。然而在将点云数据投影至图像数据进行特征提取的过程中,可能会丢失部分信息,导致提取出的每种源数据的有效特征信息不够全面,进而导致障碍物检测的准确性较低。In the traditional way, the point cloud data is projected into the image data to obtain feature information of multiple channels, and then obstacle detection is performed based on the feature information of multiple channels. However, in the process of projecting the point cloud data to the image data for feature extraction, some information may be lost, resulting in the extraction of effective feature information of each source data is not comprehensive enough, resulting in low accuracy of obstacle detection.
发明内容Summary of the invention
根据本申请公开的各种实施例,提供一种能够提高无人驾驶过程中障碍物检测准确性的基于无人驾驶技术的障碍物检测方法、装置、计算机设备和存储介质。According to various embodiments disclosed in the present application, an obstacle detection method, device, computer device, and storage medium based on an unmanned driving technology that can improve the accuracy of obstacle detection in an unmanned driving process are provided.
一种基于无人驾驶技术的障碍物检测方法,包括:An obstacle detection method based on unmanned driving technology, including:
获取当前帧点云数据和预设角度范围内的当前帧图像数据;Obtain current frame point cloud data and current frame image data within a preset angle range;
将所述当前帧点云数据在多个视角上进行投影,得到多个视角对应的二维平面;Projecting the point cloud data of the current frame on multiple viewing angles to obtain two-dimensional planes corresponding to the multiple viewing angles;
对每个视角对应的二维平面进行特征提取,得到每个视角对应的点云特征信息;Perform feature extraction on the two-dimensional plane corresponding to each perspective, and obtain the point cloud feature information corresponding to each perspective;
将每个视角对应的点云特征信息和所述当前帧图像数据输入至对应的特征提取模型中,通过相应的特征提取模型并行提取每个视角对应的空间特征信息和所述当前帧图像数据对应的图像特征信息;The point cloud feature information corresponding to each perspective and the current frame image data are input into the corresponding feature extraction model, and the spatial feature information corresponding to each perspective and the current frame image data are extracted in parallel through the corresponding feature extraction model Image feature information;
将多个视角对应的空间特征信息和所述图像特征信息进行融合,得到融合后的特征信息;及Fusing the spatial feature information corresponding to multiple viewing angles with the image feature information to obtain the fused feature information; and
将所述融合后的特征信息输入至训练后的检测模型中,通过所述检测模型对所述融合后的特征信息进行预测运算,输出障碍物检测结果。The fused feature information is input into a trained detection model, and the fused feature information is predicted and calculated through the detection model, and an obstacle detection result is output.
一种基于无人驾驶技术的障碍物检测装置,包括:An obstacle detection device based on unmanned driving technology, including:
获取模块,用于获取当前帧点云数据和预设角度范围内的当前帧图像数据;The acquisition module is used to acquire the current frame point cloud data and the current frame image data within a preset angle range;
投影模块,用于将所述当前帧点云数据在多个视角上进行投影,得到多个视角对应的二维平面;A projection module, configured to project the point cloud data of the current frame on multiple viewing angles to obtain two-dimensional planes corresponding to the multiple viewing angles;
第一提取模块,用于对每个视角对应的二维平面进行特征提取,得到每个视角对应的点云特征信息;The first extraction module is used to perform feature extraction on the two-dimensional plane corresponding to each perspective to obtain point cloud feature information corresponding to each perspective;
第二提取模块,用于将每个视角对应的点云特征信息和所述当前帧图像数据输入至对应的特征提取模型中,通过相应的特征提取模型并行提取每个视角对应的空间特征信息和所述当前帧图像数据对应的图像特征信息;The second extraction module is used to input the point cloud feature information corresponding to each perspective and the current frame image data into the corresponding feature extraction model, and extract the spatial feature information corresponding to each perspective in parallel through the corresponding feature extraction model. Image feature information corresponding to the current frame of image data;
融合模块,用于将多个视角对应的空间特征信息和所述图像特征信息进行融合,得到融合后的特征信息;及The fusion module is used to fuse the spatial feature information corresponding to multiple perspectives with the image feature information to obtain the fused feature information; and
预测模块,用于将所述融合后的特征信息输入至训练后的检测模型中,通 过所述检测模型对所述融合后的特征信息进行预测运算,输出障碍物检测结果。The prediction module is used to input the fused feature information into the trained detection model, and perform prediction operations on the fused feature information through the detection model, and output obstacle detection results.
一种计算机设备,包括存储器和一个或多个处理器,所述存储器中储存有计算机可读指令,所述计算机可读指令被所述处理器执行时,使得所述一个或多个处理器执行以下步骤:A computer device, including a memory and one or more processors, the memory stores computer readable instructions, and when the computer readable instructions are executed by the processor, the one or more processors execute The following steps:
获取当前帧点云数据和预设角度范围内的当前帧图像数据;Obtain current frame point cloud data and current frame image data within a preset angle range;
将所述当前帧点云数据在多个视角上进行投影,得到多个视角对应的二维平面;Projecting the point cloud data of the current frame on multiple viewing angles to obtain two-dimensional planes corresponding to the multiple viewing angles;
对每个视角对应的二维平面进行特征提取,得到每个视角对应的点云特征信息;Perform feature extraction on the two-dimensional plane corresponding to each perspective, and obtain the point cloud feature information corresponding to each perspective;
将每个视角对应的点云特征信息和所述当前帧图像数据输入至对应的特征提取模型中,通过相应的特征提取模型并行提取每个视角对应的空间特征信息和所述当前帧图像数据对应的图像特征信息;The point cloud feature information corresponding to each perspective and the current frame image data are input into the corresponding feature extraction model, and the spatial feature information corresponding to each perspective and the current frame image data are extracted in parallel through the corresponding feature extraction model Image feature information;
将多个视角对应的空间特征信息和所述图像特征信息进行融合,得到融合后的特征信息;及Fusing the spatial feature information corresponding to multiple viewing angles with the image feature information to obtain the fused feature information; and
将所述融合后的特征信息输入至训练后的检测模型中,通过所述检测模型对所述融合后的特征信息进行预测运算,输出障碍物检测结果。The fused feature information is input into a trained detection model, and the fused feature information is predicted and calculated through the detection model, and an obstacle detection result is output.
一个或多个存储有计算机可读指令的非易失性计算机可读存储介质,计算机可读指令被一个或多个处理器执行时,使得一个或多个处理器执行以下步骤:One or more non-volatile computer-readable storage media storing computer-readable instructions. When the computer-readable instructions are executed by one or more processors, the one or more processors perform the following steps:
获取当前帧点云数据和预设角度范围内的当前帧图像数据;Obtain current frame point cloud data and current frame image data within a preset angle range;
将所述当前帧点云数据在多个视角上进行投影,得到多个视角对应的二维平面;Projecting the point cloud data of the current frame on multiple viewing angles to obtain two-dimensional planes corresponding to the multiple viewing angles;
对每个视角对应的二维平面进行特征提取,得到每个视角对应的点云特征信息;Perform feature extraction on the two-dimensional plane corresponding to each perspective, and obtain the point cloud feature information corresponding to each perspective;
将每个视角对应的点云特征信息和所述当前帧图像数据输入至对应的特征提取模型中,通过相应的特征提取模型并行提取每个视角对应的空间特征信息和所述当前帧图像数据对应的图像特征信息;Input the point cloud feature information corresponding to each perspective and the current frame image data into the corresponding feature extraction model, and extract the spatial feature information corresponding to each perspective and the current frame image data in parallel through the corresponding feature extraction model Image feature information;
将多个视角对应的空间特征信息和所述图像特征信息进行融合,得到融合后的特征信息;及Fusing the spatial feature information corresponding to multiple viewing angles with the image feature information to obtain the fused feature information; and
将所述融合后的特征信息输入至训练后的检测模型中,通过所述检测模型对所述融合后的特征信息进行预测运算,输出障碍物检测结果。The fused feature information is input into a trained detection model, and the fused feature information is predicted and calculated through the detection model, and an obstacle detection result is output.
本申请的一个或多个实施例的细节在下面的附图和描述中提出。本申请的其它特征和优点将从说明书、附图以及权利要求书变得明显。The details of one or more embodiments of the present application are set forth in the following drawings and description. Other features and advantages of this application will become apparent from the description, drawings and claims.
附图说明Description of the drawings
为了更清楚地说明本申请实施例中的技术方案,下面将对实施例中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本申请的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其它的附图。In order to more clearly describe the technical solutions in the embodiments of the present application, the following will briefly introduce the drawings needed in the embodiments. Obviously, the drawings in the following description are only some embodiments of the present application. A person of ordinary skill in the art can obtain other drawings based on these drawings without creative work.
图1为一个或多个实施例中基于无人驾驶技术的障碍物检测方法的应用环境图。Fig. 1 is an application environment diagram of an obstacle detection method based on unmanned driving technology in one or more embodiments.
图2为一个或多个实施例中基于无人驾驶技术的障碍物检测方法的流程示意图。Fig. 2 is a schematic flowchart of an obstacle detection method based on unmanned driving technology in one or more embodiments.
图3为一个或多个实施例中将多个视角对应的空间特征信息和图像特征信息进行融合,得到融合后的特征信息步骤的流程示意图。FIG. 3 is a schematic flowchart of the step of fusing spatial feature information and image feature information corresponding to multiple viewing angles to obtain fused feature information in one or more embodiments.
图4为一个或多个实施例中基于无人驾驶技术的障碍物检测装置的框图。Fig. 4 is a block diagram of an obstacle detection device based on unmanned driving technology in one or more embodiments.
图5为一个或多个实施例中计算机设备的框图。Figure 5 is a block diagram of a computer device in one or more embodiments.
具体实施方式Detailed ways
为了使本申请的技术方案及优点更加清楚明白,以下结合附图及实施例,对本申请进行进一步详细说明。应当理解,此处描述的具体实施例仅仅用以解释本申请,并不用于限定本申请。In order to make the technical solutions and advantages of the present application clearer, the following further describes the present application in detail with reference to the accompanying drawings and embodiments. It should be understood that the specific embodiments described herein are only used to explain the present application, and are not used to limit the present application.
本申请提供的基于无人驾驶技术的障碍物检测方法,可以应用于如图1所示的无人驾驶过程中进行障碍物检测的示意图中。第一车载传感器102将采集到的当前帧点云数据发送至车载计算机设备104。例如,第一车载传感器可以是激光雷达。车载计算机设备可以称为计算机设备。第二车载传感器106将采集到的预设角度范围内的当前帧图像数据发送至计算机设备104。例如,第二车载传感器可以是车载摄像头。计算机设备104将当前帧点云数据在多个视角上进行投影,得到多个视角对应的二维平面。计算机设备104对每个视角对应的二维平面进行特征提取,得到每个视角对应的点云特征信息。计算机设备104将每个视角对应的点云特征信息和当前帧图像数据输入至对应的特征提取模型中,通过相应的特征提取模型并行提取每个视角对应的空间特征信息和当前帧图像数据对应的图像特征信息。计算机设备104将多个视角对应的空间特征信息和当前帧图像数据进行融合,得到融合后的特征信息。计算机设备104将融合后的特征信息输入至训练后的检测模型中,通过检测模型对融合后的特征信息进行预测运算,输出障碍物检测结果。The obstacle detection method based on the unmanned driving technology provided in this application can be applied to the schematic diagram of obstacle detection during the unmanned driving process as shown in FIG. 1. The first vehicle-mounted sensor 102 sends the collected point cloud data of the current frame to the vehicle-mounted computer device 104. For example, the first vehicle-mounted sensor may be a lidar. On-board computer equipment can be referred to as computer equipment. The second vehicle-mounted sensor 106 sends the collected image data of the current frame within the preset angle range to the computer device 104. For example, the second vehicle-mounted sensor may be a vehicle-mounted camera. The computer device 104 projects the point cloud data of the current frame on multiple viewing angles to obtain two-dimensional planes corresponding to the multiple viewing angles. The computer device 104 performs feature extraction on the two-dimensional plane corresponding to each view angle to obtain point cloud feature information corresponding to each view angle. The computer device 104 inputs the point cloud feature information corresponding to each perspective and the current frame image data into the corresponding feature extraction model, and extracts the spatial feature information corresponding to each perspective and the current frame image data in parallel through the corresponding feature extraction model. Image feature information. The computer device 104 fuses the spatial feature information corresponding to the multiple viewing angles with the current frame image data to obtain the fused feature information. The computer device 104 inputs the fused feature information into the trained detection model, and performs a prediction operation on the fused feature information through the detection model, and outputs an obstacle detection result.
在其中一个实施例中,如图2所示,提供了一种基于无人驾驶技术的障碍物检测方法,以该方法应用于图1中的计算机设备为例进行说明,包括以下步骤:In one of the embodiments, as shown in FIG. 2, an obstacle detection method based on unmanned driving technology is provided. Taking the method applied to the computer equipment in FIG. 1 as an example for description, the method includes the following steps:
步骤202,获取当前帧点云数据和预设角度范围内的当前帧图像数据。Step 202: Obtain current frame point cloud data and current frame image data within a preset angle range.
车辆在无人驾驶过程中,通过安装在车辆上的第一车载传感器将采集到的当前帧点云数据传送至计算机设备,通过安装在车辆上的第二车载传感器将采集到的预设角度范围内的当前帧图像数据发送至计算机设备。例如,第一车载传感器可以是激光雷达。当前帧点云数据是第一车载传感器采集到的360度范围内当前帧点云数据。例如,第二车载传感器可以是车载摄像头。预设角度范围内的当前帧图像数据可以是通过多个车载摄像头采集到的车辆周围360度范围内的当前帧图像数据。In the process of unmanned driving of the vehicle, the collected current frame point cloud data is transmitted to the computer device through the first on-board sensor installed on the vehicle, and the preset angle range collected by the second on-board sensor installed on the vehicle The image data of the current frame within is sent to the computer device. For example, the first vehicle-mounted sensor may be a lidar. The current frame point cloud data is the current frame point cloud data within a 360-degree range collected by the first vehicle-mounted sensor. For example, the second vehicle-mounted sensor may be a vehicle-mounted camera. The current frame image data within the preset angle range may be the current frame image data within a 360-degree range around the vehicle collected by multiple on-board cameras.
步骤204,将当前帧点云数据在多个视角上进行投影,得到多个视角对应的二维平面。Step 204: Project the point cloud data of the current frame on multiple viewing angles to obtain two-dimensional planes corresponding to the multiple viewing angles.
步骤206,对每个视角对应的二维平面进行特征提取,得到每个视角对应的点云特征信息。Step 206: Perform feature extraction on the two-dimensional plane corresponding to each view angle to obtain point cloud feature information corresponding to each view angle.
当前帧点云数据为3D点云数据。计算机设备通过将获取到的当前帧点云数据投影至多个视角,从而将3D点云数据投影至多个视角对应的二维平面中,实现将3D点云数据转换为二维平面中的二维数据。多个视角可以包括鸟瞰视角、正视视角。当计算机设备将当前点云数据在鸟瞰视角上进行投影时,可以得到鸟瞰视角对应的二维平面。当计算机设备将当前帧点云数据在正视视角上进行投影时,可以得到正视视角对应的二维平面。The point cloud data of the current frame is 3D point cloud data. The computer device projects the acquired point cloud data of the current frame to multiple viewing angles, thereby projecting the 3D point cloud data into the two-dimensional planes corresponding to the multiple viewing angles, and realizes the conversion of the 3D point cloud data into the two-dimensional data in the two-dimensional plane . Multiple viewing angles may include a bird's-eye view and a front view. When the computer device projects the current point cloud data on the bird's-eye view angle, a two-dimensional plane corresponding to the bird's-eye view angle can be obtained. When the computer device projects the point cloud data of the current frame on the orthographic perspective, a two-dimensional plane corresponding to the orthographic perspective can be obtained.
每个视角对应的二维平面中包括投影后的当前帧点云数据。计算机设备可以在每个视角对应的二维平面中提取每个视角对应的点云特征信息。点云特征信息可以是二维平面中每个像素对应的当前帧点云数据中每个点的局部特征信息,局部特征信息可以包括局部深度、点云密度等。计算机设备中预先存储有训练后的神经网络模型。例如,神经网络模型可以是基于注意力层的pointnet。计算机设备可以将每个视角对应的二维平面输入至训练后的神经网络模型中, 通过神经网络模型分别对每个视角对应的二维平面进行预测运算,得到每个视角对应的点云特征信息。The two-dimensional plane corresponding to each view includes the point cloud data of the current frame after projection. The computer device can extract the point cloud feature information corresponding to each perspective in the two-dimensional plane corresponding to each perspective. The point cloud feature information may be the local feature information of each point in the current frame point cloud data corresponding to each pixel in the two-dimensional plane, and the local feature information may include local depth, point cloud density, and the like. The trained neural network model is pre-stored in the computer equipment. For example, the neural network model can be a pointnet based on the attention layer. The computer device can input the two-dimensional plane corresponding to each perspective into the trained neural network model, and perform prediction operations on the two-dimensional plane corresponding to each perspective through the neural network model to obtain the point cloud feature information corresponding to each perspective .
步骤208,将每个视角对应的点云特征信息和当前帧图像数据输入至对应的特征提取模型中,通过相应的特征提取模型并行提取每个视角对应的空间特征信息和当前帧图像数据对应的图像特征信息。Step 208: Input the point cloud feature information and current frame image data corresponding to each perspective into the corresponding feature extraction model, and extract the spatial feature information corresponding to each perspective and the current frame image data in parallel through the corresponding feature extraction model. Image feature information.
计算机设备将每个视角对应的点云特征信息和当前帧图像数据进行转换,得到每个视角对应的点云特征向量和当前帧图像数据对应的图像矩阵。计算机设备中预先存储有多个特征提取模型。多个特征提取模型可以是相同类型的特征提取模型。特征提取模型是通过大量的样本数据训练得到的。例如,特征提取模型可以是2D卷积神经网络模型。计算机设备将每个视角对应的点云特征向量和当前帧数据对应的图像矩阵分别输入至对应的特征提取模型中,通过特征提取模型进行并行特征提取,得到每个视角对应的空间特征信息和当前帧图像数据对应的图像特征信息。特征提取模型中可以包括池化层,计算机设备可以通过相应的特征提取模型的池化层根据第一分辨率对每个视角对应的点云特征信息进行降维处理,进而得到每个视角对应的空间特征信息。通过相应的特征提取模型的池化层根据第二分辨率对当前帧图像数据进行降维处理,进而得到当前帧图像数据对应的图像特征信息。空间特征信息可以包括障碍物的形状等信息。图像特征信息可以包括障碍物的形状、颜色等信息。The computer device converts the point cloud feature information corresponding to each view angle and the current frame image data to obtain the point cloud feature vector corresponding to each view angle and the image matrix corresponding to the current frame image data. A plurality of feature extraction models are pre-stored in the computer equipment. The multiple feature extraction models may be the same type of feature extraction models. The feature extraction model is obtained by training a large amount of sample data. For example, the feature extraction model may be a 2D convolutional neural network model. The computer device inputs the point cloud feature vector corresponding to each perspective and the image matrix corresponding to the current frame data into the corresponding feature extraction model, and performs parallel feature extraction through the feature extraction model to obtain the spatial feature information corresponding to each perspective and the current frame data. Image feature information corresponding to the frame image data. The feature extraction model can include a pooling layer, and the computer device can perform dimensionality reduction processing on the point cloud feature information corresponding to each perspective according to the first resolution through the pooling layer of the corresponding feature extraction model, and then obtain the corresponding point cloud feature information for each perspective. Spatial feature information. The pooling layer of the corresponding feature extraction model performs dimensionality reduction processing on the current frame of image data according to the second resolution, and then obtains the image feature information corresponding to the current frame of image data. The spatial feature information may include information such as the shape of the obstacle. The image feature information may include information such as the shape and color of the obstacle.
步骤210,将多个视角对应的空间特征信息和图像特征信息进行融合,得到融合后的特征信息。Step 210: Fusion of spatial feature information and image feature information corresponding to multiple viewing angles to obtain fused feature information.
步骤212,将融合后的特征信息输入至训练后的检测模型中,通过检测模型对融合后的特征信息进行预测运算,输出障碍物检测结果。Step 212: Input the fused feature information into the trained detection model, and perform a prediction operation on the fused feature information through the detection model, and output an obstacle detection result.
计算机设备在得到多个视角对应的空间特征信息和图像特征信息后,可以 将多个视角对应的空间特征信息和图像特征信息进行融合。融合的方式可以是先根据预设参数将多个视角对应的空间特征信息和图像特征信息进行拼接,再将拼接后的特征信息对齐至预设视角上,从而得到融合后的特征信息。After obtaining the spatial feature information and image feature information corresponding to multiple viewing angles, the computer device can merge the spatial feature information and image feature information corresponding to multiple viewing angles. The way of fusion may be to first stitch the spatial feature information and image feature information corresponding to multiple viewing angles according to preset parameters, and then align the stitched feature information to the preset viewing angles to obtain the fused feature information.
计算机设备将融合后的特征信息进行转换,得到融合后的特征向量。计算机设备中预先存储有训练后的检测模型。检测模型是通过大量的样本数据训练得到的。例如,检测模型可以是2D卷积神经网络。检测模型中包括多个网络层,例如,可以包括输入层、注意力层、卷积层、池化层、全连接层等。计算机设备将融合后的特征向量输入至检测模型中,通过检测模型的注意力层计算融合后的特征向量对应的上下文向量与权重,根据上下文特征与权重生成第一提取结果。进而通过卷积层根据第一提取结果提取上下文向量对应的上下文特征,生成第二提取结果。通过检测模型的池化层对第二提取结果进行降维处理。通过全连接层对降维处理后的第二提取结果进行分类,可以得到分类结果。通过输出层将分类结果进行加权输出。计算机设备根据加权输出的分类结果得到障碍物检测结果。The computer equipment converts the fused feature information to obtain the fused feature vector. The trained detection model is pre-stored in the computer equipment. The detection model is obtained through training with a large amount of sample data. For example, the detection model may be a 2D convolutional neural network. The detection model includes multiple network layers, for example, it may include an input layer, an attention layer, a convolutional layer, a pooling layer, a fully connected layer, and so on. The computer device inputs the fused feature vector into the detection model, calculates the context vector and weight corresponding to the fused feature vector through the attention layer of the detection model, and generates a first extraction result according to the context feature and weight. Furthermore, the convolutional layer extracts the context feature corresponding to the context vector according to the first extraction result to generate the second extraction result. The second extraction result is reduced in dimensionality through the pooling layer of the detection model. The second extraction result after dimensionality reduction is classified by the fully connected layer, and the classification result can be obtained. The classification results are weighted and output through the output layer. The computer equipment obtains the obstacle detection result according to the classification result outputted by the weighting.
在本实施例中,计算机设备获取当前帧点云数据和预设角度范围内的当前帧图像数据,将当前帧点云数据在多个视角上进行投影,得到多个视角对应的二维平面。有利于后续将当前帧点云数据与当前帧图像数据进行融合。计算机设备对每个视角对应的二维平面进行特征提取,得到每个视角对应的点云特征信息,将每个视角对应的点云特征信息和当前帧图像数据输入至对应的特征提取模型中,通过相应的特征提取模型并行提取每个视角对应的空间特征信息和当前帧图像数据对应的图像特征信息。通过对每个视角对应的二维平面进行多次特征提取,能够提取出当前帧点云数据中更为全面的有效特征信息。计算机设备将多个视角对应的空间特征信息和图像特征信息进行融合,得到融合后的 特征信息。能够根据多种源数据的数据特性,将多种源数据之间进行互补,得到更为全面的障碍物特征信息。计算机设备通过通过检测模型对融合后的特征信息进行预测运算,输出障碍物检测结果。由于融合后的特征信息是全面的,且检测模型是预先训练的,从而有效提高了障碍物的检测准确性。In this embodiment, the computer device obtains the current frame point cloud data and the current frame image data within a preset angle range, and projects the current frame point cloud data on multiple viewing angles to obtain two-dimensional planes corresponding to the multiple viewing angles. It is conducive to the subsequent fusion of the point cloud data of the current frame and the image data of the current frame. The computer device performs feature extraction on the two-dimensional plane corresponding to each perspective, obtains the point cloud feature information corresponding to each perspective, and inputs the point cloud feature information corresponding to each perspective and the current frame image data into the corresponding feature extraction model. The spatial feature information corresponding to each view angle and the image feature information corresponding to the current frame image data are extracted in parallel through the corresponding feature extraction model. By performing multiple feature extraction on the two-dimensional plane corresponding to each viewing angle, it is possible to extract more comprehensive and effective feature information from the point cloud data of the current frame. The computer equipment fuses the spatial feature information and the image feature information corresponding to multiple viewing angles to obtain the fused feature information. Based on the data characteristics of multiple source data, multiple source data can be complemented to obtain more comprehensive obstacle feature information. The computer equipment predicts and calculates the fused feature information through the detection model, and outputs the obstacle detection result. Since the fused feature information is comprehensive, and the detection model is pre-trained, the accuracy of obstacle detection is effectively improved.
在其中一个实施例中,如图3所示,将多个视角对应的空间特征信息和图像特征信息进行融合,得到融合后的特征信息的步骤包括:In one of the embodiments, as shown in FIG. 3, the steps of fusing the spatial feature information and the image feature information corresponding to multiple viewing angles to obtain the fused feature information include:
步骤302,根据预设参数将多个视角对应的空间特征信息和图像特征信息进行拼接,得到拼接后的特征信息。In step 302, the spatial feature information and the image feature information corresponding to the multiple viewing angles are spliced according to preset parameters to obtain spliced feature information.
步骤304,根据预设参数将拼接后的特征信息进行对齐至预设视角,得到对齐后的特征信息,将对齐后的特征信息作为融合后的特征信息。Step 304: align the spliced feature information to a preset viewing angle according to the preset parameters to obtain the aligned feature information, and use the aligned feature information as the fused feature information.
计算机设备可以通过相应的特征提取模型的池化层根据第一分辨率对每个视角对应的点云特征信息进行降维处理,得到降维处理后的空间特征信息,即多个视角对应的空间特征信息的过程中。计算机设备可以通过相应的特征提取模型的池化层根据第二分辨率对当前帧图像数据进行降维处理,得到降维处理后的图像特征信息,即当前帧图像数据对应的图像特征信息。The computer device can perform dimensionality reduction processing on the point cloud feature information corresponding to each perspective according to the first resolution through the pooling layer of the corresponding feature extraction model, and obtain the spatial feature information after the dimensionality reduction processing, that is, the space corresponding to multiple perspectives. In the process of feature information. The computer device can perform dimensionality reduction processing on the current frame image data according to the second resolution through the pooling layer of the corresponding feature extraction model to obtain the image feature information after the dimensionality reduction processing, that is, the image feature information corresponding to the current frame image data.
预设参数可以是点云数据与图像数据之间的坐标转换关系。计算机设备根据预设参数将鸟瞰视角对应的空间特征信息、正视视角对应的空间特征信息分别和图像特征信息进行拼接。计算机设备得到拼接后的特征信息后,可根据预设参数将拼接后的特征信息对齐至预设视角。预设视角可以是鸟瞰视角。计算机设备进而得到预设视角上的对齐后的特征信息,将对齐后的特征信息作为融合后的特征信息。The preset parameter may be the coordinate conversion relationship between the point cloud data and the image data. The computer device splices the spatial feature information corresponding to the bird's-eye view angle and the spatial feature information corresponding to the front view angle with the image feature information respectively according to preset parameters. After the computer device obtains the spliced feature information, it can align the spliced feature information to a preset viewing angle according to preset parameters. The preset viewing angle may be a bird's-eye view. The computer device then obtains the aligned feature information on the preset viewing angle, and uses the aligned feature information as the fused feature information.
在本实施例中,计算机设备根据预设参数将多个视角对应的空间特征信息和图像特征信息进行拼接,进而根据预设参数将拼接后的特征信息进行对齐至 预设视角,得到对齐后的特征信息,将对齐后的特征信息作为融合后的特征信息。由于多个视角对应的空间特征信息可以提高准确的3D信息,缺少颜色信息,而图像特征信息包括分辨率更高的颜色信息,缺少3D信息,通过将空间特征信息和图像特征信息进行拼接与对齐,实现将互补的数据进行融合,从而根据融合后的特征信息进行障碍物检测,能够进一步提高障碍物的检测准确性。In this embodiment, the computer device stitches the spatial feature information and image feature information corresponding to the multiple viewing angles according to preset parameters, and then aligns the stitched feature information to the preset viewing angles according to the preset parameters to obtain the aligned Feature information, using the aligned feature information as the fused feature information. Because the spatial feature information corresponding to multiple viewing angles can improve the accurate 3D information, the lack of color information, and the image feature information includes higher-resolution color information, lacks 3D information, by splicing and aligning the spatial feature information and the image feature information , To achieve the fusion of complementary data, so as to perform obstacle detection based on the fused feature information, which can further improve the accuracy of obstacle detection.
在其中一个实施例中,二维平面包括所述当前帧点云数据中每个点对应的二维数据,对每个视角对应的二维平面进行特征提取,得到点云特征信息,包括:在当前帧点云数据中每个点对应的二维数据中提取多个数据维度;将多个数据维度输入至训练后的神经网络模型中,通过神经网络模型对多个维度的特征信息进行预测运算,得到点云特征信息。In one of the embodiments, the two-dimensional plane includes the two-dimensional data corresponding to each point in the point cloud data of the current frame, and performing feature extraction on the two-dimensional plane corresponding to each perspective to obtain point cloud feature information includes: Extract multiple data dimensions from the two-dimensional data corresponding to each point in the point cloud data of the current frame; input multiple data dimensions into the trained neural network model, and perform prediction operations on the feature information of multiple dimensions through the neural network model , Get point cloud feature information.
计算机设备可以在当前帧点云数据中每个点对应的二维数据中提取多个数据维度。多个数据维度可以包括点的坐标、反射率等维度。计算机设备中预先存储有训练后的神经网络模型。训练后的神经网络模型是通过大量样本数据训练得到的。例如,神经网络模型可以是基于注意力层的pointnet。神经网络模型中可以包括多个网络层。例如,网络层可以包括注意力层、卷积层等。计算机设备可以将提取出的多个数据维度输入至训练后的神经网络模型中,通过神经网络模型的注意力层计算多个数据维度对应的上下文向量与权重。神经网络模型将上下文向量与权重作为卷积层的输入,通过卷积层提取上下文向量对应的上下文特征。神经网络模型将上下文特征与权重作为池化层的输入,通过池化层对上下文特征进行降维处理。通过神经网络模型的输出层输出降维处理后的上下文特征与权重,将降维处理后的上下文特征作为点云特征信息。The computer device can extract multiple data dimensions from the two-dimensional data corresponding to each point in the point cloud data of the current frame. Multiple data dimensions may include the coordinates of points, reflectivity and other dimensions. The trained neural network model is pre-stored in the computer equipment. The trained neural network model is obtained by training with a large amount of sample data. For example, the neural network model can be a pointnet based on the attention layer. The neural network model can include multiple network layers. For example, the network layer may include an attention layer, a convolutional layer, and so on. The computer device can input the extracted multiple data dimensions into the trained neural network model, and calculate the context vectors and weights corresponding to the multiple data dimensions through the attention layer of the neural network model. The neural network model takes the context vector and weight as the input of the convolutional layer, and extracts the context features corresponding to the context vector through the convolutional layer. The neural network model takes the context features and weights as the input of the pooling layer, and reduces the dimensionality of the context features through the pooling layer. The output layer of the neural network model outputs the context features and weights after dimensionality reduction, and uses the context features after dimensionality reduction as point cloud feature information.
在本实施例中,计算机设备通过在当前帧点云数据中每个点对应的二维数据中提取多个数据维度,通过神经网络模型对多个数据维度进行预测运算,得 到点云特征信息。由于神经网络是预先训练的,能够通过神经网络模型准确地提取出当前帧点云数据中每个点的局部特征信息,进而有利于后续对当前帧点云数据进行空间特征信息的提取。In this embodiment, the computer device extracts multiple data dimensions from the two-dimensional data corresponding to each point in the point cloud data of the current frame, and performs prediction operations on the multiple data dimensions through a neural network model to obtain point cloud feature information. Since the neural network is pre-trained, the local feature information of each point in the current frame of point cloud data can be accurately extracted through the neural network model, which is beneficial to the subsequent extraction of spatial feature information of the current frame of point cloud data.
在其中一个实施例中,将当前帧点云数据在多个视角上进行投影,得到多个视角对应的二维平面,包括:将当前帧点云数据在鸟瞰视角上进行投影,得到鸟瞰视角对应的二维平面;将当前帧点云数据在正视视角上进行投影,得到正视视角对应的二维平面。In one of the embodiments, projecting the point cloud data of the current frame on multiple viewing angles to obtain a two-dimensional plane corresponding to the multiple viewing angles includes: projecting the point cloud data of the current frame on a bird's-eye view angle to obtain the corresponding bird's-eye view angle. The two-dimensional plane of the current frame; project the point cloud data of the current frame on the orthographic perspective to obtain the two-dimensional plane corresponding to the orthographic perspective.
计算机设备在获取当前帧点云数据后,可以将当前帧点云数据投影至多个视角。当前帧点云数据的坐标可以表示为(x,y,z)。计算机设备可以通过预设分辨率进行投影。多个视角可以包括鸟瞰视角、正视视角。例如,在进行鸟瞰视角投影的过程中,当预设分辨率为0.1m每个网格时,那么可以将-60<x<60,-60<y<60范围内的当前帧点云数据,投影至1200x 1200大小的二维平面中。将二维平面中的每个像素作为一个网格,0.05m范围内的当前帧点云数据在会都落在对应的网格上。计算机设备通过将当前帧点云数据在鸟瞰视角上进行投影,能够得到无遮挡的直观障碍物视图,避免了遮挡物体导致提取的点云特征信息不准确的问题。计算机设备通过将当前帧点云数据在正视视角上进行投影,能够对较小的目标进行更加直观地形状描述,例如,能够更加直观地描述行人。从而有利于计算机设备在多个视角对应的二维平面中提取更全面、准确的有效特征信息。After obtaining the point cloud data of the current frame, the computer device can project the point cloud data of the current frame to multiple perspectives. The coordinates of the point cloud data of the current frame can be expressed as (x, y, z). Computer equipment can project with preset resolutions. Multiple viewing angles may include a bird's-eye view and a front view. For example, in the process of bird's-eye perspective projection, when the preset resolution is 0.1m per grid, then the point cloud data of the current frame within the range of -60<x<60 and -60<y<60 can be set. Projected into a two-dimensional plane with a size of 1200x 1200. Taking each pixel in the two-dimensional plane as a grid, the point cloud data of the current frame within the range of 0.05m will all fall on the corresponding grid. By projecting the point cloud data of the current frame on the bird's-eye perspective, the computer device can obtain an unobstructed and intuitive obstruction view, avoiding the problem of inaccurate point cloud feature information extracted by obstructing objects. By projecting the point cloud data of the current frame on the front view, the computer device can describe the shape of a smaller target more intuitively, for example, can describe pedestrians more intuitively. This is beneficial for the computer equipment to extract more comprehensive and accurate effective feature information from the two-dimensional planes corresponding to multiple viewing angles.
在其中一个实施例中,检测模型包括多个网络层,通过检测模型对融合后的特征信息进行预测运算,输出障碍物检测结果,包括:将融合后的特征信息输入至检测模型的输入层;通过输入层将融合后的特征信息输入至检测模型的注意力层,通过注意力层计算融合后的特征信息对应的上下文向量与权重,生 成第一提取结果;将第一提取结果输入卷积层,通过卷积层提取上下文向量对应的上下文特征,生成第二提取结果;将第二提取结果输入池化层,通过池化层对第二提取结果进行降维处理;将降维处理后的第二提取结果输入全连接层,通过全连接层对降维处理后的第二提取结果进行分类得到分类结果,通过输出层将分类结果进行加权后输出;选取加权后输出的分类结果中权重最大的分类结果作为障碍物检测结果。In one of the embodiments, the detection model includes a plurality of network layers, and performing prediction operations on the fused feature information through the detection model, and outputting obstacle detection results, includes: inputting the fused feature information into the input layer of the detection model; The fused feature information is input to the attention layer of the detection model through the input layer, and the context vector and weight corresponding to the fused feature information are calculated through the attention layer to generate the first extraction result; the first extraction result is input to the convolutional layer , Extract the context feature corresponding to the context vector through the convolutional layer to generate the second extraction result; input the second extraction result into the pooling layer, and perform dimensionality reduction processing on the second extraction result through the pooling layer; Second, the extraction results are input to the fully connected layer, and the second extraction results after the dimensionality reduction process are classified through the fully connected layer to obtain the classification results, and the classification results are weighted through the output layer and output; the weighted classification results are selected from the weighted output The classification result is used as the obstacle detection result.
计算机设备中预先存储有训练后的检测模型。检测模型可以是预先利用大量样本数据进行训练后得到的检测模型。例如,检测模型可以是基于注意力层的2D卷积神经网络。检测模型可以包括多个网络层。例如,检测模型可以包括输入层、注意力层、卷积层、池化层、全连接层、输出层等多个网络层。计算机设备通过调用训练后的检测模型,将融合后的特征信息输入至检测模型的输入层。通过输入层将融合后的特征信息传输至注意力层,从而通过注意力层计算融合后的特征信息对应的上下文向量与权重,根据上下文向量与权重生成第一提取结果。检测模型将第一提取结果作为卷积层的输入,通过卷积层来提取上下文向量对应的上下文特征,根据上下文特征与权重生成第二提取结果。进而计算机设备将第二提取结果作为池化层的输入,通过池化层对第二提取结果进行降维处理,得到降维处理后的第二提取结果。计算机设备将降维处理后的第二提取结果作为全连接层的输入,对降维处理后的第二提取结果进行分类,得到分类结果。分类结果中可以包括障碍物的多种类别、多个位置信息等。进而通过输出层将分类结果进行加权后输出。进而计算机设备在加权后输出的分类结果中选取权重最大的分类结果,即为障碍物检测结果。障碍物检测结果中可以包括障碍物的位置信息、障碍物的大小、障碍物的形状等。The trained detection model is pre-stored in the computer equipment. The detection model may be a detection model obtained after pre-training with a large amount of sample data. For example, the detection model may be a 2D convolutional neural network based on the attention layer. The detection model can include multiple network layers. For example, the detection model may include multiple network layers such as an input layer, an attention layer, a convolutional layer, a pooling layer, a fully connected layer, and an output layer. The computer device inputs the fused feature information to the input layer of the detection model by calling the trained detection model. The fused feature information is transmitted to the attention layer through the input layer, and the context vector and weight corresponding to the fused feature information are calculated through the attention layer, and the first extraction result is generated according to the context vector and weight. The detection model uses the first extraction result as the input of the convolution layer, extracts the context feature corresponding to the context vector through the convolution layer, and generates the second extraction result according to the context feature and weight. Furthermore, the computer device uses the second extraction result as the input of the pooling layer, and performs dimensionality reduction processing on the second extraction result through the pooling layer to obtain the second extraction result after the dimensionality reduction processing. The computer device uses the second extraction result after the dimensionality reduction processing as the input of the fully connected layer, and classifies the second extraction result after the dimensionality reduction processing to obtain the classification result. The classification result can include multiple categories of obstacles, multiple location information, and so on. Furthermore, the classification results are weighted and output through the output layer. Furthermore, the computer device selects the classification result with the largest weight among the weighted classification results, which is the obstacle detection result. Obstacle detection results may include the location information of the obstacle, the size of the obstacle, the shape of the obstacle, and so on.
在本实施例中,计算机设备通过检测模型的注意力层计算融合后的特征信 息对应的上下文向量与权重,生成第一提取结果。能够将融合后的特征信息中的干扰信息进行过滤,实现对融合后的特征信息进行特征聚焦处理。通过卷积层提取上下文向量对应的上下文特征,生成第二提取结果,通过池化层对第二提取结果进行降维处理,能够提取主要的上下文特征,避免多余特征的影响。计算机设备通过对降维处理后的第二提取结果进行分类得到分类结果,并对分类结果进行加权后输出,选取加权后输出的分类结果中权重最大的分类结果作为障碍物检测结果,能够对分类结果进行归一化,进一步提高了障碍物检测的准确性。In this embodiment, the computer device calculates the context vector and weight corresponding to the fused feature information through the attention layer of the detection model, and generates the first extraction result. It can filter the interference information in the fused feature information, and realize the feature focus processing on the fused feature information. The context feature corresponding to the context vector is extracted through the convolutional layer to generate the second extraction result, and the second extraction result is reduced by the pooling layer, which can extract the main context features and avoid the influence of redundant features. The computer device classifies the second extraction result after dimensionality reduction to obtain the classification result, and then weights the classification result and outputs it. The classification result with the largest weight among the weighted classification results is selected as the obstacle detection result, which can classify The results are normalized to further improve the accuracy of obstacle detection.
在其中一个实施例中,二维平面包括多个像素,每个像素对应当前帧点云数据中多个点的二维数据,在对每个视角对应的二维平面进行特征提取之前,还包括:将每个像素对应的多个点的二维数据进行平均处理,得到平均值;根据平均值对相应的像素对应的点进行规范化处理。In one of the embodiments, the two-dimensional plane includes multiple pixels, and each pixel corresponds to the two-dimensional data of multiple points in the point cloud data of the current frame. Before feature extraction is performed on the two-dimensional plane corresponding to each view angle, it also includes : Perform average processing on the two-dimensional data of multiple points corresponding to each pixel to obtain the average value; perform normalization processing on the points corresponding to the corresponding pixels according to the average value.
计算机设备在对每个视角对应的二维平面进行特征提取之前,还可以对二维平面中当前帧点云数据中的多个点进行规范化处理。具体的,二维平面中包括多个像素,每个像素可以用一个网格来表示,在每个网格中包括当前帧点云数据中的多个点。计算机设备将每个网格中的多个点的坐标进行平均处理,得到平均值。进而计算机设备将网格中的每个点的坐标与平均值作差,实现将每个网格中的当前帧点云数据进行规范化处理。通过对二维平面中当前帧点云数据中的多个点进行规范化处理,有利于后续利用特征提取模型进行特征提取。Before performing feature extraction on the two-dimensional plane corresponding to each viewing angle, the computer device may also perform normalization processing on multiple points in the current frame point cloud data in the two-dimensional plane. Specifically, the two-dimensional plane includes multiple pixels, and each pixel may be represented by a grid, and each grid includes multiple points in the point cloud data of the current frame. The computer equipment averages the coordinates of multiple points in each grid to obtain the average value. Furthermore, the computer equipment makes the difference between the coordinates of each point in the grid and the average value to realize the normalization of the point cloud data of the current frame in each grid. By normalizing multiple points in the point cloud data of the current frame in the two-dimensional plane, it is beneficial to subsequently use the feature extraction model for feature extraction.
在其中一个实施例中,对每个视角对应的二维平面进行特征提取,得到每个视角对应的点云特征信息,还包括:调用多线程在每个视角对应的二维平面中并发提取每个视角对应的点云特征信息;在将每个视角对应的点云特征信息和当前帧图像数据输入至对应的特征提取模型中之前,还包括:利用多线程将 每个视角对应的点云特征信息和当前帧图像数据进行并行转换,得到每个视角对应的点云特征向量和当前帧图像数据对应的图像矩阵。In one of the embodiments, the feature extraction is performed on the two-dimensional plane corresponding to each perspective to obtain the point cloud feature information corresponding to each perspective, and the method further includes: invoking multiple threads to concurrently extract each perspective in the two-dimensional plane corresponding to each perspective. Point cloud feature information corresponding to each perspective; before inputting the point cloud feature information corresponding to each perspective and the current frame image data into the corresponding feature extraction model, it also includes: using multi-threaded point cloud features corresponding to each perspective The information and the current frame image data are converted in parallel to obtain the point cloud feature vector corresponding to each view angle and the image matrix corresponding to the current frame image data.
计算机设备通过调用多个线程,通过多线程在每个视角对应的二维平面中并发提取每个视角对应的点云特征信息。从而提高了点云特征信息的提取效率。计算机设备还可以在将每个视角对应的点云特征信息和当前帧图像数据输入至对应的特征提取模型中之前,利用该多线程将每个视角对应的点云特征信息和当前帧图像数据进行并行转换,进而有效减少了特征提取模型进行特征提取的耗时。The computer device uses multiple threads to concurrently extract the point cloud feature information corresponding to each perspective in the two-dimensional plane corresponding to each perspective through multiple threads. Thereby improving the extraction efficiency of point cloud feature information. The computer device can also use the multi-thread to perform the point cloud feature information corresponding to each perspective and the current frame image data before inputting the point cloud feature information corresponding to each perspective and the current frame image data into the corresponding feature extraction model. Parallel conversion effectively reduces the time-consuming feature extraction model for feature extraction.
在其中一个实施例中,计算机设备还可以根据障碍物检测结果获取当前环境中多个障碍物的历史轨迹信息,同时获取车辆的当前位置信息。根据历史轨迹信息和当前位置信息预测多个障碍物在预设时间段内的轨迹。In one of the embodiments, the computer device may also obtain historical trajectory information of multiple obstacles in the current environment according to the obstacle detection result, and at the same time obtain the current position information of the vehicle. Predict the trajectory of multiple obstacles within a preset time period based on historical trajectory information and current position information.
具体的,计算机设备对障碍物检测结果中的障碍物的运动过程进行跟踪,根据障碍物在前一时刻的位置信息来预测当前时刻的位置信息,将预测得到的当前时刻的位置信息与实际位置信息进行比较,得到误差信息。计算机设备根据误差信息对下一时刻的位置信息进行修正,从而得到多个障碍物的历史轨迹信息。计算机设备可以获取车载定位器发送的当前位置信息。从而计算机设备可以将获取到的多个障碍物的历史轨迹信息渲染至一张特征图中,得到轨迹渲染图。历史轨迹信息可以是多个障碍物的历史每帧的轨迹。计算机设备将多个障碍物的历史轨迹信息在当前帧进行渲染,进而得到轨迹渲染图。轨迹渲染图中障碍物在每一帧的颜色随着离当前帧的时间远近发生变化,离当前帧的时间越远,障碍物的颜色越浅。能够得到障碍物自身以及周围的环境信息,实现从多方面考虑轨迹的影响因素,更有利于提高轨迹预测的准确性。Specifically, the computer device tracks the movement process of the obstacle in the obstacle detection result, predicts the position information at the current time based on the position information of the obstacle at the previous time, and compares the predicted position information at the current time with the actual position. Information is compared to obtain error information. The computer device corrects the position information at the next moment according to the error information, thereby obtaining historical trajectory information of multiple obstacles. The computer equipment can obtain the current position information sent by the vehicle-mounted locator. Therefore, the computer device can render the acquired historical trajectory information of multiple obstacles into a feature map to obtain a trajectory rendering map. The historical trajectory information may be the trajectory of each frame of the history of multiple obstacles. The computer device renders the historical trajectory information of multiple obstacles in the current frame to obtain a trajectory rendering map. The color of obstacles in each frame in the trajectory rendering diagram changes with the distance from the current frame. The farther away from the current frame, the lighter the color of the obstacle. The obstacle itself and the surrounding environment information can be obtained, and the influence factors of the trajectory can be considered from various aspects, which is more conducive to improving the accuracy of trajectory prediction.
计算机设备获取车载定位器采集到的当前位置信息。当前位置信息可以是 当前时刻车辆在高精地图中的位置信息。当前位置信息可以是用经纬度的形式表示。计算机设备在当前位置信息中提取地图元素。地图元素可以包括车道线、中心线、人行道、停止线等信息。计算机设备可以根据多个通道维度将提取出来的地图元素进行渲染,将地图元素渲染至通道维度对应的地图元素渲染图中。当地图元素不同时,地图元素对应的通道维度也可以是不同的。通道维度可以包括颜色通道、元素通道等。颜色通道可以包括红色、绿色、蓝色三个通道。元素通道可以包括车道线通道、中心线通道以及人行道通道等。能够通过地图元素对应的通道维度直观、准确地将障碍物的当前位置进行渲染,有利于后续进行轨迹预测。The computer equipment obtains the current position information collected by the vehicle-mounted locator. The current location information may be the location information of the vehicle on the high-precision map at the current moment. The current location information can be expressed in the form of latitude and longitude. The computer equipment extracts map elements from the current location information. Map elements can include information such as lane lines, center lines, sidewalks, and stop lines. The computer device may render the extracted map elements according to multiple channel dimensions, and render the map elements into a map element rendering map corresponding to the channel dimensions. When the map elements are different, the channel dimensions corresponding to the map elements can also be different. Channel dimensions can include color channels, element channels, and so on. The color channel can include three channels of red, green, and blue. Elemental passages can include lane-line passages, center-line passages, and sidewalk passages. The current position of the obstacle can be rendered intuitively and accurately through the channel dimension corresponding to the map element, which is conducive to subsequent trajectory prediction.
计算机设备在得到轨迹渲染图以及地图元素渲染图后,可将轨迹渲染图以及地图元素渲染图进行拼接。计算机设备确定轨迹渲染图与地图元素渲染图的相应通道维度,将轨迹渲染图以及地图元素渲染图在相应的通道维度上进行图像拼接,进而得到拼接后的图像矩阵。拼接后的图像矩阵可以是包含有轨迹渲染图以及地图元素渲染图的完整图像。After the computer device obtains the trajectory rendering image and the map element rendering image, the trajectory rendering image and the map element rendering image can be stitched together. The computer device determines the corresponding channel dimensions of the trajectory rendering map and the map element rendering map, and performs image stitching on the trajectory rendering map and the map element rendering map in the corresponding channel dimensions to obtain a spliced image matrix. The spliced image matrix may be a complete image including the trajectory rendering map and the map element rendering map.
计算机设备在获取当前环境中多个障碍物的历史轨迹信息以及当前位置信息之前,已经预先训练有特征提取器。计算机设备调用训练后的特征提取器,将拼接后的图像矩阵输入至训练后的特征提取器中。计算机设备通过特征提取器提取出拼接后的图像矩阵对应的图像特征信息以及上下文特征信息,进而通过特征提取器的全连接层输出拼接后的图像矩阵对应的特征提取结果。实现将障碍物轨迹的多方面影响因素进行结合,进一步提高了特征提取结果的全面性。计算机设备可以通过回归预测的方式对特征提取结果进行运算,得到多个障碍物在预设时间段内的轨迹。由于障碍物检测结果更为全面、准确,且特征提取结果中包含多个障碍物的历史帧的轨迹,扩大了环境信息的范围,实现根据多 方面的影响因素进行轨迹预测,从而有效提供了轨迹预测的准确性。The computer device has pre-trained a feature extractor before acquiring the historical trajectory information and current position information of multiple obstacles in the current environment. The computer device calls the trained feature extractor, and inputs the spliced image matrix into the trained feature extractor. The computer device extracts the image feature information and context feature information corresponding to the spliced image matrix through the feature extractor, and then outputs the feature extraction result corresponding to the spliced image matrix through the fully connected layer of the feature extractor. It realizes the combination of various influence factors of the obstacle trajectory, and further improves the comprehensiveness of the feature extraction results. The computer equipment can calculate the feature extraction results by means of regression prediction to obtain the trajectories of multiple obstacles within a preset time period. Because the obstacle detection result is more comprehensive and accurate, and the feature extraction result includes the trajectory of multiple obstacles in the history frame, the scope of environmental information is expanded, and the trajectory prediction based on various influencing factors is realized, thereby effectively providing the trajectory The accuracy of the forecast.
在其中一个实施例中,如图4所示,提供了一种基于无人驾驶技术的障碍物检测装置,包括:获取模块402、投影模块404、第一提取模块406、第二提取模块408、融合模块410和预测模块412,其中:In one of the embodiments, as shown in FIG. 4, an obstacle detection device based on unmanned driving technology is provided, including: an acquisition module 402, a projection module 404, a first extraction module 406, a second extraction module 408, The fusion module 410 and the prediction module 412, where:
获取模块402,用于获取当前帧点云数据和预设角度范围内的当前帧图像数据。The acquiring module 402 is used to acquire the point cloud data of the current frame and the image data of the current frame within a preset angle range.
投影模块404,用于将当前帧点云数据在多个视角上进行投影,得到多个视角对应的二维平面。The projection module 404 is configured to project the point cloud data of the current frame on multiple viewing angles to obtain two-dimensional planes corresponding to the multiple viewing angles.
第一提取模块406,用于对每个视角对应的二维平面进行特征提取,得到每个视角对应的点云特征信息。The first extraction module 406 is configured to perform feature extraction on the two-dimensional plane corresponding to each view angle to obtain point cloud feature information corresponding to each view angle.
第二提取模块408,用于将每个视角对应的点云特征信息和当前帧图像数据输入至对应的特征提取模型中,通过相应的特征提取模型并行提取每个视角对应的空间特征信息和当前帧图像数据对应的图像特征信息。The second extraction module 408 is used to input the point cloud feature information and current frame image data corresponding to each perspective into the corresponding feature extraction model, and extract the spatial feature information corresponding to each perspective and the current frame through the corresponding feature extraction model in parallel. Image feature information corresponding to the frame image data.
融合模块410,用于将多个视角对应的空间特征信息和图像特征信息进行融合,得到融合后的特征信息。The fusion module 410 is used for fusing spatial feature information and image feature information corresponding to multiple viewing angles to obtain fused feature information.
预测模块412,用于将融合后的特征信息输入至训练后的检测模型中,通过检测模型对融合后的特征信息进行预测运算,输出障碍物检测结果。The prediction module 412 is configured to input the fused feature information into the trained detection model, perform prediction operations on the fused feature information through the detection model, and output obstacle detection results.
在其中一个实施例中,融合模块410还用于根据预设参数将多个视角对应的空间特征信息和图像特征信息进行拼接,得到拼接后的特征信息;根据预设参数将拼接后的特征信息进行对齐至预设视角,得到对齐后的特征信息,将对齐后的特征信息作为融合后的特征信息。In one of the embodiments, the fusion module 410 is further configured to splice the spatial characteristic information and image characteristic information corresponding to the multiple viewing angles according to preset parameters to obtain the spliced characteristic information; according to the preset parameters, the spliced characteristic information The alignment is performed to the preset viewing angle to obtain the aligned feature information, and the aligned feature information is used as the fused feature information.
在其中一个实施例中,第一提取模块406还用于在当前帧点云数据中每个点对应的二维数据中提取多个数据维度;将多个数据维度输入至训练后的神经 网络模型中,通过神经网络模型对多个数据维度进行预测运算,得到点云特征信息。In one of the embodiments, the first extraction module 406 is also used to extract multiple data dimensions from the two-dimensional data corresponding to each point in the point cloud data of the current frame; input the multiple data dimensions to the trained neural network model In the process, the neural network model is used to perform predictive operations on multiple data dimensions to obtain point cloud feature information.
在其中一个实施例中,投影模块404还用于将当前帧点云数据在鸟瞰视角上进行投影,得到鸟瞰视角对应的二维平面;将当前帧点云数据在正视视角上进行投影,得到正视视角对应的二维平面。In one of the embodiments, the projection module 404 is also used to project the point cloud data of the current frame on the bird's-eye view angle to obtain a two-dimensional plane corresponding to the bird's-eye view angle; Two-dimensional plane corresponding to the viewing angle
在其中一个实施例中,预测模块412还用于将融合后的特征信息输入至检测模型的输入层;通过输入层将融合后的特征信息输入至检测模型的注意力层,通过注意力层计算融合后的特征信息对应的上下文向量与权重,生成第一提取结果;将第一提取结果输入卷积层,通过卷积层提取上下文向量对应的上下文特征,生成第二提取结果;将第二提取结果输入池化层,通过池化层对第二提取结果进行降维处理;将降维处理后的第二提取结果输入全连接层,通过全连接层对降维处理后的第二提取结果进行分类得到分类结果,通过输出层将分类结果进行加权后输出;选取加权后输出的分类结果中权重最大的分类结果作为障碍物检测结果。In one of the embodiments, the prediction module 412 is also used to input the fused feature information to the input layer of the detection model; the fused feature information is input to the attention layer of the detection model through the input layer, and the attention layer is used to calculate The context vector and weight corresponding to the fused feature information are generated to generate the first extraction result; the first extraction result is input to the convolutional layer, and the context feature corresponding to the context vector is extracted through the convolutional layer to generate the second extraction result; the second extraction The result is input to the pooling layer, and the second extraction result is reduced by the pooling layer; the second extraction result after the dimensionality reduction is input into the fully connected layer, and the second extraction result after the dimensionality reduction is processed through the fully connected layer The classification results are obtained by classification, and the classification results are weighted and output through the output layer; the classification result with the largest weight among the weighted classification results is selected as the obstacle detection result.
在其中一个实施例中,上述装置还包括:规范化处理模块,用于将每个像素对应的多个点的二维数据进行平均处理,得到平均值;根据平均值对相应的像素对应的点进行规范化处理。In one of the embodiments, the above-mentioned device further includes: a normalization processing module for averaging the two-dimensional data of multiple points corresponding to each pixel to obtain an average value; according to the average value, the points corresponding to the corresponding pixels are processed Standardized processing.
在其中一个实施例中,第一提取模块406还用于调用多线程在每个视角对应的二维平面中并发提取每个视角对应的点云特征信息;上述装置还包括:转换模块,用于利用多线程将每个视角对应的点云特征信息和当前帧图像数据进行并行转换,得到每个视角对应的点云特征向量和当前帧图像数据对应的图像矩阵。In one of the embodiments, the first extraction module 406 is also used to call multiple threads to concurrently extract the point cloud feature information corresponding to each perspective in the two-dimensional plane corresponding to each perspective; the above-mentioned device further includes: a conversion module for The point cloud feature information corresponding to each view angle and the current frame image data are converted in parallel by using multiple threads to obtain the point cloud feature vector corresponding to each view angle and the image matrix corresponding to the current frame image data.
关于基于无人驾驶技术的障碍物检测装置的具体限定可以参见上文中对于 轨迹预测方法的限定,在此不再赘述。上述基于无人驾驶技术的障碍物检测装置中的各个模块可全部或部分通过软件、硬件及其组合来实现。上述各模块可以硬件形式内嵌于或独立于计算机设备中的处理器中,也可以以软件形式存储于计算机设备中的存储器中,以便于处理器调用执行以上各个模块对应的操作。Regarding the specific limitations of the obstacle detection device based on unmanned driving technology, please refer to the above limitation on the trajectory prediction method, which will not be repeated here. The various modules in the above-mentioned obstacle detection device based on unmanned driving technology can be implemented in whole or in part by software, hardware, and a combination thereof. The above-mentioned modules may be embedded in the form of hardware or independent of the processor in the computer equipment, or may be stored in the memory of the computer equipment in the form of software, so that the processor can call and execute the operations corresponding to the above-mentioned modules.
在其中一个实施例中,提供了一种计算机设备,其内部结构图可以如图5所示。该计算机设备包括通过系统总线连接的处理器、存储器、通信接口和数据库。其中,该计算机设备的处理器用于提供计算和控制能力。该计算机设备的存储器包括非易失性存储介质、内存储器。该非易失性存储介质存储有操作系统、计算机可读指令和数据库。该内存储器为非易失性存储介质中的操作系统和计算机可读指令的运行提供环境。该计算机设备的数据库用于存储障碍物检测结果。该计算机设备的通信接口用于与第一车载传感器、第二车载传感器、车载定位传感器连接通信。该计算机可读指令被处理器执行时以实现一种障碍物检测方法。In one of the embodiments, a computer device is provided, and its internal structure diagram may be as shown in FIG. 5. The computer equipment includes a processor, a memory, a communication interface and a database connected through a system bus. Among them, the processor of the computer device is used to provide calculation and control capabilities. The memory of the computer device includes a non-volatile storage medium and an internal memory. The non-volatile storage medium stores an operating system, computer readable instructions, and a database. The internal memory provides an environment for the operation of the operating system and computer-readable instructions in the non-volatile storage medium. The database of the computer equipment is used to store obstacle detection results. The communication interface of the computer device is used to connect and communicate with the first vehicle-mounted sensor, the second vehicle-mounted sensor, and the vehicle-mounted positioning sensor. The computer readable instruction is executed by the processor to realize an obstacle detection method.
本领域技术人员可以理解,图5中示出的结构,仅仅是与本申请方案相关的部分结构的框图,并不构成对本申请方案所应用于其上的计算机设备的限定,具体的计算机设备可以包括比图中所示更多或更少的部件,或者组合某些部件,或者具有不同的部件布置。Those skilled in the art can understand that the structure shown in FIG. 5 is only a block diagram of a part of the structure related to the solution of the present application, and does not constitute a limitation on the computer device to which the solution of the present application is applied. The specific computer device may Including more or fewer parts than shown in the figure, or combining some parts, or having a different arrangement of parts.
一种计算机设备,包括存储器及一个或多个处理器,存储器中储存有计算机可读指令,计算机可读指令被一个或多个处理器执行时,使得一个或多个处理器执行上述各个方法实施例中的步骤。A computer device that includes a memory and one or more processors. The memory stores computer-readable instructions. When the computer-readable instructions are executed by one or more processors, the one or more processors execute each of the foregoing method implementations. The steps in the example.
一个或多个存储有计算机可读指令的非易失性计算机可读存储介质,计算机可读指令被一个或多个处理器执行时,使得一个或多个处理器执行上述各个方法实施例中的步骤。One or more non-volatile computer-readable storage media storing computer-readable instructions. When the computer-readable instructions are executed by one or more processors, the one or more processors execute the steps in each of the foregoing method embodiments. step.
本领域普通技术人员可以理解实现上述实施例方法中的全部或部分流程,是可以通过计算机可读指令来指令相关的硬件来完成,所述的计算机可读指令可存储于一非易失性计算机可读取存储介质中,该计算机可读指令在执行时,可包括如上述各方法的实施例的流程。其中,本申请所提供的各实施例中所使用的对存储器、存储、数据库或其它介质的任何引用,均可包括非易失性和/或易失性存储器。非易失性存储器可包括只读存储器(ROM)、可编程ROM(PROM)、电可编程ROM(EPROM)、电可擦除可编程ROM(EEPROM)或闪存。易失性存储器可包括随机存取存储器(RAM)或者外部高速缓冲存储器。作为说明而非局限,RAM以多种形式可得,诸如静态RAM(SRAM)、动态RAM(DRAM)、同步DRAM(SDRAM)、双数据率SDRAM(DDRSDRAM)、增强型SDRAM(ESDRAM)、同步链路(Synchlink)DRAM(SLDRAM)、存储器总线(Rambus)直接RAM(RDRAM)、直接存储器总线动态RAM(DRDRAM)、以及存储器总线动态RAM(RDRAM)等。A person of ordinary skill in the art can understand that all or part of the processes in the above-mentioned embodiment methods can be implemented by instructing relevant hardware through computer-readable instructions. The computer-readable instructions can be stored in a non-volatile computer. In a readable storage medium, when the computer-readable instructions are executed, they may include the processes of the above-mentioned method embodiments. Wherein, any reference to memory, storage, database, or other media used in the embodiments provided in this application may include non-volatile and/or volatile memory. Non-volatile memory may include read only memory (ROM), programmable ROM (PROM), electrically programmable ROM (EPROM), electrically erasable programmable ROM (EEPROM), or flash memory. Volatile memory may include random access memory (RAM) or external cache memory. As an illustration and not a limitation, RAM is available in many forms, such as static RAM (SRAM), dynamic RAM (DRAM), synchronous DRAM (SDRAM), double data rate SDRAM (DDRSDRAM), enhanced SDRAM (ESDRAM), synchronous chain Channel (Synchlink) DRAM (SLDRAM), memory bus (Rambus) direct RAM (RDRAM), direct memory bus dynamic RAM (DRDRAM), and memory bus dynamic RAM (RDRAM), etc.
以上实施例的各技术特征可以进行任意的组合,为使描述简洁,未对上述实施例中的各个技术特征所有可能的组合都进行描述,然而,只要这些技术特征的组合不存在矛盾,都应当认为是本说明书记载的范围。The technical features of the above embodiments can be combined arbitrarily. In order to make the description concise, all possible combinations of the technical features in the above embodiments are not described. However, as long as there is no contradiction in the combination of these technical features, they should be It is considered as the range described in this specification.
以上所述实施例仅表达了本申请的几种实施方式,其描述较为具体和详细,但并不能因此而理解为对发明专利范围的限制。应当指出的是,对于本领域的普通技术人员来说,在不脱离本申请构思的前提下,还可以做出若干变形和改进,这些都属于本申请的保护范围。因此,本申请专利的保护范围应以所附权利要求为准。The above-mentioned embodiments only express several implementation manners of the present application, and the description is relatively specific and detailed, but it should not be understood as a limitation on the scope of the invention patent. It should be pointed out that for those of ordinary skill in the art, without departing from the concept of this application, several modifications and improvements can be made, and these all fall within the protection scope of this application. Therefore, the scope of protection of the patent of this application shall be subject to the appended claims.

Claims (20)

  1. 一种基于无人驾驶技术的障碍物检测方法,包括:An obstacle detection method based on unmanned driving technology, including:
    获取当前帧点云数据和预设角度范围内的当前帧图像数据;Obtain current frame point cloud data and current frame image data within a preset angle range;
    将所述当前帧点云数据在多个视角上进行投影,得到多个视角对应的二维平面;Projecting the point cloud data of the current frame on multiple viewing angles to obtain two-dimensional planes corresponding to the multiple viewing angles;
    对每个视角对应的二维平面进行特征提取,得到每个视角对应的点云特征信息;Perform feature extraction on the two-dimensional plane corresponding to each perspective, and obtain the point cloud feature information corresponding to each perspective;
    将每个视角对应的点云特征信息和所述当前帧图像数据输入至对应的特征提取模型中,通过相应的特征提取模型并行提取每个视角对应的空间特征信息和所述当前帧图像数据对应的图像特征信息;The point cloud feature information corresponding to each perspective and the current frame image data are input into the corresponding feature extraction model, and the spatial feature information corresponding to each perspective and the current frame image data are extracted in parallel through the corresponding feature extraction model Image feature information;
    将多个视角对应的空间特征信息和所述图像特征信息进行融合,得到融合后的特征信息;及Fusing the spatial feature information corresponding to multiple viewing angles with the image feature information to obtain the fused feature information; and
    将所述融合后的特征信息输入至训练后的检测模型中,通过所述检测模型对所述融合后的特征信息进行预测运算,输出障碍物检测结果。The fused feature information is input into a trained detection model, and the fused feature information is predicted and calculated through the detection model, and an obstacle detection result is output.
  2. 根据权利要求1所述的方法,其特征在于,所述将多个视角对应的空间特征信息和所述图像特征信息进行融合,得到融合后的特征信息,包括:The method according to claim 1, wherein the fusing the spatial feature information corresponding to multiple viewing angles with the image feature information to obtain the fused feature information comprises:
    根据预设参数将多个视角对应的空间特征信息和所述图像特征信息进行拼接,得到拼接后的特征信息;及Splicing the spatial feature information corresponding to the multiple viewing angles and the image feature information according to preset parameters to obtain the spliced feature information; and
    根据所述预设参数将所述拼接后的特征信息进行对齐至预设视角,得到对齐后的特征信息,将所述对齐后的特征信息作为融合后的特征信息。The spliced feature information is aligned to a preset viewing angle according to the preset parameters to obtain the aligned feature information, and the aligned feature information is used as the fused feature information.
  3. 据权利要求1所述的方法,其特征在于,所述二维平面包括所述当前帧点云数据中每个点对应的二维数据,所述对每个视角对应的二维平面进行特征提取,得到点云特征信息,包括:The method according to claim 1, wherein the two-dimensional plane includes two-dimensional data corresponding to each point in the point cloud data of the current frame, and the feature extraction is performed on the two-dimensional plane corresponding to each perspective To get point cloud feature information, including:
    在所述当前帧点云数据中每个点对应的二维数据中提取多个数据维度;及Extracting multiple data dimensions from the two-dimensional data corresponding to each point in the point cloud data of the current frame; and
    将多个数据维度输入至训练后的神经网络模型中,通过所述神经网络模型对多个数据维度进行预测运算,得到点云特征信息。Inputting multiple data dimensions into the trained neural network model, and performing prediction operations on the multiple data dimensions through the neural network model to obtain point cloud feature information.
  4. 根据权利要求1所述的方法,其特征在于,所述将所述当前帧点云数据在多个视角上进行投影,得到多个视角对应的二维平面,包括:The method according to claim 1, wherein the projecting the point cloud data of the current frame on multiple viewing angles to obtain a two-dimensional plane corresponding to the multiple viewing angles comprises:
    将所述当前帧点云数据在鸟瞰视角上进行投影,得到鸟瞰视角对应的二维平面;及Projecting the point cloud data of the current frame on a bird's-eye view angle to obtain a two-dimensional plane corresponding to the bird's-eye view angle; and
    将所述当前帧点云数据在正视视角上进行投影,得到正视视角对应的二维平面。The point cloud data of the current frame is projected on the front view angle to obtain a two-dimensional plane corresponding to the front view angle.
  5. 根据权利要求1所述的方法,其特征在于,所述检测模型包括多个网络层,所述通过所述检测模型对所述融合后的特征信息进行预测运算,输出障碍物检测结果,包括:The method according to claim 1, wherein the detection model includes a plurality of network layers, and the prediction operation on the fused feature information through the detection model to output an obstacle detection result comprises:
    将所述融合后的特征信息输入至所述检测模型的输入层;Input the fused feature information to the input layer of the detection model;
    通过所述输入层将所述融合后的特征信息输入至所述检测模型的注意力层,通过所述注意力层计算所述融合后的特征信息对应的上下文向量与权重,生成第一提取结果;The fused feature information is input to the attention layer of the detection model through the input layer, and the context vector and weight corresponding to the fused feature information are calculated through the attention layer to generate a first extraction result ;
    将所述第一提取结果输入卷积层,通过所述卷积层提取所述上下文向量对应的上下文特征,生成第二提取结果;Inputting the first extraction result into a convolutional layer, extracting context features corresponding to the context vector through the convolutional layer, and generating a second extraction result;
    将所述第二提取结果输入池化层,通过所述池化层对所述第二提取结果进行降维处理;Inputting the second extraction result into a pooling layer, and performing dimensionality reduction processing on the second extraction result through the pooling layer;
    将降维处理后的第二提取结果输入全连接层,通过所述全连接层对所述降维处理后的第二提取结果进行分类得到分类结果,通过输出层将所述分类 结果进行加权后输出;及The second extraction result after the dimensionality reduction process is input into the fully connected layer, the second extraction result after the dimensionality reduction process is classified by the fully connected layer to obtain the classification result, and the classification result is weighted by the output layer Output; and
    选取加权后输出的分类结果中权重最大的分类结果作为障碍物检测结果。The classification result with the largest weight among the weighted classification results is selected as the obstacle detection result.
  6. 根据权利要求1所述的方法,其特征在于,所述二维平面包括多个像素,每个像素对应所述当前帧点云数据中多个点的二维数据,所述在所述对每个视角对应的二维平面进行特征提取之前,还包括:The method according to claim 1, wherein the two-dimensional plane includes a plurality of pixels, and each pixel corresponds to the two-dimensional data of a plurality of points in the point cloud data of the current frame, and the Before feature extraction is performed on the two-dimensional plane corresponding to each viewing angle, it also includes:
    将每个像素对应的多个点的二维数据进行平均处理,得到平均值;及Perform averaging processing on the two-dimensional data of multiple points corresponding to each pixel to obtain the average value; and
    根据所述平均值对相应的像素对应的点进行规范化处理。The points corresponding to the corresponding pixels are normalized according to the average value.
  7. 根据权利要求1至6任意一项所述的方法,其特征在于,所述对每个视角对应的二维平面进行特征提取,得到每个视角对应的点云特征信息,还包括:The method according to any one of claims 1 to 6, wherein the performing feature extraction on the two-dimensional plane corresponding to each view angle to obtain the point cloud feature information corresponding to each view angle further comprises:
    调用多线程在每个视角对应的二维平面中并发提取每个视角对应的点云特征信息;及Calling multiple threads to concurrently extract the point cloud feature information corresponding to each viewing angle in the two-dimensional plane corresponding to each viewing angle; and
    在所述将每个视角对应的点云特征信息和所述当前帧图像数据输入至对应的特征提取模型中之前,还包括:Before inputting the point cloud feature information corresponding to each view angle and the current frame image data into the corresponding feature extraction model, the method further includes:
    利用所述多线程将每个视角对应的点云特征信息和所述当前帧图像数据进行并行转换,得到每个视角对应的点云特征向量和所述当前帧图像数据对应的图像矩阵。The point cloud feature information corresponding to each view angle and the current frame image data are converted in parallel by using the multi-thread to obtain the point cloud feature vector corresponding to each view angle and the image matrix corresponding to the current frame image data.
  8. 一种基于无人驾驶技术的障碍物检测装置,包括:An obstacle detection device based on unmanned driving technology, including:
    获取模块,用于获取当前帧点云数据和预设角度范围内的当前帧图像数据;The acquisition module is used to acquire the current frame point cloud data and the current frame image data within a preset angle range;
    投影模块,用于将所述当前帧点云数据在多个视角上进行投影,得到多个视角对应的二维平面;A projection module, configured to project the point cloud data of the current frame on multiple viewing angles to obtain two-dimensional planes corresponding to the multiple viewing angles;
    第一提取模块,用于对每个视角对应的二维平面进行特征提取,得到每个视角对应的点云特征信息;The first extraction module is used to perform feature extraction on the two-dimensional plane corresponding to each perspective to obtain point cloud feature information corresponding to each perspective;
    第二提取模块,用于将每个视角对应的点云特征信息和所述当前帧图像数据输入至对应的特征提取模型中,通过相应的特征提取模型并行提取每个视角对应的空间特征信息和所述当前帧图像数据对应的图像特征信息;The second extraction module is used to input the point cloud feature information corresponding to each perspective and the current frame image data into the corresponding feature extraction model, and extract the spatial feature information corresponding to each perspective in parallel through the corresponding feature extraction model. Image feature information corresponding to the current frame of image data;
    融合模块,用于将多个视角对应的空间特征信息和所述图像特征信息进行融合,得到融合后的特征信息;及The fusion module is used to fuse the spatial feature information corresponding to multiple perspectives with the image feature information to obtain the fused feature information; and
    预测模块,用于将所述融合后的特征信息输入至训练后的检测模型中,通过所述检测模型对所述融合后的特征信息进行预测运算,输出障碍物检测结果。The prediction module is used to input the fused feature information into a trained detection model, and perform prediction operations on the fused feature information through the detection model, and output obstacle detection results.
  9. 根据权利要求8所述的装置,其特征在于,所述融合模块还用于根据预设参数将多个视角对应的空间特征信息和所述图像特征信息进行拼接,得到拼接后的特征信息;及根据所述预设参数将所述拼接后的特征信息进行对齐至预设视角,得到对齐后的特征信息,将所述对齐后的特征信息作为融合后的特征信息。The device according to claim 8, wherein the fusion module is further configured to splice the spatial characteristic information corresponding to the multiple viewing angles and the image characteristic information according to preset parameters to obtain the spliced characteristic information; and The spliced feature information is aligned to a preset viewing angle according to the preset parameters to obtain the aligned feature information, and the aligned feature information is used as the fused feature information.
  10. 根据权利要求8所述的装置,其特征在于,所述第一提取模块还用于在所述当前帧点云数据中每个点对应的二维数据中提取多个数据维度;及将多个数据维度输入至训练后的神经网络模型中,通过所述神经网络模型对多个数据维度进行预测运算,得到点云特征信息。The device according to claim 8, wherein the first extraction module is further configured to extract multiple data dimensions from the two-dimensional data corresponding to each point in the current frame point cloud data; and The data dimensions are input into the trained neural network model, and prediction operations are performed on multiple data dimensions through the neural network model to obtain point cloud feature information.
  11. 一种计算机设备,包括存储器及一个或多个处理器,所述存储器中存储有计算机可读指令,所述计算机可读指令被所述一个或多个处理器执行时,使得所述一个或多个处理器执行以下步骤:A computer device includes a memory and one or more processors. The memory stores computer-readable instructions. When the computer-readable instructions are executed by the one or more processors, the one or more processors Each processor performs the following steps:
    获取当前帧点云数据和预设角度范围内的当前帧图像数据;Obtain current frame point cloud data and current frame image data within a preset angle range;
    将所述当前帧点云数据在多个视角上进行投影,得到多个视角对应的二维平面;Projecting the point cloud data of the current frame on multiple viewing angles to obtain two-dimensional planes corresponding to the multiple viewing angles;
    对每个视角对应的二维平面进行特征提取,得到每个视角对应的点云特征信息;Perform feature extraction on the two-dimensional plane corresponding to each perspective, and obtain the point cloud feature information corresponding to each perspective;
    将每个视角对应的点云特征信息和所述当前帧图像数据输入至对应的特征提取模型中,通过相应的特征提取模型并行提取每个视角对应的空间特征信息和所述当前帧图像数据对应的图像特征信息;The point cloud feature information corresponding to each perspective and the current frame image data are input into the corresponding feature extraction model, and the spatial feature information corresponding to each perspective and the current frame image data are extracted in parallel through the corresponding feature extraction model Image feature information;
    将多个视角对应的空间特征信息和所述图像特征信息进行融合,得到融合后的特征信息;及Fusing the spatial feature information corresponding to multiple viewing angles with the image feature information to obtain the fused feature information; and
    将所述融合后的特征信息输入至训练后的检测模型中,通过所述检测模型对所述融合后的特征信息进行预测运算,输出障碍物检测结果。The fused feature information is input into a trained detection model, and the fused feature information is predicted and calculated through the detection model, and an obstacle detection result is output.
  12. 根据权利要求11所述的计算机设备,其特征在于,所述处理器执行所述计算机可读指令时还执行以下步骤:根据预设参数将多个视角对应的空间特征信息和所述图像特征信息进行拼接,得到拼接后的特征信息;及根据所述预设参数将所述拼接后的特征信息进行对齐至预设视角,得到对齐后的特征信息,将所述对齐后的特征信息作为融合后的特征信息。The computer device according to claim 11, wherein the processor further executes the following step when executing the computer-readable instruction: according to preset parameters, the spatial feature information corresponding to the multiple viewing angles and the image feature information Perform splicing to obtain spliced feature information; and align the spliced feature information to a preset viewing angle according to the preset parameters to obtain the aligned feature information, and use the aligned feature information as the fused Characteristic information.
  13. 根据权利要求11所述的计算机设备,其特征在于,所述处理器执行所述计算机可读指令时还执行以下步骤:在所述当前帧点云数据中每个点对应的二维数据中提取多个数据维度;及将多个数据维度输入至训练后的神经网络模型中,通过所述神经网络模型对多个数据维度进行预测运算,得到点云特征信息。The computer device according to claim 11, wherein the processor further executes the following step when executing the computer-readable instructions: extracting the two-dimensional data corresponding to each point in the current frame point cloud data Multiple data dimensions; and input the multiple data dimensions into the trained neural network model, and perform prediction operations on the multiple data dimensions through the neural network model to obtain point cloud feature information.
  14. 根据权利要求11所述的计算机设备,其特征在于,所述处理器执行所述计算机可读指令时还执行以下步骤:将所述融合后的特征信息输入至所 述检测模型的输入层;通过所述输入层将所述融合后的特征信息输入至所述检测模型的注意力层,通过所述注意力层计算所述融合后的特征信息对应的上下文向量与权重,生成第一提取结果;将所述第一提取结果输入卷积层,通过所述卷积层提取所述上下文向量对应的上下文特征,生成第二提取结果;将所述第二提取结果输入池化层,通过所述池化层对所述第二提取结果进行降维处理;将降维处理后的第二提取结果输入全连接层,通过所述全连接层对所述降维处理后的第二提取结果进行分类得到分类结果,通过输出层将所述分类结果进行加权后输出;及选取加权后输出的分类结果中权重最大的分类结果作为障碍物检测结果。The computer device according to claim 11, wherein the processor further executes the following steps when executing the computer-readable instructions: inputting the fused feature information into the input layer of the detection model; The input layer inputs the fused feature information to the attention layer of the detection model, and calculates the context vector and weight corresponding to the fused feature information through the attention layer to generate a first extraction result; The first extraction result is input to the convolutional layer, the context feature corresponding to the context vector is extracted through the convolutional layer to generate a second extraction result; the second extraction result is input to the pooling layer, and the second extraction result The transformation layer performs dimensionality reduction processing on the second extraction result; the second extraction result after the dimensionality reduction processing is input to the fully connected layer, and the second extraction result after the dimensionality reduction processing is classified through the fully connected layer to obtain The classification result is output after being weighted by the output layer; and the classification result with the largest weight among the weighted classification results is selected as the obstacle detection result.
  15. 根据权利要求11至14任意一项所述的计算机设备,其特征在于,所述处理器执行所述计算机可读指令时还执行以下步骤:调用多线程在每个视角对应的二维平面中并发提取每个视角对应的点云特征信息;及在所述将每个视角对应的点云特征信息和所述当前帧图像数据输入至对应的特征提取模型中之前,还包括:利用所述多线程将每个视角对应的点云特征信息和所述当前帧图像数据进行并行转换,得到每个视角对应的点云特征向量和所述当前帧图像数据对应的图像矩阵。The computer device according to any one of claims 11 to 14, wherein the processor further executes the following steps when executing the computer-readable instructions: invoking multiple threads to concurrently execute in the two-dimensional plane corresponding to each perspective Extracting point cloud feature information corresponding to each perspective; and before inputting the point cloud feature information corresponding to each perspective and the current frame image data into the corresponding feature extraction model, the method further includes: using the multithreading The point cloud feature information corresponding to each view angle and the current frame image data are converted in parallel to obtain the point cloud feature vector corresponding to each view angle and the image matrix corresponding to the current frame image data.
  16. 一个或多个存储有计算机可读指令的非易失性计算机可读存储介质,所述计算机可读指令被一个或多个处理器执行时,使得所述一个或多个处理器执行以下步骤:One or more non-volatile computer-readable storage media storing computer-readable instructions. When the computer-readable instructions are executed by one or more processors, the one or more processors execute the following steps:
    获取当前帧点云数据和预设角度范围内的当前帧图像数据;Obtain current frame point cloud data and current frame image data within a preset angle range;
    将所述当前帧点云数据在多个视角上进行投影,得到多个视角对应的二维平面;Projecting the point cloud data of the current frame on multiple viewing angles to obtain two-dimensional planes corresponding to the multiple viewing angles;
    对每个视角对应的二维平面进行特征提取,得到每个视角对应的点云特 征信息;Perform feature extraction on the two-dimensional plane corresponding to each perspective, and obtain the point cloud feature information corresponding to each perspective;
    将每个视角对应的点云特征信息和所述当前帧图像数据输入至对应的特征提取模型中,通过相应的特征提取模型并行提取每个视角对应的空间特征信息和所述当前帧图像数据对应的图像特征信息;The point cloud feature information corresponding to each perspective and the current frame image data are input into the corresponding feature extraction model, and the spatial feature information corresponding to each perspective and the current frame image data are extracted in parallel through the corresponding feature extraction model Image feature information;
    将多个视角对应的空间特征信息和所述图像特征信息进行融合,得到融合后的特征信息;及Fusing the spatial feature information corresponding to multiple viewing angles with the image feature information to obtain the fused feature information; and
    将所述融合后的特征信息输入至训练后的检测模型中,通过所述检测模型对所述融合后的特征信息进行预测运算,输出障碍物检测结果。The fused feature information is input into a trained detection model, and the fused feature information is predicted and calculated through the detection model, and an obstacle detection result is output.
  17. 根据权利要求16所述的存储介质,其特征在于,所述计算机可读指令被所述处理器执行时还执行以下步骤:根据预设参数将多个视角对应的空间特征信息和所述图像特征信息进行拼接,得到拼接后的特征信息;及根据所述预设参数将所述拼接后的特征信息进行对齐至预设视角,得到对齐后的特征信息,将所述对齐后的特征信息作为融合后的特征信息。The storage medium according to claim 16, wherein when the computer-readable instructions are executed by the processor, the following step is further executed: according to preset parameters, the spatial feature information corresponding to the multiple viewing angles and the image feature Information is spliced to obtain spliced feature information; and the spliced feature information is aligned to a preset viewing angle according to the preset parameters to obtain the aligned feature information, and the aligned feature information is used as the fusion After the feature information.
  18. 根据权利要求16所述的存储介质,其特征在于,所述计算机可读指令被所述处理器执行时还执行以下步骤:在所述当前帧点云数据中每个点对应的二维数据中提取多个数据维度;及将多个数据维度输入至训练后的神经网络模型中,通过所述神经网络模型对多个数据维度进行预测运算,得到点云特征信息。The storage medium according to claim 16, wherein when the computer-readable instructions are executed by the processor, the following steps are further executed: in the two-dimensional data corresponding to each point in the current frame point cloud data Extracting multiple data dimensions; and inputting the multiple data dimensions into the trained neural network model, and performing prediction operations on the multiple data dimensions through the neural network model to obtain point cloud feature information.
  19. 根据权利要求16所述的存储介质,其特征在于,所述计算机可读指令被所述处理器执行时还执行以下步骤:将所述融合后的特征信息输入至所述检测模型的输入层;通过所述输入层将所述融合后的特征信息输入至所述检测模型的注意力层,通过所述注意力层计算所述融合后的特征信息对应的上下文向量与权重,生成第一提取结果;将所述第一提取结果输入卷积层, 通过所述卷积层提取所述上下文向量对应的上下文特征,生成第二提取结果;将所述第二提取结果输入池化层,通过所述池化层对所述第二提取结果进行降维处理;将降维处理后的第二提取结果输入全连接层,通过所述全连接层对所述降维处理后的第二提取结果进行分类得到分类结果,通过输出层将所述分类结果进行加权后输出;及选取加权后输出的分类结果中权重最大的分类结果作为障碍物检测结果。The storage medium according to claim 16, wherein when the computer-readable instructions are executed by the processor, the following step is further executed: inputting the fused feature information into the input layer of the detection model; The fused feature information is input to the attention layer of the detection model through the input layer, and the context vector and weight corresponding to the fused feature information are calculated through the attention layer to generate a first extraction result Input the first extraction result into the convolutional layer, extract the context feature corresponding to the context vector through the convolutional layer, and generate a second extraction result; input the second extraction result into the pooling layer, and pass the The pooling layer performs dimensionality reduction processing on the second extraction result; the second extraction result after the dimensionality reduction processing is input to the fully connected layer, and the second extraction result after the dimensionality reduction processing is classified through the fully connected layer The classification result is obtained, and the classification result is weighted and output through the output layer; and the classification result with the largest weight among the weighted classification results is selected as the obstacle detection result.
  20. 根据权利要求16至19任意一项所述的存储介质,其特征在于,所述计算机可读指令被所述处理器执行时还执行以下步骤:调用多线程在每个视角对应的二维平面中并发提取每个视角对应的点云特征信息;及在所述将每个视角对应的点云特征信息和所述当前帧图像数据输入至对应的特征提取模型中之前,还包括:利用所述多线程将每个视角对应的点云特征信息和所述当前帧图像数据进行并行转换,得到每个视角对应的点云特征向量和所述当前帧图像数据对应的图像矩阵。The storage medium according to any one of claims 16 to 19, wherein when the computer-readable instructions are executed by the processor, the following steps are further executed: calling multi-threads in a two-dimensional plane corresponding to each view angle Concurrently extracting the point cloud feature information corresponding to each perspective; and before inputting the point cloud feature information corresponding to each perspective and the current frame image data into the corresponding feature extraction model, the method further includes: using the multiple The thread performs parallel conversion of the point cloud feature information corresponding to each view angle and the current frame image data to obtain the point cloud feature vector corresponding to each view angle and the image matrix corresponding to the current frame image data.
PCT/CN2019/130155 2019-12-30 2019-12-30 Obstacle detection method and apparatus based on driverless technology and computer device WO2021134325A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
PCT/CN2019/130155 WO2021134325A1 (en) 2019-12-30 2019-12-30 Obstacle detection method and apparatus based on driverless technology and computer device
CN201980037716.XA CN113678136B (en) 2019-12-30 2019-12-30 Obstacle detection method and device based on unmanned technology and computer equipment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2019/130155 WO2021134325A1 (en) 2019-12-30 2019-12-30 Obstacle detection method and apparatus based on driverless technology and computer device

Publications (1)

Publication Number Publication Date
WO2021134325A1 true WO2021134325A1 (en) 2021-07-08

Family

ID=76686174

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2019/130155 WO2021134325A1 (en) 2019-12-30 2019-12-30 Obstacle detection method and apparatus based on driverless technology and computer device

Country Status (2)

Country Link
CN (1) CN113678136B (en)
WO (1) WO2021134325A1 (en)

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113535877A (en) * 2021-07-16 2021-10-22 上海高仙自动化科技发展有限公司 Intelligent robot map updating method, device, equipment, medium and chip
CN113570707A (en) * 2021-07-30 2021-10-29 集美大学 Three-dimensional human body reconstruction method and device, computer equipment and storage medium
CN113724393A (en) * 2021-08-12 2021-11-30 北京达佳互联信息技术有限公司 Three-dimensional reconstruction method, device, equipment and storage medium
CN114173106A (en) * 2021-12-01 2022-03-11 北京拙河科技有限公司 Real-time video stream fusion processing method and system based on light field camera
CN114399738A (en) * 2021-12-29 2022-04-26 三一专用汽车有限责任公司 Target detection method and vehicle
CN114419372A (en) * 2022-01-13 2022-04-29 南京邮电大学 Multi-scale point cloud classification method and system
CN114429631A (en) * 2022-01-27 2022-05-03 北京百度网讯科技有限公司 Three-dimensional object detection method, device, equipment and storage medium
CN114511717A (en) * 2022-02-17 2022-05-17 广发银行股份有限公司 Method, system, device and medium for background similarity identification
CN114584850A (en) * 2022-03-09 2022-06-03 合肥工业大学 User visual angle prediction method for point cloud video streaming media transmission
CN114743001A (en) * 2022-04-06 2022-07-12 合众新能源汽车有限公司 Semantic segmentation method and device, electronic equipment and storage medium
CN114913373A (en) * 2022-05-12 2022-08-16 苏州轻棹科技有限公司 Image point cloud based sequence classification method and device
CN114972165A (en) * 2022-03-24 2022-08-30 中山大学孙逸仙纪念医院 Method and device for measuring time-average shearing force
CN115471561A (en) * 2022-11-14 2022-12-13 科大讯飞股份有限公司 Object key point positioning method, cleaning robot control method and related equipment
CN117274036A (en) * 2023-08-22 2023-12-22 合肥辉羲智能科技有限公司 Parking scene detection method based on multi-view and time sequence fusion
CN117456002A (en) * 2023-12-22 2024-01-26 珠海市格努科技有限公司 Method and device for estimating pose of object in disordered grabbing process and electronic equipment

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115761702B (en) * 2022-12-01 2024-02-02 广汽埃安新能源汽车股份有限公司 Vehicle track generation method, device, electronic equipment and computer readable medium

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140340518A1 (en) * 2013-05-20 2014-11-20 Nidec Elesys Corporation External sensing device for vehicle, method of correcting axial deviation and recording medium
CN107966700A (en) * 2017-11-20 2018-04-27 天津大学 A kind of front obstacle detecting system and method for pilotless automobile
CN108269281A (en) * 2016-12-30 2018-07-10 无锡顶视科技有限公司 Avoidance technical method based on binocular vision
CN109145677A (en) * 2017-06-15 2019-01-04 百度在线网络技术(北京)有限公司 Obstacle detection method, device, equipment and storage medium
US20190220650A1 (en) * 2015-08-24 2019-07-18 Qualcomm Incorporated Systems and methods for depth map sampling
CN110371108A (en) * 2019-06-14 2019-10-25 浙江零跑科技有限公司 Cartborne ultrasound wave radar and vehicle-mounted viewing system fusion method
CN110488805A (en) * 2018-05-15 2019-11-22 武汉小狮科技有限公司 A kind of unmanned vehicle obstacle avoidance system and method based on 3D stereoscopic vision

Family Cites Families (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106951847B (en) * 2017-03-13 2020-09-29 百度在线网络技术(北京)有限公司 Obstacle detection method, apparatus, device and storage medium
CN108229366B (en) * 2017-12-28 2021-12-14 北京航空航天大学 Deep learning vehicle-mounted obstacle detection method based on radar and image data fusion
CN109948448B (en) * 2019-02-20 2021-03-12 苏州风图智能科技有限公司 Method, device and system for detecting 3D obstacle and computer storage medium
CN109948661B (en) * 2019-02-27 2023-04-07 江苏大学 3D vehicle detection method based on multi-sensor fusion
CN110045729B (en) * 2019-03-12 2022-09-13 北京小马慧行科技有限公司 Automatic vehicle driving method and device
CN110223223A (en) * 2019-04-28 2019-09-10 北京清城同衡智慧园高新技术研究院有限公司 Street scan method, device and scanner
CN110363820B (en) * 2019-06-28 2023-05-16 东南大学 Target detection method based on laser radar and pre-image fusion
CN110458112B (en) * 2019-08-14 2020-11-20 上海眼控科技股份有限公司 Vehicle detection method and device, computer equipment and readable storage medium

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140340518A1 (en) * 2013-05-20 2014-11-20 Nidec Elesys Corporation External sensing device for vehicle, method of correcting axial deviation and recording medium
US20190220650A1 (en) * 2015-08-24 2019-07-18 Qualcomm Incorporated Systems and methods for depth map sampling
CN108269281A (en) * 2016-12-30 2018-07-10 无锡顶视科技有限公司 Avoidance technical method based on binocular vision
CN109145677A (en) * 2017-06-15 2019-01-04 百度在线网络技术(北京)有限公司 Obstacle detection method, device, equipment and storage medium
CN107966700A (en) * 2017-11-20 2018-04-27 天津大学 A kind of front obstacle detecting system and method for pilotless automobile
CN110488805A (en) * 2018-05-15 2019-11-22 武汉小狮科技有限公司 A kind of unmanned vehicle obstacle avoidance system and method based on 3D stereoscopic vision
CN110371108A (en) * 2019-06-14 2019-10-25 浙江零跑科技有限公司 Cartborne ultrasound wave radar and vehicle-mounted viewing system fusion method

Cited By (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113535877A (en) * 2021-07-16 2021-10-22 上海高仙自动化科技发展有限公司 Intelligent robot map updating method, device, equipment, medium and chip
CN113570707A (en) * 2021-07-30 2021-10-29 集美大学 Three-dimensional human body reconstruction method and device, computer equipment and storage medium
CN113724393A (en) * 2021-08-12 2021-11-30 北京达佳互联信息技术有限公司 Three-dimensional reconstruction method, device, equipment and storage medium
CN113724393B (en) * 2021-08-12 2024-03-19 北京达佳互联信息技术有限公司 Three-dimensional reconstruction method, device, equipment and storage medium
CN114173106B (en) * 2021-12-01 2022-08-05 北京拙河科技有限公司 Real-time video stream fusion processing method and system based on light field camera
CN114173106A (en) * 2021-12-01 2022-03-11 北京拙河科技有限公司 Real-time video stream fusion processing method and system based on light field camera
CN114399738A (en) * 2021-12-29 2022-04-26 三一专用汽车有限责任公司 Target detection method and vehicle
CN114419372A (en) * 2022-01-13 2022-04-29 南京邮电大学 Multi-scale point cloud classification method and system
CN114429631A (en) * 2022-01-27 2022-05-03 北京百度网讯科技有限公司 Three-dimensional object detection method, device, equipment and storage medium
CN114429631B (en) * 2022-01-27 2023-11-14 北京百度网讯科技有限公司 Three-dimensional object detection method, device, equipment and storage medium
CN114511717A (en) * 2022-02-17 2022-05-17 广发银行股份有限公司 Method, system, device and medium for background similarity identification
CN114584850A (en) * 2022-03-09 2022-06-03 合肥工业大学 User visual angle prediction method for point cloud video streaming media transmission
CN114584850B (en) * 2022-03-09 2023-08-25 合肥工业大学 User visual angle prediction method for point cloud video streaming media transmission
CN114972165B (en) * 2022-03-24 2024-03-15 中山大学孙逸仙纪念医院 Method and device for measuring time average shearing force
CN114972165A (en) * 2022-03-24 2022-08-30 中山大学孙逸仙纪念医院 Method and device for measuring time-average shearing force
CN114743001A (en) * 2022-04-06 2022-07-12 合众新能源汽车有限公司 Semantic segmentation method and device, electronic equipment and storage medium
CN114913373A (en) * 2022-05-12 2022-08-16 苏州轻棹科技有限公司 Image point cloud based sequence classification method and device
CN114913373B (en) * 2022-05-12 2024-04-09 苏州轻棹科技有限公司 Classification method and device for sequence based on image point cloud
CN115471561A (en) * 2022-11-14 2022-12-13 科大讯飞股份有限公司 Object key point positioning method, cleaning robot control method and related equipment
CN117274036A (en) * 2023-08-22 2023-12-22 合肥辉羲智能科技有限公司 Parking scene detection method based on multi-view and time sequence fusion
CN117456002A (en) * 2023-12-22 2024-01-26 珠海市格努科技有限公司 Method and device for estimating pose of object in disordered grabbing process and electronic equipment
CN117456002B (en) * 2023-12-22 2024-04-02 珠海市格努科技有限公司 Method and device for estimating pose of object in disordered grabbing process and electronic equipment

Also Published As

Publication number Publication date
CN113678136B (en) 2024-09-20
CN113678136A (en) 2021-11-19

Similar Documents

Publication Publication Date Title
WO2021134325A1 (en) Obstacle detection method and apparatus based on driverless technology and computer device
CN111860695B (en) Data fusion and target detection method, device and equipment
US20230316742A1 (en) Image processing method, apparatus and device, and computer-readable storage medium
JP6031554B2 (en) Obstacle detection method and apparatus based on monocular camera
US10990836B2 (en) Method and apparatus for recognizing object, device, vehicle and medium
CN111507166A (en) Method and apparatus for learning CNN by using camera and radar together
CN111797650A (en) Obstacle identification method and device, computer equipment and storage medium
CN111563415A (en) Binocular vision-based three-dimensional target detection system and method
CN110796104A (en) Target detection method and device, storage medium and unmanned aerial vehicle
CN115004259B (en) Object recognition method, device, computer equipment and storage medium
EP4040400A1 (en) Guided inspection with object recognition models and navigation planning
CN116469079A (en) Automatic driving BEV task learning method and related device
CN105303554B (en) The 3D method for reconstructing and device of a kind of image characteristic point
CN115701864A (en) Neural network training method, target detection method, device, medium, and product
Liu et al. Vehicle-related distance estimation using customized YOLOv7
CN116681739A (en) Target motion trail generation method and device and electronic equipment
CN117707194A (en) Unmanned equipment flight operation method and device, unmanned equipment and storage medium
CN117994748A (en) Road side aerial view target detection method and device, computing equipment and storage medium
US20230401748A1 (en) Apparatus and methods to calibrate a stereo camera pair
Unger et al. Multi-camera bird’s eye view perception for autonomous driving
Zhang et al. A vision-centric approach for static map element annotation
CN116894829A (en) Weld defect detection method, device, computer equipment and storage medium
CN111460854A (en) Remote target detection method, device and system
CN112433193B (en) Multi-sensor-based mold position positioning method and system
Oh et al. Automatic Pseudo-LiDAR Annotation: Generation of Training Data for 3D Object Detection Networks

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19958320

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 19958320

Country of ref document: EP

Kind code of ref document: A1