WO2021134258A1 - Point cloud-based target tracking method and apparatus, computer device and storage medium - Google Patents

Point cloud-based target tracking method and apparatus, computer device and storage medium Download PDF

Info

Publication number
WO2021134258A1
WO2021134258A1 PCT/CN2019/130034 CN2019130034W WO2021134258A1 WO 2021134258 A1 WO2021134258 A1 WO 2021134258A1 CN 2019130034 W CN2019130034 W CN 2019130034W WO 2021134258 A1 WO2021134258 A1 WO 2021134258A1
Authority
WO
WIPO (PCT)
Prior art keywords
current frame
point cloud
cloud data
feature map
feature
Prior art date
Application number
PCT/CN2019/130034
Other languages
French (fr)
Chinese (zh)
Inventor
许家妙
何明
叶茂盛
邹晓艺
吴伟
许双杰
曹通易
Original Assignee
深圳元戎启行科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 深圳元戎启行科技有限公司 filed Critical 深圳元戎启行科技有限公司
Priority to PCT/CN2019/130034 priority Critical patent/WO2021134258A1/en
Publication of WO2021134258A1 publication Critical patent/WO2021134258A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition

Definitions

  • This application relates to a point cloud-based target tracking method, device, computer equipment, storage medium, and transportation.
  • the automatic driving equipment can be controlled by tracking the surrounding target objects or pedestrians.
  • target tracking is usually based on captured images.
  • the inventor realizes that target tracking based on images is easily affected by image quality.
  • the image quality is lower when the ambient light changes, the target moves faster, and so on. Based on low-quality images, the target cannot be tracked accurately, and the accuracy of the tracking result is low.
  • a point cloud-based target tracking method device, computer equipment, storage medium, and transportation tool are provided.
  • a point cloud-based target tracking method includes:
  • the target tracking model trained based on the point cloud data of the previous frame is called, and the target tracking area corresponding to the point cloud data of the current frame is determined according to the image characteristics.
  • a point cloud-based target tracking device includes:
  • the feature map generating module is used to generate the feature map of the current frame according to the point cloud data of the current frame;
  • a candidate region acquiring module configured to acquire a candidate region corresponding to the current frame feature map; intercept a candidate feature map matching the candidate region in the current frame feature map;
  • a feature extraction module for extracting image features corresponding to the candidate feature map
  • the target tracking module is used to call the target tracking model trained based on the point cloud data of the previous frame, and determine the target tracking area corresponding to the point cloud data of the current frame according to the image characteristics.
  • a computer device including a memory and one or more processors, the memory stores computer readable instructions, and when the computer readable instructions are executed by the processor, the one or more processors execute The following steps:
  • the target tracking model trained based on the point cloud data of the previous frame is called, and the target tracking area corresponding to the point cloud data of the current frame is determined according to the image characteristics.
  • One or more non-volatile computer-readable storage media storing computer-readable instructions.
  • the computer-readable instructions When executed by one or more processors, the one or more processors perform the following steps:
  • the target tracking model trained based on the point cloud data of the previous frame is called, and the target tracking area corresponding to the point cloud data of the current frame is determined according to the image characteristics.
  • a vehicle includes the steps of executing the above-mentioned point cloud-based target tracking method.
  • Fig. 1 is an application scenario diagram of a point cloud-based target tracking method according to one or more embodiments.
  • Fig. 2 is a schematic flowchart of a point cloud-based target tracking method according to one or more embodiments.
  • Fig. 3 is a schematic flowchart of the steps of generating a feature map of the current frame according to the point cloud data of the current frame according to one or more embodiments.
  • Fig. 4 is a schematic flowchart of a point cloud-based target tracking method in another embodiment.
  • Fig. 5 is a block diagram of a point cloud-based target tracking device according to one or more embodiments.
  • Figure 6 is a block diagram of a computer device according to one or more embodiments.
  • the point cloud-based target tracking method provided in this application can be applied to the application environment of automatic driving as shown in FIG. 1.
  • the laser sensor 102 can communicate with the computer device 104.
  • the laser sensor 102 may be a vehicle-mounted laser sensor, and the computer device 104 may be a vehicle-mounted computer device.
  • the point cloud data can be collected by the laser sensor 102, or pre-stored by a computer device.
  • the computer device 104 generates a feature map of the current frame according to the point cloud data of the current frame.
  • the computer device 104 obtains the candidate region corresponding to the feature map of the current frame, and intercepts the candidate feature map matching the candidate region in the feature map of the current frame.
  • the computer device 104 extracts the image features corresponding to the candidate feature map, calls the target tracking model trained based on the point cloud data of the previous frame, and determines the target tracking area corresponding to the point cloud data of the current frame according to the image features.
  • the laser sensor 102 may be a laser sensor carried by an automatic driving device, and may specifically include a laser radar, a laser scanner, and the like.
  • a point cloud-based target tracking method is provided. Taking the method applied to the computer device 104 in FIG. 1 as an example for description, the method includes the following steps:
  • Step 202 Generate a feature map of the current frame according to the point cloud data of the current frame.
  • the laser sensor is equipped with a device capable of autonomous driving.
  • a device capable of autonomous driving may be a laser sensor mounted on an unmanned vehicle, or a laser sensor mounted on a vehicle including an automatic driving mode.
  • Laser sensors can be used to collect surrounding environmental data.
  • the laser sensor can emit a detection signal, such as a laser beam.
  • the laser sensor compares the reflected signal with the detection signal to obtain surrounding environmental data.
  • the environmental data may specifically be point cloud data.
  • Point cloud data refers to a collection of point data corresponding to multiple points on the surface of the object in the scanning environment recorded in the form of points. Multiple can refer to two or more than two.
  • the laser sensor can collect according to a preset frequency to obtain multi-frame point cloud data.
  • the preset frequency can be preset according to actual needs.
  • the point cloud data may be three-dimensional point cloud data, and each frame of point cloud data may include point data corresponding to multiple points.
  • the point data may specifically include at least one of three-dimensional coordinates, laser reflection intensity, and color information corresponding to the point.
  • the three-dimensional coordinates may be the coordinates of the point in the Cartesian coordinate system, and specifically include the horizontal axis coordinates, the vertical axis coordinates, and the vertical axis coordinates of the point in the Cartesian coordinate system.
  • the Cartesian coordinate system is a three-dimensional space coordinate system established with a laser sensor as the origin.
  • the three-dimensional space coordinate system includes a horizontal axis (x axis), a vertical axis (y axis), and a vertical axis (z axis).
  • the three-dimensional space coordinate system established with the laser sensor as the origin satisfies the right-hand rule.
  • the computer equipment can follow the time sequence of the point cloud data collected by the laser sensor, and track the target according to the multi-frame point cloud data in turn.
  • a target refers to a living or non-living body in the surrounding environment.
  • the target can be moving or stationary.
  • the target may specifically include at least one of pedestrians, roadblocks, vehicles, and buildings.
  • the current frame of point cloud data refers to a frame of point cloud data being processed by the computer equipment. It is understandable that when the computer device finishes tracking the point cloud data of the current frame and starts to track the point cloud data of the next frame, the point cloud data of the current frame can be recorded as the point cloud data of the previous frame, and the point cloud data of the next frame can be recorded as the point cloud data of the previous frame.
  • the data is re-recorded as the point cloud data of the current frame.
  • the computer device can obtain the point data included in the point cloud data of the current frame, encode the points according to the point data, and obtain the point features corresponding to each of the multiple points.
  • the point feature can be expressed in the form of a vector, and the point feature of each point can include a point vector.
  • the point vector can be a multidimensional vector.
  • the computer device can generate a feature map according to the point features corresponding to each of the multiple points, and the feature map generated from the point cloud data of the current frame can be recorded as the current frame feature map.
  • the feature map generated from the point cloud data in this implementation does not include RGB (red, green, blue) channels.
  • the generated feature map is different from the traditionally collected image, based on the point cloud data Target tracking will not be affected by factors such as ambient lighting and target movement speed.
  • Step 204 Obtain a candidate region corresponding to the feature map of the current frame.
  • the candidate area is a feature map area used to determine the target tracking area, and the candidate area may include the target tracking area.
  • the candidate area is an area range in the feature map of the current frame, and the candidate area can be the entire area of the feature map of the current frame, or a partial area of the feature map of the current frame.
  • the candidate area may include the area size, the area shape, and the position in the current frame feature map.
  • the candidate area may be one of a variety of shapes. For example, in principle, the shape of the candidate region can generally be rectangular, and the shape of the candidate region can also be circular.
  • the computer device After the computer device generates the feature map of the current frame according to the point cloud data of the current frame, it can obtain the candidate region corresponding to the feature map of the current frame in a variety of ways. For example, the computer device may record the entire area of the feature map of the current frame as the corresponding candidate area. The computer device can also obtain the candidate area corresponding to the feature map of the current frame according to the point cloud data of the previous frame. Specifically, the computer device can obtain the target area of the previous frame corresponding to the point cloud data of the previous frame according to the tracking result of the point cloud data of the previous frame, and the computer device can determine the candidate area corresponding to the feature map of the current frame according to the target area of the previous frame. .
  • obtaining the candidate area corresponding to the feature map of the current frame includes: obtaining the target area of the previous frame corresponding to the point cloud data of the previous frame; expanding the target area of the previous frame according to a preset multiple; The target area of the previous frame is used as the candidate area corresponding to the feature map of the current frame.
  • the computer device may obtain the target area of the previous frame corresponding to the point cloud data of the previous frame according to the tracking result of the point cloud data of the previous frame.
  • the target area of the previous frame refers to the area where the target is located in the point cloud data of the previous frame.
  • the computer device can expand the target area of the previous frame according to the preset multiple to obtain the expanded target area of the previous frame, so as to ensure that the current frame feature map in the candidate area can include the target.
  • the computer equipment can expand the target area of the previous frame with a preset multiple area, or expand the side length of the target area according to the preset multiple, and determine the closed area formed by the expanded side length as the expanded target area of the previous frame .
  • the computer device may determine the expanded previous frame target area in the current frame feature map according to the expanded previous frame target area, and determine the expanded previous frame target area as the candidate area corresponding to the current frame feature map.
  • the preset multiple may be preset by the user according to actual needs, for example, the preset multiple may specifically be 2 times.
  • the candidate area corresponding to the feature map of the current frame is determined according to the target area of the previous frame, so that the target of the previous frame is tracked in the feature map of the current frame.
  • the computer device can expand the target area of the previous frame according to the preset multiple, and determine the expanded target area of the previous frame as the candidate area corresponding to the feature map of the current frame, which ensures the accuracy of target tracking and does not need to process the entire current frame
  • the feature map effectively saves the computing resources of the computer equipment.
  • Step 206 Extract a candidate feature map matching the candidate region in the current frame feature map.
  • the computer device can intercept the candidate feature map in the current frame feature map according to the candidate region corresponding to the current frame feature map, and intercept the candidate feature map corresponding to the candidate region.
  • the candidate feature map can be all feature maps of the current frame feature map, or part of the feature map of the current frame feature map.
  • the candidate feature map may include the target to be tracked, and the candidate feature map obtained by interception matches the size and shape of the candidate area.
  • Step 208 Extract image features corresponding to the candidate feature map.
  • the computer device can perform feature extraction on the selected candidate feature map, and extract the image features corresponding to the candidate feature map from the candidate feature map.
  • the image feature corresponding to the candidate feature map may include at least one of multiple feature types.
  • the image features may specifically include at least one of multiple feature types such as edge features, corner features, and regional features.
  • Image features can be recorded by means of feature vectors, etc.
  • the computer device can extract the image features corresponding to the candidate feature map, perform feature analysis on the candidate feature map according to the image features, and determine the target tracking area corresponding to the point cloud data of the current frame in the candidate feature map according to the analysis result.
  • extracting the image features corresponding to the candidate feature map includes: obtaining a feature extraction model; inputting the candidate feature map to the feature extraction model; performing feature extraction on the candidate feature map according to the feature extraction model to obtain the candidate feature map location Corresponding image characteristics.
  • the computer device can obtain the feature extraction model, and the feature extraction model can be pre-configured in the computer device.
  • the computer device can input the intercepted candidate feature maps into the feature extraction model, and perform operations on the input candidate feature maps through the feature extraction model to perform feature extraction on the candidate feature maps.
  • the feature extraction model can be one of a variety of neural network models.
  • the feature extraction model may specifically be one of neural network models such as a traditional Convolutional Neural Networks (CNN) model and a VGG (Visual Geometry Group Network) model.
  • the feature extraction model may specifically include an input layer, a convolutional layer, a pooling layer, a fully connected layer, a BN (Batch Normalization, batch normalization) layer, an output layer, and so on.
  • the computer device can sequentially perform operations corresponding to the network structure on the candidate feature maps according to the network structure of the feature extraction model to obtain image features corresponding to the candidate feature maps output by the feature extraction model.
  • the computer device can obtain the feature extraction model, input the candidate feature map into the feature extraction model, perform feature extraction on the candidate feature map according to the feature extraction model, and obtain the image features corresponding to the candidate feature map, which effectively improves The accuracy of image features. Furthermore, the target tracking area corresponding to the point cloud data of the current frame is determined according to the image characteristics, which improves the accuracy of target tracking.
  • Step 210 Call the target tracking model trained based on the point cloud data of the previous frame, and determine the target tracking area corresponding to the point cloud data of the current frame according to the image characteristics.
  • the computer device can call the target tracking model, and perform tracking processing on the image features corresponding to the candidate feature map according to the target tracking model to obtain the target tracking area corresponding to the point cloud data of the current frame.
  • the target tracking model is a tracking model obtained after training based on the point cloud data of the previous frame.
  • the target tracking model is used to track the area where the target is located in the point cloud data of the current frame to obtain the target tracking area.
  • the target tracking area refers to the location area of the target in the feature map of the current frame estimated by the tracking process.
  • the computer device can input the extracted image features into the target tracking model trained based on the point cloud data of the previous frame, and use the target tracking model to track the image features of the current frame feature map to obtain the target tracking area output by the target tracking model.
  • the computer device can obtain multiple frames of point cloud data, and process each frame of point cloud data in sequence according to the sequence of the point cloud data collection time of each frame.
  • the tracking model can be trained according to the point cloud data of the previous frame corresponding to the point cloud data of the current frame to obtain the target tracking model corresponding to the point cloud data of the current frame.
  • the computer equipment After the computer equipment finishes processing the point cloud data of the current frame, it can perform tracking processing on the point cloud data of the next frame.
  • the computer equipment can train the tracking model according to the point cloud data of the current frame, and obtain the target tracking model corresponding to the point cloud data of the next frame, so as to track the point cloud data of the next frame.
  • the computer equipment can process the point cloud data according to the point cloud data. Train the tracking model sequentially iteratively.
  • the computer device can obtain the point cloud data of the current frame, and generate a feature map of the current frame according to the point cloud data of the current frame.
  • the computer device can extract the image features of the candidate feature map in the current frame feature map, and track the image features according to the target tracking model trained based on the point cloud data of the previous frame to determine the target tracking area corresponding to the current frame point cloud data.
  • the target tracking based on point cloud data in this embodiment will not be affected by factors such as external light, target movement speed, etc., which effectively improves the accuracy of target tracking.
  • generating a feature map of the current frame according to the point cloud data of the current frame includes:
  • Step 302 Obtain the point cloud data of the current frame.
  • Step 304 Perform structured processing on the point cloud data of the current frame to obtain a processing result.
  • Step 306 Encode the points in the point cloud data of the current frame based on the processing result to obtain point features corresponding to the points.
  • Step 308 Generate a feature map of the current frame corresponding to the point cloud data of the current frame according to the point features.
  • the computer equipment can obtain the point cloud data of multiple frames within the visible range collected by the laser sensor, and process the point cloud data of each frame in sequence according to the time sequence of the point cloud data collected by the laser sensor.
  • the computer device may record the point cloud data that is being processed or being processed as the point cloud data of the current frame.
  • the computer device can process the point cloud data of the current frame to generate a feature map of the current frame corresponding to the point cloud data of the current frame.
  • the computer device may perform structured processing on the point cloud data of the current frame to obtain a processing result after the structured processing.
  • the computer device may perform rasterization processing on the current frame point cloud data, or may perform voxelization processing on the current frame point cloud data.
  • the computer equipment can rasterize the plane with the laser sensor as the origin, and divide the plane into multiple grids.
  • the structured space after the structuring process may be a columnar space, and the points may be distributed in the columnar space corresponding to the vertical axis of the grid, that is, the abscissa and ordinate of the points in the columnar space are within the corresponding grid coordinate range.
  • the computer device can encode the point in the point cloud data of the current frame according to the processing result of the structured processing to obtain the point feature corresponding to the point.
  • the point feature can be a point vector corresponding to the point.
  • the computer device can count the point data of all points in each structured space, encode the points according to the statistical point data, and obtain the point vectors corresponding to the points.
  • the computer equipment After the computer equipment rasterizes the origin plane, it counts the point data in each columnar space.
  • the point data may specifically include the three-dimensional coordinates and reflection coefficients corresponding to each point.
  • the computer equipment can encode the point according to the point data in the columnar space to obtain the point vector corresponding to the point.
  • the point vector may be a 9-dimensional vector.
  • the point vector may specifically include the horizontal axis coordinate, the vertical axis coordinate, the vertical axis coordinate, the reflection coefficient, the distance from the center of the cylindrical space, and the distance between each point and the average value of the three-dimensional coordinates of all points.
  • the distance between a point and the center of the columnar space can be represented by the horizontal axis distance and the vertical axis distance
  • the distance between the point and the average of the three-dimensional coordinates of all points can be represented by the horizontal axis distance, the vertical axis distance, and the vertical axis distance.
  • the computer equipment can record the 9-dimensional vector corresponding to the point as the point feature corresponding to the point.
  • the computer device can count the point features in multiple structured spaces, and generate the current frame feature map corresponding to the current frame point cloud data according to the multiple point features.
  • the computer device can encode the points according to the point cloud data of the current frame to obtain the point characteristics, and generate the current frame feature map corresponding to the point cloud data of the current frame according to the point characteristics.
  • the point cloud data will not be affected by ambient light or target The influence of factors such as movement speed ensures the accuracy of tracking the target.
  • the current frame feature map is generated according to the current frame point cloud data in this implementation, which can effectively use the depth features in the point cloud data, and then Effectively improve the accuracy of target tracking.
  • the computer device may collect a preset number of sampling points from the structured space, encode the sampling points, and obtain the point characteristics corresponding to the sampling points.
  • the computer device can generate the current frame feature map corresponding to the current frame point cloud data according to the point features corresponding to the sampling points in each structured space. Specifically, the computer device can count the number of points included in the structured space, and compare the number of points with the preset number.
  • the preset number can be preset according to actual needs and historical point cloud data after big data analysis. When the number of points in the structured space is greater than or equal to the preset number, the computer device may randomly select a preset number of points from the structured space as sampling points.
  • the computer device can obtain all points in the structured space as sampling points, and add virtual points as sampling points, so that the preset number of sampling points can be collected.
  • the three-dimensional coordinates of the virtual point may be located at the origin of the coordinate system.
  • the computer device collects a preset number of points from each structured space as sampling points, so that the number of sampling points in each structured space is the same, thereby balancing the point characteristics of multiple structured spaces , which helps computer equipment to generate a feature map of the current frame based on structured data.
  • generating the current frame feature map corresponding to the point cloud data of the current frame according to the point features includes: generating a point feature matrix based on multiple point features; calling the image generation model, and inputting the point feature matrix to the image generation model; Obtain the current frame feature map output by the image generation model.
  • the computer device may generate a point feature matrix based on multiple point features in multiple structured spaces, and the point feature matrix may specifically include point features, structured spaces, and corresponding points, etc.
  • the computer equipment can call the image generation model.
  • the image generation model may be pre-configured in the computer device, and the image generation model may be obtained by training a large number of point feature samples and feature map samples corresponding to the point feature samples.
  • the image generation model can be one of a variety of neural network models.
  • the image generation model may be a convolutional neural network model, and specifically may be a PointNet model.
  • the computer device can input the generated point feature matrix to the image generation model, and calculate the point feature matrix through the image generation model, and perform the maximum pooling operation on the point features of the point quantity dimension to obtain the current frame feature corresponding to the current frame point cloud data Figure.
  • the computer device generates a point feature matrix based on multiple point features, and performs operations on the point feature matrix according to the image generation model to obtain the current frame feature map output by the image generation model.
  • the computer equipment generates a feature map of the current frame according to the point cloud data of the current frame, which effectively utilizes the depth features in the point cloud data, and improves the accuracy of target tracking based on the point cloud data.
  • calling the target tracking model trained based on the point cloud data of the previous frame, and determining the target tracking area corresponding to the point cloud data of the current frame according to the image features includes: generating an image feature matrix according to the image features; The matrix is input to the target tracking model, and the area label output by the target tracking model is obtained; the target tracking area corresponding to the point cloud data of the current frame is determined according to the area label.
  • the computer device can generate a corresponding image feature matrix according to the extracted image features. Specifically, the computer device can process the extracted image features, and after each column of image features are connected to the previous column of image features, a column vector is generated according to the image features. The computer device can cyclically shift the column vector, and arrange all the image features obtained after the shift in columns to obtain an image feature matrix. Cyclic shifting may include rotating left or rotating right.
  • the computer device can input the generated image feature matrix to the target tracking model, which is obtained after training based on the point cloud data of the previous frame.
  • the target tracking model can use at least one of a variety of visual tracking algorithms. For example, KCF (Kernel Correlation Filter, kernel correlation filtering algorithm) algorithm, and KCF-based target tracking algorithm, etc. can be specifically used.
  • KCF Kernel Correlation Filter, kernel correlation filtering algorithm
  • KCF-based target tracking algorithm etc.
  • the computer device can perform operations on the image feature matrix through the target tracking model, and obtain the area label output by the target tracking model after the calculation.
  • the target tracking model can output multiple area labels, which are used to mark corresponding areas.
  • the area marked by the area label indicates the possible range of the target, and the size and shape of the area can be the same as the target area of the previous frame corresponding to the point cloud data of the previous frame.
  • the area label can indicate the possibility that the target is within the corresponding area. In one of the embodiments, the area label
  • the computer device can compare multiple area labels output by the target tracking model, and obtain the area label with the largest label value from the multiple area labels as the target area label.
  • the computer device can determine the area corresponding to the target area tag as the target tracking area.
  • the computer device can obtain the cyclic offset corresponding to the target area label, determine the area after the offset of the target area of the previous frame according to the cyclic offset, and determine the area after the offset of the target area of the previous frame as the current frame point Target tracking area corresponding to cloud data.
  • the computer device can generate an image feature matrix based on image features, call the target tracking model to perform operations on the image feature matrix, and determine the target tracking area corresponding to the point cloud data of the current frame according to the area label output by the target tracking model.
  • this embodiment uses the point cloud features corresponding to the point cloud data of the current frame, which effectively improves the accuracy of target tracking based on point cloud data. Sex.
  • the above point cloud-based target tracking method before the step of calling the target tracking model trained based on the point cloud data of the previous frame, the above point cloud-based target tracking method further includes:
  • Step 402 Generate a feature map of the previous frame according to the point cloud data of the previous frame.
  • Step 404 Take a sample of the feature map from the previous frame of feature map, and extract the sample feature corresponding to the sample feature map.
  • Step 406 Generate a sample label corresponding to the sample feature.
  • Step 408 Train the standard tracking model according to the sample features and sample labels to obtain the target tracking model.
  • the computer device Before calling the target tracking model to process the image features corresponding to the point cloud data of the current frame, the computer device also needs to train the standard tracking model according to the point cloud data of the previous frame to obtain the target tracking model.
  • the computer device may generate a feature map of the previous frame according to the point cloud data of the previous frame, cut a sample of the feature map of the previous frame of the feature map, and extract the sample feature corresponding to the sample feature map. It is understandable that since the computer device can track the point cloud data collected by the laser sensor in turn, iteratively trains the target tracking model based on multiple frames of point cloud data. Therefore, the computer device generates the feature map of the previous frame according to the point cloud data of the previous frame, cuts samples of the feature map of the previous frame of the feature map, and extracts the sample feature corresponding to the sample feature map. The device generates a feature map of the current frame according to the point cloud data of the current frame, intercepts candidate feature maps around the current frame feature map, and extracts image features in the candidate feature map in the same or similar manner, so we will not repeat them here.
  • the computer device can generate the sample label corresponding to the sample characteristic according to the sample characteristic. Specifically, the computer device may perform a cyclic shift on the sample features to obtain multiple sample features, and generate a sample feature matrix based on the multiple sample features. Determine the sample label corresponding to the sample feature according to the shift value corresponding to each sample feature in the sample feature matrix.
  • the shift value may specifically include a horizontal axis shift value and a vertical axis shift value.
  • the computer device can perform operations on the shift values corresponding to the multiple sample features in the sample feature matrix according to the preset function to obtain sample labels corresponding to each of the multiple sample features.
  • the preset function may be a function preset by the user, and the preset function may specifically be a two-dimensional Gaussian function.
  • the computer device can train the established standard tracking model according to the multiple sample features in the sample feature matrix and the sample labels corresponding to the multiple sample features to obtain the target tracking model.
  • the computer device can train the standard tracking model based on the point cloud data of the previous frame to obtain the target tracking model, so as to use the target tracking model to track the target tracking area of the point cloud data of the current frame, and perform multi-frame points.
  • Cloud data iteratively trains the target tracking model, which effectively improves the accuracy of target tracking.
  • the above-mentioned point cloud-based target tracking method further includes detecting the point cloud data of the current frame to obtain the target detection area; according to the target detection area and the target tracking area, determining the current frame point cloud data corresponding to the current Frame target area.
  • the computer equipment can detect the point cloud data of the current frame according to the point cloud data to obtain the target detection area.
  • the computer device may use at least one of multiple target detection algorithms to perform target detection on the point cloud data of the current frame to obtain the target detection area.
  • the computer device can determine the current frame target area corresponding to the current frame point cloud data according to the target detection area and the target tracking area corresponding to the current frame point cloud data. Specifically, the computer device can compare the target detection area with the target tracking area. When the target detection area is the same as the target tracking area, the computer device can determine the area corresponding to the target detection area as the current frame target area corresponding to the current frame point cloud data. When the target detection area and the target tracking area are not the same, the computer device can synthesize the target detection area and the target tracking area, and determine the comprehensive area as the current frame target area corresponding to the current frame point cloud data.
  • the computer device can obtain the detection confidence level corresponding to the target detection area and the tracking confidence level corresponding to the target tracking area.
  • the computer device may integrate the target detection area and the target tracking area based on the detection confidence and the tracking confidence, and determine the comprehensive area as the current frame target area corresponding to the current frame point cloud data.
  • the computer device can detect the target detection area according to the current frame point cloud data, adjust the target detection area according to the target tracking area, and determine that the adjusted area is the current frame target area corresponding to the current frame point cloud data. Effectively improve the accuracy of determining the target area.
  • the above-mentioned point cloud-based target tracking method further includes determining target displacement data according to the target area of the current frame and the target area of the previous frame; acquiring the point cloud acquisition frequency; and determining the target according to the point cloud acquisition frequency and the target displacement data Movement data.
  • the computer device can compare the target area of the current frame with the target area of the previous frame, and determine the target displacement data according to the comparison result.
  • the target displacement data may include the length and direction of the target displacement.
  • the computer equipment can obtain the point cloud collection frequency corresponding to the laser sensor.
  • the point cloud collection frequency can be preset by the user according to actual needs, and the laser sensor collects point cloud data according to the set point cloud collection frequency.
  • the point cloud collection frequency can be a constant.
  • a laser sensor can collect point cloud data at a frequency of 50 frames per second.
  • the point cloud collection frequency can also be a variable.
  • the laser sensor can adjust the point cloud collection frequency according to different situations or modes. For example, a laser sensor can increase the point cloud collection frequency when there are many targets in the environment and the movement speed is fast, and reduce the point cloud collection frequency when there are fewer targets in the environment and the movement speed is slow.
  • the computer device can determine the time difference between the acquisition time of the point cloud data of the previous frame and the acquisition time of the point cloud data of the current frame according to the acquired point cloud acquisition frequency. For example, when the point cloud acquisition frequency is 50 frames per second, the computer device can determine that the time difference between the two frames is 0.02 seconds.
  • the computer device can determine the target motion data corresponding to the target according to the time difference and the target displacement data.
  • the target motion data may specifically include information such as the motion speed and direction corresponding to the target, so that the computer equipment can prompt or control the unmanned driving device according to the target motion data.
  • the computer device can determine the target displacement data according to the target area of the current frame and the target area of the previous frame, and determine the target motion data according to the point cloud collection frequency and the target displacement data, which helps the computer device to determine the target motion data according to the target motion data.
  • the person drives the device to prompt or control.
  • a point cloud-based target tracking device including: a feature map generation module 502, a candidate region acquisition module 504, a feature extraction module 506, and a target tracking module 508, where :
  • the feature map generating module 502 is configured to generate a feature map of the current frame according to the point cloud data of the current frame.
  • the candidate region acquiring module 504 is used to acquire the candidate region corresponding to the feature map of the current frame; and intercept the candidate feature map matching the candidate region in the feature map of the current frame.
  • the feature extraction module 506 is used to extract image features corresponding to the candidate feature map.
  • the target tracking module 508 is configured to call the target tracking model trained based on the point cloud data of the previous frame, and determine the target tracking area corresponding to the point cloud data of the current frame according to the image characteristics.
  • the above-mentioned feature map generation module 502 is also used to obtain frame point cloud data; structure the current frame point cloud data to obtain the processing result; perform processing on the points in the current frame point cloud data based on the processing result Encode to obtain the point feature corresponding to the point; generate the current frame feature map corresponding to the point cloud data of the current frame according to the point feature.
  • the above-mentioned feature map generation module 502 is further configured to generate a point feature matrix based on multiple point features; call the image generation model, and input the point feature matrix to the image generation model; obtain the current frame features output by the image generation model Figure.
  • the above-mentioned candidate area acquisition module 504 is also used to acquire the target area of the previous frame corresponding to the point cloud data of the previous frame; expand the target area of the previous frame according to a preset multiple; and determine the expanded previous The frame target area is used as the candidate area corresponding to the feature map of the current frame.
  • the feature extraction module 506 is also used to obtain a feature extraction model; input the candidate feature map to the feature extraction model; perform feature extraction on the candidate feature map according to the feature extraction model to obtain the image corresponding to the candidate feature map feature.
  • the above-mentioned target tracking module 508 is also used to generate an image feature matrix based on image features; input the image feature matrix to the target tracking model to obtain the area label output by the target tracking model; determine the current frame point cloud according to the area label Target tracking area corresponding to the data.
  • the above-mentioned point cloud-based target tracking device further includes a model training module, which is used to generate a feature map of the previous frame according to the point cloud data of the previous frame; Sample features corresponding to the sample feature map; generate sample labels corresponding to the sample features; train the standard tracking model according to the sample features and sample labels to obtain the target tracking model.
  • a model training module which is used to generate a feature map of the previous frame according to the point cloud data of the previous frame.
  • the above-mentioned point cloud-based target tracking device further includes a target area determining module for detecting the point cloud data of the current frame to obtain the target detection area; determining the current frame according to the target detection area and the target tracking area The target area of the current frame corresponding to the point cloud data.
  • the above-mentioned point cloud-based target tracking device further includes a target data determining module for determining target displacement data according to the target area of the current frame and the target area of the previous frame; acquiring the point cloud collection frequency; The frequency and target displacement data determine the target motion data.
  • Each module in the above-mentioned point cloud-based target tracking device can be implemented in whole or in part by software, hardware, and a combination thereof.
  • the above-mentioned modules may be embedded in the form of hardware or independent of the processor in the computer equipment, or may be stored in the memory of the computer equipment in the form of software, so that the processor can call and execute the operations corresponding to the above-mentioned modules.
  • a computer device is provided.
  • the computer device may be a server, and its internal structure diagram may be as shown in FIG. 6.
  • the computer equipment includes a processor, a memory, a network interface, and a database connected through a system bus. Among them, the processor of the computer device is used to provide calculation and control capabilities.
  • the memory of the computer device includes a non-volatile storage medium and an internal memory.
  • the non-volatile storage medium stores an operating system, computer readable instructions, and a database.
  • the internal memory provides an environment for the operation of the operating system and computer-readable instructions in the non-volatile storage medium.
  • the database of the computer equipment is used to store the target tracking data based on the point cloud.
  • the network interface of the computer device is used to communicate with an external terminal through a network connection.
  • the computer-readable instructions are executed by the processor to realize a point cloud-based target tracking method.
  • FIG. 6 is only a block diagram of part of the structure related to the solution of the present application, and does not constitute a limitation on the computer device to which the solution of the present application is applied.
  • the specific computer device may Including more or fewer parts than shown in the figure, or combining some parts, or having a different arrangement of parts.
  • a computer device including a memory and one or more processors.
  • the memory stores computer-readable instructions.
  • the one or more processors When executed, the steps in the above method embodiments are implemented.
  • one or more non-volatile computer-readable storage media storing computer-readable instructions are provided.
  • the computer-readable instructions are executed by one or more processors, one or more processing
  • the steps in the above method embodiments are implemented when the device is executed.
  • a vehicle is provided.
  • the vehicle may specifically include self-driving vehicles, electric vehicles, bicycles, and aircraft.
  • the vehicle includes the above-mentioned computer equipment and can execute the above-mentioned embodiment of the point cloud-based target tracking method. Steps in.
  • the embodiments and implementation objects created by the present invention are not limited to autonomous vehicles, electric vehicles, bicycles, aircrafts, robots, etc., but also include simulation devices and test equipment related to these devices.
  • Non-volatile memory may include read only memory (ROM), programmable ROM (PROM), electrically programmable ROM (EPROM), electrically erasable programmable ROM (EEPROM), or flash memory.
  • Volatile memory may include random access memory (RAM) or external cache memory.
  • RAM is available in many forms, such as static RAM (SRAM), dynamic RAM (DRAM), synchronous DRAM (SDRAM), double data rate SDRAM (DDRSDRAM), enhanced SDRAM (ESDRAM), synchronous chain Channel (Synchlink) DRAM (SLDRAM), memory bus (Rambus) direct RAM (RDRAM), direct memory bus dynamic RAM (DRDRAM), and memory bus dynamic RAM (RDRAM), etc.

Abstract

A point-cloud based target tracking method, comprising: according to point cloud data of a current frame, generating a feature map of the current frame; acquiring a candidate region corresponding to the feature map of the current frame; in the feature map of the current frame, intercepting a candidate feature map that matches the candidate region; extracting an image feature corresponding to the candidate feature map; and calling a target tracking model obtained by training on the basis of point cloud data of the previous frame, and according to the image feature, determining a target tracking region corresponding to the point cloud data of the current frame.

Description

基于点云的目标跟踪方法、装置、计算机设备和存储介质Point cloud-based target tracking method, device, computer equipment and storage medium 技术领域Technical field
本申请涉及一种基于点云的目标跟踪方法、装置、计算机设备、存储介质和交通工具。This application relates to a point cloud-based target tracking method, device, computer equipment, storage medium, and transportation.
背景技术Background technique
随着计算机技术的发展,出现了视觉跟踪技术,通过计算机的视觉跟踪技术可以对目标的位置、速度等信息进行跟踪。例如,在自动驾驶领域中,可以通过对周围的目标物体或行人进行跟踪,以此对自动驾驶设备进行控制。在传统方式中,通常都是基于拍摄的图像进行目标跟踪。With the development of computer technology, visual tracking technology has emerged, and the position and speed of the target can be tracked through the computer's visual tracking technology. For example, in the field of automatic driving, the automatic driving equipment can be controlled by tracking the surrounding target objects or pedestrians. In traditional methods, target tracking is usually based on captured images.
然而,发明人意识到,根据图像进行目标跟踪容易受到图像质量的影响。在环境光照变化、目标运动较快等情况下,图像质量较低。基于质量较低的图像,无法准确的对目标进行跟踪,跟踪结果的准确性较低。However, the inventor realizes that target tracking based on images is easily affected by image quality. The image quality is lower when the ambient light changes, the target moves faster, and so on. Based on low-quality images, the target cannot be tracked accurately, and the accuracy of the tracking result is low.
发明内容Summary of the invention
根据本申请公开的各种实施例,提供一种基于点云的目标跟踪方法、装置、计算机设备、存储介质和交通工具。According to various embodiments disclosed in the present application, a point cloud-based target tracking method, device, computer equipment, storage medium, and transportation tool are provided.
一种基于点云的目标跟踪方法,包括:A point cloud-based target tracking method includes:
根据当前帧点云数据生成当前帧特征图;Generate a feature map of the current frame according to the point cloud data of the current frame;
获取所述当前帧特征图对应的候选区域;Acquiring the candidate area corresponding to the feature map of the current frame;
在所述当前帧特征图中截取与所述候选区域相匹配的候选特征图;Intercept a candidate feature map matching the candidate region in the current frame feature map;
提取所述候选特征图所对应的图像特征;及Extracting the image feature corresponding to the candidate feature map; and
调用基于上一帧点云数据训练得到的目标跟踪模型,根据所述图像特征确定所述当前帧点云数据对应的目标跟踪区域。The target tracking model trained based on the point cloud data of the previous frame is called, and the target tracking area corresponding to the point cloud data of the current frame is determined according to the image characteristics.
一种基于点云的目标跟踪装置,包括:A point cloud-based target tracking device includes:
特征图生成模块,用于根据当前帧点云数据生成当前帧特征图;The feature map generating module is used to generate the feature map of the current frame according to the point cloud data of the current frame;
候选区域获取模块,用于获取所述当前帧特征图对应的候选区域;在所述当前帧特征图中截取与所述候选区域相匹配的候选特征图;A candidate region acquiring module, configured to acquire a candidate region corresponding to the current frame feature map; intercept a candidate feature map matching the candidate region in the current frame feature map;
特征提取模块,用于提取所述候选特征图所对应的图像特征;及A feature extraction module for extracting image features corresponding to the candidate feature map; and
目标跟踪模块,用于调用基于上一帧点云数据训练得到的目标跟踪模型,根据所述图像特征确定所述当前帧点云数据对应的目标跟踪区域。The target tracking module is used to call the target tracking model trained based on the point cloud data of the previous frame, and determine the target tracking area corresponding to the point cloud data of the current frame according to the image characteristics.
一种计算机设备,包括存储器和一个或多个处理器,所述存储器中储存有计算机可读指令,所述计算机可读指令被所述处理器执行时,使得所述一个或多个处理器执行以下步骤:A computer device, including a memory and one or more processors, the memory stores computer readable instructions, and when the computer readable instructions are executed by the processor, the one or more processors execute The following steps:
根据当前帧点云数据生成当前帧特征图;Generate a feature map of the current frame according to the point cloud data of the current frame;
获取所述当前帧特征图对应的候选区域;Acquiring the candidate area corresponding to the feature map of the current frame;
在所述当前帧特征图中截取与所述候选区域相匹配的候选特征图;Intercept a candidate feature map matching the candidate region in the current frame feature map;
提取所述候选特征图所对应的图像特征;及Extracting the image feature corresponding to the candidate feature map; and
调用基于上一帧点云数据训练得到的目标跟踪模型,根据所述图像特征确定所述当前帧点云数据对应的目标跟踪区域。The target tracking model trained based on the point cloud data of the previous frame is called, and the target tracking area corresponding to the point cloud data of the current frame is determined according to the image characteristics.
一个或多个存储有计算机可读指令的非易失性计算机可读存储介质,计算机可读指令被一个或多个处理器执行时,使得一个或多个处理器执行以下步骤:One or more non-volatile computer-readable storage media storing computer-readable instructions. When the computer-readable instructions are executed by one or more processors, the one or more processors perform the following steps:
根据当前帧点云数据生成当前帧特征图;Generate a feature map of the current frame according to the point cloud data of the current frame;
获取所述当前帧特征图对应的候选区域;Acquiring the candidate area corresponding to the feature map of the current frame;
在所述当前帧特征图中截取与所述候选区域相匹配的候选特征图;Intercept a candidate feature map matching the candidate region in the current frame feature map;
提取所述候选特征图所对应的图像特征;及Extracting the image feature corresponding to the candidate feature map; and
调用基于上一帧点云数据训练得到的目标跟踪模型,根据所述图像特征确定所述当前帧点云数据对应的目标跟踪区域。The target tracking model trained based on the point cloud data of the previous frame is called, and the target tracking area corresponding to the point cloud data of the current frame is determined according to the image characteristics.
一种交通工具,包括执行上述基于点云的目标跟踪方法的步骤。A vehicle includes the steps of executing the above-mentioned point cloud-based target tracking method.
本申请的一个或多个实施例的细节在下面的附图和描述中提出。本申请的其它特征和优点将从说明书、附图以及权利要求书变得明显。The details of one or more embodiments of the present application are set forth in the following drawings and description. Other features and advantages of this application will become apparent from the description, drawings and claims.
附图说明Description of the drawings
为了更清楚地说明本申请实施例中的技术方案,下面将对实施例中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本申请的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其它的附图。In order to more clearly describe the technical solutions in the embodiments of the present application, the following will briefly introduce the drawings needed in the embodiments. Obviously, the drawings in the following description are only some embodiments of the present application. A person of ordinary skill in the art can obtain other drawings based on these drawings without creative work.
图1为根据一个或多个实施例中基于点云的目标跟踪方法的应用场景图。Fig. 1 is an application scenario diagram of a point cloud-based target tracking method according to one or more embodiments.
图2为根据一个或多个实施例中基于点云的目标跟踪方法的流程示意图。Fig. 2 is a schematic flowchart of a point cloud-based target tracking method according to one or more embodiments.
图3为根据一个或多个实施例中根据当前帧点云数据生成当前帧特征图步骤的流程示意图。Fig. 3 is a schematic flowchart of the steps of generating a feature map of the current frame according to the point cloud data of the current frame according to one or more embodiments.
图4为另一个实施例中基于点云的目标跟踪方法的流程示意图。Fig. 4 is a schematic flowchart of a point cloud-based target tracking method in another embodiment.
图5为根据一个或多个实施例中基于点云的目标跟踪装置的框图。Fig. 5 is a block diagram of a point cloud-based target tracking device according to one or more embodiments.
图6为根据一个或多个实施例中计算机设备的框图。Figure 6 is a block diagram of a computer device according to one or more embodiments.
具体实施方式Detailed ways
为了使本申请的技术方案及优点更加清楚明白,以下结合附图及实施例,对本申请进行进一步详细说明。应当理解,此处描述的具体实施例仅仅用以解释本申请,并不用于限定本申请。In order to make the technical solutions and advantages of the present application clearer, the following further describes the present application in detail with reference to the accompanying drawings and embodiments. It should be understood that the specific embodiments described herein are only used to explain the present application, and are not used to limit the present application.
本申请提供的基于点云的目标跟踪方法,可以应用于如图1所示的自动驾驶的应用环境中。激光传感器102可以与计算机设备104进行通信。激光传感器102可以是车载激光传感器,计算机设备104可以是车载计算机设备。点云数据可以是由激光传感器102采集的,也可以是计算机设备预先存储的。计算机设备104根据当前帧点云数据生成当前帧特征图。计算机设备104获取当前帧特征图对应的候选区域,在当前帧特征图中截取与候选区域相匹配的候选特征图。计算机设备104提取候选特征图所对应的图像特征,调用基 于上一帧点云数据训练得到的目标跟踪模型,根据图像特征确定当前帧点云数据对应的目标跟踪区域。激光传感器102可以是自动驾驶设备搭载的激光传感器,具体可以包括激光雷达、激光扫描仪等。The point cloud-based target tracking method provided in this application can be applied to the application environment of automatic driving as shown in FIG. 1. The laser sensor 102 can communicate with the computer device 104. The laser sensor 102 may be a vehicle-mounted laser sensor, and the computer device 104 may be a vehicle-mounted computer device. The point cloud data can be collected by the laser sensor 102, or pre-stored by a computer device. The computer device 104 generates a feature map of the current frame according to the point cloud data of the current frame. The computer device 104 obtains the candidate region corresponding to the feature map of the current frame, and intercepts the candidate feature map matching the candidate region in the feature map of the current frame. The computer device 104 extracts the image features corresponding to the candidate feature map, calls the target tracking model trained based on the point cloud data of the previous frame, and determines the target tracking area corresponding to the point cloud data of the current frame according to the image features. The laser sensor 102 may be a laser sensor carried by an automatic driving device, and may specifically include a laser radar, a laser scanner, and the like.
在其中一个实施例中,如图2所示,提供了一种基于点云的目标跟踪方法,以该方法应用于图1中的计算机设备104为例进行说明,包括以下步骤:In one of the embodiments, as shown in FIG. 2, a point cloud-based target tracking method is provided. Taking the method applied to the computer device 104 in FIG. 1 as an example for description, the method includes the following steps:
步骤202,根据当前帧点云数据生成当前帧特征图。Step 202: Generate a feature map of the current frame according to the point cloud data of the current frame.
激光传感器由可以进行自动驾驶的设备搭载。例如,可以是无人车上搭载的激光传感器,也可以是包括自动驾驶模式的车辆搭载的激光传感器。激光传感器可以用于采集周围的环境数据。具体的,激光传感器可以发射探测信号,例如激光束等。激光传感器将反射回的信号与探测信号进行比对,得到周围的环境数据。环境数据具体可以是点云数据。点云数据是指扫描环境中的物体以点的形式记录,物体表面多个点所对应点数据的集合。多个可以指两个或者两个以上。激光传感器可以按照预设频率进行采集,得到多帧点云数据。预设频率可以是根据实际需求预先设置的。The laser sensor is equipped with a device capable of autonomous driving. For example, it may be a laser sensor mounted on an unmanned vehicle, or a laser sensor mounted on a vehicle including an automatic driving mode. Laser sensors can be used to collect surrounding environmental data. Specifically, the laser sensor can emit a detection signal, such as a laser beam. The laser sensor compares the reflected signal with the detection signal to obtain surrounding environmental data. The environmental data may specifically be point cloud data. Point cloud data refers to a collection of point data corresponding to multiple points on the surface of the object in the scanning environment recorded in the form of points. Multiple can refer to two or more than two. The laser sensor can collect according to a preset frequency to obtain multi-frame point cloud data. The preset frequency can be preset according to actual needs.
点云数据可以是三维点云数据,每一帧点云数据可以包括多个点各自对应的点数据。点数据具体可以包括点对应的三维坐标、激光反射强度以及颜色信息等中的至少一种。其中,三维坐标可以是点在笛卡尔坐标系中的坐标,具体包括点在笛卡尔坐标系中的横轴坐标、纵轴坐标以及竖轴坐标。笛卡尔坐标系是以激光传感器为原点建立的三维空间坐标系,三维空间坐标系包括横轴(x轴)、纵轴(y轴)和竖轴(z轴)。以激光传感器为原点建立的三维空间坐标系满足右手定则。The point cloud data may be three-dimensional point cloud data, and each frame of point cloud data may include point data corresponding to multiple points. The point data may specifically include at least one of three-dimensional coordinates, laser reflection intensity, and color information corresponding to the point. Among them, the three-dimensional coordinates may be the coordinates of the point in the Cartesian coordinate system, and specifically include the horizontal axis coordinates, the vertical axis coordinates, and the vertical axis coordinates of the point in the Cartesian coordinate system. The Cartesian coordinate system is a three-dimensional space coordinate system established with a laser sensor as the origin. The three-dimensional space coordinate system includes a horizontal axis (x axis), a vertical axis (y axis), and a vertical axis (z axis). The three-dimensional space coordinate system established with the laser sensor as the origin satisfies the right-hand rule.
计算机设备可以按照激光传感器采集点云数据的时间顺序,依次根据多帧点云数据对目标进行跟踪。目标是指周围环境中的生物体或非生物体,目标可以是运动的,也可以是静止的。例如,目标具体可以包括行人、路障、车辆和建筑物等中的至少一种。当前帧点云数据是指计算机设备正在处理的一帧点云数据。可以理解的,当计算机设备对当前帧点云数据跟踪结束,开始对下一帧点云数据进行跟踪时,可以将当前帧点云数据记作上一帧点云数 据,将下一帧点云数据重新记作当前帧点云数据。计算机设备可以获取当前帧点云数据包括的点数据,根据点数据对点进行编码,得到多个点各自对应的点特征。点特征可以采用向量的形式表示,每个点的点特征可以包括点向量。点向量可以是多维向量。计算机设备可以根据多个点各自对应的点特征生成特征图,根据当前帧点云数据生成的特征图可以记作当前帧特征图。相较于传统方式采集的环境图像,本实施中根据点云数据生成的特征图不包括RGB(红、绿、蓝)通道,生成的特征图与传统采集的图像是不同的,基于点云数据进行目标跟踪不会受到环境光照、目标运动速度等因素的影响。The computer equipment can follow the time sequence of the point cloud data collected by the laser sensor, and track the target according to the multi-frame point cloud data in turn. A target refers to a living or non-living body in the surrounding environment. The target can be moving or stationary. For example, the target may specifically include at least one of pedestrians, roadblocks, vehicles, and buildings. The current frame of point cloud data refers to a frame of point cloud data being processed by the computer equipment. It is understandable that when the computer device finishes tracking the point cloud data of the current frame and starts to track the point cloud data of the next frame, the point cloud data of the current frame can be recorded as the point cloud data of the previous frame, and the point cloud data of the next frame can be recorded as the point cloud data of the previous frame. The data is re-recorded as the point cloud data of the current frame. The computer device can obtain the point data included in the point cloud data of the current frame, encode the points according to the point data, and obtain the point features corresponding to each of the multiple points. The point feature can be expressed in the form of a vector, and the point feature of each point can include a point vector. The point vector can be a multidimensional vector. The computer device can generate a feature map according to the point features corresponding to each of the multiple points, and the feature map generated from the point cloud data of the current frame can be recorded as the current frame feature map. Compared with the environmental image collected in the traditional way, the feature map generated from the point cloud data in this implementation does not include RGB (red, green, blue) channels. The generated feature map is different from the traditionally collected image, based on the point cloud data Target tracking will not be affected by factors such as ambient lighting and target movement speed.
步骤204,获取当前帧特征图对应的候选区域。Step 204: Obtain a candidate region corresponding to the feature map of the current frame.
候选区域是用于确定目标跟踪区域的特征图区域,候选区域可以包括目标跟踪区域。候选区域为当前帧特征图中的一个区域范围,候选区域可以是当前帧特征图的整个区域,也可以是当前帧特征图的部分区域。候选区域可以包括区域大小、区域形状以及在当前帧特征图中的位置。候选区域可以是多种形状的区域中的一种。例如,原则上候选区域的形状一般可以是矩形的,候选区域的形状也还可以是圆形的。The candidate area is a feature map area used to determine the target tracking area, and the candidate area may include the target tracking area. The candidate area is an area range in the feature map of the current frame, and the candidate area can be the entire area of the feature map of the current frame, or a partial area of the feature map of the current frame. The candidate area may include the area size, the area shape, and the position in the current frame feature map. The candidate area may be one of a variety of shapes. For example, in principle, the shape of the candidate region can generally be rectangular, and the shape of the candidate region can also be circular.
计算机设备根据当前帧点云数据生成当前帧特征图后,可以采用多种方式获取当前帧特征图对应的候选区域。例如,计算机设备可以将当前帧特征图的整个区域记作对应的候选区域。计算机设备还可以根据上一帧点云数据获取当前帧特征图对应的候选区域。具体的,计算机设备可以根据上一帧点云数据的跟踪结果,获取上一帧点云数据对应的上一帧目标区域,计算机设备可以根据上一帧目标区域确定当前帧特征图对应的候选区域。After the computer device generates the feature map of the current frame according to the point cloud data of the current frame, it can obtain the candidate region corresponding to the feature map of the current frame in a variety of ways. For example, the computer device may record the entire area of the feature map of the current frame as the corresponding candidate area. The computer device can also obtain the candidate area corresponding to the feature map of the current frame according to the point cloud data of the previous frame. Specifically, the computer device can obtain the target area of the previous frame corresponding to the point cloud data of the previous frame according to the tracking result of the point cloud data of the previous frame, and the computer device can determine the candidate area corresponding to the feature map of the current frame according to the target area of the previous frame. .
在其中一个实施例中,获取当前帧特征图对应的候选区域包括:获取上一帧点云数据对应的上一帧目标区域;根据预设倍数将上一帧目标区域进行扩大;确定扩大后的上一帧目标区域作为当前帧特征图对应的候选区域。In one of the embodiments, obtaining the candidate area corresponding to the feature map of the current frame includes: obtaining the target area of the previous frame corresponding to the point cloud data of the previous frame; expanding the target area of the previous frame according to a preset multiple; The target area of the previous frame is used as the candidate area corresponding to the feature map of the current frame.
具体的,计算机设备可以根据上一帧点云数据的跟踪结果,获取上一帧点云数据对应的上一帧目标区域。上一帧目标区域是指在上一帧点云数据中目标所在的区域。计算机设备可以按照预设倍数将上一帧目标区域进行扩大, 得到扩大后的上一帧目标区域,以此保证候选区域内的当前帧特征图可以包括目标。计算机设备可以扩大预设倍数面积的上一帧目标区域,也可以对目标区域的边长按照预设倍数进行扩大,确定扩大后的边长所形成的闭合区域作为扩大后的上一帧目标区域。计算机设备可以按照扩大后的上一帧目标区域,在当前帧特征图中确定扩大后的上一帧目标区域,确定扩大后的上一帧目标区域作为当前帧特征图对应的候选区域。预设倍数可以是用户根据实际需求预先设置的,例如,预设倍数具体可以是2倍。Specifically, the computer device may obtain the target area of the previous frame corresponding to the point cloud data of the previous frame according to the tracking result of the point cloud data of the previous frame. The target area of the previous frame refers to the area where the target is located in the point cloud data of the previous frame. The computer device can expand the target area of the previous frame according to the preset multiple to obtain the expanded target area of the previous frame, so as to ensure that the current frame feature map in the candidate area can include the target. The computer equipment can expand the target area of the previous frame with a preset multiple area, or expand the side length of the target area according to the preset multiple, and determine the closed area formed by the expanded side length as the expanded target area of the previous frame . The computer device may determine the expanded previous frame target area in the current frame feature map according to the expanded previous frame target area, and determine the expanded previous frame target area as the candidate area corresponding to the current frame feature map. The preset multiple may be preset by the user according to actual needs, for example, the preset multiple may specifically be 2 times.
在本实施例中,根据上一帧目标区域确定当前帧特征图对应的候选区域,实现了在当前帧特征图中跟踪上一帧目标。计算机设备可以根据预设倍数将上一帧目标区域进行扩大,确定扩大后的上一帧目标区域作为当前帧特征图对应的候选区域,保证了目标跟踪的准确性,也不需要处理整个当前帧特征图,有效的节省了计算机设备的运算资源。In this embodiment, the candidate area corresponding to the feature map of the current frame is determined according to the target area of the previous frame, so that the target of the previous frame is tracked in the feature map of the current frame. The computer device can expand the target area of the previous frame according to the preset multiple, and determine the expanded target area of the previous frame as the candidate area corresponding to the feature map of the current frame, which ensures the accuracy of target tracking and does not need to process the entire current frame The feature map effectively saves the computing resources of the computer equipment.
步骤206,在当前帧特征图中截取与候选区域相匹配的候选特征图。Step 206: Extract a candidate feature map matching the candidate region in the current frame feature map.
计算机设备可以根据当前帧特征图对应的候选区域,在当前帧特征图中对候选特征图进行截取,截取得到与候选区域相对应的候选特征图。与候选区域相对应的,候选特征图可以是当前帧特征图的全部特征图,也可以是当前帧特征图中的部分特征图。候选特征图中可以包括待跟踪的目标,截取得到的候选特征图与候选区域的大小及形状相匹配。The computer device can intercept the candidate feature map in the current frame feature map according to the candidate region corresponding to the current frame feature map, and intercept the candidate feature map corresponding to the candidate region. Corresponding to the candidate region, the candidate feature map can be all feature maps of the current frame feature map, or part of the feature map of the current frame feature map. The candidate feature map may include the target to be tracked, and the candidate feature map obtained by interception matches the size and shape of the candidate area.
步骤208,提取候选特征图所对应的图像特征。Step 208: Extract image features corresponding to the candidate feature map.
计算机设备可以对截取出的候选特征图进行特征提取,从候选特征图中提取候选特征图所对应的图像特征。候选特征图对应的图像特征可以包括多种特征类型中的至少一种。例如,图像特征具体可以包括边缘特征、角特征以及区域特征等多种特征类型中的至少一种。图像特征可以采用特征向量等方式记录。计算机设备可以提取候选特征图所对应的图像特征,根据图像特征对候选特征图进行特征分析,根据分析结果在候选特征图中确定当前帧点云数据对应的目标跟踪区域。The computer device can perform feature extraction on the selected candidate feature map, and extract the image features corresponding to the candidate feature map from the candidate feature map. The image feature corresponding to the candidate feature map may include at least one of multiple feature types. For example, the image features may specifically include at least one of multiple feature types such as edge features, corner features, and regional features. Image features can be recorded by means of feature vectors, etc. The computer device can extract the image features corresponding to the candidate feature map, perform feature analysis on the candidate feature map according to the image features, and determine the target tracking area corresponding to the point cloud data of the current frame in the candidate feature map according to the analysis result.
在其中一个实施例中,提取候选特征图所对应的图像特征包括:获取特 征提取模型;将候选特征图输入至特征提取模型;根据特征提取模型对候选特征图进行特征提取,得到候选特征图所对应的图像特征。In one of the embodiments, extracting the image features corresponding to the candidate feature map includes: obtaining a feature extraction model; inputting the candidate feature map to the feature extraction model; performing feature extraction on the candidate feature map according to the feature extraction model to obtain the candidate feature map location Corresponding image characteristics.
具体的,计算机设备可以获取特征提取模型,特征提取模型可以是预先配置在计算机设备中的。计算机设备可以将截取得到的候选特征图输入至特征提取模型中,通过特征提取模型对输入的候选特征图进行运算,对候选特征图进行特征提取。特征提取模型可以是多种神经网络模型中的一种。例如,特征提取模型具体可以是传统卷积神经网络(Convolutional Neural Networks,简称CNN)模型以及VGG(Visual Geometry Group Network)模型等神经网络模型中的一种。特征提取模型具体可以包括输入层、卷积层、池化层、全连接层、BN(Batch Normalization,批量标准化)层以及输出层等。计算机设备可以根据特征提取模型的网络结构,依次对候选特征图进行与网络结构相对应的运算,获取特征提取模型输出的候选特征图所对应的图像特征。Specifically, the computer device can obtain the feature extraction model, and the feature extraction model can be pre-configured in the computer device. The computer device can input the intercepted candidate feature maps into the feature extraction model, and perform operations on the input candidate feature maps through the feature extraction model to perform feature extraction on the candidate feature maps. The feature extraction model can be one of a variety of neural network models. For example, the feature extraction model may specifically be one of neural network models such as a traditional Convolutional Neural Networks (CNN) model and a VGG (Visual Geometry Group Network) model. The feature extraction model may specifically include an input layer, a convolutional layer, a pooling layer, a fully connected layer, a BN (Batch Normalization, batch normalization) layer, an output layer, and so on. The computer device can sequentially perform operations corresponding to the network structure on the candidate feature maps according to the network structure of the feature extraction model to obtain image features corresponding to the candidate feature maps output by the feature extraction model.
在本实施例中,计算机设备可以获取特征提取模型,将候选特征图输入至特征提取模型中,根据特征提取模型对候选特征图进行特征提取,得到候选特征图对应的图像特征,有效的提高了图像特征的准确性。进而根据图像特征确定当前帧点云数据对应的目标跟踪区域,提高了目标跟踪的准确性。In this embodiment, the computer device can obtain the feature extraction model, input the candidate feature map into the feature extraction model, perform feature extraction on the candidate feature map according to the feature extraction model, and obtain the image features corresponding to the candidate feature map, which effectively improves The accuracy of image features. Furthermore, the target tracking area corresponding to the point cloud data of the current frame is determined according to the image characteristics, which improves the accuracy of target tracking.
步骤210,调用基于上一帧点云数据训练得到的目标跟踪模型,根据图像特征确定当前帧点云数据对应的目标跟踪区域。Step 210: Call the target tracking model trained based on the point cloud data of the previous frame, and determine the target tracking area corresponding to the point cloud data of the current frame according to the image characteristics.
计算机设备可以调用目标跟踪模型,根据目标跟踪模型对候选特征图对应的图像特征进行跟踪处理,得到当前帧点云数据对应的目标跟踪区域。目标跟踪模型是基于上一帧点云数据进行训练之后所得到的的跟踪模型,目标跟踪模型用于对当前帧点云数据中目标所在的区域进行跟踪,得到目标跟踪区域。目标跟踪区域是指通过跟踪处理所预估的当前帧特征图中目标所在的位置区域。计算机设备可以将提取的图像特征输入至基于上一帧点云数据训练得到的目标跟踪模型,通过目标跟踪模型对当前帧特征图的图像特征进行跟踪处理,得到目标跟踪模型输出的目标跟踪区域。The computer device can call the target tracking model, and perform tracking processing on the image features corresponding to the candidate feature map according to the target tracking model to obtain the target tracking area corresponding to the point cloud data of the current frame. The target tracking model is a tracking model obtained after training based on the point cloud data of the previous frame. The target tracking model is used to track the area where the target is located in the point cloud data of the current frame to obtain the target tracking area. The target tracking area refers to the location area of the target in the feature map of the current frame estimated by the tracking process. The computer device can input the extracted image features into the target tracking model trained based on the point cloud data of the previous frame, and use the target tracking model to track the image features of the current frame feature map to obtain the target tracking area output by the target tracking model.
可以理解的,计算机设备可以获取多帧点云数据,根据每帧点云数据采 集时间的先后顺序,依次对每一帧点云数据进行处理。当计算机设备对当前帧点云数据进行跟踪处理时,可以根据当前帧点云数据对应的上一帧点云数据对跟踪模型进行训练,得到当前帧点云数据对应的目标跟踪模型。当计算机设备对当前帧点云数据处理结束后,可以对下一帧点云数据进行跟踪处理。计算机设备可以根据当前帧点云数据对跟踪模型进行训练,得到下一帧点云数据对应的目标跟踪模型,以此对下一帧点云数据进行跟踪处理,计算机设备可以根据点云数据的处理顺序迭代训练跟踪模型。It is understandable that the computer device can obtain multiple frames of point cloud data, and process each frame of point cloud data in sequence according to the sequence of the point cloud data collection time of each frame. When the computer device performs tracking processing on the point cloud data of the current frame, the tracking model can be trained according to the point cloud data of the previous frame corresponding to the point cloud data of the current frame to obtain the target tracking model corresponding to the point cloud data of the current frame. After the computer equipment finishes processing the point cloud data of the current frame, it can perform tracking processing on the point cloud data of the next frame. The computer equipment can train the tracking model according to the point cloud data of the current frame, and obtain the target tracking model corresponding to the point cloud data of the next frame, so as to track the point cloud data of the next frame. The computer equipment can process the point cloud data according to the point cloud data. Train the tracking model sequentially iteratively.
在本实施例中,计算机设备可以获取当前帧点云数据,根据当前帧点云数据生成当前帧特征图。计算机设备可以提取当前帧特征图中候选特征图的图像特征,根据基于上一帧点云数据训练得到的目标跟踪模型对图像特征进行跟踪处理,确定当前帧点云数据对应的目标跟踪区域。相较于传统基于图像进行目标跟踪的方式,本实施例中基于点云数据进行目标跟踪,不会受到外界光照、目标运动速度等因素的影响,有效的提高了目标跟踪的准确性。In this embodiment, the computer device can obtain the point cloud data of the current frame, and generate a feature map of the current frame according to the point cloud data of the current frame. The computer device can extract the image features of the candidate feature map in the current frame feature map, and track the image features according to the target tracking model trained based on the point cloud data of the previous frame to determine the target tracking area corresponding to the current frame point cloud data. Compared with the traditional image-based target tracking method, the target tracking based on point cloud data in this embodiment will not be affected by factors such as external light, target movement speed, etc., which effectively improves the accuracy of target tracking.
在其中一个实施例中,如图3所示,根据当前帧点云数据生成当前帧特征图,包括:In one of the embodiments, as shown in FIG. 3, generating a feature map of the current frame according to the point cloud data of the current frame includes:
步骤302,获取当前帧点云数据。Step 302: Obtain the point cloud data of the current frame.
步骤304,将当前帧点云数据进行结构化处理,得到处理结果。Step 304: Perform structured processing on the point cloud data of the current frame to obtain a processing result.
步骤306,基于处理结果对当前帧点云数据中的点进行编码,得到点对应的点特征。Step 306: Encode the points in the point cloud data of the current frame based on the processing result to obtain point features corresponding to the points.
步骤308,根据点特征生成当前帧点云数据对应的当前帧特征图。Step 308: Generate a feature map of the current frame corresponding to the point cloud data of the current frame according to the point features.
计算机设备可以获取激光传感器采集可视范围内的多帧点云数据,按照激光传感器采集点云数据的时间顺序依次对每帧点云数据进行处理。计算机设备可以将开始处理或者正在处理的点云数据记作当前帧点云数据。计算机设备可以对当前帧点云数据进行处理,生成当前帧点云数据对应的当前帧特征图。The computer equipment can obtain the point cloud data of multiple frames within the visible range collected by the laser sensor, and process the point cloud data of each frame in sequence according to the time sequence of the point cloud data collected by the laser sensor. The computer device may record the point cloud data that is being processed or being processed as the point cloud data of the current frame. The computer device can process the point cloud data of the current frame to generate a feature map of the current frame corresponding to the point cloud data of the current frame.
具体的,计算机设备可以将当前帧点云数据进行结构化处理,得到结构化处理后的处理结果。计算机设备进行结构化处理的方式包括多种。例如, 计算机设备可以对当前帧点云数据进行栅格化处理,也可以对当前帧点云数据进行体素化处理。以栅格化处理为例,计算机设备可以对激光传感器为原点的平面进行栅格化,将平面划分为多个栅格。结构化处理后的结构化空间可以为柱状空间,点可以分布在栅格对应竖轴的柱状空间中,即柱状空间中的点的横坐标与纵坐标在对应的栅格坐标范围内。Specifically, the computer device may perform structured processing on the point cloud data of the current frame to obtain a processing result after the structured processing. There are many ways for computer equipment to perform structured processing. For example, the computer device may perform rasterization processing on the current frame point cloud data, or may perform voxelization processing on the current frame point cloud data. Taking rasterization processing as an example, the computer equipment can rasterize the plane with the laser sensor as the origin, and divide the plane into multiple grids. The structured space after the structuring process may be a columnar space, and the points may be distributed in the columnar space corresponding to the vertical axis of the grid, that is, the abscissa and ordinate of the points in the columnar space are within the corresponding grid coordinate range.
计算机设备可以根据结构化处理的处理结果,对当前帧点云数据中的点进行编码,得到点对应的点特征。点特征可以是点对应的点向量。具体的,计算机设备可以统计每一个结构化空间内所有点的点数据,根据统计的点数据对点进行编码,得到点对应的点向量。The computer device can encode the point in the point cloud data of the current frame according to the processing result of the structured processing to obtain the point feature corresponding to the point. The point feature can be a point vector corresponding to the point. Specifically, the computer device can count the point data of all points in each structured space, encode the points according to the statistical point data, and obtain the point vectors corresponding to the points.
举例说明,计算机设备在对原点平面进行栅格化之后,统计每个柱状空间内的点数据。其中,点数据具体可以包括每个点对应的三维坐标和反射系数。计算机设备可以根据柱状空间内的点数据,对点进行编码,得到点对应的点向量。例如,点向量可以是9维向量。点向量具体可以包括点对应的横轴坐标、纵轴坐标、竖轴坐标、反射系数、与柱状空间中心的距离以及每个点与所有点三维坐标均值的距离。其中,点与柱状空间中心的距离可以采用横轴距离与纵轴距离表示,点与所有点三维坐标均值的距离可以采用横轴距离、纵轴距离以及竖轴距离表示。计算机设备可以将点对应的9维向量记作点对应的点特征。For example, after the computer equipment rasterizes the origin plane, it counts the point data in each columnar space. Among them, the point data may specifically include the three-dimensional coordinates and reflection coefficients corresponding to each point. The computer equipment can encode the point according to the point data in the columnar space to obtain the point vector corresponding to the point. For example, the point vector may be a 9-dimensional vector. The point vector may specifically include the horizontal axis coordinate, the vertical axis coordinate, the vertical axis coordinate, the reflection coefficient, the distance from the center of the cylindrical space, and the distance between each point and the average value of the three-dimensional coordinates of all points. Among them, the distance between a point and the center of the columnar space can be represented by the horizontal axis distance and the vertical axis distance, and the distance between the point and the average of the three-dimensional coordinates of all points can be represented by the horizontal axis distance, the vertical axis distance, and the vertical axis distance. The computer equipment can record the 9-dimensional vector corresponding to the point as the point feature corresponding to the point.
计算机设备可以统计多个结构化空间内的点特征,根据多个点特征生成当前帧点云数据对应的当前帧特征图。The computer device can count the point features in multiple structured spaces, and generate the current frame feature map corresponding to the current frame point cloud data according to the multiple point features.
在本实施例中,计算机设备可以根据当前帧点云数据对点进行编码,得到点特征,根据点特征生成当前帧点云数据对应的当前帧特征图,点云数据不会受到环境光照、目标运动速度等因素的影响,保证了对目标进行跟踪的准确性。同时,相较于传统对点云数据采用卡尔曼滤波的方式进行目标跟踪的方式,本实施中根据当前帧点云数据生成当前帧特征图,能够有效的利用点云数据中的深度特征,进而有效的提高了目标跟踪的准确性。In this embodiment, the computer device can encode the points according to the point cloud data of the current frame to obtain the point characteristics, and generate the current frame feature map corresponding to the point cloud data of the current frame according to the point characteristics. The point cloud data will not be affected by ambient light or target The influence of factors such as movement speed ensures the accuracy of tracking the target. At the same time, compared to the traditional method of using Kalman filtering to track the point cloud data, the current frame feature map is generated according to the current frame point cloud data in this implementation, which can effectively use the depth features in the point cloud data, and then Effectively improve the accuracy of target tracking.
在其中一个实施例中,计算机设备可以从结构化空间中采集预设数量的 采样点,对采样点进行编码,得到采样点对应的点特征。计算机设备可以根据每个结构化空间中采样点对应的点特征生成当前帧点云数据对应的当前帧特征图。具体的,计算机设备可以统计结构化空间包括点的数量,将点数量与预设数量进行比对。预设数量可以是根据实际需求,以及历史点云数据进行大数据分析后预先设置的。当结构化空间的点数量大于或等于预设数量时,计算机设备可以从结构化空间中随机抽取预设数量的点作为采样点。当结构化空间的点数量小于预设数量时,计算机设备可以获取结构化空间内的所有点作为采样点,并且添加虚拟点作为采样点,使得可以采集预设数量的采样点。其中,虚拟点的三维坐标可以位于坐标系原点。In one of the embodiments, the computer device may collect a preset number of sampling points from the structured space, encode the sampling points, and obtain the point characteristics corresponding to the sampling points. The computer device can generate the current frame feature map corresponding to the current frame point cloud data according to the point features corresponding to the sampling points in each structured space. Specifically, the computer device can count the number of points included in the structured space, and compare the number of points with the preset number. The preset number can be preset according to actual needs and historical point cloud data after big data analysis. When the number of points in the structured space is greater than or equal to the preset number, the computer device may randomly select a preset number of points from the structured space as sampling points. When the number of points in the structured space is less than the preset number, the computer device can obtain all points in the structured space as sampling points, and add virtual points as sampling points, so that the preset number of sampling points can be collected. Among them, the three-dimensional coordinates of the virtual point may be located at the origin of the coordinate system.
在本实施例中,计算机设备通过从每个结构化空间中采集预设数量的点作为采样点,使得每个结构化空间中采样点的数量相同,以此均衡多个结构化空间的点特征,有助于计算机设备根据结构化数据生成当前帧特征图。In this embodiment, the computer device collects a preset number of points from each structured space as sampling points, so that the number of sampling points in each structured space is the same, thereby balancing the point characteristics of multiple structured spaces , Which helps computer equipment to generate a feature map of the current frame based on structured data.
在其中一个实施例中,根据点特征生成当前帧点云数据对应的当前帧特征图,包括:根据多个点特征生成点特征矩阵;调用图像生成模型,将点特征矩阵输入至图像生成模型;获取图像生成模型输出的当前帧特征图。In one of the embodiments, generating the current frame feature map corresponding to the point cloud data of the current frame according to the point features includes: generating a point feature matrix based on multiple point features; calling the image generation model, and inputting the point feature matrix to the image generation model; Obtain the current frame feature map output by the image generation model.
计算机设备可以根据多个结构化空间中的多个点特征生成点特征矩阵,点特征矩阵具体可以包括点特征、结构化空间以及对应的点数量等。计算机设备可以调用图像生成模型。图像生成模型可以是预先配置在计算机设备中的,图像生成模型可以是通过大量点特征样本以及点特征样本对应的特征图样本进行训练得到的。图像生成模型可以是多种神经网络模型中的一种。例如,图像生成模型可以是卷积神经网络模型,具体可以为PointNet模型。计算机设备可以将生成的点特征矩阵输入至图像生成模型,通过图像生成模型对点特征矩阵进行运算,对点数量维度的点特征进行最大池化运算,得到当前帧点云数据对应的当前帧特征图。The computer device may generate a point feature matrix based on multiple point features in multiple structured spaces, and the point feature matrix may specifically include point features, structured spaces, and corresponding points, etc. The computer equipment can call the image generation model. The image generation model may be pre-configured in the computer device, and the image generation model may be obtained by training a large number of point feature samples and feature map samples corresponding to the point feature samples. The image generation model can be one of a variety of neural network models. For example, the image generation model may be a convolutional neural network model, and specifically may be a PointNet model. The computer device can input the generated point feature matrix to the image generation model, and calculate the point feature matrix through the image generation model, and perform the maximum pooling operation on the point features of the point quantity dimension to obtain the current frame feature corresponding to the current frame point cloud data Figure.
在本实施例中,计算机设备根据多个点特征生成点特征矩阵,根据图像生成模型对点特征矩阵进行运算,得到图像生成模型输出的当前帧特征图。计算机设备根据当前帧点云数据生成当前帧特征图,有效的利用了点云数据 中的深度特征,提高了基于点云数据进行目标跟踪的准确性。In this embodiment, the computer device generates a point feature matrix based on multiple point features, and performs operations on the point feature matrix according to the image generation model to obtain the current frame feature map output by the image generation model. The computer equipment generates a feature map of the current frame according to the point cloud data of the current frame, which effectively utilizes the depth features in the point cloud data, and improves the accuracy of target tracking based on the point cloud data.
在其中一个实施例中,调用基于上一帧点云数据训练得到的目标跟踪模型,根据图像特征确定当前帧点云数据对应的目标跟踪区域,包括:根据图像特征生成图像特征矩阵;将图像特征矩阵输入至目标跟踪模型,获取目标跟踪模型输出的区域标签;根据区域标签确定当前帧点云数据所对应的目标跟踪区域。In one of the embodiments, calling the target tracking model trained based on the point cloud data of the previous frame, and determining the target tracking area corresponding to the point cloud data of the current frame according to the image features, includes: generating an image feature matrix according to the image features; The matrix is input to the target tracking model, and the area label output by the target tracking model is obtained; the target tracking area corresponding to the point cloud data of the current frame is determined according to the area label.
计算机设备可以根据提取的图像特征生成对应的图像特征矩阵。具体的,计算机设备可以对提取出的图像特征进行处理,将每一列图像特征续接在上一列图像特征后,根据图像特征生成列向量。计算机设备可以对列向量进行循环移位,将移位后得到的所有图像特征按列排放,得到图像特征矩阵。循环移位可以包括循环左移或循环右移。The computer device can generate a corresponding image feature matrix according to the extracted image features. Specifically, the computer device can process the extracted image features, and after each column of image features are connected to the previous column of image features, a column vector is generated according to the image features. The computer device can cyclically shift the column vector, and arrange all the image features obtained after the shift in columns to obtain an image feature matrix. Cyclic shifting may include rotating left or rotating right.
计算机设备可以将生成的图像特征矩阵输入至目标跟踪模型,目标跟踪模型是基于上一帧点云数据进行训练后得到的。目标跟踪模型可以采用多种视觉跟踪算法中的至少一种。例如,具体可以采用KCF(Kernel Correlation Filter,核相关滤波算法)算法,以及基于KCF的目标跟踪算法等。计算机设备可以通过目标跟踪模型对图像特征矩阵进行运算,获取目标跟踪模型经过运算之后输出的区域标签。目标跟踪模型可以输出多个区域标签,区域标签用于对相对应的区域进行标记。区域标签标记的区域表示目标可能所在的范围,区域的大小和形状可以与上一帧点云数据对应的上一帧目标区域相同。区域标签可以表示目标在对应区域范围内的可能性。在其中一个实施例中,区域标签可以是目标所在对应区域的概率。The computer device can input the generated image feature matrix to the target tracking model, which is obtained after training based on the point cloud data of the previous frame. The target tracking model can use at least one of a variety of visual tracking algorithms. For example, KCF (Kernel Correlation Filter, kernel correlation filtering algorithm) algorithm, and KCF-based target tracking algorithm, etc. can be specifically used. The computer device can perform operations on the image feature matrix through the target tracking model, and obtain the area label output by the target tracking model after the calculation. The target tracking model can output multiple area labels, which are used to mark corresponding areas. The area marked by the area label indicates the possible range of the target, and the size and shape of the area can be the same as the target area of the previous frame corresponding to the point cloud data of the previous frame. The area label can indicate the possibility that the target is within the corresponding area. In one of the embodiments, the area label may be the probability of the corresponding area where the target is located.
计算机设备可以将目标跟踪模型输出的多个区域标签进行比对,从多个区域标签中获取标签值最大的区域标签作为目标区域标签。计算机设备可以确定目标区域标签所对应的区域作为目标跟踪区域。具体的,计算机设备可以获取目标区域标签所对应的循环偏移量,根据循环偏移量确定上一帧目标区域偏移后的区域,确定上一帧目标区域偏移后的区域作为当前帧点云数据对应的目标跟踪区域。The computer device can compare multiple area labels output by the target tracking model, and obtain the area label with the largest label value from the multiple area labels as the target area label. The computer device can determine the area corresponding to the target area tag as the target tracking area. Specifically, the computer device can obtain the cyclic offset corresponding to the target area label, determine the area after the offset of the target area of the previous frame according to the cyclic offset, and determine the area after the offset of the target area of the previous frame as the current frame point Target tracking area corresponding to cloud data.
在本实施例中,计算机设备可以根据图像特征生成图像特征矩阵,调用目标跟踪模型对图像特征矩阵进行运算,根据目标跟踪模型输出的区域标签确定当前帧点云数据所对应的目标跟踪区域。相较于传统对点云数据进行卡尔曼滤波未考虑点云特征的方式,本实施例中利用了当前帧点云数据对应的点云特征,有效的提高了基于点云数据进行目标跟踪的准确性。In this embodiment, the computer device can generate an image feature matrix based on image features, call the target tracking model to perform operations on the image feature matrix, and determine the target tracking area corresponding to the point cloud data of the current frame according to the area label output by the target tracking model. Compared with the traditional Kalman filtering method for point cloud data that does not consider point cloud features, this embodiment uses the point cloud features corresponding to the point cloud data of the current frame, which effectively improves the accuracy of target tracking based on point cloud data. Sex.
在其中一个实施例中,如图4所示,在调用基于上一帧点云数据训练得到的目标跟踪模型步骤之前,上述基于点云的目标跟踪方法还包括:In one of the embodiments, as shown in FIG. 4, before the step of calling the target tracking model trained based on the point cloud data of the previous frame, the above point cloud-based target tracking method further includes:
步骤402,根据上一帧点云数据生成上一帧特征图。Step 402: Generate a feature map of the previous frame according to the point cloud data of the previous frame.
步骤404,在上一帧特征图中截取样本特征图,提取样本特征图所对应的样本特征。Step 404: Take a sample of the feature map from the previous frame of feature map, and extract the sample feature corresponding to the sample feature map.
步骤406,生成与样本特征所对应的样本标签。Step 406: Generate a sample label corresponding to the sample feature.
步骤408,根据样本特征与样本标签对标准跟踪模型进行训练,得到目标跟踪模型。Step 408: Train the standard tracking model according to the sample features and sample labels to obtain the target tracking model.
在调用目标跟踪模型对当前帧点云数据对应的图像特征进行处理之前,计算机设备还需要根据上一帧点云数据对标准跟踪模型进行训练,得到目标跟踪模型。Before calling the target tracking model to process the image features corresponding to the point cloud data of the current frame, the computer device also needs to train the standard tracking model according to the point cloud data of the previous frame to obtain the target tracking model.
具体的,计算机设备可以根据上一帧点云数据生成上一帧特征图,在上一帧特征图中截取样本特征图,提取样本特征图所对应的样本特征。可以理解的,由于计算机设备可以依次对激光传感器采集的点云数据进行跟踪,根据多帧点云数据迭代训练目标跟踪模型。因此,计算机设备根据上一帧点云数据生成上一帧特征图,在上一帧特征图中截取样本特征图,以及提取样本特征图所对应的样本特征的方式,可以与上述实施例中计算机设备根据当前帧点云数据生成当前帧特征图,在当前帧特征图周截取候选特征图,以及提取候选特征图中的图像特征的方式相同或者相似,故在此不再赘述。Specifically, the computer device may generate a feature map of the previous frame according to the point cloud data of the previous frame, cut a sample of the feature map of the previous frame of the feature map, and extract the sample feature corresponding to the sample feature map. It is understandable that since the computer device can track the point cloud data collected by the laser sensor in turn, iteratively trains the target tracking model based on multiple frames of point cloud data. Therefore, the computer device generates the feature map of the previous frame according to the point cloud data of the previous frame, cuts samples of the feature map of the previous frame of the feature map, and extracts the sample feature corresponding to the sample feature map. The device generates a feature map of the current frame according to the point cloud data of the current frame, intercepts candidate feature maps around the current frame feature map, and extracts image features in the candidate feature map in the same or similar manner, so we will not repeat them here.
计算机设备可以根据样本特征生成样本特征所对应的样本标签。具体的,计算机设备可以对样本特征进行循环移位,得到多个样本特征,根据多个样本特征生成样本特征矩阵。根据样本特征矩阵中每个样本特征所对应的移位 值确定样本特征对应的样本标签。移位值具体可以包括横轴移位值和纵轴移位值。计算机设备可以根据预设函数对样本特征矩阵中多个样本特征所对应的移位值进行运算,得到多个样本特征各自对应的样本标签。其中,预设函数可以是用户预先设置的函数,预设函数具体可以是二维高斯函数。The computer device can generate the sample label corresponding to the sample characteristic according to the sample characteristic. Specifically, the computer device may perform a cyclic shift on the sample features to obtain multiple sample features, and generate a sample feature matrix based on the multiple sample features. Determine the sample label corresponding to the sample feature according to the shift value corresponding to each sample feature in the sample feature matrix. The shift value may specifically include a horizontal axis shift value and a vertical axis shift value. The computer device can perform operations on the shift values corresponding to the multiple sample features in the sample feature matrix according to the preset function to obtain sample labels corresponding to each of the multiple sample features. The preset function may be a function preset by the user, and the preset function may specifically be a two-dimensional Gaussian function.
计算机设备可以根据样本特征矩阵中的多个样本特征,以及多个样本特征各自对应的样本标签对建立的标准跟踪模型进行训练,得到目标跟踪模型。The computer device can train the established standard tracking model according to the multiple sample features in the sample feature matrix and the sample labels corresponding to the multiple sample features to obtain the target tracking model.
在本实施例中,计算机设备可以根据上一帧点云数据对标准跟踪模型进行训练,得到目标跟踪模型,从而利用目标跟踪模型对当前帧点云数据的目标跟踪区域进行跟踪,对多帧点云数据迭代训练目标跟踪模型,有效的提高了目标跟踪的准确性。In this embodiment, the computer device can train the standard tracking model based on the point cloud data of the previous frame to obtain the target tracking model, so as to use the target tracking model to track the target tracking area of the point cloud data of the current frame, and perform multi-frame points. Cloud data iteratively trains the target tracking model, which effectively improves the accuracy of target tracking.
在其中一个实施例中,上述基于点云的目标跟踪方法还包括对当前帧点云数据进行检测,得到目标检测区域;根据目标检测区域与目标跟踪区域,确定当前帧点云数据所对应的当前帧目标区域。In one of the embodiments, the above-mentioned point cloud-based target tracking method further includes detecting the point cloud data of the current frame to obtain the target detection area; according to the target detection area and the target tracking area, determining the current frame point cloud data corresponding to the current Frame target area.
计算机设备可以根据点云数据,对当前帧点云数据进行检测,得到目标检测区域。计算机设备可以采用多种目标检测算法中的至少一种对当前帧点云数据进行目标检测,得到目标检测区域。计算机设备可以根据当前帧点云数据对应的目标检测区域和目标跟踪区域,确定当前帧点云数据所对应的当前帧目标区域。具体的,计算机设备可以将目标检测区域与目标跟踪区域进行比对。当目标检测区域与目标跟踪区域相同时,计算机设备可以确定目标检测区域所对应的区域作为当前帧点云数据对应的当前帧目标区域。当目标检测区域与目标跟踪区域不相同时,计算机设备可以综合目标检测区域与目标跟踪区域,确定综合区域作为当前帧点云数据对应的当前帧目标区域。The computer equipment can detect the point cloud data of the current frame according to the point cloud data to obtain the target detection area. The computer device may use at least one of multiple target detection algorithms to perform target detection on the point cloud data of the current frame to obtain the target detection area. The computer device can determine the current frame target area corresponding to the current frame point cloud data according to the target detection area and the target tracking area corresponding to the current frame point cloud data. Specifically, the computer device can compare the target detection area with the target tracking area. When the target detection area is the same as the target tracking area, the computer device can determine the area corresponding to the target detection area as the current frame target area corresponding to the current frame point cloud data. When the target detection area and the target tracking area are not the same, the computer device can synthesize the target detection area and the target tracking area, and determine the comprehensive area as the current frame target area corresponding to the current frame point cloud data.
在其中一个实施例中,计算机设备可以获取目标检测区域对应的检测置信度,以及目标跟踪区域对应的跟踪置信度。计算机设备可以基于检测置信度和跟踪置信度,综合目标检测区域和目标跟踪区域,确定综合区域作为当前帧点云数据对应的当前帧目标区域。In one of the embodiments, the computer device can obtain the detection confidence level corresponding to the target detection area and the tracking confidence level corresponding to the target tracking area. The computer device may integrate the target detection area and the target tracking area based on the detection confidence and the tracking confidence, and determine the comprehensive area as the current frame target area corresponding to the current frame point cloud data.
在本实施例中,计算机设备可以根据当前帧点云数据检测得到目标检测 区域,根据目标跟踪区域对目标检测区域进行调整,确定调整后的区域为当前帧点云数据对应的当前帧目标区域,有效的提高了确定目标区域的准确性。In this embodiment, the computer device can detect the target detection area according to the current frame point cloud data, adjust the target detection area according to the target tracking area, and determine that the adjusted area is the current frame target area corresponding to the current frame point cloud data. Effectively improve the accuracy of determining the target area.
在其中一个实施例中,上述基于点云的目标跟踪方法还包括根据当前帧目标区域与上一帧目标区域确定目标位移数据;获取点云采集频率;根据点云采集频率与目标位移数据确定目标运动数据。In one of the embodiments, the above-mentioned point cloud-based target tracking method further includes determining target displacement data according to the target area of the current frame and the target area of the previous frame; acquiring the point cloud acquisition frequency; and determining the target according to the point cloud acquisition frequency and the target displacement data Movement data.
计算机设备可以将当前帧目标区域与上一帧目标区域进行比对,根据比对结果确定目标位移数据。目标位移数据可以包括目标位移的长度以及方向。计算机设备可以获取激光传感器对应的点云采集频率。点云采集频率可以是用户根据实际需求预先设置的,激光传感器根据设置的点云采集频率采集点云数据。点云采集频率可以是一个常量。例如,激光传感器可以按照每秒50帧的频率采集点云数据。点云采集频率还可以是一个变量。例如,激光传感器可以根据不同的情况或模式调整点云采集频率。比如激光传感器可以在环境中目标较多并且运动速度较快的情况下增大点云采集频率,在环境中目标较少并且运动速度较慢的情况下减小点云采集频率。The computer device can compare the target area of the current frame with the target area of the previous frame, and determine the target displacement data according to the comparison result. The target displacement data may include the length and direction of the target displacement. The computer equipment can obtain the point cloud collection frequency corresponding to the laser sensor. The point cloud collection frequency can be preset by the user according to actual needs, and the laser sensor collects point cloud data according to the set point cloud collection frequency. The point cloud collection frequency can be a constant. For example, a laser sensor can collect point cloud data at a frequency of 50 frames per second. The point cloud collection frequency can also be a variable. For example, the laser sensor can adjust the point cloud collection frequency according to different situations or modes. For example, a laser sensor can increase the point cloud collection frequency when there are many targets in the environment and the movement speed is fast, and reduce the point cloud collection frequency when there are fewer targets in the environment and the movement speed is slow.
计算机设备可以根据获取到的点云采集频率,确定上一帧点云数据的采集时间与当前帧点云数据的采集时间之间的时间差。例如,当点云采集频率为每秒50帧时,计算机设备可以确定两帧之间的时间差为0.02秒。计算机设备可以根据时间差和目标位移数据确定目标对应的目标运动数据。目标运动数据具体可以包括目标对应的运动速度大小、方向等信息,以便计算机设备根据目标运动数据对无人驾驶设备进行提示或控制。The computer device can determine the time difference between the acquisition time of the point cloud data of the previous frame and the acquisition time of the point cloud data of the current frame according to the acquired point cloud acquisition frequency. For example, when the point cloud acquisition frequency is 50 frames per second, the computer device can determine that the time difference between the two frames is 0.02 seconds. The computer device can determine the target motion data corresponding to the target according to the time difference and the target displacement data. The target motion data may specifically include information such as the motion speed and direction corresponding to the target, so that the computer equipment can prompt or control the unmanned driving device according to the target motion data.
在本实施例中,计算机设备可以根据当前帧目标区域和上一帧目标区域确定目标位移数据,根据点云采集频率和目标位移数据确定目标运动数据,有助于计算机设备根据目标运动数据对无人驾驶设备进行提示或控制。In this embodiment, the computer device can determine the target displacement data according to the target area of the current frame and the target area of the previous frame, and determine the target motion data according to the point cloud collection frequency and the target displacement data, which helps the computer device to determine the target motion data according to the target motion data. The person drives the device to prompt or control.
应该理解的是,虽然图2-4的流程图中的各个步骤按照箭头的指示依次显示,但是这些步骤并不是必然按照箭头指示的顺序依次执行。除非本文中有明确的说明,这些步骤的执行并没有严格的顺序限制,这些步骤可以以其它的顺序执行。而且,图2-4中的至少一部分步骤可以包括多个子步骤或者 多个阶段,这些子步骤或者阶段并不必然是在同一时刻执行完成,而是可以在不同的时刻执行,这些子步骤或者阶段的执行顺序也不必然是依次进行,而是可以与其它步骤或者其它步骤的子步骤或者阶段的至少一部分轮流或者交替地执行。It should be understood that although the various steps in the flowcharts of FIGS. 2-4 are displayed in sequence as indicated by the arrows, these steps are not necessarily performed in sequence in the order indicated by the arrows. Unless specifically stated in this article, the execution of these steps is not strictly limited in order, and these steps can be executed in other orders. Moreover, at least part of the steps in Figures 2-4 may include multiple sub-steps or multiple stages. These sub-steps or stages are not necessarily executed at the same time, but can be executed at different times. These sub-steps or stages The execution order of is not necessarily performed sequentially, but may be performed alternately or alternately with at least a part of other steps or sub-steps or stages of other steps.
在其中一个实施例中,如图5所示,提供了一种基于点云的目标跟踪装置,包括:特征图生成模块502、候选区域获取模块504、特征提取模块506和目标跟踪模块508,其中:In one of the embodiments, as shown in FIG. 5, a point cloud-based target tracking device is provided, including: a feature map generation module 502, a candidate region acquisition module 504, a feature extraction module 506, and a target tracking module 508, where :
特征图生成模块502,用于根据当前帧点云数据生成当前帧特征图。The feature map generating module 502 is configured to generate a feature map of the current frame according to the point cloud data of the current frame.
候选区域获取模块504,用于获取当前帧特征图对应的候选区域;在当前帧特征图中截取与候选区域相匹配的候选特征图。The candidate region acquiring module 504 is used to acquire the candidate region corresponding to the feature map of the current frame; and intercept the candidate feature map matching the candidate region in the feature map of the current frame.
特征提取模块506,用于提取候选特征图所对应的图像特征。The feature extraction module 506 is used to extract image features corresponding to the candidate feature map.
目标跟踪模块508,用于调用基于上一帧点云数据训练得到的目标跟踪模型,根据图像特征确定当前帧点云数据对应的目标跟踪区域。The target tracking module 508 is configured to call the target tracking model trained based on the point cloud data of the previous frame, and determine the target tracking area corresponding to the point cloud data of the current frame according to the image characteristics.
在其中一个实施例中,上述特征图生成模块502还用于获取帧点云数据;将当前帧点云数据进行结构化处理,得到处理结果;基于处理结果对当前帧点云数据中的点进行编码,得到点对应的点特征;根据点特征生成当前帧点云数据对应的当前帧特征图。In one of the embodiments, the above-mentioned feature map generation module 502 is also used to obtain frame point cloud data; structure the current frame point cloud data to obtain the processing result; perform processing on the points in the current frame point cloud data based on the processing result Encode to obtain the point feature corresponding to the point; generate the current frame feature map corresponding to the point cloud data of the current frame according to the point feature.
在其中一个实施例中,上述特征图生成模块502还用于根据多个点特征生成点特征矩阵;调用图像生成模型,将点特征矩阵输入至图像生成模型;获取图像生成模型输出的当前帧特征图。In one of the embodiments, the above-mentioned feature map generation module 502 is further configured to generate a point feature matrix based on multiple point features; call the image generation model, and input the point feature matrix to the image generation model; obtain the current frame features output by the image generation model Figure.
在其中一个实施例中,上述候选区域获取模块504还用于获取上一帧点云数据对应的上一帧目标区域;根据预设倍数将上一帧目标区域进行扩大;确定扩大后的上一帧目标区域作为当前帧特征图对应的候选区域。In one of the embodiments, the above-mentioned candidate area acquisition module 504 is also used to acquire the target area of the previous frame corresponding to the point cloud data of the previous frame; expand the target area of the previous frame according to a preset multiple; and determine the expanded previous The frame target area is used as the candidate area corresponding to the feature map of the current frame.
在其中一个实施例中,上述特征提取模块506还用于获取特征提取模型;将候选特征图输入至特征提取模型;根据特征提取模型对候选特征图进行特征提取,得到候选特征图所对应的图像特征。In one of the embodiments, the feature extraction module 506 is also used to obtain a feature extraction model; input the candidate feature map to the feature extraction model; perform feature extraction on the candidate feature map according to the feature extraction model to obtain the image corresponding to the candidate feature map feature.
在其中一个实施例中,上述目标跟踪模块508还用于根据图像特征生成 图像特征矩阵;将图像特征矩阵输入至目标跟踪模型,获取目标跟踪模型输出的区域标签;根据区域标签确定当前帧点云数据所对应的目标跟踪区域。In one of the embodiments, the above-mentioned target tracking module 508 is also used to generate an image feature matrix based on image features; input the image feature matrix to the target tracking model to obtain the area label output by the target tracking model; determine the current frame point cloud according to the area label Target tracking area corresponding to the data.
在其中一个实施例中,上述基于点云的目标跟踪装置还包括模型训练模块,用于根据上一帧点云数据生成上一帧特征图;在上一帧特征图中截取样本特征图,提取样本特征图所对应的样本特征;生成与样本特征所对应的样本标签;根据样本特征与样本标签对标准跟踪模型进行训练,得到目标跟踪模型。In one of the embodiments, the above-mentioned point cloud-based target tracking device further includes a model training module, which is used to generate a feature map of the previous frame according to the point cloud data of the previous frame; Sample features corresponding to the sample feature map; generate sample labels corresponding to the sample features; train the standard tracking model according to the sample features and sample labels to obtain the target tracking model.
在其中一个实施例中,上述基于点云的目标跟踪装置还包括目标区域确定模块,用于对当前帧点云数据进行检测,得到目标检测区域;根据目标检测区域与目标跟踪区域,确定当前帧点云数据所对应的当前帧目标区域。In one of the embodiments, the above-mentioned point cloud-based target tracking device further includes a target area determining module for detecting the point cloud data of the current frame to obtain the target detection area; determining the current frame according to the target detection area and the target tracking area The target area of the current frame corresponding to the point cloud data.
在其中一个实施例中,上述基于点云的目标跟踪装置还包括目标数据确定模块,用于根据当前帧目标区域与上一帧目标区域确定目标位移数据;获取点云采集频率;根据点云采集频率与目标位移数据确定目标运动数据。In one of the embodiments, the above-mentioned point cloud-based target tracking device further includes a target data determining module for determining target displacement data according to the target area of the current frame and the target area of the previous frame; acquiring the point cloud collection frequency; The frequency and target displacement data determine the target motion data.
关于基于点云的目标跟踪装置的具体限定可以参见上文中对于基于点云的目标跟踪方法的限定,在此不再赘述。上述基于点云的目标跟踪装置中的各个模块可全部或部分通过软件、硬件及其组合来实现。上述各模块可以硬件形式内嵌于或独立于计算机设备中的处理器中,也可以以软件形式存储于计算机设备中的存储器中,以便于处理器调用执行以上各个模块对应的操作。For the specific definition of the point cloud-based target tracking device, please refer to the above definition of the point cloud-based target tracking method, which will not be repeated here. Each module in the above-mentioned point cloud-based target tracking device can be implemented in whole or in part by software, hardware, and a combination thereof. The above-mentioned modules may be embedded in the form of hardware or independent of the processor in the computer equipment, or may be stored in the memory of the computer equipment in the form of software, so that the processor can call and execute the operations corresponding to the above-mentioned modules.
在一个实施例中,提供了一种计算机设备,该计算机设备可以是服务器,其内部结构图可以如图6所示。该计算机设备包括通过系统总线连接的处理器、存储器、网络接口和数据库。其中,该计算机设备的处理器用于提供计算和控制能力。该计算机设备的存储器包括非易失性存储介质、内存储器。该非易失性存储介质存储有操作系统、计算机可读指令和数据库。该内存储器为非易失性存储介质中的操作系统和计算机可读指令的运行提供环境。该计算机设备的数据库用于存储基于点云的目标跟踪数据。该计算机设备的网络接口用于与外部的终端通过网络连接通信。该计算机可读指令被处理器执行时以实现一种基于点云的目标跟踪方法。In one embodiment, a computer device is provided. The computer device may be a server, and its internal structure diagram may be as shown in FIG. 6. The computer equipment includes a processor, a memory, a network interface, and a database connected through a system bus. Among them, the processor of the computer device is used to provide calculation and control capabilities. The memory of the computer device includes a non-volatile storage medium and an internal memory. The non-volatile storage medium stores an operating system, computer readable instructions, and a database. The internal memory provides an environment for the operation of the operating system and computer-readable instructions in the non-volatile storage medium. The database of the computer equipment is used to store the target tracking data based on the point cloud. The network interface of the computer device is used to communicate with an external terminal through a network connection. The computer-readable instructions are executed by the processor to realize a point cloud-based target tracking method.
本领域技术人员可以理解,图6中示出的结构,仅仅是与本申请方案相关的部分结构的框图,并不构成对本申请方案所应用于其上的计算机设备的限定,具体的计算机设备可以包括比图中所示更多或更少的部件,或者组合某些部件,或者具有不同的部件布置。Those skilled in the art can understand that the structure shown in FIG. 6 is only a block diagram of part of the structure related to the solution of the present application, and does not constitute a limitation on the computer device to which the solution of the present application is applied. The specific computer device may Including more or fewer parts than shown in the figure, or combining some parts, or having a different arrangement of parts.
在其中一个实施例中,提供了一种计算机设备,包括存储器和一个或多个处理器,存储器中储存有计算机可读指令,计算机可读指令被处理器执行时,使得一个或多个处理器执行时实现上述方法实施例中的步骤。In one of the embodiments, a computer device is provided, including a memory and one or more processors. The memory stores computer-readable instructions. When the computer-readable instructions are executed by the processor, the one or more processors When executed, the steps in the above method embodiments are implemented.
在其中一个实施例中,提供了一个或多个存储有计算机可读指令的非易失性计算机可读存储介质,计算机可读指令被一个或多个处理器执行时,使得一个或多个处理器执行时实现上述方法实施例中的步骤。In one of the embodiments, one or more non-volatile computer-readable storage media storing computer-readable instructions are provided. When the computer-readable instructions are executed by one or more processors, one or more processing The steps in the above method embodiments are implemented when the device is executed.
在其中一个实施例中,提供了一种交通工具,交通工具具体可以包括自动驾驶车辆、电动车、自行车以及飞行器等,交通工具包括上述计算机设备,可以执行上述基于点云的目标跟踪方法实施例中的步骤。In one of the embodiments, a vehicle is provided. The vehicle may specifically include self-driving vehicles, electric vehicles, bicycles, and aircraft. The vehicle includes the above-mentioned computer equipment and can execute the above-mentioned embodiment of the point cloud-based target tracking method. Steps in.
本发明创造的实施例、实施对象并不局限于自动驾驶车辆、电动车、自行车、飞行器、机器人等,也包括运用到与这些装置相关的仿真模拟装置、测试设备等。The embodiments and implementation objects created by the present invention are not limited to autonomous vehicles, electric vehicles, bicycles, aircrafts, robots, etc., but also include simulation devices and test equipment related to these devices.
本领域普通技术人员可以理解实现上述实施例方法中的全部或部分流程,是可以通过计算机可读指令来指令相关的硬件来完成,所述的计算机可读指令可存储于一非易失性计算机可读取存储介质中,该计算机可读指令在执行时,可包括如上述各方法的实施例的流程。其中,本申请所提供的各实施例中所使用的对存储器、存储、数据库或其它介质的任何引用,均可包括非易失性和/或易失性存储器。非易失性存储器可包括只读存储器(ROM)、可编程ROM(PROM)、电可编程ROM(EPROM)、电可擦除可编程ROM(EEPROM)或闪存。易失性存储器可包括随机存取存储器(RAM)或者外部高速缓冲存储器。作为说明而非局限,RAM以多种形式可得,诸如静态RAM(SRAM)、动态RAM(DRAM)、同步DRAM(SDRAM)、双数据率SDRAM(DDRSDRAM)、增强型SDRAM(ESDRAM)、同步链路 (Synchlink)DRAM(SLDRAM)、存储器总线(Rambus)直接RAM(RDRAM)、直接存储器总线动态RAM(DRDRAM)、以及存储器总线动态RAM(RDRAM)等。A person of ordinary skill in the art can understand that all or part of the processes in the above-mentioned embodiment methods can be implemented by instructing relevant hardware through computer-readable instructions. The computer-readable instructions can be stored in a non-volatile computer. In a readable storage medium, when the computer-readable instructions are executed, they may include the processes of the above-mentioned method embodiments. Wherein, any reference to memory, storage, database, or other media used in the embodiments provided in this application may include non-volatile and/or volatile memory. Non-volatile memory may include read only memory (ROM), programmable ROM (PROM), electrically programmable ROM (EPROM), electrically erasable programmable ROM (EEPROM), or flash memory. Volatile memory may include random access memory (RAM) or external cache memory. As an illustration and not a limitation, RAM is available in many forms, such as static RAM (SRAM), dynamic RAM (DRAM), synchronous DRAM (SDRAM), double data rate SDRAM (DDRSDRAM), enhanced SDRAM (ESDRAM), synchronous chain Channel (Synchlink) DRAM (SLDRAM), memory bus (Rambus) direct RAM (RDRAM), direct memory bus dynamic RAM (DRDRAM), and memory bus dynamic RAM (RDRAM), etc.
以上实施例的各技术特征可以进行任意的组合,为使描述简洁,未对上述实施例中的各个技术特征所有可能的组合都进行描述,然而,只要这些技术特征的组合不存在矛盾,都应当认为是本说明书记载的范围。The technical features of the above embodiments can be combined arbitrarily. In order to make the description concise, all possible combinations of the technical features in the above embodiments are not described. However, as long as there is no contradiction in the combination of these technical features, they should be It is considered as the range described in this specification.
以上所述实施例仅表达了本申请的几种实施方式,其描述较为具体和详细,但并不能因此而理解为对发明专利范围的限制。应当指出的是,对于本领域的普通技术人员来说,在不脱离本申请构思的前提下,还可以做出若干变形和改进,这些都属于本申请的保护范围。因此,本申请专利的保护范围应以所附权利要求为准。The above-mentioned embodiments only express several implementation manners of the present application, and the description is relatively specific and detailed, but it should not be understood as a limitation on the scope of the invention patent. It should be pointed out that for those of ordinary skill in the art, without departing from the concept of this application, several modifications and improvements can be made, and these all fall within the protection scope of this application. Therefore, the scope of protection of the patent of this application shall be subject to the appended claims.

Claims (20)

  1. 一种基于点云的目标跟踪方法,包括:A point cloud-based target tracking method includes:
    根据当前帧点云数据生成当前帧特征图;Generate a feature map of the current frame according to the point cloud data of the current frame;
    获取所述当前帧特征图对应的候选区域;Acquiring the candidate area corresponding to the feature map of the current frame;
    在所述当前帧特征图中截取与所述候选区域相匹配的候选特征图;Intercept a candidate feature map matching the candidate region in the current frame feature map;
    提取所述候选特征图所对应的图像特征;及Extracting the image feature corresponding to the candidate feature map; and
    调用基于上一帧点云数据训练得到的目标跟踪模型,根据所述图像特征确定所述当前帧点云数据对应的目标跟踪区域。The target tracking model trained based on the point cloud data of the previous frame is called, and the target tracking area corresponding to the point cloud data of the current frame is determined according to the image characteristics.
  2. 根据权利要求1所述的方法,其特征在于,所述根据当前帧点云数据生成当前帧特征图,包括:The method according to claim 1, wherein the generating a feature map of the current frame according to the point cloud data of the current frame comprises:
    获取所述当前帧点云数据;Acquiring the point cloud data of the current frame;
    将所述当前帧点云数据进行结构化处理,得到处理结果;Structured processing the point cloud data of the current frame to obtain a processing result;
    基于所述处理结果对所述当前帧点云数据中的点进行编码,得到所述点对应的点特征;及Encoding the points in the point cloud data of the current frame based on the processing result to obtain the point features corresponding to the points; and
    根据所述点特征生成所述当前帧点云数据对应的当前帧特征图。The current frame feature map corresponding to the current frame point cloud data is generated according to the point feature.
  3. 根据权利要求2所述的方法,其特征在于,所述根据所述点特征生成所述当前帧点云数据对应的当前帧特征图,包括:The method according to claim 2, wherein the generating a current frame feature map corresponding to the current frame point cloud data according to the point feature comprises:
    根据多个所述点特征生成点特征矩阵;Generating a point feature matrix according to a plurality of said point features;
    调用图像生成模型,将所述点特征矩阵输入至所述图像生成模型;及Calling an image generation model, and input the point feature matrix to the image generation model; and
    获取所述图像生成模型输出的当前帧特征图。Obtain the current frame feature map output by the image generation model.
  4. 根据权利要求1所述的方法,其特征在于,所述获取所述当前帧特征图对应的候选区域,包括:The method according to claim 1, wherein said obtaining the candidate area corresponding to the feature map of the current frame comprises:
    获取上一帧点云数据对应的上一帧目标区域;Obtain the target area of the previous frame corresponding to the point cloud data of the previous frame;
    根据预设倍数将所述上一帧目标区域进行扩大;及Expand the target area of the previous frame according to a preset multiple; and
    确定扩大后的上一帧目标区域作为所述当前帧特征图对应的候选区域。The expanded target area of the previous frame is determined as the candidate area corresponding to the feature map of the current frame.
  5. 根据权利要求1所述的方法,其特征在于,所述提取所述候选特征图所对应的图像特征,包括:The method according to claim 1, wherein said extracting the image feature corresponding to the candidate feature map comprises:
    获取特征提取模型;Obtain feature extraction model;
    将所述候选特征图输入至所述特征提取模型;及Input the candidate feature map to the feature extraction model; and
    根据所述特征提取模型对所述候选特征图进行特征提取,得到所述候选特征图所对应的图像特征。Perform feature extraction on the candidate feature map according to the feature extraction model to obtain the image feature corresponding to the candidate feature map.
  6. 根据权利要求1所述的方法,其特征在于,所述调用基于上一帧点云数据训练得到的目标跟踪模型,根据所述图像特征确定所述当前帧点云数据对应的目标跟踪区域,包括:The method according to claim 1, wherein the invoking the target tracking model trained based on the point cloud data of the previous frame, and determining the target tracking area corresponding to the point cloud data of the current frame according to the image features, comprises :
    根据所述图像特征生成图像特征矩阵;Generating an image feature matrix according to the image features;
    将所述图像特征矩阵输入至所述目标跟踪模型,获取所述目标跟踪模型输出的区域标签;及Input the image feature matrix to the target tracking model, and obtain the area label output by the target tracking model; and
    根据所述区域标签确定所述当前帧点云数据所对应的目标跟踪区域。The target tracking area corresponding to the point cloud data of the current frame is determined according to the area tag.
  7. 根据权利要求1所述的方法,其特征在于,在所述调用基于上一帧点云数据训练得到的目标跟踪模型之前,所述方法还包括:The method according to claim 1, characterized in that, before the invoking the target tracking model trained based on the point cloud data of the previous frame, the method further comprises:
    根据上一帧点云数据生成上一帧特征图;Generate the feature map of the previous frame according to the point cloud data of the previous frame;
    在所述上一帧特征图中截取样本特征图,提取所述样本特征图所对应的样本特征;Take a sample of the feature map from the previous frame of feature map, and extract the sample feature corresponding to the sample feature map;
    生成与所述样本特征所对应的样本标签;及Generate a sample label corresponding to the sample feature; and
    根据所述样本特征与所述样本标签对标准跟踪模型进行训练,得到目标跟踪模型。The standard tracking model is trained according to the sample feature and the sample label to obtain a target tracking model.
  8. 根据权利要求1所述的方法,其特征在于,所述方法还包括:The method according to claim 1, wherein the method further comprises:
    对所述当前帧点云数据进行检测,得到目标检测区域;及Detect the point cloud data of the current frame to obtain the target detection area; and
    根据所述目标检测区域与所述目标跟踪区域,确定所述当前帧点云数据所对应的当前帧目标区域。According to the target detection area and the target tracking area, the current frame target area corresponding to the current frame point cloud data is determined.
  9. 根据权利要求8所述的方法,其特征在于,所述方法还包括:The method according to claim 8, wherein the method further comprises:
    根据所述当前帧目标区域与上一帧目标区域确定目标位移数据;Determining target displacement data according to the target area of the current frame and the target area of the previous frame;
    获取点云采集频率;及Obtain the point cloud collection frequency; and
    根据所述点云采集频率与所述目标位移数据确定目标运动数据。Determine target motion data according to the point cloud collection frequency and the target displacement data.
  10. 一种基于点云的目标跟踪装置,包括:A point cloud-based target tracking device includes:
    特征图生成模块,用于根据当前帧点云数据生成当前帧特征图;The feature map generating module is used to generate the feature map of the current frame according to the point cloud data of the current frame;
    候选区域获取模块,用于获取所述当前帧特征图对应的候选区域;在所述当前帧特征图中截取与所述候选区域相匹配的候选特征图;A candidate region acquiring module, configured to acquire a candidate region corresponding to the current frame feature map; intercept a candidate feature map matching the candidate region in the current frame feature map;
    特征提取模块,用于提取所述候选特征图所对应的图像特征;及A feature extraction module for extracting image features corresponding to the candidate feature map; and
    目标跟踪模块,用于调用基于上一帧点云数据训练得到的目标跟踪模型,根据所述图像特征确定所述当前帧点云数据对应的目标跟踪区域。The target tracking module is used to call the target tracking model trained based on the point cloud data of the previous frame, and determine the target tracking area corresponding to the point cloud data of the current frame according to the image characteristics.
  11. 根据权利要求10所述的装置,其特征在于,所述特征图生成模块还用于获取所述当前帧点云数据;将所述当前帧点云数据进行结构化处理,得到处理结果;基于所述处理结果对所述当前帧点云数据中的点进行编码,得到所述点对应的点特征;及根据所述点特征生成所述当前帧点云数据对应的当前帧特征图。The device according to claim 10, wherein the feature map generation module is further configured to obtain the current frame point cloud data; structure the current frame point cloud data to obtain a processing result; The processing result encodes the points in the point cloud data of the current frame to obtain a point feature corresponding to the point; and according to the point feature, a current frame feature map corresponding to the point cloud data of the current frame is generated.
  12. 一种计算机设备,包括存储器及一个或多个处理器,所述存储器中储存有计算机可读指令,所述计算机可读指令被所述一个或多个处理器执行时,使得所述一个或多个处理器执行以下步骤:A computer device includes a memory and one or more processors. The memory stores computer-readable instructions. When the computer-readable instructions are executed by the one or more processors, the one or more Each processor performs the following steps:
    根据当前帧点云数据生成当前帧特征图;Generate a feature map of the current frame according to the point cloud data of the current frame;
    获取所述当前帧特征图对应的候选区域;Acquiring the candidate area corresponding to the feature map of the current frame;
    在所述当前帧特征图中截取与所述候选区域相匹配的候选特征图;Intercept a candidate feature map matching the candidate region in the current frame feature map;
    提取所述候选特征图所对应的图像特征;及Extracting the image feature corresponding to the candidate feature map; and
    调用基于上一帧点云数据训练得到的目标跟踪模型,根据所述图像特征确定所述当前帧点云数据对应的目标跟踪区域。The target tracking model trained based on the point cloud data of the previous frame is called, and the target tracking area corresponding to the point cloud data of the current frame is determined according to the image characteristics.
  13. 根据权利要求12所述的计算机设备,其特征在于,所述处理器执行所述计算机可读指令时还执行以下步骤:The computer device according to claim 12, wherein the processor further executes the following steps when executing the computer-readable instruction:
    获取所述当前帧点云数据;Acquiring the point cloud data of the current frame;
    将所述当前帧点云数据进行结构化处理,得到处理结果;Structured processing the point cloud data of the current frame to obtain a processing result;
    基于所述处理结果对所述当前帧点云数据中的点进行编码,得到所述点对应的点特征;及Encoding the points in the point cloud data of the current frame based on the processing result to obtain the point features corresponding to the points; and
    根据所述点特征生成所述当前帧点云数据对应的当前帧特征图。The current frame feature map corresponding to the current frame point cloud data is generated according to the point feature.
  14. 根据权利要求13所述的计算机设备,其特征在于,所述处理器执行所述计算机可读指令时还执行以下步骤:The computer device according to claim 13, wherein the processor further executes the following steps when executing the computer-readable instruction:
    根据多个所述点特征生成点特征矩阵;Generating a point feature matrix according to a plurality of said point features;
    调用图像生成模型,将所述点特征矩阵输入至所述图像生成模型;及Calling an image generation model, and input the point feature matrix to the image generation model; and
    获取所述图像生成模型输出的当前帧特征图。Obtain the current frame feature map output by the image generation model.
  15. 根据权利要求12所述的计算机设备,其特征在于,所述处理器执行所述计算机可读指令时还执行以下步骤:The computer device according to claim 12, wherein the processor further executes the following steps when executing the computer-readable instruction:
    获取上一帧点云数据对应的上一帧目标区域;Obtain the target area of the previous frame corresponding to the point cloud data of the previous frame;
    根据预设倍数将所述上一帧目标区域进行扩大;及Expand the target area of the previous frame according to a preset multiple; and
    确定扩大后的上一帧目标区域作为所述当前帧特征图对应的候选区域。The expanded target area of the previous frame is determined as the candidate area corresponding to the feature map of the current frame.
  16. 一个或多个存储有计算机可读指令的非易失性计算机可读存储介质,所述计算机可读指令被一个或多个处理器执行时,使得所述一个或多个处理器执行以下步骤:One or more non-volatile computer-readable storage media storing computer-readable instructions. When the computer-readable instructions are executed by one or more processors, the one or more processors execute the following steps:
    根据当前帧点云数据生成当前帧特征图;Generate a feature map of the current frame according to the point cloud data of the current frame;
    获取所述当前帧特征图对应的候选区域;Acquiring the candidate area corresponding to the feature map of the current frame;
    在所述当前帧特征图中截取与所述候选区域相匹配的候选特征图;Intercept a candidate feature map matching the candidate region in the current frame feature map;
    提取所述候选特征图所对应的图像特征;及Extracting the image feature corresponding to the candidate feature map; and
    调用基于上一帧点云数据训练得到的目标跟踪模型,根据所述图像特征确定所述当前帧点云数据对应的目标跟踪区域。The target tracking model trained based on the point cloud data of the previous frame is called, and the target tracking area corresponding to the point cloud data of the current frame is determined according to the image characteristics.
  17. 根据权利要求16所述的存储介质,其特征在于,所述计算机可读指令被所述处理器执行时还执行以下步骤:The storage medium according to claim 16, wherein the following steps are further executed when the computer-readable instructions are executed by the processor:
    获取所述当前帧点云数据;Acquiring the point cloud data of the current frame;
    将所述当前帧点云数据进行结构化处理,得到处理结果;Structured processing the point cloud data of the current frame to obtain a processing result;
    基于所述处理结果对所述当前帧点云数据中的点进行编码,得到所述点对应的点特征;及Encoding the points in the point cloud data of the current frame based on the processing result to obtain the point features corresponding to the points; and
    根据所述点特征生成所述当前帧点云数据对应的当前帧特征图。The current frame feature map corresponding to the current frame point cloud data is generated according to the point feature.
  18. 根据权利要求17所述的存储介质,其特征在于,所述计算机可读指令被所述处理器执行时还执行以下步骤:18. The storage medium according to claim 17, wherein the following steps are further executed when the computer-readable instructions are executed by the processor:
    根据多个所述点特征生成点特征矩阵;Generating a point feature matrix according to a plurality of said point features;
    调用图像生成模型,将所述点特征矩阵输入至所述图像生成模型;及Calling an image generation model, and input the point feature matrix to the image generation model; and
    获取所述图像生成模型输出的当前帧特征图。Obtain the current frame feature map output by the image generation model.
  19. 根据权利要求16所述的存储介质,其特征在于,所述计算机可读指令被所述处理器执行时还执行以下步骤:The storage medium according to claim 16, wherein the following steps are further executed when the computer-readable instructions are executed by the processor:
    获取上一帧点云数据对应的上一帧目标区域;Obtain the target area of the previous frame corresponding to the point cloud data of the previous frame;
    根据预设倍数将所述上一帧目标区域进行扩大;及Expand the target area of the previous frame according to a preset multiple; and
    确定扩大后的上一帧目标区域作为所述当前帧特征图对应的候选区域。The expanded target area of the previous frame is determined as the candidate area corresponding to the feature map of the current frame.
  20. 一种交通工具,包括执行根据权利要求1-9任一项所述的基于点云的目标跟踪方法。A vehicle, comprising executing the point cloud-based target tracking method according to any one of claims 1-9.
PCT/CN2019/130034 2019-12-30 2019-12-30 Point cloud-based target tracking method and apparatus, computer device and storage medium WO2021134258A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/CN2019/130034 WO2021134258A1 (en) 2019-12-30 2019-12-30 Point cloud-based target tracking method and apparatus, computer device and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2019/130034 WO2021134258A1 (en) 2019-12-30 2019-12-30 Point cloud-based target tracking method and apparatus, computer device and storage medium

Publications (1)

Publication Number Publication Date
WO2021134258A1 true WO2021134258A1 (en) 2021-07-08

Family

ID=76686049

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2019/130034 WO2021134258A1 (en) 2019-12-30 2019-12-30 Point cloud-based target tracking method and apparatus, computer device and storage medium

Country Status (1)

Country Link
WO (1) WO2021134258A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113516687A (en) * 2021-07-09 2021-10-19 东软睿驰汽车技术(沈阳)有限公司 Target tracking method, device, equipment and storage medium
CN113689471A (en) * 2021-09-09 2021-11-23 中国联合网络通信集团有限公司 Target tracking method and device, computer equipment and storage medium

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170213093A1 (en) * 2016-01-27 2017-07-27 Baidu Online Network Technology (Beijing) Co., Ltd. Method and apparatus for detecting vehicle contour based on point cloud data
CN107341819A (en) * 2017-05-09 2017-11-10 深圳市速腾聚创科技有限公司 Method for tracking target and storage medium
CN109271880A (en) * 2018-08-27 2019-01-25 深圳清创新科技有限公司 Vehicle checking method, device, computer equipment and storage medium
CN110412617A (en) * 2019-09-06 2019-11-05 李娜 It is a kind of based on the unmanned plane rescue mode of self feed back laser radar scanning and application
CN110533695A (en) * 2019-09-04 2019-12-03 深圳市唯特视科技有限公司 A kind of trajectory predictions device and method based on DS evidence theory

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170213093A1 (en) * 2016-01-27 2017-07-27 Baidu Online Network Technology (Beijing) Co., Ltd. Method and apparatus for detecting vehicle contour based on point cloud data
CN107341819A (en) * 2017-05-09 2017-11-10 深圳市速腾聚创科技有限公司 Method for tracking target and storage medium
CN109271880A (en) * 2018-08-27 2019-01-25 深圳清创新科技有限公司 Vehicle checking method, device, computer equipment and storage medium
CN110533695A (en) * 2019-09-04 2019-12-03 深圳市唯特视科技有限公司 A kind of trajectory predictions device and method based on DS evidence theory
CN110412617A (en) * 2019-09-06 2019-11-05 李娜 It is a kind of based on the unmanned plane rescue mode of self feed back laser radar scanning and application

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113516687A (en) * 2021-07-09 2021-10-19 东软睿驰汽车技术(沈阳)有限公司 Target tracking method, device, equipment and storage medium
CN113689471A (en) * 2021-09-09 2021-11-23 中国联合网络通信集团有限公司 Target tracking method and device, computer equipment and storage medium
CN113689471B (en) * 2021-09-09 2023-08-18 中国联合网络通信集团有限公司 Target tracking method, device, computer equipment and storage medium

Similar Documents

Publication Publication Date Title
CN111160302B (en) Obstacle information identification method and device based on automatic driving environment
CN111191600B (en) Obstacle detection method, obstacle detection device, computer device, and storage medium
CN110163904B (en) Object labeling method, movement control method, device, equipment and storage medium
CN111328396B (en) Pose estimation and model retrieval for objects in images
CN111666921B (en) Vehicle control method, apparatus, computer device, and computer-readable storage medium
WO2021134285A1 (en) Image tracking processing method and apparatus, and computer device and storage medium
WO2021134296A1 (en) Obstacle detection method and apparatus, and computer device and storage medium
WO2020094033A1 (en) Method and system for converting point cloud data for use with 2d convolutional neural networks
US11556745B2 (en) System and method for ordered representation and feature extraction for point clouds obtained by detection and ranging sensor
CN117636331A (en) Generating a three-dimensional bounding box from two-dimensional image and point cloud data
KR102140805B1 (en) Neural network learning method and apparatus for object detection of satellite images
CN111626314B (en) Classification method and device for point cloud data, computer equipment and storage medium
CN110298281B (en) Video structuring method and device, electronic equipment and storage medium
CN113378760A (en) Training target detection model and method and device for detecting target
WO2021114777A1 (en) Target detection method, terminal device, and medium
WO2021114776A1 (en) Object detection method, object detection device, terminal device, and medium
WO2021134258A1 (en) Point cloud-based target tracking method and apparatus, computer device and storage medium
Zelener et al. Cnn-based object segmentation in urban lidar with missing points
CN116740668B (en) Three-dimensional object detection method, three-dimensional object detection device, computer equipment and storage medium
CN111239684A (en) Binocular fast distance measurement method based on YoloV3 deep learning
Cheng et al. A simple ground segmentation method for LiDAR 3D point clouds
CN115457492A (en) Target detection method and device, computer equipment and storage medium
US20210357763A1 (en) Method and device for performing behavior prediction by using explainable self-focused attention
CN115588187B (en) Pedestrian detection method, device and equipment based on three-dimensional point cloud and storage medium
US20220301176A1 (en) Object detection method, object detection device, terminal device, and medium

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19958458

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 19958458

Country of ref document: EP

Kind code of ref document: A1