WO2022188663A1 - Target detection method and apparatus - Google Patents

Target detection method and apparatus Download PDF

Info

Publication number
WO2022188663A1
WO2022188663A1 PCT/CN2022/078611 CN2022078611W WO2022188663A1 WO 2022188663 A1 WO2022188663 A1 WO 2022188663A1 CN 2022078611 W CN2022078611 W CN 2022078611W WO 2022188663 A1 WO2022188663 A1 WO 2022188663A1
Authority
WO
WIPO (PCT)
Prior art keywords
target
point cloud
image
tracking trajectory
target tracking
Prior art date
Application number
PCT/CN2022/078611
Other languages
French (fr)
Chinese (zh)
Inventor
吴家俊
梁振宝
周伟
Original Assignee
华为技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 华为技术有限公司 filed Critical 华为技术有限公司
Publication of WO2022188663A1 publication Critical patent/WO2022188663A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/246Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T11/002D [Two Dimensional] image generation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T11/002D [Two Dimensional] image generation
    • G06T11/003Reconstruction from projections, e.g. tomography
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10028Range image; Depth image; 3D point clouds

Definitions

  • the embodiments of the present application relate to the field of intelligent driving, and in particular, to a target detection method and device.
  • Most current object detection methods are based on a single type of sensor, such as only relying on lidar to obtain point clouds or only relying on cameras to obtain images.
  • the point cloud can provide the three-dimensional information of the target and can better overcome the problem of mutual occlusion of the target, but the point cloud is relatively sparse, and the recognition rate of the target features is not high.
  • images have richer information, but images are greatly affected by lighting, weather, etc., and the reliability of detection and tracking is poor.
  • the image only has two-dimensional plane information, and the information of the occluded target cannot be obtained, which is easy to lose the target or cause errors.
  • Embodiments of the present application provide a target detection method and device, so as to improve the accuracy and real-time performance of target detection.
  • an embodiment of the present application provides a target detection method, the method includes: acquiring a point cloud from a three-dimensional scanning device and an image from a vision sensor; placing the point cloud and at least one target tracking trajectory on the point cloud
  • the three-dimensional space position of the predicted target is input into the target detection model for processing, and the three-dimensional space position of at least one first target is obtained, wherein the target detection model is based on a plurality of three-dimensional space positions of the predicted target corresponding to the known target tracking trajectory.
  • the point cloud samples and the three-dimensional space position detection results of the multiple targets corresponding to the multiple point cloud samples one-to-one are obtained by training; according to the projection and the three-dimensional space position of the at least one first target in the image.
  • the at least one target tracking trajectory predicts the two-dimensional spatial position of the target in the image, and determines the two-dimensional spatial position of the at least one second target in the image; according to the two-dimensional spatial position of the at least one second target,
  • the projection in the point cloud determines the three-dimensional space position of the at least one second target in the point cloud.
  • a target tracking trajectory feedback mechanism is added.
  • a target tracking trajectory feedback mechanism When performing target detection in point clouds and images, more attention is paid to the area where the target tracking trajectory is located in the point clouds and images where the predicted target position is located, which can effectively reduce leakage. to improve the accuracy of target detection.
  • the method further includes: according to the target feature corresponding to the at least one target tracking track and the target feature of the at least one second target, performing the tracking on the at least one target tracking track and the at least one target tracking track and the at least one target tracking track.
  • a second target is matched; the matched target tracking trajectory is associated with the second target.
  • the target features include one or more of the following: position, size, speed, direction, category, number of point cloud points, numerical distribution of coordinates in each direction of point cloud, distribution of point cloud reflection intensity, appearance feature, depth features, etc.
  • the detected target can be associated with the existing target tracking trajectory based on the target feature, which is conducive to obtaining a complete target tracking trajectory and predicting the position where the target will appear at the next moment.
  • the method further includes: for the second target that is not matched to the target tracking trajectory, establishing a target tracking trajectory corresponding to the second target.
  • a new ID can be given to the target, and a target tracking trajectory corresponding to the target can be established, which is conducive to tracking all the targets that appear.
  • the method further includes: for the target tracking trajectory that is not matched to the second target, comparing the target tracking trajectory and the target tracking trajectory on the point cloud and/or the target tracking trajectory. predicted target associations in the image.
  • the target tracking trajectory can be associated with the predicted target of the target tracking trajectory in the point cloud and/or image, which is beneficial to avoid leakage due to leakage. It can detect the problem that the same target corresponds to multiple target tracking trajectories, and improve the reliability of target tracking.
  • the target tracking trajectory and the target tracking trajectory are in the point cloud and/or the image.
  • the method further includes: when the number of times the target tracking trajectory is associated with the predicted target is greater than or equal to a first threshold, deleting the target tracking trajectory.
  • deleting the target tracking trajectories for which the corresponding target is not detected in the acquired point cloud and/or image for many times is beneficial to save processing resources.
  • the method further includes: acquiring a calibration object point cloud from a three-dimensional scanning device and a calibration object image from a vision sensor; The three-dimensional coordinates and the two-dimensional coordinates in the calibration object image determine the projection matrix of the point cloud coordinate system and the image coordinate system.
  • the three-dimensional scanning device and the visual sensor can be jointly calibrated by the calibration object, and the projection matrix of the point cloud coordinate system and the image coordinate system (also called the pixel coordinate system) can be determined, which is beneficial to the point cloud and image.
  • the target detection results are fused to improve the accuracy of target detection.
  • an embodiment of the present application provides a target detection device, the device has the function of implementing the first aspect or any possible method in the design of the first aspect, and the function can be implemented by hardware or by The hardware executes the corresponding software implementation.
  • the hardware or software includes one or more units (modules) corresponding to the above functions, such as an acquisition unit and a processing unit.
  • an embodiment of the present application provides a target detection apparatus, including at least one processor and an interface, where the processor is configured to call and run a computer program from the interface, and when the processor executes the computer program,
  • a target detection apparatus including at least one processor and an interface, where the processor is configured to call and run a computer program from the interface, and when the processor executes the computer program,
  • an embodiment of the present application provides a terminal, where the terminal includes the device described in the second aspect above.
  • the terminal may be a vehicle-mounted device, a vehicle, a monitoring controller, an unmanned aerial vehicle, a robot, a roadside unit, or the like.
  • the terminal may also be a smart device that needs to perform target detection or tracking, such as smart home and smart manufacturing.
  • an embodiment of the present application provides a chip system, the chip system includes: a processor and an interface, the processor is configured to call and run a computer program from the interface, and when the processor executes the computer program , the method described in the first aspect or any possible design of the first aspect can be implemented.
  • an embodiment of the present application provides a computer-readable storage medium, the computer-readable storage medium having a computer for executing the method described in the first aspect or any possible design of the first aspect program.
  • an embodiment of the present application further provides a computer program product, including a computer program or instruction, when the computer program or instruction is executed, the first aspect or any possible design of the first aspect can be implemented method described in.
  • FIG. 1 is a schematic diagram of a target detection system provided by an embodiment of the present application.
  • FIG. 2 is a schematic flowchart of a target detection process provided by an embodiment of the present application.
  • FIG. 3 is a schematic diagram of an intelligent driving scenario provided by an embodiment of the present application.
  • FIG. 4 is one of the schematic diagrams of a target detection solution based on multi-sensor fusion provided by an embodiment of the present application
  • FIG. 5 is the second schematic diagram of the target detection solution based on multi-sensor fusion provided by the embodiment of the present application.
  • FIG. 6 is a third schematic diagram of a target detection solution based on multi-sensor fusion provided by an embodiment of the present application.
  • FIG. 7 is a schematic process diagram of a target detection method provided by an embodiment of the present application.
  • FIG. 8 is one of the schematic diagrams of the target detection apparatus provided by the embodiment of the present application.
  • FIG. 9 is a second schematic diagram of a target detection apparatus provided by an embodiment of the present application.
  • FIG. 1 is a schematic diagram of a target detection system provided for the implementation of this application, including a data preprocessing module, a joint calibration module, a point cloud detection module, an image region of interest acquisition module, a point cloud domain prediction module, an image domain prediction module, and a prediction module. Decision-making module, data association module, trajectory management module.
  • the data preprocessing module is mainly used to filter point clouds, remove ground points, and perform distortion correction on images.
  • Joint calibration module It is mainly used to jointly calibrate the point cloud and image obtained by the 3D scanning device and the vision sensor, and obtain the projection matrix between the point cloud coordinate system and the image coordinate system.
  • Point cloud detection module It is mainly used to input the point cloud obtained at the current moment and the results of the feedback target tracking trajectory management (such as at least one target tracking trajectory predicting the three-dimensional space position of the target in the point cloud obtained at the current moment) into the trained In a good target detection model (such as a deep neural network model), the target detection results are obtained.
  • a good target detection model such as a deep neural network model
  • Image ROI acquisition module It is mainly used to project the target detection results obtained based on the point cloud into the image using the projection matrix, and combine the results of the feedback target tracking trajectory management (such as at least one target tracking trajectory obtained at the current moment). The two-dimensional spatial position of the predicted target in the image) to obtain the region of interest.
  • Prediction decision module It is mainly used to back-project the target detection result of the image to the point cloud, and compare it with the target detection result of the point cloud to decide a more accurate target detection result.
  • Data association module It is mainly used to associate and match the target detection result after the prediction decision and the target tracking trajectory.
  • Trajectory management module It is mainly used to manage and update all target tracking trajectories according to the data association results.
  • Point cloud domain prediction module It is mainly used to predict the three-dimensional space position of the target in the point cloud obtained by the target tracking trajectory based on the updated target tracking trajectory at the next moment.
  • Image domain prediction module It is mainly used to predict the two-dimensional spatial position of the target in the image obtained at the next moment based on the updated target tracking trajectory and predicting the target tracking trajectory.
  • the structure of the target detection system illustrated in the embodiments of the present application does not constitute a specific limitation on the target detection system.
  • the target detection system may include more or less modules than shown, or some modules may be combined, or some modules may be split, or different modules are arranged.
  • the target detection solution provided in the embodiment of the present application can be applied to a terminal to which the target detection system shown in FIG. 1 is applied, and the terminal can be a vehicle-mounted device, a vehicle, a monitoring controller, an unmanned aerial vehicle, a robot, a roadside unit ( Road side unit, RSU) and other equipment, suitable for monitoring, intelligent driving, drone navigation, robot travel and other scenarios.
  • a terminal to which the target detection system shown in FIG. 1 is applied in an intelligent driving scenario is used as an example for description.
  • a terminal (such as vehicle A) can obtain point clouds and images of the surrounding environment through the three-dimensional scanning device(s) and visual sensor(s) set on the terminal, and can monitor the surrounding environment.
  • Vehicles such as vehicle B, vehicle C, etc.
  • pedestrians, bicycles (not shown in the figure), trees (not shown in the figure) and other objects are detected and tracked.
  • the target detection solutions based on multi-sensor fusion mainly include the following:
  • the first scheme uses a deep convolutional neural network to detect the three-dimensional spatial position of the target and extract the point cloud features after obtaining the point cloud from the lidar.
  • Acquire an image from a monocular camera project the 3D boundary of the object detected from the point cloud to the image, and use a deep convolutional neural network to extract image features of the projected area.
  • the bipartite graph matching relationship between the target and the target tracking trajectory is combined with the Kalman filter to estimate the state of the target tracking trajectory, so as to achieve the tracking of the target in the point cloud.
  • this scheme uses a deep network for feature extraction in images and point clouds at the same time, which consumes more resources, has low computational efficiency, and is poorly implemented; and once there is a missed detection in the point cloud obtained based on lidar, it cannot be retrieved through the image. Missing target, low accuracy.
  • this scheme first uses the deep learning algorithm to obtain the target detection information in the collected images and point clouds.
  • this scheme uses the deep learning image target detection algorithm for the image to obtain the two-dimensional (2-dimension, 2D) detection frame category, the pixel coordinate position of the center point and the length and width size information of the target in the image; use the deep learning point cloud target detection for the point cloud
  • the algorithm obtains the information of the three-dimensional (3-dimension, 3D) detection frame type, the spatial coordinates of the center point and the length, width and height of the target in the point cloud.
  • the Hungarian algorithm is used to optimally match the detection frame of the image obtained at the adjacent moment and the target in the point cloud to achieve target tracking, and establish the target tracking trajectory of the image and the point cloud respectively.
  • this scheme uses deep learning algorithm for feature extraction in images and point clouds at the same time, which consumes more resources and has poor real-time performance; in addition, there is no real tracking algorithm, and the detection frame and Distance matching of detection boxes is error-prone.
  • the third scheme As shown in Figure 6, this scheme collects the point cloud of the target, filters the collected point cloud, outputs the ground object point data after filtering out the ground points, and maps the obtained ground object point data to generate distance Image and based on the reflection intensity image, perform point cloud segmentation and clustering on the object point data according to the distance image, reflection intensity image and echo intensity information to obtain a plurality of point cloud regions.
  • the target point cloud area of the suspected target is screened out from the point cloud area; the feature extraction is performed on each target point cloud area, and the extracted feature vector is used to classify the target to identify the target, and obtain the first target detection result.
  • the purpose of this application is to provide a target detection solution.
  • the target detection result in the point cloud is corrected by the target detection result in the image, and the target tracking trajectory feedback mechanism is used to reduce the missed detection rate and improve the accuracy and real-time performance of target detection. .
  • Point cloud the set of point data on the surface of the object scanned by the 3D scanning device can be called a point cloud.
  • a point cloud is a collection of vectors in a three-dimensional coordinate system. These vectors are usually expressed in the form of x, y, z three-dimensional coordinates, and are generally used to represent the outer surface shape of an object. Not only that, in addition to the geometric position information represented by (x, y, z), the point cloud can also represent the RGB color, gray value, depth, intensity of the object's reflective surface, etc. of a point.
  • the point cloud coordinate system involved in the embodiments of the present application is the three-dimensional (x, y, z) coordinate system where the point cloud points in the point cloud are located.
  • the image coordinate system also known as the pixel coordinate system, is usually a two-dimensional coordinate system established with the upper left corner of the image as the origin, and the unit is pixel.
  • the two coordinate axes of the image coordinate system consist of u and v.
  • the coordinates of a point in the image coordinate system can be identified as (u, v).
  • Corner points are points with particularly prominent attributes in a certain aspect, and refer to representative and robust points in point clouds and images, such as the intersection of two sides.
  • region of interest in image processing, the area to be processed is outlined from the processed image in the form of boxes, circles, ellipses, irregular polygons, etc., which is called the region of interest.
  • the region of interest may be considered as a region in an image where a target exists.
  • FIG. 7 is a schematic diagram of a target detection method provided by an embodiment of the present application, and the method includes:
  • S701 The terminal acquires the point cloud from the three-dimensional scanning device and the image from the vision sensor.
  • the three-dimensional scanning device can be a lidar, a millimeter-wave radar, a depth camera, etc.
  • the visual sensor can be a monocular camera, a multi-eye camera, and the like.
  • At least one three-dimensional scanning device and at least one visual sensor may be installed on the terminal, and the terminal may scan objects around the terminal (or in a certain direction, such as the direction of travel) through the three-dimensional scanning device, and collect The point cloud of objects around the terminal (or in a certain direction); it is also possible to scan the objects around the terminal (or in a certain direction) through the vision sensor, and collect images of the objects around the terminal (or in a certain direction).
  • the point cloud may be a collection of point cloud points, and the information of each point cloud point in the collection includes the three-dimensional coordinates (x, y, z) of the point cloud point.
  • the information of each point cloud point can also include information such as laser reflection intensity or millimeter wave reflection intensity.
  • the terminal when the terminal starts to initially acquire the point cloud from the 3D scanning device and the image from the vision sensor, it can also obtain the acquisition time of the point cloud and the image from the 3D scanning device and the visual sensor. Therefore, according to the acquisition time of the point cloud and the image, the point cloud and the image obtained from the 3D scanning device and the vision sensor are time-aligned to ensure that the same set of point clouds and images for target detection have the same acquisition time.
  • the terminal may further perform a data preprocessing operation on the point cloud and/or the image. For example, the terminal can filter the point cloud, remove the ground point cloud points, reduce the data volume of the point cloud, and improve the target detection efficiency; it can also be based on the internal and external parameters of the visual sensor (usually provided by the visual sensor manufacturer). The barrel distortion or pincushion distortion that exists in the collected image is corrected for distortion.
  • the terminal can remove the point cloud points that meet the above conditions in the above point cloud according to the pre-given conditions that the point cloud points belonging to the ground should meet (for example, the z-coordinate of the point cloud point is less than a certain threshold), The point cloud points on the ground are filtered out, thereby reducing the data volume of the point cloud and improving the efficiency of target detection.
  • S702 The terminal inputs the point cloud and the three-dimensional space position of the target predicted in the point cloud and the at least one target tracking trajectory into the target detection model for processing, and obtains the three-dimensional space position of at least one first target.
  • the three-dimensional space position of the target includes information such as center point coordinates, length, width and height, which can also be called a three-dimensional detection box or a three-dimensional bounding box (3D BBox).
  • the target detection model is based on the prediction corresponding to the known target tracking trajectory.
  • the multiple point cloud samples of the three-dimensional spatial position of the target and the three-dimensional spatial position detection results of the multiple targets corresponding to the multiple point cloud samples one-to-one are obtained by training.
  • a target tracking track corresponds to a target, and the target tracking track records information of the target, such as an identity document (ID), target characteristics, existence time, and each frame in which the target exists.
  • ID identity document
  • target characteristics target characteristics
  • existence time time
  • each frame in which the target exists The three-dimensional space position in the point cloud, the two-dimensional space position in each frame of image where the target exists, etc.
  • the target can be tracked in the point cloud by the Kalman algorithm, etc.
  • the target can be predicted in the next The three-dimensional space position that appears in the frame point cloud (that is, the point cloud collected at the next moment), that is, the target tracking trajectory can be obtained to predict the three-dimensional space position of the target in the next frame point cloud; Tracking in the image, according to the two-dimensional space position of the target in each frame of the target image in the target tracking track corresponding to the target, through the optical flow algorithm, etc. can predict the target in the next frame of image (that is, the next moment to collect the image.
  • the two-dimensional space position that appears in the image that is, the two-dimensional space position of the target tracking trajectory in the next frame image can be obtained.
  • the existing target tracking trajectory in the current point cloud predicts the three-dimensional space position of the target in the location area where the probability of the target appearing is significantly higher than that in other location areas in the point cloud.
  • the terminal can predict the three-dimensional space position of the target in the point cloud by processing the point cloud and at least one target tracking trajectory by the target detection model.
  • the target detection model can be a plurality of point cloud samples that predict the three-dimensional spatial position of the target based on the known target tracking trajectories maintained in the sample set by the training device, and a plurality of point cloud samples corresponding to the plurality of point cloud samples one-to-one.
  • the three-dimensional space position detection result of the target is obtained by training.
  • the training device can add a three-dimensional space position label vector (such as center point coordinates, length, width, height, etc.) to each point cloud sample according to the three-dimensional space position of the target corresponding to each point cloud sample. label vector of information).
  • a three-dimensional space position label vector such as center point coordinates, length, width, height, etc.
  • label vector of information can be added to each point cloud sample, which correspond to multiple targets one-to-one.
  • the spatial location label vector can also exist in the form of a matrix.
  • the training device can input the 3D space position of the predicted target corresponding to the point cloud sample and the target tracking track(s) into the target detection model for processing , obtain the predicted value of the three-dimensional space position of the target (one or more) output by the target detection model, according to the predicted value of the three-dimensional space position of the output target and the three-dimensional space position label vector of the real target corresponding to the point cloud sample, through the loss function ( loss function) training equipment can calculate the loss of the target detection model. Adjust the parameters in the target detection model according to the loss.
  • the training process of the target detection model becomes the process of reducing the loss as much as possible.
  • the target detection model is continuously trained through the point cloud samples in the sample set. When the loss is reduced to a preset range, the trained target detection model can be obtained.
  • the target detection model may be a deep neural network or the like.
  • the point cloud samples in the training set can be obtained by pre-sampling, such as pre-collecting point cloud samples through the terminal, and predicting the three-dimensional shape of the predicted target in the collected point cloud samples according to the target tracking trajectory (one or more).
  • the spatial position is recorded, and the three-dimensional spatial position of the real target existing in the point cloud sample is marked at the same time.
  • the above training equipment can be a personal computer (PC), a notebook computer, a server, etc., or a terminal. If the training equipment and the terminal are not the same equipment, after the training equipment has completed the training of the target detection model, the training completed can be used.
  • the target detection model is imported into the terminal, so that the terminal can detect the first target in the acquired point cloud.
  • S703 The terminal predicts the two-dimensional space position of the target in the image according to the projection of the three-dimensional space position of the at least one first target in the image and the at least one target tracking trajectory in the image, and determines the position of the target in the image. two-dimensional spatial location of at least one second object.
  • the three-dimensional space position in the point cloud can be projected into the image, and the two-dimensional space position in the image can be obtained.
  • the dimensional space position is projected into the point cloud, and the 3D space position in the point cloud is obtained.
  • the projection matrix for the determination of the projection matrix, several calibration objects (such as a three-dimensional carton with multiple edges and corners) can be preset and placed in the common field of view of the 3D scanning device and the vision sensor, and the calibration object points are collected by the 3D scanning device and the vision sensor.
  • Cloud and calibration object image select multiple calibration points (such as the corners of the three-dimensional carton) in the collected calibration object point cloud and calibration object image, and obtain the three-dimensional coordinates of the multiple calibration points in the calibration object point cloud and the calibration object.
  • the projection matrix of the point cloud coordinate system and the image coordinate system can be solved according to the three-dimensional coordinates of multiple calibration points in the calibration object point cloud and the two-dimensional coordinates in the calibration object image.
  • K is the internal parameter matrix of the visual sensor.
  • the internal parameter matrix of the visual sensor is fixed after leaving the factory and is usually provided by the manufacturer or obtained through a calibration algorithm.
  • [R, T] is the external parameter matrix of the visual sensor. 3)
  • the three-dimensional coordinates of the calibration point in the point cloud of the calibration object and the two-dimensional coordinates in the image of the calibration object, the projection matrix M from the point cloud coordinate system to the image coordinate system can be solved.
  • the terminal may also add feedback on the predicted target of the target tracking trajectory when detecting the second target in the image, and convert the two-dimensional spatial position obtained by the projection of at least one first target in the image.
  • at least one target tracking trajectory predicts the two-dimensional space position of the target in the image as a target, and the two-dimensional space position obtained by projection and the two-dimensional space position of the predicted target are output as the two-dimensional space position of the second target.
  • S704 The terminal determines the three-dimensional space position of the at least one second target in the point cloud according to the projection of the two-dimensional space position of the at least one second target in the point cloud.
  • the terminal projects the two-dimensional spatial position of the at least one second target in the image into the point cloud to obtain the three-dimensional spatial position of the at least one second target in the point cloud, and obtains the final target detection result output of the point cloud.
  • the features of the second target may include target features in a three-dimensional space position in the point cloud and target features in a two-dimensional space position in the image.
  • the target features in the three-dimensional space position in the point cloud may include position (such as center point coordinates), size (such as length, width and height), speed, direction, category, number of point cloud points, coordinate value distribution in each direction of point cloud, point cloud Reflection intensity distribution (such as point cloud reflection intensity distribution histogram), depth features, etc.
  • the target features of the two-dimensional space position in the image include position (center point coordinates), size (such as length and width), speed, direction, category, appearance Features (such as image color histogram, directional gradient histogram), etc.
  • a target tracking trajectory corresponds to a target, and the target tracking trajectory records the information of the target, such as ID, target characteristics, existence time, three-dimensional space position in each frame of point cloud where the target exists, The two-dimensional spatial position in each frame of images of the target, etc., in order to achieve the tracking of the same target, in some embodiments, the terminal can detect at least one second target according to the target feature corresponding to the existing at least one target tracking trajectory The target feature is matched with the at least one target tracking trajectory and the at least one second target. The second target matched to the target tracking trajectory is associated with the target tracking trajectory to improve the existing target tracking trajectory.
  • the matching degree (or similarity) between the target feature of the at least one target tracking trajectory and the target feature of the at least one second target can be used as the cost matrix, and the Hungarian algorithm can be used to analyze the at least one target tracking trajectory and the at least one target tracking trajectory.
  • the second objective performs global optimal matching.
  • the Hungarian algorithm is a combinatorial optimization algorithm that solves the task assignment problem in polynomial time.
  • the terminal When calculating the similarity between the target feature of the target tracking trajectory and the target feature of the second target, the terminal considers the position (in the point cloud and/or in the image), size (in the point cloud and/or in the image), speed (in the point cloud and/or in the image), direction (in the point cloud and/or in the image), category (in the point cloud and/or in the image), number of points in the point cloud, numerical distribution of coordinates in each direction of the point cloud, point cloud reflection
  • One or more of the target features such as intensity distribution, appearance feature, depth feature, etc. When multiple target features are considered, different target features can be assigned different weights, and the sum of the weighted values is 1.
  • the terminal can be the target. Assign a new target tracking track ID to create a new target tracking track.
  • the terminal can associate the target tracking trajectory with the predicted target of the target tracking trajectory in the point cloud and/or image, improve the target tracking trajectory, and avoid missed detections, etc.
  • the reason is that the same target corresponds to multiple target tracking trajectories.
  • the target tracking trajectory and the target tracking trajectory in the point cloud and/or image are Before predicting the target association, if the number of times the target tracking trajectory is associated with the predicted target is greater than or equal to the first threshold, the terminal deletes the target tracking trajectory.
  • the apparatus may include corresponding hardware structures and/or software modules for performing each function.
  • the present application can be implemented in hardware or a combination of hardware and computer software with the units and algorithm steps of each example described in conjunction with the embodiments disclosed herein. Whether a function is performed by hardware or computer software driving hardware depends on the specific application and design constraints of the technical solution. Skilled artisans may implement the described functionality using different methods for each particular application, but such implementations should not be considered beyond the scope of this application.
  • FIG. 8 shows a possible exemplary block diagram of the target detection apparatus involved in the embodiment of the present application, and the target detection apparatus 800 may exist in the form of a software module or a hardware module.
  • the target detection apparatus 800 may include: an acquisition unit 803 and a processing unit 802 .
  • the device may be a chip.
  • the apparatus 800 may further include a storage unit 801 for storing program codes and/or data of the apparatus 800 .
  • the acquiring unit 803 is configured to acquire the point cloud from the three-dimensional scanning device and the image from the vision sensor;
  • the processing unit 802 is configured to input the point cloud and at least one target tracking trajectory in the point cloud to predict the three-dimensional space position of the target into the target detection model for processing, and obtain the three-dimensional space position of at least one first target, wherein the three-dimensional space position of the first target is obtained.
  • the target detection model is obtained by training based on multiple point cloud samples of the three-dimensional spatial position of the predicted target corresponding to the known target tracking trajectory, and the three-dimensional spatial position detection results of the multiple targets corresponding to the multiple point cloud samples one-to-one. of;
  • the processing unit 802 is further configured to predict the two-dimensional spatial position of the target according to the projection of the three-dimensional spatial position of the at least one first target in the image and the at least one target tracking trajectory in the image, and determine the two-dimensional spatial position of at least one second object in the image;
  • the processing unit 802 is further configured to determine the three-dimensional spatial position of the at least one second target in the point cloud according to the projection of the two-dimensional spatial position of the at least one second target in the point cloud.
  • the processing unit 802 is further configured to, according to the target feature corresponding to the at least one target tracking track and the target feature of the at least one second target, perform the tracking of the at least one target tracking track and the The at least one second target is matched; the matched target tracking trajectory is associated with the second target.
  • the processing unit 802 is further configured to establish a target tracking trajectory corresponding to the second target for the second target that is not matched to the target tracking trajectory.
  • the processing unit 802 is further configured to, for the target tracking trajectory that is not matched to the second target, place the target tracking trajectory and the target tracking trajectory in the point cloud and/or predicted target associations in the image.
  • the processing unit 802 compares the target tracking trajectory with the target tracking trajectory in the point cloud and/or the target tracking trajectory for the target tracking trajectory that is not matched to the second target. Before being associated with the predicted target in the image, it is also used for deleting the target tracking trajectory when the number of times the target tracking trajectory is associated with the predicted target is greater than or equal to a first threshold.
  • the target features include one or more of the following: position, length, width, height, speed, direction, category, number of point cloud points, coordinate value distribution in each direction of the point cloud, and point cloud reflection Intensity distribution, appearance features, depth features.
  • the acquiring unit 803 is further configured to acquire the calibration object point cloud from the three-dimensional scanning device and the calibration object image from the vision sensor;
  • the processing unit 802 is further configured to determine a point cloud coordinate system and an image coordinate system according to the three-dimensional coordinates of a plurality of calibration points in the calibration object in the calibration object point cloud and the two-dimensional coordinates in the calibration object image projection matrix.
  • an embodiment of the present application further provides a target detection apparatus 900 .
  • the target detection apparatus 900 includes at least one processor 902 and an interface circuit. Further, the apparatus further includes at least one memory 901 , and the at least one memory 901 is connected to the processor 902 .
  • the interface circuit is used to provide input and output of data and/or information for the at least one processor.
  • the memory 901 is used to store the computer-executed instructions.
  • the processor 902 executes the computer-executed instructions stored in the memory 901, so that the target detection device 900 can realize the above-mentioned target detection method.
  • the target detection device 900 executes the computer-executed instructions stored in the memory 901, so that the target detection device 900 can realize the above-mentioned target detection method.
  • a computer-readable storage medium on which a program or an instruction is stored, and when the program or instruction is executed, the target detection method in the above method embodiment can be executed.
  • a computer program product including an instruction is provided, and when the instruction is executed, the target detection method in the above method embodiment can be executed.
  • a chip is provided.
  • the chip can be coupled with a memory and is used to call a computer program product stored in the memory to implement the target detection method in the above method embodiments.
  • the embodiments of the present application may be provided as a method, a system, or a computer program product. Accordingly, the present application may take the form of an entirely hardware embodiment, an entirely software embodiment, or an embodiment combining software and hardware aspects. Furthermore, the present application may take the form of a computer program product embodied on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, etc.) having computer-usable program code embodied therein.
  • computer-usable storage media including, but not limited to, disk storage, CD-ROM, optical storage, etc.
  • These computer program instructions may also be stored in a computer-readable memory capable of directing a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory result in an article of manufacture comprising instruction means, the instructions
  • the apparatus implements the functions specified in the flow or flow of the flowcharts and/or the block or blocks of the block diagrams.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • General Engineering & Computer Science (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Biomedical Technology (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Length Measuring Devices By Optical Means (AREA)
  • Image Analysis (AREA)

Abstract

The present application relates to the field of intelligent driving. Disclosed are a target detection method and apparatus, which are used for improving the accuracy and real-time performance of target detection. The method comprises: acquiring a point cloud from a three-dimensional scanning device and an image from a visual sensor; inputting, into a target detection model, the point cloud, and the three-dimensional spatial position of a predicted target of at least one target tracking trajectory in the point cloud, and processing same, so as to obtain the three-dimensional spatial position of at least one first target; according to the projection of the three-dimensional spatial position of the at least one first target in the image and the two-dimensional spatial position of the predicted target of the at least one target tracking trajectory in the image, determining the two-dimensional spatial position of at least one second target in the image; and according to the projection of the two-dimensional spatial position of the at least one second target in the point cloud, determining the three-dimensional spatial position of the at least one second target in the point cloud.

Description

一种目标检测方法及装置A target detection method and device
相关申请的交叉引用CROSS-REFERENCE TO RELATED APPLICATIONS
本申请要求在2021年03月09日提交中华人民共和国知识产权局、申请号为202110256851.2、申请名称为“一种目标检测方法及装置”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。This application claims the priority of the Chinese patent application filed on March 09, 2021 with the Intellectual Property Office of the People's Republic of China, the application number is 202110256851.2, and the application name is "a target detection method and device", the entire contents of which are incorporated by reference in in this application.
技术领域technical field
本申请实施例涉及智能驾驶领域,尤其涉及一种目标检测方法及装置。The embodiments of the present application relate to the field of intelligent driving, and in particular, to a target detection method and device.
背景技术Background technique
随着城市的发展,交通越来越拥堵,人们驾车越来越趋于疲劳。为了满足人们的出行要求,智能驾驶(包括辅助驾驶、无人驾驶)应运而生。而如何可靠的实现对环境中的目标检测,对于智能驾驶的决策至关重要。With the development of the city, the traffic becomes more and more congested, and people tend to be more and more tired when driving. In order to meet people's travel requirements, intelligent driving (including assisted driving and unmanned driving) emerges as the times require. How to reliably detect objects in the environment is crucial to the decision-making of intelligent driving.
目前大多数目标检测方法都是基于单一类型传感器,例如仅依赖激光雷达获取点云或者仅依赖相机获取图像。点云能提供目标的三维信息,能较好地克服目标相互遮挡问题,但点云较为稀疏,对目标特征的识别率不高。而图像相比点云具有更丰富的信息,但图像受光照、天气等的影响较大,检测和跟踪的可靠性较差。而且图像只有二维平面信息,无法获取被遮挡的目标的信息,容易丢失目标或造成错误。融合点云和图像能充分发挥点云和图像的互补性,提高检测的鲁棒性。但是,目前对多传感器融合的目标检测研究较少,目标检测的准确性和实时性有待提升。Most current object detection methods are based on a single type of sensor, such as only relying on lidar to obtain point clouds or only relying on cameras to obtain images. The point cloud can provide the three-dimensional information of the target and can better overcome the problem of mutual occlusion of the target, but the point cloud is relatively sparse, and the recognition rate of the target features is not high. Compared with point clouds, images have richer information, but images are greatly affected by lighting, weather, etc., and the reliability of detection and tracking is poor. Moreover, the image only has two-dimensional plane information, and the information of the occluded target cannot be obtained, which is easy to lose the target or cause errors. The fusion of point cloud and image can give full play to the complementarity of point cloud and image, and improve the robustness of detection. However, there are few researches on target detection of multi-sensor fusion at present, and the accuracy and real-time performance of target detection need to be improved.
发明内容SUMMARY OF THE INVENTION
本申请实施例提供一种目标检测方法及装置,用以提高目标检测的准确性和实时性。Embodiments of the present application provide a target detection method and device, so as to improve the accuracy and real-time performance of target detection.
第一方面,本申请实施例提供一种目标检测方法,该方法包括:获取来自三维扫描设备的点云和来自视觉传感器的图像;将所述点云和至少一个目标跟踪轨迹在所述点云中预测目标的三维空间位置输入到目标检测模型进行处理,得到至少一个第一目标的三维空间位置,其中所述目标检测模型是基于已知目标跟踪轨迹对应的预测目标的三维空间位置的多个点云样本,以及与所述多个点云样本一一对应的多个目标的三维空间位置检测结果训练得到的;根据所述至少一个第一目标的三维空间位置在所述图像中的投影和所述至少一个目标跟踪轨迹在所述图像中预测目标的二维空间位置,确定所述图像中至少一个第二目标的二维空间位置;根据所述至少一个第二目标的二维空间位置在所述点云中的投影,确定所述点云中所述至少一个第二目标的三维空间位置。In a first aspect, an embodiment of the present application provides a target detection method, the method includes: acquiring a point cloud from a three-dimensional scanning device and an image from a vision sensor; placing the point cloud and at least one target tracking trajectory on the point cloud The three-dimensional space position of the predicted target is input into the target detection model for processing, and the three-dimensional space position of at least one first target is obtained, wherein the target detection model is based on a plurality of three-dimensional space positions of the predicted target corresponding to the known target tracking trajectory. The point cloud samples and the three-dimensional space position detection results of the multiple targets corresponding to the multiple point cloud samples one-to-one are obtained by training; according to the projection and the three-dimensional space position of the at least one first target in the image. The at least one target tracking trajectory predicts the two-dimensional spatial position of the target in the image, and determines the two-dimensional spatial position of the at least one second target in the image; according to the two-dimensional spatial position of the at least one second target, The projection in the point cloud determines the three-dimensional space position of the at least one second target in the point cloud.
在本申请实施例中,增加了目标跟踪轨迹反馈机制,在点云和图像中进行目标检测时,更关注目标跟踪轨迹在点云和图像中的预测目标位置所在的区域,可以有效减小漏检,提高目标检测的准确性。In the embodiment of the present application, a target tracking trajectory feedback mechanism is added. When performing target detection in point clouds and images, more attention is paid to the area where the target tracking trajectory is located in the point clouds and images where the predicted target position is located, which can effectively reduce leakage. to improve the accuracy of target detection.
在一种可能的设计中,所述方法还包括:根据所述至少一个目标跟踪轨迹对应的目标特征以及所述至少一个第二目标的目标特征,对所述至少一个目标跟踪轨迹和所述至少一 个第二目标进行匹配;将匹配的所述目标跟踪轨迹和所述第二目标关联。可选的,所述目标特征包括以下中的一项或多项:位置、尺寸、速度、方向、类别、点云点数、点云各方向坐标数值分布、点云反射强度分布、外观特征、深度特征等。In a possible design, the method further includes: according to the target feature corresponding to the at least one target tracking track and the target feature of the at least one second target, performing the tracking on the at least one target tracking track and the at least one target tracking track and the at least one target tracking track. A second target is matched; the matched target tracking trajectory is associated with the second target. Optionally, the target features include one or more of the following: position, size, speed, direction, category, number of point cloud points, numerical distribution of coordinates in each direction of point cloud, distribution of point cloud reflection intensity, appearance feature, depth features, etc.
上述设计中,可以基于目标特征将检测到的目标与已有的目标跟踪轨迹关联起来,有利于获取完整的目标跟踪轨迹,以及对目标下一时刻即将出现的位置进行预测。In the above design, the detected target can be associated with the existing target tracking trajectory based on the target feature, which is conducive to obtaining a complete target tracking trajectory and predicting the position where the target will appear at the next moment.
在一种可能的设计中,所述方法还包括:对于未匹配到所述目标跟踪轨迹的所述第二目标,建立所述第二目标对应的目标跟踪轨迹。In a possible design, the method further includes: for the second target that is not matched to the target tracking trajectory, establishing a target tracking trajectory corresponding to the second target.
上述设计中,对于新出现的目标,可以赋予该目标一个新的ID,并建立该目标对应的目标跟踪轨迹,有利于实现对出现的所有目标进行跟踪。In the above design, for a newly appeared target, a new ID can be given to the target, and a target tracking trajectory corresponding to the target can be established, which is conducive to tracking all the targets that appear.
在一种可能的设计中,所述方法还包括:对于未匹配到所述第二目标的所述目标跟踪轨迹,将所述目标跟踪轨迹与所述目标跟踪轨迹在所述点云和/或所述图像中的预测目标关联。In a possible design, the method further includes: for the target tracking trajectory that is not matched to the second target, comparing the target tracking trajectory and the target tracking trajectory on the point cloud and/or the target tracking trajectory. predicted target associations in the image.
上述设计中,对于未在点云和图像中检测到对应目标的目标跟踪轨迹,可以将该目标跟踪轨迹与该目标跟踪轨迹在点云和/或图像中的预测目标关联,有利于避免因漏检造成同一目标对应多个目标跟踪轨迹的问题,提高目标跟踪的可靠性。In the above design, for the target tracking trajectory that does not detect the corresponding target in the point cloud and image, the target tracking trajectory can be associated with the predicted target of the target tracking trajectory in the point cloud and/or image, which is beneficial to avoid leakage due to leakage. It can detect the problem that the same target corresponds to multiple target tracking trajectories, and improve the reliability of target tracking.
在一种可能的设计中,所述对于未匹配到所述第二目标的所述目标跟踪轨迹,将所述目标跟踪轨迹与所述目标跟踪轨迹在所述点云和/或所述图像中的预测目标关联之前,所述方法还包括:当所述目标跟踪轨迹关联预测目标的次数大于或等于第一阈值时,删除所述目标跟踪轨迹。In a possible design, for the target tracking trajectory that is not matched to the second target, the target tracking trajectory and the target tracking trajectory are in the point cloud and/or the image. Before being associated with the predicted target, the method further includes: when the number of times the target tracking trajectory is associated with the predicted target is greater than or equal to a first threshold, deleting the target tracking trajectory.
上述设计中,对于多次未在获取的点云和/或图像中检测到对应目标的目标跟踪轨迹进行删除,有利于节约处理资源。In the above design, deleting the target tracking trajectories for which the corresponding target is not detected in the acquired point cloud and/or image for many times is beneficial to save processing resources.
在一种可能的设计中,所述方法还包括:获取来自三维扫描设备的标定物点云和来自视觉传感器的标定物图像;根据标定物中多个标定点在所述标定物点云中的三维坐标以及在所述标定物图像中的二维坐标,确定点云坐标系和图像坐标系的投影矩阵。In a possible design, the method further includes: acquiring a calibration object point cloud from a three-dimensional scanning device and a calibration object image from a vision sensor; The three-dimensional coordinates and the two-dimensional coordinates in the calibration object image determine the projection matrix of the point cloud coordinate system and the image coordinate system.
上述设计中,可以通过标定物对三维扫描设备和视觉传感器进行联合标定,确定点云坐标系和图像坐标系(也可以称为像素坐标系)的投影矩阵,有利于对点云和图像中的目标检测结果进行融合,提高目标检测的准确性。In the above design, the three-dimensional scanning device and the visual sensor can be jointly calibrated by the calibration object, and the projection matrix of the point cloud coordinate system and the image coordinate system (also called the pixel coordinate system) can be determined, which is beneficial to the point cloud and image. The target detection results are fused to improve the accuracy of target detection.
第二方面,本申请实施例提供一种目标检测装置,该装置具有实现上述第一方面或者第一方面的任一种可能的设计中方法的功能,所述功能可以通过硬件实现,也可以通过硬件执行相应的软件实现。所述硬件或软件包括一个或多个与上述功能相对应的单元(模块),比如包括获取单元和处理单元。In the second aspect, an embodiment of the present application provides a target detection device, the device has the function of implementing the first aspect or any possible method in the design of the first aspect, and the function can be implemented by hardware or by The hardware executes the corresponding software implementation. The hardware or software includes one or more units (modules) corresponding to the above functions, such as an acquisition unit and a processing unit.
第三方面,本申请实施例提供一种目标检测装置,包括至少一个处理器和接口,所述处理器用于从所述接口调用并运行计算机程序,当所述处理器执行所述计算机程序时,可以实现上述第一方面或者第一方面的任一种可能的设计中所述的方法。In a third aspect, an embodiment of the present application provides a target detection apparatus, including at least one processor and an interface, where the processor is configured to call and run a computer program from the interface, and when the processor executes the computer program, The method described in the above first aspect or any possible design of the first aspect can be implemented.
第四方面,本申请实施例提供一种终端,该终端包括上述第二方面所述的装置。可选的,该终端可以为车载设备、车辆、监控控制器、无人机、机器人、路侧单元等。或者,所述终端也可以为智能家居、智能制造等需要进行目标探测或跟踪的智能设备。In a fourth aspect, an embodiment of the present application provides a terminal, where the terminal includes the device described in the second aspect above. Optionally, the terminal may be a vehicle-mounted device, a vehicle, a monitoring controller, an unmanned aerial vehicle, a robot, a roadside unit, or the like. Alternatively, the terminal may also be a smart device that needs to perform target detection or tracking, such as smart home and smart manufacturing.
第五方面,本申请实施例提供一种芯片系统,所述芯片系统包括:处理器和接口,所述处理器用于从所述接口调用并运行计算机程序,当所述处理器执行所述计算机程序时,可以实现上述第一方面或者第一方面的任一种可能的设计中所述的方法。In a fifth aspect, an embodiment of the present application provides a chip system, the chip system includes: a processor and an interface, the processor is configured to call and run a computer program from the interface, and when the processor executes the computer program , the method described in the first aspect or any possible design of the first aspect can be implemented.
第六方面,本申请实施例提供一种计算机可读存储介质,所述计算机可读存储介质具有用于执行上述第一方面或者第一方面的任一种可能的设计中所述的方法的计算机程序。In a sixth aspect, an embodiment of the present application provides a computer-readable storage medium, the computer-readable storage medium having a computer for executing the method described in the first aspect or any possible design of the first aspect program.
第七方面,本申请实施例还提供一种计算机程序产品,包括计算机程序或指令,当所述计算机程序或指令被执行时,可以实现上述第一方面或者第一方面的任一种可能的设计中所述的方法。In a seventh aspect, an embodiment of the present application further provides a computer program product, including a computer program or instruction, when the computer program or instruction is executed, the first aspect or any possible design of the first aspect can be implemented method described in.
上述第二方面至第七方面所能达到的技术效果请参照上述第一方面所能达到的技术效果,这里不再重复赘述。For the technical effects that can be achieved by the second aspect to the seventh aspect, please refer to the technical effects that can be achieved by the first aspect, which will not be repeated here.
附图说明Description of drawings
图1为本申请实施例提供的目标检测系统的示意图;1 is a schematic diagram of a target detection system provided by an embodiment of the present application;
图2为本申请实施例提供的目标检测流程示意图;2 is a schematic flowchart of a target detection process provided by an embodiment of the present application;
图3为本申请实施例提供的智能驾驶场景示意图;3 is a schematic diagram of an intelligent driving scenario provided by an embodiment of the present application;
图4为本申请实施例提供的基于多传感器融合目标检测方案示意图之一;FIG. 4 is one of the schematic diagrams of a target detection solution based on multi-sensor fusion provided by an embodiment of the present application;
图5为本申请实施例提供的基于多传感器融合目标检测方案示意图之二;FIG. 5 is the second schematic diagram of the target detection solution based on multi-sensor fusion provided by the embodiment of the present application;
图6为本申请实施例提供的基于多传感器融合目标检测方案示意图之三;FIG. 6 is a third schematic diagram of a target detection solution based on multi-sensor fusion provided by an embodiment of the present application;
图7为本申请实施例提供的目标检测方法过程示意图;FIG. 7 is a schematic process diagram of a target detection method provided by an embodiment of the present application;
图8为本申请实施例提供的目标检测装置示意图之一;FIG. 8 is one of the schematic diagrams of the target detection apparatus provided by the embodiment of the present application;
图9为本申请实施例提供的目标检测装置示意图之二。FIG. 9 is a second schematic diagram of a target detection apparatus provided by an embodiment of the present application.
具体实施方式Detailed ways
图1为本申请实施提供的一种目标检测系统的示意图,包括数据预处理模块、联合标定模块、点云检测模块、图像感兴趣区域获取模块、点云域预测模块、图像域预测模块、预测决策模块、数据关联模块、轨迹管理模块。1 is a schematic diagram of a target detection system provided for the implementation of this application, including a data preprocessing module, a joint calibration module, a point cloud detection module, an image region of interest acquisition module, a point cloud domain prediction module, an image domain prediction module, and a prediction module. Decision-making module, data association module, trajectory management module.
结合如图2所示的目标检测流程示意图,数据预处理模块:主要用于对点云滤波,去除地面点;对图像进行畸变矫正等。Combined with the schematic diagram of the target detection process shown in Figure 2, the data preprocessing module is mainly used to filter point clouds, remove ground points, and perform distortion correction on images.
联合标定模块:主要用于对三维扫描设备和视觉传感器获取的点云和图像进行联合标定,获取点云坐标系和图像坐标系之间的投影矩阵。Joint calibration module: It is mainly used to jointly calibrate the point cloud and image obtained by the 3D scanning device and the vision sensor, and obtain the projection matrix between the point cloud coordinate system and the image coordinate system.
点云检测模块:主要用于将当前时刻获取的点云和反馈的目标跟踪轨迹管理后的结果(如至少一个目标跟踪轨迹在当前时刻获取的点云中预测目标的三维空间位置)输入已训练好的目标检测模型(如深度神经网络模型)中,获得目标检测结果。Point cloud detection module: It is mainly used to input the point cloud obtained at the current moment and the results of the feedback target tracking trajectory management (such as at least one target tracking trajectory predicting the three-dimensional space position of the target in the point cloud obtained at the current moment) into the trained In a good target detection model (such as a deep neural network model), the target detection results are obtained.
图像感兴趣区域获取模块:主要用于使用投影矩阵将基于点云获取的目标检测结果投影到图像中,并结合反馈的目标跟踪轨迹管理后的结果(如至少一个目标跟踪轨迹在当前时刻获取的图像中预测目标的二维空间位置),获得感兴趣区域。Image ROI acquisition module: It is mainly used to project the target detection results obtained based on the point cloud into the image using the projection matrix, and combine the results of the feedback target tracking trajectory management (such as at least one target tracking trajectory obtained at the current moment). The two-dimensional spatial position of the predicted target in the image) to obtain the region of interest.
预测决策模块:主要用于将图像的目标检测结果反投影到点云,并和点云的目标检测结果对比,决策出更准确的目标检测结果。Prediction decision module: It is mainly used to back-project the target detection result of the image to the point cloud, and compare it with the target detection result of the point cloud to decide a more accurate target detection result.
数据关联模块:主要用于将预测决策后的目标检测结果和目标跟踪轨迹进行关联匹配。Data association module: It is mainly used to associate and match the target detection result after the prediction decision and the target tracking trajectory.
轨迹管理模块:主要用于根据数据关联结果,对所有目标跟踪轨迹进行管理更新。Trajectory management module: It is mainly used to manage and update all target tracking trajectories according to the data association results.
点云域预测模块:主要用于对基于更新后的目标跟踪轨迹,预测目标跟踪轨迹在下一时刻获取的点云中预测目标的三维空间位置。Point cloud domain prediction module: It is mainly used to predict the three-dimensional space position of the target in the point cloud obtained by the target tracking trajectory based on the updated target tracking trajectory at the next moment.
图像域预测模块:主要用于对基于更新后的目标跟踪轨迹,预测目标跟踪轨迹在下一 时刻获取的图像中预测目标的二维空间位置。Image domain prediction module: It is mainly used to predict the two-dimensional spatial position of the target in the image obtained at the next moment based on the updated target tracking trajectory and predicting the target tracking trajectory.
可以理解的是,本申请实施例示意的目标检测系统的结构并不构成对目标检测系统的具体限定。在本申请另一些实施例中,目标检测系统可以包括比图示更多或更少的模块,或者组合某些模块,或者拆分某些模块,或者不同的模块布置。It can be understood that the structure of the target detection system illustrated in the embodiments of the present application does not constitute a specific limitation on the target detection system. In other embodiments of the present application, the target detection system may include more or less modules than shown, or some modules may be combined, or some modules may be split, or different modules are arranged.
本申请实施例提供的目标检测方案,可以适用于应用有如图1所示的目标检测系统的终端,所述终端可以为车载设备、车辆、监控控制器、无人机、机器人、路侧单元(road side unit,RSU)等设备,适用于监控、智能驾驶、无人机航行、机器人行进等场景。在本申请实施例的后续说明中,以智能驾驶场景下应用有如图1所示的目标检测系统的终端为例进行说明。如图3所示,终端(如车辆A)可以通过终端上设置的三维扫描设备(一个或多个)和视觉传感器(一个或多个)获取周围环境的点云和图像,并可以对周围环境中的车辆(如车辆B、车辆C等)、行人、自行车(图中未示出)、树木(图中未示出)等目标进行检测和跟踪。The target detection solution provided in the embodiment of the present application can be applied to a terminal to which the target detection system shown in FIG. 1 is applied, and the terminal can be a vehicle-mounted device, a vehicle, a monitoring controller, an unmanned aerial vehicle, a robot, a roadside unit ( Road side unit, RSU) and other equipment, suitable for monitoring, intelligent driving, drone navigation, robot travel and other scenarios. In the subsequent description of the embodiments of the present application, a terminal to which the target detection system shown in FIG. 1 is applied in an intelligent driving scenario is used as an example for description. As shown in Figure 3, a terminal (such as vehicle A) can obtain point clouds and images of the surrounding environment through the three-dimensional scanning device(s) and visual sensor(s) set on the terminal, and can monitor the surrounding environment. Vehicles (such as vehicle B, vehicle C, etc.), pedestrians, bicycles (not shown in the figure), trees (not shown in the figure) and other objects are detected and tracked.
目前基于多传感器融合目标检测方案主要包括以下几种:At present, the target detection solutions based on multi-sensor fusion mainly include the following:
第一种方案:如图4所示,该方案从激光雷达获取点云后,使用深度卷积神经网络检测出目标的三维空间位置并提取点云特征。从单目相机获取图像,将从点云中检测出的目标的三维边界投影到图像,并使用深度卷积神经网络提取投影区域的图像特征。接着,计算检测到的目标和目标跟踪轨迹在点云三维空间位置、点云特征以及图像特征上的相似度矩阵,并对三个相似度矩阵进行合并,将合并的相似度矩阵通过匈牙利算法计算目标和目标跟踪轨迹之间的二分图匹配关系,并结合卡尔曼滤波器对目标跟踪轨迹进行状态估计,从而实现对点云中目标的跟踪。然而该方案同时在图像和点云中使用深度网络进行特征提取,资源消耗较多,计算效率较低,实现性差;并且一旦基于激光雷达获取的点云中出现漏检,则无法通过图像找回漏检目标,准确性较低。The first scheme: As shown in Figure 4, this scheme uses a deep convolutional neural network to detect the three-dimensional spatial position of the target and extract the point cloud features after obtaining the point cloud from the lidar. Acquire an image from a monocular camera, project the 3D boundary of the object detected from the point cloud to the image, and use a deep convolutional neural network to extract image features of the projected area. Next, calculate the similarity matrix of the detected target and the target tracking trajectory in the three-dimensional space position of the point cloud, the point cloud feature and the image feature, and merge the three similarity matrices, and the combined similarity matrix is calculated by the Hungarian algorithm. The bipartite graph matching relationship between the target and the target tracking trajectory is combined with the Kalman filter to estimate the state of the target tracking trajectory, so as to achieve the tracking of the target in the point cloud. However, this scheme uses a deep network for feature extraction in images and point clouds at the same time, which consumes more resources, has low computational efficiency, and is poorly implemented; and once there is a missed detection in the point cloud obtained based on lidar, it cannot be retrieved through the image. Missing target, low accuracy.
第二种方案:如图5所示,该方案首先利用深度学习算法获取采集的图像和点云中的目标检测信息。例如:对图像采用深度学习图像目标检测算法获取图像中目标的二维(2-dimension,2D)检测框类别、中心点像素坐标位置及长宽尺寸信息;对点云采用深度学习点云目标检测算法获取点云中目标的三维(3-dimension,3D)检测框类别、中心点空间坐标及长宽高尺寸的信息。然后,利用匈牙利算法,基于检测框之间的最小距离,分别对相邻时刻获取的图像和点云中目标的检测框做最优匹配,实现目标跟踪,分别建立图像及点云的目标跟踪轨迹。然而该方案同时在图像和点云中使用深度学习算法进行特征提取,资源消耗较多,实时性差;另外没有真正的跟踪算法,在目标密集或相邻时刻间位移较大的目标使用检测框和检测框的距离匹配容易出错。The second scheme: As shown in Figure 5, this scheme first uses the deep learning algorithm to obtain the target detection information in the collected images and point clouds. For example: use the deep learning image target detection algorithm for the image to obtain the two-dimensional (2-dimension, 2D) detection frame category, the pixel coordinate position of the center point and the length and width size information of the target in the image; use the deep learning point cloud target detection for the point cloud The algorithm obtains the information of the three-dimensional (3-dimension, 3D) detection frame type, the spatial coordinates of the center point and the length, width and height of the target in the point cloud. Then, based on the minimum distance between the detection frames, the Hungarian algorithm is used to optimally match the detection frame of the image obtained at the adjacent moment and the target in the point cloud to achieve target tracking, and establish the target tracking trajectory of the image and the point cloud respectively. . However, this scheme uses deep learning algorithm for feature extraction in images and point clouds at the same time, which consumes more resources and has poor real-time performance; in addition, there is no real tracking algorithm, and the detection frame and Distance matching of detection boxes is error-prone.
第三种方案:如图6所示,该方案采集目标的点云,对采集到的点云进行滤波,输出滤除地面点后的地物点数据,将得到的地物点数据映射生成距离图像以及基于反射强度图像,根据所述距离图像、反射强度图像以及回波强度信息对所述地物点数据进行点云分割聚类,得到多个点云区域,根据目标的先验知识从各个点云区域中筛选出疑似目标的目的点云区域;对各目的点云区域进行特征提取,由提取的特征向量进行分类以识别出目标,得到第一目标检测结果。采集到图像,对图像进行预处理,对预处理后的图像使用投影变换矩阵提取感兴趣区域,在感兴趣区域中进行图像特征提取,根据提取的图像特征识别出目标,得到第二目标检测结果。若所述第一目标检测结果、第二目标检测结果相同,将第 一目标检测结果或第二目标检测结果作为最终的目标检测结果输出,若第一目标检测结果、第二目标检测结果不相同,基于贝叶斯决策对第一目标检测结果、第二目标检测结果进行融合判决,得到最终的目标检测结果输出。最后,使用基于马尔科夫决策过程(markov decision procss,MDP)的多目标跟踪方法进行跟踪。然而,基于点云的目标检测依赖大量先验知识准确度较差。点云漏检时,则无法通过图像找回漏检目标,准确性较低。The third scheme: As shown in Figure 6, this scheme collects the point cloud of the target, filters the collected point cloud, outputs the ground object point data after filtering out the ground points, and maps the obtained ground object point data to generate distance Image and based on the reflection intensity image, perform point cloud segmentation and clustering on the object point data according to the distance image, reflection intensity image and echo intensity information to obtain a plurality of point cloud regions. The target point cloud area of the suspected target is screened out from the point cloud area; the feature extraction is performed on each target point cloud area, and the extracted feature vector is used to classify the target to identify the target, and obtain the first target detection result. Collect the image, preprocess the image, use the projection transformation matrix to extract the region of interest from the preprocessed image, perform image feature extraction in the region of interest, identify the target according to the extracted image features, and obtain the second target detection result . If the first target detection result and the second target detection result are the same, the first target detection result or the second target detection result is output as the final target detection result, if the first target detection result and the second target detection result are different , based on Bayesian decision-making, the first target detection result and the second target detection result are fused and judged to obtain the final target detection result output. Finally, a multi-target tracking method based on Markov decision process (MDP) is used for tracking. However, point cloud-based object detection relies on a large amount of prior knowledge with poor accuracy. When the point cloud is missed, the missed target cannot be retrieved through the image, and the accuracy is low.
本申请旨在提供一种目标检测方案,通过图像中的目标检测结果对点云中的目标检测结果进行修正,并使用目标跟踪轨迹反馈机制降低漏检率,提高目标检测的准确性以及实时性。The purpose of this application is to provide a target detection solution. The target detection result in the point cloud is corrected by the target detection result in the image, and the target tracking trajectory feedback mechanism is used to reduce the missed detection rate and improve the accuracy and real-time performance of target detection. .
在介绍本申请实施例之前,首先对本申请实施例中的部分用语进行解释说明,以便于本领域技术人员理解。Before introducing the embodiments of the present application, some terms in the embodiments of the present application will be explained first, so as to facilitate the understanding of those skilled in the art.
1)、点云,通过三维扫描设备扫描得到的物体表面上的点数据集合可称之为点云(point cloud)。点云是在一个三维坐标系统中的一组向量的集合。这些向量通常以x,y,z三维坐标的形式表示,而且一般主要用来代表一个物体的外表面形状。不仅如此,除(x,y,z)代表的几何位置信息之外,点云还可以表示一个点的RGB颜色,灰度值,深度,物体反射面强度等。在本申请实施例中涉及的点云坐标系即点云中点云点所在的三维(x,y,z)坐标系。1) Point cloud, the set of point data on the surface of the object scanned by the 3D scanning device can be called a point cloud. A point cloud is a collection of vectors in a three-dimensional coordinate system. These vectors are usually expressed in the form of x, y, z three-dimensional coordinates, and are generally used to represent the outer surface shape of an object. Not only that, in addition to the geometric position information represented by (x, y, z), the point cloud can also represent the RGB color, gray value, depth, intensity of the object's reflective surface, etc. of a point. The point cloud coordinate system involved in the embodiments of the present application is the three-dimensional (x, y, z) coordinate system where the point cloud points in the point cloud are located.
2)、图像坐标系,也可以称为像素坐标系,通常是以图像左上角为原点建立的二维坐标系,单位为像素(pixel)。所述图像坐标系的两个坐标轴由u和v构成。所述图像坐标系中某个点的坐标可以标识为(u,v)。2) The image coordinate system, also known as the pixel coordinate system, is usually a two-dimensional coordinate system established with the upper left corner of the image as the origin, and the unit is pixel. The two coordinate axes of the image coordinate system consist of u and v. The coordinates of a point in the image coordinate system can be identified as (u, v).
3)、角点,角点即在某方面属性特别突出的点,在点云和图像中指具有代表性以及稳健性的点,如两条边的交点等。3) Corner points, corner points are points with particularly prominent attributes in a certain aspect, and refer to representative and robust points in point clouds and images, such as the intersection of two sides.
4)、感兴趣区域(region of interesting,ROI),图像处理中,从被处理的图像以方框、圆、椭圆、不规则多边形等方式勾勒出需要处理的区域,称为感兴趣区域,在本申请实施例中感兴趣区域可以认为是图像中存在目标的区域。4), region of interest (ROI), in image processing, the area to be processed is outlined from the processed image in the form of boxes, circles, ellipses, irregular polygons, etc., which is called the region of interest. In this embodiment of the present application, the region of interest may be considered as a region in an image where a target exists.
另外,需要理解的是,本申请中,“至少一个”是指一个或者多个,“多个”是指两个或两个以上。“和/或”,描述关联对象的关联关系,表示可以存在三种关系,例如,A和/或B,可以表示:单独存在A,同时存在A和B,单独存在B的情况,其中A,B可以是单数或者复数。在本申请的文字描述中,字符“/”,一般表示前后关联对象是一种“或”的关系。另外,除非有相反的说明,本申请实施例提及“第一”、“第二”等序数词用于对多个对象进行区分,不用于限定多个对象的顺序、时序、优先级或者重要程度,并且“第一”、“第二”的描述也并不限定对象一定不同。在本申请中涉及的各种数字编号仅为描述方便进行的区分,并不用来限制本申请的实施例的范围。上述各过程的序号的大小并不意味着执行顺序的先后,各过程的执行顺序应以其功能和内在逻辑确定。在本申请中,“示例性的”或者“例如”等词用于表示例子、例证或说明,被描述为“示例性的”或者“例如”的任何实施例或设计方案不应被解释为比其它实施例或设计方案更优选或更具优势。使用“示例性的”或者“例如”等词旨在以具体方式呈现相关概念,便于理解。下面将结合附图,对本申请实施例进行详细描述。In addition, it should be understood that, in this application, "at least one" refers to one or more, and "a plurality" refers to two or more. "And/or", which describes the association relationship of the associated objects, indicates that there can be three kinds of relationships, for example, A and/or B, which can indicate: the existence of A alone, the existence of A and B at the same time, and the existence of B alone, where A, B can be singular or plural. In the text description of this application, the character "/" generally indicates that the contextual objects are in an "or" relationship. In addition, unless otherwise stated, the ordinal numbers such as “first” and “second” mentioned in the embodiments of the present application are used to distinguish multiple objects, and are not used to limit the order, sequence, priority, or importance of multiple objects. degree, and the descriptions of "first" and "second" do not limit the objects to be necessarily different. Various numerical numbers involved in the present application are only for the convenience of description, and are not used to limit the scope of the embodiments of the present application. The size of the sequence numbers of the above processes does not imply the sequence of execution, and the execution sequence of each process should be determined by its function and internal logic. In this application, the words "exemplary" or "such as" are used to mean an example, illustration, or illustration, and any embodiment or design described as "exemplary" or "such as" should not be construed as Other embodiments or designs are more preferred or advantageous. The use of words such as "exemplary" or "such as" is intended to present the relevant concepts in a specific manner to facilitate understanding. The embodiments of the present application will be described in detail below with reference to the accompanying drawings.
图7为本申请实施例提供的一种目标检测方法示意图,该方法包括:FIG. 7 is a schematic diagram of a target detection method provided by an embodiment of the present application, and the method includes:
S701:终端获取来自三维扫描设备的点云和来自视觉传感器的图像。S701: The terminal acquires the point cloud from the three-dimensional scanning device and the image from the vision sensor.
其中,三维扫描设备可以是激光雷达、毫米波雷达、深度相机等,视觉传感器可以为 单目摄像头、多目摄像头等。Among them, the three-dimensional scanning device can be a lidar, a millimeter-wave radar, a depth camera, etc., and the visual sensor can be a monocular camera, a multi-eye camera, and the like.
在一种可能的实施中,终端上可安装至少一个三维扫描设备以及至少一个视觉传感器,终端可以通过三维扫描设备对终端周围(或某个方向上,如行进方向上)的物体进行扫描,采集终端周围(或某个方向上)的物体的点云;也可以通过视觉传感器对终端周围(或某个方向上)的物体进行扫描,采集终端周围(或某个方向上)的物体的图像。其中,点云可以是点云点的集合,该集合中的每个点云点的信息包括点云点的三维坐标(x,y,z),在三维扫描设备为激光雷达或毫米波雷达时,每个点云点的信息还可以包括激光反射强度或毫米波反射强度等信息。In a possible implementation, at least one three-dimensional scanning device and at least one visual sensor may be installed on the terminal, and the terminal may scan objects around the terminal (or in a certain direction, such as the direction of travel) through the three-dimensional scanning device, and collect The point cloud of objects around the terminal (or in a certain direction); it is also possible to scan the objects around the terminal (or in a certain direction) through the vision sensor, and collect images of the objects around the terminal (or in a certain direction). The point cloud may be a collection of point cloud points, and the information of each point cloud point in the collection includes the three-dimensional coordinates (x, y, z) of the point cloud point. When the three-dimensional scanning device is a lidar or a millimeter-wave radar , the information of each point cloud point can also include information such as laser reflection intensity or millimeter wave reflection intensity.
另外,为了避免点云和图像的采集时间不一致,终端启动初始获取来自三维扫描设备的点云和来自视觉传感器的图像时,还可以从三维扫描设备和视觉传感器获取点云和图像的采集时间,从而根据点云和图像的采集时间,将从三维扫描设备和视觉传感器获取的点云和图像进行时间对齐,保证进行目标检测的同一组点云和图像的采集时间一致。In addition, in order to avoid inconsistency in the acquisition time of the point cloud and the image, when the terminal starts to initially acquire the point cloud from the 3D scanning device and the image from the vision sensor, it can also obtain the acquisition time of the point cloud and the image from the 3D scanning device and the visual sensor. Therefore, according to the acquisition time of the point cloud and the image, the point cloud and the image obtained from the 3D scanning device and the vision sensor are time-aligned to ensure that the same set of point clouds and images for target detection have the same acquisition time.
在一些实施中,获取到点云和图像后,终端还可以对点云和/或图像进行数据预处理操作。举例来说,终端可以对点云进行滤波,去除地面点云点,降低点云的数据量,提高目标检测效率;也可以根据视觉传感器的内参和外参(通常由视觉传感器生产厂商提供)对采集的图像存在的桶形畸变或枕形畸变等进行畸变矫正等。In some implementations, after acquiring the point cloud and the image, the terminal may further perform a data preprocessing operation on the point cloud and/or the image. For example, the terminal can filter the point cloud, remove the ground point cloud points, reduce the data volume of the point cloud, and improve the target detection efficiency; it can also be based on the internal and external parameters of the visual sensor (usually provided by the visual sensor manufacturer). The barrel distortion or pincushion distortion that exists in the collected image is corrected for distortion.
作为一种示例,终端可以根据预先给出的属于地面的点云点应该满足的条件(如点云点的z坐标小于某一阈值),将上述点云中满足上述条件的点云点去除,滤除地面的点云点,从而降低点云的数据量,提高目标检测效率。As an example, the terminal can remove the point cloud points that meet the above conditions in the above point cloud according to the pre-given conditions that the point cloud points belonging to the ground should meet (for example, the z-coordinate of the point cloud point is less than a certain threshold), The point cloud points on the ground are filtered out, thereby reducing the data volume of the point cloud and improving the efficiency of target detection.
S702:所述终端将所述点云和至少一个目标跟踪轨迹在所述点云中预测目标的三维空间位置输入到目标检测模型进行处理,得到至少一个第一目标的三维空间位置。S702: The terminal inputs the point cloud and the three-dimensional space position of the target predicted in the point cloud and the at least one target tracking trajectory into the target detection model for processing, and obtains the three-dimensional space position of at least one first target.
其中,目标的三维空间位置包括中心点坐标、长宽高尺寸等信息,也可以称为三维检测框或三维边界盒(3D BBox),所述目标检测模型是基于已知目标跟踪轨迹对应的预测目标的三维空间位置的多个点云样本,以及与所述多个点云样本一一对应的多个目标的三维空间位置检测结果训练得到的。Among them, the three-dimensional space position of the target includes information such as center point coordinates, length, width and height, which can also be called a three-dimensional detection box or a three-dimensional bounding box (3D BBox). The target detection model is based on the prediction corresponding to the known target tracking trajectory. The multiple point cloud samples of the three-dimensional spatial position of the target and the three-dimensional spatial position detection results of the multiple targets corresponding to the multiple point cloud samples one-to-one are obtained by training.
在本申请实施例中,一个目标跟踪轨迹对应一个目标,目标跟踪轨迹记录有该目标的信息,如身份标识号(identity document,ID)、目标特征、存在时间、在存在该目标的每一帧点云中的三维空间位置、在存在该目标的每一帧图像中的二维空间位置等。通过卡尔曼(kalman)算法等可以对目标在点云中进行跟踪,根据目标对应的目标跟踪轨迹中存在该目标的每一帧点云中该目标的三维空间位置,即可预测该目标在下一帧点云(即下一时刻采集的点云)中出现的三维空间位置,也即可以得到该目标跟踪轨迹在下一帧点云中预测目标的三维空间位置;通过光流算法等可以对目标在图像中进行跟踪,根据目标对应的目标跟踪轨迹中存在该目标的每一帧图像中该目标的二维空间位置,通过光流算法等即可预测该目标在下一帧图像(即下一时刻采集的图像)中出现的二维空间位置,也即可以得到该目标跟踪轨迹在下一帧图像中预测目标的二维空间位置。In the embodiment of the present application, a target tracking track corresponds to a target, and the target tracking track records information of the target, such as an identity document (ID), target characteristics, existence time, and each frame in which the target exists. The three-dimensional space position in the point cloud, the two-dimensional space position in each frame of image where the target exists, etc. The target can be tracked in the point cloud by the Kalman algorithm, etc. According to the three-dimensional space position of the target in each frame of the point cloud where the target corresponds to the target tracking track, the target can be predicted in the next The three-dimensional space position that appears in the frame point cloud (that is, the point cloud collected at the next moment), that is, the target tracking trajectory can be obtained to predict the three-dimensional space position of the target in the next frame point cloud; Tracking in the image, according to the two-dimensional space position of the target in each frame of the target image in the target tracking track corresponding to the target, through the optical flow algorithm, etc. can predict the target in the next frame of image (that is, the next moment to collect the image. The two-dimensional space position that appears in the image), that is, the two-dimensional space position of the target tracking trajectory in the next frame image can be obtained.
在对点云进行目标检测时,已有目标跟踪轨迹在当前的点云中预测目标的三维空间位置所在的位置区域出现目标的可能性明显高于点云中的其它位置区域,是对点云进行目标检测时需要重点关注的位置区域。When the target detection is performed on the point cloud, the existing target tracking trajectory in the current point cloud predicts the three-dimensional space position of the target in the location area where the probability of the target appearing is significantly higher than that in other location areas in the point cloud. The location area that needs to be focused on when performing object detection.
对于点云进行目标检测,终端可以通过目标检测模型对点云和至少一个目标跟踪轨迹在点云中预测目标的三维空间位置的处理实现。具体的,目标检测模型可以由训练设备基 于样本集中维护的已知目标跟踪轨迹对应的预测目标的三维空间位置的多个点云样本,以及与所述多个点云样本一一对应的多个目标的三维空间位置检测结果训练得到。在对目标检测模型进行训练时,训练设备可以根据每个点云样本对应的目标的三维空间位置,为每个点云样本添加三维空间位置标签向量(如由中心点坐标、长宽高尺寸等信息构成的标签向量)。另外,需要理解的是,如果点云样本中存在多个目标的三维空间位置,则为点云样本添加的三维空间位置标签向量存在多个,与多个目标一一对应,多个目标的三维空间位置标签向量也可以以矩阵的形式存在。For target detection on the point cloud, the terminal can predict the three-dimensional space position of the target in the point cloud by processing the point cloud and at least one target tracking trajectory by the target detection model. Specifically, the target detection model can be a plurality of point cloud samples that predict the three-dimensional spatial position of the target based on the known target tracking trajectories maintained in the sample set by the training device, and a plurality of point cloud samples corresponding to the plurality of point cloud samples one-to-one. The three-dimensional space position detection result of the target is obtained by training. When training the target detection model, the training device can add a three-dimensional space position label vector (such as center point coordinates, length, width, height, etc.) to each point cloud sample according to the three-dimensional space position of the target corresponding to each point cloud sample. label vector of information). In addition, it should be understood that if there are multiple three-dimensional spatial positions of targets in the point cloud sample, there are multiple three-dimensional spatial position label vectors added to the point cloud sample, which correspond to multiple targets one-to-one. The spatial location label vector can also exist in the form of a matrix.
为训练集中每个点云样本添加目标的三维空间位置标签向量后,训练设备可以将点云样本和目标跟踪轨迹(一个或多个)对应的预测目标的三维空间位置输入到目标检测模型进行处理,得到目标检测模型输出的目标(一个或多个)的三维空间位置预测值,根据输出的目标的三维空间位置预测值以及点云样本对应真实的目标的三维空间位置标签向量,通过损失函数(loss function)训练设备可以计算目标检测模型的损失(loss),loss越高表示通过目标检测模型输出的目标的三维空间位置预测值与真实的目标的三维空间位置标签向量的差异越大,训练设备根据loss调整目标检测模型中的参数,如采用随机梯度下降法更新目标检测模型中神经元的参数,那么对目标检测模型的训练过程就变成了尽可能缩小这个loss的过程。通过样本集中的点云样本不断对目标检测模型进行训练,当这个loss缩小至预设范围,即可得到训练完成的目标检测模型。其中,目标检测模型可以是深度神经网络等。After adding the 3D space position label vector of the target to each point cloud sample in the training set, the training device can input the 3D space position of the predicted target corresponding to the point cloud sample and the target tracking track(s) into the target detection model for processing , obtain the predicted value of the three-dimensional space position of the target (one or more) output by the target detection model, according to the predicted value of the three-dimensional space position of the output target and the three-dimensional space position label vector of the real target corresponding to the point cloud sample, through the loss function ( loss function) training equipment can calculate the loss of the target detection model. Adjust the parameters in the target detection model according to the loss. If the stochastic gradient descent method is used to update the parameters of the neurons in the target detection model, the training process of the target detection model becomes the process of reducing the loss as much as possible. The target detection model is continuously trained through the point cloud samples in the sample set. When the loss is reduced to a preset range, the trained target detection model can be obtained. The target detection model may be a deep neural network or the like.
需要理解的是,训练集中的点云样本可以采用预采样的方式获取,如通过终端预先采集点云样本,并根据目标跟踪轨迹(一个或多个)预测采集的点云样本中预测目标的三维空间位置,并进行记录,同时标注点云样本中存在的真实目标的三维空间位置。It should be understood that the point cloud samples in the training set can be obtained by pre-sampling, such as pre-collecting point cloud samples through the terminal, and predicting the three-dimensional shape of the predicted target in the collected point cloud samples according to the target tracking trajectory (one or more). The spatial position is recorded, and the three-dimensional spatial position of the real target existing in the point cloud sample is marked at the same time.
上述训练设备可以为个人电脑(personal computer,PC)、笔记本电脑、服务器等设备,也可以为终端,如果训练设备和终端不是同一设备,训练设备对目标检测模型训练完成后,可以将训练完成的目标检测模型导入终端中,以便终端对获取的点云中的第一目标进行检测。The above training equipment can be a personal computer (PC), a notebook computer, a server, etc., or a terminal. If the training equipment and the terminal are not the same equipment, after the training equipment has completed the training of the target detection model, the training completed can be used. The target detection model is imported into the terminal, so that the terminal can detect the first target in the acquired point cloud.
S703:所述终端根据所述至少一个第一目标的三维空间位置在所述图像中的投影和所述至少一个目标跟踪轨迹在所述图像中预测目标的二维空间位置,确定所述图像中至少一个第二目标的二维空间位置。S703: The terminal predicts the two-dimensional space position of the target in the image according to the projection of the three-dimensional space position of the at least one first target in the image and the at least one target tracking trajectory in the image, and determines the position of the target in the image. two-dimensional spatial location of at least one second object.
通过点云坐标系(三维)和图像坐标系(二维)的投影矩阵,可以点云中的三维空间位置投影到图像中,得到在图像中的二维空间位置,也可以将图像中的二维空间位置投影到点云中,得到在点云中的三维空间位置。Through the projection matrix of the point cloud coordinate system (three-dimensional) and the image coordinate system (two-dimensional), the three-dimensional space position in the point cloud can be projected into the image, and the two-dimensional space position in the image can be obtained. The dimensional space position is projected into the point cloud, and the 3D space position in the point cloud is obtained.
在一些实施中,对于投影矩阵的确定可以预先设置若干标定物(如具有多个棱角的立体纸箱)放置在三维扫描设备和视觉传感器的公共视野内,通过三维扫描设备和视觉传感器采集标定物点云和标定物图像,在采集的标定物点云和标定物图像中选取多个标定点(如立体纸箱的角点),得到多个标定点在标定物点云中的三维坐标以及在标定物图像中的二维坐标,根据多个标定点在标定物点云中的三维坐标以及在标定物图像中的二维坐标,即可求解点云坐标系和图像坐标系的投影矩阵。In some implementations, for the determination of the projection matrix, several calibration objects (such as a three-dimensional carton with multiple edges and corners) can be preset and placed in the common field of view of the 3D scanning device and the vision sensor, and the calibration object points are collected by the 3D scanning device and the vision sensor. Cloud and calibration object image, select multiple calibration points (such as the corners of the three-dimensional carton) in the collected calibration object point cloud and calibration object image, and obtain the three-dimensional coordinates of the multiple calibration points in the calibration object point cloud and the calibration object. For the two-dimensional coordinates in the image, the projection matrix of the point cloud coordinate system and the image coordinate system can be solved according to the three-dimensional coordinates of multiple calibration points in the calibration object point cloud and the two-dimensional coordinates in the calibration object image.
作为一种示例:假设(x,y,z)和(u,v)分别为标定点在点云坐标系和图像坐标系下的坐标,可以得到两坐标系间转换关系如下:As an example: Assuming that (x, y, z) and (u, v) are the coordinates of the calibration point in the point cloud coordinate system and the image coordinate system, respectively, the conversion relationship between the two coordinate systems can be obtained as follows:
Figure PCTCN2022078611-appb-000001
Figure PCTCN2022078611-appb-000001
其中,K为视觉传感器的内参矩阵,视觉传感器的内参矩阵在出厂后是固定,通常由生产厂商提供或通过标定算法获得,[R,T]是视觉传感器的外参矩阵,通过多个(至少3个)标定点在标定物点云中的三维坐标以及在标定物图像中的二维坐标,即可求解点云坐标系到图像坐标系的投影矩阵M。Among them, K is the internal parameter matrix of the visual sensor. The internal parameter matrix of the visual sensor is fixed after leaving the factory and is usually provided by the manufacturer or obtained through a calibration algorithm. [R, T] is the external parameter matrix of the visual sensor. 3) The three-dimensional coordinates of the calibration point in the point cloud of the calibration object and the two-dimensional coordinates in the image of the calibration object, the projection matrix M from the point cloud coordinate system to the image coordinate system can be solved.
另外,虽然点云中第一目标检测时,已经增加了对目标跟踪轨迹的预测目标的反馈,降低漏检率,但是目标检测模型输出的第一目标的三维空间位置检测结果仍存在漏检的可能,因此在一些实施例中,终端还可以在对图像中第二目标检测时,增加对目标跟踪轨迹的预测目标的反馈,将至少一个第一目标在图像中的投影得到的二维空间位置,以及至少一个目标跟踪轨迹在图像中预测目标的二维空间位置均认为存在目标,将投影得到的二维空间位置和预测目标的二维空间位置均作为存在第二目标的二维空间位置输出。In addition, although the feedback of the predicted target of the target tracking trajectory has been added during the detection of the first target in the point cloud to reduce the missed detection rate, the detection results of the three-dimensional space position of the first target output by the target detection model still have some missed detections. Possibly, so in some embodiments, the terminal may also add feedback on the predicted target of the target tracking trajectory when detecting the second target in the image, and convert the two-dimensional spatial position obtained by the projection of at least one first target in the image. , and at least one target tracking trajectory predicts the two-dimensional space position of the target in the image as a target, and the two-dimensional space position obtained by projection and the two-dimensional space position of the predicted target are output as the two-dimensional space position of the second target. .
S704:所述终端根据所述至少一个第二目标的二维空间位置在所述点云中的投影,确定所述点云中所述至少一个第二目标的三维空间位置。S704: The terminal determines the three-dimensional space position of the at least one second target in the point cloud according to the projection of the two-dimensional space position of the at least one second target in the point cloud.
终端将图像中至少一个第二目标的二维空间位置投影到点云中,即可得到点云中所述至少一个第二目标的三维空间位置,得到点云最终的目标检测结果输出。The terminal projects the two-dimensional spatial position of the at least one second target in the image into the point cloud to obtain the three-dimensional spatial position of the at least one second target in the point cloud, and obtains the final target detection result output of the point cloud.
对于任一个第二目标,该第二目标的特征可以包括在点云中三维空间位置中的目标特征和在图像中二维空间位置的目标特征。其中在点云中三维空间位置中的目标特征可以包括位置(如中心点坐标)、尺寸(如长宽高)、速度、方向、类别、点云点数、点云各方向坐标数值分布、点云反射强度分布(如点云反射强度分布直方图)、深度特征等,在图像中二维空间位置的目标特征包括位置(中心点坐标)、尺寸(如长宽)、速度、方向、类别、外观特征(如图像颜色直方图、方向梯度直方图)等。For any second target, the features of the second target may include target features in a three-dimensional space position in the point cloud and target features in a two-dimensional space position in the image. The target features in the three-dimensional space position in the point cloud may include position (such as center point coordinates), size (such as length, width and height), speed, direction, category, number of point cloud points, coordinate value distribution in each direction of point cloud, point cloud Reflection intensity distribution (such as point cloud reflection intensity distribution histogram), depth features, etc. The target features of the two-dimensional space position in the image include position (center point coordinates), size (such as length and width), speed, direction, category, appearance Features (such as image color histogram, directional gradient histogram), etc.
对于目标跟踪,一个目标跟踪轨迹对应一个目标,目标跟踪轨迹记录有该目标的信息,如ID、目标特征、存在时间、在存在该目标的每一帧点云中的三维空间位置、在存在该目标的每一帧图像中的二维空间位置等,为了实现对同一目标的跟踪,在一些实施例中,终端可以根据已有至少一个目标跟踪轨迹对应的目标特征以及检测到至少一个第二目标的目标特征,对所述至少一个目标跟踪轨迹和所述至少一个第二目标进行匹配。将匹配到目标跟踪轨迹的第二目标与目标跟踪轨迹进行关联,完善已有的目标跟踪轨迹。For target tracking, a target tracking trajectory corresponds to a target, and the target tracking trajectory records the information of the target, such as ID, target characteristics, existence time, three-dimensional space position in each frame of point cloud where the target exists, The two-dimensional spatial position in each frame of images of the target, etc., in order to achieve the tracking of the same target, in some embodiments, the terminal can detect at least one second target according to the target feature corresponding to the existing at least one target tracking trajectory The target feature is matched with the at least one target tracking trajectory and the at least one second target. The second target matched to the target tracking trajectory is associated with the target tracking trajectory to improve the existing target tracking trajectory.
作为一种示例,可以将至少一个目标跟踪轨迹的目标特征和至少一个第二目标的目标特征之间的匹配度(或相似度)作为成本矩阵,采用匈牙利算法对至少一个目标跟踪轨迹和至少一个第二目标进行全局最优匹配。其中匈牙利算法是一种在多项式时间内求解任务分配问题的组合优化算法。终端在计算目标跟踪轨迹的目标特征和第二目标的目标特征之间的相似度时,考虑了位置(点云中和/或图像中)、尺寸(点云中和/或图像中)、速度(点云中和/或图像中)、方向(点云中和/或图像中)、类别(点云中和/或图像中)、点云点数、点云各方向坐标数值分布、点云反射强度分布、外观特征、深度特征等目标特征中的一项或多项,当考虑了多项目标特征时,可以对不同的目标特征赋予不同的权值,且所有权值的和为1。As an example, the matching degree (or similarity) between the target feature of the at least one target tracking trajectory and the target feature of the at least one second target can be used as the cost matrix, and the Hungarian algorithm can be used to analyze the at least one target tracking trajectory and the at least one target tracking trajectory. The second objective performs global optimal matching. The Hungarian algorithm is a combinatorial optimization algorithm that solves the task assignment problem in polynomial time. When calculating the similarity between the target feature of the target tracking trajectory and the target feature of the second target, the terminal considers the position (in the point cloud and/or in the image), size (in the point cloud and/or in the image), speed (in the point cloud and/or in the image), direction (in the point cloud and/or in the image), category (in the point cloud and/or in the image), number of points in the point cloud, numerical distribution of coordinates in each direction of the point cloud, point cloud reflection One or more of the target features such as intensity distribution, appearance feature, depth feature, etc. When multiple target features are considered, different target features can be assigned different weights, and the sum of the weighted values is 1.
对于匹配上已有目标跟踪轨迹的第二目标,将第二目标赋予匹配的目标跟踪轨迹的ID, 完善已有目标跟踪轨迹;对于未匹配上目标跟踪轨迹的第二目标,终端可以为该目标赋予一个新的目标跟踪轨迹ID,新建一条目标跟踪轨迹。For the second target that matches the existing target tracking trajectory, assign the second target the ID of the matching target tracking trajectory to improve the existing target tracking trajectory; for the second target that does not match the target tracking trajectory, the terminal can be the target. Assign a new target tracking track ID to create a new target tracking track.
对于未匹配上第二目标的目标跟踪轨迹,终端可以将所述目标跟踪轨迹与所述目标跟踪轨迹在点云和/或图像中的预测目标关联,完善该目标跟踪轨迹,避免因漏检等原因,造成同一目标对应多个目标跟踪轨迹。For the target tracking trajectory that does not match the second target, the terminal can associate the target tracking trajectory with the predicted target of the target tracking trajectory in the point cloud and/or image, improve the target tracking trajectory, and avoid missed detections, etc. The reason is that the same target corresponds to multiple target tracking trajectories.
需要理解的是,虽然第二目标中已涵盖有目标跟踪轨迹的预测目标,但是如果该预测目标在点云中的三维空间位置和在图像中的二维空间位置未真实出现,则目标特征仍不会与该目标跟踪轨迹的目标特征匹配成功。It should be understood that although the predicted target of the target tracking trajectory has been covered in the second target, if the predicted target’s three-dimensional spatial position in the point cloud and the two-dimensional spatial position in the image do not actually appear, the target features still remain. The target feature of the target tracking trajectory will not be successfully matched.
另外,为了避免对已移出检测范围的目标进行检测及跟踪浪费处理资源,对于未匹配到第二目标的目标跟踪轨迹,将该目标跟踪轨迹与该目标跟踪轨迹在点云和/或图像中的预测目标关联之前,如果该目标跟踪轨迹关联预测目标的次数大于或等于第一阈值,终端删除该目标跟踪轨迹。In addition, in order to avoid wasting processing resources by detecting and tracking the target that has moved out of the detection range, for the target tracking trajectory that is not matched to the second target, the target tracking trajectory and the target tracking trajectory in the point cloud and/or image are Before predicting the target association, if the number of times the target tracking trajectory is associated with the predicted target is greater than or equal to the first threshold, the terminal deletes the target tracking trajectory.
上述主要从方法流程的角度对本申请提供的方案进行了介绍,下述将从硬件或逻辑划分模块的角度对本申请实施例的技术方案进行详细阐述。可以理解的是,为了实现上述功能,装置可以包括执行各个功能相应的硬件结构和/或软件模块。本领域技术人员应该很容易意识到,结合本文中所公开的实施例描述的各示例的单元及算法步骤,本申请能够以硬件或硬件和计算机软件的结合形式来实现。某个功能究竟以硬件还是计算机软件驱动硬件的方式来执行,取决于技术方案的特定应用和设计约束条件。专业技术人员可以对每个特定的应用来使用不同方法来实现所描述的功能,但是这种实现不应认为超出本申请的范围。The solutions provided by the present application are mainly introduced from the perspective of method flow above, and the technical solutions of the embodiments of the present application will be described in detail below from the perspective of hardware or logical division modules. It can be understood that, in order to realize the above-mentioned functions, the apparatus may include corresponding hardware structures and/or software modules for performing each function. Those skilled in the art should easily realize that the present application can be implemented in hardware or a combination of hardware and computer software with the units and algorithm steps of each example described in conjunction with the embodiments disclosed herein. Whether a function is performed by hardware or computer software driving hardware depends on the specific application and design constraints of the technical solution. Skilled artisans may implement the described functionality using different methods for each particular application, but such implementations should not be considered beyond the scope of this application.
在采用集成的单元的情况下,图8示出了本申请实施例中所涉及的目标检测装置的可能的示例性框图,该目标检测装置800可以以软件模块或硬件模块的形式存在。目标检测装置800可以包括:获取单元803和处理单元802。一种示例中,该装置可以为芯片。In the case of using an integrated unit, FIG. 8 shows a possible exemplary block diagram of the target detection apparatus involved in the embodiment of the present application, and the target detection apparatus 800 may exist in the form of a software module or a hardware module. The target detection apparatus 800 may include: an acquisition unit 803 and a processing unit 802 . In one example, the device may be a chip.
可选的,装置800还可以包括存储单元801,用于存储装置800的程序代码和/或数据。Optionally, the apparatus 800 may further include a storage unit 801 for storing program codes and/or data of the apparatus 800 .
具体地,在一个实施例中,获取单元803,用于获取来自三维扫描设备的点云和来自视觉传感器的图像;Specifically, in one embodiment, the acquiring unit 803 is configured to acquire the point cloud from the three-dimensional scanning device and the image from the vision sensor;
处理单元802,用于将所述点云和至少一个目标跟踪轨迹在所述点云中预测目标的三维空间位置输入到目标检测模型进行处理,得到至少一个第一目标的三维空间位置,其中所述目标检测模型是基于已知目标跟踪轨迹对应的预测目标的三维空间位置的多个点云样本,以及与所述多个点云样本一一对应的多个目标的三维空间位置检测结果训练得到的;The processing unit 802 is configured to input the point cloud and at least one target tracking trajectory in the point cloud to predict the three-dimensional space position of the target into the target detection model for processing, and obtain the three-dimensional space position of at least one first target, wherein the three-dimensional space position of the first target is obtained. The target detection model is obtained by training based on multiple point cloud samples of the three-dimensional spatial position of the predicted target corresponding to the known target tracking trajectory, and the three-dimensional spatial position detection results of the multiple targets corresponding to the multiple point cloud samples one-to-one. of;
所述处理单元802,还用于根据所述至少一个第一目标的三维空间位置在所述图像中的投影和所述至少一个目标跟踪轨迹在所述图像中预测目标的二维空间位置,确定所述图像中至少一个第二目标的二维空间位置;The processing unit 802 is further configured to predict the two-dimensional spatial position of the target according to the projection of the three-dimensional spatial position of the at least one first target in the image and the at least one target tracking trajectory in the image, and determine the two-dimensional spatial position of at least one second object in the image;
所述处理单元802,还用于根据所述至少一个第二目标的二维空间位置在所述点云中的投影,确定所述点云中所述至少一个第二目标的三维空间位置。The processing unit 802 is further configured to determine the three-dimensional spatial position of the at least one second target in the point cloud according to the projection of the two-dimensional spatial position of the at least one second target in the point cloud.
在一种可能的设计中,所述处理单元802,还用于根据所述至少一个目标跟踪轨迹对应的目标特征以及所述至少一个第二目标的目标特征,对所述至少一个目标跟踪轨迹和所述至少一个第二目标进行匹配;将匹配的所述目标跟踪轨迹和所述第二目标关联。In a possible design, the processing unit 802 is further configured to, according to the target feature corresponding to the at least one target tracking track and the target feature of the at least one second target, perform the tracking of the at least one target tracking track and the The at least one second target is matched; the matched target tracking trajectory is associated with the second target.
在一种可能的设计中,所述处理单元802,还用于对于未匹配到所述目标跟踪轨迹的所述第二目标,建立所述第二目标对应的目标跟踪轨迹。In a possible design, the processing unit 802 is further configured to establish a target tracking trajectory corresponding to the second target for the second target that is not matched to the target tracking trajectory.
在一种可能的设计中,所述处理单元802,还用于对于未匹配到所述第二目标的所述目标跟踪轨迹,将所述目标跟踪轨迹与所述目标跟踪轨迹在所述点云和/或所述图像中的预测目标关联。In a possible design, the processing unit 802 is further configured to, for the target tracking trajectory that is not matched to the second target, place the target tracking trajectory and the target tracking trajectory in the point cloud and/or predicted target associations in the image.
在一种可能的设计中,所述处理单元802对于未匹配到所述第二目标的所述目标跟踪轨迹,将所述目标跟踪轨迹与所述目标跟踪轨迹在所述点云和/或所述图像中的预测目标关联之前,还用于当所述目标跟踪轨迹关联预测目标的次数大于或等于第一阈值时,删除所述目标跟踪轨迹。In a possible design, the processing unit 802 compares the target tracking trajectory with the target tracking trajectory in the point cloud and/or the target tracking trajectory for the target tracking trajectory that is not matched to the second target. Before being associated with the predicted target in the image, it is also used for deleting the target tracking trajectory when the number of times the target tracking trajectory is associated with the predicted target is greater than or equal to a first threshold.
在一种可能的设计中,所述目标特征包括以下中的一项或多项:位置、长宽高尺寸、速度、方向、类别、点云点数、点云各方向坐标数值分布、点云反射强度分布、外观特征、深度特征。In a possible design, the target features include one or more of the following: position, length, width, height, speed, direction, category, number of point cloud points, coordinate value distribution in each direction of the point cloud, and point cloud reflection Intensity distribution, appearance features, depth features.
在一种可能的设计中,所述获取单元803,还用于获取来自三维扫描设备的标定物点云和来自视觉传感器的标定物图像;In a possible design, the acquiring unit 803 is further configured to acquire the calibration object point cloud from the three-dimensional scanning device and the calibration object image from the vision sensor;
所述处理单元802,还用于根据标定物中多个标定点在所述标定物点云中的三维坐标以及在所述标定物图像中的二维坐标,确定点云坐标系和图像坐标系的投影矩阵。The processing unit 802 is further configured to determine a point cloud coordinate system and an image coordinate system according to the three-dimensional coordinates of a plurality of calibration points in the calibration object in the calibration object point cloud and the two-dimensional coordinates in the calibration object image projection matrix.
如图9所示,本申请实施例还提供一种目标检测装置900,如图9所示,目标检测装置900包括至少一个处理器902以及接口电路。进一步,所述装置还包括至少一个存储器901,所述至少一个存储器901和处理器902连接。所述接口电路用于为所述至少一个处理器提供数据和/或信息的输入输出。存储器901用于存储计算机执行指令,当目标检测装置900运行时,处理器902执行存储器901中存储的计算机执行指令,以使目标检测装置900实现上述目标检测方法,具体目标检测方法的实现可参考上文及其附图的相关描述,在此不做赘述。As shown in FIG. 9 , an embodiment of the present application further provides a target detection apparatus 900 . As shown in FIG. 9 , the target detection apparatus 900 includes at least one processor 902 and an interface circuit. Further, the apparatus further includes at least one memory 901 , and the at least one memory 901 is connected to the processor 902 . The interface circuit is used to provide input and output of data and/or information for the at least one processor. The memory 901 is used to store the computer-executed instructions. When the target detection device 900 is running, the processor 902 executes the computer-executed instructions stored in the memory 901, so that the target detection device 900 can realize the above-mentioned target detection method. For the realization of the specific target detection method, please refer to The relevant descriptions of the above and the accompanying drawings are not repeated here.
作为本实施例的另一种形式,提供一种计算机可读存储介质,其上存储有程序或指令,该程序或指令被执行时可以执行上述方法实施例中的目标检测方法。As another form of this embodiment, a computer-readable storage medium is provided, on which a program or an instruction is stored, and when the program or instruction is executed, the target detection method in the above method embodiment can be executed.
作为本实施例的另一种形式,提供一种包含指令的计算机程序产品,该指令被执行时可以执行上述方法实施例中的目标检测方法。As another form of this embodiment, a computer program product including an instruction is provided, and when the instruction is executed, the target detection method in the above method embodiment can be executed.
作为本实施例的另一种形式,提供一种芯片,所述芯片可以与存储器耦合,用于调用存储器中存储的计算机程序产品,以实现上述方法实施例中的目标检测方法。As another form of this embodiment, a chip is provided. The chip can be coupled with a memory and is used to call a computer program product stored in the memory to implement the target detection method in the above method embodiments.
本领域内的技术人员应明白,本申请的实施例可提供为方法、系统、或计算机程序产品。因此,本申请可采用完全硬件实施例、完全软件实施例、或结合软件和硬件方面的实施例的形式。而且,本申请可采用在一个或多个其中包含有计算机可用程序代码的计算机可用存储介质(包括但不限于磁盘存储器、CD-ROM、光学存储器等)上实施的计算机程序产品的形式。As will be appreciated by those skilled in the art, the embodiments of the present application may be provided as a method, a system, or a computer program product. Accordingly, the present application may take the form of an entirely hardware embodiment, an entirely software embodiment, or an embodiment combining software and hardware aspects. Furthermore, the present application may take the form of a computer program product embodied on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, etc.) having computer-usable program code embodied therein.
本申请是参照根据本申请实施例的方法、设备(系统)、和计算机程序产品的流程图和/或方框图来描述的。应理解可由计算机程序指令实现流程图和/或方框图中的每一流程和/或方框、以及流程图和/或方框图中的流程和/或方框的结合。可提供这些计算机程序指令到通用计算机、专用计算机、嵌入式处理机或其他可编程数据处理设备的处理器以产生一个机器,使得通过计算机或其他可编程数据处理设备的处理器执行的指令产生用于实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能的装置。The present application is described with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to embodiments of the present application. It will be understood that each flow and/or block in the flowchart illustrations and/or block diagrams, and combinations of flows and/or blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to the processor of a general purpose computer, special purpose computer, embedded processor or other programmable data processing device to produce a machine such that the instructions executed by the processor of the computer or other programmable data processing device produce Means for implementing the functions specified in a flow or flow of a flowchart and/or a block or blocks of a block diagram.
这些计算机程序指令也可存储在能引导计算机或其他可编程数据处理设备以特定方 式工作的计算机可读存储器中,使得存储在该计算机可读存储器中的指令产生包括指令装置的制造品,该指令装置实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能。These computer program instructions may also be stored in a computer-readable memory capable of directing a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory result in an article of manufacture comprising instruction means, the instructions The apparatus implements the functions specified in the flow or flow of the flowcharts and/or the block or blocks of the block diagrams.
这些计算机程序指令也可装载到计算机或其他可编程数据处理设备上,使得在计算机或其他可编程设备上执行一系列操作步骤以产生计算机实现的处理,从而在计算机或其他可编程设备上执行的指令提供用于实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能的步骤。These computer program instructions can also be loaded on a computer or other programmable data processing device to cause a series of operational steps to be performed on the computer or other programmable device to produce a computer-implemented process such that The instructions provide steps for implementing the functions specified in the flow or blocks of the flowcharts and/or the block or blocks of the block diagrams.
尽管已描述了本申请的优选实施例,但本领域内的技术人员一旦得知了基本创造性概念,则可对这些实施例作出另外的变更和修改。所以,所附权利要求意欲解释为包括优选实施例以及落入本申请范围的所有变更和修改。While the preferred embodiments of the present application have been described, additional changes and modifications to these embodiments may occur to those skilled in the art once the basic inventive concepts are known. Therefore, the appended claims are intended to be construed to include the preferred embodiment and all changes and modifications that fall within the scope of this application.
显然,本领域的技术人员可以对本申请实施例进行各种改动和变型而不脱离本申请实施例的精神和范围。这样,倘若本申请实施例的这些修改和变型属于本申请权利要求及其等同技术的范围之内,则本申请也意图包含这些改动和变型在内。Obviously, those skilled in the art can make various changes and modifications to the embodiments of the present application without departing from the spirit and scope of the embodiments of the present application. Thus, if these modifications and variations of the embodiments of the present application fall within the scope of the claims of the present application and their equivalents, the present application is also intended to include these modifications and variations.

Claims (19)

  1. 一种目标检测方法,其特征在于,包括:A target detection method, comprising:
    获取来自三维扫描设备的点云和来自视觉传感器的图像;Acquire point clouds from 3D scanning equipment and images from vision sensors;
    将所述点云和至少一个目标跟踪轨迹在所述点云中预测目标的三维空间位置输入到目标检测模型进行处理,得到至少一个第一目标的三维空间位置;Inputting the point cloud and at least one target tracking trajectory to predict the three-dimensional space position of the target in the point cloud into the target detection model for processing, to obtain the three-dimensional space position of at least one first target;
    根据所述至少一个第一目标的三维空间位置在所述图像中的投影和所述至少一个目标跟踪轨迹在所述图像中预测目标的二维空间位置,确定所述图像中至少一个第二目标的二维空间位置;Determine at least one second target in the image according to the projection of the three-dimensional space position of the at least one first target in the image and the prediction of the two-dimensional space position of the target in the image by the at least one target tracking trajectory The two-dimensional space position of ;
    根据所述至少一个第二目标的二维空间位置在所述点云中的投影,确定所述点云中所述至少一个第二目标的三维空间位置。According to the projection of the two-dimensional spatial position of the at least one second target in the point cloud, the three-dimensional spatial position of the at least one second target in the point cloud is determined.
  2. 如权利要求1所述的方法,其特征在于,所述方法还包括:The method of claim 1, wherein the method further comprises:
    根据所述至少一个目标跟踪轨迹对应的目标特征以及所述至少一个第二目标的目标特征,对所述至少一个目标跟踪轨迹和所述至少一个第二目标进行匹配;matching the at least one target tracking trajectory and the at least one second target according to the target feature corresponding to the at least one target tracking trajectory and the target feature of the at least one second target;
    将匹配的所述目标跟踪轨迹和所述第二目标关联。Associate the matched target tracking trajectory with the second target.
  3. 如权利要求2所述的方法,其特征在于,所述方法还包括:The method of claim 2, wherein the method further comprises:
    对于未匹配到所述目标跟踪轨迹的所述第二目标,建立所述第二目标对应的目标跟踪轨迹。For the second target that is not matched to the target tracking trajectory, a target tracking trajectory corresponding to the second target is established.
  4. 如权利要求2或3所述的方法,其特征在于,所述方法还包括:The method of claim 2 or 3, wherein the method further comprises:
    对于未匹配到所述第二目标的所述目标跟踪轨迹,将所述目标跟踪轨迹与所述目标跟踪轨迹在所述点云和/或所述图像中的预测目标关联。For the target tracking trajectory that is not matched to the second target, the target tracking trajectory is associated with a predicted target of the target tracking trajectory in the point cloud and/or the image.
  5. 如权利要求4所述的方法,其特征在于,所述对于未匹配到所述第二目标的所述目标跟踪轨迹,将所述目标跟踪轨迹与所述目标跟踪轨迹在所述点云和/或所述图像中的预测目标关联之前,所述方法还包括:The method according to claim 4, wherein, for the target tracking trajectory that is not matched to the second target, the target tracking trajectory and the target tracking trajectory are compared between the point cloud and/or the target tracking trajectory. or before the prediction target in the image is associated, the method further includes:
    当所述目标跟踪轨迹关联预测目标的次数大于或等于第一阈值时,删除所述目标跟踪轨迹。When the number of times that the target tracking trajectory is associated with the predicted target is greater than or equal to a first threshold, the target tracking trajectory is deleted.
  6. 如权利要求2-5中任一项所述的方法,其特征在于,所述目标特征包括以下中的一项或多项:The method of any one of claims 2-5, wherein the target feature comprises one or more of the following:
    位置、尺寸、速度、方向、类别、点云点数、点云各方向坐标数值分布、点云反射强度分布、外观特征、深度特征。Position, size, speed, direction, category, number of point cloud points, numerical distribution of coordinates in each direction of point cloud, distribution of reflection intensity of point cloud, appearance feature, depth feature.
  7. 如权利要求1-6中任一项所述的方法,其特征在于,所述方法还包括:The method according to any one of claims 1-6, wherein the method further comprises:
    获取来自三维扫描设备的标定物点云和来自视觉传感器的标定物图像;Obtain the calibration object point cloud from the 3D scanning device and the calibration object image from the vision sensor;
    根据标定物中多个标定点在所述标定物点云中的三维坐标以及在所述标定物图像中的二维坐标,确定点云坐标系和图像坐标系的投影矩阵。The projection matrix of the point cloud coordinate system and the image coordinate system is determined according to the three-dimensional coordinates of the multiple calibration points in the calibration object in the point cloud of the calibration object and the two-dimensional coordinates in the image of the calibration object.
  8. 一种目标检测装置,其特征在于,包括:A target detection device, comprising:
    获取单元,用于获取来自三维扫描设备的点云和来自视觉传感器的图像;an acquisition unit for acquiring the point cloud from the 3D scanning device and the image from the vision sensor;
    处理单元,用于将所述点云和至少一个目标跟踪轨迹在所述点云中预测目标的三维空间位置输入到目标检测模型进行处理,得到至少一个第一目标的三维空间位置;a processing unit, configured to input the point cloud and at least one target tracking trajectory in the point cloud to predict the three-dimensional space position of the target into the target detection model for processing, and obtain the three-dimensional space position of at least one first target;
    所述处理单元,还用于根据所述至少一个第一目标的三维空间位置在所述图像中的投影和所述至少一个目标跟踪轨迹在所述图像中预测目标的二维空间位置,确定所述图像中 至少一个第二目标的二维空间位置;The processing unit is further configured to predict the two-dimensional spatial position of the target according to the projection of the three-dimensional spatial position of the at least one first target in the image and the at least one target tracking trajectory in the image, and determine the the two-dimensional spatial position of at least one second object in the image;
    所述处理单元,还用于根据所述至少一个第二目标的二维空间位置在所述点云中的投影,确定所述点云中所述至少一个第二目标的三维空间位置。The processing unit is further configured to determine the three-dimensional space position of the at least one second target in the point cloud according to the projection of the two-dimensional space position of the at least one second target in the point cloud.
  9. 如权利要求8所述的装置,其特征在于,所述处理单元,还用于根据所述至少一个目标跟踪轨迹对应的目标特征以及所述至少一个第二目标的目标特征,对所述至少一个目标跟踪轨迹和所述至少一个第二目标进行匹配;将匹配的所述目标跟踪轨迹和所述第二目标关联。The apparatus according to claim 8, wherein the processing unit is further configured to, according to the target feature corresponding to the at least one target tracking track and the target feature of the at least one second target The target tracking trajectory is matched with the at least one second target; and the matched target tracking trajectory is associated with the second target.
  10. 如权利要求9所述的装置,其特征在于,所述处理单元,还用于对于未匹配到所述目标跟踪轨迹的所述第二目标,建立所述第二目标对应的目标跟踪轨迹。The apparatus according to claim 9, wherein the processing unit is further configured to establish a target tracking trajectory corresponding to the second target for the second target that is not matched to the target tracking trajectory.
  11. 如权利要求9或10所述的装置,其特征在于,所述处理单元,还用于对于未匹配到所述第二目标的所述目标跟踪轨迹,将所述目标跟踪轨迹与所述目标跟踪轨迹在所述点云和/或所述图像中的预测目标关联。The apparatus according to claim 9 or 10, wherein the processing unit is further configured to compare the target tracking trajectory with the target tracking trajectory for the target tracking trajectory that is not matched to the second target A predicted target association of a trajectory in the point cloud and/or in the image.
  12. 如权利要求11所述的装置,其特征在于,所述处理单元对于未匹配到所述第二目标的所述目标跟踪轨迹,将所述目标跟踪轨迹与所述目标跟踪轨迹在所述点云和/或所述图像中的预测目标关联之前,还用于当所述目标跟踪轨迹关联预测目标的次数大于或等于第一阈值时,删除所述目标跟踪轨迹。The apparatus according to claim 11, wherein, for the target tracking trajectory that is not matched to the second target, the processing unit compares the target tracking trajectory and the target tracking trajectory in the point cloud And/or before the predicted target in the image is associated, it is also used to delete the target tracking track when the number of times the target tracking track is associated with the predicted target is greater than or equal to a first threshold.
  13. 如权利要求9-12中任一项所述的装置,其特征在于,所述目标特征包括以下中的一项或多项:位置、尺寸、速度、方向、类别、点云点数、点云各方向坐标数值分布、点云反射强度分布、外观特征、深度特征。The device according to any one of claims 9-12, wherein the target feature comprises one or more of the following: position, size, speed, direction, category, number of point cloud points, each point cloud Orientation coordinate numerical distribution, point cloud reflection intensity distribution, appearance features, depth features.
  14. 如权利要求8-13中任一项所述的装置,其特征在于,所述获取单元,还用于获取来自三维扫描设备的标定物点云和来自视觉传感器的标定物图像;The device according to any one of claims 8-13, wherein the acquisition unit is further configured to acquire a calibration object point cloud from a three-dimensional scanning device and a calibration object image from a vision sensor;
    所述处理单元,还用于根据标定物中多个标定点在所述标定物点云中的三维坐标以及在所述标定物图像中的二维坐标,确定点云坐标系和图像坐标系的投影矩阵。The processing unit is further configured to determine the coordinates of the point cloud coordinate system and the image coordinate system according to the three-dimensional coordinates of a plurality of calibration points in the calibration object in the calibration object point cloud and the two-dimensional coordinates in the calibration object image. Projection matrix.
  15. 一种目标检测装置,其特征在于,包括至少一个处理器和接口;A target detection device, comprising at least one processor and an interface;
    所述至少一个处理器用于从所述接口调用并运行计算机程序,当所述至少一个处理器执行所述计算机程序时,实现如权利要求1-7中任一项所述的方法。The at least one processor is configured to invoke and run a computer program from the interface, and when the computer program is executed by the at least one processor, the method according to any one of claims 1-7 is implemented.
  16. 一种芯片系统,其特征在于,所述芯片系统包括:至少一个处理器和接口;A chip system, characterized in that the chip system includes: at least one processor and an interface;
    所述至少一个处理器用于从所述接口调用并运行计算机程序,当所述至少一个处理器执行所述计算机程序时,实现如权利要求1-7中任一项所述的方法。The at least one processor is configured to invoke and run a computer program from the interface, and when the computer program is executed by the at least one processor, the method according to any one of claims 1-7 is implemented.
  17. 一种计算机可读存储介质,其特征在于,所述计算机可读存储介质中存储有计算机程序,当所述计算机程序被计算机执行时,使得所述计算机执行如权利要求1-7中任一项所述的方法。A computer-readable storage medium, characterized in that a computer program is stored in the computer-readable storage medium, and when the computer program is executed by a computer, the computer is made to execute any one of claims 1-7 the method described.
  18. 一种终端,其特征在于,所述终端包括如权利要求8-14中任一项所述的目标检测装置。A terminal, characterized in that the terminal comprises the target detection device according to any one of claims 8-14.
  19. 如权利要求18所述的终端,其特征在于,所述终端为车辆、无人机或机器人。The terminal of claim 18, wherein the terminal is a vehicle, a drone or a robot.
PCT/CN2022/078611 2021-03-09 2022-03-01 Target detection method and apparatus WO2022188663A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202110256851.2A CN115049700A (en) 2021-03-09 2021-03-09 Target detection method and device
CN202110256851.2 2021-03-09

Publications (1)

Publication Number Publication Date
WO2022188663A1 true WO2022188663A1 (en) 2022-09-15

Family

ID=83156444

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/078611 WO2022188663A1 (en) 2021-03-09 2022-03-01 Target detection method and apparatus

Country Status (2)

Country Link
CN (1) CN115049700A (en)
WO (1) WO2022188663A1 (en)

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115830079A (en) * 2023-02-15 2023-03-21 天翼交通科技有限公司 Method, device and medium for tracking trajectory of traffic participant
CN115965824A (en) * 2023-03-01 2023-04-14 安徽蔚来智驾科技有限公司 Point cloud data labeling method, point cloud target detection equipment and storage medium
CN116071231A (en) * 2022-12-16 2023-05-05 群滨智造科技(苏州)有限公司 Method, device, equipment and medium for generating ink-dispensing process track of glasses frame
CN116430338A (en) * 2023-03-20 2023-07-14 北京中科创益科技有限公司 Method, system and equipment for tracking moving target
CN116952988A (en) * 2023-09-21 2023-10-27 斯德拉马机械(太仓)有限公司 2D line scanning detection method and system for ECU (electronic control Unit) product
CN117252992A (en) * 2023-11-13 2023-12-19 整数智能信息技术(杭州)有限责任公司 4D road scene labeling method and device based on time sequence data and electronic equipment
CN117523379A (en) * 2023-11-20 2024-02-06 广东海洋大学 Underwater photographic target positioning method and system based on AI
CN117576166A (en) * 2024-01-15 2024-02-20 浙江华是科技股份有限公司 Target tracking method and system based on camera and low-frame-rate laser radar

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116664790B (en) * 2023-07-26 2023-11-17 昆明人为峰科技有限公司 Three-dimensional terrain analysis system and method based on unmanned aerial vehicle mapping

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107871129A (en) * 2016-09-27 2018-04-03 北京百度网讯科技有限公司 Method and apparatus for handling cloud data
CN110675431A (en) * 2019-10-08 2020-01-10 中国人民解放军军事科学院国防科技创新研究院 Three-dimensional multi-target tracking method fusing image and laser point cloud
US20200160542A1 (en) * 2018-11-15 2020-05-21 Toyota Research Institute, Inc. Systems and methods for registering 3d data with 2d image data
CN111709923A (en) * 2020-06-10 2020-09-25 中国第一汽车股份有限公司 Three-dimensional object detection method and device, computer equipment and storage medium
CN112102409A (en) * 2020-09-21 2020-12-18 杭州海康威视数字技术股份有限公司 Target detection method, device, equipment and storage medium
CN112270272A (en) * 2020-10-31 2021-01-26 武汉中海庭数据技术有限公司 Method and system for extracting road intersections in high-precision map making
US20210043002A1 (en) * 2018-09-11 2021-02-11 Tencent Technology (Shenzhen) Company Limited Object annotation method and apparatus, movement control method and apparatus, device, and storage medium

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107871129A (en) * 2016-09-27 2018-04-03 北京百度网讯科技有限公司 Method and apparatus for handling cloud data
US20210043002A1 (en) * 2018-09-11 2021-02-11 Tencent Technology (Shenzhen) Company Limited Object annotation method and apparatus, movement control method and apparatus, device, and storage medium
US20200160542A1 (en) * 2018-11-15 2020-05-21 Toyota Research Institute, Inc. Systems and methods for registering 3d data with 2d image data
CN110675431A (en) * 2019-10-08 2020-01-10 中国人民解放军军事科学院国防科技创新研究院 Three-dimensional multi-target tracking method fusing image and laser point cloud
CN111709923A (en) * 2020-06-10 2020-09-25 中国第一汽车股份有限公司 Three-dimensional object detection method and device, computer equipment and storage medium
CN112102409A (en) * 2020-09-21 2020-12-18 杭州海康威视数字技术股份有限公司 Target detection method, device, equipment and storage medium
CN112270272A (en) * 2020-10-31 2021-01-26 武汉中海庭数据技术有限公司 Method and system for extracting road intersections in high-precision map making

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116071231A (en) * 2022-12-16 2023-05-05 群滨智造科技(苏州)有限公司 Method, device, equipment and medium for generating ink-dispensing process track of glasses frame
CN116071231B (en) * 2022-12-16 2023-12-29 群滨智造科技(苏州)有限公司 Method, device, equipment and medium for generating ink-dispensing process track of glasses frame
CN115830079A (en) * 2023-02-15 2023-03-21 天翼交通科技有限公司 Method, device and medium for tracking trajectory of traffic participant
CN115965824A (en) * 2023-03-01 2023-04-14 安徽蔚来智驾科技有限公司 Point cloud data labeling method, point cloud target detection equipment and storage medium
CN116430338B (en) * 2023-03-20 2024-05-10 北京中科创益科技有限公司 Method, system and equipment for tracking moving target
CN116430338A (en) * 2023-03-20 2023-07-14 北京中科创益科技有限公司 Method, system and equipment for tracking moving target
CN116952988A (en) * 2023-09-21 2023-10-27 斯德拉马机械(太仓)有限公司 2D line scanning detection method and system for ECU (electronic control Unit) product
CN116952988B (en) * 2023-09-21 2023-12-08 斯德拉马机械(太仓)有限公司 2D line scanning detection method and system for ECU (electronic control Unit) product
CN117252992A (en) * 2023-11-13 2023-12-19 整数智能信息技术(杭州)有限责任公司 4D road scene labeling method and device based on time sequence data and electronic equipment
CN117252992B (en) * 2023-11-13 2024-02-23 整数智能信息技术(杭州)有限责任公司 4D road scene labeling method and device based on time sequence data and electronic equipment
CN117523379B (en) * 2023-11-20 2024-04-30 广东海洋大学 Underwater photographic target positioning method and system based on AI
CN117523379A (en) * 2023-11-20 2024-02-06 广东海洋大学 Underwater photographic target positioning method and system based on AI
CN117576166A (en) * 2024-01-15 2024-02-20 浙江华是科技股份有限公司 Target tracking method and system based on camera and low-frame-rate laser radar
CN117576166B (en) * 2024-01-15 2024-04-30 浙江华是科技股份有限公司 Target tracking method and system based on camera and low-frame-rate laser radar

Also Published As

Publication number Publication date
CN115049700A (en) 2022-09-13

Similar Documents

Publication Publication Date Title
WO2022188663A1 (en) Target detection method and apparatus
CN111337941B (en) Dynamic obstacle tracking method based on sparse laser radar data
WO2020043041A1 (en) Method and device for point cloud data partitioning, storage medium, and electronic device
CN111665842B (en) Indoor SLAM mapping method and system based on semantic information fusion
CN111563415B (en) Binocular vision-based three-dimensional target detection system and method
CN111260683A (en) Target detection and tracking method and device for three-dimensional point cloud data
US11556745B2 (en) System and method for ordered representation and feature extraction for point clouds obtained by detection and ranging sensor
CN114842438A (en) Terrain detection method, system and readable storage medium for autonomous driving vehicle
CN111213153A (en) Target object motion state detection method, device and storage medium
CN113192091A (en) Long-distance target sensing method based on laser radar and camera fusion
CN113345008A (en) Laser radar dynamic obstacle detection method considering wheel type robot position and posture estimation
CN114454875A (en) Urban road automatic parking method and system based on reinforcement learning
CN113936198A (en) Low-beam laser radar and camera fusion method, storage medium and device
CN114972968A (en) Tray identification and pose estimation method based on multiple neural networks
Choe et al. Fast point cloud segmentation for an intelligent vehicle using sweeping 2D laser scanners
CN115376109B (en) Obstacle detection method, obstacle detection device, and storage medium
CN115861968A (en) Dynamic obstacle removing method based on real-time point cloud data
CN116109601A (en) Real-time target detection method based on three-dimensional laser radar point cloud
CN114998276A (en) Robot dynamic obstacle real-time detection method based on three-dimensional point cloud
CN115201849A (en) Indoor map building method based on vector map
CN115100741A (en) Point cloud pedestrian distance risk detection method, system, equipment and medium
Zhao et al. Omni-Directional Obstacle Detection for Vehicles Based on Depth Camera
CN115685237A (en) Multi-mode three-dimensional target detection method and system combining viewing cones and geometric constraints
CN116863325A (en) Method for multiple target detection and related product
Alonso et al. Footprint-based classification of road moving objects using occupancy grids

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22766189

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 22766189

Country of ref document: EP

Kind code of ref document: A1