WO2023072055A1 - Point cloud data processing method and system - Google Patents

Point cloud data processing method and system Download PDF

Info

Publication number
WO2023072055A1
WO2023072055A1 PCT/CN2022/127326 CN2022127326W WO2023072055A1 WO 2023072055 A1 WO2023072055 A1 WO 2023072055A1 CN 2022127326 W CN2022127326 W CN 2022127326W WO 2023072055 A1 WO2023072055 A1 WO 2023072055A1
Authority
WO
WIPO (PCT)
Prior art keywords
point cloud
detection frame
cloud data
frame
detection
Prior art date
Application number
PCT/CN2022/127326
Other languages
French (fr)
Chinese (zh)
Inventor
李亚敏
晋周南
朱小天
Original Assignee
华为技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 华为技术有限公司 filed Critical 华为技术有限公司
Publication of WO2023072055A1 publication Critical patent/WO2023072055A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/774Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/80Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/60Type of objects
    • G06V20/64Three-dimensional objects

Definitions

  • the embodiments of the present application relate to the technical field of data processing, and in particular to a point cloud data processing method and system.
  • Lidar lightlaser detection and ranging, Lidar
  • the lidar Compared with traditional sensors such as cameras and ultrasonic sensors, lidar has the characteristics of high measurement accuracy, fast response speed, and strong anti-interference ability.
  • LiDAR has been widely used in the field of intelligent driving and unmanned driving.
  • the lidar transmits a detection signal to the target object during work (for example, the detection signal can be a laser beam), and after the lidar receives the reflected signal reflected from the target object, the lidar compares the detection signal with the transmitted signal and processed to obtain a point cloud set, which is a set of sampling points obtained after obtaining the spatial coordinates of each sampling point on the surface of the target object.
  • Lidar can simultaneously detect and sample multiple target objects at the same time to obtain a set of point cloud data.
  • the point cloud data can include multiple point cloud sets corresponding to multiple target objects.
  • a point cloud 3D object automatic labeling system is proposed in some designs.
  • the system takes a single frame of point cloud data or several frames of point cloud data as input , by using the 3D detection model for automatic annotation.
  • the target object is occluded or the distance is too far, due to the lack of point cloud data in the single frame point cloud data or several frames of point cloud data, it will lead to problems such as false detection, missed detection, and inaccurate detection frames.
  • the offset in the point cloud data collected by the object at different times will seriously affect the automatic labeling effect of the system due to the smearing phenomenon caused by the point cloud data processing process.
  • the embodiment of the present application provides a point cloud data processing method and system, by using the detection result of the first point cloud map obtained from the multi-frame point cloud data, and part of the frame points in the multi-frame point cloud data
  • the detection results of the first point cloud data obtained from the cloud data are fused to improve the ability of the point cloud data processing system to label part of the frame point cloud data.
  • the embodiment of the present application provides a point cloud data processing method, which can be implemented by a point cloud data processing system, and the method includes: obtaining a first detection frame, the first detection frame is used to indicate the The static object marked in the point cloud map, the first point cloud map is obtained by M frame point cloud data, M is an integer greater than or equal to 2; obtain the second detection frame, the second detection frame is used to indicate the The dynamic object marked in the first point cloud data, the first point cloud data is obtained from N frames of point cloud data, N is an integer greater than or equal to 1, and N ⁇ M; according to the second detection frame, the The single-frame point cloud data in the M frame point cloud data is fused and processed to obtain a second point cloud map; the first detection frame and the second detection frame are corrected according to the second point cloud map to obtain a label A detection frame, the target detection frame is used to indicate the marked static object and/or dynamic object marked in the single frame point cloud data.
  • the point cloud data processing system can use the M frames of point cloud data and the N frames of point cloud data in the M frames of point cloud data as data to be labeled for different processing, and then process the processed point cloud data Perform fusion and obtain the corrected labeling results of each single frame of point cloud data, so as to use M frame point cloud data to compensate N frame point cloud data, and reduce the tailing phenomenon of N frame point cloud data caused by dynamic objects , and problems such as false detection, missed detection, and inaccurate detection frames caused by too few points in the N-frame point cloud data, so as to improve the labeling ability of the point cloud data processing system.
  • the point cloud data processing system determines the specific value of N or M according to the configuration information or the application scenario, and the embodiment of the present application does not limit the specific implementation manner.
  • static objects and dynamic objects are target objects included in point cloud data or point cloud maps.
  • Static objects are generally point cloud collections collected by detecting static objects
  • dynamic objects are generally A collection of point clouds collected for the detection of dynamic objects.
  • due to the mobility of dynamic objects the positions of dynamic objects at different times are different. Therefore, in order to eliminate the impact of the mobility of dynamic objects, when splicing point cloud collections collected at different times for the same dynamic object, it will be The phenomenon of listing the same objects is the smear phenomenon mentioned in the embodiment of the present application, which may also be called the smear phenomenon.
  • the acquiring the first detection frame includes: splicing the M frames of point cloud data to obtain the first point cloud map;
  • the first point cloud map performs target detection to obtain the first detection frame, wherein the first detection model is obtained by training using first training data, and the first training data includes the third point cloud map and static Object annotation data.
  • a point cloud map (for example, the third point cloud map) can be used to train the first detection model for labeling static objects, so as to provide a high-precision static object detection model.
  • the point cloud data processing system can use the first detection model to mark the static objects in the point cloud map to be marked (for example, the first point cloud map), so that the detection results of the static objects in the point cloud map can be used in the In the M-frame point cloud data, the dynamic object is aligned with the relevant features of the static object, and the tailing phenomenon caused by the dynamic object is reduced in the N-frame point cloud data, thereby reducing the difficulty of labeling the dynamic object.
  • the third point cloud map may be the same as or different from the first point cloud map, that is to say, the first point cloud map may be used as data to be labeled or as training data. No limit.
  • the acquiring the second detection frame includes: performing object detection on the first point cloud data according to the second detection model to obtain a third detection frame, wherein the The second detection model is obtained through training using second training data, the second training data includes second point cloud data and dynamic object labeling data, and the third detection frame is used to indicate that in the first point cloud data A marked dynamic object; correcting the third detection frame according to the first detection frame to obtain the second detection frame.
  • the second point cloud data can be used to train the second detection model for labeling the dynamic object, so as to provide a high-precision dynamic object detection model.
  • the point cloud data processing system can use the second detection model to mark the dynamic objects in the N frames of point cloud data to be marked, so as to obtain the dynamic object detection results in the N frames of point cloud data, and the dynamic object detection results can be used It is used to fuse M frames of point cloud data to present dynamic objects in M frames of point cloud data in a form that eliminates smearing, so as to facilitate the subsequent correction process and improve the detection accuracy of the point cloud data processing system.
  • the second detection frame performing fusion processing on the single frame point cloud data in the M frames of point cloud data to obtain a second point cloud map, Including: according to the second detection frame, removing the dynamic object in the single frame point cloud data of the M frames of point cloud data to obtain the third point cloud data corresponding to the single frame point cloud data; according to the second detection frame Attribute information, associate the second detection frame corresponding to the same dynamic object to obtain the dynamic object association result; based on the dynamic object association result, perform fusion processing on the third point cloud data corresponding to the M frame point cloud data to obtain Second point cloud map.
  • the point cloud data processing system can use the detection results of the dynamic objects in the N-frame point cloud data to realize the object-level association of the dynamic objects in the M-frame point cloud data, and eliminate the tailing phenomenon caused by the dynamic objects , to obtain a new point cloud map, so that the dynamic objects in the new point cloud map are presented in a form that eliminates the trailing phenomenon, thereby facilitating the subsequent correction process.
  • correcting the first detection frame and the second detection frame according to the second point cloud map to obtain a target detection frame includes: In the two-point cloud map, the first attribute of the first detection frame marked with the static object is corrected, and/or the second attribute of the second detection frame marked with the dynamic object is corrected to obtain the fourth detection frame, the fourth The detection frame is used to indicate the dynamic object and/or static object marked in the second point cloud map; in the single frame point cloud data corresponding to the second point cloud map, the fourth detection frame of the marked dynamic object The third attribute is corrected to obtain the fifth detection frame; the fourth detection frame marked with the static object and the fifth detection frame are used as the target detection frame of the M frames of point cloud data.
  • the point cloud data processing system can, for example, correct the relevant attributes of the detection frame marked with static objects and/or dynamic objects in the second point cloud map, which helps to improve the detection accuracy of the point cloud data processing system .
  • the point cloud data processing system may implement the correction process by manual correction or automatic correction, and the embodiment of the present application does not limit the specific implementation manner.
  • the method further includes: combining the single frame point cloud data in the M frames of point cloud data and the target detection frame As the third training data, in the process of training a plurality of detection models based on the third training data, a sixth detection frame is determined, and the detection result of the sixth detection frame is an error; the second point cloud map and the The target detection frame is used as fourth training data, and a seventh detection frame is determined during the process of training multiple detection models based on the fourth training data; and the seventh detection frame is corrected based on the sixth detection frame.
  • the point cloud data processing system can use multi-model fusion and point cloud map to generate softening frames, provide higher-quality detection frames and softening frames that are easy to train, and help improve the accuracy of single frames using automatic labeling systems. Or the training accuracy of the model of several frames of point cloud data.
  • the embodiment of the present application provides a point cloud data processing system, including: a first acquisition unit, configured to acquire a first detection frame, and the first detection frame is used to indicate static object, the first point cloud map is obtained from M frames of point cloud data, and M is an integer greater than or equal to 2; the second acquisition unit is used to acquire a second detection frame, and the second detection frame is used to indicate The dynamic object marked in the first point cloud data, the first point cloud data is obtained from N frames of point cloud data, N is an integer greater than or equal to 1, and N ⁇ M; the processing unit is used for according to the second The detection frame is used to fuse the single-frame point cloud data in the M frames of point cloud data to obtain a second point cloud map; a correction unit is used to correct the first detection frame and the second point cloud map according to the second point cloud map.
  • the second detection frame is corrected to obtain a target detection frame, and the target detection frame is used to indicate the marked static object and/or dynamic object marked in the single frame point cloud data.
  • the first acquisition unit and the second acquisition unit may be the same unit or different units, and the embodiment of the present application does not limit the product form.
  • the first acquisition unit is configured to: stitch the M frames of point cloud data to obtain the first point cloud map;
  • the first point cloud map is used for target detection to obtain the first detection frame, wherein the first detection model is obtained by training using the first training data, and the first training data includes the third point cloud map and static objects Annotate the data.
  • the second acquisition unit is configured to: perform object detection on the first point cloud data according to a second detection model to obtain a third detection frame, wherein the The second detection model is obtained by training using the second training data, the second training data includes the second point cloud data and dynamic object labeling data, and the third detection frame is used to indicate the labeling in the first point cloud data the dynamic object; correcting the third detection frame according to the first detection frame to obtain the second detection frame.
  • the processing unit is configured to: remove dynamic objects in a single frame of point cloud data in the M frames of point cloud data according to the second detection frame, to obtain a single The third point cloud data corresponding to the frame point cloud data; according to the attribute information of the second detection frame, correlating the second detection frame corresponding to the same dynamic object to obtain a dynamic object association result; based on the dynamic object association result to The third point cloud data corresponding to the M frames of point cloud data are fused to obtain a second point cloud map.
  • the correction unit is configured to: in the second point cloud map, correct the first attribute of the first detection frame marking the static object, and/or , modifying the second attribute of the second detection frame marked with the dynamic object to obtain a fourth detection frame, the fourth detection frame is used to indicate the dynamic object and/or the static object marked in the second point cloud map; in the In the single-frame point cloud data corresponding to the second point cloud map, the third attribute of the fourth detection frame marked with the dynamic object is corrected to obtain the fifth detection frame; the fourth detection frame marked with the static object and the first Five detection frames are used as target detection frames of the M frames of point cloud data.
  • the system further includes a training unit configured to: use the single frame of point cloud data in the M frames of point cloud data and the target detection frame As the third training data, in the process of training a plurality of detection models based on the third training data, a sixth detection frame is determined, and the detection result of the sixth detection frame is an error; the second point cloud map and the The target detection frame is used as fourth training data, and a seventh detection frame is determined during the process of training multiple detection models based on the fourth training data; and the seventh detection frame is corrected based on the sixth detection frame.
  • the embodiment of the present application provides a point cloud data processing system, including a memory and a processor, the memory is used to store programs; the processor is used to execute the programs stored in the memory, so that the The device implements the method described in the foregoing first aspect and any possible implementation manner of the first aspect.
  • the embodiment of the present application provides a point cloud data processing device, including: at least one processor and an interface circuit, the interface circuit is used to provide data or code instructions for the at least one processor, and the at least one processor
  • a processor is configured to implement the method described in the first aspect and any possible implementation manner of the first aspect by using a logic circuit or executing code instructions.
  • the embodiment of the present application provides a computer-readable storage medium, the computer-readable medium stores program code, and when the program code is run on the computer, the computer executes the above-mentioned first aspect and the first aspect.
  • the method described in any possible implementation manner In one aspect, the method described in any possible implementation manner.
  • an embodiment of the present application provides a computer program product, which, when the computer program product is run on a computer, enables the computer to execute the above-mentioned first aspect and any possible implementation manner of the first aspect. method.
  • an embodiment of the present application provides a chip system, the chip system includes a processor, configured to call a computer program or a computer instruction stored in a memory, so that the processor executes the above-mentioned first aspect and the first aspect The method described in any possible implementation.
  • the processor is coupled to the memory through an interface.
  • the system on a chip further includes a memory, where computer programs or computer instructions are stored in the memory.
  • the embodiment of the present application provides a terminal device, which can be used to implement the method described in the foregoing first aspect and any possible implementation manner of the first aspect.
  • the terminal equipment includes but is not limited to: intelligent transportation equipment (such as automobiles, ships, drones, trains, trucks, etc.), intelligent manufacturing equipment (such as robots, industrial equipment, intelligent logistics, intelligent factories, etc.), intelligent terminal (Mobile phones, computers, tablets, PDAs, desktops, headsets, audio, wearable devices, car devices, etc.).
  • an embodiment of the present application provides a vehicle, which can be used to implement the method described in the first aspect and any possible implementation manner of the first aspect.
  • the embodiment of the present application provides a server, which can be used to implement the method described in the first aspect and any possible implementation manner of the first aspect.
  • Fig. 1 is the schematic diagram of a kind of point cloud data
  • Fig. 2 is a schematic diagram of a group of point cloud data with deviations in detection frame positions
  • Figure 3 is a schematic diagram of less point cloud data and trailing phenomenon
  • FIG. 4 is a schematic diagram of an application scenario applicable to the point cloud data processing method provided by the embodiment of the present application.
  • FIG. 5 is a schematic diagram of a point cloud data processing system provided by an embodiment of the present application.
  • FIG. 6 is a schematic flow chart of a point cloud data processing method provided in an embodiment of the present application.
  • FIG. 7 is a schematic diagram of the first detection frame information provided by the embodiment of the present application.
  • FIG. 8 is a schematic diagram of the second detection frame information provided by the embodiment of the present application.
  • FIG. 9 is a schematic diagram of obtaining a second point cloud map provided by an embodiment of the present application.
  • FIG. 10 is a schematic flow diagram of a point cloud data processing method provided in an embodiment of the present application.
  • FIG. 11 is a schematic diagram of a point cloud data processing system provided by an embodiment of the present application.
  • FIG. 12 is a schematic diagram of a point cloud data processing system provided by an embodiment of the present application.
  • Point cloud is the sampling point obtained after the object is detected by the measuring device.
  • Point cloud data refers to a set of vectors in a three-dimensional coordinate system.
  • a collection of point clouds obtained after the measuring device detects and samples the appearance surface of an object may be referred to as a point cloud collection.
  • the measuring device can detect and sample multiple objects at the same time, and the obtained set of point cloud data can include point cloud collections corresponding to multiple objects.
  • the point cloud data may also include color information, for example.
  • point cloud data measured based on the principle of laser measurement also referred to as laser point cloud data
  • laser point cloud data may include information such as three-dimensional coordinates and laser reflection intensity (intensity).
  • the point cloud data obtained based on the principle of photogrammetry may include information such as three-dimensional coordinates and color, wherein the color information may be color data in red, green, blue (RGB) format.
  • RGB red, green, blue
  • the measuring device is Lidar (lightlaser detection and ranging, Lidar) as an example, and the point cloud data collected by the Lidar includes at least the three-dimensional coordinates of each point cloud.
  • Lidar lightlaser detection and ranging, Lidar
  • one frame (or single frame) point cloud data refers to a group of point cloud data collected at one sampling moment
  • the point cloud map refers to the continuous collection of multiple frames
  • the image obtained by splicing point cloud data can be a two-dimensional image or a three-dimensional image.
  • the training set, verification set and test set are different sets of sample data used for model training in machine learning.
  • the training set is a sample set composed of data samples used for model fitting.
  • the validation set is a set of samples reserved during the model training process, which is used to adjust the hyperparameters of the model and to conduct a preliminary evaluation of the model's capabilities.
  • the verification set can be used to adjust the proportion of point cloud data of different data types in the training set, thereby improving the model's ability to fit point cloud data of different data types .
  • the test set is a sample set used to evaluate the final generalization ability of the model.
  • the sample data used for model training may be referred to as training data
  • the training data may include a training set, a verification set, and a test set, and data in the training set, verification set, and test set generally do not overlap.
  • the sample data can be divided according to the preset ratio, for example, according to the ratio of 7:1:1 to obtain the training set, verification set and test set, and based on the training set, verification set and test set training
  • the embodiment of the present application does not limit the specific implementation manner and training process of the model training.
  • Target detection is an important branch of image processing and computer vision disciplines, and is also the core part of an intelligent monitoring system. Target detection can perform target detection and recognition on the point cloud data, so as to determine the types of objects corresponding to the multiple point cloud sets included in the point cloud data.
  • the target detection device on the vehicle can perform target detection on the point cloud data collected by the lidar, so as to recognize the vehicles on the lane and the trees and pedestrians on the side of the road during the driving process of the vehicle. , assisting vehicles to realize route planning, obstacle avoidance and other functions, and then realize intelligent driving.
  • the embodiment of the present application may include two types of target detection models, denoted as a first detection model and a second detection model.
  • the first detection model can be obtained by performing model training on point cloud maps and static object annotation data as training data
  • the second detection model can be obtained by point cloud data (including single frame point cloud data and/or several frames of point cloud data ) and dynamic object labeling data are used as training data to carry out model training
  • the first detection model and the second detection model can be used to detect the point cloud map and point cloud data to be marked respectively, so as to determine the point cloud map , types of objects respectively corresponding to the multiple point cloud sets included in the point cloud data.
  • the data and detection results processed by the first detection model or the second detection model can also be used as training data to further train the first detection model and the second detection model to improve The accuracy of the model is not limited in this embodiment of the present application.
  • Static objects and dynamic objects are target objects contained in point cloud data or point cloud maps.
  • Static objects are generally point cloud collections collected by detecting static objects
  • dynamic objects are generally collected by detecting dynamic objects. Collection of point clouds.
  • static objects may include but not limited to roads, trees statically placed on the side of the road, buildings, street lights, road signs, parked vehicles, etc.
  • Dynamic objects may include but not limited to vehicles driving on the road, pedestrians , animals, etc. It can be understood that, in the embodiment of the present application, the dynamic or static distinction of an object only depends on whether the object moves relative to the reference object.
  • a vehicle is a dynamic object in a driving state, and the point cloud set obtained by detecting the vehicle is called As a dynamic object, the vehicle can be considered as a static object in the non-driving state, and the point cloud collection obtained by detecting the vehicle is called a static object.
  • the smearing phenomenon also known as the smearing phenomenon, is a phenomenon in which the same objects are listed when splicing point cloud collections collected at different times for the same dynamic object. Among them, due to the mobility of dynamic objects, the positions of dynamic objects at different times are different. Therefore, in order to eliminate the impact of the mobility of dynamic objects, when splicing point cloud collections collected at different times for the same dynamic object, it will be The phenomenon that the same objects are listed is formed, that is, the tailing phenomenon.
  • the object detection method based on deep learning has become the mainstream method in object detection due to its accuracy and high efficiency.
  • the target detection model is built and trained based on deep learning, and the point cloud data to be detected is input into the trained target detection model, and the target detection results output by the target detection model can be obtained.
  • the target detection model can output multiple The type of each object and the location information of each object.
  • the point cloud data may be collected by lidar, for example.
  • the laser radar transmits a detection signal to the target object (for example, the detection signal can be a laser beam), and after the laser radar receives the reflection signal reflected from the target object, the laser radar compares the detection signal with the transmitted signal and processes it to obtain
  • the point cloud set is a set of sampling points obtained after obtaining the spatial coordinates of each sampling point on the surface of the target object.
  • Lidar can detect and sample multiple target objects at the same time to obtain a set of point cloud data.
  • the point cloud data can include multiple point cloud sets corresponding to multiple target objects. Fig.
  • a set of point cloud data includes multiple point cloud collections.
  • the point clouds in each detection frame in detection frame A, detection frame B, and detection frame C in Figure 1 constitute a point cloud collection, and each point cloud collection corresponds to a target object.
  • the point cloud used The accuracy of the ensemble can have a large impact on the performance of the object detection model.
  • a set of point cloud data includes multiple point cloud sets. Therefore, before training the target detection model, it is necessary to mark the detection frame on the point cloud data used for training, so as to divide the point cloud belonging to the same object in the point cloud data. for a point cloud collection.
  • the detection frame A, detection frame B, and detection frame C in Figure 1 divide the point cloud data into three point cloud sets, and each point cloud set corresponds to a target object.
  • FIG. 2 is a schematic diagram of a group of point cloud data with deviations in the detection frame positions.
  • FIG 2 there is a deviation in the positions of the detection frame E and the detection frame F.
  • the target detection model After dividing the point cloud data according to the detection frame E and detection frame F, If the point cloud sets corresponding to the two target objects cannot be accurately obtained, the target detection model will learn the wrong point cloud set during training, which will affect the performance of the target detection model.
  • the current processing method is manual adjustment and verification, and the detection frame position and point cloud data after manual verification are used as the training data of the target detection model. It can be seen that the current point cloud data processing methods are inefficient and difficult to guarantee accuracy.
  • a three-dimensional (3Dimensions, 3D) point cloud object automatic labeling system is proposed in some designs.
  • the system uses single frame point cloud data or Several frames of point cloud data are used as input, which are automatically marked by using the 3D detection model.
  • the rectangular frame G marks the point cloud collection corresponding to the dynamic object (such as the rear of a driving vehicle).
  • the number of point clouds included in the rectangular frame G is small, which makes it impossible to accurately identify the dynamic object.
  • the dynamic object will produce smears and cause smearing.
  • the point cloud set included in the rectangular frame G is in List of different moments. Therefore, how to improve the labeling effect of the 3D point cloud object automatic labeling system is still an important problem that needs to be solved urgently.
  • the embodiment of the present application provides a point cloud data processing method and system, which can be used to correct the detection frames of a single frame of point cloud data or several frames of point cloud data, thereby improving the labeling effect of the 3D point cloud object automatic labeling system.
  • FIG. 4 is a schematic diagram of an application scenario where the point cloud data processing method provided in the embodiment of the present application is applicable.
  • the multi-frame point cloud data in the embodiment of the present application may be data collected by a radar (such as a lidar), and the radar may be located on a vehicle.
  • a radar may be installed on the vehicle 41 shown in FIG.
  • the vehicle 41 may send the collected point cloud data to an independently deployed point cloud data processing system, and the independently deployed point cloud data processing system executes the point cloud data processing method provided in the embodiment of the present application.
  • the independently deployed point cloud data processing system may be the server 42, and the server 42 may execute the point cloud data processing method provided by the embodiment of the present application on the acquired point cloud data.
  • the independently deployed point cloud data processing system may be a terminal device, such as the mobile terminal 43 shown in FIG.
  • FIG. 5 is a schematic diagram of a point cloud data processing system provided in an embodiment of the present application.
  • the system 500 may include an acquisition unit 510, a processing unit 520, a correction unit 530, a training unit 540 and an output unit 550.
  • the obtaining unit 510 may be configured to obtain multi-frame point cloud data (for example, M frames, M is an integer greater than or equal to 2), and provide the multi-frame point cloud data to the processing unit 520 .
  • the processing unit 520 can obtain the first detection model and/or the second detection model from the training unit 540, and perform the multi-frame point cloud data processing according to the first detection model and/or the second detection model Processing is performed to obtain a detection result, which can be used to indicate the static object and/or dynamic object marked in the single frame of point cloud data in the multi-frame point cloud data.
  • the processing unit 520 can provide the detection result to the correction unit 530, and the correction unit 530 can correct the detection result to obtain a target detection frame, which can be used as the target detection result of the multi-frame point cloud data via the output from the output unit 550.
  • the multi-frame point cloud data and the target detection frame information of the multi-frame point cloud data can be provided to the training unit 540 for the training unit to perform model training to improve the model precision.
  • the above unit modules are only the functional division of the point cloud data processing system 500 , and do not limit the functions of the point cloud data processing system 500 .
  • the point cloud data processing system 500 may also include other units, and the unit modules in the point cloud data processing system 500 may also be further divided into other naming methods, which is not limited in this embodiment of the present application.
  • the acquiring unit 510 may specifically include a first acquiring unit and a second acquiring unit, and the processing unit 520 may include a first processing unit and a second processing unit, which will not be repeated here.
  • the point cloud data processing method of the embodiment of the present application is introduced below.
  • Fig. 6 is a schematic flow chart of the point cloud data processing method provided by the embodiment of the present application, wherein the method can be realized by the point cloud data processing system 500 in Fig. 5 and its functional modules, as shown in Fig. 6 , the point cloud
  • the data processing method may include the following steps:
  • S610 The point cloud data processing system acquires the first detection frame.
  • the first detection frame is used to indicate the static object marked in the first point cloud map
  • the first point cloud map can be obtained from M frames of point cloud data, where M is an integer greater than or equal to 2.
  • the M frames of point cloud data can be point cloud data of continuous frames
  • the M frames of point cloud data can be collected by a data acquisition device and obtained by an acquisition unit of the point cloud data processing system, and the data acquisition device can for example is the radar on the vehicle 41 , as shown in FIG. 4 .
  • the M frames of point cloud data may also be referred to as a point cloud sequence.
  • the first point cloud map may be a point cloud map obtained by splicing the M frames of point cloud data.
  • the point cloud data processing system can convert all single-frame point cloud data in the M frame point cloud data to a unified coordinate system (such as the world coordinate system), and convert all single-frame point cloud data in the M frame point cloud data
  • the point cloud collection in the frame point cloud data is spliced into a point cloud map in this coordinate system.
  • different The frame point cloud data is spliced; or, it can also be to select a reference object (such as a vehicle positioning attitude), and based on the reference object, after adjusting the point cloud collection in the M frame point cloud data in the coordinate system , sequentially splicing the point cloud sets in different frames of point cloud data according to the collection time, and the embodiment of the present application does not limit the specific implementation manner of the splicing process.
  • a reference object such as a vehicle positioning attitude
  • the first detection frame is a detection result obtained by performing target detection on the first point cloud map, and the first detection frame may be used to indicate that in the first point cloud map Annotated static object.
  • the target detection process on the first point cloud map can be realized by the first detection model.
  • the first detection model can be obtained by the training unit 540 mentioned above using the first training data for training, the first training data is marked training data, and the first training data can include the third point Cloud map with static object annotation data.
  • the training unit 540 may only retain The static object labeling frame is used as a positive sample, that is, static object labeling data, and the first detection model is obtained by using the third point cloud map and the static object labeling data for training.
  • the third point cloud map may be the same as or different from the first point cloud map, that is, M frames of point cloud data or the first point cloud data spliced by the M frames of point cloud data.
  • the point cloud map can be used as data to be labeled or as training data, which is not limited in this application.
  • the point cloud data processing system may directly acquire the first point cloud map and related information used to describe the first point cloud map.
  • the point cloud data processing system may also obtain the M frames of point cloud data, and splicing the M frames of point cloud data to obtain the first point cloud map.
  • the first point cloud The method of obtaining the map is not limited.
  • the point cloud data processing system can perform target detection on the first point cloud map according to the first detection model, and obtain the first detection frame.
  • the first detection frame may be used to label the static object in the first point cloud map, and the attribute information of the first detection frame may be used to describe the static object marked by the first detection frame.
  • the first detection frame may include, for example, detection frame 1, detection frame 2, and detection frame 3, and the attribute information of the first detection frame
  • it may include the position of detection frame 1/detection frame 2/detection frame 3 (such as the position of detection frame 1/detection frame 2/detection frame 3 in the first point cloud map, or the relative position with other detection frames, etc. ), size (such as length, width, radius, diameter, etc.), shape (such as cylinder, cone, cube, cuboid, irregular shape, etc.), based on the attribute information of the first detection frame, the first detection frame can be detected.
  • the position, size, shape, etc. of the static object marked by a detection frame It should be noted that the first detection frame and the attribute information of the first detection frame are only exemplarily described here. Actually, it needs to be determined according to the detection and identification of the point cloud set corresponding to the target object in the point cloud data. This will not be repeated here.
  • S620 The point cloud data processing system acquires the second detection frame.
  • the second detection frame is used to indicate the dynamic object marked in the first point cloud data
  • the first point cloud data is obtained from N frames of point cloud data
  • N is an integer greater than or equal to 1
  • N ⁇ M is an integer greater than or equal to 1
  • the point cloud data processing system determines the specific value of N or M according to the configuration information or the application scenario, and the embodiment of the present application does not limit the specific implementation manner.
  • the first point cloud data can be obtained by splicing the point cloud data of these frames, and the first point cloud data only includes the first Partial information of a point cloud map can also be called a local point cloud map.
  • the second detection frame may be a detection result obtained by performing target detection on the first point cloud data, and the second detection frame may be used to indicate that the label in the first point cloud data of dynamic objects.
  • the target detection process of the first point cloud data can be realized by the second detection model.
  • the second detection model can be obtained by training the aforementioned training unit 540 using the second training data, the second training data is labeled training data, and the second training data can include the second point cloud Data and dynamic objects label data.
  • the training data when preparing the second training data, may only retain the dynamic object labeling frame as a positive sample, that is, the dynamic object labeling data, and use the second point cloud data and the The dynamic object tag data is trained to obtain the second detection model.
  • the second point cloud data can be the same as or different from the first point cloud data, that is, the first point cloud data can be used as data to be labeled or as training data. This is not limited.
  • the point cloud data processing system can perform target detection on the first point cloud data according to the second detection model to obtain a third detection frame, according to The first detection frame corrects the third detection frame to obtain the second detection frame.
  • the third detection frame is used to indicate the dynamic object marked in the first point cloud data.
  • the detection result of the static object in the first point cloud map can be fused to the first point cloud map.
  • the false detection result of the static object in the point cloud data so as to ensure that the obtained second detection frame is only used to indicate the dynamic object marked in the first point cloud data.
  • the second detection frame can be used to label the dynamic object in the first point cloud data
  • the attribute information of the second detection frame can be used to describe the dynamic object marked by the second detection frame.
  • the third detection frame may include, for example, detection frame 4, detection frame 5, and detection frame 6, and the attribute information of the third detection frame may include, for example
  • the detection frame 4 and the detection frame 5 are used to indicate the dynamic object (for example, the point cloud collection collected by detecting the vehicle, etc.), the detection frame 6 (represented by a dotted line frame, and used to distinguish from the detection frame 4 and the detection frame indicating the dynamic object.
  • the detection frame 5 is used to indicate the static object marked due to the detection error (for example, the point cloud collection collected by detecting trees, road signs, etc.). Since the detection frame 6 is also included in the first detection frame obtained by performing target detection on the first point cloud map, the point cloud data processing system corrects the third detection frame according to the first detection frame, for example, from the third detection frame A detection frame indicating a static object is removed from the detection frame, and a third detection frame remaining after processing is used as the second detection frame. As shown in FIG. 8 , after the detection frame 6 is removed, the remaining detection frame 4 and detection frame 5 are the second detection frame, and the attribute information of the detection frame 4 and the detection frame 5 is the attribute information of the second detection frame.
  • the attribute information of the first detection frame used to mark the static object is different from the attribute information of the second detection frame used to mark the dynamic object
  • the attribute information of the first detection frame used to mark the static object is different from the attribute information of the second detection frame used to mark the dynamic object
  • the point cloud data processing system performs fusion processing on the single frame of point cloud data in the M frames of point cloud data according to the second detection frame, to obtain a second point cloud map.
  • the second detection frame may also be referred to as a dynamic object detection result, and the dynamic object detection result may be used to indicate different dynamic objects.
  • the single frame of point cloud data may only include locally collected point cloud sets for the same dynamic object Rather than the point cloud collection collected as a whole for the dynamic object, this will result in too few points in a single frame of point cloud data or points in several frames of point cloud data (called sparse point cloud), and at the same time, for different frames
  • the splicing process of point cloud data may also cause smearing caused by dynamic object splicing.
  • the point cloud data processing system can associate the second detection frames corresponding to the same dynamic object according to the attribute information of the second detection frame, and point cloud data of different frames for The point cloud collection collected by the same dynamic object is spliced at the object level to obtain the dynamic object association result (called dense point cloud).
  • the point cloud data processing system can remove the dynamic object in the single frame point cloud data in the M frame point cloud data according to the second detection frame, and obtain the third point cloud data corresponding to the single frame point cloud data , the third point cloud data only includes static objects.
  • the point cloud data processing system can use the dynamic object association result to perform fusion processing on the third point cloud data corresponding to M frames of point cloud data to obtain a second point cloud map, which includes the dynamic object and/or static objects.
  • the a-frame point cloud data collected at time t1 includes the point cloud collection collected by the head of the detection vehicle 41, and the b-frame point cloud data collected at time t2 Including the collection of point clouds collected by the body of the detection vehicle 41, the b-frame point cloud data collected at time t3 includes the collection of point clouds collected by the rear of the detection vehicle 41, respectively represented by detection frame A, detection frame B, detection frame Marked by C, it is a sparse point cloud at this time.
  • the point cloud data processing system can recognize that the detection frame A in the point cloud data of frame a, the detection frame B in the point cloud data of frame b, and the detection frame C in the point cloud data of frame c all correspond to the corresponding
  • the point cloud sets respectively marked by the detection frame A, detection frame B, and detection frame C are associated with the same dynamic object.
  • the point cloud data processing system can splice point cloud sets corresponding to the same dynamic object in different frames of point cloud data based on the granularity of the dynamic object and according to the point cloud data collection time, to obtain the dynamic
  • the relatively dense point cloud collection of the object marked by the association box D shown in Figure 9, is a dense point cloud at this time.
  • the object-level splicing process can align the same dynamic objects in different frames of point cloud data based on the point acquisition time, so as to eliminate the smearing phenomenon caused by dynamic objects.
  • the point cloud data processing system performs fusion processing on the single frame point cloud data in the multi-frame point cloud data based on the dynamic object association result, the same dynamic object can be presented in the second point cloud map A relatively dense collection of point clouds.
  • the point cloud data processing system corrects the first detection frame and the second detection frame according to the second point cloud map to obtain a target detection frame, and the target detection frame is used to indicate that in the single frame Static objects and/or dynamic objects annotated in point cloud data.
  • the point cloud The data processing system may correct different attributes of different objects according to the second point cloud map, so as to obtain target detection frames of all single frames of point cloud data.
  • the point cloud data processing system may, in the second point cloud map, correct the first attribute (such as size, position, shape, etc.)
  • the second attribute (such as size, shape, etc.) of the second detection frame of the object is corrected to obtain a fourth detection frame
  • the fourth detection frame is used to indicate the dynamic object marked in the second point cloud map and/or static object
  • the fourth detection frame of the object and the fifth detection frame are used as target detection frames of the M frames of point cloud data.
  • the target detection frame is used to mark static objects and/or dynamic objects in a single frame of point cloud data
  • the attribute information of the target detection frame is used to describe the static objects and/or dynamic objects marked by the target detection frame.
  • the above-mentioned correction to the size may include, for example, the increase and decrease of the size of the inspection frame
  • the correction to the position may, for example, include correction to the absolute position and/or relative position of the detection frame (including the detection frame in any direction).
  • the correction of the shape may include, for example, the correction of the shape of the monitoring frame, for example, from a cube to a cuboid, from a cuboid to a cube, from a cone to a cylinder, and so on.
  • the point cloud data processing system may automatically implement the above correction process, or in S640, the above correction process may be implemented manually.
  • the point cloud data processing system may output the second point cloud map and the first detection frame and/or Marking the second detection frame of the dynamic object
  • the labeler can view the second point cloud map, the first detection frame and the second detection frame through the user interface.
  • the annotator can correct the first attributes such as the size, position, and shape of the first detection frame in the second point cloud map, and correct the second attributes such as the size and shape of the second detection frame to obtain the fourth detection frame. box information.
  • the annotator corrects the third attributes such as the position of the fourth detection frame of the marked dynamic object in each single frame of point cloud data corresponding to the second point cloud map to obtain the fifth detection frame.
  • the fourth detection frame and the fifth detection frame of the static object are used as the target detection frame (ie, the labeling result) of the single frame point cloud data in the M frame point cloud data, Not only can more point cloud information of each object be obtained in a single frame of point cloud data, but also the ability of the point cloud data processing system to automatically label point cloud collections corresponding to distant objects or severely occluded objects in a single frame of point cloud data can be improved. It can also reduce the smearing phenomenon caused by dynamic objects by aligning dynamic objects in multi-frame point cloud data to the relevant features of static objects. At the same time, based on the auxiliary labeling of point cloud maps containing dynamic objects and static objects, convenient auxiliary corrections can be realized, which facilitates the labeling of accurate object attributes at a lower labor cost.
  • the point cloud data processing system can also use multi-model fusion and point cloud map to generate a softening frame to provide a higher-quality detection frame and a softening frame that is easy to train. It helps to improve the training accuracy of models using automatic labeling systems for partial frame point cloud data.
  • the method may also include the following steps: using the single-frame point cloud data in the M frames of point cloud data and the target detection frame as the third training data, based on the third training In the process of data training multiple detection models, the sixth detection frame is determined, and the detection result of the sixth detection frame is an error; the second point cloud map and the target detection frame are used as the fourth training data, based on the Determining a seventh detection frame during the process of training multiple detection models with the fourth training data; correcting the seventh detection frame based on the sixth detection frame.
  • the sixth detection frame with an incorrect detection result is used instead of the seventh detection frame with a correct detection result, and the sixth detection frame with a misaligned detection result and the seventh detection frame with an incorrect detection result are eliminated.
  • the detection frame obtained after correction (for example, denoted as the eighth detection frame) includes a softening frame for training and attribute information of the softening frame, and the softening frame and its attribute information can also replace the above-mentioned single frame point
  • the target detection frame of the cloud data is used as training data to train the target detection model, thereby further improving the training accuracy of the single-frame point cloud data detection model.
  • point cloud data processing method will be introduced below by taking a vehicle driving scene as an example in conjunction with the method flow chart shown in FIG. 10 .
  • the method may include the following steps:
  • the acquisition unit of the point cloud data processing system acquires a point cloud sequence.
  • the point cloud sequence can be collected by the radar on the vehicle, for example, and the point cloud sequence includes continuous M frames of point cloud data (M is an integer greater than or equal to 2, and the value of M depends on the point cloud data processing system processing capabilities, configuration information or application scenarios, etc.).
  • the acquisition unit of the point cloud data processing system may acquire the point cloud sequence from the vehicle, it may be that the vehicle actively reports the point cloud sequence, or the vehicle responds to the point cloud data
  • the request of the processing system feeds back the sequence of point clouds.
  • the point cloud data processing system may include a storage unit, the M frames of point cloud data may be stored in the storage unit, and the acquisition unit may read the point cloud sequence from the storage unit. This example does not limit the acquisition method of the point cloud sequence.
  • the processing unit of the point cloud data processing system transforms the single frame point cloud data in the M frame point cloud data into the world coordinate system according to the positioning attitude of the vehicle, and stitches the M frame point cloud data in the world coordinate system , get the first point cloud map.
  • S1022 The processing unit of the point cloud data processing system performs object detection on the first point cloud map according to the first detection model to obtain a first detection frame.
  • S1021-S1022 refer to the relevant introduction in conjunction with S610 above, and will not be repeated here.
  • S1031 (optional): The processing unit of the point cloud data processing system splices the N frames of point cloud data in the M frames of point cloud data to obtain the first point cloud data. It can be understood that S1031 is only executed when several frames of point cloud data (that is, N ⁇ 2) in M frames of point cloud data are combined as the first point cloud data, if the M frames of point cloud data are If a single frame of point cloud data is used as the first point cloud data, S1031 does not need to be executed.
  • the processing unit of the point cloud data processing system performs object detection on the first point cloud data according to the second detection model to obtain a third detection frame.
  • the third detection frame is used to label dynamic objects in the first point cloud data. It should be understood that within the error range of the second detection model, the third detection frame may also be used to label static objects in the first point cloud data.
  • S1033 The processing unit of the point cloud data processing system corrects the third detection frame according to the first detection frame to obtain the second detection frame, and the second detection frame is used to label the first point cloud Dynamic objects in data.
  • S1031-S1033 refer to the relevant introduction in connection with S620 above, and will not be repeated here.
  • the processing unit of the point cloud data processing system removes the dynamic object in the single frame point cloud data in the M frame point cloud data according to the second detection frame obtained in S1033, and obtains the third point corresponding to the single frame point cloud data cloud data.
  • S1042 The processing unit of the point cloud data processing system performs splicing processing according to the third point cloud data corresponding to the M frames of point cloud data, to generate a fourth point cloud map.
  • S1043 The processing unit of the point cloud data processing system uses the second detection frame obtained in S1033 to correspond to the same The second detection frame of the dynamic object is associated.
  • the processing unit of the point cloud data processing system performs object-level stitching on the point cloud set marked in the second detection frame corresponding to the same dynamic object obtained in S1043 according to the M frame point cloud data, that is, in different frames of point cloud data
  • the point cloud sets associated with the same dynamic object are aligned and integrated into a dynamic object-level point cloud set at the same time to obtain the dynamic object association result.
  • the processing unit of the point cloud data processing system performs fusion processing on the fourth point cloud map obtained in S1042 according to the dynamic object association result obtained in S1044, to obtain a second point cloud map.
  • the correction unit of the point cloud data processing system obtains the second point cloud map in S1045 through the user interface output, and the labeler can correct the size, shape, position, etc. of the first detection frame of the static object in the second point cloud map
  • the first attribute, and the second attribute such as the size and shape of the second detection frame marking the dynamic object are used to obtain the fourth detection frame.
  • the correction unit of the point cloud data processing system provides the static object and/or dynamic object marked in the second point cloud map by the fourth detection frame obtained after attribute correction to the single frame point corresponding to the second point cloud map
  • the cloud data is output on the user interface, and the annotator can correct the third attribute such as the position of the fourth detection frame of the annotated dynamic object in the single frame point cloud data.
  • the labeling result of the single frame point cloud data is obtained.
  • the training unit of the point cloud data processing system uses the single frame of point cloud data in the M frames of point cloud data and the target detection frame as the third training data, and trains multiple detection models based on the third training data In the process of determining the sixth detection frame, the detection result of the sixth detection frame information is wrong.
  • the training unit of the point cloud data processing system uses the second point cloud map and the target detection frame as fourth training data, and determines a seventh detection during the process of training multiple detection models based on the fourth training data frame.
  • the correction unit of the point cloud data processing system corrects the seventh detection frame based on the sixth detection frame.
  • FIG. 11 is a schematic structural diagram of a point cloud data processing system 1100 provided in an embodiment of the present application.
  • the point cloud data processing system 1100 can be applied to the system shown in FIG. 4 A server or terminal device in an application scenario.
  • the point cloud data processing system 1100 may include a first acquisition unit 1101 , a second acquisition unit 1102 , a processing unit 1103 and a correction unit 1104 .
  • the first acquisition unit 1101 is configured to acquire a first detection frame, the first detection frame is used to indicate a static object marked in the first point cloud map, and the first point cloud map consists of M frames
  • the point cloud data is obtained, M is an integer greater than or equal to 2
  • the second acquisition unit 1102 is used to acquire second detection frame information, and the second detection frame is used to indicate the dynamic object marked in the first point cloud data
  • the first point cloud data is obtained from N frames of point cloud data, N is an integer greater than or equal to 1, and N ⁇ M
  • the processing unit 1103 is configured to process the M frames of point cloud data according to the second detection frame Fusion processing is performed on the single-frame point cloud data in to obtain a second point cloud map
  • the correction unit 1104 is configured to correct the first detection frame and the second detection frame according to the second point cloud map to obtain A target detection frame, where the target detection frame is used to indicate annotated static objects and/or dynamic objects marked in the single frame point cloud data.
  • the first acquisition unit 1101 is configured to: stitch the M frames of point cloud data to obtain the first point cloud map; Object detection is performed on the cloud map to obtain the first detection frame.
  • the first detection model is obtained through training using first training data, and the first training data includes the third point cloud map and static object labeling data.
  • the second acquisition unit 1102 is configured to: perform object detection on the first point cloud data according to a second detection model to obtain a third detection frame, wherein the second detection model Obtained by using the second training data for training, the second training data includes the second point cloud data and dynamic object labeling data, and the third detection frame is used to indicate the dynamic object marked in the first point cloud data; Correcting the third detection frame according to the first detection frame to obtain the second detection frame.
  • the processing unit 1103 is configured to: according to the second detection frame, remove the dynamic object in the single frame of point cloud data in the M frames of point cloud data to obtain the single frame of point cloud data The corresponding third point cloud data; according to the attribute information of the second detection frame, associate the second detection frame corresponding to the same dynamic object to obtain a dynamic object association result; based on the dynamic object association result, the M frame The third point cloud data corresponding to the point cloud data is fused to obtain the second point cloud map.
  • the correcting unit 1104 is configured to: in the second point cloud map, correct the first attribute of the first detection frame marking the static object, and/or correct the dynamic
  • the second attribute of the second detection frame of the object is corrected to obtain the fourth detection frame; in the single frame point cloud data corresponding to the second point cloud map, the third attribute of the fourth detection frame marked with the dynamic object is modified Correction to obtain the fifth detection frame; the fourth detection frame marked with the static object and the fifth detection frame are used as the target detection frame of the M frames of point cloud data.
  • the system further includes a training unit.
  • the training unit is configured to perform the following steps: convert the single frame point cloud in the M frames of point cloud data to Data and the target detection frame are used as the third training data, and a sixth detection frame is determined during the process of training multiple detection models based on the third training data, and the detection result of the sixth detection frame is an error; the The second point cloud map and the target detection frame are used as the fourth training data, and the seventh detection frame is determined in the process of training a plurality of detection models based on the fourth training data; the seventh detection frame is corrected based on the sixth detection frame. Seven detection boxes.
  • FIG. 12 is a schematic structural diagram of a point cloud data processing device 1200 provided in an embodiment of the present application.
  • the data processing device 1200 can be applied to the scene shown in FIG. 4 server or terminal device in the Referring to FIG. 12 , the point cloud data processing device 1200 includes: a processor 1201 , a memory 1202 and a bus 1203 . Wherein, the processor 1201 and the memory 1202 communicate through the bus 1203 , or communicate through other means such as wireless transmission.
  • the memory 1202 is used to store instructions, and the processor 1201 is used to execute the instructions stored in the memory 1202 .
  • the memory 1202 stores program codes, and the processor 1201 can call the program codes stored in the memory 1202 .
  • the processor 1201 when the data processing device 1200 is a point cloud data processing device, the processor 1201 is used to execute the above-mentioned method embodiment.
  • the relevant description above please refer to the relevant description above, which will not be repeated here. repeat.
  • the memory 1202 in FIG. 12 of the present application may be a volatile memory or a non-volatile memory, or may include both volatile and non-volatile memories.
  • the non-volatile memory can be read-only memory (Read-Only Memory, ROM), programmable read-only memory (Programmable ROM, PROM), erasable programmable read-only memory (Erasable PROM, EPROM), electronically programmable Erase Programmable Read-Only Memory (Electrically EPROM, EEPROM) or Flash.
  • the volatile memory can be Random Access Memory (RAM), which acts as external cache memory.
  • RAM Static Random Access Memory
  • SRAM Static Random Access Memory
  • DRAM Dynamic Random Access Memory
  • Synchronous Dynamic Random Access Memory Synchronous Dynamic Random Access Memory
  • SDRAM double data rate synchronous dynamic random access memory
  • Double Data Rate SDRAM, DDR SDRAM enhanced synchronous dynamic random access memory
  • Enhanced SDRAM, ESDRAM synchronous connection dynamic random access memory
  • Synchlink DRAM, SLDRAM Direct Memory Bus Random Access Memory
  • Direct Rambus RAM Direct Rambus RAM
  • the embodiments of the present application further provide a computer program, which, when the computer program is run on a computer, causes the computer to execute the above method embodiments.
  • an embodiment of the present application further provides a computer-readable storage medium, in which a computer program is stored, and when the computer program is executed by a computer, the computer executes the above-mentioned method embodiment.
  • the storage medium may be any available medium that can be accessed by a computer.
  • computer-readable media may include RAM, ROM, EEPROM, CD-ROM or other optical disk storage, magnetic disk storage media or other magnetic storage devices, or may be used to carry or store information in the form of instructions or data structures desired program code and any other medium that can be accessed by a computer.
  • an embodiment of the present application further provides a chip for reading a computer program stored in a memory to implement the above method embodiments.
  • an embodiment of the present application provides a chip system, where the chip system includes a processor, configured to support a computer device to implement the above method embodiments.
  • the chip system further includes a memory, and the memory is used to store necessary programs and data of the computer device.
  • the system-on-a-chip may consist of chips, or may include chips and other discrete devices.
  • the embodiments of the present application may be provided as methods, systems, or computer program products. Accordingly, the present application may take the form of an entirely hardware embodiment, an entirely software embodiment, or an embodiment combining software and hardware aspects. Furthermore, the present application may take the form of a computer program product embodied on one or more computer-usable storage media (including but not limited to disk storage, CD-ROM, optical storage, etc.) having computer-usable program code embodied therein.
  • computer-usable storage media including but not limited to disk storage, CD-ROM, optical storage, etc.
  • These computer program instructions may also be stored in a computer-readable memory capable of directing a computer or other programmable data processing apparatus to operate in a specific manner, such that the instructions stored in the computer-readable memory produce an article of manufacture comprising instruction means, the instructions
  • the device realizes the function specified in one or more procedures of the flowchart and/or one or more blocks of the block diagram.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Computing Systems (AREA)
  • Databases & Information Systems (AREA)
  • Evolutionary Computation (AREA)
  • General Health & Medical Sciences (AREA)
  • Medical Informatics (AREA)
  • Software Systems (AREA)
  • Image Processing (AREA)

Abstract

A point cloud data processing method and system, which relate to the technical field of data processing. The method comprises: acquiring a first detection frame, the first detection frame being used for indicating a static object labeled in a first point cloud map, and the first point cloud map being obtained by M point cloud data frames; acquiring a second detection frame, the second detection frame being used for indicating a dynamic object labeled in the first point cloud data, the first point cloud data being obtained by N point cloud data frames, and N<M; fusing single-frame point cloud data among the M point cloud data frames according to the second detection frame to obtain a second point cloud map; and correcting the first detection frame and the second detection frame according to the second point cloud map to obtain a target detection frame. By fusing the detection result of the first point cloud map obtained by multiple point cloud data frames and the detection result of the first point cloud data obtained by some point cloud data frames among the multiple point cloud data frames, the labeling capability of the point cloud data processing system is improved.

Description

一种点云数据处理方法及系统A point cloud data processing method and system
相关申请的交叉引用Cross References to Related Applications
本申请要求在2021年10月27日提交中华人民共和国知识产权局、申请号为202111253590.5、申请名称为“一种点云数据处理方法及系统”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。This application claims the priority of a Chinese patent application filed with the Intellectual Property Office of the People's Republic of China on October 27, 2021, with application number 202111253590.5, and application title "A Method and System for Processing Point Cloud Data", the entire contents of which are incorporated by reference incorporated in this application.
技术领域technical field
本申请实施例涉及数据处理技术领域,特别涉及一种点云数据处理方法及系统。The embodiments of the present application relate to the technical field of data processing, and in particular to a point cloud data processing method and system.
背景技术Background technique
激光雷达(lightlaser detection and ranging,Lidar),是激光探测及测距系统的简称。激光雷达相比于相机、超声波传感器等传统传感器,具有测量精度高、响应速度快、抗干扰能力强的特点。激光雷达在智能驾驶领域和无人驾驶领域中均得到了广泛的应用。其中,激光雷达在工作中,向目标物体发射探测信号(例如探测信号可以为激光束),在激光雷达接收到的从目标物体反射回来的反射信号之后,激光雷达将探测信号与发射信号进行比较并进行处理得到点云集合,点云集合是在获取目标物体表面每个采样点的空间坐标后得到的采样点集合。激光雷达在工作中可以同时对多个目标物体同时进行探测并采样,得到一组点云数据,点云数据可以包括多个目标物体对应的多个点云集合。Lidar (lightlaser detection and ranging, Lidar) is the abbreviation of laser detection and ranging system. Compared with traditional sensors such as cameras and ultrasonic sensors, lidar has the characteristics of high measurement accuracy, fast response speed, and strong anti-interference ability. LiDAR has been widely used in the field of intelligent driving and unmanned driving. Among them, the lidar transmits a detection signal to the target object during work (for example, the detection signal can be a laser beam), and after the lidar receives the reflected signal reflected from the target object, the lidar compares the detection signal with the transmitted signal and processed to obtain a point cloud set, which is a set of sampling points obtained after obtaining the spatial coordinates of each sampling point on the surface of the target object. Lidar can simultaneously detect and sample multiple target objects at the same time to obtain a set of point cloud data. The point cloud data can include multiple point cloud sets corresponding to multiple target objects.
目前,为了降低对点云数据进行目标检测的人工成本并提升检测精度,在一些设计中提出一种点云3D物体自动标注系统,该系统以单帧点云数据或若干帧点云数据作为输入,通过使用3D检测模型进行自动标注。然而,当目标物体受遮挡或距离过远时,由于所述单帧点云数据或若干帧点云数据中的点云数据过少会导致误检、漏检、检测框不准确等问题,动态物体在不同时刻采集的点云数据中的偏移对点云数据处理过程中引发的拖尾现象等,会严重影响系统自动标注效果。At present, in order to reduce the labor cost of object detection on point cloud data and improve the detection accuracy, a point cloud 3D object automatic labeling system is proposed in some designs. The system takes a single frame of point cloud data or several frames of point cloud data as input , by using the 3D detection model for automatic annotation. However, when the target object is occluded or the distance is too far, due to the lack of point cloud data in the single frame point cloud data or several frames of point cloud data, it will lead to problems such as false detection, missed detection, and inaccurate detection frames. The offset in the point cloud data collected by the object at different times will seriously affect the automatic labeling effect of the system due to the smearing phenomenon caused by the point cloud data processing process.
因此,如何提升点云3D物体自动标注系统的标注效果,仍为亟需解决的重要问题。Therefore, how to improve the labeling effect of the point cloud 3D object automatic labeling system is still an important problem that needs to be solved urgently.
发明内容Contents of the invention
本申请实施例提供了一种点云数据处理方法及系统,通过利用对由多帧点云数据得到的第一点云地图的检测结果,以及由所述多帧点云数据中的部分帧点云数据得到的第一点云数据的检测结果进行融合处理,提升点云数据处理系统对部分帧点云数据的标注能力。The embodiment of the present application provides a point cloud data processing method and system, by using the detection result of the first point cloud map obtained from the multi-frame point cloud data, and part of the frame points in the multi-frame point cloud data The detection results of the first point cloud data obtained from the cloud data are fused to improve the ability of the point cloud data processing system to label part of the frame point cloud data.
第一方面,本申请实施例提供了一种点云数据处理方法,该方法可由点云数据处理系统实现,所述方法包括:获取第一检测框,所述第一检测框用于指示在第一点云地图中标注的静态对象,所述第一点云地图由M帧点云数据得到,M为大于或等于2的整数;获取第二检测框,所述第二检测框用于指示在第一点云数据中标注的动态对象,所述第一点云数据由N帧点云数据得到,N为大于或等于1的整数,N<M;根据所述第二检测框,对所述M帧点云数据中的单帧点云数据进行融合处理,得到第二点云地图;根据所述第二点云地图对所述第一检测框和所述第二检测框进行校正,得到标检测框,所述目标检测框 用于指示在所述单帧点云数据中标注的标注静态对象和/或动态对象。In the first aspect, the embodiment of the present application provides a point cloud data processing method, which can be implemented by a point cloud data processing system, and the method includes: obtaining a first detection frame, the first detection frame is used to indicate the The static object marked in the point cloud map, the first point cloud map is obtained by M frame point cloud data, M is an integer greater than or equal to 2; obtain the second detection frame, the second detection frame is used to indicate the The dynamic object marked in the first point cloud data, the first point cloud data is obtained from N frames of point cloud data, N is an integer greater than or equal to 1, and N<M; according to the second detection frame, the The single-frame point cloud data in the M frame point cloud data is fused and processed to obtain a second point cloud map; the first detection frame and the second detection frame are corrected according to the second point cloud map to obtain a label A detection frame, the target detection frame is used to indicate the marked static object and/or dynamic object marked in the single frame point cloud data.
通过上述方法,点云数据处理系统可以将M帧点云数据和所述M帧点云数据中的N帧点云数据分别作为待标注数据进行不同的处理后,在将处理后的点云数据进行融合,并获得每一单帧点云数据的经过校正的标注结果,从而利用M帧点云数据对N帧点云数据进行补偿,减少N帧点云数据因动态对象带来的拖尾现象,以及N帧点云数据中点数过少导致的误检、漏检、检测框不准确等问题,以提升点云数据处理系统的标注能力。Through the above method, the point cloud data processing system can use the M frames of point cloud data and the N frames of point cloud data in the M frames of point cloud data as data to be labeled for different processing, and then process the processed point cloud data Perform fusion and obtain the corrected labeling results of each single frame of point cloud data, so as to use M frame point cloud data to compensate N frame point cloud data, and reduce the tailing phenomenon of N frame point cloud data caused by dynamic objects , and problems such as false detection, missed detection, and inaccurate detection frames caused by too few points in the N-frame point cloud data, so as to improve the labeling ability of the point cloud data processing system.
需要说明的是,本申请实施例中,N<M,该N帧点云数据为该M帧点云数据的部分帧点云数据,其中,在N=1时,该N帧点云数据为单帧点云数据,N>1时,该N帧点云数据包括若干帧点云数据。在具体实施时,该点云数据处理系统按照配置信息或者应用场景中确定该N或M的具体取值,本申请实施例对此具体实现方式不做限定。It should be noted that, in the embodiment of the present application, N<M, the N frames of point cloud data are partial frames of point cloud data of the M frames of point cloud data, wherein, when N=1, the N frames of point cloud data are For a single frame of point cloud data, when N>1, the N frames of point cloud data include several frames of point cloud data. During specific implementation, the point cloud data processing system determines the specific value of N or M according to the configuration information or the application scenario, and the embodiment of the present application does not limit the specific implementation manner.
需要说明的是,本申请实施例中,静态对象和动态对象,是包含于点云数据或者点云地图的目标对象,静态对象一般为对静态物体进行探测采集到的点云集合,动态对象一般为对动态物体进行探测采集到的点云集合。其中,由于动态物体的可移动性,动态物体在不同时刻的位置是不同的,故而为消除动态物体的移动性影响,在针对同一动态物体在不同时刻采集到的点云集合进行拼接时,会形成相同对象罗列的现象,即为本申请实施例述及的拖尾现象,也可称为拖影现象。It should be noted that in this embodiment of the application, static objects and dynamic objects are target objects included in point cloud data or point cloud maps. Static objects are generally point cloud collections collected by detecting static objects, and dynamic objects are generally A collection of point clouds collected for the detection of dynamic objects. Among them, due to the mobility of dynamic objects, the positions of dynamic objects at different times are different. Therefore, in order to eliminate the impact of the mobility of dynamic objects, when splicing point cloud collections collected at different times for the same dynamic object, it will be The phenomenon of listing the same objects is the smear phenomenon mentioned in the embodiment of the present application, which may also be called the smear phenomenon.
结合第一方面,在一种可能的实现方式中,所述获取第一检测框,包括:对所述M帧点云数据进行拼接,得到所述第一点云地图;根据第一检测模型对所述第一点云地图进行目标检测,获得所述第一检测框,其中,所述第一检测模型使用第一训练数据进行训练得到,所述第一训练数据包括第三点云地图与静态对象标注数据。With reference to the first aspect, in a possible implementation manner, the acquiring the first detection frame includes: splicing the M frames of point cloud data to obtain the first point cloud map; The first point cloud map performs target detection to obtain the first detection frame, wherein the first detection model is obtained by training using first training data, and the first training data includes the third point cloud map and static Object annotation data.
通过上述方法,可以利用点云地图(例如第三点云地图)训练用于标注静态对象的第一检测模型,提供高精准的静态对象检测模型。进一步地,点云数据处理系统可以利用该第一检测模型对待标注的点云地图(例如第一点云地图)中的静态对象进行标注,以便利用对点云地图中静态对象的检测结果,在M帧点云数据中将动态对象对准到静态对象的相关特征,减少N帧点云数据因动态对象带来的拖尾现象,从而减少对动态对象的标注难度。Through the above method, a point cloud map (for example, the third point cloud map) can be used to train the first detection model for labeling static objects, so as to provide a high-precision static object detection model. Further, the point cloud data processing system can use the first detection model to mark the static objects in the point cloud map to be marked (for example, the first point cloud map), so that the detection results of the static objects in the point cloud map can be used in the In the M-frame point cloud data, the dynamic object is aligned with the relevant features of the static object, and the tailing phenomenon caused by the dynamic object is reduced in the N-frame point cloud data, thereby reducing the difficulty of labeling the dynamic object.
应理解,本申请实施例中,第三点云地图可以与第一点云地图相同或不同,也就是说,第一点云地图既可以作为待标注数据也可以为训练数据,本申请对此不做限定。It should be understood that in the embodiment of the present application, the third point cloud map may be the same as or different from the first point cloud map, that is to say, the first point cloud map may be used as data to be labeled or as training data. No limit.
结合第一方面,在一种可能的实现方式中,所述获取第二检测框,包括:根据第二检测模型对所述第一点云数据进行目标检测,获得第三检测框,其中,所述第二检测模型使用第二训练数据进行训练得到,所述第二训练数据包括第二点云数据和动态对象标注数据,所述第三检测框用于指示在所述第一点云数据中标注的动态对象;根据所述第一检测框对所述第三检测框进行校正,得到所述第二检测框。With reference to the first aspect, in a possible implementation manner, the acquiring the second detection frame includes: performing object detection on the first point cloud data according to the second detection model to obtain a third detection frame, wherein the The second detection model is obtained through training using second training data, the second training data includes second point cloud data and dynamic object labeling data, and the third detection frame is used to indicate that in the first point cloud data A marked dynamic object; correcting the third detection frame according to the first detection frame to obtain the second detection frame.
通过上述方法,可以利用第二点云数据训练用于标注动态对象的第二检测模型,提供高精准的动态对象检测模型。进一步地,点云数据处理系统可以利用该第二检测模型对待标注的N帧点云数据中的动态对象进行标注,以便获得N帧点云数据中的动态对象检测结果,该动态对象检测结果可用于对M帧点云数据进行融合处理,以将M帧点云数据中的动态对象以消除拖尾现象的形式呈现,从而方便后续校正过程,提升该点云数据处理系统的检测精度。Through the above method, the second point cloud data can be used to train the second detection model for labeling the dynamic object, so as to provide a high-precision dynamic object detection model. Further, the point cloud data processing system can use the second detection model to mark the dynamic objects in the N frames of point cloud data to be marked, so as to obtain the dynamic object detection results in the N frames of point cloud data, and the dynamic object detection results can be used It is used to fuse M frames of point cloud data to present dynamic objects in M frames of point cloud data in a form that eliminates smearing, so as to facilitate the subsequent correction process and improve the detection accuracy of the point cloud data processing system.
结合第一方面,在一种可能的实现方式中,所述根据所述第二检测框,对所述M帧点云数据中的单帧点云数据进行融合处理,得到第二点云地图,包括:根据所述第二检测框, 去除所述M帧点云数据中单帧点云数据中的动态对象,得单帧点云数据对应的第三点云数据;根据所述第二检测框的属性信息,将对应相同动态对象的第二检测框进行关联,得到动态对象关联结果;基于所述动态对象关联结果对所述M帧点云数据对应的第三点云数据进行融合处理,得到第二点云地图。With reference to the first aspect, in a possible implementation manner, according to the second detection frame, performing fusion processing on the single frame point cloud data in the M frames of point cloud data to obtain a second point cloud map, Including: according to the second detection frame, removing the dynamic object in the single frame point cloud data of the M frames of point cloud data to obtain the third point cloud data corresponding to the single frame point cloud data; according to the second detection frame Attribute information, associate the second detection frame corresponding to the same dynamic object to obtain the dynamic object association result; based on the dynamic object association result, perform fusion processing on the third point cloud data corresponding to the M frame point cloud data to obtain Second point cloud map.
通过上述方法,点云数据处理系统可以利用对N帧点云数据中动态对象的检测结果,实现对M帧点云数据中的动态对象的对象级别关联后,消除因动态对象引起的拖尾现象,以得到新的点云地图,以便所述新的点云地图中的动态对象以消除拖尾现象的形式呈现,从而方便后续校正过程。Through the above method, the point cloud data processing system can use the detection results of the dynamic objects in the N-frame point cloud data to realize the object-level association of the dynamic objects in the M-frame point cloud data, and eliminate the tailing phenomenon caused by the dynamic objects , to obtain a new point cloud map, so that the dynamic objects in the new point cloud map are presented in a form that eliminates the trailing phenomenon, thereby facilitating the subsequent correction process.
结合第一方面,在一种可能的实现方式中,根据所述第二点云地图对所述第一检测框和所述第二检测框进行校正,得到目标检测框,包括:在所述第二点云地图中,对标注静态对象的第一检测框的第一属性进行修正,和/或,对标注动态对象的第二检测框的第二属性进行修正,得到第四检测框,第四检测框用于指示所述第二点云地图中标注的动态对象和/或静态对象;在所述第二点云地图对应的单帧点云数据中,对标注动态对象的第四检测框的第三属性进行修正,得到第五检测框;以标注静态对象的第四检测框和所述第五检测框作为所述M帧点云数据的目标检测框。With reference to the first aspect, in a possible implementation manner, correcting the first detection frame and the second detection frame according to the second point cloud map to obtain a target detection frame includes: In the two-point cloud map, the first attribute of the first detection frame marked with the static object is corrected, and/or the second attribute of the second detection frame marked with the dynamic object is corrected to obtain the fourth detection frame, the fourth The detection frame is used to indicate the dynamic object and/or static object marked in the second point cloud map; in the single frame point cloud data corresponding to the second point cloud map, the fourth detection frame of the marked dynamic object The third attribute is corrected to obtain the fifth detection frame; the fourth detection frame marked with the static object and the fifth detection frame are used as the target detection frame of the M frames of point cloud data.
通过上述方法,点云数据处理系统例如可以在该第二点云地图中对标注静态对象和/或动态对象的检测框的相关属性进行修正,有助于提升该点云数据处理系统的检测精度。其中,该点云数据处理系统可以采用人工修正方式或者自动修正方式实现该校正过程,本申请实施例对此具体实现方式不做限定。Through the above method, the point cloud data processing system can, for example, correct the relevant attributes of the detection frame marked with static objects and/or dynamic objects in the second point cloud map, which helps to improve the detection accuracy of the point cloud data processing system . Wherein, the point cloud data processing system may implement the correction process by manual correction or automatic correction, and the embodiment of the present application does not limit the specific implementation manner.
结合第一方面,在一种可能的实现方式中,在得到所述目标检测框之后,所述方法还包括:将所述M帧点云数据中的单帧点云数据以及所述目标检测框作为第三训练数据,在基于所述第三训练数据训练多个检测模型的过程中确定第六检测框,所述第六检测框的检测结果为错误;将所述第二点云地图以及所述目标检测框作为第四训练数据,在基于所述第四训练数据训练多个检测模型的过程中确定第七检测框;基于所述第六检测框校正所述第七检测框。With reference to the first aspect, in a possible implementation manner, after obtaining the target detection frame, the method further includes: combining the single frame point cloud data in the M frames of point cloud data and the target detection frame As the third training data, in the process of training a plurality of detection models based on the third training data, a sixth detection frame is determined, and the detection result of the sixth detection frame is an error; the second point cloud map and the The target detection frame is used as fourth training data, and a seventh detection frame is determined during the process of training multiple detection models based on the fourth training data; and the seventh detection frame is corrected based on the sixth detection frame.
通过上述方法,点云数据处理系统可以利用多模型融合与点云地图生成柔化框,提供更高质量的检测框和便于训练的柔化框,有助于提升使用自动标注系统的对单帧或者若干帧点云数据的模型的训练精度。Through the above method, the point cloud data processing system can use multi-model fusion and point cloud map to generate softening frames, provide higher-quality detection frames and softening frames that are easy to train, and help improve the accuracy of single frames using automatic labeling systems. Or the training accuracy of the model of several frames of point cloud data.
第二方面,本申请实施例提供了一种点云数据处理系统,包括:第一获取单元,用于获取第一检测框,所述第一检测框用于指示在第一点云地图中标注的静态对象,所述第一点云地图由M帧点云数据得到,M为大于或等于2的整数;第二获取单元,用于获取第二检测框,所述第二检测框用于指示在第一点云数据中标注的动态对象,所述第一点云数据由N帧点云数据得到,N为大于或等于1的整数,N<M;处理单元,用于根据所述第二检测框,对所述M帧点云数据中的单帧点云数据进行融合处理,得到第二点云地图;校正单元,用于根据所述第二点云地图对所述第一检测框和所述第二检测框进行校正,得到目标检测框,所述目标检测框用于指示在所述单帧点云数据中标注的标注静态对象和/或动态对象。In the second aspect, the embodiment of the present application provides a point cloud data processing system, including: a first acquisition unit, configured to acquire a first detection frame, and the first detection frame is used to indicate static object, the first point cloud map is obtained from M frames of point cloud data, and M is an integer greater than or equal to 2; the second acquisition unit is used to acquire a second detection frame, and the second detection frame is used to indicate The dynamic object marked in the first point cloud data, the first point cloud data is obtained from N frames of point cloud data, N is an integer greater than or equal to 1, and N<M; the processing unit is used for according to the second The detection frame is used to fuse the single-frame point cloud data in the M frames of point cloud data to obtain a second point cloud map; a correction unit is used to correct the first detection frame and the second point cloud map according to the second point cloud map. The second detection frame is corrected to obtain a target detection frame, and the target detection frame is used to indicate the marked static object and/or dynamic object marked in the single frame point cloud data.
需要说明的是,本申请实施例中,第一获取单元和第二获取单元可以为同一单元也可以为不同单元,本申请实施例对此产品形态不做限定。It should be noted that, in the embodiment of the present application, the first acquisition unit and the second acquisition unit may be the same unit or different units, and the embodiment of the present application does not limit the product form.
结合第二方面,在一种可能的实现方式中,所述第一获取单元用于:对所述M帧点云 数据进行拼接,得到所述第一点云地图;根据第一检测模型对所述第一点云地图进行目标检测,获得所述第一检测框,其中,所述第一检测模型使用第一训练数据进行训练得到,所述第一训练数据包括第三点云地图与静态对象标注数据。With reference to the second aspect, in a possible implementation manner, the first acquisition unit is configured to: stitch the M frames of point cloud data to obtain the first point cloud map; The first point cloud map is used for target detection to obtain the first detection frame, wherein the first detection model is obtained by training using the first training data, and the first training data includes the third point cloud map and static objects Annotate the data.
结合第二面,在一种可能的实现方式中,所述第二获取单元用于:根据第二检测模型对所述第一点云数据进行目标检测,获得第三检测框,其中,所述第二检测模型使用第二训练数据进行训练得到,所述第二训练数据包括第二点云数据和动态对象标注数据,所述第三检测框用于指示在所述第一点云数据中标注的动态对象;根据所述第一检测框对所述第三检测框进行校正,得到所述第二检测框。With reference to the second aspect, in a possible implementation manner, the second acquisition unit is configured to: perform object detection on the first point cloud data according to a second detection model to obtain a third detection frame, wherein the The second detection model is obtained by training using the second training data, the second training data includes the second point cloud data and dynamic object labeling data, and the third detection frame is used to indicate the labeling in the first point cloud data the dynamic object; correcting the third detection frame according to the first detection frame to obtain the second detection frame.
结合第二方面,在一种可能的实现方式中,所述处理单元用于:根据所述第二检测框,去除所述M帧点云数据中单帧点云数据中的动态对象,得到单帧点云数据对应的第三点云数据;根据所述第二检测框的属性信息,将对应相同动态对象的第二检测框进行关联,得到动态对象关联结果;基于所述动态对象关联结果对所述M帧点云数据对应的第三点云数据进行融合处理,得到第二点云地图。With reference to the second aspect, in a possible implementation manner, the processing unit is configured to: remove dynamic objects in a single frame of point cloud data in the M frames of point cloud data according to the second detection frame, to obtain a single The third point cloud data corresponding to the frame point cloud data; according to the attribute information of the second detection frame, correlating the second detection frame corresponding to the same dynamic object to obtain a dynamic object association result; based on the dynamic object association result to The third point cloud data corresponding to the M frames of point cloud data are fused to obtain a second point cloud map.
结合第二方面,在一种可能的实现方式中,所述校正单元用于:在所述第二点云地图中,对标注静态对象的第一检测框的第一属性进行修正,和/或,对标注动态对象的第二检测框的第二属性进行修正,得到第四检测框,第四检测框用于指示所述第二点云地图中标注的动态对象和/或静态对象;在所述第二点云地图对应的单帧点云数据中,对标注动态对象的第四检测框的第三属性进行修正,得到第五检测框;以标注静态对象的第四检测框和所述第五检测框作为所述M帧点云数据的目标检测框。With reference to the second aspect, in a possible implementation manner, the correction unit is configured to: in the second point cloud map, correct the first attribute of the first detection frame marking the static object, and/or , modifying the second attribute of the second detection frame marked with the dynamic object to obtain a fourth detection frame, the fourth detection frame is used to indicate the dynamic object and/or the static object marked in the second point cloud map; in the In the single-frame point cloud data corresponding to the second point cloud map, the third attribute of the fourth detection frame marked with the dynamic object is corrected to obtain the fifth detection frame; the fourth detection frame marked with the static object and the first Five detection frames are used as target detection frames of the M frames of point cloud data.
结合第二方面,在一种可能的实现方式中,所述系统还包括训练单元,所述训练单元用于:将所述M帧点云数据中的单帧点云数据以及所述目标检测框作为第三训练数据,在基于所述第三训练数据训练多个检测模型的过程中确定第六检测框,所述第六检测框的检测结果为错误;将所述第二点云地图以及所述目标检测框作为第四训练数据,在基于所述第四训练数据训练多个检测模型的过程中确定第七检测框;基于所述第六检测框校正所述第七检测框。With reference to the second aspect, in a possible implementation manner, the system further includes a training unit configured to: use the single frame of point cloud data in the M frames of point cloud data and the target detection frame As the third training data, in the process of training a plurality of detection models based on the third training data, a sixth detection frame is determined, and the detection result of the sixth detection frame is an error; the second point cloud map and the The target detection frame is used as fourth training data, and a seventh detection frame is determined during the process of training multiple detection models based on the fourth training data; and the seventh detection frame is corrected based on the sixth detection frame.
第三方面,本申请实施例提供了一种点云数据处理系统,包括存储器和处理器,所述存储器用于存储程序;所述处理器用于执行所述存储器所存储的程序,以使所述装置实现如上述第一方面以及第一方面任一可能实现方式所述的方法。In a third aspect, the embodiment of the present application provides a point cloud data processing system, including a memory and a processor, the memory is used to store programs; the processor is used to execute the programs stored in the memory, so that the The device implements the method described in the foregoing first aspect and any possible implementation manner of the first aspect.
第四方面,本申请实施例提供了一种点云数据处理装置,包括:至少一个处理器和接口电路,所述接口电路用于为所述至少一个处理器提供数据或者代码指令,所述至少一个处理器用于通过逻辑电路或执行代码指令实现如上第一方面以及第一方面的任一可能实现方式所述的方法。In a fourth aspect, the embodiment of the present application provides a point cloud data processing device, including: at least one processor and an interface circuit, the interface circuit is used to provide data or code instructions for the at least one processor, and the at least one processor A processor is configured to implement the method described in the first aspect and any possible implementation manner of the first aspect by using a logic circuit or executing code instructions.
第五方面,本申请实施例提供了一种计算机可读存储介质,所述计算机可读介质存储有程序代码,当所述程序代码在计算机上运行时,使得计算机执行如上述第一方面以及第一方面任一可能实现方式所述的方法。In the fifth aspect, the embodiment of the present application provides a computer-readable storage medium, the computer-readable medium stores program code, and when the program code is run on the computer, the computer executes the above-mentioned first aspect and the first aspect. In one aspect, the method described in any possible implementation manner.
第六方面,本申请实施例提供了一种计算机程序产品,当所述计算机程序产品在计算机上运行时,使得所述计算机执行上述第一方面以及第一方面的任一可能实现方式所述的方法。In a sixth aspect, an embodiment of the present application provides a computer program product, which, when the computer program product is run on a computer, enables the computer to execute the above-mentioned first aspect and any possible implementation manner of the first aspect. method.
第七方面,本申请实施例提供了一种芯片系统,该芯片系统包括处理器,用于调用存储器中存储的计算机程序或计算机指令,以使得该处理器执行上述第一方面以及第一方面 的任一可能实现方式所述的方法。In a seventh aspect, an embodiment of the present application provides a chip system, the chip system includes a processor, configured to call a computer program or a computer instruction stored in a memory, so that the processor executes the above-mentioned first aspect and the first aspect The method described in any possible implementation.
结合第七方面,在一种可能的实现方式中,该处理器通过接口与存储器耦合。With reference to the seventh aspect, in a possible implementation manner, the processor is coupled to the memory through an interface.
结合第七方面,在一种可能的实现方式中,该芯片系统还包括存储器,该存储器中存储有计算机程序或计算机指令。With reference to the seventh aspect, in a possible implementation manner, the system on a chip further includes a memory, where computer programs or computer instructions are stored in the memory.
第八方面,本申请实施例提供了一种终端设备,该终端设备可用于实现上述第一方面以及第一方面任一可能实现方式所述的方法。示例地,该终端设备包括但不限于:智能运输设备(诸如汽车、轮船、无人机、火车、货车等)、智能制造设备(诸如机器人、工业设备、智能物流、智能工厂等)、智能终端(手机、计算机、平板电脑、掌上电脑、台式机、耳机、音响、穿戴设备、车载设备等)。In an eighth aspect, the embodiment of the present application provides a terminal device, which can be used to implement the method described in the foregoing first aspect and any possible implementation manner of the first aspect. Exemplarily, the terminal equipment includes but is not limited to: intelligent transportation equipment (such as automobiles, ships, drones, trains, trucks, etc.), intelligent manufacturing equipment (such as robots, industrial equipment, intelligent logistics, intelligent factories, etc.), intelligent terminal (Mobile phones, computers, tablets, PDAs, desktops, headsets, audio, wearable devices, car devices, etc.).
第九方面,本申请实施例提供了一种车辆,该车辆可用于实现如上述第一方面以及第一方面任一可能实现方式所述的方法。In a ninth aspect, an embodiment of the present application provides a vehicle, which can be used to implement the method described in the first aspect and any possible implementation manner of the first aspect.
第十方面,本申请实施例提供了一种服务器,该服务器可用于实现如上述第一方面以及第一方面任一可能实现方式所述的方法。In a tenth aspect, the embodiment of the present application provides a server, which can be used to implement the method described in the first aspect and any possible implementation manner of the first aspect.
本申请实施例在上述各方面提供的实现的基础上,还可以进行进一步组合以提供更多实现。On the basis of the implementations provided by the foregoing aspects, the embodiments of the present application may be further combined to provide more implementations.
上述第二方面至第十方面中任一方面中的任一可能实现方式可以达到的技术效果,可以相应参照上述第一方面中任一方面中的任一可能实现方式可以达到的技术效果描述,重复之处不予论述。The technical effects that can be achieved by any possible implementation of any one of the above-mentioned second to tenth aspects can be described with reference to the technical effects that can be achieved by any possible implementation of any of the above-mentioned first aspects, Duplication will not be discussed.
附图说明Description of drawings
图1为一种点云数据的示意图;Fig. 1 is the schematic diagram of a kind of point cloud data;
图2为一组检测框位置有偏差的点云数据的示意图;Fig. 2 is a schematic diagram of a group of point cloud data with deviations in detection frame positions;
图3为点云数据较少以及拖尾现象的示意图;Figure 3 is a schematic diagram of less point cloud data and trailing phenomenon;
图4为本申请实施例提供的点云数据处理方法适用的一种应用场景的示意图;FIG. 4 is a schematic diagram of an application scenario applicable to the point cloud data processing method provided by the embodiment of the present application;
图5为本申请实施例提供的点云数据处理系统的示意图;5 is a schematic diagram of a point cloud data processing system provided by an embodiment of the present application;
图6为本申请实施例提供的点云数据处理方法的一种流程示意图;FIG. 6 is a schematic flow chart of a point cloud data processing method provided in an embodiment of the present application;
图7为本申请实施例提供的第一检测框信息的示意图;FIG. 7 is a schematic diagram of the first detection frame information provided by the embodiment of the present application;
图8为本申请实施例提供的第二检测框信息的示意图;FIG. 8 is a schematic diagram of the second detection frame information provided by the embodiment of the present application;
图9为本申请实施例提供的获得第二点云地图的示意图;FIG. 9 is a schematic diagram of obtaining a second point cloud map provided by an embodiment of the present application;
图10为本申请实施例提供的点云数据处理方法的一种流程示意图;FIG. 10 is a schematic flow diagram of a point cloud data processing method provided in an embodiment of the present application;
图11为本申请实施例提供的点云数据处理系统的示意图;11 is a schematic diagram of a point cloud data processing system provided by an embodiment of the present application;
图12为本申请实施例提供的点云数据处理系统的示意图。FIG. 12 is a schematic diagram of a point cloud data processing system provided by an embodiment of the present application.
具体实施方式Detailed ways
为了方便理解本申请实施例,下面介绍与本申请实施例相关的术语:In order to facilitate the understanding of the embodiments of the present application, the terms related to the embodiments of the present application are introduced below:
1)点云,是通过测量设备对物体进行探测后得到的采样点。点云数据(point cloud data)是指在一个三维坐标系统中的一组向量的集合。1) The point cloud is the sampling point obtained after the object is detected by the measuring device. Point cloud data refers to a set of vectors in a three-dimensional coordinate system.
测量设备对一个物体的外观表面上探测并采样后得到的点云的集合可以称为点云集合。测量设备一次可以同时对多个物体进行探测并采样,得到的一组点云数据中可以包括 多个物体对应的点云集合。A collection of point clouds obtained after the measuring device detects and samples the appearance surface of an object may be referred to as a point cloud collection. The measuring device can detect and sample multiple objects at the same time, and the obtained set of point cloud data can include point cloud collections corresponding to multiple objects.
其中,点云数据除了具有几何位置以外,例如还可以包括颜色信息。例如,基于激光测量原理测量到的点云数据(也可以称为激光点云数据)可以包括三维坐标和激光反射强度(intensity)等信息。基于摄影测量原理得到的点云数据可以包括三维坐标和颜色等信息,其中,颜色信息可以为红绿蓝(red、green、blue,RGB)格式的颜色数据。结合激光测量原理和摄影测量原理得到点云数据则可以包括三维坐标、激光反射强度和颜色等信息。Wherein, besides the geometric position, the point cloud data may also include color information, for example. For example, point cloud data measured based on the principle of laser measurement (also referred to as laser point cloud data) may include information such as three-dimensional coordinates and laser reflection intensity (intensity). The point cloud data obtained based on the principle of photogrammetry may include information such as three-dimensional coordinates and color, wherein the color information may be color data in red, green, blue (RGB) format. Combining the principles of laser measurement and photogrammetry to obtain point cloud data can include information such as three-dimensional coordinates, laser reflection intensity, and color.
本申请实施例中以测量设备为激光雷达(lightlaser detection and ranging,Lidar)为例,激光雷达采集到的点云数据至少包括每个点云的三维坐标。In the embodiment of the present application, the measuring device is Lidar (lightlaser detection and ranging, Lidar) as an example, and the point cloud data collected by the Lidar includes at least the three-dimensional coordinates of each point cloud.
需要说明的是,本申请实施例中,一帧(或者称为单帧)点云数据是指在一个采样时刻采集到的一组点云数据,点云地图是指对连续采集到的多帧点云数据进行拼接得到的图像,可以是二维图像也可以为三维图像。It should be noted that, in the embodiment of the present application, one frame (or single frame) point cloud data refers to a group of point cloud data collected at one sampling moment, and the point cloud map refers to the continuous collection of multiple frames The image obtained by splicing point cloud data can be a two-dimensional image or a three-dimensional image.
2)训练集、验证集和测试集,是机器学习中用于进行模型训练的样本数据组成的不同的集合。2) The training set, verification set and test set are different sets of sample data used for model training in machine learning.
其中,训练集为用于进行模型拟合的数据样本组成的样本集。Wherein, the training set is a sample set composed of data samples used for model fitting.
验证集为模型训练过程中预留的样本集,用于调整模型的超参数和用于对模型的能力进行初步评估。在本申请实施例提供的点云数据处理方法中,验证集可以用于调整不同数据类型的点云数据在训练集中所占的比例,从而提升模型对不同数据类型的点云数据的拟合能力。The validation set is a set of samples reserved during the model training process, which is used to adjust the hyperparameters of the model and to conduct a preliminary evaluation of the model's capabilities. In the point cloud data processing method provided in the embodiment of the present application, the verification set can be used to adjust the proportion of point cloud data of different data types in the training set, thereby improving the model's ability to fit point cloud data of different data types .
测试集为用于评估模型最终的泛化能力的样本集。The test set is a sample set used to evaluate the final generalization ability of the model.
本申请实施例中,用于模型训练的样本数据可以称为训练数据,所述训练数据可以包括训练集、验证集和测试集,训练集、验证集和测试集中的数据通常不会重叠。一般在构建模型之前,可以按照预设的比例对样本数据进行划分,例如按照7:1:1的比例划分得到训练集、验证集和测试集,并基于该训练集、验证集和测试集训练目标检测模型(例如下文中述及的第一检测模型和/或第二检测模型),本申请实施例对模型训练的具体实现方式以及训练过程不做限定。In the embodiment of the present application, the sample data used for model training may be referred to as training data, and the training data may include a training set, a verification set, and a test set, and data in the training set, verification set, and test set generally do not overlap. Generally, before building the model, the sample data can be divided according to the preset ratio, for example, according to the ratio of 7:1:1 to obtain the training set, verification set and test set, and based on the training set, verification set and test set training For the target detection model (such as the first detection model and/or the second detection model mentioned below), the embodiment of the present application does not limit the specific implementation manner and training process of the model training.
3)目标检测,是图像处理和计算机视觉学科的重要分支,也是智能监控系统的核心部分。目标检测可以对点云数据进行目标检测和识别,从而确定点云数据中包括的多个点云集合分别对应的物体的类型。3) Target detection is an important branch of image processing and computer vision disciplines, and is also the core part of an intelligent monitoring system. Target detection can perform target detection and recognition on the point cloud data, so as to determine the types of objects corresponding to the multiple point cloud sets included in the point cloud data.
例如,目标检测应用在驾驶场景时,车辆上的目标检测装置可以对激光雷达采集到的点云数据进行目标检测,从而在车辆行驶过程中,识别车道上的车辆以及路边的树木、行人等,辅助车辆实现路线规划、躲避障碍等功能,进而实现智能驾驶。For example, when the target detection is applied in the driving scene, the target detection device on the vehicle can perform target detection on the point cloud data collected by the lidar, so as to recognize the vehicles on the lane and the trees and pedestrians on the side of the road during the driving process of the vehicle. , assisting vehicles to realize route planning, obstacle avoidance and other functions, and then realize intelligent driving.
本申请实施例中可以包括两类目标检测模型,表示为第一检测模型和第二检测模型。其中,所述第一检测模型可以点云地图和静态对象标注数据作为训练数据进行模型训练得到,所述第二检测模型可以点云数据(包括单帧点云数据和/或若干帧点云数据)和动态对象标注数据作为训练数据进行模型训练得到,所述第一检测模型和所述第二检测模型可分别用于对待标注的点云地图、点云数据进行目标检测,从而确定点云地图、点云数据中包括的多个点云集合分别对应的对象的类型。The embodiment of the present application may include two types of target detection models, denoted as a first detection model and a second detection model. Wherein, the first detection model can be obtained by performing model training on point cloud maps and static object annotation data as training data, and the second detection model can be obtained by point cloud data (including single frame point cloud data and/or several frames of point cloud data ) and dynamic object labeling data are used as training data to carry out model training, and the first detection model and the second detection model can be used to detect the point cloud map and point cloud data to be marked respectively, so as to determine the point cloud map , types of objects respectively corresponding to the multiple point cloud sets included in the point cloud data.
可以理解的是,本申请实施例中,经由所述第一检测模型或第二检测模型处理的数据和检测结果也可以作为训练数据进一步训练所述第一检测模型和第二检测模型,以提升模型的精度,本申请实施例对此不做限定。It can be understood that in the embodiment of the present application, the data and detection results processed by the first detection model or the second detection model can also be used as training data to further train the first detection model and the second detection model to improve The accuracy of the model is not limited in this embodiment of the present application.
4)静态对象和动态对象,是包含于点云数据或者点云地图的目标对象,静态对象一般为对静态物体进行探测采集到的点云集合,动态对象一般为对动态物体进行探测采集到的点云集合。4) Static objects and dynamic objects are target objects contained in point cloud data or point cloud maps. Static objects are generally point cloud collections collected by detecting static objects, and dynamic objects are generally collected by detecting dynamic objects. Collection of point clouds.
示例地,在车辆驾驶场景中,静态物体可以包括但不限于道路、道路侧静态放置的树木、楼房、路灯、道路标志、停留车辆等,动态物体可以包括但不限于道路上行驶的车辆、行人、动物等。可以理解的是,本申请实施例中,对象的动态或静态的区分只在于该对象是否相对于参考物体移动,例如车辆在行驶状态下为动态物体,对该车辆进行探测得到的点云集合称为动态对象,车辆在非行驶状态下为可以认为是静态物体,对该车辆进行探测得到的点云集合称为静态对象。For example, in a vehicle driving scene, static objects may include but not limited to roads, trees statically placed on the side of the road, buildings, street lights, road signs, parked vehicles, etc. Dynamic objects may include but not limited to vehicles driving on the road, pedestrians , animals, etc. It can be understood that, in the embodiment of the present application, the dynamic or static distinction of an object only depends on whether the object moves relative to the reference object. For example, a vehicle is a dynamic object in a driving state, and the point cloud set obtained by detecting the vehicle is called As a dynamic object, the vehicle can be considered as a static object in the non-driving state, and the point cloud collection obtained by detecting the vehicle is called a static object.
5)拖尾现象,也可称为拖影现象,是针对同一动态物体在不同时刻采集到的点云集合进行拼接时,形成的相同对象罗列的现象。其中,由于动态物体的可移动性,动态物体在不同时刻的位置是不同的,故而为消除动态物体的移动性影响,在针对同一动态物体在不同时刻采集到的点云集合进行拼接时,会形成相同对象罗列的现象,即拖尾现象。5) The smearing phenomenon, also known as the smearing phenomenon, is a phenomenon in which the same objects are listed when splicing point cloud collections collected at different times for the same dynamic object. Among them, due to the mobility of dynamic objects, the positions of dynamic objects at different times are different. Therefore, in order to eliminate the impact of the mobility of dynamic objects, when splicing point cloud collections collected at different times for the same dynamic object, it will be The phenomenon that the same objects are listed is formed, that is, the tailing phenomenon.
下面将结合附图对本申请实施例作进一步地描述说明。The embodiments of the present application will be further described below in conjunction with the accompanying drawings.
随着目标检测在驾驶等领域的广泛应用,基于深度学习的目标检测方法凭借其准确性和高效率成为目标检测中的主流方法。在实施中,基于深度学习构建目标检测模型并进行训练,将待检测的点云数据输入到训练后的目标检测模型中,可以获取目标检测模型输出的目标检测结果,例如目标检测模型可以输出多个物体的类型以及每个物体的位置信息。With the wide application of object detection in driving and other fields, the object detection method based on deep learning has become the mainstream method in object detection due to its accuracy and high efficiency. In the implementation, the target detection model is built and trained based on deep learning, and the point cloud data to be detected is input into the trained target detection model, and the target detection results output by the target detection model can be obtained. For example, the target detection model can output multiple The type of each object and the location information of each object.
本申请实施例中,点云数据例如可以为激光雷达采集得到的。例如,激光雷达向目标物体发射探测信号(例如探测信号可以为激光束),在激光雷达接收到的从目标物体反射回来的反射信号之后,激光雷达将探测信号与发射信号进行比较并进行处理得到点云集合,点云集合是在获取目标物体表面每个采样点的空间坐标后得到的采样点集合。激光雷达可以同时对多个目标物体进行探测并采样,得到一组点云数据,点云数据可以包括多个目标物体对应的多个点云集合。图1为一种点云数据示意图,其中,图中的每个黑点代表一个点云,每个点云对应一组三维坐标。一组点云数据中包括多个点云集合,例如图1中检测框A、检测框B和检测框C中每个检测框中的点云构成一个点云集合,每个点云集合对应一个目标物体。In the embodiment of the present application, the point cloud data may be collected by lidar, for example. For example, the laser radar transmits a detection signal to the target object (for example, the detection signal can be a laser beam), and after the laser radar receives the reflection signal reflected from the target object, the laser radar compares the detection signal with the transmitted signal and processes it to obtain The point cloud set, the point cloud set is a set of sampling points obtained after obtaining the spatial coordinates of each sampling point on the surface of the target object. Lidar can detect and sample multiple target objects at the same time to obtain a set of point cloud data. The point cloud data can include multiple point cloud sets corresponding to multiple target objects. Fig. 1 is a schematic diagram of point cloud data, wherein each black dot in the figure represents a point cloud, and each point cloud corresponds to a set of three-dimensional coordinates. A set of point cloud data includes multiple point cloud collections. For example, the point clouds in each detection frame in detection frame A, detection frame B, and detection frame C in Figure 1 constitute a point cloud collection, and each point cloud collection corresponds to a target object.
由于训练目标检测模型时需要将目标物体的类型以及目标物体对应的点云集合作为训练样本,以使目标检测模型对目标物体的类型以及点云集合的对应关系进行学习,因此,使用的点云集合的准确性会对目标检测模型的性能产生较大影响。通常一组点云数据中包括多个点云集合,因此,在对目标检测模型训练之前,需要对训练使用的点云数据标注检测框,以将点云数据中属于同一个物体的点云划分为一个点云集合。例如图1中的检测框A、检测框B和检测框C将点云数据划分为3个点云集合,每个点云集合对应一个目标物体。Since the target object type and the point cloud set corresponding to the target object need to be used as training samples when training the target detection model, so that the target detection model can learn the corresponding relationship between the target object type and the point cloud set, the point cloud used The accuracy of the ensemble can have a large impact on the performance of the object detection model. Usually a set of point cloud data includes multiple point cloud sets. Therefore, before training the target detection model, it is necessary to mark the detection frame on the point cloud data used for training, so as to divide the point cloud belonging to the same object in the point cloud data. for a point cloud collection. For example, the detection frame A, detection frame B, and detection frame C in Figure 1 divide the point cloud data into three point cloud sets, and each point cloud set corresponds to a target object.
在对目标检测模型训练时,检测框位置准确才能保证根据检测框位置提取到的点云集合准确,当检测框位置存在偏差时,会导致提取到的点云集合不够准确。例如,图2为一组检测框位置有偏差的点云数据的示意图,图2中检测框E和检测框F的位置存在偏差,根据检测框E和检测框F对点云数据进行划分后,不能准确得到两个目标物体对应的点云集合,则目标检测模型在训练中会对错误的点云集合进行学习,进而影响目标检测模型的 性能。针对上述问题,目前的处理方式为人工进行调整校验,并将人工校验后的检测框位置以及点云数据作为目标检测模型的训练数据。可见,目前的点云数据处理方法效率较低且准确率难以保证。When training the target detection model, the accurate position of the detection frame can ensure the accuracy of the point cloud set extracted according to the position of the detection frame. When there is a deviation in the position of the detection frame, the extracted point cloud set will not be accurate enough. For example, Figure 2 is a schematic diagram of a group of point cloud data with deviations in the detection frame positions. In Figure 2, there is a deviation in the positions of the detection frame E and the detection frame F. After dividing the point cloud data according to the detection frame E and detection frame F, If the point cloud sets corresponding to the two target objects cannot be accurately obtained, the target detection model will learn the wrong point cloud set during training, which will affect the performance of the target detection model. In view of the above problems, the current processing method is manual adjustment and verification, and the detection frame position and point cloud data after manual verification are used as the training data of the target detection model. It can be seen that the current point cloud data processing methods are inefficient and difficult to guarantee accuracy.
基于上述问题,为了降低对点云数据进行目标检测的人工成本并提升检测精度,在一些设计中提出一种三维(3Dimensions,3D)点云物体自动标注系统,该系统以单帧点云数据或若干帧点云数据作为输入,通过使用3D检测模型进行自动标注。Based on the above problems, in order to reduce the labor cost of object detection on point cloud data and improve the detection accuracy, a three-dimensional (3Dimensions, 3D) point cloud object automatic labeling system is proposed in some designs. The system uses single frame point cloud data or Several frames of point cloud data are used as input, which are automatically marked by using the 3D detection model.
上述设计虽然可以在一定程度上提升检测效率和准确率,但是当目标物体受遮挡或距离过远时,由于所述单帧点云数据或若干帧点云数据中的点云数据过少会导致误检、漏检、检测框不准确等问题,并且,动态物体在不同时刻采集的点云数据中的偏移对点云数据处理过程中引发的拖尾现象等,会严重影响该系统的自动标注效果。如图3所示,矩形框G标注的为动态物体(例如行驶的车辆的车尾)对应的点云集合,该矩形框G中包括的点云数量较少,导致无法准确识别该动态物体的类型,同时,在将不同时刻帧的点云数据转换至统一坐标系后进行拼接,该动态物体会产生拖影引起拖尾现象,如图3所示,矩形框G中包括的点云集合在不同时刻的罗列。因此,如何提升3D点云物体自动标注系统的标注效果,仍为亟需解决的重要问题。Although the above design can improve the detection efficiency and accuracy to a certain extent, when the target object is blocked or the distance is too far, due to the lack of point cloud data in the single frame point cloud data or several frames of point cloud data, it will cause False detection, missed detection, inaccurate detection frame, etc., and the offset of dynamic objects in the point cloud data collected at different times will seriously affect the system's automatic Labeling effect. As shown in Figure 3, the rectangular frame G marks the point cloud collection corresponding to the dynamic object (such as the rear of a driving vehicle). The number of point clouds included in the rectangular frame G is small, which makes it impossible to accurately identify the dynamic object. At the same time, after converting the point cloud data of different frames at different times into a unified coordinate system for splicing, the dynamic object will produce smears and cause smearing. As shown in Figure 3, the point cloud set included in the rectangular frame G is in List of different moments. Therefore, how to improve the labeling effect of the 3D point cloud object automatic labeling system is still an important problem that needs to be solved urgently.
本申请实施例提供一种点云数据处理方法及系统,可用于对单帧点云数据或若干帧点云数据的检测框进行校正,从而提升3D点云物体自动标注系统的标注效果。The embodiment of the present application provides a point cloud data processing method and system, which can be used to correct the detection frames of a single frame of point cloud data or several frames of point cloud data, thereby improving the labeling effect of the 3D point cloud object automatic labeling system.
图4为本申请实施例提供的点云数据处理方法适用的一种应用场景示意图。FIG. 4 is a schematic diagram of an application scenario where the point cloud data processing method provided in the embodiment of the present application is applicable.
如图4所示,本申请实施例中的多帧点云数据可以为雷达(例如激光雷达)采集到的数据,该雷达可以位于车辆上。例如,图4中示出的车辆41上可以安装有雷达,该雷达可以将采集到的点云数据发送给车辆41,车辆41中可以部署有点云数据处理系统来对采集到的点云数据执行本申请实施例提供的点云数据处理方法。或者,所述车辆41可以将采集到的点云数据发送给独立部署的点云数据处理系统,由该独立部署的点云数据处理系统执行本申请实施例提供的点云数据处理方法。As shown in FIG. 4 , the multi-frame point cloud data in the embodiment of the present application may be data collected by a radar (such as a lidar), and the radar may be located on a vehicle. For example, a radar may be installed on the vehicle 41 shown in FIG. The point cloud data processing method provided by the embodiment of this application. Alternatively, the vehicle 41 may send the collected point cloud data to an independently deployed point cloud data processing system, and the independently deployed point cloud data processing system executes the point cloud data processing method provided in the embodiment of the present application.
其中,独立部署的点云数据处理系统可以为服务器42,服务器42可以对获取到的点云数据执行本申请实施例提供的点云数据处理方法。或者,该独立部署的点云数据处理系统可以为终端设备,例如图4中示出的移动终端43,由终端设备对采集到的点云数据执行本申请实施例提供的点云数据处理方法。Wherein, the independently deployed point cloud data processing system may be the server 42, and the server 42 may execute the point cloud data processing method provided by the embodiment of the present application on the acquired point cloud data. Alternatively, the independently deployed point cloud data processing system may be a terminal device, such as the mobile terminal 43 shown in FIG.
示例地,图5为本申请实施例提供的点云数据处理系统的示意图,如图5所示,该系统500中可以包括获取单元510、处理单元520、校正单元530、训练单元540和输出单元550。Exemplarily, FIG. 5 is a schematic diagram of a point cloud data processing system provided in an embodiment of the present application. As shown in FIG. 5 , the system 500 may include an acquisition unit 510, a processing unit 520, a correction unit 530, a training unit 540 and an output unit 550.
其中,所述获取单元510可用于获取多帧点云数据(例如M帧,M为大于或等于2的整数),并将所述多帧点云数据提供给所述处理单元520。所述处理单元520可以从所述训练单元540获取第一检测模型和/或第二检测模型,并根据所述第一检测模型和/或所述第二检测模型对所述多帧点云数据进行处理,得到检测结果,该检测结果可以用于指示在所述多帧点云数据中的单帧点云数据中标注的静态对象和/或动态对象。所述处理单元520可以将检测结果提供给校正单元530,该校正单元530可以对该检测结果进行校正得到目标检测框,该目标检测框可以作为所述多帧点云数据的目标检测结果经由所述输出单元550输出。在一种可能的实现方式中,所述多帧点云数据和所述多帧点云数据的目标检测框信息可以被提供给训练单元540,以供所述训练单元进行模型训练,以提升模型精度。Wherein, the obtaining unit 510 may be configured to obtain multi-frame point cloud data (for example, M frames, M is an integer greater than or equal to 2), and provide the multi-frame point cloud data to the processing unit 520 . The processing unit 520 can obtain the first detection model and/or the second detection model from the training unit 540, and perform the multi-frame point cloud data processing according to the first detection model and/or the second detection model Processing is performed to obtain a detection result, which can be used to indicate the static object and/or dynamic object marked in the single frame of point cloud data in the multi-frame point cloud data. The processing unit 520 can provide the detection result to the correction unit 530, and the correction unit 530 can correct the detection result to obtain a target detection frame, which can be used as the target detection result of the multi-frame point cloud data via the output from the output unit 550. In a possible implementation manner, the multi-frame point cloud data and the target detection frame information of the multi-frame point cloud data can be provided to the training unit 540 for the training unit to perform model training to improve the model precision.
可以理解的是,上述单元模块仅是对该点云数据处理系统500的功能划分,并不限定该点云数据处理系统500的功能。在其它实施例中,该点云数据处理系统500还可以包括其它单元,该点云数据处理系统500中的单元模块也可以进一步划分获取具有其它命名方式,本申请实施例对此不做限定。例如,获取单元510具体可以包括第一获取单元和第二获取单元,处理单元520可以包括第一处理单元和第二处理单元,在此不再赘述。It can be understood that the above unit modules are only the functional division of the point cloud data processing system 500 , and do not limit the functions of the point cloud data processing system 500 . In other embodiments, the point cloud data processing system 500 may also include other units, and the unit modules in the point cloud data processing system 500 may also be further divided into other naming methods, which is not limited in this embodiment of the present application. For example, the acquiring unit 510 may specifically include a first acquiring unit and a second acquiring unit, and the processing unit 520 may include a first processing unit and a second processing unit, which will not be repeated here.
下面介绍本申请实施例的点云数据处理方法。The point cloud data processing method of the embodiment of the present application is introduced below.
图6为本申请实施例提供的点云数据处理方法的一种流程示意图,其中,该方法可由图5中的点云数据处理系统500及其功能模块实现,如图6所示,该点云数据处理方法可以包括以下步骤:Fig. 6 is a schematic flow chart of the point cloud data processing method provided by the embodiment of the present application, wherein the method can be realized by the point cloud data processing system 500 in Fig. 5 and its functional modules, as shown in Fig. 6 , the point cloud The data processing method may include the following steps:
S610:点云数据处理系统获取第一检测框。S610: The point cloud data processing system acquires the first detection frame.
本申请实施例中,该第一检测框用于指示在第一点云地图中标注的静态对象,该第一点云地图可由M帧点云数据得到,M为大于或等于2的整数。其中,所述M帧点云数据可以是连续帧的点云数据,该M帧点云数据可由数据采集装置采集得到并由该点云数据处理系统的获取单元获得,所述数据采集装置例如可以为车辆41上的雷达,如图4所示。在一个示例中,为了便于区分,所述M帧点云数据也可以称为点云序列。In the embodiment of the present application, the first detection frame is used to indicate the static object marked in the first point cloud map, and the first point cloud map can be obtained from M frames of point cloud data, where M is an integer greater than or equal to 2. Wherein, the M frames of point cloud data can be point cloud data of continuous frames, and the M frames of point cloud data can be collected by a data acquisition device and obtained by an acquisition unit of the point cloud data processing system, and the data acquisition device can for example is the radar on the vehicle 41 , as shown in FIG. 4 . In an example, for the convenience of distinction, the M frames of point cloud data may also be referred to as a point cloud sequence.
示例地,所述第一点云地图可以是对所述M帧点云数据进行拼接得到的点云地图。具体实施时,点云数据处理系统可以根据将所述M帧点云数据中所有单帧点云数据变换至统一坐标系(例如世界坐标系),并将所述M帧点云数据中所有单帧点云数据中的点云集合在该坐标系中拼接成为点云地图。其中,在统一坐标系中将多帧点云数据的点云集合拼接为点云地图的过程中,可以是在所述坐标系中基于所述多帧点云数据的实际采集时刻,依次将不同帧点云数据进行拼接;或者,也可以是选择参考物(例如车辆定位姿态),并基于所述参考物,在所述坐标系中对该M帧点云数据中的点云集合进行调整后,按照采集时刻依次将不同帧点云数据中的点云集合进行拼接,本申请实施例对该拼接处理过程的具体实现方式不做限定。For example, the first point cloud map may be a point cloud map obtained by splicing the M frames of point cloud data. During specific implementation, the point cloud data processing system can convert all single-frame point cloud data in the M frame point cloud data to a unified coordinate system (such as the world coordinate system), and convert all single-frame point cloud data in the M frame point cloud data The point cloud collection in the frame point cloud data is spliced into a point cloud map in this coordinate system. Wherein, in the process of splicing the point cloud collection of multi-frame point cloud data into a point cloud map in the unified coordinate system, different The frame point cloud data is spliced; or, it can also be to select a reference object (such as a vehicle positioning attitude), and based on the reference object, after adjusting the point cloud collection in the M frame point cloud data in the coordinate system , sequentially splicing the point cloud sets in different frames of point cloud data according to the collection time, and the embodiment of the present application does not limit the specific implementation manner of the splicing process.
一种可能的实现方式中,所述第一检测框为对所述第一点云地图进行目标检测得到的检测结果,所述第一检测框可以用于指示在所述第一点云地图中标注的静态对象。In a possible implementation manner, the first detection frame is a detection result obtained by performing target detection on the first point cloud map, and the first detection frame may be used to indicate that in the first point cloud map Annotated static object.
示例地,对第一点云地图的目标检测过程可由第一检测模型实现。其中,该第一检测模型可以是前文中述及的训练单元540使用第一训练数据进行训练得到的,所述第一训练数据为已标注的训练数据,该第一训练数据可以包括第三点云地图与静态对象标注数据。Exemplarily, the target detection process on the first point cloud map can be realized by the first detection model. Wherein, the first detection model can be obtained by the training unit 540 mentioned above using the first training data for training, the first training data is marked training data, and the first training data can include the third point Cloud map with static object annotation data.
需要说明的是,本申请实施例中,在准备所述第一训练数据时,由于在所述第三点云地图上的静态对象和动态对象有较明显的形态差异,训练单元540可以仅保留静态对象标注框作为正样本,即静态对象标注数据,并利用所述第三点云地图以及所述静态对象标注数据进行训练得到所述第一检测模型。可以理解的是,本申请实施例中,该第三点云地图可以与所述第一点云地图相同或不同,即M帧点云数据或由该M帧点云数据进行拼接得到的第一点云地图既可以作为待标注数据也可以作为训练数据,本申请对此不做限定。It should be noted that, in the embodiment of the present application, when preparing the first training data, since there are obvious morphological differences between static objects and dynamic objects on the third point cloud map, the training unit 540 may only retain The static object labeling frame is used as a positive sample, that is, static object labeling data, and the first detection model is obtained by using the third point cloud map and the static object labeling data for training. It can be understood that, in the embodiment of the present application, the third point cloud map may be the same as or different from the first point cloud map, that is, M frames of point cloud data or the first point cloud data spliced by the M frames of point cloud data. The point cloud map can be used as data to be labeled or as training data, which is not limited in this application.
实施S610的过程中,所述点云数据处理系统可以直接获取所述第一点云地图以及用于描述所述第一点云地图的相关信息。或者,该点云数据处理系统也可以获取所述M帧点云数据,并对所述M帧点云数据进行拼接,得到所述第一点云地图,本申请实施例对该第一点云地图的获取方式不做限定。进一步,该点云数据处理系统可以根据所述第一检测模 型对所述第一点云地图进行目标检测,获得所述第一检测框。其中,该第一检测框可以用于标注所述第一点云地图中的静态对象,该第一检测框的属性信息可以用于描述该第一检测框所标注的静态对象。In the process of implementing S610, the point cloud data processing system may directly acquire the first point cloud map and related information used to describe the first point cloud map. Alternatively, the point cloud data processing system may also obtain the M frames of point cloud data, and splicing the M frames of point cloud data to obtain the first point cloud map. In the embodiment of the present application, the first point cloud The method of obtaining the map is not limited. Further, the point cloud data processing system can perform target detection on the first point cloud map according to the first detection model, and obtain the first detection frame. Wherein, the first detection frame may be used to label the static object in the first point cloud map, and the attribute information of the first detection frame may be used to describe the static object marked by the first detection frame.
示例地,如图7所示,在(x,y,z)三维空间中,该第一检测框例如可以包括检测框1、检测框2和检测框3,所述第一检测框的属性信息例如可以包括检测框1/检测框2/检测框3的位置(例如检测框1/检测框2/检测框3在所述第一点云地图中的位置、或与其它检测框的相对位置等)、尺寸(例如长度、宽度、半径、直径等)、形状(例如圆柱体、椎体、正方体、长方体、不规则形状等)等信息,基于第一检测框的属性信息即可检测得到该第一检测框所标注的静态对象的位置、尺寸、形状等。需要说明的是,此处仅示例性地说明第一检测框以及所述第一检测框的属性信息,实际需要根据对点云数据中与目标对象对应的点云集合的检测和识别确定,在此不再赘述。For example, as shown in FIG. 7 , in the (x, y, z) three-dimensional space, the first detection frame may include, for example, detection frame 1, detection frame 2, and detection frame 3, and the attribute information of the first detection frame For example, it may include the position of detection frame 1/detection frame 2/detection frame 3 (such as the position of detection frame 1/detection frame 2/detection frame 3 in the first point cloud map, or the relative position with other detection frames, etc. ), size (such as length, width, radius, diameter, etc.), shape (such as cylinder, cone, cube, cuboid, irregular shape, etc.), based on the attribute information of the first detection frame, the first detection frame can be detected. The position, size, shape, etc. of the static object marked by a detection frame. It should be noted that the first detection frame and the attribute information of the first detection frame are only exemplarily described here. Actually, it needs to be determined according to the detection and identification of the point cloud set corresponding to the target object in the point cloud data. This will not be repeated here.
S620:点云数据处理系统获取第二检测框。S620: The point cloud data processing system acquires the second detection frame.
本申请实施例中,该第二检测框用于指示在第一点云数据中标注的动态对象,该第一点云数据由N帧点云数据得到,N为大于或等于1的整数,N<M。该N帧点云数据为该M帧点云数据的部分帧点云数据,其中,在N=1时,该N帧点云数据为单帧点云数据,N>1时,该N帧点云数据包括若干帧点云数据。在具体实施时,该点云数据处理系统按照配置信息或者应用场景中确定该N或M的具体取值,本申请实施例对此具体实现方式不做限定。其中,在所述部分帧点云数据为若干帧的点云数据的情况下,该第一点云数据可为这些帧的点云数据进行拼接处理得到,所述第一点云数据只包含第一点云地图的部分信息,也可以称为局部点云地图。In the embodiment of the present application, the second detection frame is used to indicate the dynamic object marked in the first point cloud data, the first point cloud data is obtained from N frames of point cloud data, N is an integer greater than or equal to 1, and N <M. The N frames of point cloud data are partial frames of point cloud data of the M frames of point cloud data, wherein, when N=1, the N frames of point cloud data are single frame point cloud data, and when N>1, the N frames of point cloud data Cloud data includes several frames of point cloud data. During specific implementation, the point cloud data processing system determines the specific value of N or M according to the configuration information or the application scenario, and the embodiment of the present application does not limit the specific implementation manner. Wherein, in the case where the part of the frame point cloud data is point cloud data of several frames, the first point cloud data can be obtained by splicing the point cloud data of these frames, and the first point cloud data only includes the first Partial information of a point cloud map can also be called a local point cloud map.
一种可能的实现方式中,该第二检测框可以为对所述第一点云数据进行目标检测得到的检测结果,所述第二检测框可以用于指示所述第一点云数据中标注的动态对象。In a possible implementation manner, the second detection frame may be a detection result obtained by performing target detection on the first point cloud data, and the second detection frame may be used to indicate that the label in the first point cloud data of dynamic objects.
示例地,对该第一点云数据的目标检测过程可由第二检测模型实现。其中,该第二检测模型可以是前文述及的训练单元540使用第二训练数据进行训练得到的,所述第二训练数据为已标注的训练数据,该第二训练数据可以包括第二点云数据和动态对象标注数据。Exemplarily, the target detection process of the first point cloud data can be realized by the second detection model. Wherein, the second detection model can be obtained by training the aforementioned training unit 540 using the second training data, the second training data is labeled training data, and the second training data can include the second point cloud Data and dynamic objects label data.
需要说明的是,本申请实施例中,在准备第二训练数据时,训练数据可以仅保留动态对象标注框作为正样本,即动态对象标注数据,并利用所述第二点云数据和所述动态对象标注数据进行训练得到所述第二检测模型。可以理解的,本申请实施例中,该第二点云数据可以与所述第一点云数据相同或不同,即第一点云数据既可以作为待标注数据也可以作为训练数据,本申请对此不做限定。It should be noted that, in the embodiment of the present application, when preparing the second training data, the training data may only retain the dynamic object labeling frame as a positive sample, that is, the dynamic object labeling data, and use the second point cloud data and the The dynamic object tag data is trained to obtain the second detection model. It can be understood that in the embodiment of the present application, the second point cloud data can be the same as or different from the first point cloud data, that is, the first point cloud data can be used as data to be labeled or as training data. This is not limited.
一种可能的实现方式中,由于动态对象的移动特性,对动态对象的标注过程的实现难度大于对静态对象的标注过程的实现难度,因此训练得到的所述第二检测模型可能存在一定检测误差,利用该第二检测模型进行目标检测时,可能仍会导致检测结果中包括标注静态对象的检测框。本申请实施例中,为了减少检测误差,实施S620过程中,所述点云数据处理系统可以根据所述第二检测模型对所述第一点云数据进行目标检测,获得第三检测框,根据所述第一检测框对所述第三检测框进行校正,得到所述第二检测框。其中,所述第三检测框用于指示在所述第一点云数据中标注的动态对象,通过该校正过程,可以利用对第一点云地图中的静态对象的检测结果融合对所述第一点云数据中的静态对象的误检结果,以确保所得到的第二检测框仅用于指示所述第一点云数据中标注的动态对象。其中,该第二检测框可以用于标注所述第一点云数据中的动态对象,该第二检测框的属性信息可 以用于描述该第二检测框所标注的动态对象。In a possible implementation, due to the moving characteristics of dynamic objects, the difficulty of implementing the labeling process for dynamic objects is greater than that for static objects, so the second detection model obtained through training may have a certain detection error , when the second detection model is used for target detection, the detection result may still include a detection frame marked with a static object. In the embodiment of the present application, in order to reduce the detection error, during the implementation of S620, the point cloud data processing system can perform target detection on the first point cloud data according to the second detection model to obtain a third detection frame, according to The first detection frame corrects the third detection frame to obtain the second detection frame. Wherein, the third detection frame is used to indicate the dynamic object marked in the first point cloud data. Through this correction process, the detection result of the static object in the first point cloud map can be fused to the first point cloud map. The false detection result of the static object in the point cloud data, so as to ensure that the obtained second detection frame is only used to indicate the dynamic object marked in the first point cloud data. Wherein, the second detection frame can be used to label the dynamic object in the first point cloud data, and the attribute information of the second detection frame can be used to describe the dynamic object marked by the second detection frame.
如图8所示,在(x,y,z)三维空间中,该第三检测框例如可以包括检测框4、检测框5、检测框6,所述第三检测框的属性信息例如可以包括检测框4/检测框5/检测框6的位置、尺寸、形状等信息,详细可参见上文中结合图7的相关描述,在此不再赘述。其中,检测框4、检测框5用于指示动态对象(例如对车辆等进行探测采集到的点云集合)、检测框6(使用虚线框表示,用于区分于指示动态对象的检测框4和检测框5)用于指示由于检测误差标注出的静态对象(例如对树木、道路标志等进行探测采集到的点云集合)。由于该检测框6也包含于对第一点云地图进行目标检测得到的第一检测框,点云数据处理系统根据所述第一检测框对第三检测框进行校正,例如可以是从第三检测框中去除用于指示静态对象的检测框,以经过处理后剩余的第三检测框作为所述第二检测框。如图8所示,去除检测框6后,剩余的检测框4和检测框5即为第二检测框,检测框4和检测框5的属性信息即为第二检测框的属性信息。As shown in Figure 8, in the (x, y, z) three-dimensional space, the third detection frame may include, for example, detection frame 4, detection frame 5, and detection frame 6, and the attribute information of the third detection frame may include, for example For information such as the position, size, and shape of the detection frame 4/detection frame 5/detection frame 6, please refer to the relevant description above in conjunction with FIG. 7 for details, and will not be repeated here. Among them, the detection frame 4 and the detection frame 5 are used to indicate the dynamic object (for example, the point cloud collection collected by detecting the vehicle, etc.), the detection frame 6 (represented by a dotted line frame, and used to distinguish from the detection frame 4 and the detection frame indicating the dynamic object. The detection frame 5) is used to indicate the static object marked due to the detection error (for example, the point cloud collection collected by detecting trees, road signs, etc.). Since the detection frame 6 is also included in the first detection frame obtained by performing target detection on the first point cloud map, the point cloud data processing system corrects the third detection frame according to the first detection frame, for example, from the third detection frame A detection frame indicating a static object is removed from the detection frame, and a third detection frame remaining after processing is used as the second detection frame. As shown in FIG. 8 , after the detection frame 6 is removed, the remaining detection frame 4 and detection frame 5 are the second detection frame, and the attribute information of the detection frame 4 and the detection frame 5 is the attribute information of the second detection frame.
可以理解的是,本申请实施例中,由于动态对象与静态对象不同,相应地,用于标注静态对象的第一检测框的属性信息,与用于标注动态对象的第二检测框的属性信息可以相同或不同,本申请实施例对此不做限定。It can be understood that, in the embodiment of the present application, since the dynamic object is different from the static object, correspondingly, the attribute information of the first detection frame used to mark the static object is different from the attribute information of the second detection frame used to mark the dynamic object They may be the same or different, which is not limited in this embodiment of the present application.
S630:点云数据处理系统根据所述第二检测框,对所述M帧点云数据中的单帧点云数据进行融合处理,得到第二点云地图。S630: The point cloud data processing system performs fusion processing on the single frame of point cloud data in the M frames of point cloud data according to the second detection frame, to obtain a second point cloud map.
本申请实施例中,该第二检测框也可以称为动态对象检测结果,该动态对象检测结果可用于指示不同的动态对象。In this embodiment of the present application, the second detection frame may also be referred to as a dynamic object detection result, and the dynamic object detection result may be used to indicate different dynamic objects.
本申请实施例中,由于动态对象对应的静态物体的可移动特性,因此,在所述M帧点云数据中,单帧点云数据可能仅包括针对同一动态物体的局部采集到的点云集合而非对该动态物体整体采集到的点云集合,这会导致针对单帧点云数据中的点或若干帧点云数据中的点过少(称为稀疏点云),同时,对于不同帧点云数据进行拼接处理还可能会存在由动态对象拼接引发的拖尾现象。In the embodiment of the present application, due to the movable characteristics of the static objects corresponding to the dynamic objects, in the M frames of point cloud data, the single frame of point cloud data may only include locally collected point cloud sets for the same dynamic object Rather than the point cloud collection collected as a whole for the dynamic object, this will result in too few points in a single frame of point cloud data or points in several frames of point cloud data (called sparse point cloud), and at the same time, for different frames The splicing process of point cloud data may also cause smearing caused by dynamic object splicing.
针对这些问题,S630中,一方面,所述点云数据处理系统可以根据所述第二检测框的属性信息,将对应相同动态对象的第二检测框进行关联,对不同帧点云数据中针对同一动态物体采集到的点云集合进行对象级别拼接,得到动态对象关联结果(称为密集点云)。另一方面,所述点云数据处理系统可以根据所述第二检测框,去除M帧点云数据中单帧点云数据中的动态对象,得到单帧点云数据对应的第三点云数据,该第三点云数据中只包括静态对象。进而,所述点云数据处理系统可以利用动态对象关联结果,对M帧点云数据对应的第三点云数据进行融合处理,得到第二点云地图,该第二点云地图中包括动态对象和/或静态对象。In view of these problems, in S630, on the one hand, the point cloud data processing system can associate the second detection frames corresponding to the same dynamic object according to the attribute information of the second detection frame, and point cloud data of different frames for The point cloud collection collected by the same dynamic object is spliced at the object level to obtain the dynamic object association result (called dense point cloud). On the other hand, the point cloud data processing system can remove the dynamic object in the single frame point cloud data in the M frame point cloud data according to the second detection frame, and obtain the third point cloud data corresponding to the single frame point cloud data , the third point cloud data only includes static objects. Furthermore, the point cloud data processing system can use the dynamic object association result to perform fusion processing on the third point cloud data corresponding to M frames of point cloud data to obtain a second point cloud map, which includes the dynamic object and/or static objects.
如图9所示,以行驶的车辆41作为动态物体,t1时刻采集到的a帧点云数据中包括探测车辆41的车头采集到的点云集合,t2时刻采集到的b帧点云数据中包括探测车辆41的车身采集到的点云集合,t3时刻采集到的b帧点云数据中包括探测车辆41的车尾采集到的点云集合,分别以检测框A、检测框B、检测框C标出,此时为稀疏点云。进行对象级别关联时,点云数据处理系统可以识别出该a帧点云数据中的检测框A、b帧点云数据中的检测框B、c帧点云数据中的检测框C均对应所述车辆41,检测框A、检测框B、检测框C分别标出的点云集合即关联同一动态对象。进行对象级别拼接时,所述点云数据处理系统可以基于动态对象的粒度,根据点云数据的采集时刻,将不同帧点云数据中对应于同 一动态对象的点云集合进行拼接,得到该动态对象的较为密集的点云集合,如图9所示的关联框D标出,此时为密集点云。同时,该对象级别的拼接处理,可以基于点的采集时刻,将不同帧点云数据中相同动态对象进行对齐,以便消除动态对象引发的拖尾现象。进而,点云数据处理系统在基于所述动态对象关联结果对所述多帧点云数据中的单帧点云数据进行融合处理后,即可在所述第二点云地图中呈现相同动态对象的较为密集的点云集合。As shown in Figure 9, with the moving vehicle 41 as a dynamic object, the a-frame point cloud data collected at time t1 includes the point cloud collection collected by the head of the detection vehicle 41, and the b-frame point cloud data collected at time t2 Including the collection of point clouds collected by the body of the detection vehicle 41, the b-frame point cloud data collected at time t3 includes the collection of point clouds collected by the rear of the detection vehicle 41, respectively represented by detection frame A, detection frame B, detection frame Marked by C, it is a sparse point cloud at this time. When performing object-level association, the point cloud data processing system can recognize that the detection frame A in the point cloud data of frame a, the detection frame B in the point cloud data of frame b, and the detection frame C in the point cloud data of frame c all correspond to the corresponding For the vehicle 41 described above, the point cloud sets respectively marked by the detection frame A, detection frame B, and detection frame C are associated with the same dynamic object. When performing object-level splicing, the point cloud data processing system can splice point cloud sets corresponding to the same dynamic object in different frames of point cloud data based on the granularity of the dynamic object and according to the point cloud data collection time, to obtain the dynamic The relatively dense point cloud collection of the object, marked by the association box D shown in Figure 9, is a dense point cloud at this time. At the same time, the object-level splicing process can align the same dynamic objects in different frames of point cloud data based on the point acquisition time, so as to eliminate the smearing phenomenon caused by dynamic objects. Furthermore, after the point cloud data processing system performs fusion processing on the single frame point cloud data in the multi-frame point cloud data based on the dynamic object association result, the same dynamic object can be presented in the second point cloud map A relatively dense collection of point clouds.
S640:点云数据处理系统根据所述第二点云地图对所述第一检测框和所述第二检测框进行校正,得到目标检测框,所述目标检测框用于指示在所述单帧点云数据中标注的静态对象和/或动态对象。S640: The point cloud data processing system corrects the first detection frame and the second detection frame according to the second point cloud map to obtain a target detection frame, and the target detection frame is used to indicate that in the single frame Static objects and/or dynamic objects annotated in point cloud data.
其中,由于第一检测框用于指示静态对象,所述第二检测框用于指示动态对象,而静态对象和动态对象的属性可能相同也可能不同,S640中执行校正过程中,所述点云数据处理系统可以根据所述第二点云地图针对不同对象的不同属性进行校正,以得到所有单帧点云数据的目标检测框。Wherein, since the first detection frame is used to indicate the static object, and the second detection frame is used to indicate the dynamic object, and the attributes of the static object and the dynamic object may be the same or may be different, during the correction process in S640, the point cloud The data processing system may correct different attributes of different objects according to the second point cloud map, so as to obtain target detection frames of all single frames of point cloud data.
例如,点云数据处理系统可以在所述第二点云地图中,对标注静态对象的第一检测框的第一属性(例如尺寸、位置、形状等)进行修正,和/或,对标注动态对象的第二检测框的第二属性(例如尺寸、形状等)进行修正,得到第四检测框,所述第四检测框用于指示所述第二点云地图中标注的动态对象和/或静态对象;在所述第二点云地图对应的单帧点云数据中,对标注动态对象的第四检测框的第三属性(例如位置等)进行修正,得到第五检测框;以标注静态对象的第四检测框和所述第五检测框作为所述M帧点云数据的目标检测框。For example, the point cloud data processing system may, in the second point cloud map, correct the first attribute (such as size, position, shape, etc.) The second attribute (such as size, shape, etc.) of the second detection frame of the object is corrected to obtain a fourth detection frame, and the fourth detection frame is used to indicate the dynamic object marked in the second point cloud map and/or static object; in the single-frame point cloud data corresponding to the second point cloud map, modify the third attribute (such as position, etc.) of the fourth detection frame of the marked dynamic object to obtain the fifth detection frame; to mark the static The fourth detection frame of the object and the fifth detection frame are used as target detection frames of the M frames of point cloud data.
其中,所述目标检测框用于标注单帧点云数据中的静态对象和/或动态对象,目标检测框的属性信息用于描述该目标检测框标注的静态对象和/或动态对象。其中,上述对尺寸的修正例如可以包括对检查框的尺寸的增大、减小等,对位置的修正例如可以包括对检测框的绝对位置和/或相对位置的修正(包括检测框在任一方向的偏移)等,对形状的修正例如可以包括对监测框的形状的修正,例如从正方体修正为长方体,从长方体修正为正方体,从圆锥体修正为圆柱体等。Wherein, the target detection frame is used to mark static objects and/or dynamic objects in a single frame of point cloud data, and the attribute information of the target detection frame is used to describe the static objects and/or dynamic objects marked by the target detection frame. Wherein, the above-mentioned correction to the size may include, for example, the increase and decrease of the size of the inspection frame, and the correction to the position may, for example, include correction to the absolute position and/or relative position of the detection frame (including the detection frame in any direction). The correction of the shape may include, for example, the correction of the shape of the monitoring frame, for example, from a cube to a cuboid, from a cuboid to a cube, from a cone to a cylinder, and so on.
可以理解的是,S640可由点云数据处理系统自动实现上述校正过程,或者,S640也可以采用人工修正的方式实现上述校正过程。It can be understood that, in S640, the point cloud data processing system may automatically implement the above correction process, or in S640, the above correction process may be implemented manually.
示例地,采用人工修正方式时,所述点云数据处理系统可以通过用户界面输出所述第二点云地图以及用于标注所述第二点云地图中的静态对象第一检测框和/或标注动态对象的第二检测框,标注人员可以通过所述用户界面查看该第二点云地图、所述第一检测框以及所述第二检测框。该标注人员可以在该第二点云地图中对第一检测框的尺寸、位置、形状等第一属性进行修正,对第二检测框的尺寸、形状等第二属性进行修正,得到第四检测框信息。进一步地,该标注人员在所述第二点云地图对应的各个单帧点云数据中,对标注动态对象的第四检测框的位置等第三属性进行修正,得到第五检测框。For example, when manual correction is adopted, the point cloud data processing system may output the second point cloud map and the first detection frame and/or Marking the second detection frame of the dynamic object, the labeler can view the second point cloud map, the first detection frame and the second detection frame through the user interface. The annotator can correct the first attributes such as the size, position, and shape of the first detection frame in the second point cloud map, and correct the second attributes such as the size and shape of the second detection frame to obtain the fourth detection frame. box information. Further, the annotator corrects the third attributes such as the position of the fourth detection frame of the marked dynamic object in each single frame of point cloud data corresponding to the second point cloud map to obtain the fifth detection frame.
由此,利用第二点云地图,以标注静态对象的第四检测框和所述第五检测框作为所述M帧点云数据中单帧点云数据的目标检测框(即标注结果),不仅可以在单帧点云数据中获得每个对象的更多点云信息,提升点云数据处理系统对单帧点云数据中对应远距离物体或遮挡严重物体的点云集合的自动标注能力,还可以通过将多帧点云数据中的动态对象对准到静态对象的相关特征,减少由动态对象引发的拖尾现象。同时,基于包含动态对象和静态对象的点云地图辅助标注,可以实现便捷的辅助修正,便于以较低的人工成本标注准 确的物体属性。Thus, using the second point cloud map, the fourth detection frame and the fifth detection frame of the static object are used as the target detection frame (ie, the labeling result) of the single frame point cloud data in the M frame point cloud data, Not only can more point cloud information of each object be obtained in a single frame of point cloud data, but also the ability of the point cloud data processing system to automatically label point cloud collections corresponding to distant objects or severely occluded objects in a single frame of point cloud data can be improved. It can also reduce the smearing phenomenon caused by dynamic objects by aligning dynamic objects in multi-frame point cloud data to the relevant features of static objects. At the same time, based on the auxiliary labeling of point cloud maps containing dynamic objects and static objects, convenient auxiliary corrections can be realized, which facilitates the labeling of accurate object attributes at a lower labor cost.
此外,在一些实施方式中,在获得目标检测框之后,点云数据处理系统还可以利用多模型融合与点云地图生成柔化框,提供更高质量的检测框和便于训练的柔化框,有助于提升使用自动标注系统的对部分帧点云数据的模型的训练精度。In addition, in some implementations, after obtaining the target detection frame, the point cloud data processing system can also use multi-model fusion and point cloud map to generate a softening frame to provide a higher-quality detection frame and a softening frame that is easy to train. It helps to improve the training accuracy of models using automatic labeling systems for partial frame point cloud data.
具体实施时,在S640之后,该方法还可以包括以下步骤:将所述M帧点云数据中的单帧点云数据以及所述目标检测框作为第三训练数据,在基于所述第三训练数据训练多个检测模型的过程中确定第六检测框,所述第六检测框的检测结果为错误;将所述第二点云地图以及所述目标检测框作为第四训练数据,在基于所述第四训练数据训练多个检测模型的过程中确定第七检测框;基于所述第六检测框校正所述第七检测框。例如,针对同一对象,使用检测结果为错误的第六检测框代替检测结果为正确的第七检测框,剔除检测结果为错位的第六检测框和检测结果为错误的第七检测框。其中,经过校正后得到的检测框(例如表示为第八检测框)包括便于训练的柔化框以及所述柔化框的属性信息,该柔化框及其属性信息也可以代替上述单帧点云数据的目标检测框,作为训练数据用于训练目标检测模型,由此进一步提升对单帧点云数据检测模型的训练精度。During specific implementation, after S640, the method may also include the following steps: using the single-frame point cloud data in the M frames of point cloud data and the target detection frame as the third training data, based on the third training In the process of data training multiple detection models, the sixth detection frame is determined, and the detection result of the sixth detection frame is an error; the second point cloud map and the target detection frame are used as the fourth training data, based on the Determining a seventh detection frame during the process of training multiple detection models with the fourth training data; correcting the seventh detection frame based on the sixth detection frame. For example, for the same object, the sixth detection frame with an incorrect detection result is used instead of the seventh detection frame with a correct detection result, and the sixth detection frame with a misaligned detection result and the seventh detection frame with an incorrect detection result are eliminated. Wherein, the detection frame obtained after correction (for example, denoted as the eighth detection frame) includes a softening frame for training and attribute information of the softening frame, and the softening frame and its attribute information can also replace the above-mentioned single frame point The target detection frame of the cloud data is used as training data to train the target detection model, thereby further improving the training accuracy of the single-frame point cloud data detection model.
为了便于理解,下面以车辆驾驶场景为例,结合图10所示的方法流程图对该点云数据处理方法进行介绍。For ease of understanding, the point cloud data processing method will be introduced below by taking a vehicle driving scene as an example in conjunction with the method flow chart shown in FIG. 10 .
参阅图10所示,该方法可以包括以下步骤:Referring to Figure 10, the method may include the following steps:
S1010:点云数据处理系统的获取单元获取点云序列。其中,所述点云序列例如可由车辆上的雷达采集得到,该点云序列包括连续的M帧点云数据(M为大于或等于2的整数,M的取值取决于该点云数据处理系统的处理能力、配置信息或应用场景等)。S1010: The acquisition unit of the point cloud data processing system acquires a point cloud sequence. Wherein, the point cloud sequence can be collected by the radar on the vehicle, for example, and the point cloud sequence includes continuous M frames of point cloud data (M is an integer greater than or equal to 2, and the value of M depends on the point cloud data processing system processing capabilities, configuration information or application scenarios, etc.).
实施S1010时,可由该点云数据处理系统的获取单元可以从车辆获取所述点云序列,可以是所述车辆主动上报所述点云序列,也可以是所述车辆响应于所述点云数据处理系统的请求反馈该点云序列。或者,所述点云数据处理系统可以包括存储单元,所述M帧点云数据可以保存在所述存储单元,所述获取单元可以从所述存储单元读取所述点云序列,本申请实施例对该点云序列的获取方式不做限定。When implementing S1010, the acquisition unit of the point cloud data processing system may acquire the point cloud sequence from the vehicle, it may be that the vehicle actively reports the point cloud sequence, or the vehicle responds to the point cloud data The request of the processing system feeds back the sequence of point clouds. Alternatively, the point cloud data processing system may include a storage unit, the M frames of point cloud data may be stored in the storage unit, and the acquisition unit may read the point cloud sequence from the storage unit. This example does not limit the acquisition method of the point cloud sequence.
S1021:点云数据处理系统的处理单元根据车辆的定位姿态等,将M帧点云数据中单帧点云数据变换至世界坐标系,并在该世界坐标系中对M帧点云数据进行拼接,得到第一点云地图。S1021: The processing unit of the point cloud data processing system transforms the single frame point cloud data in the M frame point cloud data into the world coordinate system according to the positioning attitude of the vehicle, and stitches the M frame point cloud data in the world coordinate system , get the first point cloud map.
S1022:点云数据处理系统的处理单元根据第一检测模型对所述第一点云地图进行目标检测,获得第一检测框。S1021-S1022的详细实现可参见上文中结合S610的相关介绍,在此不再赘述。S1022: The processing unit of the point cloud data processing system performs object detection on the first point cloud map according to the first detection model to obtain a first detection frame. For the detailed implementation of S1021-S1022, refer to the relevant introduction in conjunction with S610 above, and will not be repeated here.
S1031(可选):点云数据处理系统的处理单元将M帧点云数据中的N帧点云数据进行拼接,得到第一点云数据。可以理解的是,S1031仅在以M帧点云数据中的若干帧点云数据(即N≥2)组合作为所述第一点云数据时执行,若以所述M帧点云数据中的单帧点云数据作为第一点云数据,则S1031无需执行。S1031 (optional): The processing unit of the point cloud data processing system splices the N frames of point cloud data in the M frames of point cloud data to obtain the first point cloud data. It can be understood that S1031 is only executed when several frames of point cloud data (that is, N≥2) in M frames of point cloud data are combined as the first point cloud data, if the M frames of point cloud data are If a single frame of point cloud data is used as the first point cloud data, S1031 does not need to be executed.
S1032:点云数据处理系统的处理单元根据第二检测模型对所述第一点云数据进行目标检测,得到第三检测框。其中,所述第三检测框用于标注所述第一点云数据中的动态对象。应理解的是,在所述第二检测模型的误差范围内,第三检测框还可能用于标注第一点云数据中的静态对象。S1032: The processing unit of the point cloud data processing system performs object detection on the first point cloud data according to the second detection model to obtain a third detection frame. Wherein, the third detection frame is used to label dynamic objects in the first point cloud data. It should be understood that within the error range of the second detection model, the third detection frame may also be used to label static objects in the first point cloud data.
S1033:点云数据处理系统的处理单元根据所述第一检测框对所述第三检测框进行校正,得到所述第二检测框,所述第二检测框用于标注所述第一点云数据中的动态对象。S1031-S1033的详细实现可参见上文中结合S620的相关介绍,在此不再赘述。S1033: The processing unit of the point cloud data processing system corrects the third detection frame according to the first detection frame to obtain the second detection frame, and the second detection frame is used to label the first point cloud Dynamic objects in data. For the detailed implementation of S1031-S1033, refer to the relevant introduction in connection with S620 above, and will not be repeated here.
S1041:点云数据处理系统的处理单元根据S1033中得到的第二检测框,去除所述M帧点云数据中单帧点云数据中的动态对象,得到单帧点云数据对应的第三点云数据。S1041: The processing unit of the point cloud data processing system removes the dynamic object in the single frame point cloud data in the M frame point cloud data according to the second detection frame obtained in S1033, and obtains the third point corresponding to the single frame point cloud data cloud data.
S1042:点云数据处理系统的处理单元根据M帧点云数据对应的第三点云数据进行拼接处理,生成第四点云地图。S1042: The processing unit of the point cloud data processing system performs splicing processing according to the third point cloud data corresponding to the M frames of point cloud data, to generate a fourth point cloud map.
S1043:点云数据处理系统的处理单元将S1033中得到的第二检测框,在S1010述及的M帧点云数据中的单帧点云数据或所述第一点云地图中,将对应相同动态对象的第二检测框进行关联。S1043: The processing unit of the point cloud data processing system uses the second detection frame obtained in S1033 to correspond to the same The second detection frame of the dynamic object is associated.
S1044:点云数据处理系统的处理单元根据M帧点云数据,将在S1043中得到的对应相同动态对象的第二检测框标出的点云集合进行对象级别拼接,即将不同帧点云数据中关联相同的动态对象的点云集合对齐,整合为同一时刻的动态对象级点云集合,得到动态对象关联结果。S1044: The processing unit of the point cloud data processing system performs object-level stitching on the point cloud set marked in the second detection frame corresponding to the same dynamic object obtained in S1043 according to the M frame point cloud data, that is, in different frames of point cloud data The point cloud sets associated with the same dynamic object are aligned and integrated into a dynamic object-level point cloud set at the same time to obtain the dynamic object association result.
S1045:点云数据处理系统的处理单元根据S1044得到的动态对象关联结果对S1042得到的第四点云地图进行融合处理,得到第二点云地图。S1045: The processing unit of the point cloud data processing system performs fusion processing on the fourth point cloud map obtained in S1042 according to the dynamic object association result obtained in S1044, to obtain a second point cloud map.
S1041-S1045的详细实现可参见上文中结合S630的相关介绍,在此不再赘述。For the detailed implementation of S1041-S1045, refer to the relevant introduction in conjunction with S630 above, and will not be repeated here.
S1051:点云数据处理系统的校正单元通过用户界面输出在S1045中得到第二点云地图,标注人员可以在第二点云地图中修正标注静态对象的第一检测框的尺寸、形状、位置等第一属性,以及标注动态对象的第二检测框的尺寸、形状等第二属性,得到第四检测框。S1051: The correction unit of the point cloud data processing system obtains the second point cloud map in S1045 through the user interface output, and the labeler can correct the size, shape, position, etc. of the first detection frame of the static object in the second point cloud map The first attribute, and the second attribute such as the size and shape of the second detection frame marking the dynamic object are used to obtain the fourth detection frame.
S1052:点云数据处理系统的校正单元将经过属性修正后得到的第四检测框在第二点云地图中标注出的静态对象和/或动态对象提供给第二点云地图对应的单帧点云数据并在用户界面输出,标注人员可以在单帧点云数据中对标注动态对象的第四检测框的位置等第三属性进行修正。由此,得到单帧点云数据的标注结果。S1052: The correction unit of the point cloud data processing system provides the static object and/or dynamic object marked in the second point cloud map by the fourth detection frame obtained after attribute correction to the single frame point corresponding to the second point cloud map The cloud data is output on the user interface, and the annotator can correct the third attribute such as the position of the fourth detection frame of the annotated dynamic object in the single frame point cloud data. Thus, the labeling result of the single frame point cloud data is obtained.
S1051-S1052的详细实现可参见上文中结合S640的相关介绍,在此不再赘述。For the detailed implementation of S1051-S1052, refer to the relevant introduction in conjunction with S640 above, and will not be repeated here.
S1061:点云数据处理系统的训练单元将所述M帧点云数据中的单帧点云数据以及所述目标检测框作为第三训练数据,在基于所述第三训练数据训练多个检测模型的过程中确定第六检测框,所述第六检测框信息的检测结果为错误。S1061: The training unit of the point cloud data processing system uses the single frame of point cloud data in the M frames of point cloud data and the target detection frame as the third training data, and trains multiple detection models based on the third training data In the process of determining the sixth detection frame, the detection result of the sixth detection frame information is wrong.
S1062:点云数据处理系统的训练单元将所述第二点云地图以及所述目标检测框作为第四训练数据,在基于所述第四训练数据训练多个检测模型的过程中确定第七检测框。S1062: The training unit of the point cloud data processing system uses the second point cloud map and the target detection frame as fourth training data, and determines a seventh detection during the process of training multiple detection models based on the fourth training data frame.
S1063:点云数据处理系统的校正单元基于所述第六检测框校正所述第七检测框。S1063: The correction unit of the point cloud data processing system corrects the seventh detection frame based on the sixth detection frame.
S1061-S1063的详细实现可参见上文中结合柔化技术的相关介绍,在此不再赘述。For the detailed implementation of S1061-S1063, please refer to the related introduction of softening technology above, and will not repeat them here.
本申请还提供了一种点云数据处理系统,图11为本申请实施例提供的一种点云数据处理系统1100的结构示意图,所述点云数据处理系统1100可以应用于图4所示的应用场景中的服务器或终端设备。参阅图11所示,所述点云数据处理系统1100可以包括第一获取单元1101、第二获取单元1102、处理单元1103和校正单元1104。The present application also provides a point cloud data processing system. FIG. 11 is a schematic structural diagram of a point cloud data processing system 1100 provided in an embodiment of the present application. The point cloud data processing system 1100 can be applied to the system shown in FIG. 4 A server or terminal device in an application scenario. Referring to FIG. 11 , the point cloud data processing system 1100 may include a first acquisition unit 1101 , a second acquisition unit 1102 , a processing unit 1103 and a correction unit 1104 .
其中,示例地,第一获取单元1101,用于获取第一检测框,所述第一检测框用于指示在第一点云地图中标注的静态对象,所述第一点云地图由M帧点云数据得到,M为大于或等于2的整数;第二获取单元1102,用于获取第二检测框信息,所述第二检测框用于指 示在第一点云数据中标注的动态对象,所述第一点云数据由N帧点云数据得到,N为大于或等于1的整数,N<M;处理单元1103,用于根据所述第二检测框,对所述M帧点云数据中的单帧点云数据进行融合处理,得到第二点云地图;校正单元1104,用于根据所述第二点云地图对所述第一检测框和所述第二检测框进行校正,得到目标检测框,所述目标检测框用于指示在所述单帧点云数据中标注的标注静态对象和/或动态对象。Wherein, for example, the first acquisition unit 1101 is configured to acquire a first detection frame, the first detection frame is used to indicate a static object marked in the first point cloud map, and the first point cloud map consists of M frames The point cloud data is obtained, M is an integer greater than or equal to 2; the second acquisition unit 1102 is used to acquire second detection frame information, and the second detection frame is used to indicate the dynamic object marked in the first point cloud data, The first point cloud data is obtained from N frames of point cloud data, N is an integer greater than or equal to 1, and N<M; the processing unit 1103 is configured to process the M frames of point cloud data according to the second detection frame Fusion processing is performed on the single-frame point cloud data in to obtain a second point cloud map; the correction unit 1104 is configured to correct the first detection frame and the second detection frame according to the second point cloud map to obtain A target detection frame, where the target detection frame is used to indicate annotated static objects and/or dynamic objects marked in the single frame point cloud data.
在一种可能的实现方式中,所述第一获取单元1101用于:对所述M帧点云数据进行拼接,得到所述第一点云地图;根据第一检测模型对所述第一点云地图进行目标检测,获得所述第一检测框其中,所述第一检测模型使用第一训练数据进行训练得到,所述第一训练数据包括第三点云地图与静态对象标注数据。In a possible implementation manner, the first acquisition unit 1101 is configured to: stitch the M frames of point cloud data to obtain the first point cloud map; Object detection is performed on the cloud map to obtain the first detection frame. The first detection model is obtained through training using first training data, and the first training data includes the third point cloud map and static object labeling data.
在一种可能的实现方式中,所述第二获取单元1102用于:根据第二检测模型对所述第一点云数据进行目标检测,获得第三检测框,其中,所述第二检测模型使用第二训练数据进行训练得到,所述第二训练数据包括第二点云数据和动态对象标注数据,所述第三检测框用于指示在所述第一点云数据中标注的动态对象;根据所述第一检测框对所述第三检测框进行校正,得到所述第二检测框。In a possible implementation manner, the second acquisition unit 1102 is configured to: perform object detection on the first point cloud data according to a second detection model to obtain a third detection frame, wherein the second detection model Obtained by using the second training data for training, the second training data includes the second point cloud data and dynamic object labeling data, and the third detection frame is used to indicate the dynamic object marked in the first point cloud data; Correcting the third detection frame according to the first detection frame to obtain the second detection frame.
在一种可能的实现方式中,所述处理单元1103用于:根据所述第二检测框,去除所述M帧点云数据中单帧点云数据中的动态对象,得到单帧点云数据对应的第三点云数据;根据所述第二检测框的属性信息,将对应相同动态对象的第二检测框进行关联,得到动态对象关联结果;基于所述动态对象关联结果对所述M帧点云数据对应的第三点云数据进行融合处理,得到第二点云地图。In a possible implementation manner, the processing unit 1103 is configured to: according to the second detection frame, remove the dynamic object in the single frame of point cloud data in the M frames of point cloud data to obtain the single frame of point cloud data The corresponding third point cloud data; according to the attribute information of the second detection frame, associate the second detection frame corresponding to the same dynamic object to obtain a dynamic object association result; based on the dynamic object association result, the M frame The third point cloud data corresponding to the point cloud data is fused to obtain the second point cloud map.
在一种可能的实现方式中,所述校正单元1104用于:在所述第二点云地图中,对标注静态对象的第一检测框的第一属性进行修正,和/或,对标注动态对象的第二检测框的第二属性进行修正,得到第四检测框;在所述第二点云地图对应的单帧点云数据中,对标注动态对象的第四检测框的第三属性进行修正,得到第五检测框;以标注静态对象的第四检测框和所述第五检测框作为所述M帧点云数据的目标检测框。In a possible implementation manner, the correcting unit 1104 is configured to: in the second point cloud map, correct the first attribute of the first detection frame marking the static object, and/or correct the dynamic The second attribute of the second detection frame of the object is corrected to obtain the fourth detection frame; in the single frame point cloud data corresponding to the second point cloud map, the third attribute of the fourth detection frame marked with the dynamic object is modified Correction to obtain the fifth detection frame; the fourth detection frame marked with the static object and the fifth detection frame are used as the target detection frame of the M frames of point cloud data.
在一种可能的实现方式中,所述系统还包括训练单元,在得到所述目标检测框之后,所述训练单元用于执行以下步骤:将所述M帧点云数据中的单帧点云数据以及所述目标检测框作为第三训练数据,在基于所述第三训练数据训练多个检测模型的过程中确定第六检测框,所述第六检测框的检测结果为错误;将所述第二点云地图以及所述目标检测框作为第四训练数据,在基于所述第四训练数据训练多个检测模型的过程中确定第七检测框;基于所述第六检测框校正所述第七检测框。In a possible implementation manner, the system further includes a training unit. After obtaining the target detection frame, the training unit is configured to perform the following steps: convert the single frame point cloud in the M frames of point cloud data to Data and the target detection frame are used as the third training data, and a sixth detection frame is determined during the process of training multiple detection models based on the third training data, and the detection result of the sixth detection frame is an error; the The second point cloud map and the target detection frame are used as the fourth training data, and the seventh detection frame is determined in the process of training a plurality of detection models based on the fourth training data; the seventh detection frame is corrected based on the sixth detection frame. Seven detection boxes.
本申请还提供了一种点云数据处理装置1200,图12为本申请实施例提供的一种点云数据处理装置1200的结构示意图,所述数据处理装置1200可以应用于图4所示的场景中的服务器或终端设备。参阅图12所示,所述点云数据处理装置1200包括:处理器1201、存储器1202和总线1203。其中,处理器1201和存储器1202通过总线1203进行通信,也可以通过无线传输等其他手段实现通信。该存储器1202用于存储指令,该处理器1201用于执行该存储器1202存储的指令。该存储器1202存储程序代码,且处理器1201可以调用存储器1202中存储的程序代码。The present application also provides a point cloud data processing device 1200. FIG. 12 is a schematic structural diagram of a point cloud data processing device 1200 provided in an embodiment of the present application. The data processing device 1200 can be applied to the scene shown in FIG. 4 server or terminal device in the Referring to FIG. 12 , the point cloud data processing device 1200 includes: a processor 1201 , a memory 1202 and a bus 1203 . Wherein, the processor 1201 and the memory 1202 communicate through the bus 1203 , or communicate through other means such as wireless transmission. The memory 1202 is used to store instructions, and the processor 1201 is used to execute the instructions stored in the memory 1202 . The memory 1202 stores program codes, and the processor 1201 can call the program codes stored in the memory 1202 .
本申请一种可选的实施例中,当数据处理装置1200为点云数据处理装置时,所述处理器1201用于执行上述方法实施例,详细可参见上文的相关描述,在此不再赘述。In an optional embodiment of the present application, when the data processing device 1200 is a point cloud data processing device, the processor 1201 is used to execute the above-mentioned method embodiment. For details, please refer to the relevant description above, which will not be repeated here. repeat.
可以理解,本申请图12中的存储器1202可以是易失性存储器或非易失性存储器,或可包括易失性和非易失性存储器两者。其中,非易失性存储器可以是只读存储器(Read-Only Memory,ROM)、可编程只读存储器(Programmable ROM,PROM)、可擦除可编程只读存储器(Erasable PROM,EPROM)、电可擦除可编程只读存储器(Electrically EPROM,EEPROM)或闪存。易失性存储器可以是随机存取存储器(Random Access Memory,RAM),其用作外部高速缓存。通过示例性但不是限制性说明,许多形式的RAM可用,例如静态随机存取存储器(Static RAM,SRAM)、动态随机存取存储器(Dynamic RAM,DRAM)、同步动态随机存取存储器(Synchronous DRAM,SDRAM)、双倍数据速率同步动态随机存取存储器(Double Data Rate SDRAM,DDR SDRAM)、增强型同步动态随机存取存储器(Enhanced SDRAM,ESDRAM)、同步连接动态随机存取存储器(Synchlink DRAM,SLDRAM)和直接内存总线随机存取存储器(Direct Rambus RAM,DR RAM)。应注意,本文描述的系统和方法的存储器旨在包括但不限于这些和任意其它适合类型的存储器。It can be understood that the memory 1202 in FIG. 12 of the present application may be a volatile memory or a non-volatile memory, or may include both volatile and non-volatile memories. Among them, the non-volatile memory can be read-only memory (Read-Only Memory, ROM), programmable read-only memory (Programmable ROM, PROM), erasable programmable read-only memory (Erasable PROM, EPROM), electronically programmable Erase Programmable Read-Only Memory (Electrically EPROM, EEPROM) or Flash. The volatile memory can be Random Access Memory (RAM), which acts as external cache memory. By way of illustration and not limitation, many forms of RAM are available, such as Static Random Access Memory (Static RAM, SRAM), Dynamic Random Access Memory (Dynamic RAM, DRAM), Synchronous Dynamic Random Access Memory (Synchronous DRAM, SDRAM), double data rate synchronous dynamic random access memory (Double Data Rate SDRAM, DDR SDRAM), enhanced synchronous dynamic random access memory (Enhanced SDRAM, ESDRAM), synchronous connection dynamic random access memory (Synchlink DRAM, SLDRAM ) and Direct Memory Bus Random Access Memory (Direct Rambus RAM, DR RAM). It should be noted that the memory of the systems and methods described herein is intended to include, but not be limited to, these and any other suitable types of memory.
基于以上实施例,本申请实施例还提供了一种计算机程序,当所述计算机程序在计算机上运行时,使得所述计算机执行上述方法实施例。Based on the above embodiments, the embodiments of the present application further provide a computer program, which, when the computer program is run on a computer, causes the computer to execute the above method embodiments.
基于以上实施例,本申请实施例还提供了一种计算机可读存储介质,该计算机可读存储介质中存储有计算机程序,所述计算机程序被计算机执行时,使得计算机执行上述方法实施例。其中,存储介质可以是计算机能够存取的任何可用介质。以此为例但不限于:计算机可读介质可以包括RAM、ROM、EEPROM、CD-ROM或其他光盘存储、磁盘存储介质或者其他磁存储设备、或者能够用于携带或存储具有指令或数据结构形式的期望的程序代码并能够由计算机存取的任何其他介质。Based on the above embodiments, an embodiment of the present application further provides a computer-readable storage medium, in which a computer program is stored, and when the computer program is executed by a computer, the computer executes the above-mentioned method embodiment. Wherein, the storage medium may be any available medium that can be accessed by a computer. By way of example but not limitation: computer-readable media may include RAM, ROM, EEPROM, CD-ROM or other optical disk storage, magnetic disk storage media or other magnetic storage devices, or may be used to carry or store information in the form of instructions or data structures desired program code and any other medium that can be accessed by a computer.
基于以上实施例,本申请实施例还提供了一种芯片,所述芯片用于读取存储器中存储的计算机程序,实现上述方法实施例。Based on the above embodiments, an embodiment of the present application further provides a chip for reading a computer program stored in a memory to implement the above method embodiments.
基于以上实施例,本申请实施例提供了一种芯片系统,该芯片系统包括处理器,用于支持计算机装置实现上述方法实施例。在一种可能的设计中,所述芯片系统还包括存储器,所述存储器用于保存该计算机装置必要的程序和数据。该芯片系统,可以由芯片构成,也可以包含芯片和其他分立器件。Based on the above embodiments, an embodiment of the present application provides a chip system, where the chip system includes a processor, configured to support a computer device to implement the above method embodiments. In a possible design, the chip system further includes a memory, and the memory is used to store necessary programs and data of the computer device. The system-on-a-chip may consist of chips, or may include chips and other discrete devices.
本领域内的技术人员应明白,本申请的实施例可提供为方法、系统、或计算机程序产品。因此,本申请可采用完全硬件实施例、完全软件实施例、或结合软件和硬件方面的实施例的形式。而且,本申请可采用在一个或多个其中包含有计算机可用程序代码的计算机可用存储介质(包括但不限于磁盘存储器、CD-ROM、光学存储器等)上实施的计算机程序产品的形式。Those skilled in the art should understand that the embodiments of the present application may be provided as methods, systems, or computer program products. Accordingly, the present application may take the form of an entirely hardware embodiment, an entirely software embodiment, or an embodiment combining software and hardware aspects. Furthermore, the present application may take the form of a computer program product embodied on one or more computer-usable storage media (including but not limited to disk storage, CD-ROM, optical storage, etc.) having computer-usable program code embodied therein.
本申请是参照根据本申请的方法、设备(系统)、和计算机程序产品的流程图和/或方框图来描述的。应理解可由计算机程序指令实现流程图和/或方框图中的每一流程和/或方框、以及流程图和/或方框图中的流程和/或方框的结合。可提供这些计算机程序指令到通用计算机、专用计算机、嵌入式处理机或其他可编程数据处理设备的处理器以产生一个机器,使得通过计算机或其他可编程数据处理设备的处理器执行的指令产生用于实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能的装置。The present application is described with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to the present application. It should be understood that each procedure and/or block in the flowchart and/or block diagram, and a combination of procedures and/or blocks in the flowchart and/or block diagram can be realized by computer program instructions. These computer program instructions may be provided to a general purpose computer, special purpose computer, embedded processor, or processor of other programmable data processing equipment to produce a machine such that the instructions executed by the processor of the computer or other programmable data processing equipment produce a An apparatus for realizing the functions specified in one or more procedures of the flowchart and/or one or more blocks of the block diagram.
这些计算机程序指令也可存储在能引导计算机或其他可编程数据处理设备以特定方式工作的计算机可读存储器中,使得存储在该计算机可读存储器中的指令产生包括指令装置的制造品,该指令装置实现在流程图一个流程或多个流程和/或方框图一个方框或多个 方框中指定的功能。These computer program instructions may also be stored in a computer-readable memory capable of directing a computer or other programmable data processing apparatus to operate in a specific manner, such that the instructions stored in the computer-readable memory produce an article of manufacture comprising instruction means, the instructions The device realizes the function specified in one or more procedures of the flowchart and/or one or more blocks of the block diagram.
这些计算机程序指令也可装载到计算机或其他可编程数据处理设备上,使得在计算机或其他可编程设备上执行一系列操作步骤以产生计算机实现的处理,从而在计算机或其他可编程设备上执行的指令提供用于实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能的步骤。These computer program instructions can also be loaded onto a computer or other programmable data processing device, causing a series of operational steps to be performed on the computer or other programmable device to produce a computer-implemented process, thereby The instructions provide steps for implementing the functions specified in the flow chart or blocks of the flowchart and/or the block or blocks of the block diagrams.
显然,本领域的技术人员可以对本申请进行各种改动和变型而不脱离本申请的保护范围。这样,倘若本申请的这些修改和变型属于本申请权利要求及其等同技术的范围之内,则本申请也意图包含这些改动和变型在内。Apparently, those skilled in the art can make various changes and modifications to this application without departing from the protection scope of this application. In this way, if these modifications and variations of the present application fall within the scope of the claims of the present application and their equivalent technologies, the present application is also intended to include these modifications and variations.

Claims (15)

  1. 一种点云数据处理方法,其特征在于,包括:A method for processing point cloud data, comprising:
    获取第一检测框,所述第一检测框用于指示在第一点云地图中标注的静态对象,所述第一点云地图由M帧点云数据得到,M为大于或等于2的整数;Obtain a first detection frame, the first detection frame is used to indicate the static object marked in the first point cloud map, the first point cloud map is obtained from M frames of point cloud data, and M is an integer greater than or equal to 2 ;
    获取第二检测框,所述第二检测框用于指示在第一点云数据中标注的动态对象,所述第一点云数据由N帧点云数据得到,N为大于或等于1的整数,N<M;Obtain a second detection frame, the second detection frame is used to indicate the dynamic object marked in the first point cloud data, the first point cloud data is obtained from N frames of point cloud data, N is an integer greater than or equal to 1 , N<M;
    根据所述第二检测框,对所述M帧点云数据中的单帧点云数据进行融合处理,得到第二点云地图;According to the second detection frame, the single-frame point cloud data in the M frames of point cloud data is fused to obtain a second point cloud map;
    根据所述第二点云地图对所述第一检测框和所述第二检测框进行校正,得到目标检测框,所述目标检测框用于指示在所述单帧点云数据中标注的静态对象和/或动态对象。Correct the first detection frame and the second detection frame according to the second point cloud map to obtain a target detection frame, and the target detection frame is used to indicate the static marked in the single frame point cloud data object and/or dynamic object.
  2. 根据权利要求1所述的方法,其特征在于,所述获取第一检测框,包括:The method according to claim 1, wherein said obtaining the first detection frame comprises:
    对所述M帧点云数据进行拼接,得到所述第一点云地图;Splicing the M frames of point cloud data to obtain the first point cloud map;
    根据第一检测模型对所述第一点云地图进行目标检测,获得所述第一检测框,其中,所述第一检测模型使用第一训练数据进行训练得到,所述第一训练数据包括第三点云地图与静态对象标注数据。Perform target detection on the first point cloud map according to the first detection model to obtain the first detection frame, wherein the first detection model is obtained by training using first training data, and the first training data includes the first Three point cloud maps with static object annotation data.
  3. 根据权利要求1或2所述的方法,其特征在于,所述获取第二检测框,包括:The method according to claim 1 or 2, wherein said obtaining the second detection frame comprises:
    根据第二检测模型对所述第一点云数据进行目标检测,得到第三检测框,其中,所述第二检测模型使用第二训练数据进行训练得到,所述第二训练数据包括第二点云数据和动态对象标注数据,所述第三检测框用于指示在所述第一点云数据中标注的动态对象;Target detection is performed on the first point cloud data according to the second detection model to obtain a third detection frame, wherein the second detection model is obtained by training using the second training data, and the second training data includes the second point Cloud data and dynamic object labeling data, the third detection frame is used to indicate the dynamic object marked in the first point cloud data;
    根据所述第一检测框对所述第三检测框进行校正,得到所述第二检测框。Correcting the third detection frame according to the first detection frame to obtain the second detection frame.
  4. 根据权利要求1-3中任一项所述的方法,其特征在于,所述根据所述第二检测框,对所述M帧点云数据中的单帧点云数据进行融合处理,得到第二点云地图,包括:The method according to any one of claims 1-3, wherein, according to the second detection frame, the single frame point cloud data in the M frames of point cloud data is fused to obtain the first Two point cloud maps, including:
    根据所述第二检测框,去除所述M帧点云数据中单帧点云数据中的动态对象,得到单帧点云数据对应的第三点云数据;According to the second detection frame, remove the dynamic object in the single-frame point cloud data in the M-frame point cloud data, and obtain the third point cloud data corresponding to the single-frame point cloud data;
    根据所述第二检测框的属性信息,将对应相同动态对象的第二检测框进行关联,得到动态对象关联结果;According to the attribute information of the second detection frame, associating the second detection frame corresponding to the same dynamic object to obtain a dynamic object association result;
    基于所述动态对象关联结果对所述M帧点云数据对应的第三点云数据进行融合处理,得到所述二点云地图。Fusion processing is performed on the third point cloud data corresponding to the M frames of point cloud data based on the dynamic object association result to obtain the two point cloud map.
  5. 根据权利要求1-4中任一项所述的方法,其特征在于,所述根据所述第二点云地图对所述第一检测框和所述第二检测框进行校正,得到目标检测框,包括:The method according to any one of claims 1-4, wherein the first detection frame and the second detection frame are corrected according to the second point cloud map to obtain a target detection frame ,include:
    在所述第二点云地图中,对标注静态对象的第一检测框的第一属性进行修正,和/或,对标注动态对象的第二检测框的第二属性进行修正,得到第四检测框,所述第四检测框用于指示所述第二点云地图中标注的动态对象和/或静态对象;In the second point cloud map, the first attribute of the first detection frame marked with the static object is corrected, and/or the second attribute of the second detection frame marked with the dynamic object is corrected to obtain the fourth detected frame, the fourth detection frame is used to indicate the dynamic object and/or static object marked in the second point cloud map;
    在所述第二点云地图对应的单帧点云数据中,对标注动态对象的第四检测框的第三属性进行修正,得到第五检测框;In the single-frame point cloud data corresponding to the second point cloud map, the third attribute of the fourth detection frame marked with the dynamic object is corrected to obtain the fifth detection frame;
    以标注静态对象的第四检测框和所述第五检测框作为所述M帧点云数据的目标检测框。The fourth detection frame and the fifth detection frame marked with a static object are used as target detection frames of the M frames of point cloud data.
  6. 根据权利要求1-5中任一项所述的方法,其特征在于,所述方法还包括:The method according to any one of claims 1-5, wherein the method further comprises:
    将所述M帧点云数据中的单帧点云数据以及所述目标检测框作为第三训练数据,在基 于所述第三训练数据训练多个检测模型的过程中确定第六检测框,所述第六检测框的检测结果为错误;Using the single-frame point cloud data in the M frames of point cloud data and the target detection frame as the third training data, the sixth detection frame is determined in the process of training multiple detection models based on the third training data, so The detection result of the sixth detection frame is wrong;
    将所述第二点云地图以及所述目标检测框作为第四训练数据,在基于所述第四训练数据训练多个检测模型的过程中确定第七检测框;Using the second point cloud map and the target detection frame as fourth training data, determining a seventh detection frame during the process of training multiple detection models based on the fourth training data;
    基于所述第六检测框校正所述第七检测框。Correcting the seventh detection frame based on the sixth detection frame.
  7. 一种点云数据处理系统,其特征在于,包括:A point cloud data processing system, characterized in that it comprises:
    第一获取单元,用于获取第一检测框,所述第一检测框用于指示第一点云地图中标注的静态对象,所述第一点云地图由M帧点云数据得到,M为大于或等于2的整数;The first acquisition unit is configured to acquire a first detection frame, the first detection frame is used to indicate the static object marked in the first point cloud map, the first point cloud map is obtained from M frames of point cloud data, and M is an integer greater than or equal to 2;
    第二获取单元,用于获取第二检测框,所述第二检测框用于指示在第一点云数据中标注的动态对象,所述第一点云数据由N帧点云数据得到,N为大于或等于1的整数,N<M;The second acquisition unit is configured to acquire a second detection frame, the second detection frame is used to indicate the dynamic object marked in the first point cloud data, the first point cloud data is obtained from N frames of point cloud data, N is an integer greater than or equal to 1, N<M;
    处理单元,用于根据所述第二检测框,对所述M帧点云数据中的单帧点云数据进行融合处理,得到第二点云地图;A processing unit, configured to perform fusion processing on the single frame of point cloud data in the M frames of point cloud data according to the second detection frame, to obtain a second point cloud map;
    校正单元,用于根据所述第二点云地图对所述第一检测框和所述第二检测框进行校正,得到目标检测框,所述目标检测框用于指示在所述单帧点云数据中标注的标注静态对象和/或动态对象。A correction unit, configured to correct the first detection frame and the second detection frame according to the second point cloud map to obtain a target detection frame, and the target detection frame is used to indicate that the single-frame point cloud Annotated static objects and/or dynamic objects labeled in the data.
  8. 根据权利要求7所述的系统,其特征在于,所述第一获取单元用于:The system according to claim 7, wherein the first acquisition unit is used for:
    对所述M点云数据进行拼接,得到所述第一点云地图;Splicing the M point cloud data to obtain the first point cloud map;
    根据第一检测模型对所述第一点云地图进行目标检测,获得所述第一检测框,其中,所述第一检测模型使用第一训练数据进行训练得到,所述第一训练数据包括第三点云地图与静态对象标注数据。Perform target detection on the first point cloud map according to the first detection model to obtain the first detection frame, wherein the first detection model is obtained by training using first training data, and the first training data includes the first Three point cloud maps with static object annotation data.
  9. 根据权利要求7或8所述的系统,其特征在于,所述第二获取单元用于:The system according to claim 7 or 8, wherein the second acquisition unit is used for:
    根据第二检测模型对所述第一点云数据进行目标检测,获得第三检测框,其中,所述第二检测模型使用第二训练数据进行训练得到,所述第二训练数据包括第二点云数据和动态对象标注数据,所述第三检测框用于指示在所述第一点云数据中标注的动态对象;Target detection is performed on the first point cloud data according to the second detection model to obtain a third detection frame, wherein the second detection model is obtained by training using the second training data, and the second training data includes the second point Cloud data and dynamic object labeling data, the third detection frame is used to indicate the dynamic object marked in the first point cloud data;
    根据所述第一检测框对所述第三检测框信息进行校正,得到所述第二检测框。Correcting the information of the third detection frame according to the first detection frame to obtain the second detection frame.
  10. 根据权利要求7-9中任一项所述的系统,其特征在于,所述处理单元用于:The system according to any one of claims 7-9, wherein the processing unit is used for:
    根据所述第二检测框,去除所述M帧点云数据中单帧点云数据中的动态对象,得到单帧点云数据对应的第三点云数据;According to the second detection frame, remove the dynamic object in the single-frame point cloud data in the M-frame point cloud data, and obtain the third point cloud data corresponding to the single-frame point cloud data;
    根据所述第二检测框的属性信息,将对应相同动态对象的第二检测框进行关联,得到动态对象关联结果;According to the attribute information of the second detection frame, associating the second detection frame corresponding to the same dynamic object to obtain a dynamic object association result;
    基于所述动态对象关联结果对所述M帧点云数据对应的第三点云数据进行融合处理,得到第二点云地图。Fusion processing is performed on the third point cloud data corresponding to the M frames of point cloud data based on the dynamic object association result to obtain a second point cloud map.
  11. 根据权利要求7-10中任一项所述的系统,其特征在于,所述校正单元用于:The system according to any one of claims 7-10, wherein the correction unit is used for:
    在所述第二点云地图中,对标注静态对象的第一检测框的第一属性进行修正,和/或,对标注动态对象的第二检测框的第二属性进行修正,得到第四检测框;In the second point cloud map, the first attribute of the first detection frame marked with the static object is corrected, and/or the second attribute of the second detection frame marked with the dynamic object is corrected to obtain the fourth detected frame;
    在所述第二点云地图对应的单帧点云数据中,对标注动态对象的第四检测框的第三属性进行修正,得到第五检测框;In the single-frame point cloud data corresponding to the second point cloud map, the third attribute of the fourth detection frame marked with the dynamic object is corrected to obtain the fifth detection frame;
    以标注静态对象的第四检测框和所述第五检测框作为所述M帧点云数据的目标检测 框。The fourth detection frame and the fifth detection frame of the marked static object are used as the target detection frame of the M frames of point cloud data.
  12. 根据权利要求7-11中任一项所述的系统,其特征在于,所述系统还包括训练单元,所述训练单元用于:The system according to any one of claims 7-11, wherein the system further comprises a training unit, the training unit is used for:
    将所述M帧点云数据中的单帧点云数据以及所述目标检测框作为第三训练数据,在基于所述第三训练数据训练多个检测模型的过程中确定第六检测框,所述第六检测框的检测结果为错误;Using the single-frame point cloud data in the M frames of point cloud data and the target detection frame as the third training data, the sixth detection frame is determined in the process of training multiple detection models based on the third training data, so The detection result of the sixth detection frame is wrong;
    将所述第二点云地图以及所述目标检测框息作为第四训练数据,在基于所述第四训练数据训练多个检测模型的过程中确定第七检测框;Using the second point cloud map and the target detection frame information as fourth training data, determining a seventh detection frame during the process of training multiple detection models based on the fourth training data;
    所述校正单元还用于:基于所述第六检测框校正所述第七检测框。The correction unit is further configured to: correct the seventh detection frame based on the sixth detection frame.
  13. 一种点云数据处理系统,其特征在于,包括存储器和处理器,A point cloud data processing system, characterized in that it includes a memory and a processor,
    所述存储器用于存储程序;The memory is used to store programs;
    所述处理器用于执行所述存储器所存储的程序,以使所述装置实现如所述权利要求1-6中任一项所述的方法。The processor is configured to execute the program stored in the memory, so that the device implements the method according to any one of claims 1-6.
  14. 一种计算机可读存储介质,其特征在于,所述计算机可读介质存储有程序代码,当所述程序代码在计算机上运行时,使得计算机执行如权利要求1至6中任一项所述的方法。A computer-readable storage medium, characterized in that the computer-readable medium stores program codes, and when the program codes run on the computer, the computer executes the computer program according to any one of claims 1 to 6. method.
  15. 一种计算机程序产品,其特征在于,当所述计算机程序产品在计算机上运行时,使得所述计算机执行如权利要求1至6中任一项所述的方法。A computer program product, characterized in that, when the computer program product is run on a computer, the computer is made to execute the method according to any one of claims 1 to 6.
PCT/CN2022/127326 2021-10-27 2022-10-25 Point cloud data processing method and system WO2023072055A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202111253590.5A CN116052155A (en) 2021-10-27 2021-10-27 Point cloud data processing method and system
CN202111253590.5 2021-10-27

Publications (1)

Publication Number Publication Date
WO2023072055A1 true WO2023072055A1 (en) 2023-05-04

Family

ID=86127801

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/127326 WO2023072055A1 (en) 2021-10-27 2022-10-25 Point cloud data processing method and system

Country Status (2)

Country Link
CN (1) CN116052155A (en)
WO (1) WO2023072055A1 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116665212A (en) * 2023-07-31 2023-08-29 福思(杭州)智能科技有限公司 Data labeling method, device, processing equipment and storage medium
CN117593685A (en) * 2024-01-19 2024-02-23 福思(杭州)智能科技有限公司 Method and device for constructing true value data and storage medium
CN117612070A (en) * 2024-01-19 2024-02-27 福思(杭州)智能科技有限公司 Static truth value data correction method and device and storage medium

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116778262B (en) * 2023-08-21 2023-11-10 江苏源驶科技有限公司 Three-dimensional target detection method and system based on virtual point cloud

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2019232075A1 (en) * 2018-06-01 2019-12-05 Magic Leap, Inc. Compression of dynamic unstructured point clouds
CN111881827A (en) * 2020-07-28 2020-11-03 浙江商汤科技开发有限公司 Target detection method and device, electronic equipment and storage medium
CN112419233A (en) * 2020-10-20 2021-02-26 腾讯科技(深圳)有限公司 Data annotation method, device, equipment and computer readable storage medium
CN113222042A (en) * 2021-05-25 2021-08-06 深圳市商汤科技有限公司 Evaluation method, evaluation device, electronic equipment and storage medium

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2019232075A1 (en) * 2018-06-01 2019-12-05 Magic Leap, Inc. Compression of dynamic unstructured point clouds
CN111881827A (en) * 2020-07-28 2020-11-03 浙江商汤科技开发有限公司 Target detection method and device, electronic equipment and storage medium
CN112419233A (en) * 2020-10-20 2021-02-26 腾讯科技(深圳)有限公司 Data annotation method, device, equipment and computer readable storage medium
CN113222042A (en) * 2021-05-25 2021-08-06 深圳市商汤科技有限公司 Evaluation method, evaluation device, electronic equipment and storage medium

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116665212A (en) * 2023-07-31 2023-08-29 福思(杭州)智能科技有限公司 Data labeling method, device, processing equipment and storage medium
CN116665212B (en) * 2023-07-31 2023-10-13 福思(杭州)智能科技有限公司 Data labeling method, device, processing equipment and storage medium
CN117593685A (en) * 2024-01-19 2024-02-23 福思(杭州)智能科技有限公司 Method and device for constructing true value data and storage medium
CN117612070A (en) * 2024-01-19 2024-02-27 福思(杭州)智能科技有限公司 Static truth value data correction method and device and storage medium
CN117593685B (en) * 2024-01-19 2024-04-26 福思(杭州)智能科技有限公司 Method and device for constructing true value data and storage medium
CN117612070B (en) * 2024-01-19 2024-05-03 福思(杭州)智能科技有限公司 Static truth value data correction method and device and storage medium

Also Published As

Publication number Publication date
CN116052155A (en) 2023-05-02

Similar Documents

Publication Publication Date Title
WO2023072055A1 (en) Point cloud data processing method and system
US20190226853A1 (en) Methods and Systems for Generating and Using Localisation Reference Data
US20190204094A1 (en) Method, apparatus, and computer readable storage medium for updating electronic map
US9123242B2 (en) Pavement marker recognition device, pavement marker recognition method and pavement marker recognition program
JP6886079B2 (en) Camera calibration systems and methods using traffic sign recognition, and computer-readable media
US10783368B2 (en) Method and apparatus for identifying intersection in electronic map
CN110426051A (en) A kind of lane line method for drafting, device and storage medium
CN109949594A (en) Real-time traffic light recognition method
CN108594244B (en) Obstacle recognition transfer learning method based on stereoscopic vision and laser radar
CN111582189B (en) Traffic signal lamp identification method and device, vehicle-mounted control terminal and motor vehicle
CN109931939A (en) Localization method, device, equipment and the computer readable storage medium of vehicle
CN107218927B (en) A kind of cargo pallet detection system and method based on TOF camera
JP7275280B2 (en) Systems and methods for automatic labeling of images for supervised machine learning
CN113034566B (en) High-precision map construction method and device, electronic equipment and storage medium
CN111912416A (en) Method, device and equipment for positioning equipment
WO2021253245A1 (en) Method and device for identifying vehicle lane changing tendency
KR20180067199A (en) Apparatus and method for recognizing object
WO2023240805A1 (en) Connected vehicle overspeed early warning method and system based on filtering correction
KR20170068937A (en) Autonomous driving vehicle navigation system using the tunnel lighting
JP2021103160A (en) Autonomous traveling vehicle meaning map establishment system and establishment method
CN115164918A (en) Semantic point cloud map construction method and device and electronic equipment
CN113988197A (en) Multi-camera and multi-laser radar based combined calibration and target fusion detection method
CN113706633B (en) Three-dimensional information determination method and device for target object
CN117079238A (en) Road edge detection method, device, equipment and storage medium
WO2023124821A1 (en) Collection method and apparatus for data for mobile robot map construction

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22885932

Country of ref document: EP

Kind code of ref document: A1