WO2022068225A1 - Point cloud annotating method and apparatus, electronic device, storage medium, and program product - Google Patents

Point cloud annotating method and apparatus, electronic device, storage medium, and program product Download PDF

Info

Publication number
WO2022068225A1
WO2022068225A1 PCT/CN2021/095444 CN2021095444W WO2022068225A1 WO 2022068225 A1 WO2022068225 A1 WO 2022068225A1 CN 2021095444 W CN2021095444 W CN 2021095444W WO 2022068225 A1 WO2022068225 A1 WO 2022068225A1
Authority
WO
WIPO (PCT)
Prior art keywords
dimensional
information
marked
view
point cloud
Prior art date
Application number
PCT/CN2021/095444
Other languages
French (fr)
Chinese (zh)
Inventor
王泰
何聪辉
王哲
石建萍
林达华
Original Assignee
上海商汤临港智能科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 上海商汤临港智能科技有限公司 filed Critical 上海商汤临港智能科技有限公司
Priority to KR1020227018072A priority Critical patent/KR20220092928A/en
Publication of WO2022068225A1 publication Critical patent/WO2022068225A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/70Labelling scene content, e.g. deriving syntactic or semantic representations
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • G06T7/75Determining position or orientation of objects or cameras using feature-based methods involving models
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S17/00Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
    • G01S17/88Lidar systems specially adapted for specific applications
    • G01S17/89Lidar systems specially adapted for specific applications for mapping or imaging
    • G01S17/8943D imaging with simultaneous measurement of time-of-flight at a 2D array of receiver pixels, e.g. time-of-flight cameras or flash lidar
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S17/00Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
    • G01S17/88Lidar systems specially adapted for specific applications
    • G01S17/93Lidar systems specially adapted for specific applications for anti-collision purposes
    • G01S17/931Lidar systems specially adapted for specific applications for anti-collision purposes of land vehicles
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • G06T19/20Editing of 3D images, e.g. changing shapes or colours, aligning objects or positioning parts
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/97Determining parameters from multiple pictures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/56Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
    • G06V20/58Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10028Range image; Depth image; 3D point clouds
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10032Satellite or aerial image; Remote sensing
    • G06T2207/10044Radar image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30248Vehicle exterior or interior
    • G06T2207/30252Vehicle exterior; Vicinity of vehicle
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2219/00Indexing scheme for manipulating 3D models or images for computer graphics
    • G06T2219/20Indexing scheme for editing of 3D models
    • G06T2219/2004Aligning objects, relative positioning of parts

Definitions

  • the present disclosure relates to the field of computer technology, and in particular, to a method, device, electronic device, storage medium and program product for point cloud labeling.
  • Lidar is widely used in various technical fields. Taking the field of autonomous driving as an example, it is particularly important to use the point cloud data collected by lidar to accurately detect obstacles around the vehicle, such as pedestrians and other vehicles.
  • the detection model In order to accurately detect obstacles, the detection model needs to be trained based on the collected point cloud data as training samples. Before training, it is necessary to accurately label the collected point cloud data to improve the accuracy of the detection results.
  • Embodiments of the present disclosure provide a method, apparatus, electronic device, storage medium, and program product for point cloud labeling.
  • an embodiment of the present disclosure provides a point cloud labeling method, the method comprising:
  • the three-dimensional annotation frame information corresponding to the object to be annotated is determined.
  • the above point cloud labeling method when the 3D point cloud data to be labelled is obtained, the corresponding bird's-eye view can be generated first. In this way, after responding to the frame selection operation for the objects to be labelled in the bird's eye view, the first One-dimensional labeling frame information, and at the same time, the initial parameter information of the to-be-labeled object can also be determined in combination with the 3D point cloud data of the frame-selected to-be-labeled object. Information about the 3D callout box corresponding to the callout object. It can be seen that the above point cloud labeling method is a frame selection operation performed in a two-dimensional view (corresponding to a bird's-eye view).
  • the complexity of performing the frame selection operation under the view is low, and the magnified view of the local point cloud facilitates the user to accurately identify the boundary of the labeling box, which can ensure the accuracy of the framed object to be labelled to a certain extent.
  • the obtained first two-dimensional labeling frame information and initial parameter information can realize automatic labeling of the three-dimensional labeling frame for the object to be labelled, and the labeling efficiency and accuracy are high.
  • the method further includes:
  • a view to be adjusted corresponding to the three-dimensional point cloud data to be marked is generated;
  • the to-be-adjusted view includes at least one of a front view and a side view;
  • second two-dimensional labeling frame information of the to-be-labeled object in the to-be-adjusted view is determined.
  • the determining, based on the first 2D annotation frame information and the initial parameter information, the 3D annotation frame information corresponding to the object to be annotated includes:
  • the 3D annotation frame information corresponding to the object to be annotated is determined.
  • the view to be adjusted and the bird's-eye view are two different 2D views that belong to the same 3D view where the 3D point cloud data is located, there is a correlation of parameter information between the two different 2D views.
  • the second two-dimensional annotation frame information determined by the view and the parameter information determined by the first two-dimensional annotation frame information determined by the bird's-eye view more accurate three-dimensional annotation frame information can be determined.
  • the 3D annotation frame information corresponding to the object to be annotated is determined based on the first 2D annotation frame information, the second 2D annotation frame information and the initial parameter information ,include:
  • the 3D annotation frame information corresponding to the object to be annotated is determined.
  • the initial parameter information can be adjusted based on the adjustment operation for the second two-dimensional annotation frame information of the object to be labeled in the view to be adjusted, so that the adjusted parameter information and the first parameter corresponding to the bird's-eye view can be adjusted.
  • One 2D annotation frame information and second 2D annotation frame information corresponding to the view to be adjusted determine the 3D annotation frame information corresponding to the to-be-annotated object. This may be due to the fact that different views focus on different information about the object to be marked.
  • the corresponding display may be the length and width of the object to be marked, and for the front view, the corresponding display The display can be the width and height of the object to be marked. Therefore, on the basis of the initial parameter information, parameter adjustment is performed in combination with the view to be adjusted, which will further improve the accuracy of the determined 3D annotation frame information.
  • determining the adjusted parameter information of the object to be annotated including:
  • the initial parameter information is adjusted based on the parameter adjustment information to obtain the adjusted parameter information.
  • the 3D annotation corresponding to the object to be annotated is determined based on the first 2D annotation frame information, the second 2D annotation frame information and the adjusted parameter information Box information, including:
  • the corresponding relationship between the three-dimensional annotation frames is determined on multiple planes, and the information of the three-dimensional annotation frames is determined by combining the information.
  • the method further includes:
  • the to-be-adjusted view marked with a second two-dimensional annotation frame corresponding to the second two-dimensional annotation frame information is displayed.
  • the method further includes:
  • mapping the coordinate system corresponding to the three-dimensional point cloud data to be marked and the coordinate system corresponding to the verification image Based on the mapping relationship between the coordinate system corresponding to the three-dimensional point cloud data to be marked and the coordinate system corresponding to the verification image, map the three-dimensional annotation frame information corresponding to the object to be marked into the verification image , determine the coordinate range of the image after mapping;
  • the accuracy of the three-dimensional annotation frame information is verified according to the mapped image coordinate range and the image coordinate range in the verification image where the object to be marked is located.
  • the accuracy of the 3D annotation frame information can be verified based on the obtained verification image, which may depend on the coordinate system corresponding to the 3D point cloud data to be annotated.
  • the mapping relationship between the coordinate systems corresponding to the verification image is realized by mapping the image coordinate range, so that the specific object to be marked can be verified more accurately.
  • the accuracy of the three-dimensional annotation frame information is checked. After verification, it also includes:
  • the adjustment operation of the two-dimensional annotation frame is to adjust the adjusted parameter information until the image coordinate range after the mapping is consistent with the image coordinate range where the object to be annotated in the verification image is located; the adjusted image coordinate range is The parameter information is obtained by adjusting the initial parameter information.
  • the method further includes:
  • the 3D annotation frame information of the marked object determine the 3D annotation frame information of other objects to be marked in the 3D view; the other objects to be marked and the marked object belong to the same object type;
  • the 3D annotation frame information of the other objects to be annotated is displayed in the 3D view.
  • the determining, according to the 3D annotation frame information of the marked object, the 3D annotation frame information of other objects to be annotated in the 3D view includes:
  • the 3D annotation frame information of the marked object can be pasted to the On other to-be-labeled objects belonging to the same object type, three-dimensional labeling of other to-be-labeled objects can be quickly realized, and the labeling efficiency is further improved.
  • the method further includes:
  • determining the first two-dimensional frame information in response to a frame selection operation on the object to be marked in the bird's-eye view, includes:
  • the reference image and the to-be-labeled 3D point cloud data originate from the same scene;
  • the first two-dimensional frame information is determined based on the reference position information.
  • the reference position information of the object to be marked in the bird's-eye view can be determined based on the obtained reference image, and based on this reference position information, the user can be instructed to perform a frame selection operation for the object to be marked in the bird's-eye view.
  • the resulting first two-dimensional labeling frame will cover the point cloud data corresponding to the object to be labelled to a greater extent, thereby improving the labeling accuracy.
  • the initial parameter information includes at least one of initial height information and initial pose information.
  • an embodiment of the present disclosure further provides a point cloud labeling device, the device comprising:
  • an acquisition module configured to acquire the 3D point cloud data to be marked
  • a generating module configured to generate a bird's-eye view corresponding to the three-dimensional point cloud data to be marked based on the three-dimensional point cloud data to be marked;
  • a first determination module configured to determine first two-dimensional labeling frame information in response to a frame selection operation for objects to be labelled in the bird's-eye view; The initial parameter information of the annotation object;
  • the labeling module is configured to determine the three-dimensional labeling frame information corresponding to the object to be labelled based on the first two-dimensional labeling frame information and the initial parameter information.
  • embodiments of the present disclosure further provide an electronic device, including: a processor, a memory, and a bus, where the memory stores machine-readable instructions executable by the processor, and the processor is configured to execute all The machine-readable instructions stored in the memory, when the electronic device is running, the processor communicates with the memory through a bus, and the machine-readable instructions are executed by the processor.
  • an electronic device including: a processor, a memory, and a bus, where the memory stores machine-readable instructions executable by the processor, and the processor is configured to execute all The machine-readable instructions stored in the memory, when the electronic device is running, the processor communicates with the memory through a bus, and the machine-readable instructions are executed by the processor.
  • an embodiment of the present disclosure further provides a computer-readable storage medium, where a computer program is stored on the computer-readable storage medium, and when the computer program is run by an electronic device, the electronic device executes the first Aspects and steps of the method for point cloud labeling described in any one of its various embodiments.
  • embodiments of the present disclosure further provide a computer program product, including a computer-readable storage medium storing program codes, where the program codes include instructions that can be configured to execute any of the first aspect and its various implementations. 1. The steps of the method for point cloud labeling.
  • FIG. 1 shows a schematic flowchart of a method for labeling a point cloud provided by an embodiment of the present disclosure
  • FIG. 2 shows a schematic flowchart of a point cloud labeling provided by an embodiment of the present disclosure
  • FIG. 3 shows a schematic structural diagram of a device for point cloud labeling provided by an embodiment of the present disclosure
  • FIG. 4 shows a schematic structural diagram of an electronic device provided by an embodiment of the present disclosure.
  • the present disclosure provides at least one point cloud labeling solution, which realizes the determination of 3D labeling frame information based on the frame selection operation of the 2D bird's-eye view, which is simple to operate, and has high labeling efficiency and accuracy.
  • Lidar is widely used in today's autonomous driving systems. It can provide accurate three-dimensional environmental spatial information to assist the system's scene understanding and decision-making process.
  • many deep learning-based perceptual algorithms have emerged to process such data, and these algorithms are significantly superior to monocular and stereo vision methods in application.
  • the rapid development of these algorithms is supported by several challenging benchmarks built on multiple open datasets.
  • actual product deployment still requires more accurately labeled data to feed algorithms.
  • publicly available annotation tools are crude, which limits research progress in this field.
  • Equipment the electronic equipment for example includes: terminal equipment or server or other processing equipment
  • the terminal equipment can be user equipment (User Equipment, UE), mobile equipment, user terminal, terminal, cellular phone, cordless phone, Personal Digital Processing (Personal Digital Assistant, PDA), handheld devices, computing devices, in-vehicle devices, wearable devices, etc.
  • the point cloud labeling method can be implemented by the processor calling computer-readable instructions stored in the memory.
  • the following describes the point cloud labeling method provided by the embodiment of the present disclosure by taking the execution subject as a terminal device as an example.
  • Fig. 1 shows a schematic flowchart of a point cloud labeling method provided by an embodiment of the present disclosure.
  • the method includes steps S101 to S105, and the method can be applied to a point cloud labeling device or a point cloud labeling device , the point cloud labeling device may be a processor or a chip, and the processor or chip may be used in the point cloud labeling device, wherein the method includes:
  • the point cloud labeling method provided by the embodiments of the present disclosure can be applied to any scene that requires 3D labeling of the object to be labelled. Based on the 3D labeling frame information after 3D labeling, training related to target detection and following models can be implemented, and can also be applied In other related application scenarios that require 3D annotation frame information, for example, it can be applied to automatic driving vehicles to identify the information of vehicles ahead to avoid traffic hazards such as vehicle collisions. In addition, it can also be applied to other application scenarios. , there is no restriction here.
  • the related technology can be manually operated in three-dimensional space for labeling, and because the three-dimensional point cloud data has more labeling degrees of freedom, for example, the point cloud of an object can be marked with its corresponding height and rotation. Angle and other information, the operation complexity is relatively high. Considering that for sparse and irregularly distributed point cloud data, it is difficult to correctly identify the object of interest, which increases the difficulty of labeling to a certain extent and reduces the efficiency of labeling.
  • the embodiments of the present disclosure provide a point cloud labeling method, which realizes the determination of 3D labeling frame information based on the frame selection operation of the 2D bird's-eye view.
  • the operation is simple, and the labeling efficiency and accuracy are high. .
  • the three-dimensional point cloud data to be labeled in the embodiment of the present disclosure may be point cloud data collected by a radar device.
  • the radar device here may use a rotary scanning laser radar, or other radar devices, which are not specifically limited.
  • the lidar can obtain the relevant three-dimensional point cloud data in the surrounding environment when it rotates and scans in the horizontal direction.
  • the lidar in the process of rotating scanning, can use a multi-line scanning method, that is, the emission uses multiple laser tubes to emit sequentially, and the structure is that multiple laser tubes are arranged longitudinally, that is, in the process of rotating and scanning in the horizontal direction, vertical scanning is performed. Orientation of multi-layer scanning.
  • each laser tube There is a certain angle between each laser tube, and the vertical emission field can be 30° to 40°. In this way, a data packet returned by multiple laser tubes can be obtained for each rotation of a scanning angle of the lidar device.
  • One frame of point cloud data can be obtained by splicing the data packets obtained from each scanning angle (corresponding to a 360° scan of one rotation), and one frame of point cloud data can be collected after one scan of the lidar.
  • 3D point cloud data to be marked in the embodiments of the present disclosure may refer to a frame of point cloud data, a half-frame point cloud data, or point cloud data determined by other scanning ranges. This does not make specific restrictions.
  • the determination of the three-dimensional labeling frame information corresponding to the to-be-labeled object may be implemented based on the frame selection operation of the to-be-labeled object in the two-dimensional bird's-eye view.
  • the three-dimensional point cloud data to be marked can be converted to the bird's-eye view, and the bird's-eye view can be is the mapping result of the three-dimensional point cloud data on the top view, which can correspond to the first two-dimensional point cloud data.
  • frame selection can be performed for the object to be marked in the bird's eye view. operation, so as to determine the corresponding first two-dimensional callout frame information.
  • Initial parameter information where the initial parameter information may be initial height information, or may be initial pose information, or may be a combination of initial height information and initial pose information, which is not specifically limited in this embodiment of the present disclosure .
  • the relevant initial height information may be determined by the difference between the coordinate information of the highest point cloud point and the coordinate information of the lowest point cloud point of the 3D point cloud data corresponding to the frame selection, and the relevant initial pose information may include the initial Position information and initial attitude information.
  • the initial position information can be determined based on the relative position of the 3D point cloud data selected by the frame in the 3D point cloud data to be marked, and the initial attitude information can be based on the two sides corresponding to the frame selection. The angle between the point cloud point set and the reference direction is determined.
  • the 3D annotation frame information of the to-be-annotated object can be determined.
  • the 3D annotation frame information here can be the initial 3D annotation frame information. Adjust the frame to synchronize the adjustment.
  • a corresponding bird's-eye view can be generated first, so that after responding to the frame selection operation for the objects to be labeled in the bird's-eye view, the first
  • the two-dimensional labeling frame information and at the same time, the initial parameter information of the to-be-labeled object can also be determined in combination with the 3D point cloud data of the frame-selected object to be labeled.
  • Information about the 3D callout box corresponding to the object is a frame selection operation performed in a two-dimensional view (corresponding to a bird's-eye view).
  • the complexity of performing the frame selection operation under the view is low, and the magnified view of the local point cloud facilitates the user to accurately identify the boundary of the labeling box, which can ensure the accuracy of the framed object to be labelled to a certain extent.
  • the obtained first two-dimensional labeling frame information and initial parameter information can realize automatic labeling of the three-dimensional labeling frame for the object to be labelled, and the labeling efficiency and accuracy are high.
  • the 3D annotation frame indicated by the 3D annotation frame information has different information on the object to be annotated observed in different views, for example, for the top view, what is observed may be the length and width of the object to be annotated, and for the front view. In other words, what is observed can be the width and height of the object to be annotated. Therefore, in order to further improve the accuracy of the annotation, the 2D annotation frame information in other views can be combined to determine the 3D annotation frame information, for example, the following steps can be included:
  • Step 1 Generate a view to be adjusted corresponding to the three-dimensional point cloud data to be labeled based on the three-dimensional point cloud data of the object to be labeled; the view to be adjusted includes at least one of a front view and a side view;
  • Step 2 According to the generated view to be adjusted, determine the second two-dimensional labeling frame information of the object to be labeled in the view to be adjusted.
  • the view to be adjusted may be generated before the second two-dimensional frame information in the view to be adjusted is determined.
  • the second two-dimensional annotation frame information here may be a mapping result of the three-dimensional point cloud data corresponding to the first two-dimensional annotation frame information on the view to be adjusted.
  • the view to be adjusted in the embodiment of the present disclosure may be a front view, a side view, or a combination of the above two types of views.
  • the generation process of the front view and the side view of the views to be adjusted may refer to the above-mentioned generation process of the two-dimensional bird's-eye view, that is, by using the conversion relationship between the corresponding coordinate systems.
  • the front view and the side view waiting to be adjusted can be understood as different perspectives for observing the three-dimensional view, and the three-dimensional view can be more accurately analyzed under different perspectives to a certain extent.
  • Download the 3D annotation frame information corresponding to the object to be annotated may be determined based on the first 2D annotation frame information, the second 2D annotation frame information, and the initial parameter information.
  • the view to be adjusted and the bird's-eye view are two different 2D views that belong to the same 3D view where the 3D point cloud data is located, there is a correlation of parameter information between the two different 2D views.
  • more accurate three-dimensional annotation frame information can be determined to a certain extent by combining the second two-dimensional annotation frame information determined by the view to be adjusted and the parameter information determined by the first two-dimensional annotation frame information determined by the bird's-eye view.
  • determining the 3D annotation frame information corresponding to the object to be annotated based on the first 2D annotation frame information, the second 2D annotation frame information and the initial parameter information can be achieved by the following steps:
  • Step 1 In response to the adjustment operation for the second two-dimensional labeling frame information, based on the initial parameter information, determine the adjusted parameter information of the object to be labelled;
  • Step 2 Based on the first two-dimensional annotation frame information, the second two-dimensional annotation frame information, and the adjusted parameter information, determine the three-dimensional annotation frame information corresponding to the object to be annotated.
  • the initial parameter information can be adjusted based on the adjustment operation for the second two-dimensional annotation frame information of the object to be labeled in the view to be adjusted.
  • the corresponding first 2D annotation frame information and the second 2D annotation frame information corresponding to the view to be adjusted are used to determine the 3D annotation frame information corresponding to the object to be annotated. This may be due to the fact that different views focus on different information about the object to be marked. For example, for the bird's-eye view, the corresponding display may be the length and width of the object to be marked, and for the front view, the corresponding display The display can be the width and height of the object to be marked. Therefore, on the basis of the initial parameter information, parameter adjustment is performed in combination with the view to be adjusted, which will further improve the accuracy of the determined 3D annotation frame information.
  • determining the adjusted parameter information of the object to be annotated based on the initial parameter information in response to the adjustment operation for the second two-dimensional annotation frame information includes: responding to the adjustment operation for the to-be-annotated object. Adjusting the adjustment operation of the second two-dimensional callout frame in the view to determine parameter adjustment information corresponding to the adjustment operation; and adjusting the initial parameter information based on the parameter adjustment information to obtain the adjusted parameter information.
  • an adjustment operation can be performed on the information of the second two-dimensional annotation frame, that is, the initial parameter information can be adjusted on the basis of the initial parameter information to obtain the adjusted parameter information of the object to be annotated.
  • the updated 3D annotation frame information can be obtained from the 2D annotation frame information and the second 2D annotation frame information.
  • the first two-dimensional annotation frame information may also be adjusted correspondingly.
  • the parameter of height information is adjusted for the second 2D annotation frame information
  • the corresponding The height information of the first 2D annotation frame information can also change accordingly, and the adjustment in the 2D view can be adaptively mapped to the 3D annotation frame indicated by the 3D annotation frame information.
  • the view to be adjusted may be displayed first.
  • the second two-dimensional frame marked on the view can be adjusted.
  • parameter adjustment information corresponding to the adjustment operation may be determined. For example, after dragging the second two-dimensional callout box to stretch outward, the parameter adjustment information can be determined according to the stretching degree. The greater the stretching degree, the greater the value of the parameter adjustment information. The parameter information can determine the adjusted parameter information.
  • determine the At least one of the parameter information of the side and front of the 3D annotation frame corresponding to the annotation object so that the determined parameter information of the bottom surface of the 3D annotation frame and at least one of the parameter information of the side and front of the 3D annotation frame Combining them, you can get the 3D annotation frame information.
  • the 2D point cloud data corresponding to the first 2D annotation frame information, and the 2D points corresponding to the second 2D annotation frame information determines the parameter information of the 3D annotation frame on multiple planes, and then determines the information of the 3D annotation frame through the combination of information.
  • the following step may be further performed: based on the determined second two-dimensional annotation of the object to be annotated Dimensional annotation frame information, displaying the to-be-adjusted view marked with a second 2D annotation frame corresponding to the second 2D annotation frame information.
  • the method for point cloud labeling provided by the embodiment of the present disclosure can be combined with a check image to check the labelled 3D label frame information, and can be implemented by the following steps:
  • Step 1 Obtain a verification image; the verification image and the 3D point cloud data to be marked originate from the same scene;
  • Step 2 Based on the mapping relationship between the coordinate system corresponding to the three-dimensional point cloud data to be marked and the coordinate system corresponding to the verification image, map the three-dimensional marked frame information corresponding to the object to be marked into the verification image, and determine the mapped image. image coordinate range;
  • Step 3 Verify the accuracy of the three-dimensional annotation frame information according to the mapped image coordinate range and the image coordinate range where the object to be marked in the verification image is located.
  • the accuracy of the 3D annotation frame information can be verified based on the obtained verification image, which may depend on the corresponding 3D point cloud data to be marked.
  • the mapping relationship between the coordinate system of the image and the coordinate system corresponding to the verification image is realized by mapping the image coordinate range, so that the specific object to be marked can be verified more accurately.
  • the verification image here may be a related image from the same scene as the 3D point cloud data to be marked, for example, an RGB image.
  • the verification image here may be collected by a camera device installed on the same vehicle as the radar device, and the camera range of the camera device may be consistent with or overlap with the scan range of the radar device.
  • the 3D labeling frame information corresponding to the object to be labelled in the world coordinate system can be mapped to the image coordinate system to obtain
  • the mapped image coordinate range in the case of determining the image coordinate range where the object to be marked in the verification image is located, can be based on the matching result between the image coordinate range and the mapped image coordinate range. Check for accuracy.
  • the point cloud labeling method can respond to the adjustment operation of the second two-dimensional labeling frame in the view to be adjusted corresponding to the to-be-labeled three-dimensional point cloud data when the two coordinate ranges are inconsistent.
  • the adjusted parameter information is adjusted until the image coordinate range after mapping is consistent with the image coordinate range where the object to be marked in the verification image is located. Adjusted to further improve the accuracy of the marked 3D annotation frame information.
  • the point cloud labeling method provided by the embodiment of the present disclosure can not only realize the calibration of the three-dimensional labeling frame information based on the verification image, but also provide a reference for the frame selection operation of the object to be labelled in the bird's-eye view based on the reference image, so as to determine the first and second Dimensional callout box information.
  • the reference image and the verification image here may be the same image, and the above description may be referred to for the process of collecting the reference image and the like.
  • the image position information of the object to be labeled in the reference image can be converted into the reference position information of the object to be labeled in the bird's-eye view, depending on Due to the reference position information, users can perform the frame selection operation of the objects to be marked in the bird's-eye view, especially for sparse 3D point clouds, the reference position information provided by the reference image can greatly improve the quality of the objects to be marked in the bird's-eye view. accuracy.
  • the embodiments of the present disclosure provide an inter-object transfer mechanism to realize the 3D annotation of unlabeled objects based on the 3D annotation frame information of the labeled objects, which can be implemented according to the following steps:
  • Step 1 Display the 3D labeling frame information corresponding to the object to be labelled in the 3D view corresponding to the 3D point cloud data to be labelled, and use the object to be labelled as the labelled object;
  • Step 2 According to the 3D annotation frame information of the marked object, determine the 3D annotation frame information of other objects to be marked in the 3D view; the other objects to be marked belong to the same object type as the marked object;
  • Step 3 Display the 3D labeling frame information of other objects to be labelled in the 3D view.
  • the object to be annotated can be regarded as a marked object.
  • the annotated object belongs to the 3D annotation frame information of other objects to be annotated of the same object type, and the 3D annotation frame information of other objects to be annotated can be displayed in the 3D view.
  • Step 1 in response to the copy instruction for the 3D annotation frame of the marked object in the 3D view, copy the 3D annotation frame information of the marked object;
  • Step 2 In response to the pasting instructions for other objects to be marked in the 3D view, according to the 3D annotation frame information of the marked objects in the 3D view, perform 3D annotation on the other objects to be annotated, and obtain the 3D annotation frame information of the other objects to be annotated. .
  • the 3D annotation frame of the marked object can be copied based on the copy operation of the 3D annotation frame of the marked object.
  • the information is pasted to other objects to be labeled that belong to the same object type, so that 3D labeling of other objects to be labeled can be quickly realized, and the labeling efficiency is further improved.
  • the 3D annotation frame information of the annotated object is copied by copying the 3D annotation frame of the annotated object.
  • the 3D annotation corresponding to the annotated object can be copied according to the
  • the frame information performs 3D labeling on other objects to be labelled to obtain corresponding 3D labeling frame information.
  • the three-dimensional labeling of each vehicle on the road can be implemented based on the above-mentioned inter-object transfer mechanism.
  • the farther the object to be labeled, the sparser the corresponding 3D point cloud data, and the lower the corresponding data integrity, and the object to be labeled with higher data integrity will be more accurate to a certain extent than the object to be labeled with lower data integrity. Easy to label.
  • the embodiment of the present disclosure can perform 3D annotation of the same object type according to the annotation method of distance, which can significantly reduce the adjustment of pose and height and other information during annotation, so as to ensure that one frame of 3D point cloud data to be annotated On the premise of the accuracy of labeling, the labeling efficiency is improved.
  • the embodiments of the present disclosure provide an inter-frame transfer mechanism to realize the 3D labeling of unlabeled objects based on the 3D labeling frame information of the labeled objects, which can be implemented according to the following steps:
  • Step 1 Display the 3D labeling frame information corresponding to the object to be labelled in the first 3D view corresponding to the 3D point cloud data to be labelled in the current frame, and use the object to be labelled as the labelled object; wherein, the 3D points to be labelled in different frames Cloud data corresponds to different 3D views;
  • Step 2 in response to the copy instruction for the 3D annotation frame of the marked object in the first 3D view, copy the 3D annotation frame information of the marked object;
  • Step 3 In the case of switching to the second 3D view corresponding to the 3D point cloud data to be labeled in the next frame, in response to the paste instruction for other objects to be labeled in the second 3D view, according to the labeled objects in the first 3D view.
  • the 3D labeling frame information in the second 3D view is 3D labelled for other objects to be labelled in the second 3D view to obtain the 3D labelling frame information of the other objects to be labelled.
  • the object to be annotated can be regarded as the marked object.
  • the marked object can be copied through the operation of copying the 3D annotation frame of the marked object. Copy the 3D annotation frame information of the object.
  • the 3D annotation frame corresponding to the marked object can be copied.
  • the information is three-dimensional marked on other objects to be marked to obtain the corresponding three-dimensional marked frame information.
  • the pairing operation based on the above copy-paste can be used to realize the next frame of the 3D point cloud data to be labeled. Annotation of the object to be annotated.
  • the to-be-labeled objects of the labeled adjacent frames may be the same object or may not be the same object.
  • it can also be combined with the target tracking or target detection mechanism in the case of switching to the 3D view where the 3D point cloud data to be marked in the next frame is located. Automatically label the objects to be labeled that belong to the same object as the labeled object in the previous frame in the 3D point cloud data to be labeled in the next frame.
  • an annotation transfer mechanism includes an inter-object transfer mechanism. mechanism and inter-frame transfer mechanism.
  • the inter-object transfer mechanism can be: in a single frame, annotations can be transferred between objects from the same class, especially on the same lane.
  • the inter-frame transfer mechanism can be that, in successive frames, annotations from previously marked frames can be transferred. Both transfer mechanisms can significantly reduce the adjustment of rotation angle and height during labeling, avoiding missing or inaccurate labels in sparse regions of the point cloud.
  • the inter-frame transfer mechanism can also realize one-to-one correspondence between the labels of consecutive frames, so that labels can be used for both 3D object detection and 3D object tracking.
  • the inter-object transfer mechanism and the inter-frame transfer mechanism through the inter-object transfer mechanism and the inter-frame transfer mechanism, most 3D labeling tasks can be simplified into labeling adjustments for bird's-eye views, which greatly reduces unnecessary operations and has little impact on the load of electronic devices. The labeling efficiency and accuracy are greatly improved.
  • determining the first two-dimensional annotation frame information in response to the frame selection operation of the object to be annotated in the bird's-eye view in S103 may be implemented by: acquiring a reference image; The three-dimensional point cloud data to be marked comes from the same scene; based on the image position information of the object to be marked in the reference image and the relationship between the coordinate system corresponding to the reference image and the coordinate system corresponding to the bird's-eye view mapping relationship, determine the reference position information of the object to be marked in the bird's-eye view; in response to the frame selection operation for the object to be marked in the bird's-eye view, determine the first and second based on the reference position information Dimensional callout box information.
  • the reference position information of the object to be marked in the bird's-eye view can be determined based on the obtained reference image, and based on the reference position information, the user is guided to perform a frame selection operation for the object to be marked in the bird's-eye view.
  • the first two-dimensional labeling frame selected by the frame will cover the point cloud data corresponding to the object to be labelled to a greater extent, thereby improving the labeling accuracy.
  • the point cloud labeling method provided by the embodiments of the present disclosure can realize view display in combination with a user interface.
  • the corresponding view switching button can be set on the user interaction interface to realize the switching between the bird's-eye view and the three-dimensional view mode.
  • the first 2D annotation frame information is determined, and in the view of switching to the 3D view, the marked 3D annotation frame information can be displayed.
  • the embodiment of the present disclosure can also display the corresponding view to be adjusted on the user interface, and in the view to be adjusted, the second two-dimensional annotation frame can be adjusted, the information of the second two-dimensional annotation frame can be determined, and The initial parameter information of the object to be annotated is adjusted.
  • This adjustment operation can be synchronized under the bird's-eye view and the three-dimensional view, which provides a guarantee for fast annotation.
  • the local point cloud that needs to be operated can be highlighted.
  • the point cloud corresponding to the frame selection can be highlighted on the bird's-eye view displayed on the user interface, and the point cloud corresponding to the view to be adjusted can also be highlighted simultaneously, so as to facilitate the adjustment operation. .
  • the embodiments of the present disclosure may also mark relevant length, width and other information for each markup frame.
  • a lightweight WAN-based labeling tool can be used.
  • the labeling tool can be less dependent on the performance of the computer, so that it can run smoothly on ordinary computers. .
  • the method for point cloud labeling involves complex interactive operations, in order to effectively realize the two technical purposes of correctly identifying objects and accurately labeling objects, in some applications, it can be divided into four steps, namely, Finding, locating, adjusting and verifying, the first two steps can be to identify and locate objects in the global view, the third step is to label accurately, and the last step is to ensure that all labels are accurate enough.
  • Finding, locating, adjusting and verifying the first two steps can be to identify and locate objects in the global view
  • the third step is to label accurately
  • the last step is to ensure that all labels are accurate enough.
  • Each stage focuses on the most important tasks, further ensuring the accuracy of the annotation results.
  • the shape information reflected by the 3D point cloud data and the semantic information in the RGB image can be combined to find the target object. Then, the bounding box is drawn in the bird's-eye view corresponding to the 3D point cloud data, and the height information is generated according to the local point cloud for preliminary positioning. Among them, using the mapping relationship between the coordinate system corresponding to the 3D point cloud data to be marked and the coordinate system corresponding to the RGB image, the 3D pyramid obtained from the 2D image can be used to estimate which points to search for belong to the target object.
  • FIG. 2 shows a schematic flowchart of a point cloud labeling provided by an embodiment of the present disclosure.
  • the three-dimensional point cloud data corresponds to the RGB image
  • the four steps (a), (b), (c) and (d) can be processed in turn for the RGB image and the bird's-eye view. These four steps correspond to search, Locate, adjust and verify.
  • the position of the closer object (the closer object, the closer object occupies a larger size in the RGB image) can be directly determined from the bird's eye view in the bird's eye view, and the farther
  • the position of the object (the farther object, the farther object occupies a smaller size in the RGB image) is not easy to determine in the bird's eye view, so the position of the farther object in the RGB image can be determined, based on the RGB image.
  • the position of distant objects which determines the position of distant objects in the bird's eye view.
  • the positions of the two objects when the positions of the two objects are easy to be determined in the bird's eye view, for example, the two objects are close, the positions of the two objects can be directly determined from the bird's eye view, or, in the case of the two objects
  • the positions of the two objects in the bird's eye view can be determined based on the positions of the two objects in the RGB image.
  • the position of the target After obtaining the position of the target in the bird's-eye view, the position of the target can be located in (b) to obtain the above-mentioned first two-dimensional labeling frame information.
  • the bird's-eye view shown in (a) and the bird's-eye view shown in (b) are bird's-eye views from different viewing angles. After that, perform the conversion between (b) and (c), obtain the position of the target in the bird's-eye view through (b), and adjust the size of the two sides of the target.
  • the two sides can be the front and the side respectively. , any two of the top-view planes, so as to obtain two side dimensions corresponding to each target, as shown in (c). In other embodiments, the dimensions of the three sides may be adjusted.
  • the 3D annotation frame information of a certain object can be verified. For example, in (d), the 3D annotation frame information of the closer object in the RGB image can be determined according to the two side dimensions of the closer object.
  • light-weight web-based annotation tools can be used to realize annotation, and the above annotation tools can run smoothly on ordinary computers. This design is very user-friendly and provides the possibility to perform in the order of each stage.
  • the writing order of each step does not mean a strict execution order and constitutes any limitation on the implementation process, and the execution order of each step should be based on its function and possible inherent Logical OK.
  • the embodiment of the present disclosure also provides a point cloud labeling device corresponding to the point cloud labeling method, because the principle of solving the problem by the device in the embodiment of the present disclosure is the same as the above-mentioned point cloud labeling method in the embodiment of the present disclosure. Similar, so the implementation of the device can refer to the implementation of the method, and the repeated places will not be described in detail.
  • a systematic labeling process is proposed to solve two key problems in 3D labeling tasks: correctly identifying objects and accurately labeling objects.
  • it provides a clear user interface and annotation delivery mechanism, which makes it easier for annotators to concentrate on simpler tasks at each stage of annotation and complete them with fewer operations.
  • FIG. 3 shows a schematic structural diagram of a point cloud labeling device provided by an embodiment of the present disclosure.
  • the device includes: an acquisition module 201 , a generation module 202 , a first determination module 203 , and a labeling module 204 ; in,
  • the obtaining module 201 is configured to obtain the three-dimensional point cloud data to be marked
  • the generating module 202 is configured to generate a bird's-eye view corresponding to the three-dimensional point cloud data to be marked based on the three-dimensional point cloud data to be marked;
  • the first determination module 203 is configured to determine the first two-dimensional annotation frame information in response to the frame selection operation for the object to be marked in the bird's-eye view; initial parameter information;
  • the labeling module 204 is configured to determine the three-dimensional labeling frame information corresponding to the object to be labelled based on the first two-dimensional labeling frame information and the initial parameter information.
  • the frame selection operation performed in the embodiment of the present disclosure in a two-dimensional view compared with the problem of high operational complexity in directly annotating 3D point cloud data in the related art, performing frame selection in a two-dimensional view
  • the complexity of the operation is low, and to a certain extent, the accuracy of the selected objects to be marked can be ensured.
  • the first two-dimensional marked frame information determined based on the frame selection can be combined with the determined initial parameter information.
  • the object is automatically marked with a 3D labeling frame, and the labeling efficiency and accuracy are high.
  • the above device further includes:
  • the second determination module 205 is configured to generate, based on the three-dimensional point cloud data of the object to be marked, a view to be adjusted corresponding to the three-dimensional point cloud data to be marked; the view to be adjusted includes at least one of a front view and a side view; Adjusting the view, and determining the second two-dimensional labeling frame information of the object to be labelled in the view to be adjusted.
  • the labeling module 204 is configured to determine the three-dimensional labeling frame information corresponding to the object to be labelled based on the first two-dimensional labeling frame information and the initial parameter information according to the following steps:
  • the 3D annotation frame information corresponding to the object to be annotated is determined.
  • the labeling module 204 is configured to determine the three-dimensional labeling frame information corresponding to the object to be labelled based on the first two-dimensional labeling frame information, the second two-dimensional labeling frame information and the initial parameter information, including:
  • the 3D annotation frame information corresponding to the object to be annotated is determined.
  • the labeling module 204 is configured to determine the adjusted parameter information of the object to be labelled according to the following steps:
  • the initial parameter information is adjusted based on the parameter adjustment information to obtain adjusted parameter information.
  • the labeling module 204 is configured to determine the 3D labeling frame corresponding to the object to be labelled based on the first 2D labeling frame information, the second 2D labeling frame information and the adjusted parameter information according to the following steps information:
  • the second determining module 205 is further configured to:
  • the above device further includes:
  • the verification module 206 is configured to obtain a verification image after determining the three-dimensional annotation frame information corresponding to the object to be marked; the verification image and the three-dimensional point cloud data to be marked originate from the same scene; based on the coordinates corresponding to the three-dimensional point cloud data to be marked system, the mapping relationship between the coordinate system corresponding to the verification image, map the three-dimensional annotation frame information corresponding to the object to be marked into the verification image, and determine the coordinate range of the image after mapping; The image coordinate range where the object to be annotated in the verification image is located, and the accuracy of the three-dimensional annotation frame information is verified.
  • the verification module 206 is further configured to:
  • the coordinate range of the image after mapping and the object to be marked in the verification image are verified.
  • the image coordinate ranges are inconsistent
  • in response to the adjustment operation for the second two-dimensional annotation frame in the to-be-adjusted view corresponding to the three-dimensional point cloud data to be marked adjust the adjusted parameter information until the mapped image coordinate range It is consistent with the image coordinate range where the object to be marked in the verification image is located; the adjusted parameter information is obtained by adjusting the initial parameter information.
  • the labeling module 204 is further configured to:
  • the 3D annotation frame information corresponding to the object to be annotated After determining the 3D annotation frame information corresponding to the object to be annotated, display the 3D annotation frame information corresponding to the object to be annotated in the 3D view corresponding to the 3D point cloud data to be annotated, and use the object to be annotated as an annotated object;
  • the 3D annotation frame information of the object determines the 3D annotation frame information of other objects to be annotated in the 3D view; other objects to be annotated and the marked object belong to the same object type; the 3D annotation frame information of other objects to be annotated is displayed in the 3D view .
  • the labeling module 204 is configured to determine the 3D labeling frame information of other objects to be labelled in the 3D view according to the 3D labeling frame information of the labelled object:
  • copy the 3D annotation frame information of the marked object In response to the copy instruction for the 3D annotation frame of the marked object in the 3D view, copy the 3D annotation frame information of the marked object; in response to the paste instruction for other objects to be marked in the 3D view, according to the marked object in the 3D view.
  • 3D labeling frame information perform 3D labeling on other objects to be labelled, and obtain 3D labeling frame information of other objects to be labelled.
  • the labeling module 204 is further configured to:
  • the 3D annotation frame information corresponding to the object to be annotated After determining the 3D annotation frame information corresponding to the object to be annotated, display the 3D annotation frame information corresponding to the object to be annotated in the first 3D view corresponding to the 3D point cloud data to be annotated in the current frame, and use the object to be annotated as the marked object ; wherein, the three-dimensional point cloud data to be marked in different frames corresponds to different three-dimensional views; in response to the copy instruction for the three-dimensional marked frame of the marked object in the first three-dimensional view, copy the three-dimensional marked frame information of the marked object; after switching to the next In the case of a second 3D view corresponding to one frame of point cloud data to be annotated, in response to the pasting instruction for other objects to be annotated in the second 3D view, according to the 3D annotation frame information of the annotated object in the first 3D view, The other objects to be marked in the second three-dimensional view are three-dimensionally marked, and the three-dimensional
  • the first determining module 203 is configured to respond to a frame selection operation for the object to be marked in the bird's-eye view according to the following steps, and determine the first two-dimensional frame information:
  • the first two-dimensional frame information is determined based on the reference position information.
  • the initial parameter information includes at least one of initial height information and initial pose information.
  • FIG. 4 shows a schematic structural diagram of an electronic device provided by an embodiment of the present disclosure.
  • the electronic device provided by an embodiment of the present disclosure is a schematic structural diagram, including: a processor 301 , a memory 302 , and a bus 303 .
  • the memory 302 stores machine-readable instructions executable by the processor 301 (for example, the execution instructions corresponding to the acquisition module 201, the generation module 202, the first determination module 203, and the labeling module 204 in the device for point cloud labeling in FIG. 3, etc.) , when the electronic device is running, the processor 301 communicates with the memory 302 through the bus 303, and the machine-readable instructions are executed by the processor 301 to perform the following processing:
  • the three-dimensional labeling frame information corresponding to the object to be labelled is determined.
  • Embodiments of the present disclosure further provide a computer-readable storage medium, where a computer program is stored on the computer-readable storage medium, and when the computer program is run by a processor, the steps of the point cloud labeling method described in the foregoing method embodiments are executed.
  • the storage medium may be a volatile or non-volatile computer-readable storage medium.
  • An embodiment of the present disclosure further provides a computer program product, including a computer-readable storage medium storing program codes, and the instructions included in the program codes can be configured to execute the steps of the point cloud labeling method described in the above method embodiments , for details, refer to the above method embodiments.
  • Embodiments of the present disclosure further provide a computer program, which implements any one of the methods in the foregoing embodiments when the computer program is executed by a processor.
  • the computer program product can be specifically implemented by hardware, software or a combination thereof.
  • the computer program product is embodied as a computer storage medium, and in another optional embodiment, the computer program product is embodied as a software product, such as a software development kit (Software Development Kit, SDK), etc. Wait.
  • the units described as separate components may or may not be physically separated, and components displayed as units may or may not be physical units, that is, may be located in one place, or may be distributed to multiple network units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution in this embodiment.
  • each functional unit in each embodiment of the present disclosure may be integrated into one processing unit, or each unit may exist physically alone, or two or more units may be integrated into one unit.
  • the functions, if implemented in the form of software functional units and sold or used as stand-alone products, may be stored in a processor-executable non-volatile computer-readable storage medium.
  • the computer software products are stored in a storage medium, including Several instructions are used to cause an electronic device (which may be a personal computer, a server, or a network device, etc.) to execute all or part of the steps of the methods described in various embodiments of the present disclosure.
  • the aforementioned storage medium includes: U disk, mobile hard disk, read-only memory (Read-Only Memory, ROM), random access memory (Random Access Memory, RAM), magnetic disk or optical disk and other media that can store program codes .
  • the first two-dimensional labeling frame information is determined by responding to a frame selection operation for the object to be labelled in the bird's-eye view corresponding to the 3D point cloud data to be labelled; Based on the cloud data, the initial parameter information of the object to be marked is determined; based on the first two-dimensional marked frame information and the initial parameter information, the three-dimensional marked frame information corresponding to the to-be-marked object is determined, so that the frame selection can be performed.
  • the obtained first two-dimensional labeling frame information and initial parameter information realize automatic labeling of the three-dimensional labeling frame for the object to be labelled, and the labeling efficiency and accuracy are high.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Multimedia (AREA)
  • Remote Sensing (AREA)
  • Radar, Positioning & Navigation (AREA)
  • Electromagnetism (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • Medical Informatics (AREA)
  • Evolutionary Computation (AREA)
  • Databases & Information Systems (AREA)
  • Computing Systems (AREA)
  • Artificial Intelligence (AREA)
  • General Health & Medical Sciences (AREA)
  • Architecture (AREA)
  • Computer Graphics (AREA)
  • Computer Hardware Design (AREA)
  • General Engineering & Computer Science (AREA)
  • Processing Or Creating Images (AREA)
  • Length Measuring Devices By Optical Means (AREA)

Abstract

A point cloud annotating method and apparatus, an electronic device, a storage medium, and a program product. The method comprises: obtaining three-dimensional point cloud data to be annotated (S101); on the basis of said three-dimensional point cloud data, generating a bird's eye view corresponding to said three-dimensional point cloud data (S102); in response to a frame selection operation for an object to be annotated in the bird's eye view, determining first two-dimensional annotation frame information (S103); determining initial parameter information of said object on the basis of three-dimensional point cloud data of said framed object (S104); and on the basis of the first two-dimensional annotation frame information and the initial parameter information, determining three-dimensional annotation frame information corresponding to said object (S105). The method determines the three-dimensional annotation frame information on the basis of the frame selection operation in the two-dimensional bird's eye view, the operation is simple, and the annotation efficiency and accuracy are high.

Description

点云标注的方法、装置、电子设备、存储介质及程序产品Method, device, electronic device, storage medium and program product for point cloud labeling
相关申请的交叉引用CROSS-REFERENCE TO RELATED APPLICATIONS
本专利申请要求2020年9月30日提交的中国专利申请号为202011061734.2,申请人为上海商汤临港智能科技有限公司,申请名称为“一种点云标注的方法、装置、电子设备及存储介质”的优先权,该申请的全文以引用的方式并入本申请中。This patent application requires the Chinese patent application number 202011061734.2 submitted on September 30, 2020, the applicant is Shanghai Shangtang Lingang Intelligent Technology Co., Ltd., and the application name is "A method, device, electronic device and storage medium for point cloud labeling ”, which is hereby incorporated by reference in its entirety.
技术领域technical field
本公开涉及计算机技术领域,尤其涉及一种点云标注的方法、装置、电子设备、存储介质及程序产品。The present disclosure relates to the field of computer technology, and in particular, to a method, device, electronic device, storage medium and program product for point cloud labeling.
背景技术Background technique
激光雷达被广泛应用于各种技术领域。以自动驾驶领域为例,利用激光雷达采集的点云数据准确检测车辆周围的障碍物,如行人、其它车辆等应用显的尤为重要。Lidar is widely used in various technical fields. Taking the field of autonomous driving as an example, it is particularly important to use the point cloud data collected by lidar to accurately detect obstacles around the vehicle, such as pedestrians and other vehicles.
为了准确检测出障碍物,需要基于采集的点云数据作为训练样本训练检测模型。在训练之前,需要对采集的点云数据进行准确标注以提升检测结果的准确度。In order to accurately detect obstacles, the detection model needs to be trained based on the collected point cloud data as training samples. Before training, it is necessary to accurately label the collected point cloud data to improve the accuracy of the detection results.
目前,对于点云数据的标注,大多依靠人工方式直接在三维空间中操作进行标注。该方式需要通过复杂而不精确的三维空间交互操作来标定待标注对象的位置、大小尺寸、前进方向和类别等训练所需信息,操作步骤复杂,标注效率和准确率均较低。At present, for the annotation of point cloud data, most of them rely on manual methods to directly operate in the three-dimensional space. This method requires complex and inaccurate three-dimensional interactive operations to demarcate the information required for training, such as the position, size, direction, and category of the object to be labeled. The operation steps are complex, and the labeling efficiency and accuracy are low.
发明内容SUMMARY OF THE INVENTION
本公开实施例提供一种点云标注的方法、装置、电子设备、存储介质及程序产品。Embodiments of the present disclosure provide a method, apparatus, electronic device, storage medium, and program product for point cloud labeling.
第一方面,本公开实施例提供了一种点云标注的方法,所述方法包括:In a first aspect, an embodiment of the present disclosure provides a point cloud labeling method, the method comprising:
获取待标注三维点云数据;Obtain the 3D point cloud data to be labeled;
基于所述待标注三维点云数据,生成所述待标注三维点云数据对应的鸟瞰图;generating a bird's-eye view corresponding to the three-dimensional point cloud data to be marked based on the three-dimensional point cloud data to be marked;
响应针对所述鸟瞰图中的待标注对象的框选操作,确定第一二维标注框信息;In response to the frame selection operation for the object to be marked in the bird's-eye view, determine the first two-dimensional frame information;
基于被框选的待标注对象的三维点云数据,确定所述待标注对象的初始参数信息;Determine the initial parameter information of the object to be marked based on the three-dimensional point cloud data of the object to be marked selected by the frame;
基于所述第一二维标注框信息,以及所述初始参数信息,确定所述待标注对象对应的三维标注框信息。Based on the first two-dimensional annotation frame information and the initial parameter information, the three-dimensional annotation frame information corresponding to the object to be annotated is determined.
采用上述点云标注的方法,在获取到待标注三维点云数据的情况下,可以先生成对应的鸟瞰图,这样,在响应针对鸟瞰图中的待标注对象的框选操作之后,可以确定第一二维标注框信息,同时,还可以结合被框选的待标注对象的三维点云数据确定待标注对象的初始参数信息,这样,结合第一二维标注框信息以及初始参数信息可以确定待标注对象对应的三维标注框信息。可知,上述点云标注的方法是在二维视图(对应鸟瞰图)下执行的框选操作,相比相关技术中直接对三维点云数据进行标注所存在的操作复杂度高的问题,二维视图下执行框选操作的复杂度较低,并且局部点云的放大视角便于用户精确地辨识标注框的边界,一定程度上可以确保所框选待标注对象的准确性,这样,结合通过框选得到的第一二维标注框信息和初始参数信息即可以实现针对待标注对象进 行三维标注框的自动标注,标注效率和准确率均较高。Using the above point cloud labeling method, when the 3D point cloud data to be labelled is obtained, the corresponding bird's-eye view can be generated first. In this way, after responding to the frame selection operation for the objects to be labelled in the bird's eye view, the first One-dimensional labeling frame information, and at the same time, the initial parameter information of the to-be-labeled object can also be determined in combination with the 3D point cloud data of the frame-selected to-be-labeled object. Information about the 3D callout box corresponding to the callout object. It can be seen that the above point cloud labeling method is a frame selection operation performed in a two-dimensional view (corresponding to a bird's-eye view). The complexity of performing the frame selection operation under the view is low, and the magnified view of the local point cloud facilitates the user to accurately identify the boundary of the labeling box, which can ensure the accuracy of the framed object to be labelled to a certain extent. In this way, combined with the frame selection The obtained first two-dimensional labeling frame information and initial parameter information can realize automatic labeling of the three-dimensional labeling frame for the object to be labelled, and the labeling efficiency and accuracy are high.
在一种可能的实施方式中,所述方法还包括:In a possible implementation, the method further includes:
基于所述待标注对象的三维点云数据,生成所述待标注三维点云数据对应的待调整视图;所述待调整视图至少包括正视图和侧视图中的一种;Based on the three-dimensional point cloud data of the object to be marked, a view to be adjusted corresponding to the three-dimensional point cloud data to be marked is generated; the to-be-adjusted view includes at least one of a front view and a side view;
根据生成的所述待调整视图,确定所述待标注对象在所述待调整视图中的第二二维标注框信息。According to the generated view to be adjusted, second two-dimensional labeling frame information of the to-be-labeled object in the to-be-adjusted view is determined.
在一种可能的实施方式中,所述基于所述第一二维标注框信息,以及所述初始参数信息,确定所述待标注对象对应的三维标注框信息,包括:In a possible implementation manner, the determining, based on the first 2D annotation frame information and the initial parameter information, the 3D annotation frame information corresponding to the object to be annotated includes:
基于所述第一二维标注框信息、所述第二二维标注框信息以及所述初始参数信息,确定所述待标注对象对应的三维标注框信息。Based on the first 2D annotation frame information, the second 2D annotation frame information and the initial parameter information, the 3D annotation frame information corresponding to the object to be annotated is determined.
这里,考虑到待调整视图和鸟瞰图是同属于三维点云数据所在三维视图下的两个不同的二维视图,两个不同的二维视图之间存在参数信息的关联性,因而,结合待调整视图所确定的第二二维标注框信息和鸟瞰图所确定的第一二维标注框信息所确定参数信息的一定程度上可以确定出更为准确性的三维标注框信息。Here, considering that the view to be adjusted and the bird's-eye view are two different 2D views that belong to the same 3D view where the 3D point cloud data is located, there is a correlation of parameter information between the two different 2D views. To a certain extent, by adjusting the second two-dimensional annotation frame information determined by the view and the parameter information determined by the first two-dimensional annotation frame information determined by the bird's-eye view, more accurate three-dimensional annotation frame information can be determined.
在一种可能的实施方式中,所述基于所述第一二维标注框信息、所述第二二维标注框信息以及所述初始参数信息,确定所述待标注对象对应的三维标注框信息,包括:In a possible implementation manner, the 3D annotation frame information corresponding to the object to be annotated is determined based on the first 2D annotation frame information, the second 2D annotation frame information and the initial parameter information ,include:
响应针对所述第二二维标注框信息的调整操作,基于所述初始参数信息,确定所述待标注对象的调整后的参数信息;In response to the adjustment operation for the second two-dimensional annotation frame information, based on the initial parameter information, determine the adjusted parameter information of the object to be annotated;
基于所述第一二维标注框信息、所述第二二维标注框信息以及所述调整后的参数信息,确定所述待标注对象对应的三维标注框信息。Based on the first 2D annotation frame information, the second 2D annotation frame information and the adjusted parameter information, the 3D annotation frame information corresponding to the object to be annotated is determined.
这里,可以基于针对待标注对象在所述待调整视图中的第二二维标注框信息的调整操作,对初始参数信息进行调整,这样,即可以基于调整后的参数信息、鸟瞰图对应的第一二维标注框信息、待调整视图对应的第二二维标注框信息,确定待标注对象对应的三维标注框信息。这可以是考虑到不同视图所侧重对应展示的待标注对象的相关信息也不同,例如,针对鸟瞰图而言,对应展示的可以是待标注对象的长和宽,而针对前视图而言,对应展示的则可以是待标注对象的宽和高,因而,在初始参数信息的基础上,再结合待调整视图进行参数调整,将进一步提升所确定的三维标注框信息的准确性。Here, the initial parameter information can be adjusted based on the adjustment operation for the second two-dimensional annotation frame information of the object to be labeled in the view to be adjusted, so that the adjusted parameter information and the first parameter corresponding to the bird's-eye view can be adjusted. One 2D annotation frame information and second 2D annotation frame information corresponding to the view to be adjusted determine the 3D annotation frame information corresponding to the to-be-annotated object. This may be due to the fact that different views focus on different information about the object to be marked. For example, for the bird's-eye view, the corresponding display may be the length and width of the object to be marked, and for the front view, the corresponding display The display can be the width and height of the object to be marked. Therefore, on the basis of the initial parameter information, parameter adjustment is performed in combination with the view to be adjusted, which will further improve the accuracy of the determined 3D annotation frame information.
在一种可能的实施方式中,所述响应针对所述第二二维标注框信息的调整操作,基于所述初始参数信息,确定所述待标注对象的调整后的参数信息,包括:In a possible implementation manner, in response to the adjustment operation for the second two-dimensional annotation frame information, based on the initial parameter information, determining the adjusted parameter information of the object to be annotated, including:
响应针对所述待调整视图中的第二二维标注框的调整操作,确定与所述调整操作对应的参数调整信息;determining parameter adjustment information corresponding to the adjustment operation in response to an adjustment operation for the second two-dimensional frame in the to-be-adjusted view;
基于所述参数调整信息对所述初始参数信息进行调整,得到所述调整后的参数信息。The initial parameter information is adjusted based on the parameter adjustment information to obtain the adjusted parameter information.
在一种可能的实施方式中,所述基于所述第一二维标注框信息、所述第二二维标注框信息以及所述调整后的参数信息,确定所述待标注对象对应的三维标注框信息,包括:In a possible implementation manner, the 3D annotation corresponding to the object to be annotated is determined based on the first 2D annotation frame information, the second 2D annotation frame information and the adjusted parameter information Box information, including:
基于所述调整后的参数信息所对应的三维点云数据,与所述第一二维标注框信息所对应的二维点云数据之间的第一对应关系,确定所述待标注对象对应的三维标注框的底面的参数信息;Based on the first correspondence between the three-dimensional point cloud data corresponding to the adjusted parameter information and the two-dimensional point cloud data corresponding to the first two-dimensional labeling frame information, determine the corresponding to-be-labeled object. The parameter information of the bottom surface of the 3D callout box;
基于所述调整后的参数信息所对应的三维点云数据,与所述第二二维标注框信息所对应的二维点云数据之间的第二对应关系,确定所述待标注对象对应的三维标注框的侧面和正面的参数信息中的至少之一;Based on the second correspondence between the three-dimensional point cloud data corresponding to the adjusted parameter information and the two-dimensional point cloud data corresponding to the second two-dimensional labeling frame information, determine the corresponding to-be-labeled object. At least one of the parameter information of the side and the front of the 3D callout frame;
将确定的所述三维标注框的底面的参数信息、以及所述三维标注框的侧面和正面的参数信息中的至少之一进行组合,得到所述三维标注框信息。Combining at least one of the determined parameter information of the bottom surface of the 3D annotation frame and the parameter information of the side surface and the front surface of the 3D annotation frame to obtain the 3D annotation frame information.
这里,可以基于调整后的参数信息所对应的三维点云数据与第一二维标注框信息所对应的二维点云数据、以及第二二维标注框信息所对应的二维点云数据之间的对应关系, 确定三维标注框在多个平面上的参数信息,进而通过信息组合,确定三维标注框信息。Here, it can be based on the difference between the 3D point cloud data corresponding to the adjusted parameter information, the 2D point cloud data corresponding to the first 2D annotation frame information, and the 2D point cloud data corresponding to the second 2D annotation frame information. The corresponding relationship between the three-dimensional annotation frames is determined on multiple planes, and the information of the three-dimensional annotation frames is determined by combining the information.
在一种可能的实施方式中,在所述确定所述待标注对象在所述待调整视图中的第二二维标注框信息之后,所述方法还包括:In a possible implementation manner, after the determining of the second two-dimensional annotation frame information of the object to be annotated in the view to be adjusted, the method further includes:
基于确定的所述待标注对象的第二二维标注框信息,展示标注有与所述第二二维标注框信息对应的第二二维标注框的所述待调整视图。Based on the determined second two-dimensional annotation frame information of the object to be annotated, the to-be-adjusted view marked with a second two-dimensional annotation frame corresponding to the second two-dimensional annotation frame information is displayed.
在一种可能的实施方式中,在所述确定所述待标注对象对应的三维标注框信息之后,所述方法还包括:In a possible implementation manner, after the determining of the three-dimensional annotation frame information corresponding to the object to be annotated, the method further includes:
获取校验图像;所述校验图像与所述待标注三维点云数据来源于同一场景;Obtaining a verification image; the verification image and the to-be-labeled 3D point cloud data originate from the same scene;
基于所述待标注三维点云数据对应的坐标系,与所述校验图像对应的坐标系之间的映射关系,将所述待标注对象对应的三维标注框信息映射至所述校验图像中,确定映射后的图像坐标范围;Based on the mapping relationship between the coordinate system corresponding to the three-dimensional point cloud data to be marked and the coordinate system corresponding to the verification image, map the three-dimensional annotation frame information corresponding to the object to be marked into the verification image , determine the coordinate range of the image after mapping;
根据所述映射后的图像坐标范围与所述校验图像中的所述待标注对象所在图像坐标范围,对所述三维标注框信息的准确性进行校验。The accuracy of the three-dimensional annotation frame information is verified according to the mapped image coordinate range and the image coordinate range in the verification image where the object to be marked is located.
这里,在针对待标注对象确定对应的三维标注框信息之后,可以基于获取的校验图像对三维标注框信息的准确性进行校验,这里可以是依赖于待标注三维点云数据对应的坐标系与校验图像对应的坐标系之间的映射关系,通过图像坐标范围映射来实现,从而可以更为准确的验证到特定的待标注对象。Here, after the corresponding 3D annotation frame information is determined for the object to be annotated, the accuracy of the 3D annotation frame information can be verified based on the obtained verification image, which may depend on the coordinate system corresponding to the 3D point cloud data to be annotated. The mapping relationship between the coordinate systems corresponding to the verification image is realized by mapping the image coordinate range, so that the specific object to be marked can be verified more accurately.
在一种可能的实施方式中,在所述根据所述映射后的图像坐标范围与所述校验图像中的所述待标注对象所在图像坐标范围,对所述三维标注框信息的准确性进行校验之后,还包括:In a possible implementation manner, according to the mapped image coordinate range and the image coordinate range where the object to be annotated in the verification image is located, the accuracy of the three-dimensional annotation frame information is checked. After verification, it also includes:
在所述映射后的图像坐标范围与所述校验图像中的所述待标注对象所在图像坐标范围不一致的情况下,响应针对所述待标注三维点云数据对应的待调整视图中的第二二维标注框的调整操作,对调整后的参数信息进行调整,直至所述映射后的图像坐标范围与所述校验图像中的所述待标注对象所在图像坐标范围一致;所述调整后的参数信息是通过对所述初始参数信息进行调整得到的。In the case where the mapped image coordinate range is inconsistent with the image coordinate range where the object to be marked in the verification image is located, respond to the second image coordinate range in the to-be-adjusted view corresponding to the three-dimensional point cloud data to be marked. The adjustment operation of the two-dimensional annotation frame is to adjust the adjusted parameter information until the image coordinate range after the mapping is consistent with the image coordinate range where the object to be annotated in the verification image is located; the adjusted image coordinate range is The parameter information is obtained by adjusting the initial parameter information.
在一种可能的实施方式中,所述确定任所述待标注对象对应的三维标注框信息之后,还包括:In a possible implementation manner, after determining the three-dimensional labeling frame information corresponding to any of the objects to be labelled, the method further includes:
将所述待标注对象对应的三维标注框信息,展示在所述待标注三维点云数据对应的三维视图中,并将所述待标注对象作为已标注对象;Displaying the 3D labeling frame information corresponding to the object to be labelled in the 3D view corresponding to the 3D point cloud data to be labelled, and using the object to be labelled as the labelled object;
根据所述已标注对象的三维标注框信息,确定所述三维视图中的其它待标注对象的三维标注框信息;所述其它待标注对象与所述已标注对象属于同一对象类型;According to the 3D annotation frame information of the marked object, determine the 3D annotation frame information of other objects to be marked in the 3D view; the other objects to be marked and the marked object belong to the same object type;
将所述其它待标注对象的三维标注框信息展示在所述三维视图中。The 3D annotation frame information of the other objects to be annotated is displayed in the 3D view.
在一种可能的实施方式中,所述根据所述已标注对象的三维标注框信息,确定所述三维视图中的其它待标注对象的三维标注框信息,包括:In a possible implementation manner, the determining, according to the 3D annotation frame information of the marked object, the 3D annotation frame information of other objects to be annotated in the 3D view includes:
响应于针对所述三维视图中已标注对象的三维标注框的复制指令,复制所述已标注对象的三维标注框信息;In response to a copy instruction for the 3D annotation frame of the marked object in the 3D view, copying the 3D annotation frame information of the marked object;
响应于针对所述三维视图中其它待标注对象的粘贴指令,根据所述已标注对象在所述三维视图中的三维标注框信息,对所述其它待标注对象进行三维标注,得到所述其它待标注对象的三维标注框信息。In response to the paste instruction for other objects to be marked in the three-dimensional view, perform three-dimensional marking on the other objects to be marked according to the three-dimensional marking frame information of the marked objects in the three-dimensional view, and obtain the other objects to be marked. The 3D callout box information of the callout object.
这里,考虑到同一对象类型的对象其对应的三维标注框信息大致是相同的,因此,本公开实施例中,可以基于已标注对象的三维标注框的复制操作,将其三维标注框信息粘贴到同属一个对象类型的其它待标注对象上,从而可以快速实现对其它待标注对象的三维标注,标注效率进一步得以提升。Here, considering that objects of the same object type have roughly the same corresponding 3D annotation frame information, in this embodiment of the present disclosure, the 3D annotation frame information of the marked object can be pasted to the On other to-be-labeled objects belonging to the same object type, three-dimensional labeling of other to-be-labeled objects can be quickly realized, and the labeling efficiency is further improved.
在一种可能的实施方式中,所述确定所述待标注对象对应的三维标注框信息之后, 还包括:In a possible implementation manner, after determining the three-dimensional labeling frame information corresponding to the object to be labelled, the method further includes:
将所述待标注对象对应的三维标注框信息,展示在当前帧待标注三维点云数据对应的第一三维视图中,并将所述待标注对象作为已标注对象;其中,不同帧待标注三维点云数据对应不同的三维视图;Display the 3D labeling frame information corresponding to the object to be labelled in the first 3D view corresponding to the 3D point cloud data to be labelled in the current frame, and use the object to be labelled as the labelled object; wherein, different frames to be labelled 3D view Point cloud data corresponds to different 3D views;
响应于针对所述第一三维视图中所述已标注对象的三维标注框的复制指令,复制所述已标注对象的三维标注框信息;In response to a copy instruction for the 3D annotation frame of the annotated object in the first 3D view, copying the 3D annotation frame information of the annotated object;
在切换到下一帧待标注三维点云数据对应的第二三维视图的情况下,响应于针对所述第二三维视图中其它待标注对象的粘贴指令,根据所述已标注对象在所述第一三维视图中的三维标注框信息,对所述第二三维视图中的其它待标注对象进行三维标注,得到其它待标注对象的三维标注框信息。In the case of switching to the second 3D view corresponding to the 3D point cloud data to be labeled in the next frame, in response to the paste instruction for other objects to be labeled in the second 3D view, according to the labeled objects in the second 3D view 3D labeling frame information in a 3D view, perform 3D labeling on other objects to be labelled in the second 3D view, and obtain 3D labeling frame information of other objects to be labelled.
这里,考虑到相邻帧待标注三维点云数据之间存在待标注对象的信息传递,因此,可以基于一帧待标注三维点云数据中已标注对象的三维标注框的复制操作,将其三维标注框信息粘贴到下一帧的其它待标注对象上,从而可以快速实现对不同帧中的相同类型待标注对象的三维标注,标注效率进一步得以提升。Here, considering that there is information transfer of the object to be labeled between the 3D point cloud data to be labeled in the adjacent frames, it can be based on the copy operation of the 3D labeled frame of the labeled object in one frame of the 3D point cloud data to be labeled, and the three-dimensional The annotation frame information is pasted to other objects to be annotated in the next frame, so that the three-dimensional annotation of the same type of objects to be annotated in different frames can be quickly realized, and the annotation efficiency is further improved.
在一种可能的实施方式中,所述响应针对所述鸟瞰图中的待标注对象的框选操作,确定第一二维标注框信息,包括:In a possible implementation manner, in response to a frame selection operation on the object to be marked in the bird's-eye view, determining the first two-dimensional frame information includes:
获取参考图像;所述参考图像与所述待标注三维点云数据来源于同一场景;obtaining a reference image; the reference image and the to-be-labeled 3D point cloud data originate from the same scene;
基于所述待标注对象在所述参考图像中的图像位置信息、以及所述参考图像对应的坐标系与所述鸟瞰图对应的坐标系之间的映射关系,确定所述待标注对象在所述鸟瞰图中的参考位置信息;Based on the image position information of the object to be labeled in the reference image and the mapping relationship between the coordinate system corresponding to the reference image and the coordinate system corresponding to the bird's-eye view, it is determined that the object to be labeled is located in the reference image. Reference location information in the bird's eye view;
响应针对所述鸟瞰图中的所述待标注对象的框选操作,基于所述参考位置信息确定所述第一二维标注框信息。In response to a frame selection operation on the object to be marked in the bird's-eye view, the first two-dimensional frame information is determined based on the reference position information.
这里,可以基于获取的参考图像,确定待标注对象在鸟瞰图中的参考位置信息,基于这一参考位置信息即可以指导用户针对鸟瞰图中的待标注对象进行框选操作,这样,所框选出来的第一二维标注框将更大程度上的覆盖待标注对象所对应的点云数据,从而可以提升标注的准确度。Here, the reference position information of the object to be marked in the bird's-eye view can be determined based on the obtained reference image, and based on this reference position information, the user can be instructed to perform a frame selection operation for the object to be marked in the bird's-eye view. The resulting first two-dimensional labeling frame will cover the point cloud data corresponding to the object to be labelled to a greater extent, thereby improving the labeling accuracy.
在一种可能的实施方式中,所述初始参数信息包括初始高度信息和初始位姿信息中的至少之一。In a possible implementation manner, the initial parameter information includes at least one of initial height information and initial pose information.
第二方面,本公开实施例还提供了一种点云标注的装置,所述装置包括:In a second aspect, an embodiment of the present disclosure further provides a point cloud labeling device, the device comprising:
获取模块,配置为获取待标注三维点云数据;an acquisition module, configured to acquire the 3D point cloud data to be marked;
生成模块,配置为基于所述待标注三维点云数据,生成所述待标注三维点云数据对应的鸟瞰图;a generating module, configured to generate a bird's-eye view corresponding to the three-dimensional point cloud data to be marked based on the three-dimensional point cloud data to be marked;
第一确定模块,配置为响应针对所述鸟瞰图中的待标注对象的框选操作,确定第一二维标注框信息;基于被框选的待标注对象的三维点云数据,确定所述待标注对象的初始参数信息;A first determination module, configured to determine first two-dimensional labeling frame information in response to a frame selection operation for objects to be labelled in the bird's-eye view; The initial parameter information of the annotation object;
标注模块,配置为基于所述第一二维标注框信息,以及所述初始参数信息,确定所述待标注对象对应的三维标注框信息。The labeling module is configured to determine the three-dimensional labeling frame information corresponding to the object to be labelled based on the first two-dimensional labeling frame information and the initial parameter information.
第三方面,本公开实施例还提供了一种电子设备,包括:处理器、存储器和总线,所述存储器存储有所述处理器可执行的机器可读指令,所述处理器配置为执行所述存储器中存储的机器可读指令,当电子设备运行时,所述处理器与所述存储器之间通过总线通信,所述机器可读指令被所述处理器执行时执行如第一方面及其各种实施方式任一所述的点云标注的方法的步骤。In a third aspect, embodiments of the present disclosure further provide an electronic device, including: a processor, a memory, and a bus, where the memory stores machine-readable instructions executable by the processor, and the processor is configured to execute all The machine-readable instructions stored in the memory, when the electronic device is running, the processor communicates with the memory through a bus, and the machine-readable instructions are executed by the processor. The steps of the point cloud labeling method described in any of the various embodiments.
第四方面,本公开实施例还提供了一种计算机可读存储介质,所述计算机可读存储介质上存储有计算机程序,所述计算机程序被电子设备运行时,所述电子设备执行如第 一方面及其各种实施方式任一所述的点云标注的方法的步骤。In a fourth aspect, an embodiment of the present disclosure further provides a computer-readable storage medium, where a computer program is stored on the computer-readable storage medium, and when the computer program is run by an electronic device, the electronic device executes the first Aspects and steps of the method for point cloud labeling described in any one of its various embodiments.
第五方面,本公开实施例还提供了一种计算机程序产品,包括存储程序代码的计算机可读存储介质,所述程序代码包括的指令可配置为执行如第一方面及其各种实施方式任一所述的点云标注的方法的步骤。In a fifth aspect, embodiments of the present disclosure further provide a computer program product, including a computer-readable storage medium storing program codes, where the program codes include instructions that can be configured to execute any of the first aspect and its various implementations. 1. The steps of the method for point cloud labeling.
关于上述点云标注的装置、电子设备、计算机可读存储介质及计算机程序产品的效果描述参见上述点云标注的方法的说明。For the description of the effects of the above point cloud labeling apparatus, electronic device, computer readable storage medium and computer program product, please refer to the description of the above point cloud labeling method.
为使本公开的上述目的、特征和优点能更明显易懂,下文特举较佳实施例,并配合所附附图,作详细说明如下。In order to make the above-mentioned objects, features and advantages of the present disclosure more obvious and easy to understand, the preferred embodiments are exemplified below, and are described in detail as follows in conjunction with the accompanying drawings.
附图说明Description of drawings
为了更清楚地说明本公开实施例的技术方案,下面将对实施例中所需要使用的附图作简单地介绍,此处的附图被并入说明书中并构成本说明书中的一部分,这些附图示出了符合本公开的实施例,并与说明书一起用于说明本公开的技术方案。应当理解,以下附图仅示出了本公开的某些实施例,因此不应被看作是对范围的限定,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其他相关的附图。In order to explain the technical solutions of the embodiments of the present disclosure more clearly, the following briefly introduces the accompanying drawings required in the embodiments, which are incorporated into the specification and constitute a part of the specification. The drawings illustrate embodiments consistent with the present disclosure, and together with the description serve to explain the technical solutions of the present disclosure. It should be understood that the following drawings only show some embodiments of the present disclosure, and therefore should not be regarded as limiting the scope. Other related figures are obtained from these figures.
图1示出了本公开实施例提供的一种点云标注的方法的流程示意图;FIG. 1 shows a schematic flowchart of a method for labeling a point cloud provided by an embodiment of the present disclosure;
图2示出了本公开实施例提供的一种点云标注的流程示意图;FIG. 2 shows a schematic flowchart of a point cloud labeling provided by an embodiment of the present disclosure;
图3示出了本公开实施例提供的一种点云标注的装置的结构示意图;FIG. 3 shows a schematic structural diagram of a device for point cloud labeling provided by an embodiment of the present disclosure;
图4示出了本公开实施例提供的一种电子设备的结构示意图。FIG. 4 shows a schematic structural diagram of an electronic device provided by an embodiment of the present disclosure.
具体实施方式Detailed ways
为使本公开实施例的目的、技术方案和优点更加清楚,下面将结合本公开实施例中附图,对本公开实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例仅仅是本公开一部分实施例,而不是全部的实施例。通常在此处描述和示出的本公开实施例的组件可以以各种不同的配置来布置和设计。因此,以下对本公开的实施例的详细描述并非旨在限制要求保护的本公开的范围,而是仅仅表示本公开的选定实施例。基于本公开的实施例,本领域技术人员在没有做出创造性劳动的前提下所获得的所有其他实施例,都属于本公开保护的范围。In order to make the purposes, technical solutions and advantages of the embodiments of the present disclosure more clear, the technical solutions in the embodiments of the present disclosure will be clearly and completely described below with reference to the accompanying drawings in the embodiments of the present disclosure. Obviously, the described embodiments are only These are some, but not all, embodiments of the present disclosure. The components of the disclosed embodiments generally described and illustrated herein may be arranged and designed in a variety of different configurations. Thus, the following detailed description of the embodiments of the present disclosure is not intended to limit the scope of the disclosure as claimed, but is merely representative of selected embodiments of the disclosure. Based on the embodiments of the present disclosure, all other embodiments obtained by those skilled in the art without creative work fall within the protection scope of the present disclosure.
经研究发现,目前,对于点云数据的标注,大多依靠人工方式直接在三维空间中操作进行标注。该方式需要通过复杂而不精确的三维空间交互操作来标定待标注对象的位置、大小尺寸、前进方向和类别等训练所需信息,操作步骤复杂,标注效率和准确率均较低。The research found that, at present, for the annotation of point cloud data, most of them rely on manual methods to directly operate in the three-dimensional space for annotation. This method requires complex and inaccurate three-dimensional interactive operations to demarcate the information required for training, such as the position, size, direction, and category of the object to be labeled. The operation steps are complicated, and the labeling efficiency and accuracy are low.
基于上述研究,本公开至少提供一种点云标注的方案,基于二维鸟瞰图的框选操作实现三维标注框信息的确定,操作简单,标注效率和准确率均较高。Based on the above research, the present disclosure provides at least one point cloud labeling solution, which realizes the determination of 3D labeling frame information based on the frame selection operation of the 2D bird's-eye view, which is simple to operate, and has high labeling efficiency and accuracy.
针对以上方案所存在的缺陷,均是发明人在经过实践并仔细研究后得出的结果,因此,上述问题的发现过程以及下文中本公开针对上述问题所提出的解决方案,都应该是发明人在本公开过程中对本公开做出的贡献。The defects existing in the above solutions are all the results obtained by the inventor after practice and careful research. Therefore, the discovery process of the above problems and the solutions to the above problems proposed by the present disclosure hereinafter should be the inventors Contributions made to this disclosure during the course of this disclosure.
应注意到:相似的标号和字母在下面的附图中表示类似项,因此,一旦某一项在一个附图中被定义,则在随后的附图中不需要对其进行进一步定义和解释。It should be noted that like numerals and letters refer to like items in the following figures, so once an item is defined in one figure, it does not require further definition and explanation in subsequent figures.
激光雷达被广泛应用于当今的自动驾驶系统中。它可以提供精确的三维环境空间信息,从而辅助系统的场景理解和决策过程。近年来出现了许多基于深度学习的感知算法 来处理这类数据,在应用上这些算法明显优于单目和立体视觉的方法。这些算法的快速发展得到了建立在多个开放数据集上的几个具有挑战性的基准测试的支持。然而,尽管已经发布了相当数量的数据,但实际的产品部署仍然需要更多带有准确标签的数据来为算法提供数据。目前可公开获取的注释工具还很粗糙,这限制了该领域的研究进展。Lidar is widely used in today's autonomous driving systems. It can provide accurate three-dimensional environmental spatial information to assist the system's scene understanding and decision-making process. In recent years, many deep learning-based perceptual algorithms have emerged to process such data, and these algorithms are significantly superior to monocular and stereo vision methods in application. The rapid development of these algorithms is supported by several challenging benchmarks built on multiple open datasets. However, despite the considerable amount of data that has been published, actual product deployment still requires more accurately labeled data to feed algorithms. Currently publicly available annotation tools are crude, which limits research progress in this field.
虽然有许多方法可以有效地注释红绿蓝(RGB)图像,但是由于三维(3 Dimensions,3D)中标注更为复杂的情况而没有多少工作集中在3D标注任务上。Although there are many methods to efficiently annotate red-green-blue (RGB) images, not much work has focused on the 3D annotation task due to the more complex situation of annotation in 3 Dimensions (3D).
在稀疏且不规则分布的点云中,很难正确识别所有感兴趣的目标。此外,考虑到标注存在较大的自由度,例如需要标注对象的高度和转向角,操作复杂度相对较高,因此需要定制用户交互界面设计来实现精确的注释。In a sparse and irregularly distributed point cloud, it is difficult to correctly identify all objects of interest. In addition, considering the large degree of freedom of annotation, such as the need to annotate the height and steering angle of the object, the operation complexity is relatively high, so the user interface design needs to be customized to achieve accurate annotation.
虽然可以采用一些算法辅助的半自动方法来提高标注的效率,但是这些方法的效率不高,特别是考虑到为注释者提供的设备。例如,有的工作需要图形处理器(Graphics Processing Unit,GPU)训练3D检测器来辅助标注,这在实际应用中很不方便。Although some algorithm-assisted semi-automatic methods can be employed to improve the efficiency of annotation, these methods are not very efficient, especially considering the equipment provided for annotators. For example, some work requires a Graphics Processing Unit (GPU) to train a 3D detector to assist in labeling, which is inconvenient in practical applications.
为便于对本实施例进行理解,首先对本公开实施例所公开的一种点云标注的方法进行详细介绍,本公开实施例所提供的点云标注的方法的执行主体一般为具有一定计算能力的电子设备,该电子设备例如包括:终端设备或服务器或其它处理设备,终端设备可以为用户设备(User Equipment,UE)、移动设备、用户终端、终端、蜂窝电话、无绳电话、个人数字处理(Personal Digital Assistant,PDA)、手持设备、计算设备、车载设备、可穿戴设备等。在一些可能的实现方式中,该点云标注的方法可以通过处理器调用存储器中存储的计算机可读指令的方式来实现。In order to facilitate the understanding of this embodiment, a method for labeling a point cloud disclosed in an embodiment of the present disclosure is first introduced in detail. Equipment, the electronic equipment for example includes: terminal equipment or server or other processing equipment, the terminal equipment can be user equipment (User Equipment, UE), mobile equipment, user terminal, terminal, cellular phone, cordless phone, Personal Digital Processing (Personal Digital Assistant, PDA), handheld devices, computing devices, in-vehicle devices, wearable devices, etc. In some possible implementations, the point cloud labeling method can be implemented by the processor calling computer-readable instructions stored in the memory.
下面以执行主体为终端设备为例对本公开实施例提供的点云标注的方法加以说明。The following describes the point cloud labeling method provided by the embodiment of the present disclosure by taking the execution subject as a terminal device as an example.
图1示出了本公开实施例提供的一种点云标注的方法的流程示意图,如图1所示,该方法包括步骤S101至S105,该方法可以应用于点云标注装置或点云标注设备,点云标注装置可以是处理器或芯片,该处理器或芯片可以用在点云标注设备中,其中,该方法包括:Fig. 1 shows a schematic flowchart of a point cloud labeling method provided by an embodiment of the present disclosure. As shown in Fig. 1 , the method includes steps S101 to S105, and the method can be applied to a point cloud labeling device or a point cloud labeling device , the point cloud labeling device may be a processor or a chip, and the processor or chip may be used in the point cloud labeling device, wherein the method includes:
S101、获取待标注三维点云数据;S101, obtaining the three-dimensional point cloud data to be marked;
S102、基于待标注三维点云数据,生成待标注三维点云数据对应的鸟瞰图;S102, generating a bird's-eye view corresponding to the three-dimensional point cloud data to be marked based on the three-dimensional point cloud data to be marked;
S103、响应针对鸟瞰图中的待标注对象的框选操作,确定第一二维标注框信息;S103, in response to a frame selection operation for the object to be marked in the bird's-eye view, determining first two-dimensional frame information;
S104、基于被框选的待标注对象的三维点云数据,确定待标注对象的初始参数信息;S104, determining the initial parameter information of the object to be marked based on the three-dimensional point cloud data of the object to be marked selected by the frame;
S105、基于第一二维标注框信息,以及初始参数信息,确定待标注对象对应的三维标注框信息。S105. Based on the first two-dimensional labeling frame information and the initial parameter information, determine the three-dimensional labeling frame information corresponding to the object to be labelled.
这里,为了便于理解本公开实施例提供的点云标注的方法,首先可以对该点云标注的方法的技术场景进行简单介绍。本公开实施例提供的点云标注的方法可以应用于任何需要对待标注对象进行三维标注的场景中,基于三维标注后的三维标注框信息可以实现有关目标检测及跟随模型等的训练,还可以应用于其它需要三维标注框信息的相关应用场景中,例如,可以应用到自动驾驶车辆中对前方车辆进行信息识别,以避免发生车辆碰撞等交通危险,除此之外,还可以应用于其它应用场景中,在此不做限制。Here, in order to facilitate the understanding of the point cloud labeling method provided by the embodiments of the present disclosure, the technical scene of the point cloud labeling method may be briefly introduced first. The point cloud labeling method provided by the embodiments of the present disclosure can be applied to any scene that requires 3D labeling of the object to be labelled. Based on the 3D labeling frame information after 3D labeling, training related to target detection and following models can be implemented, and can also be applied In other related application scenarios that require 3D annotation frame information, for example, it can be applied to automatic driving vehicles to identify the information of vehicles ahead to avoid traffic hazards such as vehicle collisions. In addition, it can also be applied to other application scenarios. , there is no restriction here.
考虑到相关技术中可以是依靠人工方式直接在三维空间中操作进行标注,又由于三维点云数据存在较多的标注自由度,例如,针对某个物体的点云可以标注其对应的高度及旋转角度等信息,操作复杂度相对较高。又考虑到针对稀疏且不规则分布的点云数据而言,很难正确识别到感兴趣的对象,这一定程度加大了标注的难度,降低了标注效率。Considering that the related technology can be manually operated in three-dimensional space for labeling, and because the three-dimensional point cloud data has more labeling degrees of freedom, for example, the point cloud of an object can be marked with its corresponding height and rotation. Angle and other information, the operation complexity is relatively high. Considering that for sparse and irregularly distributed point cloud data, it is difficult to correctly identify the object of interest, which increases the difficulty of labeling to a certain extent and reduces the efficiency of labeling.
正是为了解决上述问题,本公开实施例才提供了一种点云标注的方法,基于二维鸟瞰图的框选操作实现三维标注框信息的确定,操作简单,标注效率和准确率均较高。It is in order to solve the above problem that the embodiments of the present disclosure provide a point cloud labeling method, which realizes the determination of 3D labeling frame information based on the frame selection operation of the 2D bird's-eye view. The operation is simple, and the labeling efficiency and accuracy are high. .
本公开实施例中的待标注三维点云数据可以是利用雷达设备所采集的点云数据,这里的雷达设备可以采用旋转式扫描激光雷达,还可以采用其他雷达设备,对此不做具体 的限制。以旋转式扫描激光雷达为例,该激光雷达在水平方向旋转扫描的情况下可以获取周边环境内有关的三维点云数据。这里,在进行旋转扫描的过程中,激光雷达可以采用多线扫描方式,即发射使用多个激光管顺序发射,结构为多个激光管纵向排列,即在水平方向旋转扫描的过程中,进行垂直方向的多层扫描。每个激光管之间有一定的夹角,垂直发射视场可以在30°至40°,这样,在激光雷达设备每旋转一个扫描角度可以获取多个激光管发射激光所返回的一个数据包,将各个扫描角度获取的数据包进行拼接即可得到一帧点云数据(对应旋转一周扫描360°),在激光雷达扫描一周后,即可以进行一帧点云数据的采集。The three-dimensional point cloud data to be labeled in the embodiment of the present disclosure may be point cloud data collected by a radar device. The radar device here may use a rotary scanning laser radar, or other radar devices, which are not specifically limited. . Taking the rotary scanning lidar as an example, the lidar can obtain the relevant three-dimensional point cloud data in the surrounding environment when it rotates and scans in the horizontal direction. Here, in the process of rotating scanning, the lidar can use a multi-line scanning method, that is, the emission uses multiple laser tubes to emit sequentially, and the structure is that multiple laser tubes are arranged longitudinally, that is, in the process of rotating and scanning in the horizontal direction, vertical scanning is performed. Orientation of multi-layer scanning. There is a certain angle between each laser tube, and the vertical emission field can be 30° to 40°. In this way, a data packet returned by multiple laser tubes can be obtained for each rotation of a scanning angle of the lidar device. One frame of point cloud data can be obtained by splicing the data packets obtained from each scanning angle (corresponding to a 360° scan of one rotation), and one frame of point cloud data can be collected after one scan of the lidar.
需要说明的是,本公开实施例中的待标注三维点云数据可以指的是一帧点云数据,还可以是半帧点云数据,还可以是其它扫描范围所确定的点云数据,对此不做具体的限制。It should be noted that the 3D point cloud data to be marked in the embodiments of the present disclosure may refer to a frame of point cloud data, a half-frame point cloud data, or point cloud data determined by other scanning ranges. This does not make specific restrictions.
本公开实施例中可以是基于二维鸟瞰图中的待标注对象的框选操作,实现该待标注对象对应的三维标注框信息的确定。这里,可以基于鸟瞰图对应的坐标系和待标注三维点云数据对应的坐标系(如世界坐标系)之间的映射关系,将待标注三维点云数据转换到鸟瞰图上,该鸟瞰图可以是三维点云数据在俯视图上的映射结果,可以对应第一二维点云数据。考虑到在俯视图的视野中,对应待标注对象的轮廓是更为清晰的且不存在遮挡的问题,因而,本公开实施例中可以基于上述映射结果,针对鸟瞰图中的待标注对象执行框选操作,从而确定对应的第一二维标注框信息。In the embodiment of the present disclosure, the determination of the three-dimensional labeling frame information corresponding to the to-be-labeled object may be implemented based on the frame selection operation of the to-be-labeled object in the two-dimensional bird's-eye view. Here, based on the mapping relationship between the coordinate system corresponding to the bird's-eye view and the coordinate system (such as the world coordinate system) corresponding to the three-dimensional point cloud data to be marked, the three-dimensional point cloud data to be marked can be converted to the bird's-eye view, and the bird's-eye view can be is the mapping result of the three-dimensional point cloud data on the top view, which can correspond to the first two-dimensional point cloud data. Considering that in the field of view of the top view, the outline corresponding to the object to be marked is clearer and there is no occlusion problem, therefore, in the embodiment of the present disclosure, based on the above mapping result, frame selection can be performed for the object to be marked in the bird's eye view. operation, so as to determine the corresponding first two-dimensional callout frame information.
在针对待标注对象进行鸟瞰图下的待标注对象的框选操作的情况下,基于第一二维标注框信息所指示框选的有关待标注对象的三维点云数据,可以确定待标注对象的初始参数信息,这里的初始参数信息可以是初始高度信息,或者,可以是初始位姿信息,或者,可以是初始高度信息和初始位姿信息的结合,本公开实施例对此不做具体的限制。In the case of performing a frame selection operation of the object to be marked under the bird's eye view on the object to be marked, based on the three-dimensional point cloud data about the object to be marked selected by the frame selection indicated by the first two-dimensional markup frame information, the size of the object to be marked can be determined. Initial parameter information, where the initial parameter information may be initial height information, or may be initial pose information, or may be a combination of initial height information and initial pose information, which is not specifically limited in this embodiment of the present disclosure .
其中,有关初始高度信息可以是对应框选的三维点云数据的最高点云点的坐标信息与最低点云点的坐标信息之间的坐标之差确定的,有关初始位姿信息则可以包括初始位置信息和初始姿态信息,初始位置信息可以是基于框选的三维点云数据在待标注三维点云数据中的相对位置确定,初始姿态信息则可以是基于对应框选的两个侧边所对应点云点集合与参考方向之间的夹角确定的。Wherein, the relevant initial height information may be determined by the difference between the coordinate information of the highest point cloud point and the coordinate information of the lowest point cloud point of the 3D point cloud data corresponding to the frame selection, and the relevant initial pose information may include the initial Position information and initial attitude information. The initial position information can be determined based on the relative position of the 3D point cloud data selected by the frame in the 3D point cloud data to be marked, and the initial attitude information can be based on the two sides corresponding to the frame selection. The angle between the point cloud point set and the reference direction is determined.
基于上述第一二维标注框信息以及待标注对象的初始参数信息即可以确定待标注对象的三维标注框信息,这里的三维标注框信息可以是初始三维标注框信息,后续可以通过对二维标注框的调整来同步调整。Based on the above-mentioned first 2D annotation frame information and the initial parameter information of the object to be annotated, the 3D annotation frame information of the to-be-annotated object can be determined. The 3D annotation frame information here can be the initial 3D annotation frame information. Adjust the frame to synchronize the adjustment.
在本公开实施例中,在获取到待标注三维点云数据的情况下,可以先生成对应的鸟瞰图,这样,在响应针对鸟瞰图中的待标注对象的框选操作之后,可以确定第一二维标注框信息,同时,还可以结合被框选的待标注对象的三维点云数据确定待标注对象的初始参数信息,这样,结合第一二维标注框信息以及初始参数信息可以确定待标注对象对应的三维标注框信息。可知,上述点云标注的方法是在二维视图(对应鸟瞰图)下执行的框选操作,相比相关技术中直接对三维点云数据进行标注所存在的操作复杂度高的问题,二维视图下执行框选操作的复杂度较低,并且局部点云的放大视角便于用户精确地辨识标注框的边界,一定程度上可以确保所框选待标注对象的准确性,这样,结合通过框选得到的第一二维标注框信息和初始参数信息即可以实现针对待标注对象进行三维标注框的自动标注,标注效率和准确率均较高。In the embodiment of the present disclosure, when the 3D point cloud data to be labeled is obtained, a corresponding bird's-eye view can be generated first, so that after responding to the frame selection operation for the objects to be labeled in the bird's-eye view, the first The two-dimensional labeling frame information, and at the same time, the initial parameter information of the to-be-labeled object can also be determined in combination with the 3D point cloud data of the frame-selected object to be labeled. Information about the 3D callout box corresponding to the object. It can be seen that the above point cloud labeling method is a frame selection operation performed in a two-dimensional view (corresponding to a bird's-eye view). The complexity of performing the frame selection operation under the view is low, and the magnified view of the local point cloud facilitates the user to accurately identify the boundary of the labeling box, which can ensure the accuracy of the framed object to be labelled to a certain extent. In this way, combined with the frame selection The obtained first two-dimensional labeling frame information and initial parameter information can realize automatic labeling of the three-dimensional labeling frame for the object to be labelled, and the labeling efficiency and accuracy are high.
考虑到三维标注框信息所指示的三维标注框在不同视图所观测到的待标注对象信息也不同,例如,针对俯视图而言,所观测到的可以是待标注对象的长和宽,针对前视图而言,所观测到的可以是待标注对象的宽和高。因此,为了进一步提升标注的准确性,这里可以结合其他视图下的二维标注框信息来实现三维标注框信息的确定,例如可以包 括如下步骤:Considering that the 3D annotation frame indicated by the 3D annotation frame information has different information on the object to be annotated observed in different views, for example, for the top view, what is observed may be the length and width of the object to be annotated, and for the front view. In other words, what is observed can be the width and height of the object to be annotated. Therefore, in order to further improve the accuracy of the annotation, the 2D annotation frame information in other views can be combined to determine the 3D annotation frame information, for example, the following steps can be included:
步骤一、基于待标注对象的三维点云数据,生成待标注三维点云数据对应的待调整视图;待调整视图至少包括正视图和侧视图中的一种;Step 1: Generate a view to be adjusted corresponding to the three-dimensional point cloud data to be labeled based on the three-dimensional point cloud data of the object to be labeled; the view to be adjusted includes at least one of a front view and a side view;
步骤二、根据生成的待调整视图,确定待标注对象在待调整视图中的第二二维标注框信息。Step 2: According to the generated view to be adjusted, determine the second two-dimensional labeling frame information of the object to be labeled in the view to be adjusted.
本公开实施例中,在确定待调整视图中的第二二维标注框信息之前,可以生成待调整视图。这里的第二二维标注框信息可以是第一二维标注框信息所对应三维点云数据在待调整视图上的映射结果。In this embodiment of the present disclosure, before the second two-dimensional frame information in the view to be adjusted is determined, the view to be adjusted may be generated. The second two-dimensional annotation frame information here may be a mapping result of the three-dimensional point cloud data corresponding to the first two-dimensional annotation frame information on the view to be adjusted.
在一些应用中,本公开实施例中的待调整视图可以是正视图,也可以是侧视图,还可以是上述两类视图的结合。其中,有关正视图和侧视图这些待调整视图的生成过程可以参照上述二维鸟瞰图的生成过程,即利用相应的坐标系之间的转换关系来生成。In some applications, the view to be adjusted in the embodiment of the present disclosure may be a front view, a side view, or a combination of the above two types of views. Wherein, the generation process of the front view and the side view of the views to be adjusted may refer to the above-mentioned generation process of the two-dimensional bird's-eye view, that is, by using the conversion relationship between the corresponding coordinate systems.
本公开实施例中,相比鸟瞰图,正视图和侧视图等待调整视图可以是理解成是对三维视图进行观测的不同视角,在不同视角下一定程度上可以更为准确的分析出在三维视图下相关待标注对象对应的三维标注框信息。因此,本公开实施例中,可以基于第一二维标注框信息、第二二维标注框信息以及初始参数信息,确定待标注对象对应的三维标注框信息。通过这种方式,考虑到待调整视图和鸟瞰图是同属于三维点云数据所在三维视图下的两个不同的二维视图,两个不同的二维视图之间存在参数信息的关联性,因而,结合待调整视图所确定的第二二维标注框信息和鸟瞰图所确定的第一二维标注框信息所确定参数信息的一定程度上可以确定出更为准确性的三维标注框信息。In the embodiment of the present disclosure, compared with the bird's-eye view, the front view and the side view waiting to be adjusted can be understood as different perspectives for observing the three-dimensional view, and the three-dimensional view can be more accurately analyzed under different perspectives to a certain extent. Download the 3D annotation frame information corresponding to the object to be annotated. Therefore, in this embodiment of the present disclosure, the 3D annotation frame information corresponding to the object to be annotated may be determined based on the first 2D annotation frame information, the second 2D annotation frame information, and the initial parameter information. In this way, considering that the view to be adjusted and the bird's-eye view are two different 2D views that belong to the same 3D view where the 3D point cloud data is located, there is a correlation of parameter information between the two different 2D views. , more accurate three-dimensional annotation frame information can be determined to a certain extent by combining the second two-dimensional annotation frame information determined by the view to be adjusted and the parameter information determined by the first two-dimensional annotation frame information determined by the bird's-eye view.
在一些实施方式中,基于第一二维标注框信息、第二二维标注框信息以及初始参数信息,确定待标注对象对应的三维标注框信息,可以通过如下步骤来实现:In some embodiments, determining the 3D annotation frame information corresponding to the object to be annotated based on the first 2D annotation frame information, the second 2D annotation frame information and the initial parameter information can be achieved by the following steps:
步骤一、响应针对第二二维标注框信息的调整操作,基于初始参数信息,确定待标注对象的调整后的参数信息;Step 1: In response to the adjustment operation for the second two-dimensional labeling frame information, based on the initial parameter information, determine the adjusted parameter information of the object to be labelled;
步骤二、基于第一二维标注框信息、第二二维标注框信息以及调整后的参数信息,确定待标注对象对应的三维标注框信息。Step 2: Based on the first two-dimensional annotation frame information, the second two-dimensional annotation frame information, and the adjusted parameter information, determine the three-dimensional annotation frame information corresponding to the object to be annotated.
通过这种方式,可以基于针对待标注对象在所述待调整视图中的第二二维标注框信息的调整操作,对初始参数信息进行调整,这样,即可以基于调整后的参数信息、鸟瞰图对应的第一二维标注框信息、待调整视图对应的第二二维标注框信息,确定待标注对象对应的三维标注框信息。这可以是考虑到不同视图所侧重对应展示的待标注对象的相关信息也不同,例如,针对鸟瞰图而言,对应展示的可以是待标注对象的长和宽,而针对前视图而言,对应展示的则可以是待标注对象的宽和高,因而,在初始参数信息的基础上,再结合待调整视图进行参数调整,将进一步提升所确定的三维标注框信息的准确性。In this way, the initial parameter information can be adjusted based on the adjustment operation for the second two-dimensional annotation frame information of the object to be labeled in the view to be adjusted. The corresponding first 2D annotation frame information and the second 2D annotation frame information corresponding to the view to be adjusted are used to determine the 3D annotation frame information corresponding to the object to be annotated. This may be due to the fact that different views focus on different information about the object to be marked. For example, for the bird's-eye view, the corresponding display may be the length and width of the object to be marked, and for the front view, the corresponding display The display can be the width and height of the object to be marked. Therefore, on the basis of the initial parameter information, parameter adjustment is performed in combination with the view to be adjusted, which will further improve the accuracy of the determined 3D annotation frame information.
在一些实施方式中,所述响应针对所述第二二维标注框信息的调整操作,基于所述初始参数信息,确定所述待标注对象的调整后的参数信息,包括:响应针对所述待调整视图中的第二二维标注框的调整操作,确定与所述调整操作对应的参数调整信息;基于所述参数调整信息对所述初始参数信息进行调整,得到所述调整后的参数信息。In some implementation manners, determining the adjusted parameter information of the object to be annotated based on the initial parameter information in response to the adjustment operation for the second two-dimensional annotation frame information includes: responding to the adjustment operation for the to-be-annotated object. Adjusting the adjustment operation of the second two-dimensional callout frame in the view to determine parameter adjustment information corresponding to the adjustment operation; and adjusting the initial parameter information based on the parameter adjustment information to obtain the adjusted parameter information.
这里,首先可以通过针对第二二维标注框信息执行调整操作,即可在初始参数信息的基础上,对初始参数信息进行调整,得到待标注对象的调整后的参数信息,这样,结合第一二维标注框信息和第二二维标注框信息即可以得到更新后的三维标注框信息。Here, firstly, an adjustment operation can be performed on the information of the second two-dimensional annotation frame, that is, the initial parameter information can be adjusted on the basis of the initial parameter information to obtain the adjusted parameter information of the object to be annotated. The updated 3D annotation frame information can be obtained from the 2D annotation frame information and the second 2D annotation frame information.
需要说明的,在针对第二二维标注框信息进行调整的过程中,第一二维标注框信息也可以对应发生调整。例如,在针对第二二维标注框信息调整了高度信息这一参数的情况下,在第二二维标注框信息的高度和第一二维标注框信息指示的是同一参数的情况下,对应第一二维标注框信息的高度信息也可以随之发生变化,在二维视图中的调整可以适 应性的映射到三维标注框信息所指示的三维标注框中。It should be noted that, in the process of adjusting the second two-dimensional annotation frame information, the first two-dimensional annotation frame information may also be adjusted correspondingly. For example, in the case where the parameter of height information is adjusted for the second 2D annotation frame information, in the case that the height of the second 2D annotation frame information and the first 2D annotation frame information indicate the same parameter, the corresponding The height information of the first 2D annotation frame information can also change accordingly, and the adjustment in the 2D view can be adaptively mapped to the 3D annotation frame indicated by the 3D annotation frame information.
本公开实施例中,在进行参数信息的调整之前,可以先进行待调整视图的展示。在展示的待调整视图上即可以对该视图标注的第二二维标注框进行调整。In this embodiment of the present disclosure, before adjusting the parameter information, the view to be adjusted may be displayed first. On the displayed view to be adjusted, the second two-dimensional frame marked on the view can be adjusted.
这里,在响应针对待调整视图中的第二二维标注框的调整操作的情况下,可以确定与调整操作对应的参数调整信息。例如,在拖动第二二维标注框向外拉伸操作后,可以根据拉伸程度来确定参数调整信息,拉伸程度越大对应的参数调整信息的值也越大,这样,再结合初始参数信息即可以确定调整后的参数信息。Here, in response to an adjustment operation for the second two-dimensional callout in the view to be adjusted, parameter adjustment information corresponding to the adjustment operation may be determined. For example, after dragging the second two-dimensional callout box to stretch outward, the parameter adjustment information can be determined according to the stretching degree. The greater the stretching degree, the greater the value of the parameter adjustment information. The parameter information can determine the adjusted parameter information.
本公开实施例中,可以基于调整后的参数信息所对应的三维点云数据,与第一二维标注框信息所对应的二维点云数据之间的第一对应关系,确定待标注对象对应的三维标注框的底面的参数信息;基于调整后的参数信息所对应的三维点云数据,与第二二维标注框信息所对应的二维点云数据之间的第二对应关系,确定待标注对象对应的三维标注框的侧面和正面的参数信息中的至少之一,这样,将确定的三维标注框的底面的参数信息、以及三维标注框的侧面和正面的参数信息中的至少之一进行组合,即可得到三维标注框信息。In this embodiment of the present disclosure, it may be determined based on the first correspondence between the 3D point cloud data corresponding to the adjusted parameter information and the 2D point cloud data corresponding to the first 2D annotation frame information, to determine the corresponding object to be annotated. based on the second correspondence between the 3D point cloud data corresponding to the adjusted parameter information and the 2D point cloud data corresponding to the second 2D annotation frame information, determine the At least one of the parameter information of the side and front of the 3D annotation frame corresponding to the annotation object, so that the determined parameter information of the bottom surface of the 3D annotation frame and at least one of the parameter information of the side and front of the 3D annotation frame Combining them, you can get the 3D annotation frame information.
通过这种方式,可以基于调整后的参数信息所对应的三维点云数据与第一二维标注框信息所对应的二维点云数据、以及第二二维标注框信息所对应的二维点云数据之间的对应关系,确定三维标注框在多个平面上的参数信息,进而通过信息组合,确定三维标注框信息。In this way, based on the 3D point cloud data corresponding to the adjusted parameter information, the 2D point cloud data corresponding to the first 2D annotation frame information, and the 2D points corresponding to the second 2D annotation frame information The corresponding relationship between the cloud data determines the parameter information of the 3D annotation frame on multiple planes, and then determines the information of the 3D annotation frame through the combination of information.
在一些实施方式中,在所述确定所述待标注对象在所述待调整视图中的第二二维标注框信息之后,还可以执行以下步骤:基于确定的所述待标注对象的第二二维标注框信息,展示标注有与所述第二二维标注框信息对应的第二二维标注框的所述待调整视图。In some embodiments, after the determining of the second two-dimensional annotation frame information of the object to be annotated in the view to be adjusted, the following step may be further performed: based on the determined second two-dimensional annotation of the object to be annotated Dimensional annotation frame information, displaying the to-be-adjusted view marked with a second 2D annotation frame corresponding to the second 2D annotation frame information.
本公开实施例提供的点云标注的方法可以结合校验图像对标注的三维标注框信息进行校验,可以通过如下步骤来实现:The method for point cloud labeling provided by the embodiment of the present disclosure can be combined with a check image to check the labelled 3D label frame information, and can be implemented by the following steps:
步骤一、获取校验图像;校验图像与待标注三维点云数据来源于同一场景;Step 1: Obtain a verification image; the verification image and the 3D point cloud data to be marked originate from the same scene;
步骤二、基于待标注三维点云数据对应的坐标系,与校验图像对应的坐标系之间的映射关系,将待标注对象对应的三维标注框信息映射至校验图像中,确定映射后的图像坐标范围;Step 2: Based on the mapping relationship between the coordinate system corresponding to the three-dimensional point cloud data to be marked and the coordinate system corresponding to the verification image, map the three-dimensional marked frame information corresponding to the object to be marked into the verification image, and determine the mapped image. image coordinate range;
步骤三、根据映射后的图像坐标范围与校验图像中的待标注对象所在图像坐标范围,对三维标注框信息的准确性进行校验。Step 3: Verify the accuracy of the three-dimensional annotation frame information according to the mapped image coordinate range and the image coordinate range where the object to be marked in the verification image is located.
通过这种方式,在针对待标注对象确定对应的三维标注框信息之后,可以基于获取的校验图像对三维标注框信息的准确性进行校验,这里可以是依赖于待标注三维点云数据对应的坐标系与校验图像对应的坐标系之间的映射关系,通过图像坐标范围映射来实现,从而可以更为准确的验证到特定的待标注对象。In this way, after the corresponding 3D annotation frame information is determined for the object to be annotated, the accuracy of the 3D annotation frame information can be verified based on the obtained verification image, which may depend on the corresponding 3D point cloud data to be marked. The mapping relationship between the coordinate system of the image and the coordinate system corresponding to the verification image is realized by mapping the image coordinate range, so that the specific object to be marked can be verified more accurately.
为了基于校验图像实现对三维标注框信息的准确性的校验效果,这里的校验图像可以是与待标注三维点云数据来源于同一场景的相关图像,例如,可以是RGB图像。例如,在自动驾驶应用领域,这里的校验图像可以是与雷达设备安装于同一车辆上的摄像设备所采集的,摄像设备的摄像范围可以与雷达设备的扫描范围一致或者存在重叠。In order to realize the verification effect of the accuracy of the 3D annotation frame information based on the verification image, the verification image here may be a related image from the same scene as the 3D point cloud data to be marked, for example, an RGB image. For example, in the field of autonomous driving applications, the verification image here may be collected by a camera device installed on the same vehicle as the radar device, and the camera range of the camera device may be consistent with or overlap with the scan range of the radar device.
这里,可以基于待标注三维点云数据对应的坐标系与校验图像对应的坐标系之间的映射关系,将世界坐标系下待标注对象对应的三维标注框信息映射到图像坐标系下,得到映射后的图像坐标范围,在确定校验图像中的待标注对象所在图像坐标范围的情况下,可以基于这一图像坐标范围与映射后的图像坐标范围之间的匹配结果对三维标注框信息的准确性进行校验。Here, based on the mapping relationship between the coordinate system corresponding to the 3D point cloud data to be labeled and the coordinate system corresponding to the verification image, the 3D labeling frame information corresponding to the object to be labelled in the world coordinate system can be mapped to the image coordinate system to obtain The mapped image coordinate range, in the case of determining the image coordinate range where the object to be marked in the verification image is located, can be based on the matching result between the image coordinate range and the mapped image coordinate range. Check for accuracy.
其中,在两个坐标范围的匹配度较高的情况下,说明标注的三维标注框信息的准确性也越好,同理,在两个坐标范围的匹配度越低的情况下,说明标注的三维标注框信息 的准确性也越差。因此,本公开实施例提供的点云标注的方法可以在两个坐标范围不一致的情况下,响应针对待标注三维点云数据对应的待调整视图中的第二二维标注框的调整操作,对调整后的参数信息进行调整,直至映射后的图像坐标范围与所述校验图像中的所述待标注对象所在图像坐标范围一致,所述调整后的参数信息是通过对所述初始参数信息进行调整得到的,从而进一步提升标注的三维标注框信息的准确性。Among them, in the case where the matching degree of the two coordinate ranges is high, the accuracy of the marked 3D annotation frame information is also better. Similarly, in the case where the matching degree of the two coordinate ranges is lower, the marked The accuracy of the 3D callout box information is also worse. Therefore, the point cloud labeling method provided by the embodiment of the present disclosure can respond to the adjustment operation of the second two-dimensional labeling frame in the view to be adjusted corresponding to the to-be-labeled three-dimensional point cloud data when the two coordinate ranges are inconsistent. The adjusted parameter information is adjusted until the image coordinate range after mapping is consistent with the image coordinate range where the object to be marked in the verification image is located. Adjusted to further improve the accuracy of the marked 3D annotation frame information.
本公开实施例提供的点云标注的方法不仅可以基于校验图像实现三维标注框信息的校准,还可以基于参考图像为鸟瞰图中的待标注对象的框选操作提供参考,从而确定第一二维标注框信息。其中,这里的参考图像和校验图像可以采用的是同一图像,有关参考图像的采集等过程可以参见上述描述。The point cloud labeling method provided by the embodiment of the present disclosure can not only realize the calibration of the three-dimensional labeling frame information based on the verification image, but also provide a reference for the frame selection operation of the object to be labelled in the bird's-eye view based on the reference image, so as to determine the first and second Dimensional callout box information. The reference image and the verification image here may be the same image, and the above description may be referred to for the process of collecting the reference image and the like.
这里,基于参考图像对应的坐标系与鸟瞰图对应的坐标系之间的映射关系,可以将待标注对象在参考图像中的图像位置信息转换为待标注对象在鸟瞰图中的参考位置信息,依赖于该参考位置信息用户可以在鸟瞰图中进行待标注对象的框选操作,特别是针对稀疏三维点云而言,参考图像所提供的参考位置信息可以大大提升鸟瞰图下的待标注对象标注的准确性。Here, based on the mapping relationship between the coordinate system corresponding to the reference image and the coordinate system corresponding to the bird's-eye view, the image position information of the object to be labeled in the reference image can be converted into the reference position information of the object to be labeled in the bird's-eye view, depending on Due to the reference position information, users can perform the frame selection operation of the objects to be marked in the bird's-eye view, especially for sparse 3D point clouds, the reference position information provided by the reference image can greatly improve the quality of the objects to be marked in the bird's-eye view. accuracy.
考虑到在实际应用中,对于一帧待标注三维点云数据而言,往往具有多个待标注对象,而对于同属一个对象类型的多个待标注对象而言,其三维标注框信息存在很大程度上的一致性。基于此,本公开实施例提供了一种对象间传递机制以基于已标注对象的三维标注框信息实现未标注对象的三维标注,可以按照如下步骤实现:Considering that in practical applications, for a frame of 3D point cloud data to be labeled, there are often multiple objects to be labeled, and for multiple objects to be labeled of the same object type, there is a large amount of 3D labeling frame information. degree of consistency. Based on this, the embodiments of the present disclosure provide an inter-object transfer mechanism to realize the 3D annotation of unlabeled objects based on the 3D annotation frame information of the labeled objects, which can be implemented according to the following steps:
步骤一、将待标注对象对应的三维标注框信息,展示在待标注三维点云数据对应的三维视图中,并将待标注对象作为已标注对象;Step 1: Display the 3D labeling frame information corresponding to the object to be labelled in the 3D view corresponding to the 3D point cloud data to be labelled, and use the object to be labelled as the labelled object;
步骤二、根据已标注对象的三维标注框信息,确定三维视图中的其它待标注对象的三维标注框信息;其它待标注对象与已标注对象属于同一对象类型;Step 2: According to the 3D annotation frame information of the marked object, determine the 3D annotation frame information of other objects to be marked in the 3D view; the other objects to be marked belong to the same object type as the marked object;
步骤三、将其它待标注对象的三维标注框信息展示在三维视图中。Step 3: Display the 3D labeling frame information of other objects to be labelled in the 3D view.
这里,在展示有待标注对象对应的三维标注框信息的三维视图中,可以将该待标注对象作为已标注对象,这样,可以根据已标注对象的三维标注框信息,可以确定三维视图中与该已标注对象同属一个对象类型的其它待标注对象的三维标注框信息,并能够在三维视图中将其它待标注对象的三维标注框信息得以展示。Here, in the 3D view showing the 3D annotation frame information corresponding to the object to be annotated, the object to be annotated can be regarded as a marked object. The annotated object belongs to the 3D annotation frame information of other objects to be annotated of the same object type, and the 3D annotation frame information of other objects to be annotated can be displayed in the 3D view.
本公开实施例中,为了便于在三维视图下进行对象间的传递机制,可以基于复制-粘贴的指令配对方式来实现,可以包括如下步骤:In the embodiment of the present disclosure, in order to facilitate the transfer mechanism between objects in the three-dimensional view, it can be implemented based on a copy-paste instruction pairing method, which can include the following steps:
步骤一、响应于针对三维视图中已标注对象的三维标注框的复制指令,复制已标注对象的三维标注框信息;Step 1, in response to the copy instruction for the 3D annotation frame of the marked object in the 3D view, copy the 3D annotation frame information of the marked object;
步骤二、响应于针对三维视图中其它待标注对象的粘贴指令,根据已标注对象在三维视图中的三维标注框信息,对其它待标注对象进行三维标注,得到其它待标注对象的三维标注框信息。Step 2: In response to the pasting instructions for other objects to be marked in the 3D view, according to the 3D annotation frame information of the marked objects in the 3D view, perform 3D annotation on the other objects to be annotated, and obtain the 3D annotation frame information of the other objects to be annotated. .
通过这种方式,考虑到同一对象类型的对象其对应的三维标注框信息大致是相同的,因此,本公开实施例中,可以基于已标注对象的三维标注框的复制操作,将其三维标注框信息粘贴到同属一个对象类型的其它待标注对象上,从而可以快速实现对其它待标注对象的三维标注,标注效率进一步得以提升。In this way, considering that objects of the same object type have roughly the same 3D annotation frame information, therefore, in this embodiment of the present disclosure, the 3D annotation frame of the marked object can be copied based on the copy operation of the 3D annotation frame of the marked object. The information is pasted to other objects to be labeled that belong to the same object type, so that 3D labeling of other objects to be labeled can be quickly realized, and the labeling efficiency is further improved.
这里,通过针对已标注对象的三维标注框的复制操作,实现已标注对象的三维标注框信息的复制,这样,在针对其它待标注对象进行粘贴操作之后,即可以根据已标注对象对应的三维标注框信息对其它待标注对象进行三维标注,得到对应的三维标注框信息。Here, the 3D annotation frame information of the annotated object is copied by copying the 3D annotation frame of the annotated object. In this way, after the paste operation is performed for other objects to be annotated, the 3D annotation corresponding to the annotated object can be copied according to the The frame information performs 3D labeling on other objects to be labelled to obtain corresponding 3D labeling frame information.
例如,针对道路所拍摄的一帧点云数据而言,可以基于上述对象间传递机制实现途径道路上的各个车辆的三维标注。For example, for a frame of point cloud data captured on a road, the three-dimensional labeling of each vehicle on the road can be implemented based on the above-mentioned inter-object transfer mechanism.
考虑到雷达设备在进行三维点云数据扫描的过程中,对于距离雷达设备越近的待标 注对象其对应的三维点云数据越稠密、对应的数据完整性程度越高,反之,对距离雷达设备越远的待标注对象其对应的三维点云数据越稀疏、对应的数据完整性程度越低,而数据完整性较高的待标注对象一定程度上会比数据完整性较低的待标注对象更容易标注。基于此,本公开实施例可以是按照距离远近的标注方式进行同一对象类型的三维标注,这样可以显著减少位姿和高度等信息在标注时的调整,从而在确保一帧待标注三维点云数据的标注准确性的前提下,提升标注效率。Considering that in the process of scanning the 3D point cloud data by the radar equipment, the closer the distance to the radar equipment to the object to be marked, the denser the corresponding 3D point cloud data and the higher the corresponding data integrity. The farther the object to be labeled, the sparser the corresponding 3D point cloud data, and the lower the corresponding data integrity, and the object to be labeled with higher data integrity will be more accurate to a certain extent than the object to be labeled with lower data integrity. Easy to label. Based on this, the embodiment of the present disclosure can perform 3D annotation of the same object type according to the annotation method of distance, which can significantly reduce the adjustment of pose and height and other information during annotation, so as to ensure that one frame of 3D point cloud data to be annotated On the premise of the accuracy of labeling, the labeling efficiency is improved.
考虑到在实际应用中,对于相邻帧待标注三维点云数据而言,相邻帧的待标注对象的三维标注框信息存在很大程度上的一致性。基于此,本公开实施例提供了一种帧间传递机制以基于已标注对象的三维标注框信息实现未标注对象的三维标注,可以按照如下步骤实现:Considering that in practical applications, for the 3D point cloud data to be annotated in adjacent frames, the 3D annotation frame information of the objects to be annotated in adjacent frames is largely consistent. Based on this, the embodiments of the present disclosure provide an inter-frame transfer mechanism to realize the 3D labeling of unlabeled objects based on the 3D labeling frame information of the labeled objects, which can be implemented according to the following steps:
步骤一、将待标注对象对应的三维标注框信息,展示在当前帧待标注三维点云数据对应的第一三维视图中,并将待标注对象作为已标注对象;其中,不同帧待标注三维点云数据对应不同的三维视图;Step 1: Display the 3D labeling frame information corresponding to the object to be labelled in the first 3D view corresponding to the 3D point cloud data to be labelled in the current frame, and use the object to be labelled as the labelled object; wherein, the 3D points to be labelled in different frames Cloud data corresponds to different 3D views;
步骤二、响应于针对第一三维视图中已标注对象的三维标注框的复制指令,复制已标注对象的三维标注框信息;Step 2, in response to the copy instruction for the 3D annotation frame of the marked object in the first 3D view, copy the 3D annotation frame information of the marked object;
步骤三、在切换到下一帧待标注三维点云数据对应的第二三维视图的情况下,响应于针对第二三维视图中其它待标注对象的粘贴指令,根据已标注对象在第一三维视图中的三维标注框信息,对第二三维视图中的其它待标注对象进行三维标注,得到其它待标注对象的三维标注框信息。Step 3: In the case of switching to the second 3D view corresponding to the 3D point cloud data to be labeled in the next frame, in response to the paste instruction for other objects to be labeled in the second 3D view, according to the labeled objects in the first 3D view. The 3D labeling frame information in the second 3D view is 3D labelled for other objects to be labelled in the second 3D view to obtain the 3D labelling frame information of the other objects to be labelled.
通过这种方式,考虑到相邻帧待标注三维点云数据之间存在待标注对象的信息传递,可以基于一帧待标注三维点云数据中已标注对象的三维标注框的复制操作,将其三维标注框信息粘贴到下一帧的其它待标注对象上,从而可以快速实现对不同帧中的相同类型待标注对象的三维标注,标注效率进一步得以提升。In this way, considering that there is information transfer of the object to be labeled between the 3D point cloud data to be labeled in adjacent frames, it can be based on the copy operation of the 3D labeled frame of the labeled object in the 3D point cloud data to be labeled in one frame. The 3D annotation frame information is pasted to other objects to be annotated in the next frame, so that the 3D annotation of the same type of objects to be annotated in different frames can be quickly realized, and the annotation efficiency is further improved.
这里,在展示有待标注对象对应的三维标注框信息的第一三维视图中,可以将该待标注对象作为已标注对象,这样,可以通过针对已标注对象的三维标注框的复制操作,实现已标注对象的三维标注框信息的复制,这样,在针对下一帧待标注三维点云数据对应的第二三维视图中的其它待标注对象进行粘贴操作之后,即可以根据已标注对象对应的三维标注框信息对其它待标注对象进行三维标注,得到对应的三维标注框信息。Here, in the first 3D view showing the information of the 3D annotation frame corresponding to the object to be annotated, the object to be annotated can be regarded as the marked object. In this way, the marked object can be copied through the operation of copying the 3D annotation frame of the marked object. Copy the 3D annotation frame information of the object. In this way, after the paste operation is performed on other objects to be annotated in the second 3D view corresponding to the 3D point cloud data to be annotated in the next frame, the 3D annotation frame corresponding to the marked object can be copied. The information is three-dimensional marked on other objects to be marked to obtain the corresponding three-dimensional marked frame information.
可知的是,基于帧间传递机制可以在完成针对一帧待标注三维点云数据中待标注对象的标注后,基于上述复制-粘贴的配对操作实现针对下一帧待标注三维点云数据中的待标注对象的标注。It can be known that, based on the inter-frame transfer mechanism, after completing the labeling of the objects to be labeled in the 3D point cloud data to be labeled in one frame, the pairing operation based on the above copy-paste can be used to realize the next frame of the 3D point cloud data to be labeled. Annotation of the object to be annotated.
需要说明的是,标注的相邻帧的待标注对象可以是同一对象,也可以不是同一对象。在实际应用中,除了可以基于上述复制-粘贴的配对操作实现帧间传递,还可以是在切换到下一帧待标注三维点云数据所在三维视图的情况下,结合目标跟踪或目标检测机制,对下一帧待标注三维点云数据中与上一帧已标注对象同属一个对象的待标注对象进行自动标注。It should be noted that the to-be-labeled objects of the labeled adjacent frames may be the same object or may not be the same object. In practical applications, in addition to realizing inter-frame transfer based on the above copy-paste pairing operation, it can also be combined with the target tracking or target detection mechanism in the case of switching to the 3D view where the 3D point cloud data to be marked in the next frame is located. Automatically label the objects to be labeled that belong to the same object as the labeled object in the previous frame in the 3D point cloud data to be labeled in the next frame.
鉴于标注对象的操作复杂性,如何合理地利用已标注的标注框来减少操作次数是一个非常重要的问题,因此,本公开实施例提供的一种标注传递机制,该标注传递机制包括对象间传递机制和帧间传递机制。对象间传递机制可以是:在单帧中,可以在来自同一类别的对象之间传递标注,尤其是在同一车道上。帧间传递机制可以是:在连续帧中,可以传递来自先前已标记帧的标注。这两种传递机制都可以显著减少旋转角和高度在标注时的调整,避免点云稀疏区域的标签丢失或不准确。此外,帧间传递机制还可以实现连续帧的标注之间一对一的对应,使得标签既可以用于三维目标检测,也可以用于三维目标跟随。在本公开实施例中,通过对象间传递机制和帧间传递机制,可以将大多数三 维标注任务简化为鸟瞰图的标注调整,大大减少了不必要的操作,且对电子设备的负荷影响小,大大提升了标注效率和准确率。In view of the operation complexity of annotating objects, it is a very important issue how to reasonably use the annotated annotation frame to reduce the number of operations. Therefore, an annotation transfer mechanism provided by an embodiment of the present disclosure includes an inter-object transfer mechanism. mechanism and inter-frame transfer mechanism. The inter-object transfer mechanism can be: in a single frame, annotations can be transferred between objects from the same class, especially on the same lane. The inter-frame transfer mechanism can be that, in successive frames, annotations from previously marked frames can be transferred. Both transfer mechanisms can significantly reduce the adjustment of rotation angle and height during labeling, avoiding missing or inaccurate labels in sparse regions of the point cloud. In addition, the inter-frame transfer mechanism can also realize one-to-one correspondence between the labels of consecutive frames, so that labels can be used for both 3D object detection and 3D object tracking. In the embodiments of the present disclosure, through the inter-object transfer mechanism and the inter-frame transfer mechanism, most 3D labeling tasks can be simplified into labeling adjustments for bird's-eye views, which greatly reduces unnecessary operations and has little impact on the load of electronic devices. The labeling efficiency and accuracy are greatly improved.
在一些实施方式中,S103中的响应针对所述鸟瞰图中的待标注对象的框选操作,确定第一二维标注框信息,可以通过以下方式实现:获取参考图像;所述参考图像与所述待标注三维点云数据来源于同一场景;基于所述待标注对象在所述参考图像中的图像位置信息、以及所述参考图像对应的坐标系与所述鸟瞰图对应的坐标系之间的映射关系,确定所述待标注对象在所述鸟瞰图中的参考位置信息;响应针对所述鸟瞰图中的所述待标注对象的框选操作,基于所述参考位置信息确定所述第一二维标注框信息。In some embodiments, determining the first two-dimensional annotation frame information in response to the frame selection operation of the object to be annotated in the bird's-eye view in S103 may be implemented by: acquiring a reference image; The three-dimensional point cloud data to be marked comes from the same scene; based on the image position information of the object to be marked in the reference image and the relationship between the coordinate system corresponding to the reference image and the coordinate system corresponding to the bird's-eye view mapping relationship, determine the reference position information of the object to be marked in the bird's-eye view; in response to the frame selection operation for the object to be marked in the bird's-eye view, determine the first and second based on the reference position information Dimensional callout box information.
通过这种方式,可以基于获取的参考图像,确定待标注对象在鸟瞰图中的参考位置信息,基于这一参考位置信息来指导用户针对鸟瞰图中的待标注对象进行框选操作,这样,所框选出来的第一二维标注框将更大程度上的覆盖待标注对象所对应的点云数据,从而可以提升标注的准确度。In this way, the reference position information of the object to be marked in the bird's-eye view can be determined based on the obtained reference image, and based on the reference position information, the user is guided to perform a frame selection operation for the object to be marked in the bird's-eye view. The first two-dimensional labeling frame selected by the frame will cover the point cloud data corresponding to the object to be labelled to a greater extent, thereby improving the labeling accuracy.
为了更为方便的实现诸如鸟瞰图的框选操作、待调整视图的调整操作、三维视图的展示操作等,本公开实施例提供的点云标注的方法可以结合用户交互界面实现视图展示。In order to more conveniently implement operations such as frame selection of bird's-eye views, adjustment operations of views to be adjusted, display operations of 3D views, etc., the point cloud labeling method provided by the embodiments of the present disclosure can realize view display in combination with a user interface.
这里,在用户交互界面的布局上,由于全局场景的鸟瞰图需要大范围的展示且点云数据大范围的展示会提高标注的准确性,因此,为了使鸟瞰图和点云数据都能够有清晰的显示,在用户交互界面上可以设置相应的视图切换按钮,实现鸟瞰图和三维视图模式之间的切换,在切换到鸟瞰图这一视图下,可以进行第一二维标注框的框选,确定第一二维标注框信息,在切换到三维视图这一视图下,可以展示标注的三维标注框信息。Here, in the layout of the user interaction interface, since the bird's-eye view of the global scene needs to be displayed on a large scale and the large-scale display of point cloud data will improve the accuracy of the annotation, in order to make the bird's-eye view and point cloud data clear display, the corresponding view switching button can be set on the user interaction interface to realize the switching between the bird's-eye view and the three-dimensional view mode. The first 2D annotation frame information is determined, and in the view of switching to the 3D view, the marked 3D annotation frame information can be displayed.
除此之外,本公开实施例还可以在用户交互界面展示相应的待调整视图,在该待调整视图下,可以进行第二二维标注框的调整,确定第二二维标注框信息,并对待标注对象的初始参数信息进行调整,这一调整操作可以同步在鸟瞰图和三维视图下,为快速标注提供了保证。In addition, the embodiment of the present disclosure can also display the corresponding view to be adjusted on the user interface, and in the view to be adjusted, the second two-dimensional annotation frame can be adjusted, the information of the second two-dimensional annotation frame can be determined, and The initial parameter information of the object to be annotated is adjusted. This adjustment operation can be synchronized under the bird's-eye view and the three-dimensional view, which provides a guarantee for fast annotation.
为了进一步提升标注的准确性以及标注效率,不管是针对用户交互界面上展示的三维视图还是二维视图均可以对当前所需要操作的局部点云进行高亮显示设置,例如,可以在对鸟瞰图执行框选操作之后,可以在用户交互界面所呈现鸟瞰图上将框选所对应的点云进行高亮,还可以同步将待调整视图上所对应的点云进行高亮,从而便于执行调整操作。In order to further improve the accuracy and efficiency of labeling, whether it is for the 3D view or the 2D view displayed on the user interface, the local point cloud that needs to be operated can be highlighted. After the frame selection operation is performed, the point cloud corresponding to the frame selection can be highlighted on the bird's-eye view displayed on the user interface, and the point cloud corresponding to the view to be adjusted can also be highlighted simultaneously, so as to facilitate the adjustment operation. .
除此之外,为了更进一步的提升用户与界面的友好交互,本公开实施例还可以对各个标注框进行相关长度、宽度等信息的标注。In addition, in order to further improve the friendly interaction between the user and the interface, the embodiments of the present disclosure may also mark relevant length, width and other information for each markup frame.
在对用户交互界面中的鸟瞰图进行标注的情况下,可以采用轻量级的基于广域网(web)的标注工具,该标注工具能够较少的依赖计算机的性能,从而能够在普通计算机上平稳运行。In the case of labeling the bird's-eye view in the user interface, a lightweight WAN-based labeling tool can be used. The labeling tool can be less dependent on the performance of the computer, so that it can run smoothly on ordinary computers. .
考虑到本公开实施例提供的点云标注的方法涉及到复杂的交互操作,为了有效的实现正确识别对象和准确标注对象这两个技术目的,在一些应用中,可以分为四个步骤,即查找、定位、调整和验证,前两个步骤可以是在全局视图中识别和定位对象,第三步是准确标注,最后一步是确保所有标注都足够准确,基于上述交互操作,可以确保标注人员在每个阶段都集中精力完成最重要的任务,从而进一步保证标注结果的准确性。Considering that the method for point cloud labeling provided by the embodiments of the present disclosure involves complex interactive operations, in order to effectively realize the two technical purposes of correctly identifying objects and accurately labeling objects, in some applications, it can be divided into four steps, namely, Finding, locating, adjusting and verifying, the first two steps can be to identify and locate objects in the global view, the third step is to label accurately, and the last step is to ensure that all labels are accurate enough. Each stage focuses on the most important tasks, further ensuring the accuracy of the annotation results.
为了对用于检测和跟随的点云进行标注,需要首先正确识别待检测对象,然后再准确标注其位置、大小、方向、类别等。In order to label the point cloud for detection and following, it is necessary to first correctly identify the object to be detected, and then accurately label its position, size, orientation, category, etc.
在实施过程中,首先可以将三维点云数据所反映的形状信息和RGB图像中的语义信息结合起来,来查找目标对象。然后在三维点云数据对应的鸟瞰图中画出边界框,并根据局部点云生成高度信息来初步定位。其中,利用待标注三维点云数据对应的坐标系,与RGB图像对应的坐标系之间的映射关系,从二维图像得到的三维椎体可以用来估计 搜索哪些点属于目标物体。随后,在局部点云的前视图和侧视图中调整粗略标注的框(即画出的边界框,对应上述实施例中的第一二维标注框信息),并将二维视图中的调整映射到三维框(对应上述实施例中的三维标注框信息)中。最后,结合RGB图像,验证点云在不同立体视角下的标注结果是否准确,从而完成标注。In the implementation process, the shape information reflected by the 3D point cloud data and the semantic information in the RGB image can be combined to find the target object. Then, the bounding box is drawn in the bird's-eye view corresponding to the 3D point cloud data, and the height information is generated according to the local point cloud for preliminary positioning. Among them, using the mapping relationship between the coordinate system corresponding to the 3D point cloud data to be marked and the coordinate system corresponding to the RGB image, the 3D pyramid obtained from the 2D image can be used to estimate which points to search for belong to the target object. Then, adjust the roughly annotated frame (that is, the drawn bounding frame, corresponding to the first two-dimensional annotated frame information in the above embodiment) in the front view and side view of the local point cloud, and map the adjustment in the two-dimensional view into the 3D frame (corresponding to the 3D annotation frame information in the above embodiment). Finally, combined with RGB images, it is verified whether the annotation results of the point cloud under different stereo perspectives are accurate, so as to complete the annotation.
图2示出了本公开实施例提供的一种点云标注的流程示意图,如图2所示在,在得到图2中最左侧的RGB图像和三维点云数据对应的鸟瞰图之后,其中,三维点云数据与RGB图像相对应,可以对RGB图像和鸟瞰图依次进行(a)、(b)、(c)以及(d)的四个步骤的处理,这四个步骤依次对应查找、定位、调整和验证。在(a)中可以从鸟瞰图中直接确定出较近的物体(较近的目标物,较近的目标物在RGB图像中占用的尺寸较大)在鸟瞰图中的位置,而较远的物体(较远的目标物,较远的目标物在RGB图像中占用的尺寸较小)在鸟瞰图中的位置不容易确定,因此可以确定RGB图像中较远物体的位置,从而基于RGB图像中较远物体的位置,确定鸟瞰图中较远的物体的位置。在另一些实施方式中,在两个物体的位置都容易在鸟瞰图中确定出的情况下,例如两个物体均较近,可以从鸟瞰图中直接确定两个物体的位置,或者,在两个物体的位置都不容易在鸟瞰图中确定出的情况下,例如两个物体都较远,则可以均基于RGB图像中两个物体的位置,确定鸟瞰图中这两个物体的位置。在得到鸟瞰图中目标物的位置后,在(b)中可以定位到目标物的位置,以得到上述的第一二维标注框信息。其中,(a)所示出的鸟瞰图和(b)所示出的鸟瞰图为不同视角下的鸟瞰图。之后,再执行(b)到(c)之间的转化,通过(b)得到鸟瞰图中目标物的位置,对目标物的两个侧面的尺寸进行调整,两个侧面可以分别为正面、侧面、俯视面中的任意两个,从而得到每一目标物分别对应的两个侧面尺寸,如(c)所示。在其它实施例中,可以对三个侧面的尺寸进行调整。之后,可以对某一个物体的三维标注框信息进行验证,例如,在(d)中,可以根据较近物体的两个侧面尺寸,确定出RGB图像中较近物体的三维标注框信息。FIG. 2 shows a schematic flowchart of a point cloud labeling provided by an embodiment of the present disclosure. As shown in FIG. 2 , after obtaining the bird's-eye view corresponding to the leftmost RGB image and the three-dimensional point cloud data in FIG. 2 , wherein , the three-dimensional point cloud data corresponds to the RGB image, and the four steps (a), (b), (c) and (d) can be processed in turn for the RGB image and the bird's-eye view. These four steps correspond to search, Locate, adjust and verify. In (a), the position of the closer object (the closer object, the closer object occupies a larger size in the RGB image) can be directly determined from the bird's eye view in the bird's eye view, and the farther The position of the object (the farther object, the farther object occupies a smaller size in the RGB image) is not easy to determine in the bird's eye view, so the position of the farther object in the RGB image can be determined, based on the RGB image. The position of distant objects, which determines the position of distant objects in the bird's eye view. In other embodiments, when the positions of the two objects are easy to be determined in the bird's eye view, for example, the two objects are close, the positions of the two objects can be directly determined from the bird's eye view, or, in the case of the two objects When the positions of the two objects are not easy to determine in the bird's eye view, for example, both objects are far away, the positions of the two objects in the bird's eye view can be determined based on the positions of the two objects in the RGB image. After obtaining the position of the target in the bird's-eye view, the position of the target can be located in (b) to obtain the above-mentioned first two-dimensional labeling frame information. The bird's-eye view shown in (a) and the bird's-eye view shown in (b) are bird's-eye views from different viewing angles. After that, perform the conversion between (b) and (c), obtain the position of the target in the bird's-eye view through (b), and adjust the size of the two sides of the target. The two sides can be the front and the side respectively. , any two of the top-view planes, so as to obtain two side dimensions corresponding to each target, as shown in (c). In other embodiments, the dimensions of the three sides may be adjusted. After that, the 3D annotation frame information of a certain object can be verified. For example, in (d), the 3D annotation frame information of the closer object in the RGB image can be determined according to the two side dimensions of the closer object.
在一些应用中,可以采用轻量级的基于web的标注工具实现标注,上述标注工具可以在普通计算机上平稳运行,这一设计对用户十分友好,为以各个阶段的顺序进行提供了可能。In some applications, light-weight web-based annotation tools can be used to realize annotation, and the above annotation tools can run smoothly on ordinary computers. This design is very user-friendly and provides the possibility to perform in the order of each stage.
本领域技术人员可以理解,在一些实施方式的上述方法中,各步骤的撰写顺序并不意味着严格的执行顺序而对实施过程构成任何限定,各步骤的执行顺序应当以其功能和可能的内在逻辑确定。Those skilled in the art can understand that, in the above methods of some embodiments, the writing order of each step does not mean a strict execution order and constitutes any limitation on the implementation process, and the execution order of each step should be based on its function and possible inherent Logical OK.
基于同一发明构思,本公开实施例中还提供了与点云标注的方法对应的点云标注的装置,由于本公开实施例中的装置解决问题的原理与本公开实施例上述点云标注的方法相似,因此装置的实施可以参见方法的实施,重复之处不再进行详细说明。Based on the same inventive concept, the embodiment of the present disclosure also provides a point cloud labeling device corresponding to the point cloud labeling method, because the principle of solving the problem by the device in the embodiment of the present disclosure is the same as the above-mentioned point cloud labeling method in the embodiment of the present disclosure. Similar, so the implementation of the device can refer to the implementation of the method, and the repeated places will not be described in detail.
在本公开实施例中,提出了一个系统化的标注流程,以解决三维标注任务中的两个关键问题:正确识别物体和准确标注物体。根据数据和任务的特点,提供了一种清晰的用户交互界面和标注传递机制,使标注人员更容易在标注的每个阶段集中精力做更简单的工作,用更少的操作完成。In the embodiments of the present disclosure, a systematic labeling process is proposed to solve two key problems in 3D labeling tasks: correctly identifying objects and accurately labeling objects. According to the characteristics of data and tasks, it provides a clear user interface and annotation delivery mechanism, which makes it easier for annotators to concentrate on simpler tasks at each stage of annotation and complete them with fewer operations.
图3示出了本公开实施例提供的一种点云标注的装置的结构示意图,如图3所示,该装置包括:获取模块201、生成模块202、第一确定模块203和标注模块204;其中,FIG. 3 shows a schematic structural diagram of a point cloud labeling device provided by an embodiment of the present disclosure. As shown in FIG. 3 , the device includes: an acquisition module 201 , a generation module 202 , a first determination module 203 , and a labeling module 204 ; in,
获取模块201,配置为获取待标注三维点云数据;The obtaining module 201 is configured to obtain the three-dimensional point cloud data to be marked;
生成模块202,配置为基于待标注三维点云数据,生成待标注三维点云数据对应的鸟瞰图;The generating module 202 is configured to generate a bird's-eye view corresponding to the three-dimensional point cloud data to be marked based on the three-dimensional point cloud data to be marked;
第一确定模块203,配置为响应针对鸟瞰图中的待标注对象的框选操作,确定第一二维标注框信息;基于被框选的待标注对象的三维点云数据,确定待标注对象的初始参数信息;The first determination module 203 is configured to determine the first two-dimensional annotation frame information in response to the frame selection operation for the object to be marked in the bird's-eye view; initial parameter information;
标注模块204,配置为基于第一二维标注框信息,以及初始参数信息,确定待标注对象对应的三维标注框信息。The labeling module 204 is configured to determine the three-dimensional labeling frame information corresponding to the object to be labelled based on the first two-dimensional labeling frame information and the initial parameter information.
本公开实施例在二维视图(对应鸟瞰图)下执行的框选操作,相比相关技术中直接对三维点云数据进行标注所存在的操作复杂度高的问题,二维视图下执行框选操作的复杂度较低,且一定程度上可以确保所框选待标注对象的准确性,这样,基于框选确定的第一二维标注框信息再结合确定的初始参数信息即可以实现针对待标注对象进行三维标注框的自动标注,标注效率和准确率均较高。The frame selection operation performed in the embodiment of the present disclosure in a two-dimensional view (corresponding to a bird's-eye view), compared with the problem of high operational complexity in directly annotating 3D point cloud data in the related art, performing frame selection in a two-dimensional view The complexity of the operation is low, and to a certain extent, the accuracy of the selected objects to be marked can be ensured. In this way, the first two-dimensional marked frame information determined based on the frame selection can be combined with the determined initial parameter information. The object is automatically marked with a 3D labeling frame, and the labeling efficiency and accuracy are high.
在一种可能的实施方式中,上述装置还包括:In a possible implementation, the above device further includes:
第二确定模块205,配置为基于待标注对象的三维点云数据,生成待标注三维点云数据对应的待调整视图;待调整视图至少包括正视图和侧视图中的一种;根据生成的待调整视图,确定待标注对象在待调整视图中的第二二维标注框信息。The second determination module 205 is configured to generate, based on the three-dimensional point cloud data of the object to be marked, a view to be adjusted corresponding to the three-dimensional point cloud data to be marked; the view to be adjusted includes at least one of a front view and a side view; Adjusting the view, and determining the second two-dimensional labeling frame information of the object to be labelled in the view to be adjusted.
在一种可能的实施方式中,标注模块204,配置为按照以下步骤基于第一二维标注框信息,以及初始参数信息,确定待标注对象对应的三维标注框信息:In a possible implementation, the labeling module 204 is configured to determine the three-dimensional labeling frame information corresponding to the object to be labelled based on the first two-dimensional labeling frame information and the initial parameter information according to the following steps:
基于第一二维标注框信息、第二二维标注框信息以及初始参数信息,确定待标注对象对应的三维标注框信息。Based on the first 2D annotation frame information, the second 2D annotation frame information and the initial parameter information, the 3D annotation frame information corresponding to the object to be annotated is determined.
在一种可能的实施方式中,标注模块204,配置为基于第一二维标注框信息、第二二维标注框信息以及初始参数信息,确定待标注对象对应的三维标注框信息,包括:In a possible implementation, the labeling module 204 is configured to determine the three-dimensional labeling frame information corresponding to the object to be labelled based on the first two-dimensional labeling frame information, the second two-dimensional labeling frame information and the initial parameter information, including:
响应针对第二二维标注框信息的调整操作,基于初始参数信息,确定待标注对象的调整后的参数信息;In response to the adjustment operation for the second two-dimensional annotation frame information, based on the initial parameter information, determine the adjusted parameter information of the object to be annotated;
基于第一二维标注框信息、第二二维标注框信息以及调整后的参数信息,确定待标注对象对应的三维标注框信息。Based on the first 2D annotation frame information, the second 2D annotation frame information and the adjusted parameter information, the 3D annotation frame information corresponding to the object to be annotated is determined.
在一种可能的实施方式中,标注模块204,配置为按照以下步骤确定待标注对象的调整后的参数信息:In a possible implementation, the labeling module 204 is configured to determine the adjusted parameter information of the object to be labelled according to the following steps:
响应针对待调整视图中的第二二维标注框的调整操作,确定与调整操作对应的参数调整信息;determining parameter adjustment information corresponding to the adjustment operation in response to an adjustment operation for the second two-dimensional callout frame in the view to be adjusted;
基于参数调整信息对初始参数信息进行调整,得到调整后的参数信息。The initial parameter information is adjusted based on the parameter adjustment information to obtain adjusted parameter information.
在一种可能的实施方式中,标注模块204,配置为按照以下步骤基于第一二维标注框信息、第二二维标注框信息以及调整后的参数信息,确定待标注对象对应的三维标注框信息:In a possible implementation, the labeling module 204 is configured to determine the 3D labeling frame corresponding to the object to be labelled based on the first 2D labeling frame information, the second 2D labeling frame information and the adjusted parameter information according to the following steps information:
基于调整后的参数信息所对应的三维点云数据,与第一二维标注框信息所对应的二维点云数据之间的第一对应关系,确定待标注对象对应的三维标注框的底面的参数信息;Based on the first correspondence between the 3D point cloud data corresponding to the adjusted parameter information and the 2D point cloud data corresponding to the first 2D annotation frame information, determine the size of the bottom surface of the 3D annotation frame corresponding to the object to be annotated. Parameter information;
基于调整后的参数信息所对应的三维点云数据,与第二二维标注框信息所对应的二维点云数据之间的第二对应关系,确定待标注对象对应的三维标注框的侧面和正面的参数信息中的至少之一;Based on the second correspondence between the 3D point cloud data corresponding to the adjusted parameter information and the 2D point cloud data corresponding to the second 2D annotation frame information, determine the side surface and the 3D annotation frame corresponding to the object to be annotated. At least one of the positive parameter information;
将确定的三维标注框的底面的参数信息、以及三维标注框的侧面和正面的参数信息中的至少之一进行组合,得到三维标注框信息。Combining at least one of the determined parameter information of the bottom surface of the 3D annotation frame and the parameter information of the side surface and the front surface of the 3D annotation frame to obtain the 3D annotation frame information.
在一种可能的实施方式中,第二确定模块205,还配置为:In a possible implementation manner, the second determining module 205 is further configured to:
在确定待标注对象在待调整视图中的第二二维标注框信息之后,基于确定的待标注对象的第二二维标注框信息,展示标注有与所述第二二维标注框信息对应的第二二维标注框的待调整视图。After determining the second two-dimensional callout frame information of the object to be marked in the view to be adjusted, based on the determined second two-dimensional callout frame information of the to-be-marked object, display the information marked with the second two-dimensional callout frame corresponding to the second two-dimensional callout frame information. The to-be-adjusted view of the second 2D callout box.
在一种可能的实施方式中,上述装置还包括:In a possible implementation, the above device further includes:
校验模块206,配置为在确定待标注对象对应的三维标注框信息之后,获取校验图像;校验图像与待标注三维点云数据来源于同一场景;基于待标注三维点云数据对应的坐标系,与校验图像对应的坐标系之间的映射关系,将待标注对象对应的三维标注框信 息映射至校验图像中,确定映射后的图像坐标范围;根据映射后的图像坐标范围与校验图像中的待标注对象所在图像坐标范围,对三维标注框信息的准确性进行校验。The verification module 206 is configured to obtain a verification image after determining the three-dimensional annotation frame information corresponding to the object to be marked; the verification image and the three-dimensional point cloud data to be marked originate from the same scene; based on the coordinates corresponding to the three-dimensional point cloud data to be marked system, the mapping relationship between the coordinate system corresponding to the verification image, map the three-dimensional annotation frame information corresponding to the object to be marked into the verification image, and determine the coordinate range of the image after mapping; The image coordinate range where the object to be annotated in the verification image is located, and the accuracy of the three-dimensional annotation frame information is verified.
在一种可能的实施方式中,校验模块206,还配置为:In a possible implementation manner, the verification module 206 is further configured to:
根据映射后的图像坐标范围与校验图像中的待标注对象所在图像坐标范围,对三维标注框信息的准确性进行校验之后,在映射后的图像坐标范围与校验图像中的待标注对象所在图像坐标范围不一致的情况下,响应针对待标注三维点云数据对应的待调整视图中的第二二维标注框的调整操作,对调整后的参数信息进行调整,直至映射后的图像坐标范围与校验图像中的待标注对象所在图像坐标范围一致;所述调整后的参数信息是通过对所述初始参数信息进行调整得到的。After checking the accuracy of the three-dimensional annotation frame information according to the coordinate range of the image after mapping and the coordinate range of the image where the object to be marked in the verification image is located, the coordinate range of the image after mapping and the object to be marked in the verification image are verified. In the case where the image coordinate ranges are inconsistent, in response to the adjustment operation for the second two-dimensional annotation frame in the to-be-adjusted view corresponding to the three-dimensional point cloud data to be marked, adjust the adjusted parameter information until the mapped image coordinate range It is consistent with the image coordinate range where the object to be marked in the verification image is located; the adjusted parameter information is obtained by adjusting the initial parameter information.
在一种可能的实施方式中,标注模块204,还配置为:In a possible implementation manner, the labeling module 204 is further configured to:
确定待标注对象对应的三维标注框信息之后,将待标注对象对应的三维标注框信息,展示在待标注三维点云数据对应的三维视图中,并将待标注对象作为已标注对象;根据已标注对象的三维标注框信息,确定三维视图中的其它待标注对象的三维标注框信息;其它待标注对象与已标注对象属于同一对象类型;将其它待标注对象的三维标注框信息展示在三维视图中。After determining the 3D annotation frame information corresponding to the object to be annotated, display the 3D annotation frame information corresponding to the object to be annotated in the 3D view corresponding to the 3D point cloud data to be annotated, and use the object to be annotated as an annotated object; The 3D annotation frame information of the object determines the 3D annotation frame information of other objects to be annotated in the 3D view; other objects to be annotated and the marked object belong to the same object type; the 3D annotation frame information of other objects to be annotated is displayed in the 3D view .
在一种可能的实施方式中,标注模块204,配置为根据已标注对象的三维标注框信息,确定三维视图中的其它待标注对象的三维标注框信息:In a possible implementation, the labeling module 204 is configured to determine the 3D labeling frame information of other objects to be labelled in the 3D view according to the 3D labeling frame information of the labelled object:
响应于针对三维视图中已标注对象的三维标注框的复制指令,复制已标注对象的三维标注框信息;响应于针对三维视图中其它待标注对象的粘贴指令,根据已标注对象在三维视图中的三维标注框信息,对其它待标注对象进行三维标注,得到其它待标注对象的三维标注框信息。In response to the copy instruction for the 3D annotation frame of the marked object in the 3D view, copy the 3D annotation frame information of the marked object; in response to the paste instruction for other objects to be marked in the 3D view, according to the marked object in the 3D view. 3D labeling frame information, perform 3D labeling on other objects to be labelled, and obtain 3D labeling frame information of other objects to be labelled.
在一种可能的实施方式中,标注模块204,还配置为:In a possible implementation manner, the labeling module 204 is further configured to:
确定待标注对象对应的三维标注框信息之后,将待标注对象对应的三维标注框信息,展示在当前帧待标注三维点云数据对应的第一三维视图中,并将待标注对象作为已标注对象;其中,不同帧待标注三维点云数据对应不同的三维视图;响应于针对第一三维视图中已标注对象的三维标注框的复制指令,复制已标注对象的三维标注框信息;在切换到下一帧待标注点云数据对应的第二三维视图的情况下,响应于针对第二三维视图中其它待标注对象的粘贴指令,根据已标注对象在第一三维视图中的三维标注框信息,对第二三维视图中的其它待标注对象进行三维标注,得到其它待标注对象的三维标注框信息。After determining the 3D annotation frame information corresponding to the object to be annotated, display the 3D annotation frame information corresponding to the object to be annotated in the first 3D view corresponding to the 3D point cloud data to be annotated in the current frame, and use the object to be annotated as the marked object ; wherein, the three-dimensional point cloud data to be marked in different frames corresponds to different three-dimensional views; in response to the copy instruction for the three-dimensional marked frame of the marked object in the first three-dimensional view, copy the three-dimensional marked frame information of the marked object; after switching to the next In the case of a second 3D view corresponding to one frame of point cloud data to be annotated, in response to the pasting instruction for other objects to be annotated in the second 3D view, according to the 3D annotation frame information of the annotated object in the first 3D view, The other objects to be marked in the second three-dimensional view are three-dimensionally marked, and the three-dimensional marking frame information of the other objects to be marked is obtained.
在一种可能的实施方式中,第一确定模块203,配置为按照如下步骤响应针对鸟瞰图中的待标注对象的框选操作,确定第一二维标注框信息:In a possible implementation, the first determining module 203 is configured to respond to a frame selection operation for the object to be marked in the bird's-eye view according to the following steps, and determine the first two-dimensional frame information:
获取参考图像;Get a reference image;
基于待标注对象在参考图像中的图像位置信息、以及参考图像对应的坐标系与鸟瞰图对应的坐标系之间的映射关系,确定待标注对象在鸟瞰图中的参考位置信息;Determine the reference position information of the object to be labeled in the bird's-eye view based on the image position information of the object to be labeled in the reference image and the mapping relationship between the coordinate system corresponding to the reference image and the coordinate system corresponding to the bird's-eye view;
响应针对鸟瞰图中的待标注对象的框选操作,基于参考位置信息确定第一二维标注框信息。In response to the frame selection operation on the object to be marked in the bird's-eye view, the first two-dimensional frame information is determined based on the reference position information.
在一种可能的实施方式中,初始参数信息包括初始高度信息和初始位姿信息中的至少之一。In a possible implementation manner, the initial parameter information includes at least one of initial height information and initial pose information.
关于装置中的各模块的处理流程、以及各模块之间的交互流程的描述可以参照上述方法实施例中的相关说明,这里不再详述。For the description of the processing flow of each module in the apparatus and the interaction flow between the modules, reference may be made to the relevant descriptions in the foregoing method embodiments, which will not be described in detail here.
图4示出了本公开实施例提供的一种电子设备的结构示意图,如图4所示,为本公开实施例提供的电子设备结构示意图,包括:处理器301、存储器302、和总线303。存储器302存储有处理器301可执行的机器可读指令(比如,图3中的点云标注的装置中获取模块201、生成模块202、第一确定模块203和标注模块204对应的执行指令等), 当电子设备运行时,处理器301与存储器302之间通过总线303通信,机器可读指令被处理器301执行时执行如下处理:FIG. 4 shows a schematic structural diagram of an electronic device provided by an embodiment of the present disclosure. As shown in FIG. 4 , the electronic device provided by an embodiment of the present disclosure is a schematic structural diagram, including: a processor 301 , a memory 302 , and a bus 303 . The memory 302 stores machine-readable instructions executable by the processor 301 (for example, the execution instructions corresponding to the acquisition module 201, the generation module 202, the first determination module 203, and the labeling module 204 in the device for point cloud labeling in FIG. 3, etc.) , when the electronic device is running, the processor 301 communicates with the memory 302 through the bus 303, and the machine-readable instructions are executed by the processor 301 to perform the following processing:
获取待标注三维点云数据;Obtain the 3D point cloud data to be labeled;
基于待标注三维点云数据,生成待标注三维点云数据对应的鸟瞰图;Based on the 3D point cloud data to be labeled, generate a bird's-eye view corresponding to the 3D point cloud data to be labeled;
响应针对鸟瞰图中的待标注对象的框选操作,确定第一二维标注框信息;In response to a frame selection operation for the object to be marked in the bird's-eye view, determining the first two-dimensional frame information;
基于被框选的待标注对象的三维点云数据,确定待标注对象的初始参数信息;Determine the initial parameter information of the object to be marked based on the 3D point cloud data of the object to be marked selected by the frame;
基于第一二维标注框信息,以及初始参数信息,确定待标注对象对应的三维标注框信息。Based on the first two-dimensional labeling frame information and the initial parameter information, the three-dimensional labeling frame information corresponding to the object to be labelled is determined.
上述指令的执行过程可以参考本公开实施例中所述的点云标注的方法的步骤。For the execution process of the above instruction, reference may be made to the steps of the method for labeling a point cloud described in the embodiments of the present disclosure.
本公开实施例还提供一种计算机可读存储介质,该计算机可读存储介质上存储有计算机程序,该计算机程序被处理器运行时执行上述方法实施例中所述的点云标注的方法的步骤。其中,该存储介质可以是易失性或非易失的计算机可读取存储介质。Embodiments of the present disclosure further provide a computer-readable storage medium, where a computer program is stored on the computer-readable storage medium, and when the computer program is run by a processor, the steps of the point cloud labeling method described in the foregoing method embodiments are executed. . Wherein, the storage medium may be a volatile or non-volatile computer-readable storage medium.
本公开实施例还提供的一种计算机程序产品,包括存储程序代码的计算机可读存储介质,所述程序代码包括的指令可配置为执行上述方法实施例中所述的点云标注的方法的步骤,具体可参见上述方法实施例。An embodiment of the present disclosure further provides a computer program product, including a computer-readable storage medium storing program codes, and the instructions included in the program codes can be configured to execute the steps of the point cloud labeling method described in the above method embodiments , for details, refer to the above method embodiments.
本公开实施例还提供一种计算机程序,该计算机程序被处理器执行时实现前述实施例的任意一种方法。该计算机程序产品可以具体通过硬件、软件或其结合的方式实现。在一个可选实施例中,所述计算机程序产品具体体现为计算机存储介质,在另一个可选实施例中,计算机程序产品具体体现为软件产品,例如软件开发包(Software Development Kit,SDK)等等。Embodiments of the present disclosure further provide a computer program, which implements any one of the methods in the foregoing embodiments when the computer program is executed by a processor. The computer program product can be specifically implemented by hardware, software or a combination thereof. In an optional embodiment, the computer program product is embodied as a computer storage medium, and in another optional embodiment, the computer program product is embodied as a software product, such as a software development kit (Software Development Kit, SDK), etc. Wait.
所属领域的技术人员可以清楚地了解到,为描述的方便和简洁,上述描述的系统和装置的具体工作过程,可以参考前述方法实施例中的对应过程。在本公开所提供的几个实施例中,应该理解到,所揭露的系统、装置和方法,可以通过其它的方式实现。以上所描述的装置实施例仅仅是示意性的,例如,所述单元的划分,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分方式,又例如,多个单元或组件可以结合或者可以集成到另一个系统,或一些特征可以忽略,或不执行。另一点,所显示或讨论的相互之间的耦合或直接耦合或通信连接可以是通过一些通信接口,装置或单元的间接耦合或通信连接,可以是电性,机械或其它的形式。Those skilled in the art can clearly understand that, for the convenience and brevity of description, for the specific working process of the system and device described above, reference may be made to the corresponding process in the foregoing method embodiments. In the several embodiments provided by the present disclosure, it should be understood that the disclosed system, apparatus and method may be implemented in other manners. The apparatus embodiments described above are only illustrative. For example, the division of the units is only a logical function division. In actual implementation, there may be other division methods. For example, multiple units or components may be combined or Can be integrated into another system, or some features can be ignored, or not implemented. On the other hand, the shown or discussed mutual coupling or direct coupling or communication connection may be through some communication interfaces, indirect coupling or communication connection of devices or units, which may be in electrical, mechanical or other forms.
所述作为分离部件说明的单元可以是或者也可以不是物理上分开的,作为单元显示的部件可以是或者也可以不是物理单元,即可以位于一个地方,或者也可以分布到多个网络单元上。可以根据实际的需要选择其中的部分或者全部单元来实现本实施例方案的目的。The units described as separate components may or may not be physically separated, and components displayed as units may or may not be physical units, that is, may be located in one place, or may be distributed to multiple network units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution in this embodiment.
另外,在本公开各个实施例中的各功能单元可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中。In addition, each functional unit in each embodiment of the present disclosure may be integrated into one processing unit, or each unit may exist physically alone, or two or more units may be integrated into one unit.
所述功能如果以软件功能单元的形式实现并作为独立的产品销售或使用时,可以存储在一个处理器可执行的非易失的计算机可读取存储介质中。基于这样的理解,本公开的技术方案本质上或者说对现有技术做出贡献的部分或者该技术方案的部分可以以软件产品的形式体现出来,该计算机软件产品存储在一个存储介质中,包括若干指令用以使得一台电子设备(可以是个人计算机,服务器,或者网络设备等)执行本公开各个实施例所述方法的全部或部分步骤。而前述的存储介质包括:U盘、移动硬盘、只读存储器(Read-Only Memory,ROM)、随机存取存储器(Random Access Memory,RAM)、磁碟或者光盘等各种可以存储程序代码的介质。The functions, if implemented in the form of software functional units and sold or used as stand-alone products, may be stored in a processor-executable non-volatile computer-readable storage medium. Based on this understanding, the technical solutions of the present disclosure can be embodied in the form of software products in essence, or the parts that make contributions to the prior art or the parts of the technical solutions. The computer software products are stored in a storage medium, including Several instructions are used to cause an electronic device (which may be a personal computer, a server, or a network device, etc.) to execute all or part of the steps of the methods described in various embodiments of the present disclosure. The aforementioned storage medium includes: U disk, mobile hard disk, read-only memory (Read-Only Memory, ROM), random access memory (Random Access Memory, RAM), magnetic disk or optical disk and other media that can store program codes .
最后应说明的是:以上所述实施例,仅为本公开的具体实施方式,用以说明本公开的技术方案,而非对其限制,本公开的保护范围并不局限于此,尽管参照前述实施例对 本公开进行了详细的说明,本领域的普通技术人员应当理解:任何熟悉本技术领域的技术人员在本公开揭露的技术范围内,其依然可以对前述实施例所记载的技术方案进行修改或可轻易想到变化,或者对其中部分技术特征进行等同替换;而这些修改、变化或者替换,并不使相应技术方案的本质脱离本公开实施例技术方案的精神和范围,都应涵盖在本公开的保护范围之内。因此,本公开的保护范围应所述以权利要求的保护范围为准。Finally, it should be noted that the above-mentioned embodiments are only specific implementations of the present disclosure, and are used to illustrate the technical solutions of the present disclosure rather than limit them. The protection scope of the present disclosure is not limited thereto, although referring to the foregoing The embodiments describe the present disclosure in detail, and those skilled in the art should understand that: any person skilled in the art can still modify the technical solutions described in the foregoing embodiments within the technical scope disclosed by the present disclosure. Or can easily think of changes, or equivalently replace some of the technical features; and these modifications, changes or replacements do not make the essence of the corresponding technical solutions deviate from the spirit and scope of the technical solutions of the embodiments of the present disclosure, and should be covered in the present disclosure. within the scope of protection. Therefore, the protection scope of the present disclosure should be based on the protection scope of the claims.
工业实用性Industrial Applicability
本实施例中,通过响应针对所述待标注三维点云数据对应的鸟瞰图中的待标注对象的框选操作,确定第一二维标注框信息;基于被框选的待标注对象的三维点云数据,确定所述待标注对象的初始参数信息;基于所述第一二维标注框信息,以及所述初始参数信息,确定所述待标注对象对应的三维标注框信息,从而能够通过框选得到的第一二维标注框信息和初始参数信息,实现针对待标注对象进行三维标注框的自动标注,标注效率和准确率均较高。In this embodiment, the first two-dimensional labeling frame information is determined by responding to a frame selection operation for the object to be labelled in the bird's-eye view corresponding to the 3D point cloud data to be labelled; Based on the cloud data, the initial parameter information of the object to be marked is determined; based on the first two-dimensional marked frame information and the initial parameter information, the three-dimensional marked frame information corresponding to the to-be-marked object is determined, so that the frame selection can be performed. The obtained first two-dimensional labeling frame information and initial parameter information realize automatic labeling of the three-dimensional labeling frame for the object to be labelled, and the labeling efficiency and accuracy are high.

Claims (31)

  1. 一种点云标注的方法,所述方法包括:A method for point cloud labeling, the method comprising:
    获取待标注三维点云数据;Obtain the 3D point cloud data to be labeled;
    基于所述待标注三维点云数据,生成所述待标注三维点云数据对应的鸟瞰图;generating a bird's-eye view corresponding to the three-dimensional point cloud data to be marked based on the three-dimensional point cloud data to be marked;
    响应针对所述鸟瞰图中的待标注对象的框选操作,确定第一二维标注框信息;In response to the frame selection operation for the object to be marked in the bird's-eye view, determine the first two-dimensional frame information;
    基于被框选的待标注对象的三维点云数据,确定所述待标注对象的初始参数信息;Determine the initial parameter information of the object to be marked based on the three-dimensional point cloud data of the object to be marked selected by the frame;
    基于所述第一二维标注框信息,以及所述初始参数信息,确定所述待标注对象对应的三维标注框信息。Based on the first two-dimensional annotation frame information and the initial parameter information, the three-dimensional annotation frame information corresponding to the object to be annotated is determined.
  2. 根据权利要求1所述的方法,其中,所述方法还包括:The method of claim 1, wherein the method further comprises:
    基于所述待标注对象的三维点云数据,生成所述待标注三维点云数据对应的待调整视图;所述待调整视图至少包括正视图和侧视图中的一种;Based on the three-dimensional point cloud data of the object to be marked, a view to be adjusted corresponding to the three-dimensional point cloud data to be marked is generated; the to-be-adjusted view includes at least one of a front view and a side view;
    根据生成的所述待调整视图,确定所述待标注对象在所述待调整视图中的第二二维标注框信息。According to the generated view to be adjusted, second two-dimensional labeling frame information of the to-be-labeled object in the to-be-adjusted view is determined.
  3. 根据权利要求2所述的方法,其中,所述基于所述第一二维标注框信息,以及所述初始参数信息,确定所述待标注对象对应的三维标注框信息,包括:The method according to claim 2, wherein the determining, based on the first 2D annotation frame information and the initial parameter information, the 3D annotation frame information corresponding to the object to be annotated comprises:
    基于所述第一二维标注框信息、所述第二二维标注框信息以及所述初始参数信息,确定所述待标注对象对应的三维标注框信息。Based on the first 2D annotation frame information, the second 2D annotation frame information and the initial parameter information, the 3D annotation frame information corresponding to the object to be annotated is determined.
  4. 根据权利要求3所述的方法,其中,所述基于所述第一二维标注框信息、所述第二二维标注框信息以及所述初始参数信息,确定所述待标注对象对应的三维标注框信息,包括:The method according to claim 3, wherein the 3D annotation corresponding to the object to be annotated is determined based on the first 2D annotation frame information, the second 2D annotation frame information and the initial parameter information Box information, including:
    响应针对所述第二二维标注框信息的调整操作,基于所述初始参数信息,确定所述待标注对象的调整后的参数信息;In response to the adjustment operation for the second two-dimensional annotation frame information, based on the initial parameter information, determine the adjusted parameter information of the object to be annotated;
    基于所述第一二维标注框信息、所述第二二维标注框信息以及所述调整后的参数信息,确定所述待标注对象对应的三维标注框信息。Based on the first 2D annotation frame information, the second 2D annotation frame information and the adjusted parameter information, the 3D annotation frame information corresponding to the object to be annotated is determined.
  5. 根据权利要求4所述的方法,其中,所述响应针对所述第二二维标注框信息的调整操作,基于所述初始参数信息,确定所述待标注对象的调整后的参数信息,包括:The method according to claim 4, wherein, in response to an adjustment operation for the second two-dimensional annotation frame information, determining the adjusted parameter information of the object to be annotated based on the initial parameter information, comprising:
    响应针对所述待调整视图中的第二二维标注框的调整操作,确定与所述调整操作对应的参数调整信息;determining parameter adjustment information corresponding to the adjustment operation in response to an adjustment operation for the second two-dimensional frame in the to-be-adjusted view;
    基于所述参数调整信息对所述初始参数信息进行调整,得到所述调整后的参数信息。The initial parameter information is adjusted based on the parameter adjustment information to obtain the adjusted parameter information.
  6. 根据权利要求4或5所述的方法,其中,所述基于所述第一二维标注框信息、所述第二二维标注框信息以及所述调整后的参数信息,确定所述待标注对象对应的三维标注框信息,包括:The method according to claim 4 or 5, wherein the object to be annotated is determined based on the first two-dimensional annotation frame information, the second two-dimensional annotation frame information and the adjusted parameter information Corresponding 3D callout box information, including:
    基于所述调整后的参数信息所对应的三维点云数据,与所述第一二维标注框信息所对应的二维点云数据之间的第一对应关系,确定所述待标注对象对应的三维标注框的底面的参数信息;Based on the first correspondence between the three-dimensional point cloud data corresponding to the adjusted parameter information and the two-dimensional point cloud data corresponding to the first two-dimensional labeling frame information, determine the corresponding to-be-labeled object. The parameter information of the bottom surface of the 3D callout box;
    基于所述调整后的参数信息所对应的三维点云数据,与所述第二二维标注框信息所对应的二维点云数据之间的第二对应关系,确定所述待标注对象对应的三维标注框的侧面和正面的参数信息中的至少之一;Based on the second correspondence between the three-dimensional point cloud data corresponding to the adjusted parameter information and the two-dimensional point cloud data corresponding to the second two-dimensional labeling frame information, determine the corresponding to-be-labeled object. At least one of the parameter information of the side and the front of the 3D callout frame;
    将确定的所述三维标注框的底面的参数信息、以及所述三维标注框的侧面和正 面的参数信息中的至少之一进行组合,得到所述三维标注框信息。Combining at least one of the determined parameter information of the bottom surface of the 3D annotation frame and the parameter information of the side surface and the front surface of the 3D annotation frame to obtain the 3D annotation frame information.
  7. 根据权利要求2至6任一所述的方法,其中,在所述确定所述待标注对象在所述待调整视图中的第二二维标注框信息之后,所述方法还包括:The method according to any one of claims 2 to 6, wherein after the determining the second two-dimensional annotation frame information of the object to be annotated in the view to be adjusted, the method further comprises:
    基于确定的所述待标注对象的第二二维标注框信息,展示标注有与所述第二二维标注框信息对应的第二二维标注框的所述待调整视图。Based on the determined second two-dimensional annotation frame information of the object to be annotated, the to-be-adjusted view marked with a second two-dimensional annotation frame corresponding to the second two-dimensional annotation frame information is displayed.
  8. 根据权利要求1至7任一所述的方法,其中,在所述确定所述待标注对象对应的三维标注框信息之后,所述方法还包括:The method according to any one of claims 1 to 7, wherein after the determining the three-dimensional annotation frame information corresponding to the object to be annotated, the method further comprises:
    获取校验图像;所述校验图像与所述待标注三维点云数据来源于同一场景;Obtaining a verification image; the verification image and the to-be-labeled 3D point cloud data originate from the same scene;
    基于所述待标注三维点云数据对应的坐标系,与所述校验图像对应的坐标系之间的映射关系,将所述待标注对象对应的三维标注框信息映射至所述校验图像中,确定映射后的图像坐标范围;Based on the mapping relationship between the coordinate system corresponding to the three-dimensional point cloud data to be marked and the coordinate system corresponding to the verification image, map the three-dimensional annotation frame information corresponding to the object to be marked into the verification image , determine the coordinate range of the image after mapping;
    根据所述映射后的图像坐标范围与所述校验图像中的所述待标注对象所在图像坐标范围,对所述三维标注框信息的准确性进行校验。The accuracy of the three-dimensional annotation frame information is verified according to the mapped image coordinate range and the image coordinate range in the verification image where the object to be marked is located.
  9. 根据权利要求8所述的方法,其中,在所述根据所述映射后的图像坐标范围与所述校验图像中的所述待标注对象所在图像坐标范围,对所述三维标注框信息的准确性进行校验之后,还包括:The method according to claim 8, wherein, according to the image coordinate range after the mapping and the image coordinate range where the object to be annotated in the verification image is located, the accuracy of the three-dimensional annotation frame information is accurate. After verification, it also includes:
    在所述映射后的图像坐标范围与所述校验图像中的所述待标注对象所在图像坐标范围不一致的情况下,响应针对所述待标注三维点云数据对应的待调整视图中的第二二维标注框的调整操作,对调整后的参数信息进行调整,直至所述映射后的图像坐标范围与所述校验图像中的所述待标注对象所在图像坐标范围一致;所述调整后的参数信息是通过对所述初始参数信息进行调整得到的。In the case where the mapped image coordinate range is inconsistent with the image coordinate range where the object to be marked in the verification image is located, respond to the second image coordinate range in the to-be-adjusted view corresponding to the three-dimensional point cloud data to be marked. The adjustment operation of the two-dimensional annotation frame is to adjust the adjusted parameter information until the image coordinate range after the mapping is consistent with the image coordinate range where the object to be annotated in the verification image is located; the adjusted image coordinate range is The parameter information is obtained by adjusting the initial parameter information.
  10. 根据权利要求1至9任一所述的方法,其中,所述确定所述待标注对象对应的三维标注框信息之后,还包括:The method according to any one of claims 1 to 9, wherein after the determining the three-dimensional labeling frame information corresponding to the object to be labelled, the method further comprises:
    将所述待标注对象对应的三维标注框信息,展示在所述待标注三维点云数据对应的三维视图中,并将所述待标注对象作为已标注对象;Displaying the 3D labeling frame information corresponding to the object to be labelled in the 3D view corresponding to the 3D point cloud data to be labelled, and using the object to be labelled as the labelled object;
    根据所述已标注对象的三维标注框信息,确定所述三维视图中的其它待标注对象的三维标注框信息;所述其它待标注对象与所述已标注对象属于同一对象类型;According to the 3D annotation frame information of the marked object, determine the 3D annotation frame information of other objects to be marked in the 3D view; the other objects to be marked and the marked object belong to the same object type;
    将所述其它待标注对象的三维标注框信息展示在所述三维视图中。The 3D annotation frame information of the other objects to be annotated is displayed in the 3D view.
  11. 根据权利要求10所述的方法,其中,所述根据所述已标注对象的三维标注框信息,确定所述三维视图中的其它待标注对象的三维标注框信息,包括:The method according to claim 10, wherein the determining, according to the 3D annotation frame information of the marked object, the 3D annotation frame information of other objects to be annotated in the 3D view comprises:
    响应于针对所述三维视图中已标注对象的三维标注框的复制指令,复制所述已标注对象的三维标注框信息;In response to a copy instruction for the three-dimensional call-out frame of the marked object in the three-dimensional view, copying the three-dimensional call-out frame information of the marked object;
    响应于针对所述三维视图中其它待标注对象的粘贴指令,根据所述已标注对象在所述三维视图中的三维标注框信息,对所述其它待标注对象进行三维标注,得到所述其它待标注对象的三维标注框信息。In response to the paste instruction for other objects to be marked in the three-dimensional view, perform three-dimensional marking on the other objects to be marked according to the three-dimensional marking frame information of the marked objects in the three-dimensional view, and obtain the other objects to be marked. The 3D callout box information of the callout object.
  12. 根据权利要求1至11任一所述的方法,其中,所述确定所述待标注对象对应的三维标注框信息之后,还包括:The method according to any one of claims 1 to 11, wherein after the determining the three-dimensional labeling frame information corresponding to the object to be labelled, the method further comprises:
    将所述待标注对象对应的三维标注框信息,展示在当前帧待标注三维点云数据对应的第一三维视图中,并将所述待标注对象作为已标注对象;其中,不同帧待标注三维点云数据对应不同的三维视图;Display the 3D labeling frame information corresponding to the object to be labelled in the first 3D view corresponding to the 3D point cloud data to be labelled in the current frame, and use the object to be labelled as the labelled object; wherein, different frames to be labelled 3D view Point cloud data corresponds to different 3D views;
    响应于针对所述第一三维视图中所述已标注对象的三维标注框的复制指令,复制所述已标注对象的三维标注框信息;In response to a copy instruction for the 3D annotation frame of the annotated object in the first 3D view, copying the 3D annotation frame information of the annotated object;
    在切换到下一帧待标注三维点云数据对应的第二三维视图的情况下,响应于针对所述第二三维视图中其它待标注对象的粘贴指令,根据所述已标注对象在所述第一 三维视图中的三维标注框信息,对所述第二三维视图中的其它待标注对象进行三维标注,得到其它待标注对象的三维标注框信息。In the case of switching to the second 3D view corresponding to the 3D point cloud data to be labeled in the next frame, in response to the paste instruction for other objects to be labeled in the second 3D view, according to the labeled objects in the second 3D view 3D labeling frame information in a 3D view, perform 3D labeling on other objects to be labelled in the second 3D view, and obtain 3D labeling frame information of other objects to be labelled.
  13. 根据权利要求1至12任一所述的方法,其中,所述响应针对所述鸟瞰图中的待标注对象的框选操作,确定第一二维标注框信息,包括:The method according to any one of claims 1 to 12, wherein, in response to a frame selection operation on an object to be marked in the bird's-eye view, determining the first two-dimensional frame information includes:
    获取参考图像;所述参考图像与所述待标注三维点云数据来源于同一场景;obtaining a reference image; the reference image and the to-be-labeled 3D point cloud data originate from the same scene;
    基于所述待标注对象在所述参考图像中的图像位置信息、以及所述参考图像对应的坐标系与所述鸟瞰图对应的坐标系之间的映射关系,确定所述待标注对象在所述鸟瞰图中的参考位置信息;Based on the image position information of the object to be labeled in the reference image and the mapping relationship between the coordinate system corresponding to the reference image and the coordinate system corresponding to the bird's-eye view, it is determined that the object to be labeled is located in the reference image. Reference location information in the bird's eye view;
    响应针对所述鸟瞰图中的所述待标注对象的框选操作,基于所述参考位置信息确定所述第一二维标注框信息。In response to a frame selection operation on the object to be marked in the bird's-eye view, the first two-dimensional frame information is determined based on the reference position information.
  14. 根据权利要求1至13任一所述的方法,其中,所述初始参数信息包括初始高度信息和初始位姿信息中的至少之一。The method according to any one of claims 1 to 13, wherein the initial parameter information includes at least one of initial height information and initial pose information.
  15. 一种点云标注的装置,所述装置包括:A device for point cloud labeling, the device comprising:
    获取模块,配置为获取待标注三维点云数据;an acquisition module, configured to acquire the 3D point cloud data to be marked;
    生成模块,配置为基于所述待标注三维点云数据,生成所述待标注三维点云数据对应的鸟瞰图;a generating module, configured to generate a bird's-eye view corresponding to the three-dimensional point cloud data to be marked based on the three-dimensional point cloud data to be marked;
    第一确定模块,配置为响应针对所述鸟瞰图中的待标注对象的框选操作,确定第一二维标注框信息;基于被框选的待标注对象的三维点云数据,确定所述待标注对象的初始参数信息;A first determination module, configured to determine first two-dimensional labeling frame information in response to a frame selection operation for objects to be labelled in the bird's-eye view; The initial parameter information of the annotation object;
    标注模块,配置为基于所述第一二维标注框信息,以及所述初始参数信息,确定所述待标注对象对应的三维标注框信息。The labeling module is configured to determine the three-dimensional labeling frame information corresponding to the object to be labelled based on the first two-dimensional labeling frame information and the initial parameter information.
  16. 根据权利要求15所述的装置,其中,所述装置还包括:The apparatus of claim 15, wherein the apparatus further comprises:
    第二确定模块,配置为:基于所述待标注对象的三维点云数据,生成所述待标注三维点云数据对应的待调整视图;所述待调整视图至少包括正视图和侧视图中的一种;根据生成的所述待调整视图,确定所述待标注对象在所述待调整视图中的第二二维标注框信息。The second determination module is configured to: generate a view to be adjusted corresponding to the three-dimensional point cloud data to be labeled based on the three-dimensional point cloud data of the object to be labeled; the view to be adjusted includes at least one of a front view and a side view. Type; according to the generated view to be adjusted, determine the second two-dimensional annotation frame information of the to-be-annotated object in the to-be-adjusted view.
  17. 根据权利要求16所述的装置,其中,所述标注模块,还配置为:基于所述第一二维标注框信息、所述第二二维标注框信息以及所述初始参数信息,确定所述待标注对象对应的三维标注框信息。The apparatus according to claim 16, wherein the labeling module is further configured to: determine the Information about the 3D annotation frame corresponding to the object to be annotated.
  18. 根据权利要求17所述的装置,其中,所述标注模块,还配置为:响应针对所述第二二维标注框信息的调整操作,基于所述初始参数信息,确定所述待标注对象的调整后的参数信息;基于所述第一二维标注框信息、所述第二二维标注框信息以及所述调整后的参数信息,确定所述待标注对象对应的三维标注框信息。The apparatus according to claim 17, wherein the labeling module is further configured to: in response to an adjustment operation for the second two-dimensional labeling frame information, determine the adjustment of the object to be labelled based on the initial parameter information and based on the first 2D annotation frame information, the second 2D annotation frame information and the adjusted parameter information, determine the 3D annotation frame information corresponding to the object to be annotated.
  19. 根据权利要求18所述的装置,其中,所述标注模块,还配置为:响应针对所述待调整视图中的第二二维标注框的调整操作,确定与所述调整操作对应的参数调整信息;基于所述参数调整信息对所述初始参数信息进行调整,得到所述调整后的参数信息。The apparatus according to claim 18, wherein the annotation module is further configured to: in response to an adjustment operation for the second two-dimensional annotation frame in the to-be-adjusted view, determine parameter adjustment information corresponding to the adjustment operation ; Adjust the initial parameter information based on the parameter adjustment information to obtain the adjusted parameter information.
  20. 根据权利要求18或19所述的装置,其中,所述标注模块,还配置为:基于所述调整后的参数信息所对应的三维点云数据,与所述第一二维标注框信息所对应的二维点云数据之间的第一对应关系,确定所述待标注对象对应的三维标注框的底面的参数信息;基于所述调整后的参数信息所对应的三维点云数据,与所述第二二维标注框信息所对应的二维点云数据之间的第二对应关系,确定所述待标注对象对应的三维标注框的侧面和正面的参数信息中的至少之一;将确定的所述三维标注框的底面的参数信息、以及所述三维标注框的侧面和正面的参数信息中的至少之一进行组合,得 到所述三维标注框信息。The device according to claim 18 or 19, wherein the labeling module is further configured to: based on the three-dimensional point cloud data corresponding to the adjusted parameter information, corresponding to the first two-dimensional labeling frame information The first corresponding relationship between the two-dimensional point cloud data, determine the parameter information of the bottom surface of the three-dimensional mark frame corresponding to the object to be marked; based on the three-dimensional point cloud data corresponding to the adjusted parameter information, and the The second correspondence between the two-dimensional point cloud data corresponding to the second two-dimensional labeling frame information is to determine at least one of the side and front parameter information of the three-dimensional labeling frame corresponding to the object to be labelled; At least one of the parameter information of the bottom surface of the 3D annotation frame and the parameter information of the side surface and the front surface of the 3D annotation frame are combined to obtain the 3D annotation frame information.
  21. 根据权利要求16至20任一所述的装置,其中,所述第二确定模块,还配置为:基于确定的所述待标注对象的第二二维标注框信息,展示标注有与所述第二二维标注框信息对应的第二二维标注框的所述待调整视图。The device according to any one of claims 16 to 20, wherein the second determining module is further configured to: display a label with the The to-be-adjusted view of the second two-dimensional callout frame corresponding to the two-dimensional callout frame information.
  22. 根据权利要求15至22任一所述的装置,其中,所述装置还包括:The apparatus of any one of claims 15 to 22, wherein the apparatus further comprises:
    校验模块,配置为:获取校验图像;所述校验图像与所述待标注三维点云数据来源于同一场景;基于所述待标注三维点云数据对应的坐标系,与所述校验图像对应的坐标系之间的映射关系,将所述待标注对象对应的三维标注框信息映射至所述校验图像中,确定映射后的图像坐标范围;根据所述映射后的图像坐标范围与所述校验图像中的所述待标注对象所在图像坐标范围,对所述三维标注框信息的准确性进行校验。a verification module, configured to: obtain a verification image; the verification image and the three-dimensional point cloud data to be marked originate from the same scene; based on the coordinate system corresponding to the three-dimensional point cloud data to be marked, the verification image and the verification The mapping relationship between the coordinate systems corresponding to the images, the three-dimensional annotation frame information corresponding to the object to be marked is mapped to the verification image, and the mapped image coordinate range is determined; according to the mapped image coordinate range and In the verification image, the image coordinate range where the object to be marked is located, and the accuracy of the three-dimensional marked frame information is verified.
  23. 根据权利要求22所述的装置,其中,所述校验模块,还配置为:在所述映射后的图像坐标范围与所述校验图像中的所述待标注对象所在图像坐标范围不一致的情况下,响应针对所述待标注三维点云数据对应的待调整视图中的第二二维标注框的调整操作,对调整后的参数信息进行调整,直至所述映射后的图像坐标范围与所述校验图像中的所述待标注对象所在图像坐标范围一致;所述调整后的参数信息是通过对所述初始参数信息进行调整得到的。The apparatus according to claim 22, wherein the verification module is further configured to: in the case where the coordinate range of the image after the mapping is inconsistent with the coordinate range of the image where the object to be marked in the verification image is located Next, in response to the adjustment operation for the second two-dimensional annotation frame in the view to be adjusted corresponding to the three-dimensional point cloud data to be marked, the adjusted parameter information is adjusted until the mapped image coordinate range is the same as the described image coordinate range. Verify that the image coordinate ranges of the objects to be marked in the verification image are consistent; the adjusted parameter information is obtained by adjusting the initial parameter information.
  24. 根据权利要求15至23任一所述的装置,其中,所述标注模块,还配置为:将所述待标注对象对应的三维标注框信息,展示在所述待标注三维点云数据对应的三维视图中,并将所述待标注对象作为已标注对象;根据所述已标注对象的三维标注框信息,确定所述三维视图中的其它待标注对象的三维标注框信息;所述其它待标注对象与所述已标注对象属于同一对象类型;将所述其它待标注对象的三维标注框信息展示在所述三维视图中。The device according to any one of claims 15 to 23, wherein the labeling module is further configured to: display the 3D labeling frame information corresponding to the to-be-labeled object in the 3D corresponding to the to-be-labeled 3D point cloud data In the view, the object to be marked is regarded as the marked object; according to the three-dimensional markup frame information of the marked object, the three-dimensional markup frame information of other objects to be marked in the three-dimensional view is determined; the other objects to be marked belonging to the same object type as the marked object; displaying the three-dimensional marked frame information of the other objects to be marked in the three-dimensional view.
  25. 根据权利要求24所述的装置,其中,所述标注模块,还配置为:响应于针对所述三维视图中已标注对象的三维标注框的复制指令,复制所述已标注对象的三维标注框信息;响应于针对所述三维视图中其它待标注对象的粘贴指令,根据所述已标注对象在所述三维视图中的三维标注框信息,对所述其它待标注对象进行三维标注,得到所述其它待标注对象的三维标注框信息。The apparatus according to claim 24, wherein the annotation module is further configured to: in response to a copy instruction for the 3D annotation frame of the annotated object in the 3D view, copy the 3D annotation frame information of the annotated object ; In response to the paste instruction for other objects to be marked in the three-dimensional view, according to the three-dimensional markup frame information of the marked objects in the three-dimensional view, perform three-dimensional marking on the other objects to be marked, and obtain the other objects to be marked. Information about the 3D annotation frame of the object to be annotated.
  26. 根据权利要求15至25任一所述的装置,其中,所述标注模块,还配置为:将所述待标注对象对应的三维标注框信,息展示在当前帧待标注三维点云数据对应的第一三维视图中,并将所述待标注对象作为已标注对象;其中,不同帧待标注三维点云数据对应不同的三维视图;响应于针对所述第一三维视图中所述已标注对象的三维标注框的复制指令,复制所述已标注对象的三维标注框信息;在切换到下一帧待标注三维点云数据对应的第二三维视图的情况下,响应于针对所述第二三维视图中其它待标注对象的粘贴指令,根据所述已标注对象在所述第一三维视图中的三维标注框信息,对所述第二三维视图中的其它待标注对象进行三维标注,得到其它待标注对象的三维标注框信息。The device according to any one of claims 15 to 25, wherein the labeling module is further configured to: display the information of the 3D labeling frame corresponding to the object to be labelled in the current frame corresponding to the 3D point cloud data to be labelled In the first three-dimensional view, the object to be marked is regarded as the marked object; wherein, different frames of three-dimensional point cloud data to be marked correspond to different three-dimensional views; in response to the marked object in the first three-dimensional view The instruction for copying the 3D annotation frame, copying the 3D annotation frame information of the marked object; in the case of switching to the second 3D view corresponding to the 3D point cloud data to be marked in the next frame, in response to the second 3D view The pasting instruction of other objects to be annotated in the second 3D view, according to the 3D annotation frame information of the marked object in the first 3D view, perform 3D annotation on the other objects to be annotated in the second 3D view, and obtain other objects to be annotated. The 3D callout box information of the object.
  27. 根据权利要求15至25任一所述的装置,其中,所述第一确定模块,还配置为获取参考图像;所述参考图像与所述待标注三维点云数据来源于同一场景;基于所述待标注对象在所述参考图像中的图像位置信息、以及所述参考图像对应的坐标系与所述鸟瞰图对应的坐标系之间的映射关系,确定所述待标注对象在所述鸟瞰图中的参考位置信息;响应针对所述鸟瞰图中的所述待标注对象的框选操作,基于所述参考位置信息确定所述第一二维标注框信息。The apparatus according to any one of claims 15 to 25, wherein the first determining module is further configured to acquire a reference image; the reference image and the to-be-labeled 3D point cloud data originate from the same scene; based on the The image position information of the object to be labeled in the reference image, and the mapping relationship between the coordinate system corresponding to the reference image and the coordinate system corresponding to the bird's-eye view, determine that the object to be labeled is in the bird's-eye view The reference position information; in response to the frame selection operation on the object to be marked in the bird's-eye view, the first two-dimensional frame information is determined based on the reference position information.
  28. 根据权利要求22至26任一所述的装置,其中,所述初始参数信息包括初始高度信息和初始位姿信息中的至少之一。The apparatus according to any one of claims 22 to 26, wherein the initial parameter information includes at least one of initial height information and initial pose information.
  29. 一种电子设备,包括:处理器、存储器和总线,所述存储器存储有所述处理器可执行的机器可读指令,所述处理器配置为执行所述存储器中存储的机器可读指令,在所述电子设备运行的情况下,所述处理器与所述存储器之间通过总线通信,所述机器可读指令被所述处理器执行时执行如权利要求1至14任一所述的点云标注的方法的步骤。An electronic device comprising: a processor, a memory and a bus, the memory stores machine-readable instructions executable by the processor, the processor is configured to execute the machine-readable instructions stored in the memory, When the electronic device is running, the processor and the memory communicate through a bus, and the machine-readable instructions are executed by the processor to execute the point cloud according to any one of claims 1 to 14 Annotated method steps.
  30. 一种计算机可读存储介质,所述计算机可读存储介质上存储有计算机程序,所述计算机程序被电子设备运行时,所述电子设备执行如权利要求1至14任一所述的点云标注的方法的步骤。A computer-readable storage medium on which a computer program is stored, and when the computer program is run by an electronic device, the electronic device executes the point cloud labeling according to any one of claims 1 to 14 steps of the method.
  31. 一种计算机程序产品,包括存储程序代码的计算机可读存储介质,所述程序代码包括的指令可配置为执行如权利要求1至14任一所述的点云标注的方法的步骤。A computer program product comprising a computer-readable storage medium storing program code, the program code comprising instructions configurable to perform the steps of the method for point cloud labeling as claimed in any one of claims 1 to 14.
PCT/CN2021/095444 2020-09-30 2021-05-24 Point cloud annotating method and apparatus, electronic device, storage medium, and program product WO2022068225A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
KR1020227018072A KR20220092928A (en) 2020-09-30 2021-05-24 Point cloud labeling methods, devices, electronic devices, storage media and program products

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202011061734.2 2020-09-30
CN202011061734.2A CN112132901A (en) 2020-09-30 2020-09-30 Point cloud labeling method and device, electronic equipment and storage medium

Publications (1)

Publication Number Publication Date
WO2022068225A1 true WO2022068225A1 (en) 2022-04-07

Family

ID=73843536

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/095444 WO2022068225A1 (en) 2020-09-30 2021-05-24 Point cloud annotating method and apparatus, electronic device, storage medium, and program product

Country Status (3)

Country Link
KR (1) KR20220092928A (en)
CN (1) CN112132901A (en)
WO (1) WO2022068225A1 (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115423934A (en) * 2022-08-12 2022-12-02 北京城市网邻信息技术有限公司 House type graph generation method and device, electronic equipment and storage medium
CN115542338A (en) * 2022-11-30 2022-12-30 湖南大学 Laser radar data learning method based on point cloud space distribution mapping
CN115880536A (en) * 2023-02-15 2023-03-31 北京百度网讯科技有限公司 Data processing method, training method, target object detection method and device
CN116152783A (en) * 2023-04-18 2023-05-23 安徽蔚来智驾科技有限公司 Target element labeling data acquisition method, computer equipment and storage medium

Families Citing this family (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111259980B (en) * 2020-02-10 2023-10-03 北京小马慧行科技有限公司 Method and device for processing annotation data
CN112132901A (en) * 2020-09-30 2020-12-25 上海商汤临港智能科技有限公司 Point cloud labeling method and device, electronic equipment and storage medium
CN113689508B (en) * 2021-09-09 2024-02-02 北京地平线机器人技术研发有限公司 Point cloud labeling method and device, storage medium and electronic equipment
CN113763572B (en) * 2021-09-17 2023-06-27 北京京航计算通讯研究所 3D entity labeling method based on AI intelligent recognition and storage medium
CN113763573B (en) * 2021-09-17 2023-07-11 北京京航计算通讯研究所 Digital labeling method and device for three-dimensional object
CN115423933B (en) * 2022-08-12 2023-09-29 北京城市网邻信息技术有限公司 House type diagram generation method and device, electronic equipment and storage medium
CN115620199B (en) * 2022-10-24 2023-06-13 四川警察学院 Traffic safety risk diagnosis method and device
CN116309962B (en) * 2023-05-10 2023-09-26 倍基智能科技(四川)有限公司 Laser radar point cloud data labeling method, system and application
CN117557999A (en) * 2023-11-20 2024-02-13 镁佳(北京)科技有限公司 Image joint labeling method, computer equipment and medium

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109727312A (en) * 2018-12-10 2019-05-07 广州景骐科技有限公司 Point cloud mask method, device, computer equipment and storage medium
CN110136273A (en) * 2019-03-29 2019-08-16 初速度(苏州)科技有限公司 A kind of sample data mask method and device in machine learning
CN110135453A (en) * 2019-03-29 2019-08-16 初速度(苏州)科技有限公司 A kind of laser point cloud data mask method and device
CN111310667A (en) * 2020-02-18 2020-06-19 北京小马慧行科技有限公司 Method, device, storage medium and processor for determining whether annotation is accurate
US20200200907A1 (en) * 2018-12-19 2020-06-25 Dalong Li Semi-automatic lidar annotation system for autonomous driving
CN112132901A (en) * 2020-09-30 2020-12-25 上海商汤临港智能科技有限公司 Point cloud labeling method and device, electronic equipment and storage medium

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109727312A (en) * 2018-12-10 2019-05-07 广州景骐科技有限公司 Point cloud mask method, device, computer equipment and storage medium
US20200200907A1 (en) * 2018-12-19 2020-06-25 Dalong Li Semi-automatic lidar annotation system for autonomous driving
CN110136273A (en) * 2019-03-29 2019-08-16 初速度(苏州)科技有限公司 A kind of sample data mask method and device in machine learning
CN110135453A (en) * 2019-03-29 2019-08-16 初速度(苏州)科技有限公司 A kind of laser point cloud data mask method and device
CN111310667A (en) * 2020-02-18 2020-06-19 北京小马慧行科技有限公司 Method, device, storage medium and processor for determining whether annotation is accurate
CN112132901A (en) * 2020-09-30 2020-12-25 上海商汤临港智能科技有限公司 Point cloud labeling method and device, electronic equipment and storage medium

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115423934A (en) * 2022-08-12 2022-12-02 北京城市网邻信息技术有限公司 House type graph generation method and device, electronic equipment and storage medium
CN115423934B (en) * 2022-08-12 2024-03-08 北京城市网邻信息技术有限公司 House type diagram generation method and device, electronic equipment and storage medium
CN115542338A (en) * 2022-11-30 2022-12-30 湖南大学 Laser radar data learning method based on point cloud space distribution mapping
CN115880536A (en) * 2023-02-15 2023-03-31 北京百度网讯科技有限公司 Data processing method, training method, target object detection method and device
CN115880536B (en) * 2023-02-15 2023-09-01 北京百度网讯科技有限公司 Data processing method, training method, target object detection method and device
CN116152783A (en) * 2023-04-18 2023-05-23 安徽蔚来智驾科技有限公司 Target element labeling data acquisition method, computer equipment and storage medium
CN116152783B (en) * 2023-04-18 2023-08-04 安徽蔚来智驾科技有限公司 Target element labeling data acquisition method, computer equipment and storage medium

Also Published As

Publication number Publication date
CN112132901A (en) 2020-12-25
KR20220092928A (en) 2022-07-04

Similar Documents

Publication Publication Date Title
WO2022068225A1 (en) Point cloud annotating method and apparatus, electronic device, storage medium, and program product
EP3690815B1 (en) Method, medium and apparatus for automatically labeling target object within image
US20210279503A1 (en) Image processing method, apparatus, and device, and storage medium
CN112861653B (en) Method, system, equipment and storage medium for detecting fused image and point cloud information
US10824862B2 (en) Three-dimensional object detection for autonomous robotic systems using image proposals
WO2020102944A1 (en) Point cloud processing method and device and storage medium
US10977818B2 (en) Machine learning based model localization system
Li et al. Sustech points: A portable 3d point cloud interactive annotation platform system
US11978243B2 (en) System and method using augmented reality for efficient collection of training data for machine learning
WO2021114884A1 (en) Point cloud labeling method, apparatus, and system, device, and storage medium
WO2019042426A1 (en) Augmented reality scene processing method and apparatus, and computer storage medium
CN110986969B (en) Map fusion method and device, equipment and storage medium
US20110248995A1 (en) System and methods for creating interactive virtual content based on machine analysis of freeform physical markup
EP2560145A2 (en) Methods and systems for enabling the creation of augmented reality content
US20150138193A1 (en) Method and device for panorama-based inter-viewpoint walkthrough, and machine readable medium
CN108876706A (en) It is generated according to the thumbnail of panoramic picture
CN110070578B (en) Loop detection method
WO2022021782A1 (en) Method and system for automatically generating six-dimensional posture data set, and terminal and storage medium
KR20150082195A (en) Annotation method and apparatus
CN108597034B (en) Method and apparatus for generating information
CN116978010A (en) Image labeling method and device, storage medium and electronic equipment
CN112948605B (en) Point cloud data labeling method, device, equipment and readable storage medium
JP6304815B2 (en) Image processing apparatus and image feature detection method, program and apparatus thereof
CN114089836B (en) Labeling method, terminal, server and storage medium
CN113920191B (en) 6D data set construction method based on depth camera

Legal Events

Date Code Title Description
ENP Entry into the national phase

Ref document number: 2022514231

Country of ref document: JP

Kind code of ref document: A

121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21873887

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 20227018072

Country of ref document: KR

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 28.08.2023)

122 Ep: pct application non-entry in european phase

Ref document number: 21873887

Country of ref document: EP

Kind code of ref document: A1