WO2022246724A1 - Point cloud decoding and upsampling and model training methods and apparatus - Google Patents

Point cloud decoding and upsampling and model training methods and apparatus Download PDF

Info

Publication number
WO2022246724A1
WO2022246724A1 PCT/CN2021/096287 CN2021096287W WO2022246724A1 WO 2022246724 A1 WO2022246724 A1 WO 2022246724A1 CN 2021096287 W CN2021096287 W CN 2021096287W WO 2022246724 A1 WO2022246724 A1 WO 2022246724A1
Authority
WO
WIPO (PCT)
Prior art keywords
feature
point cloud
information
block
feature information
Prior art date
Application number
PCT/CN2021/096287
Other languages
French (fr)
Chinese (zh)
Inventor
元辉
刘昊
王婷婷
李明
Original Assignee
Oppo广东移动通信有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Oppo广东移动通信有限公司 filed Critical Oppo广东移动通信有限公司
Priority to CN202180096083.7A priority Critical patent/CN117242493A/en
Priority to PCT/CN2021/096287 priority patent/WO2022246724A1/en
Publication of WO2022246724A1 publication Critical patent/WO2022246724A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects

Definitions

  • the present application relates to the field of point cloud technology, and in particular to a point cloud decoding, upsampling and model training method and device.
  • the surface of the object is collected by the collection device to form point cloud data, which includes hundreds of thousands or more points.
  • point cloud data is transmitted between the point cloud encoding device and the point cloud decoding device in the form of point cloud media files.
  • point cloud encoding equipment needs to compress the point cloud data before transmission.
  • the point cloud decoding end decodes the point cloud code stream to obtain the reconstructed point cloud.
  • a lot of post-processing is required to improve the accuracy of point clouds to improve driving performance. safety.
  • current point cloud upsampling methods have poor point cloud upsampling effects and low accuracy.
  • the embodiment of the present application provides a point cloud decoding, upsampling and model training method and device, so as to improve the accuracy of point cloud upsampling.
  • the embodiment of the present application provides a point cloud decoding method, including:
  • the generator includes: a feature extraction module, a feature upsampling module and a geometry generation module, the feature extraction module is used to extract the first feature information of the point cloud block, and the feature sampling module is used to extract the The first feature information of the point cloud block is up-sampled to the second feature information, and the geometry generation module is used to map the second feature information of the point cloud block into a geometric space, so as to obtain the up-sampling of the point cloud block geometric information.
  • the present application provides a point cloud upsampling method, including:
  • the generator includes: a feature extraction module, a feature upsampling module and a geometry generation module, the feature extraction module is used to extract the first feature information of the point cloud block, and the feature sampling module is used to extract the The first feature information of the point cloud block is up-sampled to the second feature information, and the geometry generation module is used to map the second feature information of the point cloud block into a geometric space, so as to obtain the up-sampling of the point cloud block geometric information.
  • the present application provides a model training method, including:
  • the feature extraction module of the geometric information input generator of described training point cloud block is carried out feature extraction, obtains the first feature information of described training point cloud block;
  • the feature extraction module, feature upsampling module and geometry generation module in the generator are trained to obtain the trained generator.
  • a point cloud decoder configured to execute the method in the above first aspect or various implementations thereof.
  • the point cloud decoder includes a functional unit for executing the method in the above first aspect or its implementations.
  • a point cloud decoder including a processor and a memory.
  • the memory is used to store a computer program, and the processor is used to call and run the computer program stored in the memory to execute the method in the above first aspect or its various implementations.
  • a device for upsampling a point cloud configured to execute the method in the above second aspect or its various implementations.
  • the point cloud upsampling device includes a functional unit for executing the method in the above second aspect or its various implementations.
  • a point cloud upsampling device including a processor and a memory.
  • the memory is used to store a computer program
  • the processor is used to invoke and run the computer program stored in the memory, so as to execute the method in the above second aspect or its various implementations.
  • a model training device configured to execute the method in the above third aspect or various implementations thereof.
  • the model training device includes a functional unit for executing the method in the above third aspect or its various implementations.
  • a model training device including a processor and a memory.
  • the memory is used to store a computer program, and the processor is used to call and run the computer program stored in the memory, so as to execute the method in the above third aspect or its various implementations.
  • a chip configured to implement any one of the foregoing first to third aspects or the method in each implementation manner thereof.
  • the chip includes: a processor, configured to call and run a computer program from the memory, so that the device installed with the chip executes any one of the above-mentioned first to third aspects or any of the implementations thereof. method.
  • a computer-readable storage medium for storing a computer program, and the computer program causes a computer to execute any one of the above-mentioned first to third aspects or the method in each implementation manner thereof.
  • a twelfth aspect provides a computer program product, including computer program instructions, the computer program instructions cause a computer to execute any one of the above first to third aspects or the method in each implementation manner.
  • a thirteenth aspect provides a computer program, which, when running on a computer, causes the computer to execute any one of the above first to third aspects or the method in each implementation manner.
  • the point cloud is divided into at least one point cloud block through the geometric information of the point cloud; the geometric information of the point cloud block is input into the generator for up-sampling, and the up-sampling geometric information of the point cloud block is obtained; the generator Including: a feature extraction module, a feature upsampling module and a geometry generation module, the feature extraction module is used to extract the first feature information of the point cloud block, and the feature sampling module is used to upsample the first feature information of the point cloud block into the second feature Information, the geometry generation module is used to map the second feature information of the point cloud block into the geometric space, so as to obtain the upsampling geometric information of the point cloud block.
  • the generator in the embodiment of the present application is a generator based on deep learning, through which more characteristic information of the point cloud can be learned, and then when the generator is used for upsampling of the point cloud, a high-precision point cloud can be generated, Moreover, the features of the high-precision point cloud are close to the true value of the point cloud, thereby improving the accuracy of point cloud upsampling.
  • FIG. 1 is a schematic block diagram of a point cloud encoding and decoding system involved in an embodiment of the present application
  • Fig. 2 is a schematic block diagram of a point cloud encoder provided by an embodiment of the present application
  • Fig. 3 is a schematic block diagram of a point cloud decoder provided by an embodiment of the present application.
  • FIG. 4 is a schematic flow chart of a model training method provided by an embodiment of the present application.
  • FIG. 5 is a schematic diagram of a network of a generator according to an embodiment of the present application.
  • FIG. 6A is a schematic structural diagram of a feature extraction module involved in an embodiment of the present application.
  • FIG. 6B is a schematic structural diagram of a feature extraction block involved in an embodiment of the present application.
  • FIG. 6C is a schematic structural diagram of the second feature extraction unit HRA involved in the embodiment of the present application.
  • FIG. 6D is a schematic structural diagram of a residual block involved in the embodiment of the present application.
  • FIG. 6E is a schematic structural diagram of the second feature extraction unit HRA involved in the embodiment of the present application.
  • FIG. 6F is a schematic structural diagram of the second feature extraction unit HRA involved in the embodiment of the present application.
  • FIG. 6G is a schematic structural diagram of a gating unit involved in an embodiment of the present application.
  • FIG. 7A is a schematic structural diagram of a feature upsampling module involved in an embodiment of the present application.
  • FIG. 7B is another schematic structural diagram of the feature upsampling module involved in the embodiment of the present application.
  • FIG. 7C is a schematic structural diagram of the feature extraction submodule involved in the embodiment of the present application.
  • FIG. 7D is a schematic diagram of a specific network structure of the feature upsampling module provided by the embodiment of the present application.
  • FIG. 8 is a schematic diagram of a specific network structure of the geometry generation module provided by the embodiment of the present application.
  • FIG. 9 is a schematic diagram of a training process involving a generator according to an embodiment of the present application.
  • FIG. 10 is another schematic diagram of the training process involving the generator according to the embodiment of the present application.
  • FIG. 11 is a schematic diagram of a network structure of a discriminator
  • FIG. 12 is a schematic flowchart of a model training method provided by an embodiment of the present application.
  • FIG. 13 is a schematic diagram of a specific network structure of the discriminator provided in the embodiment of the present application.
  • FIG. 14 is a schematic flow diagram of a point cloud upsampling method provided in an embodiment of the present application.
  • FIG. 15 is a schematic diagram of a network structure of a generator involved in an embodiment of the present application.
  • FIG. 16 is a schematic flow diagram of a point cloud decoding method provided in an embodiment of the present application.
  • Fig. 17 is a schematic block diagram of a point cloud decoder provided by an embodiment of the present application.
  • Fig. 18 is a schematic block diagram of a point cloud upsampling device provided by an embodiment of the present application.
  • Fig. 19 is a schematic block diagram of a model training device provided by an embodiment of the present application.
  • Fig. 20 is a schematic block diagram of an electronic device provided by an embodiment of the present application.
  • the present application can be applied to the technical field of point cloud upsampling, for example, can be applied to the technical field of point cloud compression.
  • Point cloud refers to a set of discrete point sets randomly distributed in space, expressing the spatial structure and surface properties of 3D objects or 3D scenes.
  • Point cloud data is a specific record form of point cloud, and the points in the point cloud can include point location information and point attribute information.
  • the point position information may be three-dimensional coordinate information of the point.
  • the location information of a point may also be referred to as geometric information of a point.
  • the attribute information of a point may include color information and/or reflectivity and the like.
  • the color information may be information on any color space.
  • the color information may be (RGB).
  • the color information may be luminance and chrominance (YcbCr, YUV) information.
  • Y represents brightness (Luma)
  • Cb (U) represents blue color difference
  • Cr (V) represents red color
  • U and V are expressed as chromaticity (Chroma) for describing color difference information.
  • the points in the point cloud may include the three-dimensional coordinate information of the point and the laser reflection intensity (reflectance) of the point.
  • the points in the point cloud may include three-dimensional coordinate information and color information of the point.
  • the points in the point cloud may include the three-dimensional coordinate information of the point, the laser reflection intensity (reflectance) of the point, and the color information of the point.
  • Ways to obtain point cloud data may include but not limited to at least one of the following: (1) Generated by computer equipment.
  • the computer device can generate point cloud data according to virtual three-dimensional objects and virtual three-dimensional scenes.
  • Point cloud data of static real-world 3D objects or 3D scenes can be obtained through 3D laser scanning, and millions of point cloud data can be obtained per second;
  • 3D photography equipment that is, a group of cameras or camera equipment with multiple lenses and sensors
  • 3D photography can obtain dynamic real world three-dimensional objects Or point cloud data of a 3D scene.
  • point cloud data of biological tissues and organs can be obtained through magnetic resonance imaging (Magnetic Resonance Imaging, MRI), electronic computer tomography (Computed Tomography, CT), electromagnetic positioning information and other medical equipment.
  • Magnetic Resonance Imaging Magnetic Resonance Imaging
  • CT electronic computer tomography
  • electromagnetic positioning information and other medical equipment.
  • Point clouds can be divided into dense point clouds and sparse point clouds according to the way of acquisition.
  • point cloud is divided into:
  • the first type of static point cloud that is, the object is stationary, and the device for obtaining the point cloud is also stationary;
  • the second type of dynamic point cloud the object is moving, but the device for obtaining the point cloud is still;
  • the third type of dynamic acquisition of point clouds the equipment for acquiring point clouds is in motion.
  • point cloud According to the purpose of point cloud, it can be divided into two categories:
  • Category 1 Machine perception point cloud, which can be used in scenarios such as autonomous navigation systems, real-time inspection systems, geographic information systems, visual sorting robots, and emergency rescue robots;
  • Category 2 Human eyes perceive point clouds, which can be used in point cloud application scenarios such as digital cultural heritage, free viewpoint broadcasting, 3D immersive communication, and 3D immersive interaction.
  • the point cloud upsampling method provided by the embodiment of the present application can be applied to the point cloud encoding and decoding framework, for example, the geometric information of the point cloud parsed from the code stream by the point cloud decoder is upsampled to obtain Upsampled point clouds with higher accuracy.
  • FIG. 1 is a schematic block diagram of a point cloud encoding and decoding system involved in an embodiment of the present application. It should be noted that FIG. 1 is just an example, and the point cloud encoding and decoding system in the embodiment of the present application includes but is not limited to what is shown in FIG. 1 .
  • the point cloud encoding and decoding system 100 includes an encoding device 110 and a decoding device 120 .
  • the encoding device is used to encode the point cloud data (which can be understood as compression) to generate a code stream, and transmit the code stream to the decoding device.
  • the decoding device decodes the code stream generated by the encoding device to obtain decoded point cloud data.
  • the encoding device 110 in the embodiment of the present application can be understood as a device having a point cloud encoding function
  • the decoding device 120 can be understood as a device having a point cloud decoding function.
  • Devices including, for example, smartphones, desktop computers, mobile computing devices, notebook (e.g., laptop) computers, tablet computers, set-top boxes, televisions, cameras, display devices, digital media players, point cloud gaming consoles, vehicle-mounted computers, etc. .
  • the encoding device 110 can transmit the encoded point cloud data (eg code stream) to the decoding device 120 via the channel 130 .
  • Channel 130 may include one or more media and/or devices capable of transmitting encoded point cloud data from encoding device 110 to decoding device 120 .
  • channel 130 includes one or more communication media that enable encoding device 110 to transmit encoded point cloud data directly to decoding device 120 in real-time.
  • the encoding device 110 may modulate the encoded point cloud data according to the communication standard, and transmit the modulated point cloud data to the decoding device 120 .
  • the communication medium includes a wireless communication medium, such as a radio frequency spectrum.
  • the communication medium may also include a wired communication medium, such as one or more physical transmission lines.
  • the channel 130 includes a storage medium, which can store the point cloud data encoded by the encoding device 110 .
  • the storage medium includes a variety of local access data storage media, such as optical discs, DVDs, flash memory, and the like.
  • the decoding device 120 can acquire encoded point cloud data from the storage medium.
  • the channel 130 may include a storage server, and the storage server may store the point cloud data encoded by the encoding device 110 .
  • the decoding device 120 may download the stored encoded point cloud data from the storage server.
  • the storage server can store the encoded point cloud data and can transmit the encoded point cloud data to the decoding device 120, such as a web server (for example, for a website), a file transfer protocol (FTP) server, etc. .
  • a web server for example, for a website
  • FTP file transfer protocol
  • the encoding device 110 includes a point cloud encoder 112 and an output interface 113 .
  • the output interface 113 may include a modulator/demodulator (modem) and/or a transmitter.
  • the encoding device 110 may include a point cloud source 111 in addition to the point cloud encoder 112 and the input interface 113 .
  • the point cloud source 111 may include at least one of a point cloud acquisition device (for example, a scanner), a point cloud archive, a point cloud input interface, and a computer graphics system, wherein the point cloud input interface is used to receive from a point cloud content provider Point cloud data, computer graphics system is used to generate point cloud data.
  • a point cloud acquisition device for example, a scanner
  • a point cloud archive for example, a point cloud archive
  • a point cloud input interface for example, a point cloud archive
  • point cloud input interface for example, a point cloud input interface
  • computer graphics system is used to generate point cloud data.
  • the point cloud encoder 112 encodes the point cloud data from the point cloud source 111 to generate a code stream.
  • the point cloud encoder 112 directly transmits the encoded point cloud data to the decoding device 120 via the output interface 113 .
  • the encoded point cloud data can also be stored on a storage medium or a storage server for subsequent reading by the decoding device 120 .
  • the decoding device 120 includes an input interface 121 and a point cloud decoder 122 .
  • the decoding device 120 may further include a display device 123 in addition to the input interface 121 and the point cloud decoder 122 .
  • the input interface 121 includes a receiver and/or a modem.
  • the input interface 121 can receive the encoded point cloud data through the channel 130 .
  • the point cloud decoder 122 is used to decode the encoded point cloud data to obtain decoded point cloud data, and transmit the decoded point cloud data to the display device 123 .
  • the display device 123 displays the decoded point cloud data.
  • the display device 123 may be integrated with the decoding device 120 or external to the decoding device 120 .
  • the display device 123 may include various display devices, such as a liquid crystal display (LCD), a plasma display, an organic light emitting diode (OLED) display, or other types of display devices.
  • LCD liquid crystal display
  • plasma display a plasma display
  • OLED organic light emitting diode
  • FIG. 1 is only an example, and the technical solution of the embodiment of the present application is not limited to FIG. 1 .
  • the technology of the present application can also be applied to one-sided point cloud encoding or one-sided point cloud decoding.
  • the current point cloud encoder can use the Geometry Point Cloud Compression (G-PCC) codec framework provided by the Moving Picture Experts Group (MPEG) or the video-based point cloud compression (Video Point Cloud Compression, V-PCC) codec framework, or the AVS-PCC codec framework provided by Audio Video Standard (AVS). Both G-PCC and AVS-PCC are aimed at static sparse point clouds, and their coding frameworks are roughly the same.
  • the G-PCC codec framework can be used to compress the first static point cloud and the third type of dynamically acquired point cloud, and the V-PCC codec framework can be used to compress the second type of dynamic point cloud.
  • the G-PCC codec framework is also called point cloud codec TMC13, and the V-PCC codec framework is also called point cloud codec TMC2.
  • Fig. 2 is a schematic block diagram of a point cloud encoder provided by an embodiment of the present application.
  • the points in the point cloud can include the position information of the point and the attribute information of the point, therefore, the encoding of the point in the point cloud mainly includes the position encoding and the attribute encoding.
  • the position information of the points in the point cloud is also called geometric information, and the corresponding position codes of the points in the point cloud may also be called geometric codes.
  • the process of position encoding includes: preprocessing the points in the point cloud, such as coordinate transformation, quantization, and removing duplicate points, etc.; then, geometrically encoding the preprocessed point cloud, such as constructing an octree, based on the constructed The octree performs geometric encoding to form a geometric code stream. At the same time, based on the position information output by the constructed octree, the position information of each point in the point cloud data is reconstructed to obtain the reconstruction value of the position information of each point.
  • the attribute encoding process includes: by given the reconstruction information of the position information of the input point cloud and the original value of the attribute information, select one of the three prediction modes for point cloud prediction, quantify the predicted results, and perform arithmetic coding to form property stream.
  • position coding can be achieved by the following units:
  • Coordinate conversion (Tanmsform coordinates) unit 201, quantization and removal of repeated points (Quantize and remove points) unit 202, octree analysis (Analyze octree) unit 203, geometric reconstruction (Reconstruct geometry) unit 204 and first arithmetic coding (Arithmetic enconde) unit 205.
  • the coordinate transformation unit 201 can be used to transform the world coordinates of points in the point cloud into relative coordinates. For example, subtracting the minimum values of the xyz coordinate axes from the geometric coordinates of the point is equivalent to a DC operation to convert the coordinates of the points in the point cloud from world coordinates to relative coordinates.
  • Quantization and removal of duplicate points unit 202 can reduce the number of coordinates by quantization; original different points may be given the same coordinates after quantization, based on this, duplicate points can be deleted through de-duplication operations; for example, with the same quantization position and Multiple clouds with different attribute information can be merged into one cloud through attribute conversion.
  • the Quantize and Remove Duplicate Points unit 202 is an optional unit module.
  • the octree analysis unit 203 may use an octree encoding method to encode the position information of the quantized points.
  • the point cloud is divided in the form of an octree, so that the position of the point can be in one-to-one correspondence with the position of the octree, and the position of the point in the octree is counted, and its flag (flag) is recorded as 1 for geometric encoding.
  • the geometry reconstruction unit 204 may perform position reconstruction based on the position information output by the octree analysis unit 203 to obtain reconstruction values of the position information of each point in the point cloud data.
  • the first arithmetic coding unit 205 can arithmetically encode the position information output by the octree analysis unit 203 in an entropy coding manner, that is, the position information output by the octree analysis unit 203 is generated using an arithmetic coding method to generate a geometric code stream; the geometric code stream is also Can be called geometry bitstream (geometry bitstream).
  • Attribute coding can be achieved by the following units:
  • Color space conversion (Transform colors) unit 210 attribute conversion (Transfer attributes) unit 211, region adaptive layered transformation (Region Adaptive Hierarchical Transform, RAHT) unit 212, prediction change (predicting transform) unit 213 and lifting transform (lifting transform) ) unit 214, a quantization coefficient (Quantize coefficients) unit 215, and a second arithmetic coding unit 216.
  • Transform colors attribute conversion
  • Transfer attributes region adaptive layered transformation
  • RAHT Region adaptive layered transformation
  • prediction change predicting transform
  • lifting transform lifting transform
  • quantization coefficient quantization coefficient
  • point cloud encoder 200 may include more, less or different functional components than those shown in FIG. 2 .
  • the color space conversion unit 210 can be used to convert the RGB color space of points in the point cloud into YCbCr format or other formats.
  • the attribute conversion unit 211 can be used to convert attribute information of points in the point cloud to minimize attribute distortion.
  • the attribute conversion unit 211 can be used to obtain the original value of the attribute information of the point.
  • the attribute information may be color information of dots.
  • any prediction unit can be selected to predict the point in the point cloud.
  • the prediction unit may include: RAHT 212, predicting transform unit 213, and lifting transform unit 214.
  • any one of the RAHT 212, the predicting transform unit 213 and the lifting transform unit 214 can be used to predict the attribute information of the points in the point cloud, so as to obtain the predicted values of the attribute information of the points, Furthermore, the residual value of the attribute information of the point is obtained based on the predicted value of the attribute information of the point.
  • the residual value of the point's attribute information may be the original value of the point's attribute information minus the predicted value of the point's attribute information.
  • the predictive transformation unit 213 may also be used to generate a level of detail (LOD).
  • LOD level of detail
  • the generation process of LOD includes: according to the position information of the points in the point cloud, the Euclidean distance between the points is obtained; according to the Euclidean distance, the points are divided into different detail expression layers.
  • the Euclidean distances in different ranges can be divided into different detail expression layers. For example, a point can be randomly selected as the first detail expression layer. Then calculate the Euclidean distance between the remaining points and this point, and classify the points whose Euclidean distance meets the first threshold requirement as the second detailed expression layer.
  • the point cloud can be directly divided into one or more detail expression layers, or the point cloud can be divided into multiple point cloud slices first, and then each point cloud slice can be divided into one or more Multiple LOD layers.
  • the point cloud can be divided into multiple point cloud cutouts, and the number of points in each point cloud cutout can be between 550,000 and 1.1 million.
  • Each point cloud slice can be regarded as a separate point cloud.
  • Each point cloud slice can be divided into multiple detail expression layers, and each detail expression layer includes multiple points.
  • the detail expression layer can be divided according to the Euclidean distance between points.
  • the quantization unit 215 may be used to quantize residual values of attribute information of points. For example, if the quantization unit 215 is connected to the predictive transformation unit 213, the quantization unit may be used to quantize the residual value of the point attribute information output by the predictive transformation unit 213.
  • the residual value of the point attribute information output by the predictive transformation unit 213 is quantized using the quantization step size, so as to improve system performance.
  • the second arithmetic coding unit 216 may use zero run length coding to perform entropy coding on the residual value of the attribute information of the point to obtain an attribute code stream.
  • the attribute code stream may be bit stream information.
  • Fig. 3 is a schematic block diagram of a point cloud decoder provided by an embodiment of the present application.
  • the decoder 300 can obtain the point cloud code stream from the encoding device, and obtain the position information and attribute information of the points in the point cloud by parsing the code.
  • the decoding of point cloud includes position decoding and attribute decoding.
  • the process of position decoding includes: performing arithmetic decoding on the geometric code stream; merging after constructing the octree, and reconstructing the position information of the point to obtain the reconstruction information of the position information of the point; Transform to get the position information of the point.
  • the location information of a point may also be referred to as geometric information of a point.
  • the attribute decoding process includes: obtaining the residual value of the attribute information of the point cloud by parsing the attribute code stream; dequantizing the residual value of the attribute information of the point to obtain the residual value of the attribute information of the dequantized point value; based on the reconstruction information of the position information of the point obtained in the position decoding process, select one of the following three prediction modes: RAHT, prediction change and promotion change to predict the point cloud, and obtain the predicted value, which is consistent with the residual value Add the reconstruction value of the attribute information of the point; perform color space inverse transformation on the reconstruction value of the attribute information of the point to obtain the decoded point cloud.
  • position decoding can be achieved by the following units:
  • a first arithmetic decoding unit 301 an octree analysis unit 302 , a geometry reconstruction unit 304 and an inverse transform coordinates unit 305 .
  • Attribute coding can be achieved by the following units:
  • decompression is an inverse process of compression
  • the functions of each unit in the decoder 300 may refer to the functions of corresponding units in the encoder 200 .
  • the point cloud decoder 300 may include more, fewer or different functional components than in FIG. 3 .
  • the decoder 300 can divide the point cloud into multiple LODs according to the Euclidean distance between points in the point cloud; then, decode the attribute information of the points in the LOD in sequence; number (zero_cnt), to decode the residual based on zero_cnt; then, the decoding framework 200 can perform dequantization based on the decoded residual value, and add the dequantized residual value to the predicted value of the current point to obtain the Point cloud reconstruction values until all point clouds are decoded.
  • the current point will be used as the nearest neighbor point of the subsequent LOD midpoint, and the attribute information of the subsequent point will be predicted by using the reconstructed value of the current point.
  • the current point cloud encoding and decoding method reconstructs the point cloud to the original scale, but in some application scenarios, it is necessary to use high-quality point clouds with higher precision than the original ones, for example, sparse point clouds collected by radar in areas such as autonomous driving , often need to do a lot of post-processing work to improve the accuracy of the point cloud to improve driving safety.
  • the embodiment of the present application provides a point cloud upsampling method, which uses deep learning to upsample point cloud geometric information to obtain a higher resolution (or precision) point cloud, thereby meeting the task requirements for high-precision point clouds.
  • the point cloud upsampling method provided in this application uses a generator after deep learning to upsample the geometric information of the point cloud, and the generator is a piece of software code or a chip with data processing functions. Based on this, the training process of the generator is firstly introduced.
  • Fig. 4 is a schematic flow chart of the model training method provided by an embodiment of the present application. As shown in Fig. 4, the training process of the generator includes:
  • the embodiment of the present application records the point cloud used for generator training as a training point cloud.
  • the above-mentioned training point cloud is a point cloud in the training set, which includes multiple point clouds, and the process of using each point cloud in the training set to train the generator is consistent.
  • the embodiment of the present application uses a training set Take point clouds as an example.
  • the point cloud in the process of up-sampling the point cloud, the point cloud is divided into point cloud blocks, and the point cloud geometric information is up-sampled with the point cloud blocks as objects.
  • the ways of dividing the training point cloud into at least one training point cloud block in S402 include but are not limited to the following ways:
  • Method 1 Divide the training point cloud into at least one training point cloud block of equal size according to the geometric information of the training point cloud. That is to say, the geometric scale of each point cloud block is the same.
  • Method 2 Divide the training point cloud into at least one training point cloud block according to the geometric information of the training point cloud, and each training point cloud block includes the same number of points.
  • Method 3 Obtain at least one seed point from the training point cloud according to the geometric information of the training point cloud, for example, randomly sample a specified number of seed points from the training point cloud by using Monte Carlo random sampling method. For each seed point, determine the neighboring points of the seed point, divide the seed point and the neighboring points of the seed point into a training point cloud block, and obtain at least one training point cloud block.
  • the obtained training point cloud blocks are also called point cloud patches (Patch), and the number of points included in each training point cloud block in the training point cloud blocks obtained in this way is the same.
  • the training point cloud blocks obtained above are recorded as Where N is the number of points included in the training point cloud block, and 3 is the geometric information dimension of the training point cloud block.
  • the network structure of the generator involved in the embodiment of the present application will be introduced below in conjunction with FIG. 5. It should be noted that the network structure of the generator in the embodiment of the present application includes but is not limited to the modules shown in FIG. more or less modules.
  • Fig. 5 is a kind of network schematic diagram of the generator of the embodiment of the present application, as shown in Fig. 5, generator includes feature extraction module, feature upsampling module and geometry generation module, and feature extraction module is used to extract the first of training point cloud block One feature information, the feature sampling module is used to upsample the first feature information of the training point cloud block into the second feature information, and the geometry generation module is used to map the second feature information of the training point cloud block into the geometric space, so as to obtain Upsampled geometric information for training point cloud blocks.
  • the feature extraction module is used to extract the expressive features of each point, such as the geometric information of the low-resolution training point cloud block
  • the feature extraction module is used to extract the expressive feature information of each point in the training point cloud block, and output the feature information of the training point cloud block
  • N is the number of points in the training point cloud block
  • 3 is the geometric information dimension
  • C is the feature dimension.
  • the feature information output by the feature extraction module is recorded as the first feature information of the training point cloud block.
  • the present application proposes a feature extraction module based on a Dynamic Graph Hierarchical Residual Aggregation (DGHRA) unit, as shown in FIG. 6A , the feature extraction module includes: M densely connected feature extraction blocks (Feature Extraction Block, referred to as FEB), that is, the output of the previous FEB is used as the input of each subsequent FEB, as shown in Figure 6A.
  • DGHRA Dynamic Graph Hierarchical Residual Aggregation
  • the above S403 includes the following S403-A1 to S403-A4:
  • i is a positive integer smaller than M.
  • the embodiment of the present application further includes: determining the initial feature information of the training point cloud block according to the geometric information of the training point cloud block; inputting the initial feature information of the training point cloud block into the first In the first feature extraction block, the first third feature information of the training point cloud block extracted by the first feature extraction block is obtained.
  • the above S403-A2 includes: acquiring the third feature information extracted by each feature extraction block before the i-th feature extraction block among the M feature extraction blocks;
  • the third feature information extracted by each feature extraction block located before the i-th feature extraction block is concatenated with the third feature information extracted by the i-th feature extraction block, as the i-th feature information of the training point cloud block Four feature information.
  • the first third feature information extracted by the first feature extraction block in the M feature extraction units is used as the first fourth feature information of the training point cloud block .
  • the feature extraction module includes 4 feature extraction blocks FEB, as shown in Figure 6A, first determine the initial feature information of the training point cloud block according to the geometric information of the training point cloud block; the training point cloud The initial feature information of the block is input into the first feature extraction block to obtain the first third feature information of the training point cloud block extracted by the first feature extraction block. Since there is no other feature extraction block before the first feature extraction block, the first third feature information is input into the second FEB as the first fourth feature information, and the second FEB is based on the first fourth feature information Output the second and third feature information.
  • the second piece of fourth characteristic information outputs the third third characteristic information. Concatenate the third third characteristic information, the second third characteristic information and the first third characteristic information as the third fourth characteristic information, and input the third fourth characteristic information into the fourth FEB Among them, the fourth FEB outputs the fourth third feature information according to the third fourth feature information, and the fourth third feature information is used as the first feature information of the training point cloud block.
  • convolution is set between FEBs Network, such as a convolutional network with a convolution kernel set to 1X1, to reduce the feature dimension in the input FEB.
  • each feature extraction block FEB includes a first feature extraction unit and at least one second feature extraction unit connected in series, wherein the first feature extraction unit is a dynamic map layered residual Aggregation (Dynamic graph hierarchical residual aggregation, DGHRA) unit, the second feature extraction unit is hierarchical residual aggregation block (Hierarchical residual aggregation, referred to as HRA), used to extract more detailed features.
  • the processing process of each FEB is the same, and they are iterated with each other.
  • the processing process of each FEB is the same. It is convenient to describe as an example. Take the i+1th feature extraction block as an example.
  • the above S403-A3 includes S403-A31 To S403-A32:
  • the size of the i-th fourth feature information of the training point cloud block is NXC
  • the i-th fourth feature information of the training point cloud block with a size of NXC is input in the first feature extraction unit, for the training point cloud block
  • the first feature extraction unit searches K neighboring points of the current point, for example, dynamically searches the K neighboring points of the current point through a feature space nearest neighbor search method.
  • the fourth feature information of the current point from the ith fourth feature information of the training point cloud block, and copy K copies of the fourth feature information of the current point, and compare the fourth feature information of the current point with K neighboring
  • the fourth feature information of each adjacent point in the point is subtracted to obtain K residual feature information, the size of which is 1XKXC.
  • the K residual feature information of the current point is concatenated with the fourth feature information of the current point to obtain the ith concatenated feature information of the current point.
  • the i-th concatenated feature information of each point in the training point cloud block can be obtained, and then the i-th concatenated feature information of the training point cloud block can be obtained, and its size is NXKX2C.
  • the i+1th feature extraction block includes 3 second feature extraction units, and the i-th cascaded feature information of the above-mentioned training point cloud block size of NXKX2C is input into the first i+1th feature extraction block
  • the second feature extraction unit, the first second feature extraction unit extracts the detailed features of the training point cloud block, outputs the first fifth feature information of the training point cloud block, and inputs the first fifth feature information
  • the second second feature extraction unit, the second second feature extraction unit outputs the second fifth feature information of the training point cloud block according to the first fifth feature information, and inputs the second fifth feature information into the second Three second feature extraction units
  • the third second feature extraction unit outputs the third fifth feature information of the training point cloud block according to the second fifth feature information
  • the third fifth feature information is used as the training point The i+1th third characteristic information of the cloud block.
  • the size of the i+1th third feature information is NXC.
  • the second feature extraction unit includes P residual blocks (Residual block, RB for short), and P is a positive integer.
  • the above S403-A32 includes:
  • the network structure of the residual block is as shown in Figure 6D
  • the residual block includes multiple linear layers with a linear rectification function (Relu)
  • the residual block used in the embodiment of the present application is used for Feature mining helps the network converge.
  • the first The RB outputs the first residual information 1, and adds the first residual information 1 and the i-th concatenated feature information to the second RB, and the second RB outputs the first residual information 2, and the first residual information
  • the information 2 and the i-th concatenated feature information are added and input to the third RB to obtain the first residual information 3 output by the third RB, and the first residual information 3 and the i-th concatenated feature information are added and input to the first Four RBs, obtain the first residual information 4 output by the fourth RB, and then determine the fifth feature output by the first feature extraction unit according to the first residual information output by each RB and the i-th concatenated feature information information.
  • the ways of outputting the fifth feature information include but not limited to the following:
  • Way 1 add the first residual information output by the last residual block to the i-th concatenated feature information, and use it as the fifth feature information output by the first second feature extraction unit.
  • Method 2 the above S403-A323 includes:
  • Step B1 combine the first residual information output by the last residual block among the P residual blocks in the first second feature extraction unit, and the first residual information output by at least one residual block among the P-1 residual blocks.
  • the residual information is concatenated, wherein the P-1 residual blocks are the residual blocks except the last residual block among the P residual blocks;
  • the first residual information output by the last residual block in the P residual blocks is compared with the first residual information output by each residual block in the P-1 residual blocks. Cascade to obtain the feature information after cascading.
  • Step B2 according to the concatenated feature information and the i-th concatenated feature information, determine the fifth feature information output by the first second feature extraction unit.
  • step B2 The implementation methods of the above step B2 include but are not limited to the following:
  • Way 1 add the concatenated feature information to the i-th concatenated feature information, and use it as the fifth feature information output by the first second feature extraction unit.
  • the second feature extraction unit also includes a gating unit, and at this time the above step B2 includes inputting the cascaded feature information into the gating unit for de-redundancy, and obtaining the de-redundant feature information ; Add the feature information after de-redundancy to the i-th cascaded feature information, and use it as the fifth feature information output by the first second feature extraction unit.
  • the residual block of the embodiment of the present application provides detailed residual information, and inputs the detailed residual information obtained by each residual block into the gating unit to collect more feature details and realize full learning of the network.
  • the embodiment of the present application does not limit the network structure of the above-mentioned gate control unit.
  • the gating unit includes Squeeze-and-excitation networks (SE-Net for short) and a linear layer, where SE-Net includes a global average Pooling layer, full connection (Full connection, referred to as FC) layer, the SE-Net network can perform global average pooling of feature information with a size of NXKXC, and obtain feature information with a size of 1X1XC.
  • SE-Net Squeeze-and-excitation networks
  • FC Full connection
  • the size is the feature information of 1X1XC, multiply the feature information of 1X1XC processed by the fully connected layer with the feature information of size NXKXC on the channel, and obtain the feature information of size NXKXC after de-redundancy, and finally de-redundancy
  • the feature information whose size is the feature information of NXKXC is input to the linear layer, and finally the feature information after de-redundancy is output.
  • the process of inputting the geometric information of the training point cloud block into the feature extraction module to obtain the first feature information of the training point cloud block is described in detail.
  • After obtaining the first feature information of the training point cloud block perform the following S404 to up-sample the first feature information to obtain the second feature information of the training point cloud.
  • the feature upsampling module is used to upsample the first feature information of the training point cloud block to obtain the second feature information of the training point cloud block, for example, the first feature information of the training point cloud block Upsampling is the second feature information
  • r is a preset sampling rate, which is a positive integer
  • C′ is the feature dimension of the second feature information of the training point cloud block after upsampling.
  • the feature upsampling module includes: a feature upsampling submodule and a feature extraction submodule, wherein the feature upsampling submodule is used to upsample the first feature information of the training point cloud block , to obtain the upsampled feature information of the training point cloud block.
  • the feature extraction sub-module is used to perform feature extraction on the upsampled feature information of the training point cloud block to obtain expressive features of the training point cloud block, and use the expressive feature as the second feature information of the training point cloud block.
  • S404 includes S404-A1 and S404-A2:
  • copy r copies of the first feature information F of the training point cloud block and for each assigned feature, add an n-dimensional vector to its feature dimension, so that there is a clear difference between each copied feature
  • the difference, at this time the feature dimension of each point is C+n.
  • the feature information whose feature dimension is C+n is recorded as the upsampling feature information of the training point cloud block.
  • the upsampled feature information of the training point cloud block is input into the feature extraction sub-module to perform detail feature extraction to obtain the second feature information of the training point cloud block.
  • the feature upsampling module further includes a first autocorrelation attention network.
  • the above S404-A2 includes: inputting the first upsampling feature information of the training point cloud block into the first The autocorrelation attention network performs feature interaction to obtain the upsampling feature information of the training point cloud block after the feature interaction; the upsampling feature information of the training point cloud block after the feature interaction is input into the feature extraction sub-module for feature extraction to obtain the training point The second characteristic information of the cloud block.
  • the feature dimension of the upsampled feature information of the training point cloud block after the feature interaction is the same as the feature dimension of the upsampled feature information of the training point cloud block. That is, the first autocorrelation attention network is used for feature interaction, allowing the network to learn more detailed features.
  • the first autocorrelation attention network in the embodiment of the present application has a reduced left and right to reduce the feature dimension of the feature information, so that the feature dimension of the upsampled feature information of the training point cloud block after feature interaction is lower than that of the training point.
  • the first autocorrelation attention network is not only used for feature interaction, but also used to reduce the feature dimension to reduce the training complexity of the network, thereby improving the training speed of the network.
  • the feature extraction submodule includes Q third feature extraction units connected in series, Q is a positive integer, wherein the feature extraction process of each third feature extraction unit is the same, at this time,
  • the above S404-A2 includes:
  • the upsampling feature information of the training point cloud block into the first third feature extraction unit for feature extraction, obtain the first enhanced upsampling feature information of the training point cloud block, and use the first enhanced Input the upsampling feature information into the second third feature extraction unit for feature extraction, obtain the second enhanced upsampling feature information of the training point cloud block, and input the second enhanced upsampling feature information into the third third feature extraction unit Perform feature extraction to obtain the third enhanced upsampling feature information of the training point cloud block, and record the third enhanced upsampling feature information of the training point cloud block as the second feature information of the training point cloud block.
  • the network structure of the above-mentioned third feature extraction unit is the same as that of the above-mentioned second feature extraction unit.
  • the network structures of the third feature extraction unit and the second feature extraction unit are not completely the same.
  • the third feature extraction unit includes L residual blocks, where L is a positive integer.
  • S404-A22 includes:
  • S404-A221 input the kth enhanced upsampling feature information of the training point cloud block into the k+1th third feature extraction unit, and obtain the output of the lth residual block in the k+1th third feature extraction unit
  • the second residual information, l is a positive integer less than or equal to L;
  • the network structure of the residual block is as shown in Figure 6D
  • the residual block includes multiple linear layers with a linear rectification function (Relu)
  • the residual block used in the embodiment of the present application is used for Feature mining helps the network converge.
  • the kth enhanced upsampling feature information of the training point cloud block is input into the first k+1th third feature extraction unit.
  • a residual block RB the first RB outputs the second residual information 1, and the second residual information 1 and the kth enhanced upsampling feature information are added to the second RB, and the second RB outputs the first Two residual information 2, the second residual information 2 and the kth enhanced upsampling feature information are added to the third RB to obtain the second residual information 3 output by the third RB, and the second residual information 3 Adding the kth enhanced upsampling feature information to the fourth RB to obtain the second residual information 4 output by the fourth RB, and then according to the second residual information output by each RB and the kth enhanced upsampling feature information , to get the k+1 enhanced upsampling feature information of the training point cloud block.
  • the k+1th enhanced upsampling feature of the training point cloud block is obtained.
  • Information methods include but are not limited to the following:
  • Method 1 add the second residual information output by the last residual block in the L residual blocks to the kth enhanced upsampling feature information, and use it as the i+1th third feature information of the training point cloud block .
  • Method 2 the above S404-A223 includes step C1 and step C2:
  • Step C1 concatenate the second residual information output by the last residual block in the L residual blocks with the second residual information output by at least one residual block in the L-1 residual blocks, where L -1 residual block is a residual block except the last residual block among the L residual blocks.
  • the second residual information output by the last residual block in the L residual blocks is compared with the second residual information output by each residual block in the L-1 residual blocks. Cascade to obtain the feature information after cascading.
  • Step C2 according to the concatenated feature information and the kth upsampling feature information, determine the k+1th enhanced upsampling feature information of the training point cloud block.
  • step C2 The implementation methods of the above step C2 include but are not limited to the following:
  • Way 1 add the cascaded feature information to the kth enhanced upsampling feature information, and use it as the k+1th enhanced upsampling feature information of the training point cloud block.
  • the third feature extraction unit also includes a gating unit, and at this time, the above step C2 includes: inputting the cascaded feature information into the gating unit for de-redundancy to obtain de-redundant feature information; The final feature information is added to the k-th enhanced up-sampled feature information, and used as the k+1-th enhanced up-sampled feature information of the training point cloud block.
  • FIG. 7D is a schematic diagram of a specific network structure of the feature upsampling module provided in the embodiment of the present application.
  • the feature upsampling submodule upsamples the first feature information with a size of NXC to a size of rNX(C+ 2)
  • the upsampled feature information of the size rNX(C+2) is input into the first self-correlation attention network (Self-attetion), and the upsampled feature information after feature interaction is obtained, and the feature interaction is
  • the embodiment of the present application does not limit the network structure of the above-mentioned gate control unit.
  • the network structure of the above-mentioned gate control unit is as shown in FIG. 6G , and for details, refer to the above-mentioned description of S403 .
  • the above network structure combined with the feature upsampling module describes in detail the process of inputting the first feature information of the training point cloud block into the feature upsampling module to obtain the second feature information of the training point cloud block. After obtaining the second feature information of the training point cloud block, perform the following S405 to perform spatial conversion on the second feature information to obtain upsampled geometric information of the training point cloud block.
  • the function of the geometry generation module in the embodiment of the present application is to obtain the second feature information of the training point cloud by upsampling Remap from the feature space back to the geometric space, and finally obtain the upsampled point cloud, that is, return F up to the geometric space, and obtain the upsampled geometric information of the training point cloud block Among them, 3 refers to the geometric information dimension, and rN is the number of points included in the training point cloud block after upsampling.
  • the embodiment of the present application does not limit the specific network structure of the geometry generation module.
  • the geometry generation module includes a plurality of fully connected layers
  • the above S405 includes: inputting the second feature information of the training point cloud block into a plurality of fully connected layers for space conversion, and obtaining the upsampled geometry of the training point cloud block information.
  • directly outputting the upsampled geometric information cannot well generate a uniformly distributed point cloud and suppress noise at the boundary.
  • this application will upsample the point cloud by r+m times; then generate the upsampled geometric information through FC; then use the high-pass image filter to explicitly remove multiple high-frequency points (ie noise) in each upsampled Patch, and finally pass the farthest point sampling (Farthest point sampling , referred to as FPS) algorithm, downsampling the point cloud to r times, output
  • FPS farthest point sampling
  • the geometry generation module includes: a geometry reconstruction unit, a filter unit, and a downsampling unit, wherein the geometry reconstruction unit includes multiple fully connected layers.
  • the above S405 includes:
  • the filtering unit may be a high-pass image filter, which explicitly removes a plurality of, for example, 5 high-frequency points (ie, noise points) in each upsampling patch.
  • FPS furthest point sampling
  • the implementation of the above S406 includes but is not limited to the following methods:
  • Method 1 According to the loss between the predicted upsampled geometric information of the training point cloud block and the upsampled true value of the geometric information of the training point cloud block, the feature extraction module, feature upsampling module and geometry generation module in the reverse training generator module to get the trained generator.
  • training process of the embodiment of the present application is an iterative process, and each training process is consistent, and the parameters in the generator (such as the weight matrix) are updated once during each training process until the model training end condition is reached until.
  • the model training end condition includes that the number of training times reaches a preset number of times, or the prediction error of the generator reaches a preset value, and the like.
  • Fig. 9 is a schematic diagram of the training process involving the generator in the embodiment of the present application.
  • the geometric information of the training point cloud block is input into the generator, and the predicted upsampling geometry of the training point cloud block output by the generator is obtained.
  • Information, according to the predicted upsampled geometric information of the training point cloud and the upsampled true value of the geometric information of the training point cloud adjust the parameters of the feature extraction module, feature upsampling module and geometry generation module in the generator, for example, according to the training
  • the loss between the upsampled geometric information of the point cloud and the upsampled true value of the geometric information of the training point cloud updates the parameter matrix of the feature extraction module, feature upsampling module and geometry generation module to obtain a trained generator .
  • the upsampled true value of the geometric information of the training point cloud can be understood as the data included in the training data after upsampling the geometric information of the training point cloud.
  • the resolution of the upsampled true value of the geometric information of the training point cloud is lower than the upsampled geometric information of the training point cloud output by the generator.
  • Method 2 training the generator with the help of the judger, at this time the above S406 includes:
  • the discriminator may be a piece of software code or a chip with data processing function.
  • Fig. 10 is another schematic diagram of the training process involving the generator in the embodiment of the present application.
  • the geometric information of the training point cloud block is input into the generator to obtain the upsampled geometry of the training point cloud block output by the generator
  • the upsampled geometric information of the training point cloud block is input into the discriminator, and obtain the first discriminant result output by the discriminator.
  • the parameter matrix of the feature extraction module, the feature upsampling module and the geometry generation module in the generator are adjusted to realize the training of the generator.
  • the training point cloud is divided into at least one training point cloud block, and for each training point cloud block in at least one training point cloud block, the geometric information of the training point cloud block is input in the generator, and the generator is The geometric information of the training point cloud block is up-sampled to obtain the predicted up-sampling geometric information of the training point cloud block. After the geometric information of the training point cloud block is up-sampled, it becomes a dense training point cloud block.
  • the dense training point cloud block should have the same geometric distribution as the upsampled true value of the training point cloud block, that is, if the generator has high precision, the geometric distribution of the training point cloud block after the generator upsampling should be close to The geometric distribution of the upsampled ground truth for training point cloud blocks.
  • the predicted upsampled geometric information of the training point cloud block is input to the discriminator, so that the discriminator judges whether the data input to the discriminator is the upsampled true value of the training point cloud block, and outputs the first discrimination result.
  • the first discrimination result is the first value, such as 0, it means that the discriminator judges that the data input to the discriminator is an upsampled training point cloud block, indicating that the generator has not been trained, and the parameter matrix in the generator is reversed. Adjustment.
  • the first discriminant result is the second value, such as 1, it means that the discriminator judges that the data input to the discriminator is the upsampled ground truth of the training point cloud block.
  • the training of the generator is completed, and then the training The completed generator upsamples the geometric information of the point cloud.
  • the above S406-A2 includes:
  • the embodiment of the present application does not limit the specific type of the loss function used when determining the first loss according to the first discrimination result.
  • a least squares loss function is used to determine the first loss of the generator.
  • the first loss of the generator is determined according to the following formula (1):
  • L gen (P up ) is the first loss
  • P up is the upsampling geometric information of the training point cloud block
  • D(P up ) is the input of the upsampling geometric information of the training point cloud block to the discriminator
  • the output of the discriminator The first judgment result of .
  • the ways of determining the parameter matrix of the feature extraction module, feature upsampling module and geometry generation module in the generator include but are not limited to the following:
  • Method 1 Determine the parameter matrix of the feature extraction module, feature upsampling module, and geometry generation module in the generator based on the first loss. For example, when the first loss is greater than a certain preset value, it means that the accuracy of the generator has not reached Preset requirements, inversely adjust the parameter matrix of the feature extraction module, feature upsampling module and geometry generation module in the generator. If the first loss is less than a certain preset value, it means that the accuracy of the generator meets the preset requirements, and the parameter matrix of the feature extraction module, feature upsampling module and geometry generation module in the generator is fixed at this time.
  • Step A Determine at least one second loss of the generator
  • Step A2. Determine the target loss of the generator according to the first loss of the generator and at least one second loss of the generator;
  • Step A3 according to the target loss of the generator, determine the parameter matrix of the feature extraction module, feature upsampling module and geometry generation module in the generator.
  • At least one second loss of the generator is determined, and according to the first loss and at least one second loss of the generator, the feature extraction module, feature The parameter matrices of the upsampling module and the geometry generation module are tuned to improve the training accuracy of the generator.
  • the embodiment of the present application does not limit the manner of determining at least one second loss of the generator in the above step A1, which is specifically determined according to actual needs.
  • the above step A1 includes: according to the upsampled geometric information of the training point cloud block and the upsampled true value of the geometric information of the training point cloud block, using the ground motion distance method to determine a second loss of the generator.
  • the reconstruction loss function Using the ground motion distance method to determine a second loss of the generator is also called the reconstruction loss function.
  • the purpose is to make the upsampled training point cloud block and the upsampled true value of the training point cloud block have a consistent geometric distribution.
  • a second loss of the generator is determined:
  • L rec is the second loss
  • EMD represents the ground motion distance method
  • P up is the geometric information of the upsampled training point cloud block
  • P T is the upsampled true value of the geometric information of the training point cloud block
  • ⁇ :P up ⁇ P T is a bijection composed of two equal-sized subsets P up and P T
  • p i is the i-th point in P up
  • ⁇ (p i ) means that in P T according to the bijective relationship ⁇ Find the corresponding point of pi .
  • the above step A1 includes: determining at least one second loss of the generator according to a uniform loss function.
  • a second loss of the generator is determined:
  • L uni is the second loss
  • S i refers to the local surface i obtained by the radius sphere (radius r q ) method
  • T is the number of seed points obtained
  • d i,j represents the distance between the jth point and its nearest neighbor in the i-th local surface distance
  • d i,j represents the distance between the jth point and its nearest neighbor in the i-th local surface distance
  • step A1 includes:
  • Step A11 downsampling the upsampled geometric information of the training point cloud block to obtain a downsampled training point cloud block with the same number of points as the training point cloud block.
  • FPS Farthest point sampling
  • Step A12 According to the geometric information of the downsampled training point cloud block and the geometric information of the training point cloud block, a second loss of the generator is determined by using the ground motion distance method.
  • a second loss of the generator is determined:
  • L id is the second loss of the generator
  • P ori is the low-resolution training point cloud block
  • P low is the training point cloud block after downsampling
  • ⁇ :P low ⁇ P ori means that it is composed of P low and P ori In the bijection formed, there is one and only one way to move P low and P ori to the minimum distance between the point sets of each other, is the kth point in P low , for Corresponding point in P ori .
  • the target loss of the generator is determined according to the first loss of the generator and at least one second loss of the generator, for example, the first loss of the generator and at least one second loss of the generator A weighted average of the second loss, which determines the target loss for the generator.
  • the target loss of the generator is determined according to the following formula (5):
  • L G w gen L gen (P up )+w rec L rec +w uni L uni +w id L id (5)
  • L G is the target loss of the generator
  • L gen (P up ) is the first loss of the generator
  • L rec , L uni , L id are the second losses of the generator respectively
  • w gen is the first loss of the generator Weights
  • w rec , w uni , and w id are weights corresponding to the second losses, respectively.
  • w gen 1.
  • w rec 100.
  • the embodiment of the present application uses the training point cloud to train the generator to obtain the trained generator, so that in practical applications, the trained generator can be used to upsample the geometric information of the point cloud to obtain a high-precision point cloud. Further, the embodiment of the present application divides the training point cloud into training point cloud blocks, uses the training point cloud blocks to train the generator, and uses the discriminator to supervise the training process of the generator, thereby improving the training accuracy and accuracy of the generator. reliability.
  • the training process of the generator is introduced above in combination with the network structure of the generator, and the discriminator involved in the above S406-A1 is introduced below.
  • the above discriminator is a pre-trained discriminator.
  • the discriminator is not pre-trained, that is, the embodiment of the present application also involves a training process of the discriminator.
  • the discriminator before using the geometric information of the training point cloud block to train the generator, the discriminator is trained once, and then S406-A1 is executed using the trained discriminator.
  • the discriminator and the generator are alternately trained, that is, in the training process, the geometric information of the training point cloud block is used to train the discriminator first, and the discriminator After the training, the generator is trained using the geometric information of the training point cloud block, and the training process of the discriminator and the generator is carried out alternately until the training of the generator and the discriminator is completed.
  • the training process of the discriminator specifically includes the following steps:
  • Step 21 Input the predicted upsampled geometric information of the training point cloud block generated by the generator into the discriminator, and obtain the second discrimination result output by the discriminator; input the upsampled true value of the geometric information of the training point cloud block into the discriminator, and obtain the third discrimination result output by the discriminator;
  • Step 22 Determine the loss of the discriminator according to the second discrimination result and the third discrimination result
  • Step 23 Adjust the parameters in the discriminator according to the loss of the discriminator.
  • the embodiment of the present application does not limit the type of loss function used to determine the loss of the discriminator according to the second discrimination result and the third discrimination result in step 21 .
  • step 21 includes: according to the second discrimination result and the third discrimination result, using a least squares loss function to determine the loss of the discriminator.
  • the loss of the discriminator is determined:
  • L dis (P up , P T ) represents the loss of the discriminator
  • P T is the upsampled true value of the training point cloud block
  • P up refers to the point cloud obtained by the upsampling of the generator, that is, the training after upsampling Point cloud blocks.
  • the discriminator is trained according to the difference between the discriminator's discriminant result of the predicted upsampled geometric information of the training point cloud block and the upsampled true value of the geometric information of the training point cloud block, thereby improving the discriminator. training accuracy.
  • the process of the discriminator obtaining the discriminant result based on the geometric information of the point cloud block will be described in detail, that is, the discriminator generates the first discriminant result, the second discriminant result and the third discriminant result.
  • the process is introduced.
  • Figure 11 is a schematic diagram of a network structure of the discriminator.
  • the discriminator includes a global discriminant module, a boundary discriminant module and a fully connected module, wherein the global discriminant module is used to extract the global feature information of the point cloud, and the The module is used to extract the boundary feature information of the point cloud, and the fully connected module is used to process the global feature information and boundary feature information of the point cloud to obtain the discrimination result.
  • Fig. 12 is a schematic flow chart of the model training method provided by an embodiment of the present application. As shown in Fig. 11 and Fig. 12, the process for the discriminator to obtain the discriminant result includes:
  • a high-pass graph filter is used to extract the geometric information of the boundary points of the target point cloud block.
  • S602. Input the geometric information of the boundary points of the target point cloud block into the boundary discrimination module to perform boundary feature extraction, and obtain boundary feature information of the target point cloud block.
  • the global feature information and boundary feature information of the target point cloud block are concatenated; the concatenated global feature information and boundary feature information are input into the fully connected module to obtain the target discrimination result of the discriminator.
  • the discriminator in the embodiment of the present application can be understood as a double-headed discriminator, which can realize the judgment on the two dimensions of the whole world and the boundary, thereby improving the accuracy of the judgment.
  • each input target point cloud First extract the boundary points of the target point cloud block, for example, extract the R boundary points of the target point cloud block through a high-pass image filter R ⁇ N', and then explicitly send the complete target point cloud block and P b into the double-headed discriminator shown in Figure 12, and obtain the global feature information of the target point cloud block output by the global discriminant module, and the boundary
  • the boundary discrimination feature of the target point cloud block output by the discriminant module, the global feature information and boundary feature information of the target point cloud block are input into the full connection module, and the target discrimination result of the discriminator is obtained.
  • the discriminator obtains the first discriminant result, the second discriminant result, and the third discriminant result in the same process.
  • the target point cloud block is a training point cloud block sampled by the generator, and the judger is trained by the training point cloud block, then the above target discrimination result is the first discrimination result. If the target point cloud block is a training point cloud block sampled by the generator, and the judger has not been trained by the training point cloud block, then the above target discrimination result is the second discrimination result. If the target point cloud block is the upsampled true value of the training point cloud block, the target discrimination result is the third discrimination result.
  • the training process of the discriminator is to input the training point cloud block into the generator that has not been trained by the training point cloud block, and the generator generates the above
  • the up-sampled training point cloud block 1 generated by the generator is input into a judger that has not been trained by the training point cloud block, and the judger outputs a second judgment result.
  • input the upsampling true value of the training point cloud block into the discriminator, and the discriminator outputs the third discriminant result, and update the parameter matrix of the discriminator according to the second discriminant result and the third discriminant result to realize the discriminant A training session of the machine.
  • the upsampled training point cloud block 1 generated by the generator is input into the above-mentioned discriminator trained by the training point cloud block, and the discriminator outputs the first discrimination result.
  • the following are the global discrimination module and the boundary discrimination module in the discriminator respectively.
  • the global discriminant module sequentially includes along the network depth direction: a first number of multi-layer perceptrons, a first maximum pooling layer, a second autocorrelation attention network, a second number of A multilayer perceptron and a second max pooling layer.
  • the above S603 includes:
  • the first global feature information and the second global feature information are concatenated; the concatenated first global feature information and the second global feature information are input into the second autocorrelation attention network Perform feature interaction to obtain the third global feature information of the target point cloud block.
  • the geometric information of the target point cloud is input into the first number of multilayer perception machines (Multilayer perception, MLP for short) for feature extraction, and the first global feature information of the target point cloud is obtained; then, the first global The feature information is input into the first maximum pooling layer for dimension reduction processing, and the second global feature information of the target point cloud block is obtained through the maximum pooling operation; then, the first global feature information and the second global feature information are input into the second autocorrelation attention
  • the force (Self-attetion) network performs feature interaction, improves the feature interaction between each point, and obtains the third global feature information of the target point cloud block; then, the third global feature information is input into the second number of multi-layer perceptrons (MLP) and further feature extraction to obtain the fourth global feature information of the target point cloud block; finally, input the fourth global feature information into the second maximum pooling layer for dimensionality reduction processing to obtain the global feature information of the target point cloud block.
  • MLP multilayer perception
  • MLP multi-layer perceptrons
  • the first quantity is equal to the second quantity.
  • both the first quantity and the second quantity are equal to 2.
  • the first number of multilayer perceptrons includes a first layer of multilayer perceptrons and a second layer of multilayer perceptrons
  • the second number of multilayer perceptrons includes a third layer of multilayer perceptrons and a fourth layer of multilayer perceptrons.
  • Layer perceptron, the feature dimension of the first layer of multi-layer perceptron, the second layer of multi-layer perceptron, the third layer of multi-layer perceptron and the fourth layer of multi-layer perceptron gradually increases.
  • the feature dimension of the first layer of multi-layer perceptron is 32
  • the feature dimension of the second layer of multi-layer perceptron is 64
  • the feature dimension of the third layer of multi-layer perceptron is 128,
  • the fourth layer of multi-layer perceptron The feature dimension of is 256.
  • the boundary discrimination module sequentially includes along the network depth direction: a third number of multi-layer perceptrons, a third maximum pooling layer, a third autocorrelation attention network, a third Four multi-layer perceptrons and the fourth maximum pooling layer, at this time, the above S602 includes:
  • S602-A3 includes: concatenating the first boundary feature information and the second boundary feature information; inputting the concatenated first boundary feature information and the second boundary feature information into a third self- The relevant attention network performs feature interaction to obtain the third boundary feature information of the target point cloud block.
  • the third maximum pooling layer performs dimension reduction processing, and obtains the second boundary feature information of the target point cloud block through the maximum pooling operation; then, the first boundary feature information and the second boundary feature information are input into the third autocorrelation attention (Self -attetion) network for feature interaction, enhance the feature interaction between each point, and obtain the third boundary feature information of the target point cloud block; then, input the third boundary feature information into the fourth number of multi-layer perceptrons (MLP) Further feature extraction is performed to obtain the fourth boundary feature information of the target point cloud block; finally, the fourth boundary feature information is input into the fourth maximum pooling layer for dimensionality reduction processing to obtain the boundary feature information of the target point cloud block.
  • MLP multi-layer perceptrons
  • the third quantity is equal to the fourth quantity.
  • both the third quantity and the fourth quantity are equal to 2.
  • the third number of multi-layer perceptrons includes a fifth-layer multi-layer perceptron and a sixth-layer multi-layer perceptron
  • the fourth number of multi-layer perceptrons includes a seventh-layer multi-layer perceptron and an eighth-layer multi-layer perceptron.
  • Layer perceptron, the feature dimension of the fifth layer multilayer perceptron, sixth layer multilayer perceptron, seventh layer multilayer perceptron and eighth layer multilayer perceptron gradually increases.
  • the feature dimension of the eighth-layer multi-layer perceptron is greater than or equal to the feature dimension of the seventh-layer multi-layer perceptron, and smaller than or equal to the feature dimension of the fourth-layer multi-layer perceptron.
  • the feature dimension of the eighth-layer multilayer perceptron is greater than or equal to 128 and less than or equal to 256.
  • the feature dimension of the fifth-layer multi-layer perceptron is 32
  • the feature dimension of the sixth-layer multi-layer perceptron is 64
  • the feature dimension of the seventh-layer multi-layer perceptron is 128,
  • the eighth-layer multi-layer perceptron The feature dimension of is 192.
  • the global feature information of the target point cloud block output by the global discrimination module and the boundary feature information output by the boundary discrimination module are cascaded, and the connected global feature information and boundary feature information are input into the fully connected layer
  • the module obtains the confidence value of the discriminator through three fully connected layers (FC), that is, the discriminant result of the discriminator. If the input of the discriminator is the upsampled point cloud output by the generator, the confidence value is close to 0. If the discriminator The input is the upsampled true value of the point cloud, and the confidence value is close to 1.
  • the training of the generator can be supervised according to the discrimination results of the discriminator, thereby improving the training accuracy of the generator, so that the distribution of the upsampled point cloud of the trained generator is close to the true value of the upsampled point cloud , to ensure the accuracy of the upsampled point cloud.
  • the discriminator includes a global discrimination module and a boundary discrimination module, and discriminates the global information and boundary information of the point cloud, thereby improving the discrimination.
  • the discriminative accuracy of the discriminator improves the training accuracy of the generator when the discriminator is used to assist the training of the generator.
  • the training process of the generator is introduced above, and the upsampling process of the geometric information of the point cloud using the trained generator is introduced below.
  • the above trained generator can realize the upsampling of the geometric information of the point cloud.
  • Fig. 14 is a schematic flow chart of the point cloud upsampling method provided by the embodiment of the present application. As shown in Fig. 14, the point cloud upsampling process includes:
  • the point cloud to be upsampled may be collected in real time by a point cloud collection device.
  • the point cloud to be upsampled may be obtained from other storage devices.
  • the point cloud to be upsampled is decoded by the decoding device from the code stream obtained by the editing device.
  • the embodiment of the present application does not limit the specific process of obtaining the point cloud to be processed.
  • the methods of dividing the point cloud to be upsampled into at least one point cloud block in S702 include but are not limited to the following methods:
  • Method 1 Divide the point cloud to be upsampled into at least one point cloud block of equal size according to the geometric information of the point cloud to be upsampled. That is to say, the geometric scale of each point cloud block is the same.
  • Method 2 Divide the point cloud to be upsampled into at least one point cloud block according to the geometric information of the point cloud to be upsampled, and each point cloud block includes the same number of points.
  • Method 3 Obtain at least one seed point from the point cloud to be upsampled according to the geometric information of the point cloud to be upsampled, for example, use Monte Carlo random sampling method to randomly sample a specified number of seed points from the point cloud to be upsampled . For each seed point, determine the neighboring points of the seed point, divide the seed point and the neighboring points of the seed point into a point cloud block, and then obtain at least one point cloud block. In the third method, the obtained point cloud blocks are also called point cloud patches (Patch), and the number of points included in each point cloud block in the obtained point cloud blocks is the same.
  • Monte Carlo random sampling method to randomly sample a specified number of seed points from the point cloud to be upsampled . For each seed point, determine the neighboring points of the seed point, divide the seed point and the neighboring points of the seed point into a point cloud block, and then obtain at least one point cloud block.
  • the obtained point cloud blocks are also called point cloud patches (Patch), and the number of points included in each point
  • Fig. 15 is a schematic diagram of a network structure of the generator involved in the embodiment of the present application.
  • the generator includes: a feature extraction module, a feature upsampling module and a geometry generation module, wherein the feature extraction module is used to extract points The first feature information of the cloud block, the feature sampling module is used to upsample the first feature information of the point cloud block into the second feature information, and the geometry generation module is used to map the second feature information of the point cloud block into the geometric space, In order to obtain the upsampled geometric information of the point cloud block.
  • the network structure of the feature extraction module, feature upsampling module and geometry generation module in the generator is introduced below.
  • the feature extraction module includes densely connected M feature extraction blocks
  • the i+1th feature extraction block is used to output the i+1th third feature information according to the input i-th fourth feature information, and the ith
  • the fourth feature information is determined according to the i-th third feature information output by the i-th feature extraction block, and the first feature information of the point cloud block is based on the output of the M-th feature extraction block in the M feature extraction blocks
  • i is a positive integer smaller than M.
  • the i-th fourth feature information is the third feature information extracted by each feature extraction block before the i-th feature extraction block in the M feature extraction blocks, and The feature information obtained by cascading the third feature information extracted by the i feature extraction blocks. If i is equal to 1, the i-th fourth feature information is the first third feature information output by the first feature extraction block in the M feature extraction blocks.
  • the feature extraction block includes: a first feature extraction unit and S second feature extraction units connected in series, where S is a positive integer;
  • the first extraction unit in the i+1th feature extraction block is used to search for K neighboring points of the current point for the current point in the point cloud block, and based on the i-th feature extraction unit of the point cloud block Four feature information, the fourth feature information of the current point is subtracted from the fourth feature information of the adjacent point to obtain K residual feature information, and the K residual feature information is graded with the fourth feature information of the current point According to the i-th cascade feature information of the current point, get the i-th cascade feature information of the point cloud block, and concatenate the i-th cascade feature information of the point cloud block The feature information is input to the first second feature extraction unit in the S second feature extraction units;
  • the first second feature extraction unit is used to output the first fifth feature information to the second second feature extraction unit according to the ith cascaded feature information of the point cloud block, wherein the i+1th of the point cloud block
  • the third feature information is the fifth feature information output by the last second feature extraction unit among the S second feature extraction units.
  • the second feature extraction unit includes P residual blocks, where P is a positive integer
  • the j+1th residual block in the sth second feature extraction unit is used according to the output of the jth residual block in the sth second feature extraction unit
  • the j-th first residual information and the fifth feature information input to the s-th second feature extraction unit output the j+1-th first residual information, where j is a positive integer less than P, and s is less than or A positive integer equal to S.
  • the fifth feature information output by the sth second feature extraction unit is based on the first residual information information output by at least one residual block in the sth second feature extraction unit, and input to the sth second feature extraction unit.
  • the fifth feature information is determined.
  • the fifth feature information output by the sth second feature extraction unit is based on the first residual information output by the last residual block in the sth second feature extraction unit, and The feature information after concatenation of the first residual information output by at least one residual block in the P-1 residual blocks and the fifth feature information input to the s-th second feature extraction unit are determined, wherein, P The -1 residual block is a residual block except the last residual block among the P residual blocks of the s-th second feature extraction unit.
  • the fifth feature information output by the sth second feature extraction unit is based on the first residual information output by the last residual block in the sth second feature extraction unit, and The first residual information output by at least one residual block in the P-1 residual blocks is concatenated and then determined by adding the fifth feature information input to the s-th second feature extraction unit.
  • the second feature extraction unit further includes a gating unit
  • the gating unit in the sth second feature extraction unit is used for the first residual information output by the last residual block in the sth second feature extraction unit, and P-1 De-redundancy is performed on the feature information after concatenation of the first residual information output by at least one residual block in the residual block, and the de-redundant feature information is output; the fifth feature information output by the sth second feature extraction unit It is determined after adding the feature information after de-redundancy to the fifth feature information input to the sth second feature extraction unit.
  • the network structure of the feature extraction module in the generator is introduced above with reference to FIGS. 6A to 6F
  • the network structure of the feature upsampling module in the generator is introduced below with reference to FIGS. 7A to 7D .
  • the feature upsampling module includes: a feature upsampling submodule and a feature extraction submodule;
  • the feature upsampling submodule is used to copy r copies of the first feature information of the point cloud block according to the preset upsampling rate r, and add an n-dimensional vector to the feature dimension of the copied first feature information, Obtain the upsampling feature information of the point cloud block, and input the upsampling feature information of the point cloud block into the feature extraction submodule, wherein the values of the n-dimensional vectors corresponding to different first feature information are different;
  • the feature extraction sub-module is used to output the second feature information of the point cloud block according to the up-sampled feature information of the point cloud block.
  • the feature extraction submodule includes Q third feature extraction units, where Q is a positive integer
  • the k+1th third feature extraction unit is used to extract the point cloud block according to the kth third feature extraction unit.
  • k enhanced upsampling feature information the k+1th enhanced upsampling feature information of the output point cloud block, k is a positive integer less than Q;
  • the second feature information of the point cloud block is the Qth enhanced upsampling feature information of the point cloud block extracted by the last third feature extraction unit among the Q third feature extraction units.
  • the third feature extraction unit is the HRA in Figure 7D, the third feature extraction unit includes L residual blocks, L is a positive integer, for example, the third feature extraction unit includes 4 residual block RB;
  • the l+1th residual block in the k+1th third feature extraction unit is used according to the lth residual in the k+1th third feature extraction unit
  • the lth second residual information output by the block and the kth enhanced upsampling feature information input to the k+1th third feature extraction unit, and the l+1th second residual information is output, where l is less than L Positive integer; optionally, after adding the l-th second residual information output by the l-th residual block and the k-th enhanced upsampling feature information, input the l+1th residual block.
  • the k+1th enhanced upsampling feature information of the point cloud block is determined according to the second residual information output by at least one residual block in the k+1th third feature extraction unit, and the kth enhanced upsampling feature information of.
  • the k+1th enhanced upsampled feature information of the above point cloud block is based on the second residual information output by the last residual block in the L residual blocks, and the L-1 residual
  • the second residual information output by at least one residual block in the difference block is determined by concatenating the feature information and the kth enhanced upsampling feature information, wherein, the L-1 residual block is the k+1th A residual block except the last residual block among the L residual blocks of the three-feature extraction unit.
  • the k+1th enhanced upsampled feature information of the above point cloud block is based on the second residual information output by the last residual block in the L residual blocks, and the L-1 residual It is determined by adding the concatenated feature information of the second residual information output by at least one residual block in the difference block and the kth enhanced upsampling feature information.
  • the third feature extraction unit further includes a gating unit
  • the gating unit is used to output the second residual information of the last residual block in the k+1th third feature extraction unit, and L -
  • the second feature information output by at least one residual block in the 1 residual block is de-redundant after concatenating the feature information, and outputting the feature information after de-redundancy;
  • the k+1th enhanced upsampling feature information of the point cloud block is determined after adding the deredundant feature information to the kth enhanced upsampling feature information.
  • the network structure of the third feature extraction unit is the same as that of the above-mentioned second feature extraction unit.
  • the feature upsampling module further includes a first autocorrelation attention network
  • the first autocorrelation attention network is used to perform feature interaction on the upsampling feature information of the point cloud block output by the feature upsampling submodule, and output the upsampling feature information of the point cloud block after feature interaction to the feature extraction submodule;
  • the feature extraction submodule is configured to output second feature information of the point cloud block according to the upsampled feature information of the point cloud block after feature interaction.
  • the feature dimension of the upsampled feature information of the point cloud block after the feature interaction is lower than the feature dimension of the upsampled feature information of the point cloud block.
  • the network structure of the feature extraction module in the generator is introduced above with reference to FIG. 7A to FIG. 7D , and the network structure of the geometry generation module in the generator is introduced below in conjunction with FIG. 8 and FIG. 15 .
  • the geometry generation module includes a plurality of fully connected layers
  • the multiple fully connected layers are used to output upsampled geometric information of the point cloud block according to the second feature information of the point cloud block.
  • the geometry generation module includes: a geometry reconstruction unit, a filtering unit and a downsampling unit;
  • the geometric reconstruction unit is used to geometrically reconstruct the second feature information of the point cloud block, and output the initial upsampling geometric information of the point cloud block to the filtering unit;
  • the filter unit is used to denoise the initial upsampling geometric information of the point cloud block, and output the initial upsampling geometric information of the point cloud block to filter out the noise to the downsampling unit;
  • the down-sampling unit is used for down-sampling the initial up-sampling geometric information of the point cloud block to filter noise to a target up-sampling rate, and output the up-sampling geometric information of the point cloud block.
  • the target upsampling rate is less than or equal to the upsampling rate of the feature upsampling module.
  • the scheme proposed in the embodiment of the present application is implemented on the test platform, and the chamfering distance (CD), the Hausdorff distance (HD), and the point-to-surface distance (P2FD) are used respectively ) to measure the similarity between the upsampled point cloud and the upsampled ground truth of the point cloud.
  • the upsampling rate r is set to 4.
  • the technical solution of the embodiment of the application was tested with the optimization-based method EAR, the most advanced point cloud upsampling network PU-Net, MPU, and PU-GAN, and the results on the test data set are shown in Table 1:
  • the difference between the point cloud generated by the method proposed in this application and the upsampled true value of the point cloud is the smallest, for example, Chamfer distance (CD for short), Hausdorff distance (Hausdorff distance , referred to as HD), and the point-to-face distance (Point to face distance, referred to as P2FD) are 0.258, 3.571 and 2.392 respectively. Therefore, it is shown that the point cloud upsampling method proposed in this application can realize effective upsampling of the point cloud.
  • the point cloud upsampling method of the embodiment of the present application obtains the geometric information of the point cloud to be upsampled, divides the point cloud to be upsampled into at least one point cloud block according to the geometric information of the point cloud to be upsampled, and divides the point cloud block into The geometric information of the input generator is upsampled, and the upsampled geometric information of the point cloud block is obtained.
  • the generator includes: a feature extraction module, a feature upsampling module and a geometry generation module.
  • the feature extraction module is used to extract the first point cloud block.
  • the feature sampling module is used to upsample the first feature information of the point cloud block into the second feature information
  • the geometry generation module is used to map the second feature information of the point cloud block into the geometric space to obtain the point cloud
  • the generator in the embodiment of the present application is a generator based on deep learning, through which more characteristic information of the point cloud can be learned, and then when the generator is used for upsampling of the point cloud, a high-precision point cloud can be generated, Moreover, the feature of the high-precision point cloud is close to the true value of the upsampling of the point cloud, thereby improving the accuracy of the upsampling of the point cloud.
  • the point cloud upsampling method provided in the embodiment of the present application can also be applied to a point cloud encoding and decoding framework, for example, it can be applied to a point cloud decoding end.
  • Fig. 16 is a schematic flow chart of the point cloud decoding method provided by the embodiment of the present application. As shown in Fig. 16, the point cloud decoding method includes:
  • the point cloud code stream includes attribute code stream and geometry code stream. By decoding the geometry code stream, the geometric information of the point cloud can be obtained, and by decoding the attribute code stream, the attribute information of the point cloud can be obtained.
  • the generator includes: a feature extraction module, a feature upsampling module and a geometry generation module, the feature extraction module is used to extract the first feature information of the point cloud block, and the feature sampling module is used to extract the point cloud block The first feature information of the cloud block is upsampled to the second feature information, and the geometry generation module is used to map the second feature information of the point cloud block into a geometric space, so as to obtain the upsampled geometric information of the point cloud block.
  • the network structure of the feature extraction module, feature upsampling module and geometry generation module in the generator is introduced below.
  • the network structure of the feature extraction module is introduced with reference to FIG. 6A to FIG. 6F .
  • the feature extraction module includes densely connected M feature extraction blocks
  • the i+1th feature extraction block is used to output the i+1th third feature information according to the input i-th fourth feature information, and the ith
  • the fourth feature information is determined according to the i-th third feature information output by the i-th feature extraction block, and the first feature information of the point cloud block is based on the output of the M-th feature extraction block in the M feature extraction blocks
  • i is a positive integer smaller than M.
  • the i-th fourth feature information is the third feature information extracted by each feature extraction block before the i-th feature extraction block in the M feature extraction blocks, and The feature information obtained by cascading the third feature information extracted by the i feature extraction blocks. If i is equal to 1, the i-th fourth feature information is the first third feature information output by the first feature extraction block in the M feature extraction blocks.
  • the feature extraction block includes: a first feature extraction unit and S second feature extraction units connected in series, where S is a positive integer;
  • the first extraction unit in the i+1th feature extraction block is used to search for K neighboring points of the current point for the current point in the point cloud block, and based on the i-th feature extraction unit of the point cloud block Four feature information, the fourth feature information of the current point is subtracted from the fourth feature information of the adjacent point to obtain K residual feature information, and the K residual feature information is graded with the fourth feature information of the current point According to the i-th cascade feature information of the current point, get the i-th cascade feature information of the point cloud block, and concatenate the i-th cascade feature information of the point cloud block The feature information is input to the first second feature extraction unit in the S second feature extraction units;
  • the first second feature extraction unit is used to output the first fifth feature information to the second second feature extraction unit according to the ith cascaded feature information of the point cloud block, wherein the i+1th of the point cloud block
  • the third feature information is the fifth feature information output by the last second feature extraction unit among the S second feature extraction units.
  • the second feature extraction unit includes P residual blocks, where P is a positive integer
  • the j+1th residual block in the sth second feature extraction unit is used according to the output of the jth residual block in the sth second feature extraction unit
  • the j-th first residual information and the fifth feature information input to the s-th second feature extraction unit output the j+1-th first residual information, where j is a positive integer less than P, and s is less than or A positive integer equal to S.
  • the fifth feature information output by the sth second feature extraction unit is based on the first residual information information output by at least one residual block in the sth second feature extraction unit, and input to the sth second feature extraction unit.
  • the fifth feature information is determined.
  • the fifth feature information output by the sth second feature extraction unit is based on the first residual information output by the last residual block in the sth second feature extraction unit, and The feature information after concatenation of the first residual information output by at least one residual block in the P-1 residual blocks and the fifth feature information input to the s-th second feature extraction unit are determined, wherein, P The -1 residual block is a residual block except the last residual block among the P residual blocks of the s-th second feature extraction unit.
  • the fifth feature information output by the sth second feature extraction unit is based on the first residual information output by the last residual block in the sth second feature extraction unit, and The first residual information output by at least one residual block in the P-1 residual blocks is concatenated and then determined by adding the fifth feature information input to the s-th second feature extraction unit.
  • the second feature extraction unit further includes a gating unit
  • the gating unit in the sth second feature extraction unit is used for the first residual information output by the last residual block in the sth second feature extraction unit, and P-1 De-redundancy is performed on the feature information after concatenation of the first residual information output by at least one residual block in the residual block, and the de-redundant feature information is output; the fifth feature information output by the sth second feature extraction unit It is determined after adding the feature information after de-redundancy to the fifth feature information input to the sth second feature extraction unit.
  • the network structure of the feature extraction module in the generator is introduced above with reference to FIGS. 6A to 6F
  • the network structure of the feature upsampling module in the generator is introduced below with reference to FIGS. 7A to 7D .
  • the feature upsampling module includes: a feature upsampling submodule and a feature extraction submodule;
  • the feature upsampling submodule is used to copy r copies of the first feature information of the point cloud block according to the preset upsampling rate r, and add an n-dimensional vector to the feature dimension of the copied first feature information, Obtain the upsampling feature information of the point cloud block, and input the upsampling feature information of the point cloud block into the feature extraction submodule, wherein the values of the n-dimensional vectors corresponding to different first feature information are different;
  • the feature extraction sub-module is used to output the second feature information of the point cloud block according to the up-sampled feature information of the point cloud block.
  • the feature extraction submodule includes Q third feature extraction units, where Q is a positive integer
  • the k+1th third feature extraction unit is used to extract the point cloud block according to the kth third feature extraction unit.
  • k enhanced upsampling feature information the k+1th enhanced upsampling feature information of the output point cloud block, k is a positive integer less than Q;
  • the second feature information of the point cloud block is the Qth enhanced upsampling feature information of the point cloud block extracted by the last third feature extraction unit among the Q third feature extraction units.
  • the third feature extraction unit is the HRA in Figure 7D, the third feature extraction unit includes L residual blocks, L is a positive integer, for example, the third feature extraction unit includes 4 residual block RB;
  • the l+1th residual block in the k+1th third feature extraction unit is used according to the lth residual in the k+1th third feature extraction unit
  • the lth second residual information output by the block and the kth enhanced upsampling feature information input to the k+1th third feature extraction unit, and the l+1th second residual information is output, where l is less than L Positive integer; optionally, after adding the l-th second residual information output by the l-th residual block and the k-th enhanced upsampling feature information, input the l+1th residual block.
  • the k+1th enhanced upsampling feature information of the point cloud block is determined according to the second residual information output by at least one residual block in the k+1th third feature extraction unit, and the kth enhanced upsampling feature information of.
  • the k+1th enhanced upsampled feature information of the above point cloud block is based on the second residual information output by the last residual block in the L residual blocks, and the L-1 residual
  • the second residual information output by at least one residual block in the difference block is determined by concatenating the feature information and the kth enhanced upsampling feature information, wherein, the L-1 residual block is the k+1th A residual block except the last residual block among the L residual blocks of the three-feature extraction unit.
  • the k+1th enhanced upsampled feature information of the above point cloud block is based on the second residual information output by the last residual block in the L residual blocks, and the L-1 residual It is determined by adding the concatenated feature information of the second residual information output by at least one residual block in the difference block and the kth enhanced upsampling feature information.
  • the third feature extraction unit further includes a gating unit
  • the gating unit is used to output the second residual information of the last residual block in the k+1th third feature extraction unit, and L -
  • the second feature information output by at least one residual block in the 1 residual block is de-redundant after concatenating the feature information, and outputting the feature information after de-redundancy;
  • the k+1th enhanced upsampling feature information of the point cloud block is determined after adding the deredundant feature information to the kth enhanced upsampling feature information.
  • the network structure of the third feature extraction unit is the same as that of the above-mentioned second feature extraction unit.
  • the feature upsampling module further includes a first autocorrelation attention network
  • the first autocorrelation attention network is used to perform feature interaction on the upsampling feature information of the point cloud block output by the feature upsampling submodule, and output the upsampling feature information of the point cloud block after feature interaction to the feature extraction submodule;
  • the feature extraction submodule is configured to output second feature information of the point cloud block according to the upsampled feature information of the point cloud block after feature interaction.
  • the feature dimension of the upsampled feature information of the point cloud block after the feature interaction is lower than the feature dimension of the upsampled feature information of the point cloud block.
  • the network structure of the feature extraction module in the generator is introduced above with reference to FIG. 7A to FIG. 7D , and the network structure of the geometry generation module in the generator is introduced below in conjunction with FIG. 8 and FIG. 15 .
  • the geometry generation module includes a plurality of fully connected layers
  • the multiple fully connected layers are used to output upsampled geometric information of the point cloud block according to the second feature information of the point cloud block.
  • the geometry generation module includes: a geometry reconstruction unit, a filtering unit and a downsampling unit;
  • the geometric reconstruction unit is used to geometrically reconstruct the second feature information of the point cloud block, and output the initial upsampling geometric information of the point cloud block to the filtering unit;
  • the filter unit is used to denoise the initial upsampling geometric information of the point cloud block, and output the initial upsampling geometric information of the point cloud block to filter out the noise to the downsampling unit;
  • the down-sampling unit is used for down-sampling the initial up-sampling geometric information of the point cloud block to filter noise to a target up-sampling rate, and output the up-sampling geometric information of the point cloud block.
  • the target upsampling rate is less than or equal to the upsampling rate of the feature upsampling module.
  • the target upsampling rate is a preset value.
  • the above target upsampling rate is parsed from the point cloud code stream.
  • the embodiment of the present application upsamples the geometric information of the point cloud generated by the point cloud decoding end to generate a high-precision reconstruction point cloud, which can meet the application scenarios of high-precision point clouds, and further improves the diversity of point cloud decoding.
  • sequence numbers of the above-mentioned processes do not mean the order of execution, and the order of execution of the processes should be determined by their functions and internal logic, and should not be used in this application.
  • the implementation of the examples constitutes no limitation.
  • the term "and/or" is only an association relationship describing associated objects, indicating that there may be three relationships. Specifically, A and/or B may mean: A exists alone, A and B exist simultaneously, and B exists alone.
  • the character "/" in this article generally indicates that the contextual objects are an "or" relationship.
  • Fig. 17 is a schematic block diagram of a point cloud decoder provided by an embodiment of the present application.
  • this point point cloud decoder 20 can comprise:
  • the decoding unit 21 is used to decode the point cloud code stream to obtain the geometric information of the point cloud;
  • a division unit 22 configured to divide the point cloud into at least one point cloud block according to the geometric information of the point cloud
  • An up-sampling unit 23 configured to input the geometric information of the point cloud block into the generator for up-sampling, to obtain the up-sampled geometric information of the point cloud block;
  • the generator includes: a feature extraction module, a feature upsampling module and a geometry generation module, the feature extraction module is used to extract the first feature information of the point cloud block, and the feature sampling module is used to extract the The first feature information of the point cloud block is up-sampled to the second feature information, and the geometry generation module is used to map the second feature information of the point cloud block into a geometric space, so as to obtain the up-sampling of the point cloud block geometric information.
  • the feature extraction module comprises densely connected M feature extraction blocks
  • the i+1th feature extraction block in the M feature extraction blocks is used to output the i+1th third feature according to the input i-th fourth feature information information, the i-th fourth feature information is determined according to the i-th third feature information output by the i-th feature extraction block, and the first feature information of the point cloud block is determined according to the M feature extraction blocks Determined by the Mth third feature information output by the Mth feature extraction block, the i is a positive integer smaller than M.
  • the i-th fourth feature information is the first feature extracted by each feature extraction block before the i-th feature extraction block among the M feature extraction blocks.
  • the ith fourth feature information is the first third feature information output by the first feature extraction block in the M feature extraction blocks.
  • the feature extraction block includes: a first feature extraction unit and S second feature extraction units connected in series, wherein S is a positive integer;
  • the first extraction unit in the i+1th feature extraction block is used to search for K neighboring points of the current point for the current point in the point cloud block, and based on For the ith fourth feature information of the point cloud block, the fourth feature information of the current point is subtracted from the fourth feature information of the adjacent point to obtain K residual feature information, and the The K residual feature information is concatenated with the fourth feature information of the current point to obtain the i-th concatenated feature information of the current point, and according to the i-th concatenated feature information of the current point, the The i-th concatenated feature information of the point cloud block, and input the i-th concatenated feature information of the point cloud block into the first second feature extraction unit in the S second feature extraction units;
  • the first second feature extraction unit is used to output the first fifth feature information to the second second feature extraction unit according to the i-th cascaded feature information of the point cloud block, wherein the point cloud
  • the i+1th third feature information of the block is the fifth feature information output by the last second feature extraction unit among the S second feature extraction units.
  • the second feature extraction unit includes P residual blocks, where P is a positive integer
  • the j+1th residual block in the sth second feature extraction unit is used according to the jth residual in the sth second feature extraction unit
  • the j-th first residual information output by the block and the fifth feature information input to the s-th second feature extraction unit, and the j+1-th first residual information is output, wherein the j is less than P A positive integer, the s is a positive integer less than or equal to S;
  • the fifth feature information output by the sth second feature extraction unit is based on the first residual information output by at least one residual block in the sth second feature extraction unit, and input to the sth second feature extraction unit determined by the fifth feature information of the second feature extraction unit.
  • the up-sampling unit 23 is further configured to input the j-th first residual information and the j-th residual information output by the j-th residual block in the s-th second feature extraction unit to the s-th After the fifth feature information of the second feature extraction unit is added, it is input to the j+1th residual block in the sth second feature extraction unit.
  • the fifth feature information output by the s th second feature extraction unit is based on the first residual information output by the last residual block in the s th second feature extraction unit, and The feature information after the concatenation of the first residual information output by at least one residual block in the P-1 residual blocks is determined from the fifth feature information input to the s-th second feature extraction unit, wherein , the P-1 residual blocks are residual blocks except the last residual block among the P residual blocks of the s-th second feature extraction unit.
  • the fifth feature information output by the s th second feature extraction unit is based on the first residual information output by the last residual block in the s th second feature extraction unit, and The first residual information output by at least one residual block in the P-1 residual blocks is concatenated and then determined after being added to the fifth feature information input to the s-th second feature extraction unit of.
  • the second feature extraction unit further includes a gating unit
  • the gating unit in the s th second feature extraction unit is used for the first residual information output by the last residual block in the s th second feature extraction unit, De-redundancy is performed on the feature information concatenated with the first residual information output by at least one residual block in the P-1 residual blocks, and the de-redundant feature information is output; the s second The fifth feature information output by the feature extraction unit is determined after adding the de-redundant feature information and the fifth feature information input to the s-th second feature extraction unit.
  • the feature upsampling module includes: a feature upsampling submodule and a feature extraction submodule;
  • the feature upsampling submodule is used to copy r copies of the first feature information of the point cloud block according to the preset upsampling rate r, and add an n dimension to the feature dimension of the copied first feature information vector, obtain the upsampling feature information of the point cloud block, and input the upsampling feature information of the point cloud block into the feature extraction submodule, wherein the values of n-dimensional vectors corresponding to different first feature information are different;
  • the feature extraction submodule is configured to output second feature information of the point cloud block according to the upsampled feature information of the point cloud block.
  • the feature extraction submodule includes Q third feature extraction units, where Q is a positive integer
  • the k+1th third feature extraction unit is used to extract all The kth enhanced upsampling feature information of the point cloud block, output the k+1th enhanced upsampling feature information of the point cloud block, and the k is a positive integer less than Q;
  • the second feature information of the point cloud block is the Qth enhanced upsampling feature information of the point cloud block extracted by the last third feature extraction unit among the Q third feature extraction units.
  • the third feature extraction unit includes L residual blocks, and the L is a positive integer
  • the l+1th residual block is used according to the k+1th third feature extraction unit
  • the lth second residual information output by the lth residual block and the kth enhanced upsampling feature information input to the k+1th third feature extraction unit, output the l+1th second residual difference information, the l is a positive integer less than L;
  • the k+1th enhanced upsampling feature information of the point cloud block is the second residual information output from at least one residual block in the k+1th third feature extraction unit, and the kth Enhanced upsampling feature information determined.
  • the upsampling unit 23 is further configured to: after adding the lth second residual information output by the lth residual block and the kth enhanced upsampling feature information, input The l+1th residual block.
  • the k+1th enhanced upsampling feature information of the point cloud block is based on the second residual information output by the last residual block in the L residual blocks, and the L-1 residual
  • the second residual information output by at least one residual block in the difference block is determined by concatenating the feature information and the kth enhanced upsampling feature information, wherein the L-1 residual blocks are the A residual block except the last residual block among the L residual blocks of the k+1 third feature extraction unit.
  • the k+1th enhanced upsampling feature information of the point cloud block is based on the second residual information output by the last residual block in the L residual blocks, and the L-1 residual
  • the second residual information output by at least one residual block in the difference block is determined by adding the concatenated feature information to the kth enhanced upsampling feature information.
  • the third feature extraction unit further includes a gating unit
  • the gating unit is used for the second output of the last residual block in the k+1th third feature extraction unit performing de-redundancy on the residual information and the feature information concatenated with the second feature information output by at least one of the L-1 residual blocks, and outputting the de-redundant feature information;
  • the k+1th enhanced upsampling feature information of the point cloud block is determined after adding the deredundant feature information to the kth enhanced upsampling feature information.
  • the feature upsampling module further includes a first autocorrelation attention network
  • the first autocorrelation attention network is used to perform feature interaction on the upsampling feature information of the point cloud block output by the feature upsampling submodule, and output the upsampling feature information of the point cloud block after feature interaction to the feature extraction submodule;
  • the feature extraction submodule is configured to output second feature information of the point cloud block according to the upsampled feature information of the point cloud block after feature interaction.
  • the feature dimension of the upsampled feature information of the point cloud block after the feature interaction is lower than the feature dimension of the upsampled feature information of the point cloud block.
  • the geometry generation module includes a plurality of fully connected layers
  • the multiple fully connected layers are used to output upsampled geometric information of the point cloud block according to the second feature information of the point cloud block.
  • the geometry generation module includes: a geometry reconstruction unit, a filtering unit, and a downsampling unit;
  • the geometric reconstruction unit is used to geometrically reconstruct the second feature information of the point cloud block, and output the initial upsampling geometric information of the point cloud block to the filtering unit;
  • the filtering unit is used to denoise the initial upsampling geometric information of the point cloud block, and output the initial upsampling geometric information of the point cloud block to filter noise to the downsampling unit;
  • the down-sampling unit is configured to down-sample the initial up-sampled geometric information of the point cloud block after filtering noise to a target up-sampling rate, and output the up-sampled geometric information of the point cloud block.
  • the target upsampling rate is less than or equal to the upsampling rate of the feature upsampling module.
  • the decoding unit 21 is further configured to: decode the point cloud code stream to obtain the target upsampling rate.
  • the device embodiment and the method embodiment may correspond to each other, and similar descriptions may refer to the method embodiment. To avoid repetition, details are not repeated here.
  • the point cloud decoder 20 shown in FIG. 17 may correspond to the corresponding subject in the point cloud decoding method of the embodiment of the present application, and the aforementioned and other operations and/or functions of each unit in the point cloud decoder 20 In order to realize the corresponding processes in the point cloud decoding method respectively, for the sake of brevity, details are not repeated here.
  • Fig. 18 is a schematic block diagram of a point cloud upsampling device provided by an embodiment of the present application.
  • the model training device 40 includes:
  • An acquisition unit 41 configured to acquire geometric information of the point cloud to be upsampled
  • a division unit 42 configured to divide the point cloud to be upsampled into at least one point cloud block according to the geometric information of the point cloud to be upsampled;
  • An up-sampling unit 43 configured to input the geometric information of the point cloud block into the generator for up-sampling, to obtain the up-sampling geometric information of the point cloud block;
  • the generator includes: a feature extraction module, a feature upsampling module and a geometry generation module, the feature extraction module is used to extract the first feature information of the point cloud block, and the feature sampling module is used to extract the The first feature information of the point cloud block is up-sampled to the second feature information, and the geometry generation module is used to map the second feature information of the point cloud block into a geometric space, so as to obtain the up-sampling of the point cloud block geometric information.
  • the feature extraction module comprises densely connected M feature extraction blocks
  • the i+1th feature extraction block in the M feature extraction blocks is used to output the i+1th third feature according to the input i-th fourth feature information information, the i-th fourth feature information is determined according to the i-th third feature information output by the i-th feature extraction block, and the first feature information of the point cloud block is determined according to the M feature extraction blocks Determined by the Mth third feature information output by the Mth feature extraction block, the i is a positive integer smaller than M.
  • the i-th fourth feature information is the first feature extracted by each feature extraction block before the i-th feature extraction block among the M feature extraction blocks.
  • the ith fourth feature information is the first third feature information output by the first feature extraction block in the M feature extraction blocks.
  • the feature extraction block includes: a first feature extraction unit and S second feature extraction units connected in series, wherein S is a positive integer;
  • the first extraction unit in the i+1th feature extraction block is used to search for K neighboring points of the current point for the current point in the point cloud block, and based on For the ith fourth feature information of the point cloud block, the fourth feature information of the current point is subtracted from the fourth feature information of the adjacent point to obtain K residual feature information, and the The K residual feature information is concatenated with the fourth feature information of the current point to obtain the i-th concatenated feature information of the current point, and according to the i-th concatenated feature information of the current point, the The i-th concatenated feature information of the point cloud block, and input the i-th concatenated feature information of the point cloud block into the first second feature extraction unit in the S second feature extraction units;
  • the first second feature extraction unit is used to output the first fifth feature information to the second second feature extraction unit according to the i-th cascaded feature information of the point cloud block, wherein the point cloud
  • the i+1th third feature information of the block is the fifth feature information output by the last second feature extraction unit among the S second feature extraction units.
  • the second feature extraction unit includes P residual blocks, where P is a positive integer
  • the j+1th residual block in the sth second feature extraction unit is used according to the jth residual in the sth second feature extraction unit
  • the j-th first residual information output by the block and the fifth feature information input to the s-th second feature extraction unit, and the j+1-th first residual information is output, wherein the j is less than P A positive integer, the s is a positive integer less than or equal to S;
  • the fifth feature information output by the sth second feature extraction unit is based on the first residual information output by at least one residual block in the sth second feature extraction unit, and input to the sth second feature extraction unit determined by the fifth feature information of the second feature extraction unit.
  • the up-sampling unit 43 is further configured to input the j-th first residual information and the j-th residual information output by the j-th residual block in the s-th second feature extraction unit to the s-th After the fifth feature information of the second feature extraction unit is added, it is input to the j+1th residual block in the sth second feature extraction unit.
  • the fifth feature information output by the s th second feature extraction unit is based on the first residual information output by the last residual block in the s th second feature extraction unit, and The feature information after the concatenation of the first residual information output by at least one residual block in the P-1 residual blocks is determined from the fifth feature information input to the s-th second feature extraction unit, wherein , the P-1 residual blocks are residual blocks except the last residual block among the P residual blocks of the s-th second feature extraction unit.
  • the fifth feature information output by the s th second feature extraction unit is based on the first residual information output by the last residual block in the s th second feature extraction unit, and The first residual information output by at least one residual block in the P-1 residual blocks is concatenated and then determined after being added to the fifth feature information input to the s-th second feature extraction unit of.
  • the second feature extraction unit further includes a gating unit
  • the gating unit in the s th second feature extraction unit is used for the first residual information output by the last residual block in the s th second feature extraction unit, De-redundancy is performed on the feature information concatenated with the first residual information output by at least one residual block in the P-1 residual blocks, and the de-redundant feature information is output; the s second The fifth feature information output by the feature extraction unit is determined after adding the de-redundant feature information and the fifth feature information input to the s-th second feature extraction unit.
  • the feature upsampling module includes: a feature upsampling submodule and a feature extraction submodule;
  • the feature upsampling submodule is used to copy r copies of the first feature information of the point cloud block according to the preset upsampling rate r, and add an n dimension to the feature dimension of the copied first feature information vector, obtain the upsampling feature information of the point cloud block, and input the upsampling feature information of the point cloud block into the feature extraction submodule, wherein the values of n-dimensional vectors corresponding to different first feature information are different;
  • the feature extraction submodule is configured to output second feature information of the point cloud block according to the upsampled feature information of the point cloud block.
  • the feature extraction submodule includes Q third feature extraction units, where Q is a positive integer
  • the k+1th third feature extraction unit is used to extract all The kth enhanced upsampling feature information of the point cloud block, output the k+1th enhanced upsampling feature information of the point cloud block, and the k is a positive integer less than Q;
  • the second feature information of the point cloud block is the Qth enhanced upsampling feature information of the point cloud block extracted by the last third feature extraction unit among the Q third feature extraction units.
  • the third feature extraction unit includes L residual blocks, and the L is a positive integer
  • the l+1th residual block is used according to the k+1th third feature extraction unit
  • the lth second residual information output by the lth residual block and the kth enhanced upsampling feature information input to the k+1th third feature extraction unit, output the l+1th second residual difference information, the l is a positive integer less than L;
  • the k+1th enhanced upsampling feature information of the point cloud block is the second residual information output from at least one residual block in the k+1th third feature extraction unit, and the kth Enhanced upsampling feature information determined.
  • the upsampling unit 43 is further configured to: after adding the lth second residual information output by the lth residual block and the kth enhanced upsampling feature information, input The l+1th residual block.
  • the k+1th enhanced upsampling feature information of the point cloud block is based on the second residual information output by the last residual block in the L residual blocks, and the L-1 residual
  • the second residual information output by at least one residual block in the difference block is determined by concatenating the feature information and the kth enhanced upsampling feature information, wherein the L-1 residual blocks are the A residual block except the last residual block among the L residual blocks of the k+1 third feature extraction unit.
  • the k+1th enhanced upsampling feature information of the point cloud block is based on the second residual information output by the last residual block in the L residual blocks, and the L-1 residual
  • the second residual information output by at least one residual block in the difference block is determined by adding the concatenated feature information to the kth enhanced upsampling feature information.
  • the third feature extraction unit further includes a gating unit
  • the gating unit is used for the second output of the last residual block in the k+1th third feature extraction unit performing de-redundancy on the residual information and the feature information concatenated with the second feature information output by at least one of the L-1 residual blocks, and outputting the de-redundant feature information;
  • the k+1th enhanced upsampling feature information of the point cloud block is determined after adding the deredundant feature information to the kth enhanced upsampling feature information.
  • the feature upsampling module further includes a first autocorrelation attention network
  • the first autocorrelation attention network is used to perform feature interaction on the upsampling feature information of the point cloud block output by the feature upsampling submodule, and output the upsampling feature information of the point cloud block after feature interaction to the feature extraction submodule;
  • the feature extraction submodule is configured to output second feature information of the point cloud block according to the upsampled feature information of the point cloud block after feature interaction.
  • the feature dimension of the upsampled feature information of the point cloud block after the feature interaction is lower than the feature dimension of the upsampled feature information of the point cloud block.
  • the geometry generation module includes a plurality of fully connected layers
  • the multiple fully connected layers are used to output upsampled geometric information of the point cloud block according to the second feature information of the point cloud block.
  • the geometry generation module includes: a geometry reconstruction unit, a filtering unit, and a downsampling unit;
  • the geometric reconstruction unit is used to geometrically reconstruct the second feature information of the point cloud block, and output the initial upsampling geometric information of the point cloud block to the filtering unit;
  • the filtering unit is used to denoise the initial upsampling geometric information of the point cloud block, and output the initial upsampling geometric information of the point cloud block to filter noise to the downsampling unit;
  • the down-sampling unit is configured to down-sample the initial up-sampled geometric information of the point cloud block after filtering noise to a target up-sampling rate, and output the up-sampled geometric information of the point cloud block.
  • the target upsampling rate is less than or equal to the upsampling rate of the feature upsampling module.
  • the device embodiment and the method embodiment may correspond to each other, and similar descriptions may refer to the method embodiment. To avoid repetition, details are not repeated here.
  • the point cloud upsampling device 40 shown in FIG. 18 may correspond to the corresponding subject in the point cloud upsampling method of the embodiment of the present application, and the aforementioned and other operations of each unit in the point cloud upsampling device 40 and The /or functions are respectively used to realize the corresponding process in the point cloud upsampling method, for the sake of brevity, no more details are given here.
  • Fig. 19 is a schematic block diagram of a model training device provided by an embodiment of the present application.
  • the model training device 10 includes:
  • Acquisition unit 11 for obtaining the geometric information of training point cloud
  • a division unit 12 configured to divide the training point cloud into at least one training point cloud block according to the geometric information of the training point cloud
  • the training unit 13 is used to input the geometric information of the training point cloud block into the feature extraction module of the generator for feature extraction, and obtain the first feature information of the training point cloud block; the first feature information of the training point cloud block is Feature information is input into the feature upsampling module of the generator for upsampling to obtain the second feature information of the training point cloud block; the second feature information of the training point cloud block is input into the geometry generation module of the generator Perform geometric reconstruction to obtain the predicted upsampling geometric information of the training point cloud block; according to the predicted upsampling geometric information of the training point cloud block, the feature extraction module, feature upsampling module and geometric generation in the generator The module is trained to obtain the trained generator.
  • the training unit 12 is specifically configured to input the predicted upsampled geometric information of the training point cloud block into a discriminator to obtain the first discriminant result of the discriminator, and the discriminator is used to judge the input Whether the data of the discriminator is the upsampled true value of the training point cloud block; according to the first discrimination result of the discriminator, the feature extraction module, feature upsampling module and geometry generation module in the generator are trained , to get the trained generator.
  • the feature extraction module includes M densely connected feature extraction blocks
  • the training unit 12 is specifically configured to input the geometric information of the training point cloud block into the feature extraction module, and obtain the M
  • the i-th third feature information of the training point cloud block extracted by the i-th feature extraction block in the feature extraction block the i is a positive integer less than M; according to the i-th feature information of the training point cloud block
  • the third feature information is to obtain the i-th fourth feature information of the training point cloud block; the i-th fourth feature information of the training point cloud block is input in the i+1 feature extraction block to obtain the i+1 feature extraction block.
  • the i+1th third feature information of the training point cloud block; the Mth third feature information extracted by the Mth feature extraction block of the training point cloud block is used as the first feature information of the training point cloud block characteristic information.
  • the training unit 12 is specifically configured to, if i is not equal to 1, obtain the third feature extracted by each feature extraction block located before the ith feature extraction block among the M feature extraction blocks information; and the third feature information extracted by each feature extraction block located before the ith feature extraction block is concatenated with the third feature information extracted by the ith feature extraction block, as the The i-th fourth feature information of the training point cloud block;
  • the first third feature information extracted by the first feature extraction block in the M feature extraction units is used as the ith fourth feature information of the training point cloud block.
  • the feature extraction block includes: a first feature extraction unit and at least one second feature extraction unit connected in series, and a training unit 12, which is specifically used to convert the i-th fourth feature of the training point cloud block
  • the feature information is input to the first feature extraction unit in the i+1th feature extraction block, so that the first feature extraction unit searches for K of the current point for the current point in the training point cloud block.
  • the K residual feature information is concatenated with the fourth feature information of the current point to obtain the i-th concatenated feature information of the current point, and according to the i-th concatenated feature information of the current point, it is obtained
  • the i-th concatenated feature information of the training point cloud block; the i-th concatenated feature information of the training point cloud block is input into the first second feature extraction in the i+1 feature extraction block unit to obtain the first, first, and fifth feature information, and input the first, first, and fifth feature information into the second, second feature extraction unit in the i+1th feature extraction block , to obtain the second fifth feature information; the fifth feature information extracted by the last second feature extraction unit in the i+1 feature extraction block is used as the i+1th feature information of the training point cloud block Three characteristic information.
  • the second feature extraction unit includes P residual blocks, where P is a positive integer
  • the training unit 12 is specifically configured to input the i-th concatenated feature information into the i+th
  • the first second feature extraction unit in one feature extraction block obtains the first residual information output by the jth residual block in the first second feature extraction unit, where j is less than or equal to P is a positive integer; input the first residual information output by the jth residual block and the ith concatenated feature information into the j+1th residual in the first second feature extraction unit In the block, the first residual information output by the j+1th residual block is obtained; according to the first residual information output by at least one of the P residual blocks in the first second feature extraction unit The residual information, as well as the i-th concatenated feature information, determine fifth feature information output by the first second feature extraction unit.
  • the training unit 12 is specifically configured to add the first residual information output by the jth residual block and the ith concatenated feature information, and add the added feature The information is input into the j+1th residual block, and the first residual information output by the j+1th residual block is obtained.
  • the training unit 12 is specifically configured to output the first residual information output by the last residual block in the P residual blocks and at least one residual block in the P-1 residual blocks
  • the first residual information is concatenated, wherein the P-1 residual blocks are residual blocks except the last residual block among the P residual blocks; according to the concatenated features information and the i-th concatenated feature information to determine the fifth feature information output by the first second feature extraction unit.
  • the training unit 12 is specifically configured to add the concatenated feature information and the i-th concatenated feature information as the first output of the first second feature extraction unit. Five characteristic information.
  • the second feature extraction unit further includes a gating unit, a training unit 12, specifically configured to input the cascaded feature information into the gating unit for de-redundancy, to obtain de-redundancy
  • the feature information after de-redundancy is added to the i-th concatenated feature information as the fifth feature information output by the first second feature extraction unit.
  • the feature upsampling module includes: a feature upsampling submodule and a feature extraction submodule, a training unit 12, specifically for inputting the first feature information of the training point cloud block into the feature upsampling sub-module, so that the feature up-sampling sub-module copies r shares of the first feature information of the training point cloud block according to the preset up-sampling rate r, and performs a feature dimension on the copied first feature information Add an n-dimensional vector to obtain the upsampling feature information of the training point cloud block, wherein the values of the n-dimensional vectors corresponding to different first feature information are different; the upsampling feature information of the training point cloud block is input into the The feature extraction sub-module obtains the second feature information of the training point cloud block extracted by the feature extraction sub-module.
  • the feature extraction submodule includes Q third feature extraction units connected in series, the Q is a positive integer, and the training unit 12 is specifically used to convert the upsampled feature information of the training point cloud block to Input the feature extraction submodule to obtain the kth enhancement upsampling feature information of the training point cloud block extracted by the kth third feature extraction unit; the kth enhancement upsampling feature information of the training point cloud block
  • the feature information is input into the k+1th third feature extraction unit to obtain the k+1th enhanced upsampling feature information of the training point cloud block extracted by the k+1th third feature extraction unit;
  • the Qth enhanced upsampled feature information of the training point cloud block extracted by the last third feature extraction unit among the Q third feature extraction units is used as the second feature information of the training point cloud block.
  • the third feature extraction unit includes L residual blocks, where L is a positive integer
  • the training unit 12 is specifically used to enhance and upsample the feature information of the k th training point cloud block
  • the second residual information output by the residual block according to the second residual information output by at least one residual block in the L residual blocks, and the kth enhanced upsampling feature information, the training point is obtained The k+1th enhanced upsampled feature information of the cloud block.
  • the training unit 12 is specifically configured to add the second residual information output by the lth residual block and the kth enhanced upsampling feature information, and add the added The characteristic information is input into the l+1th residual block, and the second residual information output by the l+1th residual block is determined.
  • the training unit 12 is specifically configured to output the second residual information output by the last residual block in the L residual blocks, and at least one residual block output in the L-1 residual blocks
  • the second residual information is concatenated, wherein the L-1 residual blocks are residual blocks except the last residual block among the L residual blocks; according to the concatenated features information and the kth enhanced upsampling feature information to determine the k+1th enhanced upsampling feature information of the training point cloud block.
  • the training unit 12 is specifically configured to add the concatenated feature information and the kth enhanced upsampling feature information as the k+1th of the training point cloud block Enhance upsampled feature information.
  • the third feature extraction unit further includes a gating unit, a training unit 12, specifically configured to input the cascaded feature information into the gating unit for de-redundancy, to obtain de-redundancy
  • the feature information after de-redundancy is added to the k-th enhanced up-sampling feature information, and used as the k+1th enhanced up-sampling feature information of the training point cloud block.
  • the feature upsampling module further includes a first autocorrelation attention network, a training unit 12, specifically for inputting the upsampling feature information of the training point cloud block into the first autocorrelation attention
  • the network performs feature interaction to obtain the upsampling feature information of the training point cloud block after the feature interaction; the upsampling feature information of the training point cloud block after the feature interaction is input into the feature extraction submodule to perform feature extraction, and obtain The second feature information of the training point cloud block.
  • the feature dimension of the upsampled feature information of the training point cloud block after the feature interaction is lower than the feature dimension of the upsampled feature information of the training point cloud block.
  • the geometry generation module includes multiple fully connected layers
  • the training unit 12 is specifically configured to input the second feature information of the training point cloud block into the multiple fully connected layers to obtain the training Prediction of point cloud blocks upsamples geometric information.
  • the geometry generation module includes: a geometry reconstruction unit, a filter unit, and a downsampling unit, and a training unit 12, specifically configured to input the second feature information of the training point cloud block into the geometry reconstruction unit for further processing.
  • Geometric reconstruction to obtain the initial upsampling geometric information of the training point cloud block; input the initial upsampling geometric information of the training point cloud block into a filter unit for noise removal, and obtain the initial upsampling of the training point cloud block to filter out noise Sampling geometric information; inputting the initial upsampling geometric information of the training point cloud block to filter out noise into the downsampling unit for downsampling, to obtain the predicted upsampling geometric information of the training point cloud block.
  • the upsampling rate corresponding to the upsampling geometric information of the training point cloud block is less than or equal to the upsampling rate of the feature upsampling module.
  • the discriminator is a pre-trained discriminator.
  • the training unit 12 is further configured to use the geometric information of the training point cloud block to train the discriminator.
  • the training unit 12 is specifically configured to input the predicted upsampling geometric information of the training point cloud block generated by the generator into the discriminator, and obtain a second discrimination result of the discriminator; Inputting the upsampling true value of the geometric information of the training point cloud block into the discriminator to obtain a third discriminant result of the discriminator; according to the second discriminant result and the third discriminant result, determine the discriminator The loss; according to the loss of the discriminator, the discriminator is trained.
  • the training unit 12 is specifically configured to determine the loss of the discriminator by using a least square loss function according to the second discrimination result and the third discrimination result.
  • the discriminator includes: a global discriminant module, a boundary discriminant module and a fully connected module, and a training unit 12, which is specifically used to obtain the geometric information of the boundary points of the target point cloud block;
  • the geometric information of the boundary point is input into the boundary discrimination module to obtain the boundary feature information of the target point cloud block;
  • the geometric information of the target point cloud block is input to the global discrimination module to obtain the target point cloud block Global feature information;
  • the training unit 12 is specifically configured to use a high-pass image filter to extract the geometric information of the boundary points of the target point cloud block.
  • the training unit 12 is specifically configured to concatenate the global feature information and boundary feature information of the target point cloud block; input the concatenated global feature information and boundary feature information into the full connection module , to obtain the target discrimination result of the discriminator.
  • the global discrimination module includes sequentially along the network depth direction: a first number of multi-layer perceptrons, a first maximum pooling layer, a second autocorrelation attention network, and a second number of multi-layer perceptrons machine and the second maximum pooling layer;
  • the training unit 12 is specifically used to input the geometric information of the target point cloud block into the first number of multi-layer perceptrons for feature extraction, and obtain the first number of the target point cloud block A global feature information; inputting the first global feature information into the first maximum pooling layer for dimensionality reduction processing to obtain the second global feature information of the target point cloud block; combining the first global feature information and The second global feature information is input into the second autocorrelation attention network for feature interaction to obtain the third global feature information of the target point cloud block; the third global feature information is input into the second number of The multi-layer perceptron then extracts features to obtain the fourth global feature information of the target point cloud block; input the fourth global feature information into the second maximum pooling layer for dimensionality reduction processing to obtain the target
  • the training unit 12 is specifically configured to concatenate the first global feature information and the second global feature information; combine the concatenated first global feature information and the second global feature information
  • the global feature information is input into the second autocorrelation attention network for feature interaction to obtain the third global feature information of the target point cloud block.
  • the first quantity is equal to the second quantity.
  • the first number and the second number are both equal to 2.
  • the first number of multilayer perceptrons includes a first layer of multilayer perceptrons and a second layer of multilayer perceptrons
  • the second number of multilayer perceptrons includes a third layer of multilayer perceptrons machine and the fourth layer of multi-layer perceptron, the first layer of multi-layer perceptron, the second layer of multi-layer perceptron, the third layer of multi-layer perceptron and the fourth layer of multi-layer perceptron
  • the feature dimension gradually increases sequentially.
  • the feature dimension of the first layer of multi-layer perceptron is 32
  • the feature dimension of the second layer of multi-layer perceptron is 64
  • the feature dimension of the third layer of multi-layer perceptron is 128, so
  • the feature dimension of the fourth layer multilayer perceptron is 256.
  • the boundary discrimination module sequentially includes along the network depth direction: a third number of multi-layer perceptrons, a third maximum pooling layer, a third autocorrelation attention network, and a fourth number of multi-layer perceptrons machine and the fourth maximum pooling layer;
  • the training unit 12 is specifically used to input the geometric information of the boundary points of the target point cloud block into the third number of multi-layer perceptrons for feature extraction, and obtain the target point
  • the first boundary feature information of the cloud block the first boundary feature information is input into the third maximum pooling layer for dimension reduction processing, and the second boundary feature information of the target point cloud block is obtained;
  • the first The boundary feature information and the second boundary feature information are input into the third autocorrelation attention network for feature interaction to obtain the third boundary feature information of the target point cloud block;
  • the third boundary feature information is input into the A fourth number of multi-layer perceptrons perform feature extraction to obtain fourth boundary feature information of the target point cloud block; input the fourth boundary feature information into the fourth maximum pooling layer for dimensionality reduction processing to obtain
  • the training unit 12 is specifically configured to concatenate the first boundary feature information and the second boundary feature information; combine the concatenated first boundary feature information and the second boundary feature information
  • the boundary feature information is input into the third autocorrelation attention network for feature interaction to obtain the third boundary feature information of the target point cloud block.
  • the third quantity is equal to the fourth quantity.
  • both the third quantity and the fourth quantity are equal to 2.
  • the third number of multilayer perceptrons includes a fifth layer of multilayer perceptrons and a sixth layer of multilayer perceptrons
  • the fourth number of multilayer perceptrons includes a seventh layer of multilayer perceptrons machine and the eighth layer multi-layer perceptron, the fifth layer multi-layer perceptron, the sixth layer multi-layer perceptron, the seventh layer multi-layer perceptron and the eighth layer multi-layer perceptron
  • the feature dimension gradually increases sequentially.
  • the feature dimension of the eighth-layer multi-layer perceptron is greater than or equal to the feature dimension of the seventh-layer multi-layer perceptron, and smaller than or equal to the feature dimension of the fourth-layer multi-layer perceptron.
  • the feature dimension of the fifth-layer multi-layer perceptron is 32
  • the feature dimension of the sixth-layer multi-layer perceptron is 64
  • the feature dimension of the seventh-layer multi-layer perceptron is 128, so
  • the feature dimension of the eighth layer multilayer perceptron is 192.
  • the training unit 12 is specifically configured to determine the first loss of the generator according to the first discrimination result; and determine the feature extraction module and feature of the generator according to the first loss. Parameter matrix for the upsampling block and the geometry generation block.
  • the training unit 12 is specifically configured to determine the first loss of the generator by using a least squares loss function according to the first discrimination result.
  • the training unit 12 is specifically configured to determine at least one second loss of the generator; according to the first loss of the generator and at least one second loss of the generator, determine the generation The target loss of the generator; according to the target loss of the generator, determine the parameter matrix of the feature extraction module, feature upsampling module and geometry generation module in the generator.
  • the training unit 12 is specifically configured to determine the A second loss for the generator.
  • the training unit 12 is specifically configured to downsample the upsampled geometric information of the training point cloud block to obtain a downsampled training point cloud block with the same number of points as the training point cloud block; according to the Downsampling the geometric information of the training point cloud block and the geometric information of the training point cloud block, sampling the ground motion distance method, and determining a second loss of the generator.
  • the training unit 12 is specifically configured to determine a second loss of the generator according to the following formula:
  • the L id is the second loss of the generator
  • the P ori is a training point cloud block
  • the P low is a downsampled training point cloud block
  • ⁇ :P low ⁇ P ori means that P For the bijection composed of low and P ori , there is only one and only way to move P low and P ori to the minimum distance between the point sets of each other. is the kth point in the P low , the for the said Corresponding point in the P ori .
  • the training unit 12 is specifically configured to determine at least one second loss of the generator according to a uniform loss function.
  • the training unit 12 is specifically configured to use a weighted average of the first loss of the generator and the at least one second loss to determine the target loss of the generator.
  • the device embodiment and the method embodiment may correspond to each other, and similar descriptions may refer to the method embodiment. To avoid repetition, details are not repeated here.
  • the point cloud upsampling device 10 shown in FIG. 18 may correspond to the corresponding subject in the model training method of the embodiment of the present application, and the foregoing and other operations and/or The functions are to realize the corresponding processes in each method such as the model training method, and for the sake of brevity, details are not repeated here.
  • the functional unit may be implemented in the form of hardware, may also be implemented by instructions in the form of software, and may also be implemented by a combination of hardware and software units.
  • each step of the method embodiment in the embodiment of the present application can be completed by an integrated logic circuit of the hardware in the processor and/or instructions in the form of software, and the steps of the method disclosed in the embodiment of the present application can be directly embodied as hardware
  • the decoding processor is executed, or the combination of hardware and software units in the decoding processor is used to complete the execution.
  • the software unit may be located in a mature storage medium in the field such as random access memory, flash memory, read-only memory, programmable read-only memory, electrically erasable programmable memory, and registers.
  • the storage medium is located in the memory, and the processor reads the information in the memory, and completes the steps in the above method embodiments in combination with its hardware.
  • Fig. 20 is a schematic block diagram of an electronic device provided by an embodiment of the present application.
  • the electronic device 30 may be the point cloud upsampling device described in the embodiment of the present application, or a point cloud decoder, or a model training device, and the electronic device 30 may include:
  • a memory 33 and a processor 32 the memory 33 is used to store a computer program 34 and transmit the program code 34 to the processor 32 .
  • the processor 32 can call and run the computer program 34 from the memory 33 to implement the method in the embodiment of the present application.
  • the processor 32 can be used to execute the steps in the above-mentioned method 200 according to the instructions in the computer program 34 .
  • the processor 32 may include, but is not limited to:
  • DSP Digital Signal Processor
  • ASIC Application Specific Integrated Circuit
  • FPGA Field Programmable Gate Array
  • the memory 33 includes but is not limited to:
  • non-volatile memory can be read-only memory (Read-Only Memory, ROM), programmable read-only memory (Programmable ROM, PROM), erasable programmable read-only memory (Erasable PROM, EPROM), electronically programmable Erase Programmable Read-Only Memory (Electrically EPROM, EEPROM) or Flash.
  • the volatile memory can be Random Access Memory (RAM), which acts as external cache memory.
  • RAM Static Random Access Memory
  • SRAM Static Random Access Memory
  • DRAM Dynamic Random Access Memory
  • Synchronous Dynamic Random Access Memory Synchronous Dynamic Random Access Memory
  • SDRAM double data rate synchronous dynamic random access memory
  • Double Data Rate SDRAM, DDR SDRAM double data rate synchronous dynamic random access memory
  • Enhanced SDRAM, ESDRAM enhanced synchronous dynamic random access memory
  • SLDRAM synchronous connection dynamic random access memory
  • Direct Rambus RAM Direct Rambus RAM
  • the computer program 34 can be divided into one or more units, and the one or more units are stored in the memory 33 and executed by the processor 32 to complete the present application.
  • the one or more units may be a series of computer program instruction segments capable of accomplishing specific functions, and the instruction segments are used to describe the execution process of the computer program 34 in the electronic device 30 .
  • the electronic device 30 may also include:
  • a transceiver 33 the transceiver 33 can be connected to the processor 32 or the memory 33 .
  • the processor 32 can control the transceiver 33 to communicate with other devices, specifically, can send information or data to other devices, or receive information or data sent by other devices.
  • Transceiver 33 may include a transmitter and a receiver.
  • the transceiver 33 may further include antennas, and the number of antennas may be one or more.
  • bus system includes not only a data bus, but also a power bus, a control bus and a status signal bus.
  • the present application also provides a computer storage medium, on which a computer program is stored, and when the computer program is executed by a computer, the computer can execute the methods of the above method embodiments.
  • the embodiments of the present application further provide a computer program product including instructions, and when the instructions are executed by a computer, the computer executes the methods of the foregoing method embodiments.
  • the computer program product includes one or more computer instructions.
  • the computer can be a general purpose computer, a special purpose computer, a computer network, or other programmable device.
  • the computer instructions may be stored in or transmitted from one computer-readable storage medium to another computer-readable storage medium, for example, the computer instructions may be transferred from a website, computer, server, or data center by wire (such as coaxial cable, optical fiber, digital subscriber line (DSL)) or wireless (such as infrared, wireless, microwave, etc.) to another website site, computer, server or data center.
  • the computer-readable storage medium may be any available medium that can be accessed by a computer, or a data storage device such as a server or a data center integrated with one or more available media.
  • the available medium may be a magnetic medium (such as a floppy disk, a hard disk, or a magnetic tape), an optical medium (such as a digital point cloud disc (digital video disc, DVD)), or a semiconductor medium (such as a solid state disk (solid state disk, SSD)), etc. .
  • a magnetic medium such as a floppy disk, a hard disk, or a magnetic tape
  • an optical medium such as a digital point cloud disc (digital video disc, DVD)
  • a semiconductor medium such as a solid state disk (solid state disk, SSD)
  • the disclosed systems, devices and methods may be implemented in other ways.
  • the device embodiments described above are only illustrative.
  • the division of the units is only a logical function division. In actual implementation, there may be other division methods.
  • multiple units or components can be combined or can be Integrate into another system, or some features may be ignored, or not implemented.
  • the mutual coupling or direct coupling or communication connection shown or discussed may be through some interfaces, and the indirect coupling or communication connection of devices or units may be in electrical, mechanical or other forms.
  • a unit described as a separate component may or may not be physically separated, and a component displayed as a unit may or may not be a physical unit, that is, it may be located in one place, or may be distributed to multiple network units. Part or all of the units can be selected according to actual needs to achieve the purpose of the solution of this embodiment.
  • each functional unit in each embodiment of the present application may be integrated into one processing unit, each unit may exist separately physically, or two or more units may be integrated into one unit.

Abstract

The present application provides point cloud decoding and upsampling and model training methods and apparatus. The point cloud decoding method comprises: obtaining geometric information of a point cloud; according to the geometric information of the point cloud, dividing the point cloud into at least one point cloud block; and inputting the geometric information of the point cloud blocks into a generator for upsampling, so as to obtain upsampled geometric information of the point cloud blocks, wherein the generator comprises: a feature extraction module, a feature upsampling module, and a geometric generation module, the feature extraction module is used to extract first feature information of the point cloud blocks, the feature upsampling module is used to upsample the first feature information of the point cloud blocks to second feature information, and the geometric generation module is used to map the second feature information of the point cloud blocks into a geometric space, so as to obtain the upsampled geometric information of the point cloud blocks. That is, the generator of the present application is a deep learning-based generator, which is used in turn to generate a high-precision point cloud having high accuracy.

Description

点云解码、上采样及模型训练方法与装置Point cloud decoding, upsampling and model training method and device 技术领域technical field
本申请涉及点云技术领域,尤其涉及一种点云解码、上采样及模型训练方法与装置。The present application relates to the field of point cloud technology, and in particular to a point cloud decoding, upsampling and model training method and device.
背景技术Background technique
通过采集设备对物体表面进行采集,形成点云数据,点云数据包括几十万甚至更多的点。在视频制作过程中,将点云数据以点云媒体文件的形式在点云编码设备和点云解码设备之间传输。但是,如此庞大的点给传输带来了挑战,因此,点云编码设备需要对点云数据进行压缩后传输。The surface of the object is collected by the collection device to form point cloud data, which includes hundreds of thousands or more points. During the video production process, the point cloud data is transmitted between the point cloud encoding device and the point cloud decoding device in the form of point cloud media files. However, such a large number of points brings challenges to transmission, therefore, point cloud encoding equipment needs to compress the point cloud data before transmission.
点云解码端对点云码流进行解码,得到重建点云。但是,在一些应用场景中,需要使用比原始精度更高的高质量点云,比如,自动驾驶领域对雷达采集的稀疏点云,需要做大量的后处理提升点云的精度,以提升驾驶的安全性。但是,目前的点云上采样方法的点云上采样效果差,准确性低。The point cloud decoding end decodes the point cloud code stream to obtain the reconstructed point cloud. However, in some application scenarios, it is necessary to use high-quality point clouds with higher accuracy than the original ones. For example, for the sparse point clouds collected by radar in the field of autonomous driving, a lot of post-processing is required to improve the accuracy of point clouds to improve driving performance. safety. However, current point cloud upsampling methods have poor point cloud upsampling effects and low accuracy.
发明内容Contents of the invention
本申请实施例提供了一种点云解码、上采样及模型训练方法与装置,以提高点云上采样的准确性。The embodiment of the present application provides a point cloud decoding, upsampling and model training method and device, so as to improve the accuracy of point cloud upsampling.
第一方面,本申请实施例提供一种点云解码方法,包括:In the first aspect, the embodiment of the present application provides a point cloud decoding method, including:
解码点云码流,得到点云的几何信息;Decode the point cloud code stream to obtain the geometric information of the point cloud;
根据所述点云的几何信息,将所述点云划分成至少一个点云块;dividing the point cloud into at least one point cloud block according to the geometric information of the point cloud;
将所述点云块的几何信息输入生成器中进行上采样,得到所述点云块的上采样几何信息;Input the geometric information of the point cloud block into the generator for upsampling, and obtain the upsampling geometric information of the point cloud block;
其中,所述生成器包括:特征提取模块、特征上采样模块和几何生成模块,所述特征提取模块用于提取所述点云块的第一特征信息,所述特征采样模块用于将所述点云块的第一特征信息上采样为第二特征信息,所述几何生成模块用于将所述点云块的第二特征信息映射至几何空间中,以得到所述点云块的上采样几何信息。Wherein, the generator includes: a feature extraction module, a feature upsampling module and a geometry generation module, the feature extraction module is used to extract the first feature information of the point cloud block, and the feature sampling module is used to extract the The first feature information of the point cloud block is up-sampled to the second feature information, and the geometry generation module is used to map the second feature information of the point cloud block into a geometric space, so as to obtain the up-sampling of the point cloud block geometric information.
第二方面,本申请提供了一种点云上采样方法,包括:In a second aspect, the present application provides a point cloud upsampling method, including:
获取待上采样点云的几何信息;Obtain the geometric information of the point cloud to be upsampled;
根据所述待上采样点云的几何信息,将所述待上采样点云划分成至少一个点云块;Divide the point cloud to be upsampled into at least one point cloud block according to the geometric information of the point cloud to be upsampled;
将所述点云块的几何信息输入生成器中进行上采样,得到所述点云块的上采样几何信息;Input the geometric information of the point cloud block into the generator for upsampling, and obtain the upsampling geometric information of the point cloud block;
其中,所述生成器包括:特征提取模块、特征上采样模块和几何生成模块,所述特征提取模块用于提取所述点云块的第一特征信息,所述特征采样模块用于将所述点云块的第一特征信息上采样为第二特征信息,所述几何生成模块用于将所述点云块的第二特征信息映射至几何空间中,以得到所述点云块的上采样几何信息。Wherein, the generator includes: a feature extraction module, a feature upsampling module and a geometry generation module, the feature extraction module is used to extract the first feature information of the point cloud block, and the feature sampling module is used to extract the The first feature information of the point cloud block is up-sampled to the second feature information, and the geometry generation module is used to map the second feature information of the point cloud block into a geometric space, so as to obtain the up-sampling of the point cloud block geometric information.
第三方面,本申请提供了一种模型训练方法,包括:In a third aspect, the present application provides a model training method, including:
获取训练点云的几何信息,并根据所述训练点云的几何信息,将所述训练点云划分成至少一个训练点云块;Obtaining geometric information of the training point cloud, and dividing the training point cloud into at least one training point cloud block according to the geometric information of the training point cloud;
将所述训练点云块的几何信息输入生成器的特征提取模块进行特征提取,得到所述训练点云块的第一特征信息;The feature extraction module of the geometric information input generator of described training point cloud block is carried out feature extraction, obtains the first feature information of described training point cloud block;
将所述训练点云块的第一特征信息输入所述生成器的特征上采样模块进行上采样,得到所述训练点云块的第二特征信息;Inputting the first feature information of the training point cloud block into the feature upsampling module of the generator for upsampling to obtain the second feature information of the training point cloud block;
将所述训练点云块的第二特征信息输入所述生成器的几何生成模块进行几何重建,得到所述训练点云块的预测上采样几何信息;Inputting the second feature information of the training point cloud block into the geometric generation module of the generator for geometric reconstruction, and obtaining the predicted upsampling geometric information of the training point cloud block;
根据所述训练点云块的预测上采样几何信息,对所述生成器中的特征提取模块、特征上采样模块和几何生成模块进行训练,得到训练后的生成器。According to the predicted upsampling geometric information of the training point cloud block, the feature extraction module, feature upsampling module and geometry generation module in the generator are trained to obtain the trained generator.
第四方面,提供了一种点云解码器,用于执行上述第一方面或其各实现方式中的方法。具体地,该点云解码器包括用于执行上述第一方面或其各实现方式中的方法的功能单元。In a fourth aspect, a point cloud decoder is provided, configured to execute the method in the above first aspect or various implementations thereof. Specifically, the point cloud decoder includes a functional unit for executing the method in the above first aspect or its implementations.
第五方面,提供了一种点云解码器,包括处理器和存储器。该存储器用于存储计算机程序,该处 理器用于调用并运行该存储器中存储的计算机程序,以执行上述第一方面或其各实现方式中的方法。In a fifth aspect, a point cloud decoder is provided, including a processor and a memory. The memory is used to store a computer program, and the processor is used to call and run the computer program stored in the memory to execute the method in the above first aspect or its various implementations.
第六方面,提供了一种点云上采样装置,用于执行上述第二方面或其各实现方式中的方法。具体地,该点云上采样装置包括用于执行上述第二方面或其各实现方式中的方法的功能单元。In a sixth aspect, a device for upsampling a point cloud is provided, configured to execute the method in the above second aspect or its various implementations. Specifically, the point cloud upsampling device includes a functional unit for executing the method in the above second aspect or its various implementations.
第七方面,提供了一种点云上采样设备,包括处理器和存储器。该存储器用于存储计算机程序,该处理器用于调用并运行该存储器中存储的计算机程序,以执行上述第二方面或其各实现方式中的方法。In a seventh aspect, a point cloud upsampling device is provided, including a processor and a memory. The memory is used to store a computer program, and the processor is used to invoke and run the computer program stored in the memory, so as to execute the method in the above second aspect or its various implementations.
第八方面,提供了一种模型训练装置,用于执行上述第三方面或其各实现方式中的方法。具体地,该模型训练装置包括用于执行上述第三方面或其各实现方式中的方法的功能单元。In an eighth aspect, a model training device is provided, configured to execute the method in the above third aspect or various implementations thereof. Specifically, the model training device includes a functional unit for executing the method in the above third aspect or its various implementations.
第九方面,提供了一种模型训练设备,包括处理器和存储器。该存储器用于存储计算机程序,该处理器用于调用并运行该存储器中存储的计算机程序,以执行上述第三方面或其各实现方式中的方法。In a ninth aspect, a model training device is provided, including a processor and a memory. The memory is used to store a computer program, and the processor is used to call and run the computer program stored in the memory, so as to execute the method in the above third aspect or its various implementations.
第十方面,提供了一种芯片,用于实现上述第一方面至第三方面中的任一方面或其各实现方式中的方法。具体地,该芯片包括:处理器,用于从存储器中调用并运行计算机程序,使得安装有该芯片的设备执行如上述第一方面至第三方面中的任一方面或其各实现方式中的方法。In a tenth aspect, a chip is provided, configured to implement any one of the foregoing first to third aspects or the method in each implementation manner thereof. Specifically, the chip includes: a processor, configured to call and run a computer program from the memory, so that the device installed with the chip executes any one of the above-mentioned first to third aspects or any of the implementations thereof. method.
第十一方面,提供了一种计算机可读存储介质,用于存储计算机程序,该计算机程序使得计算机执行上述第一方面至第三方面中的任一方面或其各实现方式中的方法。In an eleventh aspect, there is provided a computer-readable storage medium for storing a computer program, and the computer program causes a computer to execute any one of the above-mentioned first to third aspects or the method in each implementation manner thereof.
第十二方面,提供了一种计算机程序产品,包括计算机程序指令,该计算机程序指令使得计算机执行上述第一方面至第三方面中的任一方面或其各实现方式中的方法。A twelfth aspect provides a computer program product, including computer program instructions, the computer program instructions cause a computer to execute any one of the above first to third aspects or the method in each implementation manner.
第十三方面,提供了一种计算机程序,当其在计算机上运行时,使得计算机执行上述第一方面至第三方面中的任一方面或其各实现方式中的方法。A thirteenth aspect provides a computer program, which, when running on a computer, causes the computer to execute any one of the above first to third aspects or the method in each implementation manner.
基于以上技术方案,通过点云的几何信息,将点云划分成至少一个点云块;将点云块的几何信息输入生成器中进行上采样,得到点云块的上采样几何信息;生成器包括:特征提取模块、特征上采样模块和几何生成模块,特征提取模块用于提取点云块的第一特征信息,特征采样模块用于将点云块的第一特征信息上采样为第二特征信息,几何生成模块用于将点云块的第二特征信息映射至几何空间中,以得到点云块的上采样几何信息。即本申请实施例的生成器为基于深度学习的生成器,通过深度学习可以学习到点云的更多特征信息,进而使用该生成器进行点云上采样时,可以生成高精度的点云,且该高精度的点云的特征与点云的真值接近,进而提高了点云上采样的准确性。Based on the above technical scheme, the point cloud is divided into at least one point cloud block through the geometric information of the point cloud; the geometric information of the point cloud block is input into the generator for up-sampling, and the up-sampling geometric information of the point cloud block is obtained; the generator Including: a feature extraction module, a feature upsampling module and a geometry generation module, the feature extraction module is used to extract the first feature information of the point cloud block, and the feature sampling module is used to upsample the first feature information of the point cloud block into the second feature Information, the geometry generation module is used to map the second feature information of the point cloud block into the geometric space, so as to obtain the upsampling geometric information of the point cloud block. That is, the generator in the embodiment of the present application is a generator based on deep learning, through which more characteristic information of the point cloud can be learned, and then when the generator is used for upsampling of the point cloud, a high-precision point cloud can be generated, Moreover, the features of the high-precision point cloud are close to the true value of the point cloud, thereby improving the accuracy of point cloud upsampling.
附图说明Description of drawings
图1为本申请实施例涉及的一种点云编解码系统的示意性框图;FIG. 1 is a schematic block diagram of a point cloud encoding and decoding system involved in an embodiment of the present application;
图2是本申请实施例提供的点云编码器的示意性框图;Fig. 2 is a schematic block diagram of a point cloud encoder provided by an embodiment of the present application;
图3是本申请实施例提供的点云解码器的示意性框图;Fig. 3 is a schematic block diagram of a point cloud decoder provided by an embodiment of the present application;
图4为本申请一实施例提供的模型训练方法流程示意图;FIG. 4 is a schematic flow chart of a model training method provided by an embodiment of the present application;
图5为本申请实施例的生成器的一种网络示意图;FIG. 5 is a schematic diagram of a network of a generator according to an embodiment of the present application;
图6A为本申请实施例涉及的特征提取模块的一种结构示意图;FIG. 6A is a schematic structural diagram of a feature extraction module involved in an embodiment of the present application;
图6B为本申请实施例涉及的特征提取块的一种结构示意图;FIG. 6B is a schematic structural diagram of a feature extraction block involved in an embodiment of the present application;
图6C为本申请实施例涉及的第二特征提取单元HRA的一种结构示意图;FIG. 6C is a schematic structural diagram of the second feature extraction unit HRA involved in the embodiment of the present application;
图6D为本申请实施例涉及的残差块的一种结构示意图;FIG. 6D is a schematic structural diagram of a residual block involved in the embodiment of the present application;
图6E为本申请实施例涉及的第二特征提取单元HRA的一种结构示意图;FIG. 6E is a schematic structural diagram of the second feature extraction unit HRA involved in the embodiment of the present application;
图6F为本申请实施例涉及的第二特征提取单元HRA的一种结构示意图;FIG. 6F is a schematic structural diagram of the second feature extraction unit HRA involved in the embodiment of the present application;
图6G为本申请实施例涉及的门控单元的一种结构示意图;FIG. 6G is a schematic structural diagram of a gating unit involved in an embodiment of the present application;
图7A为本申请实施例涉及的特征上采样模块的一种结构示意图;FIG. 7A is a schematic structural diagram of a feature upsampling module involved in an embodiment of the present application;
图7B为本申请实施例涉及的特征上采样模块的另一种结构示意图;FIG. 7B is another schematic structural diagram of the feature upsampling module involved in the embodiment of the present application;
图7C为本申请实施例涉及的特征提取子模块的一种结构示意图;FIG. 7C is a schematic structural diagram of the feature extraction submodule involved in the embodiment of the present application;
图7D为本申请实施例提供的特征上采样模块的一种具体网络结构示意图;FIG. 7D is a schematic diagram of a specific network structure of the feature upsampling module provided by the embodiment of the present application;
图8为本申请实施例提供的几何生成模块的一种具体网络结构示意图;FIG. 8 is a schematic diagram of a specific network structure of the geometry generation module provided by the embodiment of the present application;
图9为本申请实施例涉及生成器的训练过程的一种示意图;FIG. 9 is a schematic diagram of a training process involving a generator according to an embodiment of the present application;
图10为本申请实施例涉及生成器的训练过程的另一种示意图;FIG. 10 is another schematic diagram of the training process involving the generator according to the embodiment of the present application;
图11为判别器的一种网络结构示意图;FIG. 11 is a schematic diagram of a network structure of a discriminator;
图12为本申请一实施例提供的模型训练方法的流程示意图;FIG. 12 is a schematic flowchart of a model training method provided by an embodiment of the present application;
图13为本申请实施例提供的判别器的一种具体网络结构示意图;FIG. 13 is a schematic diagram of a specific network structure of the discriminator provided in the embodiment of the present application;
图14为本申请实施例提供的点云上采样方法的流程示意图;FIG. 14 is a schematic flow diagram of a point cloud upsampling method provided in an embodiment of the present application;
图15为本申请实施例涉及的生成器的一种网络结构示意图;FIG. 15 is a schematic diagram of a network structure of a generator involved in an embodiment of the present application;
图16为本申请实施例提供的点云解码方法的流程示意图;FIG. 16 is a schematic flow diagram of a point cloud decoding method provided in an embodiment of the present application;
图17是本申请实施例提供的点云解码器的示意性框图;Fig. 17 is a schematic block diagram of a point cloud decoder provided by an embodiment of the present application;
图18是本申请实施例提供的点云上采样装置的示意性框图;Fig. 18 is a schematic block diagram of a point cloud upsampling device provided by an embodiment of the present application;
图19是本申请实施例提供的模型训练装置的示意性框图;Fig. 19 is a schematic block diagram of a model training device provided by an embodiment of the present application;
图20是本申请实施例提供的电子设备的示意性框图。Fig. 20 is a schematic block diagram of an electronic device provided by an embodiment of the present application.
具体实施方式Detailed ways
本申请可应用于点云上采样技术领域,例如可以应用于点云压缩技术领域。The present application can be applied to the technical field of point cloud upsampling, for example, can be applied to the technical field of point cloud compression.
为了便于理解本申请的实施例,首先对本申请实施例涉及到的相关概念进行如下简单介绍:In order to facilitate the understanding of the embodiments of the present application, firstly, the relevant concepts involved in the embodiments of the present application are briefly introduced as follows:
点云(Point Cloud)是指空间中一组无规则分布的、表达三维物体或三维场景的空间结构及表面属性的离散点集。Point cloud refers to a set of discrete point sets randomly distributed in space, expressing the spatial structure and surface properties of 3D objects or 3D scenes.
点云数据(Point Cloud Data)是点云的具体记录形式,点云中的点可以包括点的位置信息和点的属性信息。例如,点的位置信息可以是点的三维坐标信息。点的位置信息也可称为点的几何信息。例如,点的属性信息可包括颜色信息和/或反射率等等。例如,所述颜色信息可以是任意一种色彩空间上的信息。例如,所述颜色信息可以是(RGB)。再如,所述颜色信息可以是于亮度色度(YcbCr,YUV)信息。例如,Y表示明亮度(Luma),Cb(U)表示蓝色色差,Cr(V)表示红色,U和V表示为色度(Chroma)用于描述色差信息。例如,根据激光测量原理得到的点云,所述点云中的点可以包括点的三维坐标信息和点的激光反射强度(reflectance)。再如,根据摄影测量原理得到的点云,所述点云中的点可以可包括点的三维坐标信息和点的颜色信息。再如,结合激光测量和摄影测量原理得到点云,所述点云中的点可以可包括点的三维坐标信息、点的激光反射强度(reflectance)和点的颜色信息。Point cloud data (Point Cloud Data) is a specific record form of point cloud, and the points in the point cloud can include point location information and point attribute information. For example, the point position information may be three-dimensional coordinate information of the point. The location information of a point may also be referred to as geometric information of a point. For example, the attribute information of a point may include color information and/or reflectivity and the like. For example, the color information may be information on any color space. For example, the color information may be (RGB). For another example, the color information may be luminance and chrominance (YcbCr, YUV) information. For example, Y represents brightness (Luma), Cb (U) represents blue color difference, Cr (V) represents red color, and U and V are expressed as chromaticity (Chroma) for describing color difference information. For example, according to the point cloud obtained according to the principle of laser measurement, the points in the point cloud may include the three-dimensional coordinate information of the point and the laser reflection intensity (reflectance) of the point. For another example, in the point cloud obtained according to the principle of photogrammetry, the points in the point cloud may include three-dimensional coordinate information and color information of the point. For another example, combining the principles of laser measurement and photogrammetry to obtain a point cloud, the points in the point cloud may include the three-dimensional coordinate information of the point, the laser reflection intensity (reflectance) of the point, and the color information of the point.
点云数据的获取途径可以包括但不限于以下至少一种:(1)计算机设备生成。计算机设备可以根据虚拟三维物体及虚拟三维场景的生成点云数据。(2)3D(3-Dimension,三维)激光扫描获取。通过3D激光扫描可以获取静态现实世界三维物体或三维场景的点云数据,每秒可以获取百万级点云数据;(3)3D摄影测量获取。通过3D摄影设备(即一组摄像机或具有多个镜头和传感器的摄像机设备)对现实世界的视觉场景进行采集以获取现实世界的视觉场景的点云数据,通过3D摄影可以获得动态现实世界三维物体或三维场景的点云数据。(4)通过医学设备获取生物组织器官的点云数据。在医学领域可以通过磁共振成像(Magnetic Resonance Imaging,MRI)、电子计算机断层扫描(Computed Tomography,CT)、电磁定位信息等医学设备获取生物组织器官的点云数据。Ways to obtain point cloud data may include but not limited to at least one of the following: (1) Generated by computer equipment. The computer device can generate point cloud data according to virtual three-dimensional objects and virtual three-dimensional scenes. (2) 3D (3-Dimension, three-dimensional) laser scanning acquisition. Point cloud data of static real-world 3D objects or 3D scenes can be obtained through 3D laser scanning, and millions of point cloud data can be obtained per second; (3) 3D photogrammetry acquisition. Through 3D photography equipment (that is, a group of cameras or camera equipment with multiple lenses and sensors) to collect the visual scene of the real world to obtain the point cloud data of the visual scene of the real world, through 3D photography can obtain dynamic real world three-dimensional objects Or point cloud data of a 3D scene. (4) Obtain point cloud data of biological tissues and organs through medical equipment. In the medical field, point cloud data of biological tissues and organs can be obtained through magnetic resonance imaging (Magnetic Resonance Imaging, MRI), electronic computer tomography (Computed Tomography, CT), electromagnetic positioning information and other medical equipment.
点云可以按获取的途径分为:密集型点云和稀疏性点云。Point clouds can be divided into dense point clouds and sparse point clouds according to the way of acquisition.
点云按照数据的时序类型划分为:According to the time series type of data, point cloud is divided into:
第一类静态点云:即物体是静止的,获取点云的设备也是静止的;The first type of static point cloud: that is, the object is stationary, and the device for obtaining the point cloud is also stationary;
第二类动态点云:物体是运动的,但获取点云的设备是静止的;The second type of dynamic point cloud: the object is moving, but the device for obtaining the point cloud is still;
第三类动态获取点云:获取点云的设备是运动的。The third type of dynamic acquisition of point clouds: the equipment for acquiring point clouds is in motion.
按点云的用途分为两大类:According to the purpose of point cloud, it can be divided into two categories:
类别一:机器感知点云,其可以用于自主导航系统、实时巡检系统、地理信息系统、视觉分拣机器人、抢险救灾机器人等场景;Category 1: Machine perception point cloud, which can be used in scenarios such as autonomous navigation systems, real-time inspection systems, geographic information systems, visual sorting robots, and emergency rescue robots;
类别二:人眼感知点云,其可以用于数字文化遗产、自由视点广播、三维沉浸通信、三维沉浸交互等点云应用场景。Category 2: Human eyes perceive point clouds, which can be used in point cloud application scenarios such as digital cultural heritage, free viewpoint broadcasting, 3D immersive communication, and 3D immersive interaction.
在一些实施例中,本申请实施例提供的点云上采样方法,可以应用于点云编解码框架中,例如对点云解码端从码流中解析出点云的几何信息进行上采样,得到精度更高的上采样点云。In some embodiments, the point cloud upsampling method provided by the embodiment of the present application can be applied to the point cloud encoding and decoding framework, for example, the geometric information of the point cloud parsed from the code stream by the point cloud decoder is upsampled to obtain Upsampled point clouds with higher accuracy.
下面对点云编解码的相关知识进行介绍。The following is an introduction to the relevant knowledge of point cloud encoding and decoding.
图1为本申请实施例涉及的一种点云编解码系统的示意性框图。需要说明的是,图1只是一种示例,本申请实施例的点云编解码系统包括但不限于图1所示。如图1所示,该点云编解码系统100包含编码设备110和解码设备120。其中编码设备用于对点云数据进行编码(可以理解成压缩)产生码流,并将码流传输给解码设备。解码设备对编码设备编码产生的码流进行解码,得到解码后的点云数据。FIG. 1 is a schematic block diagram of a point cloud encoding and decoding system involved in an embodiment of the present application. It should be noted that FIG. 1 is just an example, and the point cloud encoding and decoding system in the embodiment of the present application includes but is not limited to what is shown in FIG. 1 . As shown in FIG. 1 , the point cloud encoding and decoding system 100 includes an encoding device 110 and a decoding device 120 . The encoding device is used to encode the point cloud data (which can be understood as compression) to generate a code stream, and transmit the code stream to the decoding device. The decoding device decodes the code stream generated by the encoding device to obtain decoded point cloud data.
本申请实施例的编码设备110可以理解为具有点云编码功能的设备,解码设备120可以理解为具有点云解码功能的设备,即本申请实施例对编码设备110和解码设备120包括更广泛的装置,例如包含智能手机、台式计算机、移动计算装置、笔记本(例如,膝上型)计算机、平板计算机、机顶盒、电视、相机、显示装置、数字媒体播放器、点云游戏控制台、车载计算机等。The encoding device 110 in the embodiment of the present application can be understood as a device having a point cloud encoding function, and the decoding device 120 can be understood as a device having a point cloud decoding function. Devices, including, for example, smartphones, desktop computers, mobile computing devices, notebook (e.g., laptop) computers, tablet computers, set-top boxes, televisions, cameras, display devices, digital media players, point cloud gaming consoles, vehicle-mounted computers, etc. .
在一些实施例中,编码设备110可以经由信道130将编码后的点云数据(如码流)传输给解码设备120。信道130可以包括能够将编码后的点云数据从编码设备110传输到解码设备120的一个或多个媒体和/或装置。In some embodiments, the encoding device 110 can transmit the encoded point cloud data (eg code stream) to the decoding device 120 via the channel 130 . Channel 130 may include one or more media and/or devices capable of transmitting encoded point cloud data from encoding device 110 to decoding device 120 .
在一个实例中,信道130包括使编码设备110能够实时地将编码后的点云数据直接发射到解码设备120的一个或多个通信媒体。在此实例中,编码设备110可根据通信标准来调制编码后的点云数据,且将调制后的点云数据发射到解码设备120。其中通信媒体包含无线通信媒体,例如射频频谱,可选的,通信媒体还可以包含有线通信媒体,例如一根或多根物理传输线。In one example, channel 130 includes one or more communication media that enable encoding device 110 to transmit encoded point cloud data directly to decoding device 120 in real-time. In this instance, the encoding device 110 may modulate the encoded point cloud data according to the communication standard, and transmit the modulated point cloud data to the decoding device 120 . The communication medium includes a wireless communication medium, such as a radio frequency spectrum. Optionally, the communication medium may also include a wired communication medium, such as one or more physical transmission lines.
在另一实例中,信道130包括存储介质,该存储介质可以存储编码设备110编码后的点云数据。存储介质包含多种本地存取式数据存储介质,例如光盘、DVD、快闪存储器等。在该实例中,解码设备120可从该存储介质中获取编码后的点云数据。In another example, the channel 130 includes a storage medium, which can store the point cloud data encoded by the encoding device 110 . The storage medium includes a variety of local access data storage media, such as optical discs, DVDs, flash memory, and the like. In this example, the decoding device 120 can acquire encoded point cloud data from the storage medium.
在另一实例中,信道130可包含存储服务器,该存储服务器可以存储编码设备110编码后的点云数据。在此实例中,解码设备120可以从该存储服务器中下载存储的编码后的点云数据。可选的,该存储服务器可以存储编码后的点云数据且可以将该编码后的点云数据发射到解码设备120,例如web服务器(例如,用于网站)、文件传送协议(FTP)服务器等。In another example, the channel 130 may include a storage server, and the storage server may store the point cloud data encoded by the encoding device 110 . In this instance, the decoding device 120 may download the stored encoded point cloud data from the storage server. Optionally, the storage server can store the encoded point cloud data and can transmit the encoded point cloud data to the decoding device 120, such as a web server (for example, for a website), a file transfer protocol (FTP) server, etc. .
一些实施例中,编码设备110包含点云编码器112及输出接口113。其中,输出接口113可以包含调制器/解调器(调制解调器)和/或发射器。In some embodiments, the encoding device 110 includes a point cloud encoder 112 and an output interface 113 . Wherein, the output interface 113 may include a modulator/demodulator (modem) and/or a transmitter.
在一些实施例中,编码设备110除了包括点云编码器112和输入接口113外,还可以包括点云源111。In some embodiments, the encoding device 110 may include a point cloud source 111 in addition to the point cloud encoder 112 and the input interface 113 .
点云源111可包含点云采集装置(例如,扫描仪)、点云存档、点云输入接口、计算机图形系统中的至少一个,其中,点云输入接口用于从点云内容提供者处接收点云数据,计算机图形系统用于产生点云数据。The point cloud source 111 may include at least one of a point cloud acquisition device (for example, a scanner), a point cloud archive, a point cloud input interface, and a computer graphics system, wherein the point cloud input interface is used to receive from a point cloud content provider Point cloud data, computer graphics system is used to generate point cloud data.
点云编码器112对来自点云源111的点云数据进行编码,产生码流。点云编码器112经由输出接口113将编码后的点云数据直接传输到解码设备120。编码后的点云数据还可存储于存储介质或存储服务器上,以供解码设备120后续读取。The point cloud encoder 112 encodes the point cloud data from the point cloud source 111 to generate a code stream. The point cloud encoder 112 directly transmits the encoded point cloud data to the decoding device 120 via the output interface 113 . The encoded point cloud data can also be stored on a storage medium or a storage server for subsequent reading by the decoding device 120 .
在一些实施例中,解码设备120包含输入接口121和点云解码器122。In some embodiments, the decoding device 120 includes an input interface 121 and a point cloud decoder 122 .
在一些实施例中,解码设备120除包括输入接口121和点云解码器122外,还可以包括显示装置123。In some embodiments, the decoding device 120 may further include a display device 123 in addition to the input interface 121 and the point cloud decoder 122 .
其中,输入接口121包含接收器及/或调制解调器。输入接口121可通过信道130接收编码后的点云数据。Wherein, the input interface 121 includes a receiver and/or a modem. The input interface 121 can receive the encoded point cloud data through the channel 130 .
点云解码器122用于对编码后的点云数据进行解码,得到解码后的点云数据,并将解码后的点云数据传输至显示装置123。The point cloud decoder 122 is used to decode the encoded point cloud data to obtain decoded point cloud data, and transmit the decoded point cloud data to the display device 123 .
显示装置123显示解码后的点云数据。显示装置123可与解码设备120整合或在解码设备120外部。显示装置123可包括多种显示装置,例如液晶显示器(LCD)、等离子体显示器、有机发光二极管(OLED)显示器或其它类型的显示装置。The display device 123 displays the decoded point cloud data. The display device 123 may be integrated with the decoding device 120 or external to the decoding device 120 . The display device 123 may include various display devices, such as a liquid crystal display (LCD), a plasma display, an organic light emitting diode (OLED) display, or other types of display devices.
此外,图1仅为实例,本申请实施例的技术方案不限于图1,例如本申请的技术还可以应用于单侧的点云编码或单侧的点云解码。In addition, FIG. 1 is only an example, and the technical solution of the embodiment of the present application is not limited to FIG. 1 . For example, the technology of the present application can also be applied to one-sided point cloud encoding or one-sided point cloud decoding.
目前的点云编码器可以采用运动图像专家组(Moving Picture Experts Group,MPEG)提供的基于几何的点云压缩(Geometry Point Cloud Compression,G-PCC)编解码框架或基于视频的点云压缩(Video Point Cloud Compression,V-PCC)编解码框架,也可以采用音视频编码标准(Audio Video Standard,AVS)提供的AVS-PCC编解码框架。G-PCC及AVS-PCC均针对静态的稀疏型点云,其编码框架大致相同。G-PCC编解码框架可用于针对第一静态点云和第三类动态获取点云进行压缩,V-PCC编解码框架可用于针对第二类动态点云进行压缩。G-PCC编解码框架也称为点云编解码器TMC13,V-PCC编解码框架也称为点云编解码器TMC2。The current point cloud encoder can use the Geometry Point Cloud Compression (G-PCC) codec framework provided by the Moving Picture Experts Group (MPEG) or the video-based point cloud compression (Video Point Cloud Compression, V-PCC) codec framework, or the AVS-PCC codec framework provided by Audio Video Standard (AVS). Both G-PCC and AVS-PCC are aimed at static sparse point clouds, and their coding frameworks are roughly the same. The G-PCC codec framework can be used to compress the first static point cloud and the third type of dynamically acquired point cloud, and the V-PCC codec framework can be used to compress the second type of dynamic point cloud. The G-PCC codec framework is also called point cloud codec TMC13, and the V-PCC codec framework is also called point cloud codec TMC2.
下面以G-PCC编解码框架为例,对本申请实施例可适用的点云编码器和点云解码器进行说明。The following uses the G-PCC codec framework as an example to describe the applicable point cloud encoder and point cloud decoder in this embodiment of the present application.
图2是本申请实施例提供的点云编码器的示意性框图。Fig. 2 is a schematic block diagram of a point cloud encoder provided by an embodiment of the present application.
由上述可知点云中的点可以包括点的位置信息和点的属性信息,因此,点云中的点的编码主要包括位置编码和属性编码。在一些示例中点云中点的位置信息又称为几何信息,对应的点云中点的位置编码也可以称为几何编码。From the above, it can be seen that the points in the point cloud can include the position information of the point and the attribute information of the point, therefore, the encoding of the point in the point cloud mainly includes the position encoding and the attribute encoding. In some examples, the position information of the points in the point cloud is also called geometric information, and the corresponding position codes of the points in the point cloud may also be called geometric codes.
位置编码的过程包括:对点云中的点进行预处理,例如坐标变换、量化和移除重复点等;接着,对预处理后的点云进行几何编码,例如构建八叉树,基于构建的八叉树进行几何编码形成几何码流。同时,基于构建的八叉树输出的位置信息,对点云数据中各点的位置信息进行重建,得到各点的位置信息的重建值。The process of position encoding includes: preprocessing the points in the point cloud, such as coordinate transformation, quantization, and removing duplicate points, etc.; then, geometrically encoding the preprocessed point cloud, such as constructing an octree, based on the constructed The octree performs geometric encoding to form a geometric code stream. At the same time, based on the position information output by the constructed octree, the position information of each point in the point cloud data is reconstructed to obtain the reconstruction value of the position information of each point.
属性编码过程包括:通过给定输入点云的位置信息的重建信息和属性信息的原始值,选择三种预测模式的一种进行点云预测,对预测后的结果进行量化,并进行算术编码形成属性码流。The attribute encoding process includes: by given the reconstruction information of the position information of the input point cloud and the original value of the attribute information, select one of the three prediction modes for point cloud prediction, quantify the predicted results, and perform arithmetic coding to form property stream.
如图2所示,位置编码可通过以下单元实现:As shown in Figure 2, position coding can be achieved by the following units:
坐标转换(Tanmsform coordinates)单元201、量化和移除重复点(Quantize and remove points)单元202、八叉树分析(Analyze octree)单元203、几何重建(Reconstruct geometry)单元204以及第一算术编码(Arithmetic enconde)单元205。Coordinate conversion (Tanmsform coordinates) unit 201, quantization and removal of repeated points (Quantize and remove points) unit 202, octree analysis (Analyze octree) unit 203, geometric reconstruction (Reconstruct geometry) unit 204 and first arithmetic coding (Arithmetic enconde) unit 205.
坐标转换单元201可用于将点云中点的世界坐标变换为相对坐标。例如,点的几何坐标分别减去xyz坐标轴的最小值,相当于去直流操作,以实现将点云中的点的坐标从世界坐标转换为相对坐标。The coordinate transformation unit 201 can be used to transform the world coordinates of points in the point cloud into relative coordinates. For example, subtracting the minimum values of the xyz coordinate axes from the geometric coordinates of the point is equivalent to a DC operation to convert the coordinates of the points in the point cloud from world coordinates to relative coordinates.
量化和移除重复点单元202可通过量化减少坐标的数目;量化后原先不同的点可能被赋予相同的坐标,基于此,可通过去重操作将重复的点删除;例如,具有相同量化位置和不同属性信息的多个云可通过属性转换合并到一个云中。在本申请的一些实施例中,量化和移除重复点单元202为可选的单元模块。Quantization and removal of duplicate points unit 202 can reduce the number of coordinates by quantization; original different points may be given the same coordinates after quantization, based on this, duplicate points can be deleted through de-duplication operations; for example, with the same quantization position and Multiple clouds with different attribute information can be merged into one cloud through attribute conversion. In some embodiments of the present application, the Quantize and Remove Duplicate Points unit 202 is an optional unit module.
八叉树分析单元203可利用八叉树(octree)编码方式编码量化的点的位置信息。例如,将点云按照八叉树的形式进行划分,由此,点的位置可以和八叉树的位置一一对应,通过统计八叉树中有点的位置,并将其标识(flag)记为1,以进行几何编码。The octree analysis unit 203 may use an octree encoding method to encode the position information of the quantized points. For example, the point cloud is divided in the form of an octree, so that the position of the point can be in one-to-one correspondence with the position of the octree, and the position of the point in the octree is counted, and its flag (flag) is recorded as 1 for geometric encoding.
几何重建单元204可以基于八叉树分析单元203输出的位置信息进行位置重建,得到点云数据中各点的位置信息的重建值。The geometry reconstruction unit 204 may perform position reconstruction based on the position information output by the octree analysis unit 203 to obtain reconstruction values of the position information of each point in the point cloud data.
第一算术编码单元205可以采用熵编码方式对八叉树分析单元203输出的位置信息进行算术编码,即将八叉树分析单元203输出的位置信息利用算术编码方式生成几何码流;几何码流也可称为几何比特流(geometry bitstream)。The first arithmetic coding unit 205 can arithmetically encode the position information output by the octree analysis unit 203 in an entropy coding manner, that is, the position information output by the octree analysis unit 203 is generated using an arithmetic coding method to generate a geometric code stream; the geometric code stream is also Can be called geometry bitstream (geometry bitstream).
属性编码可通过以下单元实现:Attribute coding can be achieved by the following units:
颜色空间转换(Transform colors)单元210、属性转化(Transfer attributes)单元211、区域自适应分层变换(Region Adaptive Hierarchical Transform,RAHT)单元212、预测变化(predicting transform)单元213以及提升变化(lifting transform)单元214、量化系数(Quantize coefficients)单元215以及第二算术编码单元216。Color space conversion (Transform colors) unit 210, attribute conversion (Transfer attributes) unit 211, region adaptive layered transformation (Region Adaptive Hierarchical Transform, RAHT) unit 212, prediction change (predicting transform) unit 213 and lifting transform (lifting transform) ) unit 214, a quantization coefficient (Quantize coefficients) unit 215, and a second arithmetic coding unit 216.
需要说明的是,点云编码器200可包含比图2更多、更少或不同的功能组件。It should be noted that the point cloud encoder 200 may include more, less or different functional components than those shown in FIG. 2 .
颜色空间转换单元210可用于将点云中点的RGB色彩空间变换为YCbCr格式或其他格式。The color space conversion unit 210 can be used to convert the RGB color space of points in the point cloud into YCbCr format or other formats.
属性转化单元211可用于转换点云中点的属性信息,以最小化属性失真。例如,属性转化单元211可用于得到点的属性信息的原始值。例如,所述属性信息可以是点的颜色信息。The attribute conversion unit 211 can be used to convert attribute information of points in the point cloud to minimize attribute distortion. For example, the attribute conversion unit 211 can be used to obtain the original value of the attribute information of the point. For example, the attribute information may be color information of dots.
经过属性转化单元211转换得到点的属性信息的原始值后,可选择任一种预测单元,对点云中的点进行预测。预测单元可包括:RAHT 212、预测变化(predicting transform)单元213以及提升变化 (lifting transform)单元214。After the original value of the attribute information of the point is converted by the attribute conversion unit 211, any prediction unit can be selected to predict the point in the point cloud. The prediction unit may include: RAHT 212, predicting transform unit 213, and lifting transform unit 214.
换言之,RAHT 212、预测变化(predicting transform)单元213以及提升变化(lifting transform)单元214中的任一项可用于对点云中点的属性信息进行预测,以得到点的属性信息的预测值,进而基于点的属性信息的预测值得到点的属性信息的残差值。例如,点的属性信息的残差值可以是点的属性信息的原始值减去点的属性信息的预测值。In other words, any one of the RAHT 212, the predicting transform unit 213 and the lifting transform unit 214 can be used to predict the attribute information of the points in the point cloud, so as to obtain the predicted values of the attribute information of the points, Furthermore, the residual value of the attribute information of the point is obtained based on the predicted value of the attribute information of the point. For example, the residual value of the point's attribute information may be the original value of the point's attribute information minus the predicted value of the point's attribute information.
在本申请的一实施例中,预测变换单元213还可用于生成细节层(level of detail,LOD)。LOD的生成过程包括:根据点云中点的位置信息,获取点与点之间的欧式距离;根据欧式距离,将点分为不同的细节表达层。在一个实施例中,可以将欧式距离进行排序后,将不同范围的欧式距离划分为不同的细节表达层。例如,可以随机挑选一个点,作为第一细节表达层。然后计算剩余点与该点的欧式距离,并将欧式距离符合第一阈值要求的点,归为第二细节表达层。获取第二细节表达层中点的质心,计算除第一、第二细节表达层以外的点与该质心的欧式距离,并将欧式距离符合第二阈值的点,归为第三细节表达层。以此类推,将所有的点都归到细节表达层中。通过调整欧式距离的阈值,可以使得每层LOD层的点的数量是递增的。应理解,LOD划分的方式还可以采用其它方式,本申请对此不进行限制。In an embodiment of the present application, the predictive transformation unit 213 may also be used to generate a level of detail (LOD). The generation process of LOD includes: according to the position information of the points in the point cloud, the Euclidean distance between the points is obtained; according to the Euclidean distance, the points are divided into different detail expression layers. In one embodiment, after sorting the Euclidean distances, the Euclidean distances in different ranges can be divided into different detail expression layers. For example, a point can be randomly selected as the first detail expression layer. Then calculate the Euclidean distance between the remaining points and this point, and classify the points whose Euclidean distance meets the first threshold requirement as the second detailed expression layer. Obtain the centroid of the point in the second detail expression layer, calculate the Euclidean distance between points other than the first and second detail expression layer and the centroid, and classify the points whose Euclidean distance meets the second threshold as the third detail expression layer. By analogy, all points are classified into the detail expression layer. By adjusting the threshold of the Euclidean distance, the number of points in each LOD layer can be increased. It should be understood that other manners may also be used for LOD division, which is not limited in this application.
需要说明的是,可以直接将点云划分为一个或多个细节表达层,也可以先将点云划分为多个点云切块(slice),再将每一个点云切块划分为一个或多个LOD层。It should be noted that the point cloud can be directly divided into one or more detail expression layers, or the point cloud can be divided into multiple point cloud slices first, and then each point cloud slice can be divided into one or more Multiple LOD layers.
例如,可将点云划分为多个点云切块,每个点云切块的点的个数可以在55万-110万之间。每个点云切块可看成单独的点云。每个点云切块又可以划分为多个细节表达层,每个细节表达层包括多个点。在一个实施例中,可根据点与点之间的欧式距离,进行细节表达层的划分。For example, the point cloud can be divided into multiple point cloud cutouts, and the number of points in each point cloud cutout can be between 550,000 and 1.1 million. Each point cloud slice can be regarded as a separate point cloud. Each point cloud slice can be divided into multiple detail expression layers, and each detail expression layer includes multiple points. In one embodiment, the detail expression layer can be divided according to the Euclidean distance between points.
量化单元215可用于量化点的属性信息的残差值。例如,若所述量化单元215和所述预测变换单元213相连,则所述量化单元可用于量化所述预测变换单元213输出的点的属性信息的残差值。The quantization unit 215 may be used to quantize residual values of attribute information of points. For example, if the quantization unit 215 is connected to the predictive transformation unit 213, the quantization unit may be used to quantize the residual value of the point attribute information output by the predictive transformation unit 213.
例如,对预测变换单元213输出的点的属性信息的残差值使用量化步长进行量化,以实现提升系统性能。For example, the residual value of the point attribute information output by the predictive transformation unit 213 is quantized using the quantization step size, so as to improve system performance.
第二算术编码单元216可使用零行程编码(Zero run length coding)对点的属性信息的残差值进行熵编码,以得到属性码流。所述属性码流可以是比特流信息。The second arithmetic coding unit 216 may use zero run length coding to perform entropy coding on the residual value of the attribute information of the point to obtain an attribute code stream. The attribute code stream may be bit stream information.
图3是本申请实施例提供的点云解码器的示意性框图。Fig. 3 is a schematic block diagram of a point cloud decoder provided by an embodiment of the present application.
如图3所示,解码器300可以从编码设备获取点云码流,通过解析码得到点云中的点的位置信息和属性信息。点云的解码包括位置解码和属性解码。As shown in FIG. 3 , the decoder 300 can obtain the point cloud code stream from the encoding device, and obtain the position information and attribute information of the points in the point cloud by parsing the code. The decoding of point cloud includes position decoding and attribute decoding.
位置解码的过程包括:对几何码流进行算术解码;构建八叉树后进行合并,对点的位置信息进行重建,以得到点的位置信息的重建信息;对点的位置信息的重建信息进行坐标变换,得到点的位置信息。点的位置信息也可称为点的几何信息。The process of position decoding includes: performing arithmetic decoding on the geometric code stream; merging after constructing the octree, and reconstructing the position information of the point to obtain the reconstruction information of the position information of the point; Transform to get the position information of the point. The location information of a point may also be referred to as geometric information of a point.
属性解码过程包括:通过解析属性码流,获取点云中点的属性信息的残差值;通过对点的属性信息的残差值进行反量化,得到反量化后的点的属性信息的残差值;基于位置解码过程中获取的点的位置信息的重建信息,选择如下RAHT、预测变化和提升变化三种预测模式中的一种进行点云预测,得到预测值,预测值与残差值相加得到点的属性信息的重建值;对点的属性信息的重建值进行颜色空间反转化,以得到解码点云。The attribute decoding process includes: obtaining the residual value of the attribute information of the point cloud by parsing the attribute code stream; dequantizing the residual value of the attribute information of the point to obtain the residual value of the attribute information of the dequantized point value; based on the reconstruction information of the position information of the point obtained in the position decoding process, select one of the following three prediction modes: RAHT, prediction change and promotion change to predict the point cloud, and obtain the predicted value, which is consistent with the residual value Add the reconstruction value of the attribute information of the point; perform color space inverse transformation on the reconstruction value of the attribute information of the point to obtain the decoded point cloud.
如图3所示,位置解码可通过以下单元实现:As shown in Figure 3, position decoding can be achieved by the following units:
第一算数解码单元301、八叉树分析(synthesize octree)单元302、几何重建(Reconstruct geometry)单元304以及坐标反转换(inverse transform coordinates)单元305。A first arithmetic decoding unit 301 , an octree analysis unit 302 , a geometry reconstruction unit 304 and an inverse transform coordinates unit 305 .
属性编码可通过以下单元实现:Attribute coding can be achieved by the following units:
第二算数解码单元310、反量化(inverse quantize)单元311、RAHT单元312、预测变化(predicting transform)单元313、提升变化(lifting transform)单元314以及颜色空间反转换(inverse trasform colors)单元315。A second arithmetic decoding unit 310, an inverse quantize unit 311, an RAHT unit 312, a predicting transform unit 313, a lifting transform unit 314, and an inverse trasform colors unit 315.
需要说明的是,解压缩是压缩的逆过程,类似的,解码器300中的各个单元的功能可参见编码器200中相应的单元的功能。另外,点云解码器300可包含比图3更多、更少或不同的功能组件。It should be noted that decompression is an inverse process of compression, and similarly, the functions of each unit in the decoder 300 may refer to the functions of corresponding units in the encoder 200 . In addition, the point cloud decoder 300 may include more, fewer or different functional components than in FIG. 3 .
例如,解码器300可根据点云中点与点之间的欧式距离将点云划分为多个LOD;然后,依次对 LOD中点的属性信息进行解码;例如,计算零行程编码技术中零的数量(zero_cnt),以基于zero_cnt对残差进行解码;接着,解码框架200可基于解码出的残差值进行反量化,并基于反量化后的残差值与当前点的预测值相加得到该点云的重建值,直到解码完所有的点云。当前点将会作为后续LOD中点的最邻近点,并利用当前点的重建值对后续点的属性信息进行预测。For example, the decoder 300 can divide the point cloud into multiple LODs according to the Euclidean distance between points in the point cloud; then, decode the attribute information of the points in the LOD in sequence; number (zero_cnt), to decode the residual based on zero_cnt; then, the decoding framework 200 can perform dequantization based on the decoded residual value, and add the dequantized residual value to the predicted value of the current point to obtain the Point cloud reconstruction values until all point clouds are decoded. The current point will be used as the nearest neighbor point of the subsequent LOD midpoint, and the attribute information of the subsequent point will be predicted by using the reconstructed value of the current point.
上述是基于G-PCC编解码框架下的点云编解码器的基本流程,随着技术的发展,该框架或流程的一些模块或步骤可能会被优化,本申请适用于该基于G-PCC编解码框架下的点云编解码器的基本流程,但不限于该框架及流程。The above is the basic process of the point cloud codec based on the G-PCC codec framework. With the development of technology, some modules or steps of the framework or process may be optimized. This application is applicable to the G-PCC codec-based The basic process of the point cloud codec under the decoding framework, but not limited to the framework and process.
目前的点云编解码方式,将点云重建为原始尺度,但是在一些应用场景中,需要使用比原始精度更高的高质量点云,例如,在自动驾驶等领域对雷达采集的稀疏点云,经常需要做大量的后处理工作,来提升点云的精度,以提升驾驶的安全性。The current point cloud encoding and decoding method reconstructs the point cloud to the original scale, but in some application scenarios, it is necessary to use high-quality point clouds with higher precision than the original ones, for example, sparse point clouds collected by radar in areas such as autonomous driving , often need to do a lot of post-processing work to improve the accuracy of the point cloud to improve driving safety.
本申请实施例提供一种点云上采样方法,使用深度学习来上采样点云几何信息以获取更高分辨率(或精度)的点云,进而满足对高精度点云的任务需求。The embodiment of the present application provides a point cloud upsampling method, which uses deep learning to upsample point cloud geometric information to obtain a higher resolution (or precision) point cloud, thereby meeting the task requirements for high-precision point clouds.
下面结合具体的实施例,对本申请实施例涉及的点云上采样方法进行介绍。The point cloud upsampling method involved in the embodiment of the present application will be introduced below in combination with specific embodiments.
本申请提供的点云上采样方法是使用深度学习后的生成器来进行点云的几何信息进行上采样的,该生成器为一段软件代码或者为具有数据处理功能的芯片。基于此,首先对生成器的训练过程进行介绍。The point cloud upsampling method provided in this application uses a generator after deep learning to upsample the geometric information of the point cloud, and the generator is a piece of software code or a chip with data processing functions. Based on this, the training process of the generator is firstly introduced.
图4为本申请一实施例提供的模型训练方法流程示意图,如图4所示,生成器的训练过程包括:Fig. 4 is a schematic flow chart of the model training method provided by an embodiment of the present application. As shown in Fig. 4, the training process of the generator includes:
S401、获取训练点云的几何信息。S401. Obtain geometric information of the training point cloud.
需要说明的是,为了便于描述,本申请实施例将用于生成器训练的点云记为训练点云。It should be noted that, for ease of description, the embodiment of the present application records the point cloud used for generator training as a training point cloud.
上述训练点云为训练集中的一个点云,该训练集中包括多个点云,其中使用训练集中的每个点云对生成器进行训练的过程一致,为了便于描述,本申请实施例以一个训练点云为例。The above-mentioned training point cloud is a point cloud in the training set, which includes multiple point clouds, and the process of using each point cloud in the training set to train the generator is consistent. For the convenience of description, the embodiment of the present application uses a training set Take point clouds as an example.
S402、根据训练点云的几何信息,将训练点云划分成至少一个训练点云块。S402. Divide the training point cloud into at least one training point cloud block according to the geometric information of the training point cloud.
本申请实施例在点云上采样的过程中,是将点云划分成点云块,以点云块为对象进行点云几何信息的上采样。In the embodiment of the present application, in the process of up-sampling the point cloud, the point cloud is divided into point cloud blocks, and the point cloud geometric information is up-sampled with the point cloud blocks as objects.
在一些实施例中,上述S402中将训练点云划分为至少一个训练点云块的方式包括但不限于如下几种方式:In some embodiments, the ways of dividing the training point cloud into at least one training point cloud block in S402 include but are not limited to the following ways:
方式一,根据训练点云的几何信息,将训练点云划分成至少一个大小相等的训练点云块。也就是说每个点云块的几何尺度相同。Method 1: Divide the training point cloud into at least one training point cloud block of equal size according to the geometric information of the training point cloud. That is to say, the geometric scale of each point cloud block is the same.
方式二,根据训练点云的几何信息,将训练点云划分为至少一个训练点云块,每个训练点云块中包括相同数量个点。Method 2: Divide the training point cloud into at least one training point cloud block according to the geometric information of the training point cloud, and each training point cloud block includes the same number of points.
方式三,根据训练点云的几何信息,从训练点云中获取至少一个种子点,例如采用蒙特卡洛随机采样法随机地从训练点云中采样指定个数的种子点。对于每个种子点,确定该种子点的邻近点,将该种子点与该种子点的邻近点划分为一个训练点云块,得到至少一个训练点云块。在该方式三中,得到训练点云块也称为点云补丁(Patch),该方式得到的训练点云块中每个训练点云块所包括的点的个数相同。Method 3: Obtain at least one seed point from the training point cloud according to the geometric information of the training point cloud, for example, randomly sample a specified number of seed points from the training point cloud by using Monte Carlo random sampling method. For each seed point, determine the neighboring points of the seed point, divide the seed point and the neighboring points of the seed point into a training point cloud block, and obtain at least one training point cloud block. In the third way, the obtained training point cloud blocks are also called point cloud patches (Patch), and the number of points included in each training point cloud block in the training point cloud blocks obtained in this way is the same.
在一些实施例中,将上述得到的训练点云块记为
Figure PCTCN2021096287-appb-000001
其中N为训练点云块中所包括的点的个数,3为训练点云块的几何信息维度。
In some embodiments, the training point cloud blocks obtained above are recorded as
Figure PCTCN2021096287-appb-000001
Where N is the number of points included in the training point cloud block, and 3 is the geometric information dimension of the training point cloud block.
S403、将训练点云块的几何信息输入生成器的特征提取模块进行特征提取,得到训练点云块的第一特征信息。S403. Input the geometric information of the training point cloud block into the feature extraction module of the generator to perform feature extraction, and obtain the first feature information of the training point cloud block.
下面结合图5对本申请实施例涉及的生成器的网络结构进行介绍,需要说明的是,本申请实施例的生成器的网络结构包括但不限于图5所示的模块,还可以包括比图5更多或更少的模块。The network structure of the generator involved in the embodiment of the present application will be introduced below in conjunction with FIG. 5. It should be noted that the network structure of the generator in the embodiment of the present application includes but is not limited to the modules shown in FIG. more or less modules.
图5为本申请实施例的生成器的一种网络示意图,如图5所示,生成器包括特征提取模块、特征上采样模块和几何生成模块,特征提取模块用于提取训练点云块的第一特征信息,特征采样模块用于将训练点云块的第一特征信息上采样为第二特征信息,几何生成模块用于将训练点云块的第二特征信息映射至几何空间中,以得到训练点云块的上采样几何信息。Fig. 5 is a kind of network schematic diagram of the generator of the embodiment of the present application, as shown in Fig. 5, generator includes feature extraction module, feature upsampling module and geometry generation module, and feature extraction module is used to extract the first of training point cloud block One feature information, the feature sampling module is used to upsample the first feature information of the training point cloud block into the second feature information, and the geometry generation module is used to map the second feature information of the training point cloud block into the geometric space, so as to obtain Upsampled geometric information for training point cloud blocks.
如图5所示,特征提取模块用于提取每个点的有表现力特征,例如将低分分辨率的训练点云块的几何信息
Figure PCTCN2021096287-appb-000002
输入特征提取模块,特征提取模块用于提取训练点云块中每个点有表现力的特征信息,输出训练点云块的特征信息
Figure PCTCN2021096287-appb-000003
其中N为训练点云块中点的个数,3为几何信息维度,C为特征维度。
As shown in Figure 5, the feature extraction module is used to extract the expressive features of each point, such as the geometric information of the low-resolution training point cloud block
Figure PCTCN2021096287-appb-000002
Input the feature extraction module, the feature extraction module is used to extract the expressive feature information of each point in the training point cloud block, and output the feature information of the training point cloud block
Figure PCTCN2021096287-appb-000003
Among them, N is the number of points in the training point cloud block, 3 is the geometric information dimension, and C is the feature dimension.
为了便于描述,将特征提取模块输出的特征信息记为训练点云块的第一特征信息。For the convenience of description, the feature information output by the feature extraction module is recorded as the first feature information of the training point cloud block.
在一些实施例中,本申请提出了一个基于动态图分层残差聚合(DGHRA)单元的特征提取模块,如图6A所示,上述特征提取模块包括:M个密集连接的特征提取块(Feature Extraction Block,简称FEB),即前一个FEB的输出作为后面每一个FEB的输入,具体的如图6A所示。In some embodiments, the present application proposes a feature extraction module based on a Dynamic Graph Hierarchical Residual Aggregation (DGHRA) unit, as shown in FIG. 6A , the feature extraction module includes: M densely connected feature extraction blocks (Feature Extraction Block, referred to as FEB), that is, the output of the previous FEB is used as the input of each subsequent FEB, as shown in Figure 6A.
在一些实施例中,上述S403包括如下S403-A1至S403-A4:In some embodiments, the above S403 includes the following S403-A1 to S403-A4:
S403-A1、将训练点云块的几何信息输入特征提取模块中,获取M个特征提取块中第i个特征提取块所提取的训练点云块的第i个第三特征信息。S403-A1. Input the geometric information of the training point cloud block into the feature extraction module, and acquire the i-th third feature information of the training point cloud block extracted by the i-th feature extraction block among the M feature extraction blocks.
其中,i为小于M的正整数。Wherein, i is a positive integer smaller than M.
在一些实施例中,若i=1,则本申请实施例还包括:根据训练点云块的几何信息,确定训练点云块的初始特征信息;将训练点云块的初始特征信息输入第一个特征提取块中,得到第一特征提取块所提取的训练点云块的第一个第三特征信息。In some embodiments, if i=1, the embodiment of the present application further includes: determining the initial feature information of the training point cloud block according to the geometric information of the training point cloud block; inputting the initial feature information of the training point cloud block into the first In the first feature extraction block, the first third feature information of the training point cloud block extracted by the first feature extraction block is obtained.
S403-A2、根据训练点云块的第i个第三特征信息,得到训练点云块的第i个第四特征信息。S403-A2. Obtain the i-th fourth feature information of the training point cloud block according to the i-th third feature information of the training point cloud block.
在一种可能的实现方式中,若i不等于1,则上述S403-A2包括:获取M个特征提取块中位于第i个特征提取块之前的各特征提取块所提取的第三特征信息;将位于第i个特征提取块之前的各特征提取块所提取的第三特征信息、与第i个特征提取块所提取的第三特征信息进行级联,作为训练点云块的第i个第四特征信息。In a possible implementation, if i is not equal to 1, the above S403-A2 includes: acquiring the third feature information extracted by each feature extraction block before the i-th feature extraction block among the M feature extraction blocks; The third feature information extracted by each feature extraction block located before the i-th feature extraction block is concatenated with the third feature information extracted by the i-th feature extraction block, as the i-th feature information of the training point cloud block Four feature information.
在一种可能的实现方式中,若i等于1,则M个特征提取单元中第一特征提取块所提取的第一个第三特征信息,作为训练点云块的第一个第四特征信息。In a possible implementation, if i is equal to 1, the first third feature information extracted by the first feature extraction block in the M feature extraction units is used as the first fourth feature information of the training point cloud block .
S403-A3、将训练点云块的第i个第四特征信息输入第i+1个特征提取块中,得到训练点云块的第i+1个第三特征信息。S403-A3. Input the ith fourth feature information of the training point cloud block into the i+1th feature extraction block to obtain the i+1th third feature information of the training point cloud block.
S403-A4、将训练点云块的第M个特征提取块所提取的第M个第三特征信息,作为训练点云块的第一特征信息。S403-A4. Use the Mth third feature information extracted by the Mth feature extraction block of the training point cloud block as the first feature information of the training point cloud block.
举例说明,假设M=4,即特征提取模块包括4个特征提取块FEB,如图6A所示,首先根据训练点云块的几何信息,确定训练点云块的初始特征信息;将训练点云块的初始特征信息输入第一个特征提取块中,得到第一特征提取块所提取的训练点云块的第一个第三特征信息。由于第一个特征提取块之前没有其他的特征提取块,因此将第一个第三特征信息作为第一个第四特征信息输入第二个FEB,第二个FEB根据第一个第四特征信息输出第二个第三特征信息。将第二个第三特征信息和第一个第三特征信息级联,作为第二个第四特征信息,将该第二个第四特征信息输入第三个FEB中,第三个FEB根据第二个第四特征信息输出第三个第三特征信息。将第三个第三特征信息、第二个第三特征信息和第一个第三特征信息级联,作为第三个第四特征信息,将该第三个第四特征信息输入第四个FEB中,第四个FEB根据第三个第四特征信息输出第四个第三特征信息,将该第四个第三特征信息作为训练点云块的第一特征信息。For example, assume that M=4, that is, the feature extraction module includes 4 feature extraction blocks FEB, as shown in Figure 6A, first determine the initial feature information of the training point cloud block according to the geometric information of the training point cloud block; the training point cloud The initial feature information of the block is input into the first feature extraction block to obtain the first third feature information of the training point cloud block extracted by the first feature extraction block. Since there is no other feature extraction block before the first feature extraction block, the first third feature information is input into the second FEB as the first fourth feature information, and the second FEB is based on the first fourth feature information Output the second and third feature information. Concatenate the second third feature information with the first third feature information as the second fourth feature information, input the second fourth feature information into the third FEB, and the third FEB according to the first The second piece of fourth characteristic information outputs the third third characteristic information. Concatenate the third third characteristic information, the second third characteristic information and the first third characteristic information as the third fourth characteristic information, and input the third fourth characteristic information into the fourth FEB Among them, the fourth FEB outputs the fourth third feature information according to the third fourth feature information, and the fourth third feature information is used as the first feature information of the training point cloud block.
在一些实施例中,如图6A所示,随着特征提取模块网络的深入,位于网络深处的FEB的特征维度过多时,为了降低网络的训练复杂度,则在FEB之间设置有卷积网络,例如设置卷积核为1X1的卷积网络,以降低输入FEB中的特征维度。In some embodiments, as shown in Figure 6A, with the deepening of the feature extraction module network, when the feature dimensions of FEBs located deep in the network are too large, in order to reduce the training complexity of the network, convolution is set between FEBs Network, such as a convolutional network with a convolution kernel set to 1X1, to reduce the feature dimension in the input FEB.
在一些实施例中,如图6B所示,每个特征提取块FEB包括第一特征提取单元和串联连接的至少一个第二特征提取单元,其中,第一特征提取单元为动态图分层残差聚合(Dynamic graph hierarchical residual aggregation,DGHRA)单元,第二特征提取单元为分层残差聚合块(Hierarchical residual aggregation,简称HRA),用于提取更多细节特征。各FEB的处理过程相同,且之间是相互迭代,每一个FEB的处理过程相同,为例便于描述,以一个第i+1个特征提取块为例,此时上述S403-A3包括S403-A31至S403-A32:In some embodiments, as shown in FIG. 6B, each feature extraction block FEB includes a first feature extraction unit and at least one second feature extraction unit connected in series, wherein the first feature extraction unit is a dynamic map layered residual Aggregation (Dynamic graph hierarchical residual aggregation, DGHRA) unit, the second feature extraction unit is hierarchical residual aggregation block (Hierarchical residual aggregation, referred to as HRA), used to extract more detailed features. The processing process of each FEB is the same, and they are iterated with each other. The processing process of each FEB is the same. It is convenient to describe as an example. Take the i+1th feature extraction block as an example. At this time, the above S403-A3 includes S403-A31 To S403-A32:
S403-A31、将训练点云块的第i个第四特征信息输入第i+1个特征提取块中的第一特征提取单元,以使第一特征提取单元针对训练点云块中的当前点,搜索当前点的K个邻近点,并基于第i个第四特 征信息,将当前点的第四特征信息与邻近点的第四特征信息进行相减,得到K个残差特征信息;将K个残差特征信息与当前点的第四特征信息进行级联,得到当前点的第i个级联特征信息,并根据当前点的第i个级联特征信息,得到训练点云块的第i个级联特征信息。S403-A31. Input the i-th fourth feature information of the training point cloud block into the first feature extraction unit in the i+1th feature extraction block, so that the first feature extraction unit targets the current point in the training point cloud block , search for K neighboring points of the current point, and based on the ith fourth feature information, subtract the fourth feature information of the current point from the fourth feature information of the neighboring points to obtain K residual feature information; The residual feature information is concatenated with the fourth feature information of the current point to obtain the i-th concatenated feature information of the current point, and according to the i-th concatenated feature information of the current point, the i-th concatenated feature information of the training point cloud block is obtained A cascade of feature information.
例如,训练点云块的第i个第四特征信息的大小为NXC,将大小为NXC的训练点云块的第i个第四特征信息输入第一特征提取单元中,针对训练点云块中的当前点,第一特征提取单元搜索当前点的K个邻近点,例如通过特征空间最近邻搜索方法动态地搜索当前点的K个邻近点。接着,从训练点云块的第i个第四特征信息中获取当前点的第四特征信息,并将当前点的第四特征信息复制K份,将当前点的第四特征信息与K个邻近点中每个邻近点的第四特征信息进行相减,得到K个残差特征信息,大小为1XKXC。将当前点的K个残差特征信息与当前点的第四特征信息进行级联,得到当前点的第i个级联特征信息。参照上述方式可以得到训练点云块中每个点的第i个级联特征信息,进而得到训练点云块的第i个级联特征信息,其大小为NXKX2C。For example, the size of the i-th fourth feature information of the training point cloud block is NXC, and the i-th fourth feature information of the training point cloud block with a size of NXC is input in the first feature extraction unit, for the training point cloud block For the current point of , the first feature extraction unit searches K neighboring points of the current point, for example, dynamically searches the K neighboring points of the current point through a feature space nearest neighbor search method. Next, obtain the fourth feature information of the current point from the ith fourth feature information of the training point cloud block, and copy K copies of the fourth feature information of the current point, and compare the fourth feature information of the current point with K neighboring The fourth feature information of each adjacent point in the point is subtracted to obtain K residual feature information, the size of which is 1XKXC. The K residual feature information of the current point is concatenated with the fourth feature information of the current point to obtain the ith concatenated feature information of the current point. Referring to the above method, the i-th concatenated feature information of each point in the training point cloud block can be obtained, and then the i-th concatenated feature information of the training point cloud block can be obtained, and its size is NXKX2C.
S403-A32、将训练点云块的第i个级联特征信息输入第i+1个特征提取块中的第一个第二特征提取单元,得到第一个第一个第五特征信息,并将第一个第一个第五特征信息输入第i+1个特征提取块中的第二个第二特征提取单元中,得到第二个第五特征信息,依次进行,将第i+1个特征提取块中最后一个第二特征提取单元提取的第五特征信息,作为训练点云块的第i+1个第三特征信息。S403-A32. Input the i-th cascaded feature information of the training point cloud block into the first second feature extraction unit in the i+1th feature extraction block to obtain the first, first, and fifth feature information, and Input the first and fifth feature information into the second second feature extraction unit in the i+1th feature extraction block to obtain the second fifth feature information, and proceed sequentially, and the i+1th The fifth feature information extracted by the last second feature extraction unit in the feature extraction block is used as the i+1th third feature information of the training point cloud block.
例如,第i+1个特征提取块包括3个第二特征提取单元,将上述训练点云块大小为NXKX2C的第i个级联特征信息输入第i+1个特征提取块中的第一个第二特征提取单元,该第一个第二特征提取单元对训练点云块的细节特征进行提取,输出训练点云块的第一个第五特征信息,并将第一个第五特征信息输入第二个第二特征提取单元,第二个第二特征提取单元根据第一个第五特征信息输出训练点云块的第二个第五特征信息,并将第二个第五特征信息输入第三个第二特征提取单元,第三个第二特征提取单元根据第二个第五特征信息输出训练点云块的第三个第五特征信息,将该第三个第五特征信息作为训练点云块的第i+1个第三特征信息。For example, the i+1th feature extraction block includes 3 second feature extraction units, and the i-th cascaded feature information of the above-mentioned training point cloud block size of NXKX2C is input into the first i+1th feature extraction block The second feature extraction unit, the first second feature extraction unit extracts the detailed features of the training point cloud block, outputs the first fifth feature information of the training point cloud block, and inputs the first fifth feature information The second second feature extraction unit, the second second feature extraction unit outputs the second fifth feature information of the training point cloud block according to the first fifth feature information, and inputs the second fifth feature information into the second Three second feature extraction units, the third second feature extraction unit outputs the third fifth feature information of the training point cloud block according to the second fifth feature information, and the third fifth feature information is used as the training point The i+1th third characteristic information of the cloud block.
可选的,该第i+1个第三特征信息的大小为NXC。Optionally, the size of the i+1th third feature information is NXC.
在一些实施例中,如图6C所示,第二特征提取单元包括P个残差块(Residual block,简称RB),P为正整数,此时,上述S403-A32包括:In some embodiments, as shown in FIG. 6C, the second feature extraction unit includes P residual blocks (Residual block, RB for short), and P is a positive integer. At this time, the above S403-A32 includes:
S403-A321、将第i个级联特征信息输入第i+1个特征提取块中的第一个第二特征提取单元,获得第一个第二特征提取单元中第j个残差块输出的第一残差信息,j为小于P的正整数。S403-A321. Input the i-th cascaded feature information into the first second feature extraction unit in the i+1th feature extraction block, and obtain the output of the j-th residual block in the first second feature extraction unit. For the first residual information, j is a positive integer smaller than P.
S403-A322、将第j个残差块输出的第一残差信息和第i个级联特征信息输入第一个第二特征提取单元中的第j+1个残差块中,得到第j+1个残差块输出的第一残差信息。S403-A322. Input the first residual information and the i-th concatenated feature information output by the j-th residual block into the j+1-th residual block in the first second feature extraction unit to obtain the j-th residual block The first residual information output by +1 residual block.
在一种可能的实现方式中,对第j个残差块输出的第一残差信息和第i个级联特征信息进行相加,并将相加后的特征信息输入第j+1个残差块中,得到第j+1个残差块输出的第一残差信息。In a possible implementation, add the first residual information output by the jth residual block and the ith concatenated feature information, and input the added feature information into the j+1th residual block In the difference block, the first residual information output by the j+1th residual block is obtained.
S403-A323、根据第一个第二特征提取单元中的P个残差块中至少一个残差块输出的第一残差信息,以及第i个级联特征信息,确定第一个第二特征提取单元输出的第五特征信息。S403-A323, according to the first residual information output by at least one of the P residual blocks in the first second feature extraction unit, and the i-th concatenated feature information, determine the first second feature The fifth feature information output by the extraction unit.
接着,将第一个第二特征提取单元输出的第五特征信息输入第二个第二特征提取单元,依次进行,直到最后一个第二特征提取单元执行完为止;将第i+1个特征提取块中最后一个第二特征提取单元输出的第五特征信息,确定为训练点云块的第i+1个第三特征信息。Then, input the fifth feature information output by the first second feature extraction unit into the second second feature extraction unit, and proceed sequentially until the last second feature extraction unit is executed; extract the i+1th feature The fifth feature information output by the last second feature extraction unit in the block is determined as the i+1th third feature information of the training point cloud block.
需要说明的是,本申请实施例对残差块的具体网络结构不做限制。It should be noted that the embodiment of the present application does not limit the specific network structure of the residual block.
在一些可能的实现方式中,残差块的网络结构如图6D所示,残差块包括多个带有线性整流函数(Relu)的线性层,用于本申请实施例的残差块用于特征挖掘,帮助网络收敛。In some possible implementations, the network structure of the residual block is as shown in Figure 6D, the residual block includes multiple linear layers with a linear rectification function (Relu), and the residual block used in the embodiment of the present application is used for Feature mining helps the network converge.
举例说明,假设上述P=4,即第二特征提取单元4个残差块,将第i个级联特征信息输入第二特征提取单元HRA中的第一个残差块RB中,该第一RB输出第一残差信息1,将该第一残差信息1和第i个级联特征信息相加输入第二RB,得到第二个RB输出第一残差信息2,将第一残差信息2和第i个级联特征信息相加输入第三RB,得到第三个RB输出的第一残差信息3,将第一残差信息3和第i个级联特征信息相加输入第四RB,得到第四个RB输出的第一残差信息4,进而根据各RB输出的第一残差信息以及第i个级联特征信息,确定该第一个特征提取单元输出的第五特征信息。For example, assuming that the above P=4, that is, there are 4 residual blocks in the second feature extraction unit, and the i-th concatenated feature information is input into the first residual block RB in the second feature extraction unit HRA, the first The RB outputs the first residual information 1, and adds the first residual information 1 and the i-th concatenated feature information to the second RB, and the second RB outputs the first residual information 2, and the first residual information The information 2 and the i-th concatenated feature information are added and input to the third RB to obtain the first residual information 3 output by the third RB, and the first residual information 3 and the i-th concatenated feature information are added and input to the first Four RBs, obtain the first residual information 4 output by the fourth RB, and then determine the fifth feature output by the first feature extraction unit according to the first residual information output by each RB and the i-th concatenated feature information information.
上述S403-A323中根据第一个第二特征提取单元中的P个残差块中至少一个残差块输出的特征信息,以及第i个级联特征信息,确定第一个第二特征提取单元输出的第五特征信息的方式包括但不 限于如下几种:In the above S403-A323, according to the feature information output by at least one of the P residual blocks in the first second feature extraction unit, and the i-th cascaded feature information, determine the first second feature extraction unit The ways of outputting the fifth feature information include but not limited to the following:
方式一,将最后一个残差块输出的第一残差信息与第i个级联特征信息进行相加,作为第一个第二特征提取单元输出的第五特征信息。Way 1: add the first residual information output by the last residual block to the i-th concatenated feature information, and use it as the fifth feature information output by the first second feature extraction unit.
方式二,上述S403-A323包括: Method 2, the above S403-A323 includes:
步骤B1、将第一个第二特征提取单元中P个残差块中最后一个残差块输出的第一残差信息、与P-1个残差块中至少一个残差块输出的第一残差信息进行级联,其中P-1个残差块为P个残差块中除最后一个残差块之外的残差块;Step B1, combine the first residual information output by the last residual block among the P residual blocks in the first second feature extraction unit, and the first residual information output by at least one residual block among the P-1 residual blocks. The residual information is concatenated, wherein the P-1 residual blocks are the residual blocks except the last residual block among the P residual blocks;
例如,如图6E所示,将P个残差块中最后一个残差块输出的第一残差信息、与P-1个残差块中每一个残差块输出的第一残差信息进行级联,得到级联后的特征信息。For example, as shown in FIG. 6E, the first residual information output by the last residual block in the P residual blocks is compared with the first residual information output by each residual block in the P-1 residual blocks. Cascade to obtain the feature information after cascading.
步骤B2、根据级联后的特征信息和第i个级联特征信息,确定第一个第二特征提取单元输出的第五特征信息。Step B2, according to the concatenated feature information and the i-th concatenated feature information, determine the fifth feature information output by the first second feature extraction unit.
上述步骤B2的实现方式包括但不限于如下几种:The implementation methods of the above step B2 include but are not limited to the following:
方式一,将级联后的特征信息和第i个级联特征信息进行相加,作为第一个第二特征提取单元输出的第五特征信息。Way 1: add the concatenated feature information to the i-th concatenated feature information, and use it as the fifth feature information output by the first second feature extraction unit.
方式二,如图6F所示,第二特征提取单元还包括门控单元,此时上述步骤B2包括将级联后的特征信息输入门控单元进行去冗余,得到去冗余后的特征信息;将去冗余后的特征信息与第i个级联特征信息进行相加,作为第一个第二特征提取单元输出的第五特征信息。 Mode 2, as shown in Figure 6F, the second feature extraction unit also includes a gating unit, and at this time the above step B2 includes inputting the cascaded feature information into the gating unit for de-redundancy, and obtaining the de-redundant feature information ; Add the feature information after de-redundancy to the i-th cascaded feature information, and use it as the fifth feature information output by the first second feature extraction unit.
本申请实施例的残差块,提供细节残差信息,并将每个残差块得到的细节残差信息输入门控单元,以收集更多特征细节,实现网络的充分学习。The residual block of the embodiment of the present application provides detailed residual information, and inputs the detailed residual information obtained by each residual block into the gating unit to collect more feature details and realize full learning of the network.
本申请实施例对上述门控单元的网络结构不做限制。The embodiment of the present application does not limit the network structure of the above-mentioned gate control unit.
在一种可能的实现方式,如图6G所示,门控单元包括挤压与激发网络(Squeeze-and-excitation networks,简称SE-Net)和1个线性层组成,其中SE-Net包括全局平均池化层、全连接(Full connection,简称FC)层,SE-Net网络可以将大小为NXKXC的特征信息进行全局平均池化进行,得到大小为1X1XC的特征信息,经过全连接层后,得到大小为1X1XC的特征信息,将全连接层处理后的1X1XC的特征信息与大小为NXKXC的特征信息在信道上进行相乘,得到去冗余后的大小为NXKXC的特征信息,最后将去冗余后的大小为NXKXC的特征信息的特征信息输入线性层,最后输出去冗余后的特征信息。In a possible implementation, as shown in Figure 6G, the gating unit includes Squeeze-and-excitation networks (SE-Net for short) and a linear layer, where SE-Net includes a global average Pooling layer, full connection (Full connection, referred to as FC) layer, the SE-Net network can perform global average pooling of feature information with a size of NXKXC, and obtain feature information with a size of 1X1XC. After passing through the full connection layer, the size is the feature information of 1X1XC, multiply the feature information of 1X1XC processed by the fully connected layer with the feature information of size NXKXC on the channel, and obtain the feature information of size NXKXC after de-redundancy, and finally de-redundancy The feature information whose size is the feature information of NXKXC is input to the linear layer, and finally the feature information after de-redundancy is output.
上述结合特征提取模块的网络结构,详细描述了将训练点云块的几何信息输入特征提取模块,得到训练点云块的第一特征信息的过程。得到训练点云块的第一特征信息后,执行如下S404,以对第一特征信息进行上采样,得到训练点云的第二特征信息。In combination with the network structure of the feature extraction module, the process of inputting the geometric information of the training point cloud block into the feature extraction module to obtain the first feature information of the training point cloud block is described in detail. After obtaining the first feature information of the training point cloud block, perform the following S404 to up-sample the first feature information to obtain the second feature information of the training point cloud.
S404、将训练点云块的第一特征信息输入生成器的特征上采样模块进行上采样,得到训练点云块的第二特征信息。S404. Input the first feature information of the training point cloud block into the feature up-sampling module of the generator for up-sampling, and obtain the second feature information of the training point cloud block.
如图5所示,特征上采样模块用于对训练点云块的第一特征信息进行上采样,得到训练点云块的第二特征信息,例如,将训练点云块的第一特征信息
Figure PCTCN2021096287-appb-000004
上采样为第二特征信息
Figure PCTCN2021096287-appb-000005
其中r为预设的采样率,为正整数,C′是经过上采样后训练点云块的第二特征信息的特征维度。
As shown in Figure 5, the feature upsampling module is used to upsample the first feature information of the training point cloud block to obtain the second feature information of the training point cloud block, for example, the first feature information of the training point cloud block
Figure PCTCN2021096287-appb-000004
Upsampling is the second feature information
Figure PCTCN2021096287-appb-000005
Where r is a preset sampling rate, which is a positive integer, and C′ is the feature dimension of the second feature information of the training point cloud block after upsampling.
在一些实施例中,如图7A所示,特征上采样模块包括:特征上采样子模块和特征提取子模块,其中特征上采样子模块用于对训练点云块的第一特征信息进行上采样,得到训练点云块的上采样特征信息。特征提取子模块用于对训练点云块的上采样特征信息进行特征提取,以得到训练点云块具有表现力的特征,将该具有表现力的特征作为训练点云块的第二特征信息。In some embodiments, as shown in FIG. 7A , the feature upsampling module includes: a feature upsampling submodule and a feature extraction submodule, wherein the feature upsampling submodule is used to upsample the first feature information of the training point cloud block , to obtain the upsampled feature information of the training point cloud block. The feature extraction sub-module is used to perform feature extraction on the upsampled feature information of the training point cloud block to obtain expressive features of the training point cloud block, and use the expressive feature as the second feature information of the training point cloud block.
在上述图7A所示的基础上,上述S404包括S404-A1和S404-A2:Based on the above shown in Figure 7A, the above S404 includes S404-A1 and S404-A2:
S404-A1、将训练点云块的第一特征信息输入特征上采样子模块,以使特征上采样子模块按照预设的上采样率r,将训练点云块的第一特征信息复制r份,并对复制后的第一特征信息在特征维度上增加一个n维向量,得到训练点云块的上采样特征信息,其中不同第一特征信息对应的n维向量的值不相同;S404-A1. Input the first feature information of the training point cloud block into the feature upsampling submodule, so that the feature upsampling submodule copies r copies of the first feature information of the training point cloud block according to the preset upsampling rate r , and add an n-dimensional vector on the feature dimension to the first feature information after copying, to obtain the upsampling feature information of the training point cloud block, wherein the values of the n-dimensional vectors corresponding to different first feature information are different;
S404-A2、将训练点云块的上采样特征信息输入特征提取子模块,得到特征提取子模块提取的训练点云块的第二特征信息。S404-A2. Input the upsampled feature information of the training point cloud block into the feature extraction submodule to obtain the second feature information of the training point cloud block extracted by the feature extraction submodule.
举例说明,将训练点云块的第一特征信息F复制r份,对于每一份被赋值的特征,在其特征维度上增加一个n维的向量,使得每份被复制的特征之间有明显的差异化,此时每个点的特征维度是C+n。例如n=2,即在每个点的特征维度增加一个2维向量,其中向量的值等间隔分布,例如向量的值从-0.2至0.2等间隔分布,此时每个点的特征维度是C+2。将特征维度为C+n的特征信息记为训练点云块的上采样特征信息。接着,将训练点云块的上采样特征信息输入特征提取子模块进行细节特征提取,得到训练点云块的第二特征信息。For example, copy r copies of the first feature information F of the training point cloud block, and for each assigned feature, add an n-dimensional vector to its feature dimension, so that there is a clear difference between each copied feature The difference, at this time the feature dimension of each point is C+n. For example, n=2, that is, add a 2-dimensional vector to the feature dimension of each point, where the values of the vector are equally spaced, for example, the value of the vector is equally spaced from -0.2 to 0.2, and the feature dimension of each point is C +2. The feature information whose feature dimension is C+n is recorded as the upsampling feature information of the training point cloud block. Next, the upsampled feature information of the training point cloud block is input into the feature extraction sub-module to perform detail feature extraction to obtain the second feature information of the training point cloud block.
在一些实施例中,如图7B所示,特征上采样模块还包括第一自相关注意力网络,此时,上述S404-A2包括:将训练点云块的第一上采样特征信息输入第一自相关注意力网络进行特征交互,得到特征交互后的训练点云块的上采样特征信息;将特征交互后的训练点云块的上采样特征信息输入特征提取子模块进行特征提取,得到训练点云块的第二特征信息。In some embodiments, as shown in FIG. 7B , the feature upsampling module further includes a first autocorrelation attention network. At this time, the above S404-A2 includes: inputting the first upsampling feature information of the training point cloud block into the first The autocorrelation attention network performs feature interaction to obtain the upsampling feature information of the training point cloud block after the feature interaction; the upsampling feature information of the training point cloud block after the feature interaction is input into the feature extraction sub-module for feature extraction to obtain the training point The second characteristic information of the cloud block.
可选的,特征交互后的训练点云块的上采样特征信息的特征维度与训练点云块的上采样特征信息的特征维度相同。即第一自相关注意力网络用于特征交互,使得网络学习到更详细的特征。Optionally, the feature dimension of the upsampled feature information of the training point cloud block after the feature interaction is the same as the feature dimension of the upsampled feature information of the training point cloud block. That is, the first autocorrelation attention network is used for feature interaction, allowing the network to learn more detailed features.
可选的,由于上采样特征信息的特征维度为C+n,其C+n的值较大,例如为700左右,使用该特征维度较大的数据进行网络训练时,处理效率低,甚至无法处理。基于此,本申请实施例的第一自相关注意力网络具有降微的左右,以降低特征信息的特征维度,使得特征交互后的训练点云块的上采样特征信息的特征维度低于训练点云块的上采样特征信息的特征维度。即第一自相关注意力网络不仅用于特征交互,并且用于降低特征维度,以降低网络的训练复杂度,进而提高网络的训练速度。Optionally, since the feature dimension of the upsampled feature information is C+n, the value of C+n is relatively large, for example, about 700. When using data with a large feature dimension for network training, the processing efficiency is low, or even impossible. deal with. Based on this, the first autocorrelation attention network in the embodiment of the present application has a reduced left and right to reduce the feature dimension of the feature information, so that the feature dimension of the upsampled feature information of the training point cloud block after feature interaction is lower than that of the training point The feature dimension of the upsampled feature information of the cloud block. That is, the first autocorrelation attention network is not only used for feature interaction, but also used to reduce the feature dimension to reduce the training complexity of the network, thereby improving the training speed of the network.
在一些实施例中,如图7C所示,特征提取子模块包括串联连接的Q个第三特征提取单元,Q为正整数,其中每个第三特征提取单元的特征提取过程相同,此时,上述S404-A2包括:In some embodiments, as shown in Figure 7C, the feature extraction submodule includes Q third feature extraction units connected in series, Q is a positive integer, wherein the feature extraction process of each third feature extraction unit is the same, at this time, The above S404-A2 includes:
S404-A21、将训练点云块的上采样特征信息输入特征提取子模块,获得第k个第三特征提取单元所提取的训练点云块的第k个增强上采样特征信息;S404-A21. Input the upsampling feature information of the training point cloud block into the feature extraction submodule, and obtain the kth enhanced upsampling feature information of the training point cloud block extracted by the kth third feature extraction unit;
S404-A22、将训练点云块的第k个增强上采样特征信息输入第k+1个第三特征提取单元进行特征提取,得到训练点云块的第k+1个增强上采样特征信息;S404-A22. Input the kth enhanced upsampling feature information of the training point cloud block into the k+1th third feature extraction unit for feature extraction, and obtain the k+1th enhanced upsampling feature information of the training point cloud block;
S404-A23、将Q个第三特征提取单元中最后一个第三特征提取单元所提取的训练点云块的第Q个增强上采样特征信息,作为训练点云块的第二特征信息。S404-A23. Using the Qth enhanced upsampled feature information of the training point cloud block extracted by the last third feature extraction unit among the Q third feature extraction units as the second feature information of the training point cloud block.
举例说明,Q=3,将训练点云块的上采样特征信息输入第一个第三特征提取单元进行特征提取,得到训练点云块的第1个增强上采样特征信息,将第1个增强上采样特征信息输入第二个第三特征提取单元进行特征提取,得到训练点云块的第2个增强上采样特征信息,将第2个增强上采样特征信息输入第三个第三特征提取单元进行特征提取,得到训练点云块的第3个增强上采样特征信息,将训练点云块的第3个增强上采样特征信息记为训练点云块的第二特征信息。For example, if Q=3, input the upsampling feature information of the training point cloud block into the first third feature extraction unit for feature extraction, obtain the first enhanced upsampling feature information of the training point cloud block, and use the first enhanced Input the upsampling feature information into the second third feature extraction unit for feature extraction, obtain the second enhanced upsampling feature information of the training point cloud block, and input the second enhanced upsampling feature information into the third third feature extraction unit Perform feature extraction to obtain the third enhanced upsampling feature information of the training point cloud block, and record the third enhanced upsampling feature information of the training point cloud block as the second feature information of the training point cloud block.
在一些实施例中,上述第三特征提取单元与上述第二特征提取单元的网络结构相同。In some embodiments, the network structure of the above-mentioned third feature extraction unit is the same as that of the above-mentioned second feature extraction unit.
在一些实施例中,上述第三特征提取单元与上述第二特征提取单元的网络结构不完全相同。In some embodiments, the network structures of the third feature extraction unit and the second feature extraction unit are not completely the same.
在一些实施例中,第三特征提取单元包括L个残差块,L为正整数。此时,S404-A22包括:In some embodiments, the third feature extraction unit includes L residual blocks, where L is a positive integer. At this time, S404-A22 includes:
S404-A221、将训练点云块的第k个增强上采样特征信息输入第k+1个第三特征提取单元,获得第k+1个第三特征提取单元中第l个残差块输出的第二残差信息,l为小于或等于L的正整数;S404-A221, input the kth enhanced upsampling feature information of the training point cloud block into the k+1th third feature extraction unit, and obtain the output of the lth residual block in the k+1th third feature extraction unit The second residual information, l is a positive integer less than or equal to L;
S404-A222、将第l个残差块输出的第二残差信息和第k个增强上采样特征信息输入第l+1个残差块中,得到第l+1个残差块输出的第二残差信息;S404-A222, input the second residual information output by the lth residual block and the kth enhanced upsampling feature information into the l+1th residual block, and obtain the first outputted by the l+1th residual block Second residual information;
例如,对第l个残差块输出的第二残差信息和第k个增强上采样特征信息进行相加,并将相加后的特征信息输入第l+1个残差块中,确定第l+1个残差块输出的第二残差信息。For example, add the second residual information output by the lth residual block and the kth enhanced upsampling feature information, and input the added feature information into the l+1th residual block to determine the The second residual information output by l+1 residual blocks.
S404-A223、根据L个残差块中至少一个残差块输出的第二残差信息,以及第k个增强上采样特征信息,得到训练点云块的第k+1个增强上采样特征信息。S404-A223. Obtain the k+1th enhanced upsampling feature information of the training point cloud block according to the second residual information output by at least one of the L residual blocks and the kth enhanced upsampling feature information .
需要说明的是,本申请实施例对残差块的具体网络结构不做限制。It should be noted that the embodiment of the present application does not limit the specific network structure of the residual block.
在一些可能的实现方式中,残差块的网络结构如图6D所示,残差块包括多个带有线性整流函数(Relu)的线性层,用于本申请实施例的残差块用于特征挖掘,帮助网络收敛。In some possible implementations, the network structure of the residual block is as shown in Figure 6D, the residual block includes multiple linear layers with a linear rectification function (Relu), and the residual block used in the embodiment of the present application is used for Feature mining helps the network converge.
举例说明,假设上述P=4,即第三特征提取单元4个残差块,将训练点云块的第k个增强上采样特征信息输入第k+1个第三特征提取单元中的第一个残差块RB中,该第一RB输出第二残差信息1,将该第二残差信息1和第k个增强上采样特征信息相加输入第二RB,得到第二个RB输出第二残差 信息2,将第二残差信息2和第k个增强上采样特征信息相加输入第三RB,得到第三个RB输出的第二残差信息3,将第二残差信息3和第k个增强上采样特征信息相加输入第四RB,得到第四个RB输出的第二残差信息4,进而根据各RB输出的第二残差信息以及第k个增强上采样特征信息,得到训练点云块的第k+1个增强上采样特征信息。For example, assuming that the above P=4, that is, there are 4 residual blocks in the third feature extraction unit, and the kth enhanced upsampling feature information of the training point cloud block is input into the first k+1th third feature extraction unit. In a residual block RB, the first RB outputs the second residual information 1, and the second residual information 1 and the kth enhanced upsampling feature information are added to the second RB, and the second RB outputs the first Two residual information 2, the second residual information 2 and the kth enhanced upsampling feature information are added to the third RB to obtain the second residual information 3 output by the third RB, and the second residual information 3 Adding the kth enhanced upsampling feature information to the fourth RB to obtain the second residual information 4 output by the fourth RB, and then according to the second residual information output by each RB and the kth enhanced upsampling feature information , to get the k+1 enhanced upsampling feature information of the training point cloud block.
上述S404-A223中根据L个残差块中至少一个残差块输出的第二残差信息,以及第k个增强上采样特征信息,得到训练点云块的第k+1个增强上采样特征信息的方式包括但不限于如下几种:In the above S404-A223, according to the second residual information output by at least one of the L residual blocks and the kth enhanced upsampling feature information, the k+1th enhanced upsampling feature of the training point cloud block is obtained Information methods include but are not limited to the following:
方式一,将L个残差块中最后一个残差块输出的第二残差信息与第k个增强上采样特征信息进行相加,作为训练点云块的第i+1个第三特征信息。Method 1: add the second residual information output by the last residual block in the L residual blocks to the kth enhanced upsampling feature information, and use it as the i+1th third feature information of the training point cloud block .
方式二,上述S404-A223包括步骤C1和步骤C2: Method 2, the above S404-A223 includes step C1 and step C2:
步骤C1、将L个残差块中最后一个残差块输出的第二残差信息、与L-1个残差块中至少一个残差块输出的第二残差信息进行级联,其中L-1个残差块为L个残差块中除最后一个残差块之外的残差块。Step C1, concatenate the second residual information output by the last residual block in the L residual blocks with the second residual information output by at least one residual block in the L-1 residual blocks, where L -1 residual block is a residual block except the last residual block among the L residual blocks.
例如,如图7E所示,将L个残差块中最后一个残差块输出的第二残差信息、与L-1个残差块中每一个残差块输出的第二残差信息进行级联,得到级联后的特征信息。For example, as shown in FIG. 7E , the second residual information output by the last residual block in the L residual blocks is compared with the second residual information output by each residual block in the L-1 residual blocks. Cascade to obtain the feature information after cascading.
步骤C2、根据级联后的特征信息和第k个上采样特征信息,确定训练点云块的第k+1个增强上采样特征信息。Step C2, according to the concatenated feature information and the kth upsampling feature information, determine the k+1th enhanced upsampling feature information of the training point cloud block.
上述步骤C2的实现方式包括但不限于如下几种:The implementation methods of the above step C2 include but are not limited to the following:
方式一,将级联后的特征信息和第k个增强上采样特征信息进行相加,作为训练点云块的第k+1个增强上采样特征信息。Way 1: add the cascaded feature information to the kth enhanced upsampling feature information, and use it as the k+1th enhanced upsampling feature information of the training point cloud block.
方式二,第三特征提取单元还包括门控单元,此时上述步骤C2包括:将级联后的特征信息输入门控单元进行去冗余,得到去冗余后的特征信息;将去冗余后的特征信息与第k个增强上采样特征信息进行相加,作为训练点云块的第k+1个增强上采样特征信息。 Mode 2, the third feature extraction unit also includes a gating unit, and at this time, the above step C2 includes: inputting the cascaded feature information into the gating unit for de-redundancy to obtain de-redundant feature information; The final feature information is added to the k-th enhanced up-sampled feature information, and used as the k+1-th enhanced up-sampled feature information of the training point cloud block.
图7D为本申请实施例提供的特征上采样模块的一种具体网络结构示意图,如图7D所示,特征上采样子模块将大小为NXC的第一特征信息上采样为大小为rNX(C+2)的上采样特征信息,将大小为rNX(C+2)的上采样特征信息输入第一自相关注意力网络(Self-attetion),得到特征交互后的上采样特征信息,将特征交互后的上采样特征信息输入特征提取子模块,该特征提取子模块包括多个第三特征提取单元,可选的第三特征提取单元与第二特征提取单元HRA的网络结构一致,输出训练点云块的第二特征信息。FIG. 7D is a schematic diagram of a specific network structure of the feature upsampling module provided in the embodiment of the present application. As shown in FIG. 7D , the feature upsampling submodule upsamples the first feature information with a size of NXC to a size of rNX(C+ 2) The upsampled feature information of the size rNX(C+2) is input into the first self-correlation attention network (Self-attetion), and the upsampled feature information after feature interaction is obtained, and the feature interaction is The upsampled feature information of the input feature extraction submodule, the feature extraction submodule includes a plurality of third feature extraction units, the optional third feature extraction unit is consistent with the network structure of the second feature extraction unit HRA, and outputs the training point cloud block The second characteristic information of .
本申请实施例对上述门控单元的网络结构不做限制。The embodiment of the present application does not limit the network structure of the above-mentioned gate control unit.
在一种可能的实现方式,上述门控单元的网络结构如图6G所示,具体参照上述S403的描述。In a possible implementation manner, the network structure of the above-mentioned gate control unit is as shown in FIG. 6G , and for details, refer to the above-mentioned description of S403 .
上述结合特征上采样模块的网络结构,详细描述了将训练点云块的第一特征信息输入特征上采样模块,得到训练点云块的第二特征信息的过程。得到训练点云块的第二特征信息后,执行如下S405,以对第二特征信息进行空间转换,得到训练点云块的上采样几何信息。The above network structure combined with the feature upsampling module describes in detail the process of inputting the first feature information of the training point cloud block into the feature upsampling module to obtain the second feature information of the training point cloud block. After obtaining the second feature information of the training point cloud block, perform the following S405 to perform spatial conversion on the second feature information to obtain upsampled geometric information of the training point cloud block.
S405、将训练点云块的第二特征信息输入生成器中的几何生成模块,得到训练点云块的上采样几何信息。S405. Input the second characteristic information of the training point cloud block into the geometry generation module in the generator, and obtain the upsampled geometric information of the training point cloud block.
本申请实施例的几何生成模块的作用是将上采样得到训练点云块的第二特征信息
Figure PCTCN2021096287-appb-000006
从特征空间重新映射回几何空间,最终获得上采样的点云,即把F up回归到几何空间,得到训练点云块的上采样几何信息
Figure PCTCN2021096287-appb-000007
其中3指的是几何信息维度,rN为上采样后训练点云块所包括的点的数量。
The function of the geometry generation module in the embodiment of the present application is to obtain the second feature information of the training point cloud by upsampling
Figure PCTCN2021096287-appb-000006
Remap from the feature space back to the geometric space, and finally obtain the upsampled point cloud, that is, return F up to the geometric space, and obtain the upsampled geometric information of the training point cloud block
Figure PCTCN2021096287-appb-000007
Among them, 3 refers to the geometric information dimension, and rN is the number of points included in the training point cloud block after upsampling.
本申请实施例对几何生成模块的具体网络结构不做限制。The embodiment of the present application does not limit the specific network structure of the geometry generation module.
在一些实施例中,几何生成模块包括多个全连接层,则上述S405包括:将训练点云块的第二特征信息输入多个全连接层进行空间转换,得到训练点云块的上采样几何信息。In some embodiments, the geometry generation module includes a plurality of fully connected layers, then the above S405 includes: inputting the second feature information of the training point cloud block into a plurality of fully connected layers for space conversion, and obtaining the upsampled geometry of the training point cloud block information.
在一些实施例中,直接输出上采样的几何信息并不能很好地生成均匀分布的点云以及抑制边界处的噪声,此时,为了解决该技术问题,本申请将点云上采样r+m倍;然后通过FC生成上采样的几何信息;再使用高通图滤波器显式地移除每个上采样Patch中的多个高频点(即噪点),最后通过最远点采样(Farthest point sampling,简称FPS)算法,将点云下采样至r倍,输出
Figure PCTCN2021096287-appb-000008
In some embodiments, directly outputting the upsampled geometric information cannot well generate a uniformly distributed point cloud and suppress noise at the boundary. At this time, in order to solve this technical problem, this application will upsample the point cloud by r+m times; then generate the upsampled geometric information through FC; then use the high-pass image filter to explicitly remove multiple high-frequency points (ie noise) in each upsampled Patch, and finally pass the farthest point sampling (Farthest point sampling , referred to as FPS) algorithm, downsampling the point cloud to r times, output
Figure PCTCN2021096287-appb-000008
基于此,如图8所示,几何生成模块包括:几何重建单元、滤波单元和下采样单元,其中几何重建单元包括多个全连接层,此时,上述S405包括:Based on this, as shown in FIG. 8, the geometry generation module includes: a geometry reconstruction unit, a filter unit, and a downsampling unit, wherein the geometry reconstruction unit includes multiple fully connected layers. At this time, the above S405 includes:
S405-A1、将训练点云块的第二特征信息输入几何重建单元进行几何重建,得到训练点云块的初始上采样几何信息。S405-A1. Input the second characteristic information of the training point cloud block into the geometric reconstruction unit to perform geometric reconstruction, and obtain the initial upsampling geometric information of the training point cloud block.
S405-A2、将训练点云块的初始上采样几何信息输入滤波单元进行除噪,得到训练点云块滤除噪点的初始上采样几何信息。S405-A2. Input the initial upsampling geometric information of the training point cloud block into the filtering unit for denoising, and obtain the initial upsampling geometric information of the training point cloud block for filtering noise.
可选的,滤波单元可以是高通图滤波器,显式地移除每个上采样Patch中的多个,例如5个高频点(即噪点)。Optionally, the filtering unit may be a high-pass image filter, which explicitly removes a plurality of, for example, 5 high-frequency points (ie, noise points) in each upsampling patch.
S405-A3、将训练点云块滤除噪点的初始上采样几何信息输入下采样单元中进行下采样,得到训练点云块的预测上采样几何信息。S405-A3. Input the initial upsampling geometric information of the training point cloud block to filter out the noise into the downsampling unit for downsampling, and obtain the predicted upsampling geometric information of the training point cloud block.
例如,最后通过最远点采样(FPS)算法,将训练点云块滤除噪点的初始上采样几何信息由r+m倍下采样至r倍,输出
Figure PCTCN2021096287-appb-000009
m为正整数,例如为m=2。即训练点云块的上采样几何信息对应的上采样率小于或等于所述特征上采样模块的上采样率。
For example, at the end, the furthest point sampling (FPS) algorithm is used to downsample the initial upsampling geometric information of the training point cloud block to filter out noise from r+m times to r times, and output
Figure PCTCN2021096287-appb-000009
m is a positive integer, for example m=2. That is, the upsampling rate corresponding to the upsampling geometric information of the training point cloud block is less than or equal to the upsampling rate of the feature upsampling module.
S406、根据训练点云块的预测上采样几何信息,对生成器中的特征提取模块、特征上采样模块和几何生成模块进行训练,得到训练后的生成器。S406. According to the predicted upsampling geometric information of the training point cloud block, train the feature extraction module, feature upsampling module and geometry generation module in the generator to obtain a trained generator.
本申请实施例中,上述S406的实现方式包括但不限于如下几种方式:In the embodiment of this application, the implementation of the above S406 includes but is not limited to the following methods:
方式一,根据训练点云块的预测上采样几何信息与训练点云块的几何信息的上采样真值之间的损失,反向训练生成器中的特征提取模块、特征上采样模块和几何生成模块,得到训练后的生成器。Method 1: According to the loss between the predicted upsampled geometric information of the training point cloud block and the upsampled true value of the geometric information of the training point cloud block, the feature extraction module, feature upsampling module and geometry generation module in the reverse training generator module to get the trained generator.
需要说明的是,本申请实施例的训练过程为一个迭代过程,每次训练过程一致,且每次训练过程中对生成器中的参数(例如权重矩阵)进行一次更新,直到达到模型训练结束条件为止。It should be noted that the training process of the embodiment of the present application is an iterative process, and each training process is consistent, and the parameters in the generator (such as the weight matrix) are updated once during each training process until the model training end condition is reached until.
可选的,模型训练结束条件包括训练次数到达预设次数,或者生成器的预测误差到达预设值等。Optionally, the model training end condition includes that the number of training times reaches a preset number of times, or the prediction error of the generator reaches a preset value, and the like.
图9为本申请实施例涉及生成器的训练过程的一种示意图,如图9所示,将训练点云块的几何信息输入生成器,得到生成器输出的训练点云块的预测上采样几何信息,根据训练点云的预测上采样几何信息和训练点云的几何信息的上采样真值,对生成器中的特征提取模块、特征上采样模块和几何生成模块的参数进行调整,例如根据训练点云的上采样几何信息和训练点云的几何信息的上采样真值之间的损失,对特征提取模块、特征上采样模块和几何生成模块的参数矩阵进行更新,以得到训练好的生成器。Fig. 9 is a schematic diagram of the training process involving the generator in the embodiment of the present application. As shown in Fig. 9, the geometric information of the training point cloud block is input into the generator, and the predicted upsampling geometry of the training point cloud block output by the generator is obtained. Information, according to the predicted upsampled geometric information of the training point cloud and the upsampled true value of the geometric information of the training point cloud, adjust the parameters of the feature extraction module, feature upsampling module and geometry generation module in the generator, for example, according to the training The loss between the upsampled geometric information of the point cloud and the upsampled true value of the geometric information of the training point cloud updates the parameter matrix of the feature extraction module, feature upsampling module and geometry generation module to obtain a trained generator .
其中,训练点云的几何信息的上采样真值可以理解为训练数据中已包括的对训练点云的几何信息进行上采样后的数据。Wherein, the upsampled true value of the geometric information of the training point cloud can be understood as the data included in the training data after upsampling the geometric information of the training point cloud.
可选的,训练点云的几何信息的上采样真值的分辨率低于生成器输出的训练点云的上采样几何信息。Optionally, the resolution of the upsampled true value of the geometric information of the training point cloud is lower than the upsampled geometric information of the training point cloud output by the generator.
方式二,借助判断器对生成器进行训练,此时上述S406包括: Method 2, training the generator with the help of the judger, at this time the above S406 includes:
S406-A1、将训练点云块的预测上采样几何信息输入判别器,得到判别器的第一判别结果,判别器用于判断输入判别器的数据是否为训练点云块的上采样真值。S406-A1. Input the predicted upsampled geometric information of the training point cloud block into the discriminator to obtain a first discrimination result of the discriminator, and the discriminator is used to judge whether the data input to the discriminator is the upsampled true value of the training point cloud block.
S406-A2、根据判别器的第一判别结果,对所述生成器中的特征提取模块、特征上采样模块和几何生成模块进行训练,得到训练后的生成器。S406-A2. According to the first discrimination result of the discriminator, train the feature extraction module, feature upsampling module and geometry generation module in the generator to obtain a trained generator.
其中,判别器可以为一段软件代码或者为具有数据处理功能的芯片。Wherein, the discriminator may be a piece of software code or a chip with data processing function.
图10为本申请实施例涉及生成器的训练过程的另一种示意图,如图10所示,将训练点云块的几何信息输入生成器,得到生成器输出的训练点云块的上采样几何信息,接着,将训练点云块的上采样几何信息输入判别器中,得到判别器输出的第一判别结果。根据判别器的第一判别结果,对生成器中的特征提取模块、特征上采样模块和几何生成模块的参数矩阵进行调整,以实现对生成器的训练。Fig. 10 is another schematic diagram of the training process involving the generator in the embodiment of the present application. As shown in Fig. 10, the geometric information of the training point cloud block is input into the generator to obtain the upsampled geometry of the training point cloud block output by the generator Next, input the upsampled geometric information of the training point cloud block into the discriminator, and obtain the first discriminant result output by the discriminator. According to the first discrimination result of the discriminator, the parameter matrix of the feature extraction module, the feature upsampling module and the geometry generation module in the generator are adjusted to realize the training of the generator.
具体的,将训练点云划分为至少一个训练点云块,针对至少一个训练点云块中的每一个训练点云块,将该训练点云块的几何信息输入生成器中,该生成器对该训练点云块的几何信息进行上采样,得到训练点云块的预测上采样几何信息,训练点云块的几何信息经过上采样后,变为稠密的训练点云块。该稠密的训练点云块应该与训练点云块的上采样真值具有一致的几何分布,也就是说,若生成器精度高时,生成器上采样后的训练点云块的几何分布应该接近训练点云块的上采样真值的几何分布。Specifically, the training point cloud is divided into at least one training point cloud block, and for each training point cloud block in at least one training point cloud block, the geometric information of the training point cloud block is input in the generator, and the generator is The geometric information of the training point cloud block is up-sampled to obtain the predicted up-sampling geometric information of the training point cloud block. After the geometric information of the training point cloud block is up-sampled, it becomes a dense training point cloud block. The dense training point cloud block should have the same geometric distribution as the upsampled true value of the training point cloud block, that is, if the generator has high precision, the geometric distribution of the training point cloud block after the generator upsampling should be close to The geometric distribution of the upsampled ground truth for training point cloud blocks.
基于此,本申请实施例将训练点云块的预测上采样几何信息输入判别器,使得判别器判断输入判别器的数据是否为训练点云块的上采样真值,并输出第一判别结果。当第一判别结果为第一数值,例如0时,表示判别器判断输入判别器的数据为上采样后的训练点云块,说明生成器未训练完成,对生成器中的参数矩阵进行反向调整。当第一判别结果为第二数值,例如1时,表示判别器判断输入判别器的数据为训练点云块的上采样真值(Ground Truth),此时说明该生成器训练完成,进而使用训练完成的生成器对点云的几何信息进行上采样。Based on this, in the embodiment of the present application, the predicted upsampled geometric information of the training point cloud block is input to the discriminator, so that the discriminator judges whether the data input to the discriminator is the upsampled true value of the training point cloud block, and outputs the first discrimination result. When the first discrimination result is the first value, such as 0, it means that the discriminator judges that the data input to the discriminator is an upsampled training point cloud block, indicating that the generator has not been trained, and the parameter matrix in the generator is reversed. Adjustment. When the first discriminant result is the second value, such as 1, it means that the discriminator judges that the data input to the discriminator is the upsampled ground truth of the training point cloud block. At this time, the training of the generator is completed, and then the training The completed generator upsamples the geometric information of the point cloud.
在一些实施例中,上述S406-A2包括:In some embodiments, the above S406-A2 includes:
S406-A21、根据所述第一判别结果,确定所述生成器的第一损失。S406-A21. Determine a first loss of the generator according to the first discrimination result.
本申请实施例对根据第一判别结果,确定第一损失时所使用的损失函数的具体类型不做限制。The embodiment of the present application does not limit the specific type of the loss function used when determining the first loss according to the first discrimination result.
在一种可能的实现方式中,根据第一判别结果,采用最小二乘损失函数,确定生成器的第一损失。In a possible implementation manner, according to the first discrimination result, a least squares loss function is used to determine the first loss of the generator.
例如,根据如下公式(1)确定生成器的第一损失:For example, the first loss of the generator is determined according to the following formula (1):
Figure PCTCN2021096287-appb-000010
Figure PCTCN2021096287-appb-000010
其中,L gen(P up)为第一损失,P up为训练点云块的上采样几何信息,D(P up)为将训练点云块的上采样几何信息输入判别器,判别器所输出的第一判断结果。 Among them, L gen (P up ) is the first loss, P up is the upsampling geometric information of the training point cloud block, D(P up ) is the input of the upsampling geometric information of the training point cloud block to the discriminator, and the output of the discriminator The first judgment result of .
S406-A22、根据第一损失,确定所述生成器中的特征提取模块、特征上采样模块和几何生成模块的参数矩阵。S406-A22. According to the first loss, determine a parameter matrix of a feature extraction module, a feature upsampling module, and a geometry generation module in the generator.
上述S406-A22中根据第一损失,确定所述生成器中的特征提取模块、特征上采样模块和几何生成模块的参数矩阵的方式包括但不限于如下几种:According to the first loss in the above S406-A22, the ways of determining the parameter matrix of the feature extraction module, feature upsampling module and geometry generation module in the generator include but are not limited to the following:
方式一,基于第一损失确定所述生成器中的特征提取模块、特征上采样模块和几何生成模块的参数矩阵,例如当第一损失大于某一预设值时,说明生成器的精度未达到预设要求,对生成器中的特征提取模块、特征上采样模块和几何生成模块的参数矩阵进行反向调整。若第一损失小于某一预设值时,说明生成器的精度达到预设要求,则固定生成器中特征提取模块、特征上采样模块和几何生成模块在此时的参数矩阵。Method 1: Determine the parameter matrix of the feature extraction module, feature upsampling module, and geometry generation module in the generator based on the first loss. For example, when the first loss is greater than a certain preset value, it means that the accuracy of the generator has not reached Preset requirements, inversely adjust the parameter matrix of the feature extraction module, feature upsampling module and geometry generation module in the generator. If the first loss is less than a certain preset value, it means that the accuracy of the generator meets the preset requirements, and the parameter matrix of the feature extraction module, feature upsampling module and geometry generation module in the generator is fixed at this time.
方式二,上述S406-A22包括如下步骤: Mode 2, the above S406-A22 includes the following steps:
步骤A1、确定生成器的至少一个第二损失;Step A1. Determine at least one second loss of the generator;
步骤A2、根据生成器的第一损失和生成器的至少一个第二损失,确定生成器的目标损失;Step A2. Determine the target loss of the generator according to the first loss of the generator and at least one second loss of the generator;
步骤A3、根据生成器的目标损失,确定生成器中的特征提取模块、特征上采样模块和几何生成模块的参数矩阵。Step A3, according to the target loss of the generator, determine the parameter matrix of the feature extraction module, feature upsampling module and geometry generation module in the generator.
在该方式二中,为了进一步提高生成器的训练精度,确定出生成器的至少一个第二损失,根据生成器的第一损失和至少一个第二损失,对生成器中的特征提取模块、特征上采样模块和几何生成模块的参数矩阵进行调整,以提高生成器的训练准确性。In the second method, in order to further improve the training accuracy of the generator, at least one second loss of the generator is determined, and according to the first loss and at least one second loss of the generator, the feature extraction module, feature The parameter matrices of the upsampling module and the geometry generation module are tuned to improve the training accuracy of the generator.
本申请实施例对上述步骤A1中确定生成器的至少一个第二损失的方式不做限制,具体根据实际需要确定。The embodiment of the present application does not limit the manner of determining at least one second loss of the generator in the above step A1, which is specifically determined according to actual needs.
在一种示例中,上述步骤A1包括:根据训练点云块的上采样几何信息和训练点云块的几何信息的上采样真值,采用地动距离方式,确定生成器的一个第二损失。In an example, the above step A1 includes: according to the upsampled geometric information of the training point cloud block and the upsampled true value of the geometric information of the training point cloud block, using the ground motion distance method to determine a second loss of the generator.
采用地动距离方式确定生成器的一个第二损失也称为重建损失函数,目的是让上采样后的训练点云块与训练点云块的上采样真值具有一致的几何分布。Using the ground motion distance method to determine a second loss of the generator is also called the reconstruction loss function. The purpose is to make the upsampled training point cloud block and the upsampled true value of the training point cloud block have a consistent geometric distribution.
例如,根据如下公式(2),确定生成器的一个第二损失:For example, according to the following formula (2), a second loss of the generator is determined:
Figure PCTCN2021096287-appb-000011
Figure PCTCN2021096287-appb-000011
其中,L rec为第二损失,EMD表示地动距离方式,P up为上采样后的训练点云块的几何信息,P T为训练点云块的几何信息的上采样真值,φ:P up→P T是由2个等大小的子集P up和P T构成的双射,p i为P up为中的第i个点,φ(p i)表示P T中按照双射关系φ找到的p i的对应点。 Among them, L rec is the second loss, EMD represents the ground motion distance method, P up is the geometric information of the upsampled training point cloud block, P T is the upsampled true value of the geometric information of the training point cloud block, φ:P up → P T is a bijection composed of two equal-sized subsets P up and P T , p i is the i-th point in P up , φ(p i ) means that in P T according to the bijective relationship φ Find the corresponding point of pi .
在一种示例中,上述步骤A1包括:根据均匀损失函数,确定所述生成器的至少一个第二损失。In an example, the above step A1 includes: determining at least one second loss of the generator according to a uniform loss function.
例如,根据如下公式(3),确定生成器的一个第二损失:For example, according to the following formula (3), a second loss of the generator is determined:
Figure PCTCN2021096287-appb-000012
Figure PCTCN2021096287-appb-000012
其中,L uni为第二损失,
Figure PCTCN2021096287-appb-000013
S i指的是通过半径球(半径为r q)方法获得的局部表面i,T是获得的种子点个数,d i,j表示第i个局部表面内第j个点与其最近邻点的距离,
Figure PCTCN2021096287-appb-000014
是期望的每个局部表面包含的点的个数,
Figure PCTCN2021096287-appb-000015
是期望的局部表面内每个点与其最近邻点的空间距离。
Among them, L uni is the second loss,
Figure PCTCN2021096287-appb-000013
S i refers to the local surface i obtained by the radius sphere (radius r q ) method, T is the number of seed points obtained, and d i,j represents the distance between the jth point and its nearest neighbor in the i-th local surface distance,
Figure PCTCN2021096287-appb-000014
is the desired number of points per local surface,
Figure PCTCN2021096287-appb-000015
is the spatial distance between each point and its nearest neighbor in the desired local surface.
在一种示例中,上述步骤A1包括:In an example, the above step A1 includes:
步骤A11、将训练点云块的上采样几何信息进行下采样,得到与训练点云块相同点数的下采样训练点云块。Step A11, downsampling the upsampled geometric information of the training point cloud block to obtain a downsampled training point cloud block with the same number of points as the training point cloud block.
例如,使用最远点采样(Farthest point sampling,FPS)将上采样的训练点云块P up下采样到与低分辨率的训练点云块P ori相同的点数,得到下采样训练点云块P low,即
Figure PCTCN2021096287-appb-000016
For example, use Farthest point sampling (FPS) to downsample the upsampled training point cloud block P up to the same number of points as the low-resolution training point cloud block P ori to obtain the downsampled training point cloud block P low , ie
Figure PCTCN2021096287-appb-000016
步骤A12、根据下采样训练点云块的几何信息和训练点云块的几何信息,采用地动距离方式,确定生成器的一个第二损失。Step A12. According to the geometric information of the downsampled training point cloud block and the geometric information of the training point cloud block, a second loss of the generator is determined by using the ground motion distance method.
例如,根据如下公式(4),确定生成器的一个第二损失:For example, according to the following formula (4), a second loss of the generator is determined:
Figure PCTCN2021096287-appb-000017
Figure PCTCN2021096287-appb-000017
其中,L id为生成器的第二损失,P ori为低分辨率的训练点云块,P low为下采样后的训练点云块,φ:P low→P ori表示由P low和P ori构成的双射,有且只有唯一的一种移动方式让P low与P ori移动到彼此点集的距离最小,
Figure PCTCN2021096287-appb-000018
为P low中的第k个点,
Figure PCTCN2021096287-appb-000019
Figure PCTCN2021096287-appb-000020
在P ori中对应的点。
Among them, L id is the second loss of the generator, P ori is the low-resolution training point cloud block, P low is the training point cloud block after downsampling, φ:P low → P ori means that it is composed of P low and P ori In the bijection formed, there is one and only one way to move P low and P ori to the minimum distance between the point sets of each other,
Figure PCTCN2021096287-appb-000018
is the kth point in P low ,
Figure PCTCN2021096287-appb-000019
for
Figure PCTCN2021096287-appb-000020
Corresponding point in P ori .
根据上述方式确定出生成器的至少一个第二损失后,根据生成器的第一损失和生成器的至少一个第二损失,确定生成器的目标损失,例如将生成器的第一损失和至少一个第二损失的加权平均值,确定生成器的目标损失。After at least one second loss of the generator is determined according to the above method, the target loss of the generator is determined according to the first loss of the generator and at least one second loss of the generator, for example, the first loss of the generator and at least one second loss of the generator A weighted average of the second loss, which determines the target loss for the generator.
示例性的,根据如下公式(5)确定生成器的目标损失:Exemplarily, the target loss of the generator is determined according to the following formula (5):
L G=w genL gen(P up)+w recL rec+w uniL uni+w idL id  (5) L G =w gen L gen (P up )+w rec L rec +w uni L uni +w id L id (5)
其中,L G为生成器的目标损失,L gen(P up)为生成器的第一损失,L rec、L uni、L id分别为生成器的各第二损失,w gen为第一损失的权重,w rec、w uni、w id分别为各第二损失对应的权重。 Among them, L G is the target loss of the generator, L gen (P up ) is the first loss of the generator, L rec , L uni , L id are the second losses of the generator respectively, w gen is the first loss of the generator Weights, w rec , w uni , and w id are weights corresponding to the second losses, respectively.
需要说明的是,本申请实施例对上述各损失对应的权重的具体取值不做限制,具体根据实际需要确定。It should be noted that, the embodiments of the present application do not limit the specific values of the weights corresponding to the above losses, which are specifically determined according to actual needs.
可选的,w gen=1。 Optionally, w gen =1.
可选的,w rec=100。 Optionally, w rec =100.
可选的,w uni=10。 Optionally, w uni =10.
可选的,w id=1。 Optionally, w id =1.
本申请实施例采用训练点云对生成器进行训练,得到训练后的生成器,以便在实际应用中使用训练好的生成器进行点云的几何信息上采样,得到高精度的点云。进一步的,本申请实施例将训练点云划分为训练点云块使用训练点云块对生成器进行训练,使用判别器对生成器的训练过程进行监督,进而提高了生成器的训练准确性和可靠性。The embodiment of the present application uses the training point cloud to train the generator to obtain the trained generator, so that in practical applications, the trained generator can be used to upsample the geometric information of the point cloud to obtain a high-precision point cloud. Further, the embodiment of the present application divides the training point cloud into training point cloud blocks, uses the training point cloud blocks to train the generator, and uses the discriminator to supervise the training process of the generator, thereby improving the training accuracy and accuracy of the generator. reliability.
上文结合生成器的网络结构,对生成器的训练过程进行介绍,下面对上述S406-A1中涉及的判别器进行介绍。The training process of the generator is introduced above in combination with the network structure of the generator, and the discriminator involved in the above S406-A1 is introduced below.
在一些实施例中,上述判别器为预先训练好的判别器。In some embodiments, the above discriminator is a pre-trained discriminator.
在一些实施例中,上述判别器不是预先训练好的,即本申请实施例还涉及判别器的训练过程。In some embodiments, the discriminator is not pre-trained, that is, the embodiment of the present application also involves a training process of the discriminator.
本申请实施例在使用训练点云块的几何信息对生成器进行训练之前,先对判别器进行一次训练,接着,使用训练后的判别器执行S406-A1。In the embodiment of the present application, before using the geometric information of the training point cloud block to train the generator, the discriminator is trained once, and then S406-A1 is executed using the trained discriminator.
在一种可能的训练方式中,在判别器的训练过程中,判别器与生成器交替进行训练,即在训练过程中,先使用训练点云块的几何信息对判别器进行训练,对判别器训练结束后,再使用该训练点云块的几何信息对生成器进行训练,判别器与生成器的训练过程交替进行,直到生成器和判别器训练完成为止。In a possible training method, during the training process of the discriminator, the discriminator and the generator are alternately trained, that is, in the training process, the geometric information of the training point cloud block is used to train the discriminator first, and the discriminator After the training, the generator is trained using the geometric information of the training point cloud block, and the training process of the discriminator and the generator is carried out alternately until the training of the generator and the discriminator is completed.
在一些实施例中,判别器的训练过程具体包括如下几个步骤:In some embodiments, the training process of the discriminator specifically includes the following steps:
步骤21、将生成器生成的训练点云块的预测上采样几何信息输入判别器,得到判别器输出的第二判别结果;将训练点云块的几何信息的上采样真值输入判别器,得到判别器输出的第三判别结果;Step 21. Input the predicted upsampled geometric information of the training point cloud block generated by the generator into the discriminator, and obtain the second discrimination result output by the discriminator; input the upsampled true value of the geometric information of the training point cloud block into the discriminator, and obtain the third discrimination result output by the discriminator;
步骤22、根据第二判别结果和第三判别结果,确定判别器的损失;Step 22. Determine the loss of the discriminator according to the second discrimination result and the third discrimination result;
步骤23、根据判别器的损失,对判别器中的参数进行调整。Step 23. Adjust the parameters in the discriminator according to the loss of the discriminator.
本申请实施例对步骤21中根据第二判别结果和第三判别结果,确定判别器的损失所使用的损失函数的类型不做限制。The embodiment of the present application does not limit the type of loss function used to determine the loss of the discriminator according to the second discrimination result and the third discrimination result in step 21 .
在一种可能的实现方式中,步骤21包括:根据第二判别结果和第三判别结果,采用最小二乘损失函数,确定判别器的损失。In a possible implementation manner, step 21 includes: according to the second discrimination result and the third discrimination result, using a least squares loss function to determine the loss of the discriminator.
例如,根据如下公式(7),确定判别器的损失:For example, according to the following formula (7), the loss of the discriminator is determined:
Figure PCTCN2021096287-appb-000021
Figure PCTCN2021096287-appb-000021
其中,L dis(P up,P T)表示判别器的损失,P T是训练点云块的上采样真值,P up指的是生成器上采样得到的点云,即上采样后的训练点云块。 Among them, L dis (P up , P T ) represents the loss of the discriminator, P T is the upsampled true value of the training point cloud block, P up refers to the point cloud obtained by the upsampling of the generator, that is, the training after upsampling Point cloud blocks.
本申请实施例根据判别器对训练点云块的预测上采样几何信息和训练点云块的几何信息的上采样真值的判别结果之间的差异,对判别器进行训练,进而提高了判别器的训练准确性。In the embodiment of the present application, the discriminator is trained according to the difference between the discriminator's discriminant result of the predicted upsampled geometric information of the training point cloud block and the upsampled true value of the geometric information of the training point cloud block, thereby improving the discriminator. training accuracy.
下面结合判别器的网络结构,对判别器根据点云块的几何信息,得到判别结果的过程进行详细,也就是说,对判别器生成第一判别结果、第二判别结果和第三判别结果的过程进行介绍。In the following, combined with the network structure of the discriminator, the process of the discriminator obtaining the discriminant result based on the geometric information of the point cloud block will be described in detail, that is, the discriminator generates the first discriminant result, the second discriminant result and the third discriminant result. The process is introduced.
图11为判别器的一种网络结构示意图,如图11所示,判别器包括全局判别模块、边界判别模块和全连接模块,其中,全局判别模块用于提取点云的全局特征信息,边界判别模块用于提取点云的边界特征信息,全连接模块用于对点云的全局特征信息和边界特征信息进行处理,得到判别结果。Figure 11 is a schematic diagram of a network structure of the discriminator. As shown in Figure 11, the discriminator includes a global discriminant module, a boundary discriminant module and a fully connected module, wherein the global discriminant module is used to extract the global feature information of the point cloud, and the The module is used to extract the boundary feature information of the point cloud, and the fully connected module is used to process the global feature information and boundary feature information of the point cloud to obtain the discrimination result.
图12为本申请一实施例提供的模型训练方法的流程示意图,如图11和图12所示,判别器得到判别结果的过程包括:Fig. 12 is a schematic flow chart of the model training method provided by an embodiment of the present application. As shown in Fig. 11 and Fig. 12, the process for the discriminator to obtain the discriminant result includes:
S601、获取目标点云块的边界点的几何信息。S601. Acquire geometric information of boundary points of a target point cloud block.
例如,使用高通图滤波器(high-pass graph filter)提取目标点云块的边界点的几何信息。For example, a high-pass graph filter is used to extract the geometric information of the boundary points of the target point cloud block.
S602、将目标点云块的边界点的几何信息输入边界判别模块进行边界特征提取,得到目标点云块的边界特征信息。S602. Input the geometric information of the boundary points of the target point cloud block into the boundary discrimination module to perform boundary feature extraction, and obtain boundary feature information of the target point cloud block.
S603、将目标点云块的几何信息输入全局判别模块进行全局特征提取,得到目标点云块的全局特征信息。S603. Input the geometric information of the target point cloud block into the global discrimination module for global feature extraction, and obtain the global feature information of the target point cloud block.
S604、将目标点云块的全局特征信息和边界特征信息输入全连接模块,得到判别器的目标判别结果。S604. Input the global feature information and boundary feature information of the target point cloud block into the fully connected module to obtain the target discrimination result of the discriminator.
例如,将目标点云块的全局特征信息和边界特征信息进行级联;将级联后的全局特征信息和边界特征信息输入全连接模块,得到判别器的目标判别结果。For example, the global feature information and boundary feature information of the target point cloud block are concatenated; the concatenated global feature information and boundary feature information are input into the fully connected module to obtain the target discrimination result of the discriminator.
本申请实施例的判别器可以理解为双头判别器,可以实现对全局和边界两个维度上的判断,进而提高了判别准确性。The discriminator in the embodiment of the present application can be understood as a double-headed discriminator, which can realize the judgment on the two dimensions of the whole world and the boundary, thereby improving the accuracy of the judgment.
具体的,为了高效地抑制生成点云的边界噪点,对于每个输入目标点云
Figure PCTCN2021096287-appb-000022
首先提取目标点云块的边界点,例如通过高通图滤波器提取目标点云块的R个边界点
Figure PCTCN2021096287-appb-000023
R<N’,然后显式地把完整的目标点云块和P b送入图12所示的双头判别器中,分别得到全局判别模块输出的目标点云块的全局特征信息,以及边界判别模块输出的目标点云块的边界判别特征,将目标点云块的全局特征信息和边界特征信息输入全连接模块,得到判别器的目标判别结果。
Specifically, in order to efficiently suppress the boundary noise of the generated point cloud, for each input target point cloud
Figure PCTCN2021096287-appb-000022
First extract the boundary points of the target point cloud block, for example, extract the R boundary points of the target point cloud block through a high-pass image filter
Figure PCTCN2021096287-appb-000023
R<N', and then explicitly send the complete target point cloud block and P b into the double-headed discriminator shown in Figure 12, and obtain the global feature information of the target point cloud block output by the global discriminant module, and the boundary The boundary discrimination feature of the target point cloud block output by the discriminant module, the global feature information and boundary feature information of the target point cloud block are input into the full connection module, and the target discrimination result of the discriminator is obtained.
本申请实施例中,判别器得到上述第一判别结果、第二判别结果和第三判别结果的过程一致。In the embodiment of the present application, the discriminator obtains the first discriminant result, the second discriminant result, and the third discriminant result in the same process.
其中,若目标点云块为生成器上采样后的训练点云块,且判断器经过该训练点云块训练,则上述目标判别结果为第一判别结果。若目标点云块为生成器上采样后的训练点云块,且判断器未经过该训练点云块训练,则上述目标判别结果为第二判别结果。若目标点云块为训练点云块的上采样真值,则目标判别结果为第三判别结果。也就是说,首先使用该训练点云块对判别器进行训练,对判别器的训练过程是,将该训练点云块输入未经该训练点云块训练的生成器中,该生成器生成上采样后的训练点云块1,将生成器生成的上采样后的训练点云块1输入未经该训练点云块训练的判断器中,判断器输出第二判断结果。接着,将该训练点云块的上采样真值输入该判别器中,该判别器输出第三判别结果, 根据第二判别结果和第三判别结果对判别器的参数矩阵进行更新,实现对判别器的一次训练。接着,将生成器生成上采样后的训练点云块1输入上述经过该训练点云块训练后的判别器中,该判别器输出第一判别结果。Wherein, if the target point cloud block is a training point cloud block sampled by the generator, and the judger is trained by the training point cloud block, then the above target discrimination result is the first discrimination result. If the target point cloud block is a training point cloud block sampled by the generator, and the judger has not been trained by the training point cloud block, then the above target discrimination result is the second discrimination result. If the target point cloud block is the upsampled true value of the training point cloud block, the target discrimination result is the third discrimination result. That is to say, first use the training point cloud block to train the discriminator, the training process of the discriminator is to input the training point cloud block into the generator that has not been trained by the training point cloud block, and the generator generates the above For the sampled training point cloud block 1, the up-sampled training point cloud block 1 generated by the generator is input into a judger that has not been trained by the training point cloud block, and the judger outputs a second judgment result. Next, input the upsampling true value of the training point cloud block into the discriminator, and the discriminator outputs the third discriminant result, and update the parameter matrix of the discriminator according to the second discriminant result and the third discriminant result to realize the discriminant A training session of the machine. Next, the upsampled training point cloud block 1 generated by the generator is input into the above-mentioned discriminator trained by the training point cloud block, and the discriminator outputs the first discrimination result.
下面分别为判别器中的全局判别模块和边界判别模块分别进行介绍。The following are the global discrimination module and the boundary discrimination module in the discriminator respectively.
在一些实施例中,如图13所示,全局判别模块沿着网络深度方向依次包括:第一数量个多层感知机、第一最大池化层、第二自相关注意力网络、第二数量个多层感知机和第二最大池化层。此时,上述S603包括:In some embodiments, as shown in Figure 13, the global discriminant module sequentially includes along the network depth direction: a first number of multi-layer perceptrons, a first maximum pooling layer, a second autocorrelation attention network, a second number of A multilayer perceptron and a second max pooling layer. At this time, the above S603 includes:
S603-A1、将目标点云块的几何信息输入第一数量个多层感知机进行特征提取,得到目标点云块的第一全局特征信息;S603-A1. Input the geometric information of the target point cloud block into the first number of multi-layer perceptrons for feature extraction, and obtain the first global feature information of the target point cloud block;
S603-A2、将第一全局特征信息输入第一最大池化层进行降维处理,得到目标点云块的第二全局特征信息;S603-A2. Input the first global feature information into the first maximum pooling layer to perform dimension reduction processing, and obtain the second global feature information of the target point cloud block;
S603-A3、将第一全局特征信息和第二全局特征信息输入第二自相关注意力网络进行特征交互,得到目标点云块的第三全局特征信息;S603-A3. Input the first global feature information and the second global feature information into the second autocorrelation attention network to perform feature interaction, and obtain the third global feature information of the target point cloud block;
在一种可能的实现方式中,将第一全局特征信息和第二全局特征信息进行级联;将级联后的第一全局特征信息和第二全局特征信息,输入第二自相关注意力网络进行特征交互,得到目标点云块的第三全局特征信息。In a possible implementation, the first global feature information and the second global feature information are concatenated; the concatenated first global feature information and the second global feature information are input into the second autocorrelation attention network Perform feature interaction to obtain the third global feature information of the target point cloud block.
S603-A4、将第三全局特征信息输入第二数量个多层感知机进而特征提取,得到目标点云块的第四全局特征信息;S603-A4. Input the third global feature information into the second number of multi-layer perceptrons for feature extraction, and obtain the fourth global feature information of the target point cloud block;
S603-A5、将第四全局特征信息输入第二最大池化层进行降维处理,得到目标点云块的全局特征信息。S603-A5. Input the fourth global feature information into the second maximum pooling layer for dimensionality reduction processing to obtain the global feature information of the target point cloud block.
具体是,首先将目标点云块的几何信息输入第一数量个多层感知机(Multilayer perception,简称MLP)进行特征提取,得到目标点云块的第一全局特征信息;接着,将第一全局特征信息输入第一最大池化层进行降维处理,通过最大池操作得到目标点云块的第二全局特征信息;随后,将第一全局特征信息和第二全局特征信息输入第二自相关注意力(Self-attetion)网络进行特征交互,提升每个点之间的特征交互,得到目标点云块的第三全局特征信息;接着,将第三全局特征信息输入第二数量个多层感知机(MLP)进而特征提取,得到目标点云块的第四全局特征信息;最后,将第四全局特征信息输入第二最大池化层进行降维处理,得到目标点云块的全局特征信息。Specifically, first, the geometric information of the target point cloud is input into the first number of multilayer perception machines (Multilayer perception, MLP for short) for feature extraction, and the first global feature information of the target point cloud is obtained; then, the first global The feature information is input into the first maximum pooling layer for dimension reduction processing, and the second global feature information of the target point cloud block is obtained through the maximum pooling operation; then, the first global feature information and the second global feature information are input into the second autocorrelation attention The force (Self-attetion) network performs feature interaction, improves the feature interaction between each point, and obtains the third global feature information of the target point cloud block; then, the third global feature information is input into the second number of multi-layer perceptrons (MLP) and further feature extraction to obtain the fourth global feature information of the target point cloud block; finally, input the fourth global feature information into the second maximum pooling layer for dimensionality reduction processing to obtain the global feature information of the target point cloud block.
可选的,第一数量等于第二数量。Optionally, the first quantity is equal to the second quantity.
可选的,第一数量与第二数量均等于2。Optionally, both the first quantity and the second quantity are equal to 2.
可选的,第一数量个多层感知机包括第一层多层感知机和第二层多层感知机,第二数量个多层感知机包括第三层多层感知机和第四层多层感知机,第一层多层感知机、第二层多层感知机、第三层多层感知机和第四层多层感知机的特征维度依次逐渐增加。Optionally, the first number of multilayer perceptrons includes a first layer of multilayer perceptrons and a second layer of multilayer perceptrons, and the second number of multilayer perceptrons includes a third layer of multilayer perceptrons and a fourth layer of multilayer perceptrons. Layer perceptron, the feature dimension of the first layer of multi-layer perceptron, the second layer of multi-layer perceptron, the third layer of multi-layer perceptron and the fourth layer of multi-layer perceptron gradually increases.
可选的,第一层多层感知机的特征维度为32,第二层多层感知机的特征维度为64,第三层多层感知机的特征维度为128,第四层多层感知机的特征维度为256。Optionally, the feature dimension of the first layer of multi-layer perceptron is 32, the feature dimension of the second layer of multi-layer perceptron is 64, the feature dimension of the third layer of multi-layer perceptron is 128, and the fourth layer of multi-layer perceptron The feature dimension of is 256.
在一些实施例中,继续参照如图13所示,边界判别模块沿着网络深度方向依次包括:第三数量个多层感知机、第三最大池化层、第三自相关注意力网络、第四数量个多层感知机和第四最大池化层,此时,上述S602包括:In some embodiments, continuing to refer to FIG. 13 , the boundary discrimination module sequentially includes along the network depth direction: a third number of multi-layer perceptrons, a third maximum pooling layer, a third autocorrelation attention network, a third Four multi-layer perceptrons and the fourth maximum pooling layer, at this time, the above S602 includes:
S602-A1、将目标点云块的边界点的几何信息输入第三数量个多层感知机中进行特征提取,得到目标点云块的第一边界特征信息;S602-A1. Input the geometric information of the boundary points of the target point cloud block into a third number of multi-layer perceptrons for feature extraction, and obtain the first boundary feature information of the target point cloud block;
S602-A2、将第一边界特征信息输入第三最大池化层进行降维处理,得到目标点云块的第二边界特征信息;S602-A2. Input the first boundary feature information into the third maximum pooling layer to perform dimension reduction processing, and obtain the second boundary feature information of the target point cloud block;
S602-A3、将第一边界特征信息和第二边界特征信息输入第三自相关注意力网络进行特征交互,得到目标点云块的第三边界特征信息。S602-A3. Input the first boundary feature information and the second boundary feature information into the third autocorrelation attention network to perform feature interaction, and obtain the third boundary feature information of the target point cloud block.
在一种可能实现方式中,S602-A3包括:将第一边界特征信息和第二边界特征信息进行级联;将级联后的第一边界特征信息和第二边界特征信息,输入第三自相关注意力网络进行特征交互,得到目标点云块的第三边界特征信息。In a possible implementation manner, S602-A3 includes: concatenating the first boundary feature information and the second boundary feature information; inputting the concatenated first boundary feature information and the second boundary feature information into a third self- The relevant attention network performs feature interaction to obtain the third boundary feature information of the target point cloud block.
S602-A4、将第三边界特征信息输入第四数量个多层感知机进行特征提取,得到目标点云块的第 四边界特征信息;S602-A4. Input the third boundary feature information into the fourth number of multi-layer perceptrons for feature extraction, and obtain the fourth boundary feature information of the target point cloud block;
S602-A5、将第四边界特征信息输入第四最大池化层进行降维处理,得到目标点云块的边界特征信息。S602-A5. Input the fourth boundary feature information into the fourth maximum pooling layer to perform dimensionality reduction processing to obtain boundary feature information of the target point cloud block.
具体是,首先将目标点云块的边界几何信息输入第三数量个多层感知机(MLP)进行特征提取,得到目标点云块的第一边界特征信息;接着,将第一边界特征信息输入第三最大池化层进行降维处理,通过最大池操作得到目标点云块的第二边界特征信息;随后,将第一边界特征信息和第二边界特征信息输入第三自相关注意力(Self-attetion)网络进行特征交互,提升每个点之间的特征交互,得到目标点云块的第三边界特征信息;接着,将第三边界特征信息输入第四数量个多层感知机(MLP)进而特征提取,得到目标点云块的第四边界特征信息;最后,将第四边界特征信息输入第四最大池化层进行降维处理,得到目标点云块的边界特征信息。Specifically, first input the boundary geometric information of the target point cloud block into a third number of multi-layer perceptrons (MLP) for feature extraction, and obtain the first boundary feature information of the target point cloud block; then, input the first boundary feature information The third maximum pooling layer performs dimension reduction processing, and obtains the second boundary feature information of the target point cloud block through the maximum pooling operation; then, the first boundary feature information and the second boundary feature information are input into the third autocorrelation attention (Self -attetion) network for feature interaction, enhance the feature interaction between each point, and obtain the third boundary feature information of the target point cloud block; then, input the third boundary feature information into the fourth number of multi-layer perceptrons (MLP) Further feature extraction is performed to obtain the fourth boundary feature information of the target point cloud block; finally, the fourth boundary feature information is input into the fourth maximum pooling layer for dimensionality reduction processing to obtain the boundary feature information of the target point cloud block.
可选的,第三数量等于第四数量。Optionally, the third quantity is equal to the fourth quantity.
可选的,第三数量与第四数量均等于2。Optionally, both the third quantity and the fourth quantity are equal to 2.
可选的,第三数量个多层感知机包括第五层多层感知机和第六层多层感知机,第四数量个多层感知机包括第七层多层感知机和第八层多层感知机,第五层多层感知机、第六层多层感知机、第七层多层感知机和第八层多层感知机的特征维度依次逐渐增加。Optionally, the third number of multi-layer perceptrons includes a fifth-layer multi-layer perceptron and a sixth-layer multi-layer perceptron, and the fourth number of multi-layer perceptrons includes a seventh-layer multi-layer perceptron and an eighth-layer multi-layer perceptron. Layer perceptron, the feature dimension of the fifth layer multilayer perceptron, sixth layer multilayer perceptron, seventh layer multilayer perceptron and eighth layer multilayer perceptron gradually increases.
可选的,第八层多层感知机的特征维度大于或等于第七层多层感知机的特征维度,且小于或等于第四层多层感知机的特征维度。例如,第八层多层感知机的特征维度大于或等于128且小于或等于256。Optionally, the feature dimension of the eighth-layer multi-layer perceptron is greater than or equal to the feature dimension of the seventh-layer multi-layer perceptron, and smaller than or equal to the feature dimension of the fourth-layer multi-layer perceptron. For example, the feature dimension of the eighth-layer multilayer perceptron is greater than or equal to 128 and less than or equal to 256.
可选的,第五层多层感知机的特征维度为32,第六层多层感知机的特征维度为64,第七层多层感知机的特征维度为128,第八层多层感知机的特征维度为192。Optionally, the feature dimension of the fifth-layer multi-layer perceptron is 32, the feature dimension of the sixth-layer multi-layer perceptron is 64, the feature dimension of the seventh-layer multi-layer perceptron is 128, and the eighth-layer multi-layer perceptron The feature dimension of is 192.
继续参照图13所示,将全局判别模块输出的目标点云块的全局特征信息和边界判别模块输出的边界特征信息进行级联,将连接在一起的全局特征信息和边界特征信息输入全连接层模块,通过3个全连接层(FC)获得判别器的置信值,即判别器的判别结果,若判别器输入的是生成器所输出的上采样点云,则置信值接近0,若判别器输入的为点云的上采样真值,则置信值接近1。Continuing to refer to Figure 13, the global feature information of the target point cloud block output by the global discrimination module and the boundary feature information output by the boundary discrimination module are cascaded, and the connected global feature information and boundary feature information are input into the fully connected layer The module obtains the confidence value of the discriminator through three fully connected layers (FC), that is, the discriminant result of the discriminator. If the input of the discriminator is the upsampled point cloud output by the generator, the confidence value is close to 0. If the discriminator The input is the upsampled true value of the point cloud, and the confidence value is close to 1.
基于此,可以根据判别器的判别结果,来监督生成器的训练,进而提高了生成器的训练准确性,使得训练完成的生成器上采样后的点云的分布接近点云的上采样真值,保证上采样后的点云的准确性。Based on this, the training of the generator can be supervised according to the discrimination results of the discriminator, thereby improving the training accuracy of the generator, so that the distribution of the upsampled point cloud of the trained generator is close to the true value of the upsampled point cloud , to ensure the accuracy of the upsampled point cloud.
本申请实施例,为了提高判别器的判断准确性,提出一种新的判别器,该判别器包括全局判别模块和边界判别模块,分布对点云的全局信息和边界信息进行判别,进而提高判别器的判别准确性,从而提高了使用该判别器辅助训练生成器时,提高生成器的训练精度。In the embodiment of the present application, in order to improve the judgment accuracy of the discriminator, a new discriminator is proposed. The discriminator includes a global discrimination module and a boundary discrimination module, and discriminates the global information and boundary information of the point cloud, thereby improving the discrimination. The discriminative accuracy of the discriminator improves the training accuracy of the generator when the discriminator is used to assist the training of the generator.
上文对生成器的训练过程进行了介绍,下面对使用训练好的生成器进行点云的几何信息的上采样过程进行介绍。上述训练好的生成器可以实现对点云的几何信息进行上采样。The training process of the generator is introduced above, and the upsampling process of the geometric information of the point cloud using the trained generator is introduced below. The above trained generator can realize the upsampling of the geometric information of the point cloud.
图14为本申请实施例提供的点云上采样方法的流程示意图,如图14所示,点云上采样过程包括:Fig. 14 is a schematic flow chart of the point cloud upsampling method provided by the embodiment of the present application. As shown in Fig. 14, the point cloud upsampling process includes:
S701、获取待上采样点云的几何信息。S701. Acquire geometric information of the point cloud to be upsampled.
可选的,该待上采样点云可以为点云采集设备实时采集的。Optionally, the point cloud to be upsampled may be collected in real time by a point cloud collection device.
可选的,上述待上采样点云可以是从其他存储设备中获取的。Optionally, the point cloud to be upsampled may be obtained from other storage devices.
可选的,上述待上采样点云为解码设备从编辑设备获取的码流中解码出的。Optionally, the point cloud to be upsampled is decoded by the decoding device from the code stream obtained by the editing device.
本申请实施例对获取待处理的点云的具体过程不做限制。The embodiment of the present application does not limit the specific process of obtaining the point cloud to be processed.
S702、根据待上采样点云的几何信息,将待上采样点云划分成至少一个点云块。S702. Divide the point cloud to be upsampled into at least one point cloud block according to the geometric information of the point cloud to be upsampled.
在一些实施例中,上述S702中将待上采样点云划分为至少一个点云块的方式包括但不限于如下几种方式:In some embodiments, the methods of dividing the point cloud to be upsampled into at least one point cloud block in S702 include but are not limited to the following methods:
方式一,根据待上采样点云的几何信息,将待上采样点云划分成至少一个大小相等的点云块。也就是说每个点云块的几何尺度相同。Method 1: Divide the point cloud to be upsampled into at least one point cloud block of equal size according to the geometric information of the point cloud to be upsampled. That is to say, the geometric scale of each point cloud block is the same.
方式二,根据待上采样点云的几何信息,将待上采样点云划分为至少一个点云块,每个点云块中包括相同数量个点。Method 2: Divide the point cloud to be upsampled into at least one point cloud block according to the geometric information of the point cloud to be upsampled, and each point cloud block includes the same number of points.
方式三,根据待上采样点云的几何信息,从待上采样点云中获取至少一个种子点,例如采用蒙特 卡洛随机采样法随机地从待上采样点云中采样指定个数的种子点。对于每个种子点,确定该种子点的邻近点,将该种子点与该种子点的邻近点划分为一个点云块,进而得到至少一个点云块。在该方式三中,得到点云块也称为点云补丁(Patch),该方式得到的点云块中每个点云块所包括的点的个数相同。Method 3: Obtain at least one seed point from the point cloud to be upsampled according to the geometric information of the point cloud to be upsampled, for example, use Monte Carlo random sampling method to randomly sample a specified number of seed points from the point cloud to be upsampled . For each seed point, determine the neighboring points of the seed point, divide the seed point and the neighboring points of the seed point into a point cloud block, and then obtain at least one point cloud block. In the third method, the obtained point cloud blocks are also called point cloud patches (Patch), and the number of points included in each point cloud block in the obtained point cloud blocks is the same.
S703、将点云块的几何信息输入生成器中进行上采样,得到点云块的上采样几何信息。S703. Input the geometric information of the point cloud block into the generator for up-sampling, and obtain the up-sampling geometric information of the point cloud block.
图15为本申请实施例涉及的生成器的一种网络结构示意图,如图15所示,生成器包括:特征提取模块、特征上采样模块和几何生成模块,其中,特征提取模块用于提取点云块的第一特征信息,特征采样模块用于将点云块的第一特征信息上采样为第二特征信息,几何生成模块用于将点云块的第二特征信息映射至几何空间中,以得到点云块的上采样几何信息。Fig. 15 is a schematic diagram of a network structure of the generator involved in the embodiment of the present application. As shown in Fig. 15, the generator includes: a feature extraction module, a feature upsampling module and a geometry generation module, wherein the feature extraction module is used to extract points The first feature information of the cloud block, the feature sampling module is used to upsample the first feature information of the point cloud block into the second feature information, and the geometry generation module is used to map the second feature information of the point cloud block into the geometric space, In order to obtain the upsampled geometric information of the point cloud block.
下面对生成器中特征提取模块、特征上采样模块和几何生成模块的网络结构进行介绍。The network structure of the feature extraction module, feature upsampling module and geometry generation module in the generator is introduced below.
在一些实施例中,如图6A所示,特征提取模块包括密集连接的M个特征提取块;In some embodiments, as shown in Figure 6A, the feature extraction module includes densely connected M feature extraction blocks;
对于M个特征提取块中的第i+1个特征提取块,第i+1个特征提取块用于根据输入的第i个第四特征信息输出第i+1个第三特征信息,第i个第四特征信息是根据第i个特征提取块输出的第i个第三特征信息确定的,点云块的第一特征信息是根据M个特征提取块中第M个特征提取块所输出的第M个第三特征信息确定的,i为小于M的正整数,具体参照上述上述S403的描述,在此不再赘述。For the i+1th feature extraction block in the M feature extraction blocks, the i+1th feature extraction block is used to output the i+1th third feature information according to the input i-th fourth feature information, and the ith The fourth feature information is determined according to the i-th third feature information output by the i-th feature extraction block, and the first feature information of the point cloud block is based on the output of the M-th feature extraction block in the M feature extraction blocks As determined by the Mth third feature information, i is a positive integer smaller than M. For details, refer to the description of S403 above, and details will not be repeated here.
在一些实施例中,若i不等于1,则第i个第四特征信息为M个特征提取块中位于第i个特征提取块之前的各特征提取块所提取的第三特征信息、与第i个特征提取块所提取的第三特征信息进行级联后的特征信息。若i等于1,则第i个第四特征信息为M个特征提取块中第一个特征提取块所输出的第一个第三特征信息。In some embodiments, if i is not equal to 1, the i-th fourth feature information is the third feature information extracted by each feature extraction block before the i-th feature extraction block in the M feature extraction blocks, and The feature information obtained by cascading the third feature information extracted by the i feature extraction blocks. If i is equal to 1, the i-th fourth feature information is the first third feature information output by the first feature extraction block in the M feature extraction blocks.
在一些实施例中,如图6B所示,特征提取块包括:第一特征提取单元和串联连接的S个第二特征提取单元,S为正整数;In some embodiments, as shown in Figure 6B, the feature extraction block includes: a first feature extraction unit and S second feature extraction units connected in series, where S is a positive integer;
对于第i+1个特征提取块中的第一提取单元,第一提取单元用于针对点云块中的当前点,搜索当前点的K个邻近点,并基于点云块的第i个第四特征信息,将当前点的第四特征信息与邻近点的第四特征信息进行相减,得到K个残差特征信息,并将K个残差特征信息与当前点的第四特征信息进行级联,得到当前点的第i个级联特征信息,根据当前点的第i个级联特征信息,得到点云块的第i个级联特征信息,并将点云块的第i个级联特征信息输入S个第二特征提取单元中的第一个第二特征提取单元;For the first extraction unit in the i+1th feature extraction block, the first extraction unit is used to search for K neighboring points of the current point for the current point in the point cloud block, and based on the i-th feature extraction unit of the point cloud block Four feature information, the fourth feature information of the current point is subtracted from the fourth feature information of the adjacent point to obtain K residual feature information, and the K residual feature information is graded with the fourth feature information of the current point According to the i-th cascade feature information of the current point, get the i-th cascade feature information of the point cloud block, and concatenate the i-th cascade feature information of the point cloud block The feature information is input to the first second feature extraction unit in the S second feature extraction units;
第一个第二特征提取单元用于根据点云块的第i个级联特征信息,输出第一个第五特征信息至第二个第二特征提取单元,其中点云块的第i+1个第三特征信息为S个第二特征提取单元中最后一个第二特征提取单元输出的第五特征信息,具体参照上述上述S403-A31的描述,在此不再赘述。The first second feature extraction unit is used to output the first fifth feature information to the second second feature extraction unit according to the ith cascaded feature information of the point cloud block, wherein the i+1th of the point cloud block The third feature information is the fifth feature information output by the last second feature extraction unit among the S second feature extraction units. For details, refer to the above-mentioned description of S403-A31, which will not be repeated here.
在一些实施例中,如图6C所示,第二特征提取单元包括P个残差块,P为正整数;In some embodiments, as shown in FIG. 6C, the second feature extraction unit includes P residual blocks, where P is a positive integer;
对于第s个第二特征提取单元中的第j+1个残差块,第j+1个残差块用于根据第s个第二特征提取单元中的第j个残差块所输出的第j个第一残差信息和输入第s个第二特征提取单元的第五特征信息,输出第j+1个第一残差信息,其中,j为小于P的正整数,s为小于或等于S的正整数。可选的,将第s个第二特征提取单元中的第j个残差块所输出的第j个第一残差信息和输入第s个第二特征提取单元的第五特征信息进行相加后,输入第s个第二特征提取单元中的第j+1个残差块。For the j+1th residual block in the sth second feature extraction unit, the j+1th residual block is used according to the output of the jth residual block in the sth second feature extraction unit The j-th first residual information and the fifth feature information input to the s-th second feature extraction unit output the j+1-th first residual information, where j is a positive integer less than P, and s is less than or A positive integer equal to S. Optionally, adding the j-th first residual information output by the j-th residual block in the s-th second feature extraction unit to the fifth feature information input to the s-th second feature extraction unit After that, input the j+1th residual block in the sth second feature extraction unit.
第s个第二特征提取单元输出的第五特征信息是根据第s个第二特征提取单元中至少一个残差块输出的第一残差信息信息,以及输入第s个第二特征提取单元的第五特征信息确定的。The fifth feature information output by the sth second feature extraction unit is based on the first residual information information output by at least one residual block in the sth second feature extraction unit, and input to the sth second feature extraction unit. The fifth feature information is determined.
在一种可能的实现方式中,上述第s个第二特征提取单元输出的第五特征信息是根据第s个第二特征提取单元中最后一个残差块输出的第一残差信息信息、与P-1个残差块中至少一个残差块输出的第一残差信息信息进行级联后的特征信息、与输入第s个第二特征提取单元的第五特征信息确定的,其中,P-1个残差块为第s个第二特征提取单元的P个残差块中除最后一个残差块之外的残差块。In a possible implementation manner, the fifth feature information output by the sth second feature extraction unit is based on the first residual information output by the last residual block in the sth second feature extraction unit, and The feature information after concatenation of the first residual information output by at least one residual block in the P-1 residual blocks and the fifth feature information input to the s-th second feature extraction unit are determined, wherein, P The -1 residual block is a residual block except the last residual block among the P residual blocks of the s-th second feature extraction unit.
在一种可能的实现方式中,上述第s个第二特征提取单元输出的第五特征信息是根据第s个第二特征提取单元中最后一个残差块输出的第一残差信息信息、与P-1个残差块中至少一个残差块输出的第一残差信息信息进行级联后特征信息、与输入第s个第二特征提取单元的第五特征信息进行相加后确定的。In a possible implementation manner, the fifth feature information output by the sth second feature extraction unit is based on the first residual information output by the last residual block in the sth second feature extraction unit, and The first residual information output by at least one residual block in the P-1 residual blocks is concatenated and then determined by adding the fifth feature information input to the s-th second feature extraction unit.
在一些实施例中,如图6F所示,第二特征提取单元还包括门控单元,In some embodiments, as shown in FIG. 6F, the second feature extraction unit further includes a gating unit,
对于第s个第二特征提取单元中的门控单元,该门控单元用于对第s个第二特征提取单元中的最 后一个残差块输出的第一残差信息、与P-1个残差块中至少一个残差块输出的第一残差信息级联后的特征信息进行去冗余,输出去冗余后的特征信息;第s个第二特征提取单元输出的第五特征信息是根据去冗余后的特征信息和输入第s个第二特征提取单元的第五特征信息进行相加后确定的。For the gating unit in the sth second feature extraction unit, the gating unit is used for the first residual information output by the last residual block in the sth second feature extraction unit, and P-1 De-redundancy is performed on the feature information after concatenation of the first residual information output by at least one residual block in the residual block, and the de-redundant feature information is output; the fifth feature information output by the sth second feature extraction unit It is determined after adding the feature information after de-redundancy to the fifth feature information input to the sth second feature extraction unit.
上文结合图6A至图6F对生成器中的特征提取模块的网络结构进行了介绍,下面结合图7A至图7D对生成器中的特征上采样模块的网络结构进行介绍。The network structure of the feature extraction module in the generator is introduced above with reference to FIGS. 6A to 6F , and the network structure of the feature upsampling module in the generator is introduced below with reference to FIGS. 7A to 7D .
在一些实施例中,如图7A所示,特征上采样模块包括:特征上采样子模块和特征提取子模块;In some embodiments, as shown in FIG. 7A, the feature upsampling module includes: a feature upsampling submodule and a feature extraction submodule;
其中,特征上采样子模块用于按照预设的上采样率r,将点云块的第一特征信息复制r份,并对复制后的第一特征信息在特征维度上增加一个n维向量,得到点云块的上采样特征信息,并将点云块的上采样特征信息输入特征提取子模块,其中不同第一特征信息对应的n维向量的值不相同;Wherein, the feature upsampling submodule is used to copy r copies of the first feature information of the point cloud block according to the preset upsampling rate r, and add an n-dimensional vector to the feature dimension of the copied first feature information, Obtain the upsampling feature information of the point cloud block, and input the upsampling feature information of the point cloud block into the feature extraction submodule, wherein the values of the n-dimensional vectors corresponding to different first feature information are different;
特征提取子模块用于根据点云块的上采样特征信息,输出点云块的第二特征信息。The feature extraction sub-module is used to output the second feature information of the point cloud block according to the up-sampled feature information of the point cloud block.
在一些实施例中,如图7C所示,上述特征提取子模块包括Q个第三特征提取单元,Q为正整数;In some embodiments, as shown in FIG. 7C, the feature extraction submodule includes Q third feature extraction units, where Q is a positive integer;
针对Q个第三特征提取单元中的第k+1个第三特征提取单元,第k+1个第三特征提取单元用于根据第k个第三特征提取单元所提取的点云块的第k个增强上采样特征信息,输出点云块的第k+1个增强上采样特征信息,k为小于Q的正整数;For the k+1th third feature extraction unit among the Q third feature extraction units, the k+1th third feature extraction unit is used to extract the point cloud block according to the kth third feature extraction unit. k enhanced upsampling feature information, the k+1th enhanced upsampling feature information of the output point cloud block, k is a positive integer less than Q;
点云块的第二特征信息为Q个第三特征提取单元中最后一个第三特征提取单元所提取的点云块的第Q个增强上采样特征信息。The second feature information of the point cloud block is the Qth enhanced upsampling feature information of the point cloud block extracted by the last third feature extraction unit among the Q third feature extraction units.
在一些实施例中,如图7D所示,第三特征提取单元为图7D中的HRA,第三特征提取单元包括L个残差块,L为正整数,例如第三特征提取单元包括4个残差块RB;In some embodiments, as shown in Figure 7D, the third feature extraction unit is the HRA in Figure 7D, the third feature extraction unit includes L residual blocks, L is a positive integer, for example, the third feature extraction unit includes 4 residual block RB;
对于第k+1个第三特征提取单元中的第l+1个残差块,第l+1个残差块用于根据第k+1个第三特征提取单元中的第l个残差块输出的第l个第二残差信息和输入第k+1个第三特征提取单元的第k个增强上采样特征信息,输出第l+1个第二残差信息,l为小于L的正整数;可选的,对第l个残差块输出的第l个第二残差信息和第k个增强上采样特征信息进行相加后,输入第l+1个残差块。For the l+1th residual block in the k+1th third feature extraction unit, the l+1th residual block is used according to the lth residual in the k+1th third feature extraction unit The lth second residual information output by the block and the kth enhanced upsampling feature information input to the k+1th third feature extraction unit, and the l+1th second residual information is output, where l is less than L Positive integer; optionally, after adding the l-th second residual information output by the l-th residual block and the k-th enhanced upsampling feature information, input the l+1th residual block.
点云块的第k+1个增强上采样特征信息是根据第k+1个第三特征提取单元中至少一个残差块输出的第二残差信息,以及第k个增强上采样特征信息确定的。The k+1th enhanced upsampling feature information of the point cloud block is determined according to the second residual information output by at least one residual block in the k+1th third feature extraction unit, and the kth enhanced upsampling feature information of.
在一种可能的实现方式中,上述点云块的第k+1个增强上采样特征信息根据L个残差块中最后一个残差块输出的第二残差信息、与L-1个残差块中至少一个残差块输出的第二残差信息进行级联后的特征信息和第k个增强上采样特征信息确定的,其中,L-1个残差块为第k+1个第三特征提取单元的L个残差块中除最后一个残差块之外的残差块。In a possible implementation, the k+1th enhanced upsampled feature information of the above point cloud block is based on the second residual information output by the last residual block in the L residual blocks, and the L-1 residual The second residual information output by at least one residual block in the difference block is determined by concatenating the feature information and the kth enhanced upsampling feature information, wherein, the L-1 residual block is the k+1th A residual block except the last residual block among the L residual blocks of the three-feature extraction unit.
在一种可能的实现方式中,上述点云块的第k+1个增强上采样特征信息根据L个残差块中最后一个残差块输出的第二残差信息、与L-1个残差块中至少一个残差块输出的第二残差信息进行级联后的特征信息和第k个增强上采样特征信息进行相加后确定的。In a possible implementation, the k+1th enhanced upsampled feature information of the above point cloud block is based on the second residual information output by the last residual block in the L residual blocks, and the L-1 residual It is determined by adding the concatenated feature information of the second residual information output by at least one residual block in the difference block and the kth enhanced upsampling feature information.
在一些实施例中,如图7D所示,第三特征提取单元还包括门控单元;In some embodiments, as shown in FIG. 7D, the third feature extraction unit further includes a gating unit;
对于第k+1个第三特征提取单元中的门控单元,门控单元用于对第k+1个第三特征提取单元中的最后一个残差块输出的第二残差信息、与L-1个残差块中至少一个残差块输出的第二特征信息进行级联后的特征信息进行去冗余,输出去冗余后的特征信息;For the gating unit in the k+1th third feature extraction unit, the gating unit is used to output the second residual information of the last residual block in the k+1th third feature extraction unit, and L - The second feature information output by at least one residual block in the 1 residual block is de-redundant after concatenating the feature information, and outputting the feature information after de-redundancy;
点云块的第k+1个增强上采样特征信息是根据去冗余后的特征信息与第k个增强上采样特征信息进行相加后确定的。The k+1th enhanced upsampling feature information of the point cloud block is determined after adding the deredundant feature information to the kth enhanced upsampling feature information.
在一些实施例中,第三特征提取单元与上述第二特征提取单元的网络结构相同。In some embodiments, the network structure of the third feature extraction unit is the same as that of the above-mentioned second feature extraction unit.
在一些实施例中,如图7B所示,特征上采样模块还包括第一自相关注意力网络;In some embodiments, as shown in Figure 7B, the feature upsampling module further includes a first autocorrelation attention network;
其中,第一自相关注意力网络用于对特征上采样子模块输出的点云块的上采样特征信息进行特征交互,输出特征交互后的点云块的上采样特征信息至特征提取子模块;Wherein, the first autocorrelation attention network is used to perform feature interaction on the upsampling feature information of the point cloud block output by the feature upsampling submodule, and output the upsampling feature information of the point cloud block after feature interaction to the feature extraction submodule;
此时,特征提取子模块用于根据特征交互后的点云块的上采样特征信息,输出点云块的第二特征信息。At this time, the feature extraction submodule is configured to output second feature information of the point cloud block according to the upsampled feature information of the point cloud block after feature interaction.
可选的,特征交互后的点云块的上采样特征信息的特征维度低于点云块的上采样特征信息的特征维度。Optionally, the feature dimension of the upsampled feature information of the point cloud block after the feature interaction is lower than the feature dimension of the upsampled feature information of the point cloud block.
上文结合图7A至图7D对生成器中的特征提取模块的网络结构进行了介绍,下面结合图8和图15对生成器中的几何生成模块的网络结构进行介绍。The network structure of the feature extraction module in the generator is introduced above with reference to FIG. 7A to FIG. 7D , and the network structure of the geometry generation module in the generator is introduced below in conjunction with FIG. 8 and FIG. 15 .
在一些实施例中,几何生成模块包括多个全连接层;In some embodiments, the geometry generation module includes a plurality of fully connected layers;
该多个全连接层用于根据点云块的第二特征信息,输出点云块的上采样几何信息。The multiple fully connected layers are used to output upsampled geometric information of the point cloud block according to the second feature information of the point cloud block.
在一些实施例中,如图8和图15所示,几何生成模块包括:几何重建单元、滤波单元和下采样单元;In some embodiments, as shown in Figure 8 and Figure 15, the geometry generation module includes: a geometry reconstruction unit, a filtering unit and a downsampling unit;
其中,几何重建单元用于对将点云块的第二特征信息进行几何重建,输出点云块的初始上采样几何信息至滤波单元;Wherein, the geometric reconstruction unit is used to geometrically reconstruct the second feature information of the point cloud block, and output the initial upsampling geometric information of the point cloud block to the filtering unit;
滤波单元用于对点云块的初始上采样几何信息进行除噪,输出点云块滤除噪点的初始上采样几何信息至下采样单元;The filter unit is used to denoise the initial upsampling geometric information of the point cloud block, and output the initial upsampling geometric information of the point cloud block to filter out the noise to the downsampling unit;
下采样单元用于对点云块滤除噪点的初始上采样几何信息下采样至目标上采样率,输出点云块的上采样几何信息。The down-sampling unit is used for down-sampling the initial up-sampling geometric information of the point cloud block to filter noise to a target up-sampling rate, and output the up-sampling geometric information of the point cloud block.
可选的,目标上采样率小于或等于特征上采样模块的上采样率。Optionally, the target upsampling rate is less than or equal to the upsampling rate of the feature upsampling module.
为了进一步说明本申请的技术效果,将本申请实施例提出的方案在测试平台上进行实现,且分别使用倒角距离(CD),豪斯多夫距离(HD),以及点到面距离(P2FD)来衡量上采样的点云与点云上采样真值之间的相似程度。可选的,上采样率r设置为4。将本申请实施例的技术方案与基于优化的方法EAR、最先进的点云上采样网络PU-Net、MPU、PU-GAN分别进行测试,在测试数据集上的结果如表1所示:In order to further illustrate the technical effect of the present application, the scheme proposed in the embodiment of the present application is implemented on the test platform, and the chamfering distance (CD), the Hausdorff distance (HD), and the point-to-surface distance (P2FD) are used respectively ) to measure the similarity between the upsampled point cloud and the upsampled ground truth of the point cloud. Optionally, the upsampling rate r is set to 4. The technical solution of the embodiment of the application was tested with the optimization-based method EAR, the most advanced point cloud upsampling network PU-Net, MPU, and PU-GAN, and the results on the test data set are shown in Table 1:
表1Table 1
Figure PCTCN2021096287-appb-000024
Figure PCTCN2021096287-appb-000024
如表1所示,本申请提出的方法生成的点云与点云的上采样真值之间的差异最小,例如,倒角距离(Chamfer distance,简称CD)、豪斯多夫距离(Hausdorff distance,简称HD)、以及点到面距离(Point to face distance,简称P2FD)分别为0.258、3.571和2.392,因此,说明本申请提出的点云上采样方法可以实现点云的有效上采样。As shown in Table 1, the difference between the point cloud generated by the method proposed in this application and the upsampled true value of the point cloud is the smallest, for example, Chamfer distance (CD for short), Hausdorff distance (Hausdorff distance , referred to as HD), and the point-to-face distance (Point to face distance, referred to as P2FD) are 0.258, 3.571 and 2.392 respectively. Therefore, it is shown that the point cloud upsampling method proposed in this application can realize effective upsampling of the point cloud.
本申请实施例的点云上采样方法,通过获取待上采样点云的几何信息,根据待上采样点云的几何信息,将待上采样点云划分成至少一个点云块,将点云块的几何信息输入生成器中进行上采样,得到点云块的上采样几何信息,其中生成器包括:特征提取模块、特征上采样模块和几何生成模块,特征提取模块用于提取点云块的第一特征信息,特征采样模块用于将点云块的第一特征信息上采样为第二特征信息,几何生成模块用于将点云块的第二特征信息映射至几何空间中,以得到点云块的上采样几何信息。即本申请实施例的生成器为基于深度学习的生成器,通过深度学习可以学习到点云的更多特征信息,进而使用该生成器进行点云上采样时,可以生成高精度的点云,且该高精度的点云的特征与点云的上采样真值接近,进而提高了点云上采样的准确性。The point cloud upsampling method of the embodiment of the present application obtains the geometric information of the point cloud to be upsampled, divides the point cloud to be upsampled into at least one point cloud block according to the geometric information of the point cloud to be upsampled, and divides the point cloud block into The geometric information of the input generator is upsampled, and the upsampled geometric information of the point cloud block is obtained. The generator includes: a feature extraction module, a feature upsampling module and a geometry generation module. The feature extraction module is used to extract the first point cloud block. A feature information, the feature sampling module is used to upsample the first feature information of the point cloud block into the second feature information, and the geometry generation module is used to map the second feature information of the point cloud block into the geometric space to obtain the point cloud The block's upsampled geometry information. That is, the generator in the embodiment of the present application is a generator based on deep learning, through which more characteristic information of the point cloud can be learned, and then when the generator is used for upsampling of the point cloud, a high-precision point cloud can be generated, Moreover, the feature of the high-precision point cloud is close to the true value of the upsampling of the point cloud, thereby improving the accuracy of the upsampling of the point cloud.
在一些实施例中,本申请实施例提供的点云的上采样方法还可以应用于点云编解码框架中,例如可以应用于点云解码端。In some embodiments, the point cloud upsampling method provided in the embodiment of the present application can also be applied to a point cloud encoding and decoding framework, for example, it can be applied to a point cloud decoding end.
图16为本申请实施例提供的点云解码方法的流程示意图,如图16所示,点云解码方法包括:Fig. 16 is a schematic flow chart of the point cloud decoding method provided by the embodiment of the present application. As shown in Fig. 16, the point cloud decoding method includes:
S801、解码点云码流,得到点云的几何信息。S801. Decode the point cloud code stream to obtain geometric information of the point cloud.
点云码流包括属性码流和几何码流,解码几何码流可以得到点云的几何信息,解码属性码流可以得到点云的属性信息。The point cloud code stream includes attribute code stream and geometry code stream. By decoding the geometry code stream, the geometric information of the point cloud can be obtained, and by decoding the attribute code stream, the attribute information of the point cloud can be obtained.
其中解码几何码流,得到点云的几何信息的过程参照已有技术,本申请实施例在此不再赘述。The process of decoding the geometric code stream and obtaining the geometric information of the point cloud refers to the prior art, and will not be repeated in this embodiment of the present application.
S802、根据点云的几何信息,将点云划分成至少一个点云块。S802. Divide the point cloud into at least one point cloud block according to the geometric information of the point cloud.
上述S802的执行过程与上述S702一致,参照上述S702的描述,在此不再赘述。The execution process of the above S802 is consistent with that of the above S702, refer to the description of the above S702, and will not be repeated here.
S803、将点云块的几何信息输入生成器中进行上采样,得到点云块的上采样几何信息。S803. Input the geometric information of the point cloud block into the generator for up-sampling, and obtain the up-sampled geometric information of the point cloud block.
参照上述图15所示的生成器,该生成器包括:特征提取模块、特征上采样模块和几何生成模块,特征提取模块用于提取点云块的第一特征信息,特征采样模块用于将点云块的第一特征信息上采样为第二特征信息,几何生成模块用于将点云块的第二特征信息映射至几何空间中,以得到点云块的上采样几何信息。Referring to the above-mentioned generator shown in Figure 15, the generator includes: a feature extraction module, a feature upsampling module and a geometry generation module, the feature extraction module is used to extract the first feature information of the point cloud block, and the feature sampling module is used to extract the point cloud block The first feature information of the cloud block is upsampled to the second feature information, and the geometry generation module is used to map the second feature information of the point cloud block into a geometric space, so as to obtain the upsampled geometric information of the point cloud block.
下面对生成器中特征提取模块、特征上采样模块和几何生成模块的网络结构进行介绍。The network structure of the feature extraction module, feature upsampling module and geometry generation module in the generator is introduced below.
首先,结合图6A至图6F对特征提取模块的网络结构进行介绍。First, the network structure of the feature extraction module is introduced with reference to FIG. 6A to FIG. 6F .
在一些实施例中,如图6A所示,特征提取模块包括密集连接的M个特征提取块;In some embodiments, as shown in Figure 6A, the feature extraction module includes densely connected M feature extraction blocks;
对于M个特征提取块中的第i+1个特征提取块,第i+1个特征提取块用于根据输入的第i个第四特征信息输出第i+1个第三特征信息,第i个第四特征信息是根据第i个特征提取块输出的第i个第三特征信息确定的,点云块的第一特征信息是根据M个特征提取块中第M个特征提取块所输出的第M个第三特征信息确定的,i为小于M的正整数,具体参照上述上述S403的描述,在此不再赘述。For the i+1th feature extraction block in the M feature extraction blocks, the i+1th feature extraction block is used to output the i+1th third feature information according to the input i-th fourth feature information, and the ith The fourth feature information is determined according to the i-th third feature information output by the i-th feature extraction block, and the first feature information of the point cloud block is based on the output of the M-th feature extraction block in the M feature extraction blocks As determined by the Mth third feature information, i is a positive integer smaller than M. For details, refer to the description of S403 above, and details will not be repeated here.
在一些实施例中,若i不等于1,则第i个第四特征信息为M个特征提取块中位于第i个特征提取块之前的各特征提取块所提取的第三特征信息、与第i个特征提取块所提取的第三特征信息进行级联后的特征信息。若i等于1,则第i个第四特征信息为M个特征提取块中第一个特征提取块所输出的第一个第三特征信息。In some embodiments, if i is not equal to 1, the i-th fourth feature information is the third feature information extracted by each feature extraction block before the i-th feature extraction block in the M feature extraction blocks, and The feature information obtained by cascading the third feature information extracted by the i feature extraction blocks. If i is equal to 1, the i-th fourth feature information is the first third feature information output by the first feature extraction block in the M feature extraction blocks.
在一些实施例中,如图6B所示,特征提取块包括:第一特征提取单元和串联连接的S个第二特征提取单元,S为正整数;In some embodiments, as shown in Figure 6B, the feature extraction block includes: a first feature extraction unit and S second feature extraction units connected in series, where S is a positive integer;
对于第i+1个特征提取块中的第一提取单元,第一提取单元用于针对点云块中的当前点,搜索当前点的K个邻近点,并基于点云块的第i个第四特征信息,将当前点的第四特征信息与邻近点的第四特征信息进行相减,得到K个残差特征信息,并将K个残差特征信息与当前点的第四特征信息进行级联,得到当前点的第i个级联特征信息,根据当前点的第i个级联特征信息,得到点云块的第i个级联特征信息,并将点云块的第i个级联特征信息输入S个第二特征提取单元中的第一个第二特征提取单元;For the first extraction unit in the i+1th feature extraction block, the first extraction unit is used to search for K neighboring points of the current point for the current point in the point cloud block, and based on the i-th feature extraction unit of the point cloud block Four feature information, the fourth feature information of the current point is subtracted from the fourth feature information of the adjacent point to obtain K residual feature information, and the K residual feature information is graded with the fourth feature information of the current point According to the i-th cascade feature information of the current point, get the i-th cascade feature information of the point cloud block, and concatenate the i-th cascade feature information of the point cloud block The feature information is input to the first second feature extraction unit in the S second feature extraction units;
第一个第二特征提取单元用于根据点云块的第i个级联特征信息,输出第一个第五特征信息至第二个第二特征提取单元,其中点云块的第i+1个第三特征信息为S个第二特征提取单元中最后一个第二特征提取单元输出的第五特征信息,具体参照上述上述S403-A31的描述,在此不再赘述。The first second feature extraction unit is used to output the first fifth feature information to the second second feature extraction unit according to the ith cascaded feature information of the point cloud block, wherein the i+1th of the point cloud block The third feature information is the fifth feature information output by the last second feature extraction unit among the S second feature extraction units. For details, refer to the above-mentioned description of S403-A31, which will not be repeated here.
在一些实施例中,如图6C所示,第二特征提取单元包括P个残差块,P为正整数;In some embodiments, as shown in FIG. 6C, the second feature extraction unit includes P residual blocks, where P is a positive integer;
对于第s个第二特征提取单元中的第j+1个残差块,第j+1个残差块用于根据第s个第二特征提取单元中的第j个残差块所输出的第j个第一残差信息和输入第s个第二特征提取单元的第五特征信息,输出第j+1个第一残差信息,其中,j为小于P的正整数,s为小于或等于S的正整数。可选的,将第s个第二特征提取单元中的第j个残差块所输出的第j个第一残差信息和输入第s个第二特征提取单元的第五特征信息进行相加后,输入第s个第二特征提取单元中的第j+1个残差块。For the j+1th residual block in the sth second feature extraction unit, the j+1th residual block is used according to the output of the jth residual block in the sth second feature extraction unit The j-th first residual information and the fifth feature information input to the s-th second feature extraction unit output the j+1-th first residual information, where j is a positive integer less than P, and s is less than or A positive integer equal to S. Optionally, adding the j-th first residual information output by the j-th residual block in the s-th second feature extraction unit to the fifth feature information input to the s-th second feature extraction unit After that, input the j+1th residual block in the sth second feature extraction unit.
第s个第二特征提取单元输出的第五特征信息是根据第s个第二特征提取单元中至少一个残差块输出的第一残差信息信息,以及输入第s个第二特征提取单元的第五特征信息确定的。The fifth feature information output by the sth second feature extraction unit is based on the first residual information information output by at least one residual block in the sth second feature extraction unit, and input to the sth second feature extraction unit. The fifth feature information is determined.
在一种可能的实现方式中,上述第s个第二特征提取单元输出的第五特征信息是根据第s个第二特征提取单元中最后一个残差块输出的第一残差信息信息、与P-1个残差块中至少一个残差块输出的第一残差信息信息进行级联后的特征信息、与输入第s个第二特征提取单元的第五特征信息确定的,其中,P-1个残差块为第s个第二特征提取单元的P个残差块中除最后一个残差块之外的残差块。In a possible implementation manner, the fifth feature information output by the sth second feature extraction unit is based on the first residual information output by the last residual block in the sth second feature extraction unit, and The feature information after concatenation of the first residual information output by at least one residual block in the P-1 residual blocks and the fifth feature information input to the s-th second feature extraction unit are determined, wherein, P The -1 residual block is a residual block except the last residual block among the P residual blocks of the s-th second feature extraction unit.
在一种可能的实现方式中,上述第s个第二特征提取单元输出的第五特征信息是根据第s个第二特征提取单元中最后一个残差块输出的第一残差信息信息、与P-1个残差块中至少一个残差块输出的第一残差信息信息进行级联后特征信息、与输入第s个第二特征提取单元的第五特征信息进行相加后确定的。In a possible implementation manner, the fifth feature information output by the sth second feature extraction unit is based on the first residual information output by the last residual block in the sth second feature extraction unit, and The first residual information output by at least one residual block in the P-1 residual blocks is concatenated and then determined by adding the fifth feature information input to the s-th second feature extraction unit.
在一些实施例中,如图6F所示,第二特征提取单元还包括门控单元,In some embodiments, as shown in FIG. 6F, the second feature extraction unit further includes a gating unit,
对于第s个第二特征提取单元中的门控单元,该门控单元用于对第s个第二特征提取单元中的最 后一个残差块输出的第一残差信息、与P-1个残差块中至少一个残差块输出的第一残差信息级联后的特征信息进行去冗余,输出去冗余后的特征信息;第s个第二特征提取单元输出的第五特征信息是根据去冗余后的特征信息和输入第s个第二特征提取单元的第五特征信息进行相加后确定的。For the gating unit in the sth second feature extraction unit, the gating unit is used for the first residual information output by the last residual block in the sth second feature extraction unit, and P-1 De-redundancy is performed on the feature information after concatenation of the first residual information output by at least one residual block in the residual block, and the de-redundant feature information is output; the fifth feature information output by the sth second feature extraction unit It is determined after adding the feature information after de-redundancy to the fifth feature information input to the sth second feature extraction unit.
上文结合图6A至图6F对生成器中的特征提取模块的网络结构进行了介绍,下面结合图7A至图7D对生成器中的特征上采样模块的网络结构进行介绍。The network structure of the feature extraction module in the generator is introduced above with reference to FIGS. 6A to 6F , and the network structure of the feature upsampling module in the generator is introduced below with reference to FIGS. 7A to 7D .
在一些实施例中,如图7A所示,特征上采样模块包括:特征上采样子模块和特征提取子模块;In some embodiments, as shown in FIG. 7A, the feature upsampling module includes: a feature upsampling submodule and a feature extraction submodule;
其中,特征上采样子模块用于按照预设的上采样率r,将点云块的第一特征信息复制r份,并对复制后的第一特征信息在特征维度上增加一个n维向量,得到点云块的上采样特征信息,并将点云块的上采样特征信息输入特征提取子模块,其中不同第一特征信息对应的n维向量的值不相同;Wherein, the feature upsampling submodule is used to copy r copies of the first feature information of the point cloud block according to the preset upsampling rate r, and add an n-dimensional vector to the feature dimension of the copied first feature information, Obtain the upsampling feature information of the point cloud block, and input the upsampling feature information of the point cloud block into the feature extraction submodule, wherein the values of the n-dimensional vectors corresponding to different first feature information are different;
特征提取子模块用于根据点云块的上采样特征信息,输出点云块的第二特征信息。The feature extraction sub-module is used to output the second feature information of the point cloud block according to the up-sampled feature information of the point cloud block.
在一些实施例中,如图7C所示,上述特征提取子模块包括Q个第三特征提取单元,Q为正整数;In some embodiments, as shown in FIG. 7C, the feature extraction submodule includes Q third feature extraction units, where Q is a positive integer;
针对Q个第三特征提取单元中的第k+1个第三特征提取单元,第k+1个第三特征提取单元用于根据第k个第三特征提取单元所提取的点云块的第k个增强上采样特征信息,输出点云块的第k+1个增强上采样特征信息,k为小于Q的正整数;For the k+1th third feature extraction unit among the Q third feature extraction units, the k+1th third feature extraction unit is used to extract the point cloud block according to the kth third feature extraction unit. k enhanced upsampling feature information, the k+1th enhanced upsampling feature information of the output point cloud block, k is a positive integer less than Q;
点云块的第二特征信息为Q个第三特征提取单元中最后一个第三特征提取单元所提取的点云块的第Q个增强上采样特征信息。The second feature information of the point cloud block is the Qth enhanced upsampling feature information of the point cloud block extracted by the last third feature extraction unit among the Q third feature extraction units.
在一些实施例中,如图7D所示,第三特征提取单元为图7D中的HRA,第三特征提取单元包括L个残差块,L为正整数,例如第三特征提取单元包括4个残差块RB;In some embodiments, as shown in Figure 7D, the third feature extraction unit is the HRA in Figure 7D, the third feature extraction unit includes L residual blocks, L is a positive integer, for example, the third feature extraction unit includes 4 residual block RB;
对于第k+1个第三特征提取单元中的第l+1个残差块,第l+1个残差块用于根据第k+1个第三特征提取单元中的第l个残差块输出的第l个第二残差信息和输入第k+1个第三特征提取单元的第k个增强上采样特征信息,输出第l+1个第二残差信息,l为小于L的正整数;可选的,对第l个残差块输出的第l个第二残差信息和第k个增强上采样特征信息进行相加后,输入第l+1个残差块。For the l+1th residual block in the k+1th third feature extraction unit, the l+1th residual block is used according to the lth residual in the k+1th third feature extraction unit The lth second residual information output by the block and the kth enhanced upsampling feature information input to the k+1th third feature extraction unit, and the l+1th second residual information is output, where l is less than L Positive integer; optionally, after adding the l-th second residual information output by the l-th residual block and the k-th enhanced upsampling feature information, input the l+1th residual block.
点云块的第k+1个增强上采样特征信息是根据第k+1个第三特征提取单元中至少一个残差块输出的第二残差信息,以及第k个增强上采样特征信息确定的。The k+1th enhanced upsampling feature information of the point cloud block is determined according to the second residual information output by at least one residual block in the k+1th third feature extraction unit, and the kth enhanced upsampling feature information of.
在一种可能的实现方式中,上述点云块的第k+1个增强上采样特征信息根据L个残差块中最后一个残差块输出的第二残差信息、与L-1个残差块中至少一个残差块输出的第二残差信息进行级联后的特征信息和第k个增强上采样特征信息确定的,其中,L-1个残差块为第k+1个第三特征提取单元的L个残差块中除最后一个残差块之外的残差块。In a possible implementation, the k+1th enhanced upsampled feature information of the above point cloud block is based on the second residual information output by the last residual block in the L residual blocks, and the L-1 residual The second residual information output by at least one residual block in the difference block is determined by concatenating the feature information and the kth enhanced upsampling feature information, wherein, the L-1 residual block is the k+1th A residual block except the last residual block among the L residual blocks of the three-feature extraction unit.
在一种可能的实现方式中,上述点云块的第k+1个增强上采样特征信息根据L个残差块中最后一个残差块输出的第二残差信息、与L-1个残差块中至少一个残差块输出的第二残差信息进行级联后的特征信息和第k个增强上采样特征信息进行相加后确定的。In a possible implementation, the k+1th enhanced upsampled feature information of the above point cloud block is based on the second residual information output by the last residual block in the L residual blocks, and the L-1 residual It is determined by adding the concatenated feature information of the second residual information output by at least one residual block in the difference block and the kth enhanced upsampling feature information.
在一些实施例中,如图7D所示,第三特征提取单元还包括门控单元;In some embodiments, as shown in FIG. 7D, the third feature extraction unit further includes a gating unit;
对于第k+1个第三特征提取单元中的门控单元,门控单元用于对第k+1个第三特征提取单元中的最后一个残差块输出的第二残差信息、与L-1个残差块中至少一个残差块输出的第二特征信息进行级联后的特征信息进行去冗余,输出去冗余后的特征信息;For the gating unit in the k+1th third feature extraction unit, the gating unit is used to output the second residual information of the last residual block in the k+1th third feature extraction unit, and L - The second feature information output by at least one residual block in the 1 residual block is de-redundant after concatenating the feature information, and outputting the feature information after de-redundancy;
点云块的第k+1个增强上采样特征信息是根据去冗余后的特征信息与第k个增强上采样特征信息进行相加后确定的。The k+1th enhanced upsampling feature information of the point cloud block is determined after adding the deredundant feature information to the kth enhanced upsampling feature information.
在一些实施例中,第三特征提取单元与上述第二特征提取单元的网络结构相同。In some embodiments, the network structure of the third feature extraction unit is the same as that of the above-mentioned second feature extraction unit.
在一些实施例中,如图7B所示,特征上采样模块还包括第一自相关注意力网络;In some embodiments, as shown in Figure 7B, the feature upsampling module further includes a first autocorrelation attention network;
其中,第一自相关注意力网络用于对特征上采样子模块输出的点云块的上采样特征信息进行特征交互,输出特征交互后的点云块的上采样特征信息至特征提取子模块;Wherein, the first autocorrelation attention network is used to perform feature interaction on the upsampling feature information of the point cloud block output by the feature upsampling submodule, and output the upsampling feature information of the point cloud block after feature interaction to the feature extraction submodule;
此时,特征提取子模块用于根据特征交互后的点云块的上采样特征信息,输出点云块的第二特征信息。At this time, the feature extraction submodule is configured to output second feature information of the point cloud block according to the upsampled feature information of the point cloud block after feature interaction.
可选的,特征交互后的点云块的上采样特征信息的特征维度低于点云块的上采样特征信息的特征维度。Optionally, the feature dimension of the upsampled feature information of the point cloud block after the feature interaction is lower than the feature dimension of the upsampled feature information of the point cloud block.
上文结合图7A至图7D对生成器中的特征提取模块的网络结构进行了介绍,下面结合图8和图15对生成器中的几何生成模块的网络结构进行介绍。The network structure of the feature extraction module in the generator is introduced above with reference to FIG. 7A to FIG. 7D , and the network structure of the geometry generation module in the generator is introduced below in conjunction with FIG. 8 and FIG. 15 .
在一些实施例中,几何生成模块包括多个全连接层;In some embodiments, the geometry generation module includes a plurality of fully connected layers;
该多个全连接层用于根据点云块的第二特征信息,输出点云块的上采样几何信息。The multiple fully connected layers are used to output upsampled geometric information of the point cloud block according to the second feature information of the point cloud block.
在一些实施例中,如图8和图15所示,几何生成模块包括:几何重建单元、滤波单元和下采样单元;In some embodiments, as shown in Figure 8 and Figure 15, the geometry generation module includes: a geometry reconstruction unit, a filtering unit and a downsampling unit;
其中,几何重建单元用于对将点云块的第二特征信息进行几何重建,输出点云块的初始上采样几何信息至滤波单元;Wherein, the geometric reconstruction unit is used to geometrically reconstruct the second feature information of the point cloud block, and output the initial upsampling geometric information of the point cloud block to the filtering unit;
滤波单元用于对点云块的初始上采样几何信息进行除噪,输出点云块滤除噪点的初始上采样几何信息至下采样单元;The filter unit is used to denoise the initial upsampling geometric information of the point cloud block, and output the initial upsampling geometric information of the point cloud block to filter out the noise to the downsampling unit;
下采样单元用于对点云块滤除噪点的初始上采样几何信息下采样至目标上采样率,输出点云块的上采样几何信息。The down-sampling unit is used for down-sampling the initial up-sampling geometric information of the point cloud block to filter noise to a target up-sampling rate, and output the up-sampling geometric information of the point cloud block.
可选的,目标上采样率小于或等于特征上采样模块的上采样率。Optionally, the target upsampling rate is less than or equal to the upsampling rate of the feature upsampling module.
可选的,上述目标上采样率为预设值。Optionally, the target upsampling rate is a preset value.
可选的,上述目标上采样率为从点云码流中解析出。Optionally, the above target upsampling rate is parsed from the point cloud code stream.
本申请实施例对点云解码端生成的点云的几何信息进行上采样,生成高精度的重建点云,可以满足高精度点云的应用场景,进而提高了点云解码的多样性。The embodiment of the present application upsamples the geometric information of the point cloud generated by the point cloud decoding end to generate a high-precision reconstruction point cloud, which can meet the application scenarios of high-precision point clouds, and further improves the diversity of point cloud decoding.
应理解,图4至图16仅为本申请的示例,不应理解为对本申请的限制。It should be understood that Fig. 4 to Fig. 16 are only examples of the present application, and should not be construed as limiting the present application.
以上结合附图详细描述了本申请的优选实施方式,但是,本申请并不限于上述实施方式中的具体细节,在本申请的技术构思范围内,可以对本申请的技术方案进行多种简单变型,这些简单变型均属于本申请的保护范围。例如,在上述具体实施方式中所描述的各个具体技术特征,在不矛盾的情况下,可以通过任何合适的方式进行组合,为了避免不必要的重复,本申请对各种可能的组合方式不再另行说明。又例如,本申请的各种不同的实施方式之间也可以进行任意组合,只要其不违背本申请的思想,其同样应当视为本申请所公开的内容。The preferred embodiments of the present application have been described in detail above in conjunction with the accompanying drawings. However, the present application is not limited to the specific details in the above embodiments. Within the scope of the technical concept of the present application, various simple modifications can be made to the technical solutions of the present application. These simple modifications all belong to the protection scope of the present application. For example, the various specific technical features described in the above specific implementation manners can be combined in any suitable manner if there is no contradiction. Separately. As another example, any combination of various implementations of the present application can also be made, as long as they do not violate the idea of the present application, they should also be regarded as the content disclosed in the present application.
还应理解,在本申请的各种方法实施例中,上述各过程的序号的大小并不意味着执行顺序的先后,各过程的执行顺序应以其功能和内在逻辑确定,而不应对本申请实施例的实施过程构成任何限定。另外,本申请实施例中,术语“和/或”,仅仅是一种描述关联对象的关联关系,表示可以存在三种关系。具体地,A和/或B可以表示:单独存在A,同时存在A和B,单独存在B这三种情况。另外,本文中字符“/”,一般表示前后关联对象是一种“或”的关系。It should also be understood that in the various method embodiments of the present application, the sequence numbers of the above-mentioned processes do not mean the order of execution, and the order of execution of the processes should be determined by their functions and internal logic, and should not be used in this application. The implementation of the examples constitutes no limitation. In addition, in the embodiment of the present application, the term "and/or" is only an association relationship describing associated objects, indicating that there may be three relationships. Specifically, A and/or B may mean: A exists alone, A and B exist simultaneously, and B exists alone. In addition, the character "/" in this article generally indicates that the contextual objects are an "or" relationship.
上文结合图4至图16,详细描述了本申请的方法实施例,下文结合图17至图20,详细描述本申请的装置实施例。The method embodiment of the present application is described in detail above with reference to FIG. 4 to FIG. 16 , and the device embodiment of the present application is described in detail below in conjunction with FIG. 17 to FIG. 20 .
图17是本申请实施例提供的点云解码器的示意性框图。Fig. 17 is a schematic block diagram of a point cloud decoder provided by an embodiment of the present application.
如图17所示,该点点云解码器20可包括:As shown in Figure 17, this point point cloud decoder 20 can comprise:
解码单元21,用于解码点云码流,得到点云的几何信息;The decoding unit 21 is used to decode the point cloud code stream to obtain the geometric information of the point cloud;
划分单元22,用于根据所述点云的几何信息,将所述点云划分成至少一个点云块;A division unit 22, configured to divide the point cloud into at least one point cloud block according to the geometric information of the point cloud;
上采样单元23,用于将所述点云块的几何信息输入生成器中进行上采样,得到所述点云块的上采样几何信息;An up-sampling unit 23, configured to input the geometric information of the point cloud block into the generator for up-sampling, to obtain the up-sampled geometric information of the point cloud block;
其中,所述生成器包括:特征提取模块、特征上采样模块和几何生成模块,所述特征提取模块用于提取所述点云块的第一特征信息,所述特征采样模块用于将所述点云块的第一特征信息上采样为第二特征信息,所述几何生成模块用于将所述点云块的第二特征信息映射至几何空间中,以得到所述点云块的上采样几何信息。Wherein, the generator includes: a feature extraction module, a feature upsampling module and a geometry generation module, the feature extraction module is used to extract the first feature information of the point cloud block, and the feature sampling module is used to extract the The first feature information of the point cloud block is up-sampled to the second feature information, and the geometry generation module is used to map the second feature information of the point cloud block into a geometric space, so as to obtain the up-sampling of the point cloud block geometric information.
在一些实施例中,所述特征提取模块包括密集连接的M个特征提取块;In some embodiments, the feature extraction module comprises densely connected M feature extraction blocks;
对于所述M个特征提取块中的第i+1个特征提取块,所述第i+1个特征提取块用于根据输入的第i个第四特征信息输出第i+1个第三特征信息,所述第i个第四特征信息是根据第i个特征提取块输出的第i个第三特征信息确定的,所述点云块的第一特征信息是根据所述M个特征提取块中第M个特征提取块所输出的第M个第三特征信息确定的,所述i为小于M的正整数。For the i+1th feature extraction block in the M feature extraction blocks, the i+1th feature extraction block is used to output the i+1th third feature according to the input i-th fourth feature information information, the i-th fourth feature information is determined according to the i-th third feature information output by the i-th feature extraction block, and the first feature information of the point cloud block is determined according to the M feature extraction blocks Determined by the Mth third feature information output by the Mth feature extraction block, the i is a positive integer smaller than M.
在一些实施例中,若i不等于1,则所述第i个第四特征信息为所述M个特征提取块中位于所述第i个特征提取块之前的各特征提取块所提取的第三特征信息、与所述第i个特征提取块所提取的第三特征信息进行级联后的特征信息;In some embodiments, if i is not equal to 1, the i-th fourth feature information is the first feature extracted by each feature extraction block before the i-th feature extraction block among the M feature extraction blocks. Three feature information, feature information concatenated with the third feature information extracted by the ith feature extraction block;
若i等于1,则所述第i个第四特征信息为所述M个特征提取块中第一个特征提取块所输出的第一个第三特征信息。If i is equal to 1, the ith fourth feature information is the first third feature information output by the first feature extraction block in the M feature extraction blocks.
在一些实施例中,所述特征提取块包括:第一特征提取单元和串联连接的S个第二特征提取单元,所述S为正整数;In some embodiments, the feature extraction block includes: a first feature extraction unit and S second feature extraction units connected in series, wherein S is a positive integer;
对于所述第i+1个特征提取块中的第一提取单元,所述第一提取单元用于针对所述点云块中的当前点,搜索所述当前点的K个邻近点,并基于所述点云块的第i个第四特征信息,将所述当前点的第四特征信息与所述邻近点的第四特征信息进行相减,得到K个残差特征信息,并将所述K个残差特征信息与所述当前点的第四特征信息进行级联,得到所述当前点的第i个级联特征信息,根据所述当前点的第i个级联特征信息,得到所述点云块的第i个级联特征信息,并将所述点云块的第i个级联特征信息输入所述S个第二特征提取单元中的第一个第二特征提取单元;For the first extraction unit in the i+1th feature extraction block, the first extraction unit is used to search for K neighboring points of the current point for the current point in the point cloud block, and based on For the ith fourth feature information of the point cloud block, the fourth feature information of the current point is subtracted from the fourth feature information of the adjacent point to obtain K residual feature information, and the The K residual feature information is concatenated with the fourth feature information of the current point to obtain the i-th concatenated feature information of the current point, and according to the i-th concatenated feature information of the current point, the The i-th concatenated feature information of the point cloud block, and input the i-th concatenated feature information of the point cloud block into the first second feature extraction unit in the S second feature extraction units;
所述第一个第二特征提取单元用于根据所述点云块的第i个级联特征信息,输出第一个第五特征信息至第二个第二特征提取单元,其中所述点云块的第i+1个第三特征信息为所述S个第二特征提取单元中最后一个第二特征提取单元输出的第五特征信息。The first second feature extraction unit is used to output the first fifth feature information to the second second feature extraction unit according to the i-th cascaded feature information of the point cloud block, wherein the point cloud The i+1th third feature information of the block is the fifth feature information output by the last second feature extraction unit among the S second feature extraction units.
在一些实施例中,所述第二特征提取单元包括P个残差块,所述P为正整数;In some embodiments, the second feature extraction unit includes P residual blocks, where P is a positive integer;
对于第s个第二特征提取单元中的第j+1个残差块,所述第j+1个残差块用于根据所述第s个第二特征提取单元中的第j个残差块所输出的第j个第一残差信息和输入所述第s个第二特征提取单元的第五特征信息,输出第j+1个第一残差信息,其中,所述j为小于P的正整数,所述s为小于或等于S的正整数;For the j+1th residual block in the sth second feature extraction unit, the j+1th residual block is used according to the jth residual in the sth second feature extraction unit The j-th first residual information output by the block and the fifth feature information input to the s-th second feature extraction unit, and the j+1-th first residual information is output, wherein the j is less than P A positive integer, the s is a positive integer less than or equal to S;
所述第s个第二特征提取单元输出的第五特征信息是根据所述第s个第二特征提取单元中至少一个残差块输出的第一残差信息信息,以及输入所述第s个第二特征提取单元的第五特征信息确定的。The fifth feature information output by the sth second feature extraction unit is based on the first residual information output by at least one residual block in the sth second feature extraction unit, and input to the sth second feature extraction unit determined by the fifth feature information of the second feature extraction unit.
在一些实施例中,上采样单元23还用于将所述第s个第二特征提取单元中的第j个残差块所输出的第j个第一残差信息和输入所述第s个第二特征提取单元的第五特征信息进行相加后,输入所述第s个第二特征提取单元中的第j+1个残差块。In some embodiments, the up-sampling unit 23 is further configured to input the j-th first residual information and the j-th residual information output by the j-th residual block in the s-th second feature extraction unit to the s-th After the fifth feature information of the second feature extraction unit is added, it is input to the j+1th residual block in the sth second feature extraction unit.
在一些实施例中,所述第s个第二特征提取单元输出的第五特征信息是根据所述第s个第二特征提取单元中最后一个残差块输出的第一残差信息信息、与P-1个残差块中至少一个残差块输出的第一残差信息信息进行级联后的特征信息、与输入所述第s个第二特征提取单元的第五特征信息确定的,其中,所述P-1个残差块为所述第s个第二特征提取单元的P个残差块中除最后一个残差块之外的残差块。In some embodiments, the fifth feature information output by the s th second feature extraction unit is based on the first residual information output by the last residual block in the s th second feature extraction unit, and The feature information after the concatenation of the first residual information output by at least one residual block in the P-1 residual blocks is determined from the fifth feature information input to the s-th second feature extraction unit, wherein , the P-1 residual blocks are residual blocks except the last residual block among the P residual blocks of the s-th second feature extraction unit.
在一些实施例中,所述第s个第二特征提取单元输出的第五特征信息是根据所述第s个第二特征提取单元中最后一个残差块输出的第一残差信息信息、与P-1个残差块中至少一个残差块输出的第一残差信息信息进行级联后特征信息、与输入所述第s个第二特征提取单元的第五特征信息进行相加后确定的。In some embodiments, the fifth feature information output by the s th second feature extraction unit is based on the first residual information output by the last residual block in the s th second feature extraction unit, and The first residual information output by at least one residual block in the P-1 residual blocks is concatenated and then determined after being added to the fifth feature information input to the s-th second feature extraction unit of.
在一些实施例中,所述第二特征提取单元还包括门控单元,In some embodiments, the second feature extraction unit further includes a gating unit,
对于所述第s个第二特征提取单元中的门控单元,所述门控单元用于对所述第s个第二特征提取单元中的最后一个残差块输出的第一残差信息、与所述P-1个残差块中至少一个残差块输出的第一残差信息级联后的特征信息进行去冗余,输出去冗余后的特征信息;所述第s个第二特征提取单元输出的第五特征信息是根据所述去冗余后的特征信息和所述输入所述第s个第二特征提取单元的第五特征信息进行相加后确定的。For the gating unit in the s th second feature extraction unit, the gating unit is used for the first residual information output by the last residual block in the s th second feature extraction unit, De-redundancy is performed on the feature information concatenated with the first residual information output by at least one residual block in the P-1 residual blocks, and the de-redundant feature information is output; the s second The fifth feature information output by the feature extraction unit is determined after adding the de-redundant feature information and the fifth feature information input to the s-th second feature extraction unit.
在一些实施例中,所述特征上采样模块包括:特征上采样子模块和特征提取子模块;In some embodiments, the feature upsampling module includes: a feature upsampling submodule and a feature extraction submodule;
所述特征上采样子模块用于按照预设的上采样率r,将所述点云块的第一特征信息复制r份,并对复制后的第一特征信息在特征维度上增加一个n维向量,得到所述点云块的上采样特征信息,并将所述点云块的上采样特征信息输入特征提取子模块,其中不同第一特征信息对应的n维向量的值不相同;The feature upsampling submodule is used to copy r copies of the first feature information of the point cloud block according to the preset upsampling rate r, and add an n dimension to the feature dimension of the copied first feature information vector, obtain the upsampling feature information of the point cloud block, and input the upsampling feature information of the point cloud block into the feature extraction submodule, wherein the values of n-dimensional vectors corresponding to different first feature information are different;
所述特征提取子模块用于根据所述点云块的上采样特征信息,输出所述点云块的第二特征信息。The feature extraction submodule is configured to output second feature information of the point cloud block according to the upsampled feature information of the point cloud block.
在一些实施例中,所述特征提取子模块包括Q个第三特征提取单元,所述Q为正整数;In some embodiments, the feature extraction submodule includes Q third feature extraction units, where Q is a positive integer;
针对所述Q个第三特征提取单元中的第k+1个第三特征提取单元,所述第k+1个第三特征提取单元用于根据第k个第三特征提取单元所提取的所述点云块的第k个增强上采样特征信息,输出所述点云块的第k+1个增强上采样特征信息,所述k为小于Q的正整数;For the k+1th third feature extraction unit among the Q third feature extraction units, the k+1th third feature extraction unit is used to extract all The kth enhanced upsampling feature information of the point cloud block, output the k+1th enhanced upsampling feature information of the point cloud block, and the k is a positive integer less than Q;
所述点云块的第二特征信息为所述Q个第三特征提取单元中最后一个第三特征提取单元所提取的所述点云块的第Q个增强上采样特征信息。The second feature information of the point cloud block is the Qth enhanced upsampling feature information of the point cloud block extracted by the last third feature extraction unit among the Q third feature extraction units.
在一些实施例中,所述第三特征提取单元包括L个残差块,所述L为正整数;In some embodiments, the third feature extraction unit includes L residual blocks, and the L is a positive integer;
对于所述第k+1个第三特征提取单元中的第l+1个残差块,所述第l+1个残差块用于根据所述第k+1个第三特征提取单元中的第l个残差块输出的第l个第二残差信息和输入所述第k+1个第三特征提取单元的第k个增强上采样特征信息,输出第l+1个第二残差信息,所述l为小于L的正整数;For the l+1th residual block in the k+1th third feature extraction unit, the l+1th residual block is used according to the k+1th third feature extraction unit The lth second residual information output by the lth residual block and the kth enhanced upsampling feature information input to the k+1th third feature extraction unit, output the l+1th second residual difference information, the l is a positive integer less than L;
所述点云块的第k+1个增强上采样特征信息是根据所述第k+1个第三特征提取单元中至少一个残差块输出的第二残差信息,以及所述第k个增强上采样特征信息确定的。The k+1th enhanced upsampling feature information of the point cloud block is the second residual information output from at least one residual block in the k+1th third feature extraction unit, and the kth Enhanced upsampling feature information determined.
在一些实施例中,上采样单元23还用于:对所述第l个残差块输出的第l个第二残差信息和所述第k个增强上采样特征信息进行相加后,输入所述第l+1个残差块。In some embodiments, the upsampling unit 23 is further configured to: after adding the lth second residual information output by the lth residual block and the kth enhanced upsampling feature information, input The l+1th residual block.
在一些实施例中,所述点云块的第k+1个增强上采样特征信息根据所述L个残差块中最后一个残差块输出的第二残差信息、与L-1个残差块中至少一个残差块输出的第二残差信息进行级联后的特征信息和所述第k个增强上采样特征信息确定的,其中,所述L-1个残差块为所述第k+1个第三特征提取单元的L个残差块中除最后一个残差块之外的残差块。In some embodiments, the k+1th enhanced upsampling feature information of the point cloud block is based on the second residual information output by the last residual block in the L residual blocks, and the L-1 residual The second residual information output by at least one residual block in the difference block is determined by concatenating the feature information and the kth enhanced upsampling feature information, wherein the L-1 residual blocks are the A residual block except the last residual block among the L residual blocks of the k+1 third feature extraction unit.
在一些实施例中,所述点云块的第k+1个增强上采样特征信息根据所述L个残差块中最后一个残差块输出的第二残差信息、与L-1个残差块中至少一个残差块输出的第二残差信息进行级联后的特征信息和所述第k个增强上采样特征信息进行相加后确定的。In some embodiments, the k+1th enhanced upsampling feature information of the point cloud block is based on the second residual information output by the last residual block in the L residual blocks, and the L-1 residual The second residual information output by at least one residual block in the difference block is determined by adding the concatenated feature information to the kth enhanced upsampling feature information.
在一些实施例中,所述第三特征提取单元还包括门控单元;In some embodiments, the third feature extraction unit further includes a gating unit;
对于所述第k+1个第三特征提取单元中的门控单元,所述门控单元用于对所述第k+1个第三特征提取单元中的最后一个残差块输出的第二残差信息、与所述L-1个残差块中至少一个残差块输出的第二特征信息进行级联后的特征信息进行去冗余,输出去冗余后的特征信息;For the gating unit in the k+1th third feature extraction unit, the gating unit is used for the second output of the last residual block in the k+1th third feature extraction unit performing de-redundancy on the residual information and the feature information concatenated with the second feature information output by at least one of the L-1 residual blocks, and outputting the de-redundant feature information;
所述点云块的第k+1个增强上采样特征信息是根据去冗余后的特征信息与所述第k个增强上采样特征信息进行相加后确定的。The k+1th enhanced upsampling feature information of the point cloud block is determined after adding the deredundant feature information to the kth enhanced upsampling feature information.
在一些实施例中,所述特征上采样模块还包括第一自相关注意力网络;In some embodiments, the feature upsampling module further includes a first autocorrelation attention network;
所述第一自相关注意力网络用于对所述特征上采样子模块输出的所述点云块的上采样特征信息进行特征交互,输出特征交互后的所述点云块的上采样特征信息至所述特征提取子模块;The first autocorrelation attention network is used to perform feature interaction on the upsampling feature information of the point cloud block output by the feature upsampling submodule, and output the upsampling feature information of the point cloud block after feature interaction to the feature extraction submodule;
所述特征提取子模块用于根据特征交互后的所述点云块的上采样特征信息,输出所述点云块的第二特征信息。The feature extraction submodule is configured to output second feature information of the point cloud block according to the upsampled feature information of the point cloud block after feature interaction.
在一些实施例中,所述特征交互后的所述点云块的上采样特征信息的特征维度低于所述点云块的上采样特征信息的特征维度。In some embodiments, the feature dimension of the upsampled feature information of the point cloud block after the feature interaction is lower than the feature dimension of the upsampled feature information of the point cloud block.
在一些实施例中,所述几何生成模块包括多个全连接层;In some embodiments, the geometry generation module includes a plurality of fully connected layers;
所述多个全连接层用于根据所述点云块的第二特征信息,输出所述点云块的上采样几何信息。The multiple fully connected layers are used to output upsampled geometric information of the point cloud block according to the second feature information of the point cloud block.
在一些实施例中,所述几何生成模块包括:几何重建单元、滤波单元和下采样单元;In some embodiments, the geometry generation module includes: a geometry reconstruction unit, a filtering unit, and a downsampling unit;
所述几何重建单元用于对所述将所述点云块的第二特征信息进行几何重建,输出所述点云块的初始上采样几何信息至所述滤波单元;The geometric reconstruction unit is used to geometrically reconstruct the second feature information of the point cloud block, and output the initial upsampling geometric information of the point cloud block to the filtering unit;
所述滤波单元用于对所述点云块的初始上采样几何信息进行除噪,输出所述点云块滤除噪点的初始上采样几何信息至所述下采样单元;The filtering unit is used to denoise the initial upsampling geometric information of the point cloud block, and output the initial upsampling geometric information of the point cloud block to filter noise to the downsampling unit;
所述下采样单元用于对所述点云块滤除噪点的初始上采样几何信息下采样至目标上采样率,输出所述点云块的上采样几何信息。The down-sampling unit is configured to down-sample the initial up-sampled geometric information of the point cloud block after filtering noise to a target up-sampling rate, and output the up-sampled geometric information of the point cloud block.
在一些实施例中,所述目标上采样率小于或等于所述特征上采样模块的上采样率。In some embodiments, the target upsampling rate is less than or equal to the upsampling rate of the feature upsampling module.
在一些实施例中,所述解码单元21还用于:解码所述点云码流,得到所述目标上采样率。In some embodiments, the decoding unit 21 is further configured to: decode the point cloud code stream to obtain the target upsampling rate.
应理解,装置实施例与方法实施例可以相互对应,类似的描述可以参照方法实施例。为避免重复,此处不再赘述。具体地,图17所示的点云解码器20可以对应于执行本申请实施例的点云解码方法中 的相应主体,并且点云解码器20中的各个单元的前述和其它操作和/或功能分别为了实现点云解码方法中的相应流程,为了简洁,在此不再赘述。It should be understood that the device embodiment and the method embodiment may correspond to each other, and similar descriptions may refer to the method embodiment. To avoid repetition, details are not repeated here. Specifically, the point cloud decoder 20 shown in FIG. 17 may correspond to the corresponding subject in the point cloud decoding method of the embodiment of the present application, and the aforementioned and other operations and/or functions of each unit in the point cloud decoder 20 In order to realize the corresponding processes in the point cloud decoding method respectively, for the sake of brevity, details are not repeated here.
图18是本申请实施例提供的点云上采样装置的示意性框图。Fig. 18 is a schematic block diagram of a point cloud upsampling device provided by an embodiment of the present application.
如图18所示,模型训练装置40包括:As shown in Figure 18, the model training device 40 includes:
获取单元41,用于获取待上采样点云的几何信息;An acquisition unit 41, configured to acquire geometric information of the point cloud to be upsampled;
划分单元42,用于根据所述待上采样点云的几何信息,将所述待上采样点云划分成至少一个点云块;A division unit 42, configured to divide the point cloud to be upsampled into at least one point cloud block according to the geometric information of the point cloud to be upsampled;
上采样单元43,用于将所述点云块的几何信息输入生成器中进行上采样,得到所述点云块的上采样几何信息;An up-sampling unit 43, configured to input the geometric information of the point cloud block into the generator for up-sampling, to obtain the up-sampling geometric information of the point cloud block;
其中,所述生成器包括:特征提取模块、特征上采样模块和几何生成模块,所述特征提取模块用于提取所述点云块的第一特征信息,所述特征采样模块用于将所述点云块的第一特征信息上采样为第二特征信息,所述几何生成模块用于将所述点云块的第二特征信息映射至几何空间中,以得到所述点云块的上采样几何信息。Wherein, the generator includes: a feature extraction module, a feature upsampling module and a geometry generation module, the feature extraction module is used to extract the first feature information of the point cloud block, and the feature sampling module is used to extract the The first feature information of the point cloud block is up-sampled to the second feature information, and the geometry generation module is used to map the second feature information of the point cloud block into a geometric space, so as to obtain the up-sampling of the point cloud block geometric information.
在一些实施例中,所述特征提取模块包括密集连接的M个特征提取块;In some embodiments, the feature extraction module comprises densely connected M feature extraction blocks;
对于所述M个特征提取块中的第i+1个特征提取块,所述第i+1个特征提取块用于根据输入的第i个第四特征信息输出第i+1个第三特征信息,所述第i个第四特征信息是根据第i个特征提取块输出的第i个第三特征信息确定的,所述点云块的第一特征信息是根据所述M个特征提取块中第M个特征提取块所输出的第M个第三特征信息确定的,所述i为小于M的正整数。For the i+1th feature extraction block in the M feature extraction blocks, the i+1th feature extraction block is used to output the i+1th third feature according to the input i-th fourth feature information information, the i-th fourth feature information is determined according to the i-th third feature information output by the i-th feature extraction block, and the first feature information of the point cloud block is determined according to the M feature extraction blocks Determined by the Mth third feature information output by the Mth feature extraction block, the i is a positive integer smaller than M.
在一些实施例中,若i不等于1,则所述第i个第四特征信息为所述M个特征提取块中位于所述第i个特征提取块之前的各特征提取块所提取的第三特征信息、与所述第i个特征提取块所提取的第三特征信息进行级联后的特征信息;In some embodiments, if i is not equal to 1, the i-th fourth feature information is the first feature extracted by each feature extraction block before the i-th feature extraction block among the M feature extraction blocks. Three feature information, feature information concatenated with the third feature information extracted by the ith feature extraction block;
若i等于1,则所述第i个第四特征信息为所述M个特征提取块中第一个特征提取块所输出的第一个第三特征信息。If i is equal to 1, the ith fourth feature information is the first third feature information output by the first feature extraction block in the M feature extraction blocks.
在一些实施例中,所述特征提取块包括:第一特征提取单元和串联连接的S个第二特征提取单元,所述S为正整数;In some embodiments, the feature extraction block includes: a first feature extraction unit and S second feature extraction units connected in series, wherein S is a positive integer;
对于所述第i+1个特征提取块中的第一提取单元,所述第一提取单元用于针对所述点云块中的当前点,搜索所述当前点的K个邻近点,并基于所述点云块的第i个第四特征信息,将所述当前点的第四特征信息与所述邻近点的第四特征信息进行相减,得到K个残差特征信息,并将所述K个残差特征信息与所述当前点的第四特征信息进行级联,得到所述当前点的第i个级联特征信息,根据所述当前点的第i个级联特征信息,得到所述点云块的第i个级联特征信息,并将所述点云块的第i个级联特征信息输入所述S个第二特征提取单元中的第一个第二特征提取单元;For the first extraction unit in the i+1th feature extraction block, the first extraction unit is used to search for K neighboring points of the current point for the current point in the point cloud block, and based on For the ith fourth feature information of the point cloud block, the fourth feature information of the current point is subtracted from the fourth feature information of the adjacent point to obtain K residual feature information, and the The K residual feature information is concatenated with the fourth feature information of the current point to obtain the i-th concatenated feature information of the current point, and according to the i-th concatenated feature information of the current point, the The i-th concatenated feature information of the point cloud block, and input the i-th concatenated feature information of the point cloud block into the first second feature extraction unit in the S second feature extraction units;
所述第一个第二特征提取单元用于根据所述点云块的第i个级联特征信息,输出第一个第五特征信息至第二个第二特征提取单元,其中所述点云块的第i+1个第三特征信息为所述S个第二特征提取单元中最后一个第二特征提取单元输出的第五特征信息。The first second feature extraction unit is used to output the first fifth feature information to the second second feature extraction unit according to the i-th cascaded feature information of the point cloud block, wherein the point cloud The i+1th third feature information of the block is the fifth feature information output by the last second feature extraction unit among the S second feature extraction units.
在一些实施例中,所述第二特征提取单元包括P个残差块,所述P为正整数;In some embodiments, the second feature extraction unit includes P residual blocks, where P is a positive integer;
对于第s个第二特征提取单元中的第j+1个残差块,所述第j+1个残差块用于根据所述第s个第二特征提取单元中的第j个残差块所输出的第j个第一残差信息和输入所述第s个第二特征提取单元的第五特征信息,输出第j+1个第一残差信息,其中,所述j为小于P的正整数,所述s为小于或等于S的正整数;For the j+1th residual block in the sth second feature extraction unit, the j+1th residual block is used according to the jth residual in the sth second feature extraction unit The j-th first residual information output by the block and the fifth feature information input to the s-th second feature extraction unit, and the j+1-th first residual information is output, wherein the j is less than P A positive integer, the s is a positive integer less than or equal to S;
所述第s个第二特征提取单元输出的第五特征信息是根据所述第s个第二特征提取单元中至少一个残差块输出的第一残差信息信息,以及输入所述第s个第二特征提取单元的第五特征信息确定的。The fifth feature information output by the sth second feature extraction unit is based on the first residual information output by at least one residual block in the sth second feature extraction unit, and input to the sth second feature extraction unit determined by the fifth feature information of the second feature extraction unit.
在一些实施例中,上采样单元43还用于将所述第s个第二特征提取单元中的第j个残差块所输出的第j个第一残差信息和输入所述第s个第二特征提取单元的第五特征信息进行相加后,输入所述第s个第二特征提取单元中的第j+1个残差块。In some embodiments, the up-sampling unit 43 is further configured to input the j-th first residual information and the j-th residual information output by the j-th residual block in the s-th second feature extraction unit to the s-th After the fifth feature information of the second feature extraction unit is added, it is input to the j+1th residual block in the sth second feature extraction unit.
在一些实施例中,所述第s个第二特征提取单元输出的第五特征信息是根据所述第s个第二特征提取单元中最后一个残差块输出的第一残差信息信息、与P-1个残差块中至少一个残差块输出的第一 残差信息信息进行级联后的特征信息、与输入所述第s个第二特征提取单元的第五特征信息确定的,其中,所述P-1个残差块为所述第s个第二特征提取单元的P个残差块中除最后一个残差块之外的残差块。In some embodiments, the fifth feature information output by the s th second feature extraction unit is based on the first residual information output by the last residual block in the s th second feature extraction unit, and The feature information after the concatenation of the first residual information output by at least one residual block in the P-1 residual blocks is determined from the fifth feature information input to the s-th second feature extraction unit, wherein , the P-1 residual blocks are residual blocks except the last residual block among the P residual blocks of the s-th second feature extraction unit.
在一些实施例中,所述第s个第二特征提取单元输出的第五特征信息是根据所述第s个第二特征提取单元中最后一个残差块输出的第一残差信息信息、与P-1个残差块中至少一个残差块输出的第一残差信息信息进行级联后特征信息、与输入所述第s个第二特征提取单元的第五特征信息进行相加后确定的。In some embodiments, the fifth feature information output by the s th second feature extraction unit is based on the first residual information output by the last residual block in the s th second feature extraction unit, and The first residual information output by at least one residual block in the P-1 residual blocks is concatenated and then determined after being added to the fifth feature information input to the s-th second feature extraction unit of.
在一些实施例中,所述第二特征提取单元还包括门控单元,In some embodiments, the second feature extraction unit further includes a gating unit,
对于所述第s个第二特征提取单元中的门控单元,所述门控单元用于对所述第s个第二特征提取单元中的最后一个残差块输出的第一残差信息、与所述P-1个残差块中至少一个残差块输出的第一残差信息级联后的特征信息进行去冗余,输出去冗余后的特征信息;所述第s个第二特征提取单元输出的第五特征信息是根据所述去冗余后的特征信息和所述输入所述第s个第二特征提取单元的第五特征信息进行相加后确定的。For the gating unit in the s th second feature extraction unit, the gating unit is used for the first residual information output by the last residual block in the s th second feature extraction unit, De-redundancy is performed on the feature information concatenated with the first residual information output by at least one residual block in the P-1 residual blocks, and the de-redundant feature information is output; the s second The fifth feature information output by the feature extraction unit is determined after adding the de-redundant feature information and the fifth feature information input to the s-th second feature extraction unit.
在一些实施例中,所述特征上采样模块包括:特征上采样子模块和特征提取子模块;In some embodiments, the feature upsampling module includes: a feature upsampling submodule and a feature extraction submodule;
所述特征上采样子模块用于按照预设的上采样率r,将所述点云块的第一特征信息复制r份,并对复制后的第一特征信息在特征维度上增加一个n维向量,得到所述点云块的上采样特征信息,并将所述点云块的上采样特征信息输入特征提取子模块,其中不同第一特征信息对应的n维向量的值不相同;The feature upsampling submodule is used to copy r copies of the first feature information of the point cloud block according to the preset upsampling rate r, and add an n dimension to the feature dimension of the copied first feature information vector, obtain the upsampling feature information of the point cloud block, and input the upsampling feature information of the point cloud block into the feature extraction submodule, wherein the values of n-dimensional vectors corresponding to different first feature information are different;
所述特征提取子模块用于根据所述点云块的上采样特征信息,输出所述点云块的第二特征信息。The feature extraction submodule is configured to output second feature information of the point cloud block according to the upsampled feature information of the point cloud block.
在一些实施例中,所述特征提取子模块包括Q个第三特征提取单元,所述Q为正整数;In some embodiments, the feature extraction submodule includes Q third feature extraction units, where Q is a positive integer;
针对所述Q个第三特征提取单元中的第k+1个第三特征提取单元,所述第k+1个第三特征提取单元用于根据第k个第三特征提取单元所提取的所述点云块的第k个增强上采样特征信息,输出所述点云块的第k+1个增强上采样特征信息,所述k为小于Q的正整数;For the k+1th third feature extraction unit among the Q third feature extraction units, the k+1th third feature extraction unit is used to extract all The kth enhanced upsampling feature information of the point cloud block, output the k+1th enhanced upsampling feature information of the point cloud block, and the k is a positive integer less than Q;
所述点云块的第二特征信息为所述Q个第三特征提取单元中最后一个第三特征提取单元所提取的所述点云块的第Q个增强上采样特征信息。The second feature information of the point cloud block is the Qth enhanced upsampling feature information of the point cloud block extracted by the last third feature extraction unit among the Q third feature extraction units.
在一些实施例中,所述第三特征提取单元包括L个残差块,所述L为正整数;In some embodiments, the third feature extraction unit includes L residual blocks, and the L is a positive integer;
对于所述第k+1个第三特征提取单元中的第l+1个残差块,所述第l+1个残差块用于根据所述第k+1个第三特征提取单元中的第l个残差块输出的第l个第二残差信息和输入所述第k+1个第三特征提取单元的第k个增强上采样特征信息,输出第l+1个第二残差信息,所述l为小于L的正整数;For the l+1th residual block in the k+1th third feature extraction unit, the l+1th residual block is used according to the k+1th third feature extraction unit The lth second residual information output by the lth residual block and the kth enhanced upsampling feature information input to the k+1th third feature extraction unit, output the l+1th second residual difference information, the l is a positive integer less than L;
所述点云块的第k+1个增强上采样特征信息是根据所述第k+1个第三特征提取单元中至少一个残差块输出的第二残差信息,以及所述第k个增强上采样特征信息确定的。The k+1th enhanced upsampling feature information of the point cloud block is the second residual information output from at least one residual block in the k+1th third feature extraction unit, and the kth Enhanced upsampling feature information determined.
在一些实施例中,上采样单元43还用于:对所述第l个残差块输出的第l个第二残差信息和所述第k个增强上采样特征信息进行相加后,输入所述第l+1个残差块。In some embodiments, the upsampling unit 43 is further configured to: after adding the lth second residual information output by the lth residual block and the kth enhanced upsampling feature information, input The l+1th residual block.
在一些实施例中,所述点云块的第k+1个增强上采样特征信息根据所述L个残差块中最后一个残差块输出的第二残差信息、与L-1个残差块中至少一个残差块输出的第二残差信息进行级联后的特征信息和所述第k个增强上采样特征信息确定的,其中,所述L-1个残差块为所述第k+1个第三特征提取单元的L个残差块中除最后一个残差块之外的残差块。In some embodiments, the k+1th enhanced upsampling feature information of the point cloud block is based on the second residual information output by the last residual block in the L residual blocks, and the L-1 residual The second residual information output by at least one residual block in the difference block is determined by concatenating the feature information and the kth enhanced upsampling feature information, wherein the L-1 residual blocks are the A residual block except the last residual block among the L residual blocks of the k+1 third feature extraction unit.
在一些实施例中,所述点云块的第k+1个增强上采样特征信息根据所述L个残差块中最后一个残差块输出的第二残差信息、与L-1个残差块中至少一个残差块输出的第二残差信息进行级联后的特征信息和所述第k个增强上采样特征信息进行相加后确定的。In some embodiments, the k+1th enhanced upsampling feature information of the point cloud block is based on the second residual information output by the last residual block in the L residual blocks, and the L-1 residual The second residual information output by at least one residual block in the difference block is determined by adding the concatenated feature information to the kth enhanced upsampling feature information.
在一些实施例中,所述第三特征提取单元还包括门控单元;In some embodiments, the third feature extraction unit further includes a gating unit;
对于所述第k+1个第三特征提取单元中的门控单元,所述门控单元用于对所述第k+1个第三特征提取单元中的最后一个残差块输出的第二残差信息、与所述L-1个残差块中至少一个残差块输出的第二特征信息进行级联后的特征信息进行去冗余,输出去冗余后的特征信息;For the gating unit in the k+1th third feature extraction unit, the gating unit is used for the second output of the last residual block in the k+1th third feature extraction unit performing de-redundancy on the residual information and the feature information concatenated with the second feature information output by at least one of the L-1 residual blocks, and outputting the de-redundant feature information;
所述点云块的第k+1个增强上采样特征信息是根据去冗余后的特征信息与所述第k个增强上采样特征信息进行相加后确定的。The k+1th enhanced upsampling feature information of the point cloud block is determined after adding the deredundant feature information to the kth enhanced upsampling feature information.
在一些实施例中,所述特征上采样模块还包括第一自相关注意力网络;In some embodiments, the feature upsampling module further includes a first autocorrelation attention network;
所述第一自相关注意力网络用于对所述特征上采样子模块输出的所述点云块的上采样特征信息进行特征交互,输出特征交互后的所述点云块的上采样特征信息至所述特征提取子模块;The first autocorrelation attention network is used to perform feature interaction on the upsampling feature information of the point cloud block output by the feature upsampling submodule, and output the upsampling feature information of the point cloud block after feature interaction to the feature extraction submodule;
所述特征提取子模块用于根据特征交互后的所述点云块的上采样特征信息,输出所述点云块的第二特征信息。The feature extraction submodule is configured to output second feature information of the point cloud block according to the upsampled feature information of the point cloud block after feature interaction.
在一些实施例中,所述特征交互后的所述点云块的上采样特征信息的特征维度低于所述点云块的上采样特征信息的特征维度。In some embodiments, the feature dimension of the upsampled feature information of the point cloud block after the feature interaction is lower than the feature dimension of the upsampled feature information of the point cloud block.
在一些实施例中,所述几何生成模块包括多个全连接层;In some embodiments, the geometry generation module includes a plurality of fully connected layers;
所述多个全连接层用于根据所述点云块的第二特征信息,输出所述点云块的上采样几何信息。The multiple fully connected layers are used to output upsampled geometric information of the point cloud block according to the second feature information of the point cloud block.
在一些实施例中,所述几何生成模块包括:几何重建单元、滤波单元和下采样单元;In some embodiments, the geometry generation module includes: a geometry reconstruction unit, a filtering unit, and a downsampling unit;
所述几何重建单元用于对所述将所述点云块的第二特征信息进行几何重建,输出所述点云块的初始上采样几何信息至所述滤波单元;The geometric reconstruction unit is used to geometrically reconstruct the second feature information of the point cloud block, and output the initial upsampling geometric information of the point cloud block to the filtering unit;
所述滤波单元用于对所述点云块的初始上采样几何信息进行除噪,输出所述点云块滤除噪点的初始上采样几何信息至所述下采样单元;The filtering unit is used to denoise the initial upsampling geometric information of the point cloud block, and output the initial upsampling geometric information of the point cloud block to filter noise to the downsampling unit;
所述下采样单元用于对所述点云块滤除噪点的初始上采样几何信息下采样至目标上采样率,输出所述点云块的上采样几何信息。The down-sampling unit is configured to down-sample the initial up-sampled geometric information of the point cloud block after filtering noise to a target up-sampling rate, and output the up-sampled geometric information of the point cloud block.
在一些实施例中,所述目标上采样率小于或等于所述特征上采样模块的上采样率。In some embodiments, the target upsampling rate is less than or equal to the upsampling rate of the feature upsampling module.
应理解,装置实施例与方法实施例可以相互对应,类似的描述可以参照方法实施例。为避免重复,此处不再赘述。具体地,图18所示的点云上采样装置40可以对应于执行本申请实施例的点云上采样方法中的相应主体,并且点点云上采样装置40中的各个单元的前述和其它操作和/或功能分别为了实现点云上采样方法中的相应流程,为了简洁,在此不再赘述。It should be understood that the device embodiment and the method embodiment may correspond to each other, and similar descriptions may refer to the method embodiment. To avoid repetition, details are not repeated here. Specifically, the point cloud upsampling device 40 shown in FIG. 18 may correspond to the corresponding subject in the point cloud upsampling method of the embodiment of the present application, and the aforementioned and other operations of each unit in the point cloud upsampling device 40 and The /or functions are respectively used to realize the corresponding process in the point cloud upsampling method, for the sake of brevity, no more details are given here.
图19是本申请实施例提供的模型训练装置的示意性框图。Fig. 19 is a schematic block diagram of a model training device provided by an embodiment of the present application.
如图19所示,模型训练装置10包括:As shown in Figure 19, the model training device 10 includes:
获取单元11,用于获取训练点云的几何信息;Acquisition unit 11, for obtaining the geometric information of training point cloud;
划分单元12,用于根据所述训练点云的几何信息,将所述训练点云划分成至少一个训练点云块;A division unit 12, configured to divide the training point cloud into at least one training point cloud block according to the geometric information of the training point cloud;
训练单元13,用于将所述训练点云块的几何信息输入生成器的特征提取模块进行特征提取,得到所述训练点云块的第一特征信息;将所述训练点云块的第一特征信息输入所述生成器的特征上采样模块进行上采样,得到所述训练点云块的第二特征信息;将所述训练点云块的第二特征信息输入所述生成器的几何生成模块进行几何重建,得到所述训练点云块的预测上采样几何信息;根据所述训练点云块的预测上采样几何信息,对所述生成器中的特征提取模块、特征上采样模块和几何生成模块进行训练,得到训练后的生成器。The training unit 13 is used to input the geometric information of the training point cloud block into the feature extraction module of the generator for feature extraction, and obtain the first feature information of the training point cloud block; the first feature information of the training point cloud block is Feature information is input into the feature upsampling module of the generator for upsampling to obtain the second feature information of the training point cloud block; the second feature information of the training point cloud block is input into the geometry generation module of the generator Perform geometric reconstruction to obtain the predicted upsampling geometric information of the training point cloud block; according to the predicted upsampling geometric information of the training point cloud block, the feature extraction module, feature upsampling module and geometric generation in the generator The module is trained to obtain the trained generator.
在一些实施例中,训练单元12,具体用于将所述训练点云块的预测上采样几何信息输入判别器,得到所述判别器的第一判别结果,所述判别器用于判断输入所述判别器的数据是否为所述训练点云块的上采样真值;根据所述判别器的第一判别结果,对所述生成器中的特征提取模块、特征上采样模块和几何生成模块进行训练,得到训练后的生成器。In some embodiments, the training unit 12 is specifically configured to input the predicted upsampled geometric information of the training point cloud block into a discriminator to obtain the first discriminant result of the discriminator, and the discriminator is used to judge the input Whether the data of the discriminator is the upsampled true value of the training point cloud block; according to the first discrimination result of the discriminator, the feature extraction module, feature upsampling module and geometry generation module in the generator are trained , to get the trained generator.
在一些实施例中,所述特征提取模块包括M个密集连接的特征提取块,训练单元12,具体用于将所述训练点云块的几何信息输入所述特征提取模块中,获取所述M个特征提取块中第i个特征提取块所提取的所述训练点云块的第i个第三特征信息,所述i为小于M的正整数;根据所述训练点云块的第i个第三特征信息,得到所述训练点云块的第i个第四特征信息;将所述训练点云块的第i个第四特征信息输入第i+1个特征提取块中,得到所述训练点云块的第i+1个第三特征信息;将所述训练点云块的第M个特征提取块所提取的第M个第三特征信息,作为所述训练点云块的第一特征信息。In some embodiments, the feature extraction module includes M densely connected feature extraction blocks, and the training unit 12 is specifically configured to input the geometric information of the training point cloud block into the feature extraction module, and obtain the M The i-th third feature information of the training point cloud block extracted by the i-th feature extraction block in the feature extraction block, the i is a positive integer less than M; according to the i-th feature information of the training point cloud block The third feature information is to obtain the i-th fourth feature information of the training point cloud block; the i-th fourth feature information of the training point cloud block is input in the i+1 feature extraction block to obtain the i+1 feature extraction block. The i+1th third feature information of the training point cloud block; the Mth third feature information extracted by the Mth feature extraction block of the training point cloud block is used as the first feature information of the training point cloud block characteristic information.
在一些实施例中,训练单元12,具体用于若i不等于1,则获取所述M个特征提取块中位于所述第i个特征提取块之前的各特征提取块所提取的第三特征信息;并将位于所述第i个特征提取块之前的各特征提取块所提取的第三特征信息、与所述第i个特征提取块所提取的第三特征信息进行级联,作为所述训练点云块的第i个第四特征信息;In some embodiments, the training unit 12 is specifically configured to, if i is not equal to 1, obtain the third feature extracted by each feature extraction block located before the ith feature extraction block among the M feature extraction blocks information; and the third feature information extracted by each feature extraction block located before the ith feature extraction block is concatenated with the third feature information extracted by the ith feature extraction block, as the The i-th fourth feature information of the training point cloud block;
若i等于1,则所述M个特征提取单元中第一特征提取块所提取的第一个第三特征信息,作为所述训练点云块的第i个第四特征信息。If i is equal to 1, the first third feature information extracted by the first feature extraction block in the M feature extraction units is used as the ith fourth feature information of the training point cloud block.
在一些实施例中,所述特征提取块包括:第一特征提取单元和串联连接的至少一个第二特征提取单元,训练单元12,具体用于将所述训练点云块的第i个第四特征信息输入所述第i+1个特征提取块中的第一特征提取单元,以使所述第一特征提取单元针对所述训练点云块中的当前点,搜索所述当前点的K个邻近点,并基于所述第i个第四特征信息,将所述当前点的第四特征信息与所述邻近点的第四特征信息进行相减,得到K个残差特征信息;将所述K个残差特征信息与所述当前点的第四特征信息进行级联,得到所述当前点的第i个级联特征信息,并根据所述当前点的第i个级联特征信息,得到所述训练点云块的第i个级联特征信息;将所述训练点云块的第i个级联特征信息输入所述第i+1个特征提取块中的第一个第二特征提取单元,得到第一个第一个第五特征信息,并将所述第一个第一个第五特征信息输入所述第i+1个特征提取块中的第二个第二特征提取单元中,得到第二个第五特征信息;将所述第i+1个特征提取块中最后一个第二特征提取单元提取的第五特征信息,作为所述训练点云块的第i+1个第三特征信息。In some embodiments, the feature extraction block includes: a first feature extraction unit and at least one second feature extraction unit connected in series, and a training unit 12, which is specifically used to convert the i-th fourth feature of the training point cloud block The feature information is input to the first feature extraction unit in the i+1th feature extraction block, so that the first feature extraction unit searches for K of the current point for the current point in the training point cloud block. neighboring points, and based on the ith fourth characteristic information, subtracting the fourth characteristic information of the current point from the fourth characteristic information of the neighboring points to obtain K residual characteristic information; The K residual feature information is concatenated with the fourth feature information of the current point to obtain the i-th concatenated feature information of the current point, and according to the i-th concatenated feature information of the current point, it is obtained The i-th concatenated feature information of the training point cloud block; the i-th concatenated feature information of the training point cloud block is input into the first second feature extraction in the i+1 feature extraction block unit to obtain the first, first, and fifth feature information, and input the first, first, and fifth feature information into the second, second feature extraction unit in the i+1th feature extraction block , to obtain the second fifth feature information; the fifth feature information extracted by the last second feature extraction unit in the i+1 feature extraction block is used as the i+1th feature information of the training point cloud block Three characteristic information.
在一些实施例中,所述第二特征提取单元包括P个残差块,所述P为正整数,训练单元12,具体用于将所述第i个级联特征信息输入所述第i+1个特征提取块中的第一个第二特征提取单元,获得所述第一个第二特征提取单元中第j个残差块输出的第一残差信息,所述j为小于或等于P的正整数;将所述第j个残差块输出的第一残差信息和所述第i个级联特征信息输入所述第一个第二特征提取单元中的第j+1个残差块中,得到所述第j+1个残差块输出的第一残差信息;根据所述第一个第二特征提取单元中的P个残差块中至少一个残差块输出的第一残差信息,以及所述第i个级联特征信息,确定所述第一个第二特征提取单元输出的第五特征信息。In some embodiments, the second feature extraction unit includes P residual blocks, where P is a positive integer, and the training unit 12 is specifically configured to input the i-th concatenated feature information into the i+th The first second feature extraction unit in one feature extraction block obtains the first residual information output by the jth residual block in the first second feature extraction unit, where j is less than or equal to P is a positive integer; input the first residual information output by the jth residual block and the ith concatenated feature information into the j+1th residual in the first second feature extraction unit In the block, the first residual information output by the j+1th residual block is obtained; according to the first residual information output by at least one of the P residual blocks in the first second feature extraction unit The residual information, as well as the i-th concatenated feature information, determine fifth feature information output by the first second feature extraction unit.
在一些实施例中,训练单元12,具体用于对所述第j个残差块输出的第一残差信息和所述第i个级联特征信息进行相加,并将相加后的特征信息输入所述第j+1个残差块中,得到所述第j+1个残差块输出的第一残差信息。In some embodiments, the training unit 12 is specifically configured to add the first residual information output by the jth residual block and the ith concatenated feature information, and add the added feature The information is input into the j+1th residual block, and the first residual information output by the j+1th residual block is obtained.
在一些实施例中,训练单元12,具体用于将所述P个残差块中最后一个残差块输出的第一残差信息、与P-1个残差块中至少一个残差块输出的第一残差信息进行级联,其中所述P-1个残差块为所述P个残差块中除所述最后一个残差块之外的残差块;根据级联后的特征信息和所述第i个级联特征信息,确定所述第一个第二特征提取单元输出的第五特征信息。In some embodiments, the training unit 12 is specifically configured to output the first residual information output by the last residual block in the P residual blocks and at least one residual block in the P-1 residual blocks The first residual information is concatenated, wherein the P-1 residual blocks are residual blocks except the last residual block among the P residual blocks; according to the concatenated features information and the i-th concatenated feature information to determine the fifth feature information output by the first second feature extraction unit.
在一些实施例中,训练单元12,具体用于将所述级联后的特征信息和所述第i个级联特征信息进行相加,作为所述第一个第二特征提取单元输出的第五特征信息。In some embodiments, the training unit 12 is specifically configured to add the concatenated feature information and the i-th concatenated feature information as the first output of the first second feature extraction unit. Five characteristic information.
在一些实施例中,所述第二特征提取单元还包括门控单元,训练单元12,具体用于将所述级联后的特征信息输入所述门控单元进行去冗余,得到去冗余后的特征信息;将去冗余后的特征信息与所述第i个级联特征信息进行相加,作为所述第一个第二特征提取单元输出的第五特征信息。In some embodiments, the second feature extraction unit further includes a gating unit, a training unit 12, specifically configured to input the cascaded feature information into the gating unit for de-redundancy, to obtain de-redundancy The feature information after de-redundancy is added to the i-th concatenated feature information as the fifth feature information output by the first second feature extraction unit.
在一些实施例中,所述特征上采样模块包括:特征上采样子模块和特征提取子模块,训练单元12,具体用于将所述训练点云块的第一特征信息输入所述特征上采样子模块,以使所述特征上采样子模块按照预设的上采样率r,将所述训练点云块的第一特征信息复制r份,并对复制后的第一特征信息在特征维度上增加一个n维向量,得到所述训练点云块的上采样特征信息,其中不同第一特征信息对应的n维向量的值不相同;将所述训练点云块的上采样特征信息输入所述特征提取子模块,得到所述特征提取子模块提取的所述训练点云块的第二特征信息。In some embodiments, the feature upsampling module includes: a feature upsampling submodule and a feature extraction submodule, a training unit 12, specifically for inputting the first feature information of the training point cloud block into the feature upsampling sub-module, so that the feature up-sampling sub-module copies r shares of the first feature information of the training point cloud block according to the preset up-sampling rate r, and performs a feature dimension on the copied first feature information Add an n-dimensional vector to obtain the upsampling feature information of the training point cloud block, wherein the values of the n-dimensional vectors corresponding to different first feature information are different; the upsampling feature information of the training point cloud block is input into the The feature extraction sub-module obtains the second feature information of the training point cloud block extracted by the feature extraction sub-module.
在一些实施例中,所述特征提取子模块包括串联连接的Q个第三特征提取单元,所述Q为正整数,训练单元12,具体用于将所述训练点云块的上采样特征信息输入所述特征提取子模块,获得第k个第三特征提取单元所提取的所述训练点云块的第k个增强上采样特征信息;将所述训练点云块的第k个增强上采样特征信息输入第k+1个第三特征提取单元,得到所述第k+1个第三特征提取单元提取的所述训练点云块的第k+1个增强上采样特征信息;将所述Q个第三特征提取单元中最后一个第三特征提取单元所提取的所述训练点云块的第Q个增强上采样特征信息,作为所述训练点云块的第二特征信息。In some embodiments, the feature extraction submodule includes Q third feature extraction units connected in series, the Q is a positive integer, and the training unit 12 is specifically used to convert the upsampled feature information of the training point cloud block to Input the feature extraction submodule to obtain the kth enhancement upsampling feature information of the training point cloud block extracted by the kth third feature extraction unit; the kth enhancement upsampling feature information of the training point cloud block The feature information is input into the k+1th third feature extraction unit to obtain the k+1th enhanced upsampling feature information of the training point cloud block extracted by the k+1th third feature extraction unit; The Qth enhanced upsampled feature information of the training point cloud block extracted by the last third feature extraction unit among the Q third feature extraction units is used as the second feature information of the training point cloud block.
在一些实施例中,所述第三特征提取单元包括L个残差块,所述L为正整数,训练单元12,具体用于将所述训练点云块的第k个增强上采样特征信息输入所述第k+1个第三特征提取单元,获得所述第k+1个第三特征提取单元中第l个残差块输出的第二残差信息,所述l为小于或等于L的正整数;将所述第l个残差块输出的第二残差信息和所述第k个增强上采样特征信息输入第l+1个残差块中, 得到所述第l+1个残差块输出的第二残差信息;根据所述L个残差块中至少一个残差块输出的第二残差信息,以及所述第k个增强上采样特征信息,得到所述训练点云块的第k+1个增强上采样特征信息。In some embodiments, the third feature extraction unit includes L residual blocks, where L is a positive integer, and the training unit 12 is specifically used to enhance and upsample the feature information of the k th training point cloud block Input the k+1th third feature extraction unit to obtain the second residual information output by the lth residual block in the k+1th third feature extraction unit, and the l is less than or equal to L is a positive integer; input the second residual information output by the lth residual block and the kth enhanced upsampling feature information into the l+1th residual block to obtain the l+1th The second residual information output by the residual block; according to the second residual information output by at least one residual block in the L residual blocks, and the kth enhanced upsampling feature information, the training point is obtained The k+1th enhanced upsampled feature information of the cloud block.
在一些实施例中,训练单元12,具体用于对所述第l个残差块输出的第二残差信息和所述第k个增强上采样特征信息进行相加,并将相加后的特征信息输入所述第l+1个残差块中,确定所述第l+1个残差块输出的第二残差信息。In some embodiments, the training unit 12 is specifically configured to add the second residual information output by the lth residual block and the kth enhanced upsampling feature information, and add the added The characteristic information is input into the l+1th residual block, and the second residual information output by the l+1th residual block is determined.
在一些实施例中,训练单元12,具体用于将所述L个残差块中最后一个残差块输出的第二残差信息、与L-1个残差块中至少一个残差块输出的第二残差信息进行级联,其中所述L-1个残差块为所述L个残差块中除所述最后一个残差块之外的残差块;根据级联后的特征信息和所述第k个增强上采样特征信息,确定所述训练点云块的第k+1个增强上采样特征信息。In some embodiments, the training unit 12 is specifically configured to output the second residual information output by the last residual block in the L residual blocks, and at least one residual block output in the L-1 residual blocks The second residual information is concatenated, wherein the L-1 residual blocks are residual blocks except the last residual block among the L residual blocks; according to the concatenated features information and the kth enhanced upsampling feature information to determine the k+1th enhanced upsampling feature information of the training point cloud block.
在一些实施例中,训练单元12,具体用于将所述级联后的特征信息和所述第k个增强上采样特征信息进行相加,作为所述训练点云块的第k+1个增强上采样特征信息。In some embodiments, the training unit 12 is specifically configured to add the concatenated feature information and the kth enhanced upsampling feature information as the k+1th of the training point cloud block Enhance upsampled feature information.
在一些实施例中,所述第三特征提取单元还包括门控单元,训练单元12,具体用于将所述级联后的特征信息输入所述门控单元进行去冗余,得到去冗余后的特征信息;将去冗余后的特征信息与所述第k个增强上采样特征信息进行相加,作为所述训练点云块的第k+1个增强上采样特征信息。In some embodiments, the third feature extraction unit further includes a gating unit, a training unit 12, specifically configured to input the cascaded feature information into the gating unit for de-redundancy, to obtain de-redundancy The feature information after de-redundancy is added to the k-th enhanced up-sampling feature information, and used as the k+1th enhanced up-sampling feature information of the training point cloud block.
在一些实施例中,所述特征上采样模块还包括第一自相关注意力网络,训练单元12,具体用于将所述训练点云块的上采样特征信息输入所述第一自相关注意力网络进行特征交互,得到特征交互后的所述训练点云块的上采样特征信息;将特征交互后的所述训练点云块的上采样特征信息输入所述特征提取子模块进行特征提取,得到所述训练点云块的第二特征信息。In some embodiments, the feature upsampling module further includes a first autocorrelation attention network, a training unit 12, specifically for inputting the upsampling feature information of the training point cloud block into the first autocorrelation attention The network performs feature interaction to obtain the upsampling feature information of the training point cloud block after the feature interaction; the upsampling feature information of the training point cloud block after the feature interaction is input into the feature extraction submodule to perform feature extraction, and obtain The second feature information of the training point cloud block.
可选的,所述特征交互后的所述训练点云块的上采样特征信息的特征维度低于所述训练点云块的上采样特征信息的特征维度。Optionally, the feature dimension of the upsampled feature information of the training point cloud block after the feature interaction is lower than the feature dimension of the upsampled feature information of the training point cloud block.
在一些实施例中,所述几何生成模块包括多个全连接层,训练单元12,具体用于将所述训练点云块的第二特征信息输入所述多个全连接层,得到所述训练点云块的预测上采样几何信息。In some embodiments, the geometry generation module includes multiple fully connected layers, and the training unit 12 is specifically configured to input the second feature information of the training point cloud block into the multiple fully connected layers to obtain the training Prediction of point cloud blocks upsamples geometric information.
在一些实施例中,所述几何生成模块包括:几何重建单元、滤波单元和下采样单元,训练单元12,具体用于将所述训练点云块的第二特征信息输入所述几何重建单元进行几何重建,得到所述训练点云块的初始上采样几何信息;将所述训练点云块的初始上采样几何信息输入滤波单元进行除噪,得到所述训练点云块滤除噪点的初始上采样几何信息;将所述训练点云块滤除噪点的初始上采样几何信息输入所述下采样单元中进行下采样,得到所述训练点云块的预测上采样几何信息。In some embodiments, the geometry generation module includes: a geometry reconstruction unit, a filter unit, and a downsampling unit, and a training unit 12, specifically configured to input the second feature information of the training point cloud block into the geometry reconstruction unit for further processing. Geometric reconstruction to obtain the initial upsampling geometric information of the training point cloud block; input the initial upsampling geometric information of the training point cloud block into a filter unit for noise removal, and obtain the initial upsampling of the training point cloud block to filter out noise Sampling geometric information; inputting the initial upsampling geometric information of the training point cloud block to filter out noise into the downsampling unit for downsampling, to obtain the predicted upsampling geometric information of the training point cloud block.
可选的,所述训练点云块的上采样几何信息对应的上采样率小于或等于所述特征上采样模块的上采样率。Optionally, the upsampling rate corresponding to the upsampling geometric information of the training point cloud block is less than or equal to the upsampling rate of the feature upsampling module.
可选的,所述判别器为预先训练好的判别器。Optionally, the discriminator is a pre-trained discriminator.
在一些实施例中,训练单元12,还用于使用所述训练点云块的几何信息对所述判别器进行训练。In some embodiments, the training unit 12 is further configured to use the geometric information of the training point cloud block to train the discriminator.
在一些实施例中,训练单元12,具体用于将所述生成器生成的所述训练点云块的预测上采样几何信息输入所述判别器中,得到所述判别器的第二判别结果;将所述训练点云块的几何信息的上采样真值输入所述判别器,得到所述判别器的第三判别结果;根据所述第二判别结果和第三判别结果,确定所述判别器的损失;根据所述判别器的损失,对所述判别器进行训练。In some embodiments, the training unit 12 is specifically configured to input the predicted upsampling geometric information of the training point cloud block generated by the generator into the discriminator, and obtain a second discrimination result of the discriminator; Inputting the upsampling true value of the geometric information of the training point cloud block into the discriminator to obtain a third discriminant result of the discriminator; according to the second discriminant result and the third discriminant result, determine the discriminator The loss; according to the loss of the discriminator, the discriminator is trained.
在一些实施例中,训练单元12,具体用于根据所述第二判别结果和第三判别结果,采用最小二乘损失函数,确定所述判别器的损失。In some embodiments, the training unit 12 is specifically configured to determine the loss of the discriminator by using a least square loss function according to the second discrimination result and the third discrimination result.
在一些实施例中,所述判别器包括:全局判别模块、边界判别模块和全连接模块,训练单元12,具体用于获取目标点云块的边界点的几何信息;将所述目标点云块的边界点的几何信息输入所述边界判别模块,得到所述目标点云块的边界特征信息;将所述目标点云块的几何信息输入所述全局判别模块,得到所述目标点云块的全局特征信息;将所述目标点云块的全局特征信息和边界特征信息输入所述全连接模块,得到所述判别器的目标判别结果;其中,若所述目标点云块为所述生成器上采样后的训练点云块,且所述判断器未经过所述训练点云块训练,则所述目标判别结果为所述第二判别结果;若所述目标点云块为所述训练点云块的上采样真值,则所述目标判别结果为所述第三判别结果;若所述目标点云块为所述生成器上采样后的训练点云块,且所述判断器经过所述训练点云块训练,则所述目标判别结果为所述第一判别结果。In some embodiments, the discriminator includes: a global discriminant module, a boundary discriminant module and a fully connected module, and a training unit 12, which is specifically used to obtain the geometric information of the boundary points of the target point cloud block; The geometric information of the boundary point is input into the boundary discrimination module to obtain the boundary feature information of the target point cloud block; the geometric information of the target point cloud block is input to the global discrimination module to obtain the target point cloud block Global feature information; input the global feature information and boundary feature information of the target point cloud block into the full connection module to obtain the target discrimination result of the discriminator; wherein, if the target point cloud block is the generator Upsampled training point cloud block, and the judge has not been trained by the training point cloud block, then the target discrimination result is the second discrimination result; if the target point cloud block is the training point If the upsampling true value of the cloud block, the target discrimination result is the third discrimination result; if the target point cloud block is the training point cloud block after the generator upsamples, and the judger passes the If the training point cloud is trained, the target discrimination result is the first discrimination result.
在一些实施例中,训练单元12,具体用于使用高通图滤波器提取所述目标点云块的边界点的几何信息。In some embodiments, the training unit 12 is specifically configured to use a high-pass image filter to extract the geometric information of the boundary points of the target point cloud block.
在一些实施例中,训练单元12,具体用于将所述目标点云块的全局特征信息和边界特征信息进行级联;将级联后的全局特征信息和边界特征信息输入所述全连接模块,得到所述判别器的目标判别结果。In some embodiments, the training unit 12 is specifically configured to concatenate the global feature information and boundary feature information of the target point cloud block; input the concatenated global feature information and boundary feature information into the full connection module , to obtain the target discrimination result of the discriminator.
在一些实施例中,所述全局判别模块沿着网络深度方向依次包括:第一数量个多层感知机、第一最大池化层、第二自相关注意力网络、第二数量个多层感知机和第二最大池化层;训练单元12,具体用于将所述目标点云块的几何信息输入所述第一数量个多层感知机进行特征提取,得到所述目标点云块的第一全局特征信息;将所述第一全局特征信息输入所述第一最大池化层进行降维处理,得到所述目标点云块的第二全局特征信息;将所述第一全局特征信息和所述第二全局特征信息输入所述第二自相关注意力网络进行特征交互,得到所述目标点云块的第三全局特征信息;将所述第三全局特征信息输入所述第二数量个多层感知机进而特征提取,得到所述目标点云块的第四全局特征信息;将所述第四全局特征信息输入所述第二最大池化层进行降维处理,得到所述目标点云块的全局特征信息。In some embodiments, the global discrimination module includes sequentially along the network depth direction: a first number of multi-layer perceptrons, a first maximum pooling layer, a second autocorrelation attention network, and a second number of multi-layer perceptrons machine and the second maximum pooling layer; the training unit 12 is specifically used to input the geometric information of the target point cloud block into the first number of multi-layer perceptrons for feature extraction, and obtain the first number of the target point cloud block A global feature information; inputting the first global feature information into the first maximum pooling layer for dimensionality reduction processing to obtain the second global feature information of the target point cloud block; combining the first global feature information and The second global feature information is input into the second autocorrelation attention network for feature interaction to obtain the third global feature information of the target point cloud block; the third global feature information is input into the second number of The multi-layer perceptron then extracts features to obtain the fourth global feature information of the target point cloud block; input the fourth global feature information into the second maximum pooling layer for dimensionality reduction processing to obtain the target point cloud The global feature information of the block.
在一些实施例中,训练单元12,具体用于将所述第一全局特征信息和所述第二全局特征信息进行级联;将级联后的所述第一全局特征信息和所述第二全局特征信息,输入所述第二自相关注意力网络进行特征交互,得到所述目标点云块的第三全局特征信息。In some embodiments, the training unit 12 is specifically configured to concatenate the first global feature information and the second global feature information; combine the concatenated first global feature information and the second global feature information The global feature information is input into the second autocorrelation attention network for feature interaction to obtain the third global feature information of the target point cloud block.
可选的,所述第一数量等于所述第二数量。Optionally, the first quantity is equal to the second quantity.
可选的,所述第一数量与所述第二数量均等于2。Optionally, the first number and the second number are both equal to 2.
在一些实施例中,所述第一数量个多层感知机包括第一层多层感知机和第二层多层感知机,所述第二数量个多层感知机包括第三层多层感知机和第四层多层感知机,所述第一层多层感知机、所述第二层多层感知机、所述第三层多层感知机和所述第四层多层感知机的特征维度依次逐渐增加。In some embodiments, the first number of multilayer perceptrons includes a first layer of multilayer perceptrons and a second layer of multilayer perceptrons, and the second number of multilayer perceptrons includes a third layer of multilayer perceptrons machine and the fourth layer of multi-layer perceptron, the first layer of multi-layer perceptron, the second layer of multi-layer perceptron, the third layer of multi-layer perceptron and the fourth layer of multi-layer perceptron The feature dimension gradually increases sequentially.
可选的,所述第一层多层感知机的特征维度为32,所述第二层多层感知机的特征维度为64,所述第三层多层感知机的特征维度为128,所述第四层多层感知机的特征维度为256。Optionally, the feature dimension of the first layer of multi-layer perceptron is 32, the feature dimension of the second layer of multi-layer perceptron is 64, and the feature dimension of the third layer of multi-layer perceptron is 128, so The feature dimension of the fourth layer multilayer perceptron is 256.
在一些实施例中,所述边界判别模块沿着网络深度方向依次包括:第三数量个多层感知机、第三最大池化层、第三自相关注意力网络、第四数量个多层感知机和第四最大池化层;训练单元12,具体用于将所述目标点云块的边界点的几何信息输入所述第三数量个多层感知机中进行特征提取,得到所述目标点云块的第一边界特征信息;将所述第一边界特征信息输入所述第三最大池化层进行降维处理,得到所述目标点云块的第二边界特征信息;将所述第一边界特征信息和所述第二边界特征信息输入所述第三自相关注意力网络进行特征交互,得到所述目标点云块的第三边界特征信息;将所述第三边界特征信息输入所述第四数量个多层感知机进行特征提取,得到所述目标点云块的第四边界特征信息;将所述第四边界特征信息输入所述第四最大池化层进行降维处理,得到所述目标点云块的边界特征信息。In some embodiments, the boundary discrimination module sequentially includes along the network depth direction: a third number of multi-layer perceptrons, a third maximum pooling layer, a third autocorrelation attention network, and a fourth number of multi-layer perceptrons machine and the fourth maximum pooling layer; the training unit 12 is specifically used to input the geometric information of the boundary points of the target point cloud block into the third number of multi-layer perceptrons for feature extraction, and obtain the target point The first boundary feature information of the cloud block; the first boundary feature information is input into the third maximum pooling layer for dimension reduction processing, and the second boundary feature information of the target point cloud block is obtained; the first The boundary feature information and the second boundary feature information are input into the third autocorrelation attention network for feature interaction to obtain the third boundary feature information of the target point cloud block; the third boundary feature information is input into the A fourth number of multi-layer perceptrons perform feature extraction to obtain fourth boundary feature information of the target point cloud block; input the fourth boundary feature information into the fourth maximum pooling layer for dimensionality reduction processing to obtain the Describe the boundary feature information of the target point cloud block.
在一些实施例中,训练单元12,具体用于将所述第一边界特征信息和所述第二边界特征信息进行级联;将级联后的所述第一边界特征信息和所述第二边界特征信息,输入所述第三自相关注意力网络进行特征交互,得到所述目标点云块的第三边界特征信息。In some embodiments, the training unit 12 is specifically configured to concatenate the first boundary feature information and the second boundary feature information; combine the concatenated first boundary feature information and the second boundary feature information The boundary feature information is input into the third autocorrelation attention network for feature interaction to obtain the third boundary feature information of the target point cloud block.
可选的,所述第三数量等于所述第四数量。Optionally, the third quantity is equal to the fourth quantity.
可选的,所述第三数量与所述第四数量均等于2。Optionally, both the third quantity and the fourth quantity are equal to 2.
在一些实施例中,所述第三数量个多层感知机包括第五层多层感知机和第六层多层感知机,所述第四数量个多层感知机包括第七层多层感知机和第八层多层感知机,所述第五层多层感知机、所述第六层多层感知机、所述第七层多层感知机和所述第八层多层感知机的特征维度依次逐渐增加。In some embodiments, the third number of multilayer perceptrons includes a fifth layer of multilayer perceptrons and a sixth layer of multilayer perceptrons, and the fourth number of multilayer perceptrons includes a seventh layer of multilayer perceptrons machine and the eighth layer multi-layer perceptron, the fifth layer multi-layer perceptron, the sixth layer multi-layer perceptron, the seventh layer multi-layer perceptron and the eighth layer multi-layer perceptron The feature dimension gradually increases sequentially.
可选的,所述第八层多层感知机的特征维度大于或等于所述第七层多层感知机的特征维度,且小于或等于第四层多层感知机的特征维度。Optionally, the feature dimension of the eighth-layer multi-layer perceptron is greater than or equal to the feature dimension of the seventh-layer multi-layer perceptron, and smaller than or equal to the feature dimension of the fourth-layer multi-layer perceptron.
可选的,所述第五层多层感知机的特征维度为32,所述第六层多层感知机的特征维度为64,所述第七层多层感知机的特征维度为128,所述第八层多层感知机的特征维度为192。Optionally, the feature dimension of the fifth-layer multi-layer perceptron is 32, the feature dimension of the sixth-layer multi-layer perceptron is 64, and the feature dimension of the seventh-layer multi-layer perceptron is 128, so The feature dimension of the eighth layer multilayer perceptron is 192.
在一些实施例中,训练单元12,具体用于根据所述第一判别结果,确定所述生成器的第一损失;根据所述第一损失,确定所述生成器中的特征提取模块、特征上采样模块和几何生成模块的参数矩阵。In some embodiments, the training unit 12 is specifically configured to determine the first loss of the generator according to the first discrimination result; and determine the feature extraction module and feature of the generator according to the first loss. Parameter matrix for the upsampling block and the geometry generation block.
在一些实施例中,训练单元12,具体用于根据所述第一判别结果,采用最小二乘损失函数,确定 所述生成器的第一损失。In some embodiments, the training unit 12 is specifically configured to determine the first loss of the generator by using a least squares loss function according to the first discrimination result.
在一些实施例中,训练单元12,具体用于确定所述生成器的至少一个第二损失;根据所述生成器的第一损失和所述生成器的至少一个第二损失,确定所述生成器的目标损失;根据所述生成器的目标损失,确定所述生成器中的特征提取模块、特征上采样模块和几何生成模块的参数矩阵。In some embodiments, the training unit 12 is specifically configured to determine at least one second loss of the generator; according to the first loss of the generator and at least one second loss of the generator, determine the generation The target loss of the generator; according to the target loss of the generator, determine the parameter matrix of the feature extraction module, feature upsampling module and geometry generation module in the generator.
在一些实施例中,训练单元12,具体用于根据所述训练点云块的上采样几何信息和所述训练点云块的几何信息的上采样真值,采用地动距离方式,确定所述生成器的一个第二损失。In some embodiments, the training unit 12 is specifically configured to determine the A second loss for the generator.
在一些实施例中,训练单元12,具体用于将所述训练点云块的上采样几何信息进行下采样,得到与所述训练点云块相同点数的下采样训练点云块;根据所述下采样训练点云块的几何信息和所述训练点云块的几何信息,采样地动距离方式,确定所述生成器的一个第二损失。In some embodiments, the training unit 12 is specifically configured to downsample the upsampled geometric information of the training point cloud block to obtain a downsampled training point cloud block with the same number of points as the training point cloud block; according to the Downsampling the geometric information of the training point cloud block and the geometric information of the training point cloud block, sampling the ground motion distance method, and determining a second loss of the generator.
在一些实施例中,训练单元12,具体用于根据如下公式,确定所述生成器的一个第二损失:In some embodiments, the training unit 12 is specifically configured to determine a second loss of the generator according to the following formula:
Figure PCTCN2021096287-appb-000025
Figure PCTCN2021096287-appb-000025
其中,所述L id为所述生成器的第二损失,所述P ori为训练点云块,所述P low为下采样后的训练点云块,φ:P low→P ori表示由P low和P ori构成的双射,有且只有唯一的一种移动方式让P low与P ori移动到彼此点集的距离最小,所述
Figure PCTCN2021096287-appb-000026
为所述P low中的第k个点,所述
Figure PCTCN2021096287-appb-000027
为所述
Figure PCTCN2021096287-appb-000028
在所述P ori中对应的点。
Wherein, the L id is the second loss of the generator, the P ori is a training point cloud block, and the P low is a downsampled training point cloud block, and φ:P low →P ori means that P For the bijection composed of low and P ori , there is only one and only way to move P low and P ori to the minimum distance between the point sets of each other.
Figure PCTCN2021096287-appb-000026
is the kth point in the P low , the
Figure PCTCN2021096287-appb-000027
for the said
Figure PCTCN2021096287-appb-000028
Corresponding point in the P ori .
在一些实施例中,训练单元12,具体用于根据均匀损失函数,确定所述生成器的至少一个第二损失。In some embodiments, the training unit 12 is specifically configured to determine at least one second loss of the generator according to a uniform loss function.
在一些实施例中,训练单元12,具体用于将所述生成器的第一损失和所述至少一个第二损失的加权平均值,确定所述生成器的目标损失。In some embodiments, the training unit 12 is specifically configured to use a weighted average of the first loss of the generator and the at least one second loss to determine the target loss of the generator.
应理解,装置实施例与方法实施例可以相互对应,类似的描述可以参照方法实施例。为避免重复,此处不再赘述。具体地,图18所示的点云上采样装置10可以对应于执行本申请实施例的模型训练方法中的相应主体,并且点云上采样装置10中的各个单元的前述和其它操作和/或功能分别为了实现模型训练方法等各个方法中的相应流程,为了简洁,在此不再赘述。It should be understood that the device embodiment and the method embodiment may correspond to each other, and similar descriptions may refer to the method embodiment. To avoid repetition, details are not repeated here. Specifically, the point cloud upsampling device 10 shown in FIG. 18 may correspond to the corresponding subject in the model training method of the embodiment of the present application, and the foregoing and other operations and/or The functions are to realize the corresponding processes in each method such as the model training method, and for the sake of brevity, details are not repeated here.
上文中结合附图从功能单元的角度描述了本申请实施例的装置和系统。应理解,该功能单元可以通过硬件形式实现,也可以通过软件形式的指令实现,还可以通过硬件和软件单元组合实现。具体地,本申请实施例中的方法实施例的各步骤可以通过处理器中的硬件的集成逻辑电路和/或软件形式的指令完成,结合本申请实施例公开的方法的步骤可以直接体现为硬件译码处理器执行完成,或者用译码处理器中的硬件及软件单元组合执行完成。可选地,软件单元可以位于随机存储器,闪存、只读存储器、可编程只读存储器、电可擦写可编程存储器、寄存器等本领域的成熟的存储介质中。该存储介质位于存储器,处理器读取存储器中的信息,结合其硬件完成上述方法实施例中的步骤。The device and system of the embodiments of the present application are described above from the perspective of functional units with reference to the accompanying drawings. It should be understood that the functional unit may be implemented in the form of hardware, may also be implemented by instructions in the form of software, and may also be implemented by a combination of hardware and software units. Specifically, each step of the method embodiment in the embodiment of the present application can be completed by an integrated logic circuit of the hardware in the processor and/or instructions in the form of software, and the steps of the method disclosed in the embodiment of the present application can be directly embodied as hardware The decoding processor is executed, or the combination of hardware and software units in the decoding processor is used to complete the execution. Optionally, the software unit may be located in a mature storage medium in the field such as random access memory, flash memory, read-only memory, programmable read-only memory, electrically erasable programmable memory, and registers. The storage medium is located in the memory, and the processor reads the information in the memory, and completes the steps in the above method embodiments in combination with its hardware.
图20是本申请实施例提供的电子设备的示意性框图。Fig. 20 is a schematic block diagram of an electronic device provided by an embodiment of the present application.
如图20所示,该电子设备30可以为本申请实施例所述的点云上采样装置,或者点云解码器,或者为模型训练装置,该电子设备30可包括:As shown in Figure 20, the electronic device 30 may be the point cloud upsampling device described in the embodiment of the present application, or a point cloud decoder, or a model training device, and the electronic device 30 may include:
存储器33和处理器32,该存储器33用于存储计算机程序34,并将该程序代码34传输给该处理器32。换言之,该处理器32可以从存储器33中调用并运行计算机程序34,以实现本申请实施例中的方法。A memory 33 and a processor 32 , the memory 33 is used to store a computer program 34 and transmit the program code 34 to the processor 32 . In other words, the processor 32 can call and run the computer program 34 from the memory 33 to implement the method in the embodiment of the present application.
例如,该处理器32可用于根据该计算机程序34中的指令执行上述方法200中的步骤。For example, the processor 32 can be used to execute the steps in the above-mentioned method 200 according to the instructions in the computer program 34 .
在本申请的一些实施例中,该处理器32可以包括但不限于:In some embodiments of the present application, the processor 32 may include, but is not limited to:
通用处理器、数字信号处理器(Digital Signal Processor,DSP)、专用集成电路(Application Specific Integrated Circuit,ASIC)、现场可编程门阵列(Field Programmable Gate Array,FPGA)或者其他可编程逻辑器件、分立门或者晶体管逻辑器件、分立硬件组件等等。General-purpose processors, digital signal processors (Digital Signal Processor, DSP), application specific integrated circuits (Application Specific Integrated Circuit, ASIC), field programmable gate arrays (Field Programmable Gate Array, FPGA) or other programmable logic devices, discrete gates Or transistor logic devices, discrete hardware components, and so on.
在本申请的一些实施例中,该存储器33包括但不限于:In some embodiments of the present application, the memory 33 includes but is not limited to:
易失性存储器和/或非易失性存储器。其中,非易失性存储器可以是只读存储器(Read-Only Memory,ROM)、可编程只读存储器(Programmable ROM,PROM)、可擦除可编程只读存储器(Erasable PROM,EPROM)、电可擦除可编程只读存储器(Electrically EPROM,EEPROM)或闪存。易失性存储器可以是随机存取存储器(Random Access Memory,RAM),其用作外部高速缓存。通过示例性但不是限制性说明,许多形式的RAM可用,例如静态随机存取存储器(Static RAM,SRAM)、动态 随机存取存储器(Dynamic RAM,DRAM)、同步动态随机存取存储器(Synchronous DRAM,SDRAM)、双倍数据速率同步动态随机存取存储器(Double Data Rate SDRAM,DDR SDRAM)、增强型同步动态随机存取存储器(Enhanced SDRAM,ESDRAM)、同步连接动态随机存取存储器(synch link DRAM,SLDRAM)和直接内存总线随机存取存储器(Direct Rambus RAM,DR RAM)。volatile memory and/or non-volatile memory. Among them, the non-volatile memory can be read-only memory (Read-Only Memory, ROM), programmable read-only memory (Programmable ROM, PROM), erasable programmable read-only memory (Erasable PROM, EPROM), electronically programmable Erase Programmable Read-Only Memory (Electrically EPROM, EEPROM) or Flash. The volatile memory can be Random Access Memory (RAM), which acts as external cache memory. By way of illustration and not limitation, many forms of RAM are available, such as Static Random Access Memory (Static RAM, SRAM), Dynamic Random Access Memory (Dynamic RAM, DRAM), Synchronous Dynamic Random Access Memory (Synchronous DRAM, SDRAM), double data rate synchronous dynamic random access memory (Double Data Rate SDRAM, DDR SDRAM), enhanced synchronous dynamic random access memory (Enhanced SDRAM, ESDRAM), synchronous connection dynamic random access memory (synch link DRAM, SLDRAM) and Direct Memory Bus Random Access Memory (Direct Rambus RAM, DR RAM).
在本申请的一些实施例中,该计算机程序34可以被分割成一个或多个单元,该一个或者多个单元被存储在该存储器33中,并由该处理器32执行,以完成本申请提供的方法。该一个或多个单元可以是能够完成特定功能的一系列计算机程序指令段,该指令段用于描述该计算机程序34在该电子设备30中的执行过程。In some embodiments of the present application, the computer program 34 can be divided into one or more units, and the one or more units are stored in the memory 33 and executed by the processor 32 to complete the present application. Methods. The one or more units may be a series of computer program instruction segments capable of accomplishing specific functions, and the instruction segments are used to describe the execution process of the computer program 34 in the electronic device 30 .
如图20所示,该电子设备30还可包括:As shown in Figure 20, the electronic device 30 may also include:
收发器33,该收发器33可连接至该处理器32或存储器33。A transceiver 33 , the transceiver 33 can be connected to the processor 32 or the memory 33 .
其中,处理器32可以控制该收发器33与其他设备进行通信,具体地,可以向其他设备发送信息或数据,或接收其他设备发送的信息或数据。收发器33可以包括发射机和接收机。收发器33还可以进一步包括天线,天线的数量可以为一个或多个。Wherein, the processor 32 can control the transceiver 33 to communicate with other devices, specifically, can send information or data to other devices, or receive information or data sent by other devices. Transceiver 33 may include a transmitter and a receiver. The transceiver 33 may further include antennas, and the number of antennas may be one or more.
应当理解,该电子设备30中的各个组件通过总线系统相连,其中,总线系统除包括数据总线之外,还包括电源总线、控制总线和状态信号总线。It should be understood that the various components in the electronic device 30 are connected through a bus system, wherein the bus system includes not only a data bus, but also a power bus, a control bus and a status signal bus.
本申请还提供了一种计算机存储介质,其上存储有计算机程序,该计算机程序被计算机执行时使得该计算机能够执行上述方法实施例的方法。或者说,本申请实施例还提供一种包含指令的计算机程序产品,该指令被计算机执行时使得计算机执行上述方法实施例的方法。The present application also provides a computer storage medium, on which a computer program is stored, and when the computer program is executed by a computer, the computer can execute the methods of the above method embodiments. In other words, the embodiments of the present application further provide a computer program product including instructions, and when the instructions are executed by a computer, the computer executes the methods of the foregoing method embodiments.
当使用软件实现时,可以全部或部分地以计算机程序产品的形式实现。该计算机程序产品包括一个或多个计算机指令。在计算机上加载和执行该计算机程序指令时,全部或部分地产生按照本申请实施例该的流程或功能。该计算机可以是通用计算机、专用计算机、计算机网络、或者其他可编程装置。该计算机指令可以存储在计算机可读存储介质中,或者从一个计算机可读存储介质向另一个计算机可读存储介质传输,例如,该计算机指令可以从一个网站站点、计算机、服务器或数据中心通过有线(例如同轴电缆、光纤、数字用户线(digital subscriber line,DSL))或无线(例如红外、无线、微波等)方式向另一个网站站点、计算机、服务器或数据中心进行传输。该计算机可读存储介质可以是计算机能够存取的任何可用介质或者是包含一个或多个可用介质集成的服务器、数据中心等数据存储设备。该可用介质可以是磁性介质(例如,软盘、硬盘、磁带)、光介质(例如数字点云光盘(digital video disc,DVD))、或者半导体介质(例如固态硬盘(solid state disk,SSD))等。When implemented using software, it may be implemented in whole or in part in the form of a computer program product. The computer program product includes one or more computer instructions. When the computer program instructions are loaded and executed on the computer, the processes or functions according to the embodiments of the present application will be generated in whole or in part. The computer can be a general purpose computer, a special purpose computer, a computer network, or other programmable device. The computer instructions may be stored in or transmitted from one computer-readable storage medium to another computer-readable storage medium, for example, the computer instructions may be transferred from a website, computer, server, or data center by wire (such as coaxial cable, optical fiber, digital subscriber line (DSL)) or wireless (such as infrared, wireless, microwave, etc.) to another website site, computer, server or data center. The computer-readable storage medium may be any available medium that can be accessed by a computer, or a data storage device such as a server or a data center integrated with one or more available media. The available medium may be a magnetic medium (such as a floppy disk, a hard disk, or a magnetic tape), an optical medium (such as a digital point cloud disc (digital video disc, DVD)), or a semiconductor medium (such as a solid state disk (solid state disk, SSD)), etc. .
本领域普通技术人员可以意识到,结合本文中所公开的实施例描述的各示例的单元及算法步骤,能够以电子硬件、或者计算机软件和电子硬件的结合来实现。这些功能究竟以硬件还是软件方式来执行,取决于技术方案的特定应用和设计约束条件。专业技术人员可以对每个特定的应用来使用不同方法来实现所描述的功能,但是这种实现不应认为超出本申请的范围。Those skilled in the art can appreciate that the units and algorithm steps of the examples described in conjunction with the embodiments disclosed herein can be implemented by electronic hardware, or a combination of computer software and electronic hardware. Whether these functions are executed by hardware or software depends on the specific application and design constraints of the technical solution. Skilled artisans may use different methods to implement the described functions for each specific application, but such implementation should not be regarded as exceeding the scope of the present application.
在本申请所提供的几个实施例中,应该理解到,所揭露的系统、装置和方法,可以通过其它的方式实现。例如,以上所描述的装置实施例仅仅是示意性的,例如,该单元的划分,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分方式,例如多个单元或组件可以结合或者可以集成到另一个系统,或一些特征可以忽略,或不执行。另一点,所显示或讨论的相互之间的耦合或直接耦合或通信连接可以是通过一些接口,装置或单元的间接耦合或通信连接,可以是电性,机械或其它的形式。In the several embodiments provided in this application, it should be understood that the disclosed systems, devices and methods may be implemented in other ways. For example, the device embodiments described above are only illustrative. For example, the division of the units is only a logical function division. In actual implementation, there may be other division methods. For example, multiple units or components can be combined or can be Integrate into another system, or some features may be ignored, or not implemented. In another point, the mutual coupling or direct coupling or communication connection shown or discussed may be through some interfaces, and the indirect coupling or communication connection of devices or units may be in electrical, mechanical or other forms.
作为分离部件说明的单元可以是或者也可以不是物理上分开的,作为单元显示的部件可以是或者也可以不是物理单元,即可以位于一个地方,或者也可以分布到多个网络单元上。可以根据实际的需要选择其中的部分或者全部单元来实现本实施例方案的目的。例如,在本申请各个实施例中的各功能单元可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中。A unit described as a separate component may or may not be physically separated, and a component displayed as a unit may or may not be a physical unit, that is, it may be located in one place, or may be distributed to multiple network units. Part or all of the units can be selected according to actual needs to achieve the purpose of the solution of this embodiment. For example, each functional unit in each embodiment of the present application may be integrated into one processing unit, each unit may exist separately physically, or two or more units may be integrated into one unit.
以上内容,仅为本申请的具体实施方式,但本申请的保护范围并不局限于此,任何熟悉本技术领域的技术人员在本申请揭露的技术范围内,可轻易想到变化或替换,都应涵盖在本申请的保护范围之内。因此,本申请的保护范围应以该权利要求的保护范围为准。The above content is only the specific implementation of the application, but the scope of protection of the application is not limited thereto. Anyone familiar with the technical field can easily think of changes or substitutions within the technical scope disclosed in the application, and should covered within the scope of protection of this application. Therefore, the protection scope of the present application should be based on the protection scope of the claims.

Claims (100)

  1. 一种点云解码方法,其特征在于,包括:A point cloud decoding method is characterized in that, comprising:
    解码点云码流,得到点云的几何信息;Decode the point cloud code stream to obtain the geometric information of the point cloud;
    根据所述点云的几何信息,将所述点云划分成至少一个点云块;dividing the point cloud into at least one point cloud block according to the geometric information of the point cloud;
    将所述点云块的几何信息输入生成器中进行上采样,得到所述点云块的上采样几何信息;Input the geometric information of the point cloud block into the generator for upsampling, and obtain the upsampling geometric information of the point cloud block;
    其中,所述生成器包括:特征提取模块、特征上采样模块和几何生成模块,所述特征提取模块用于提取所述点云块的第一特征信息,所述特征采样模块用于将所述点云块的第一特征信息上采样为第二特征信息,所述几何生成模块用于将所述点云块的第二特征信息映射至几何空间中,以得到所述点云块的上采样几何信息。Wherein, the generator includes: a feature extraction module, a feature upsampling module and a geometry generation module, the feature extraction module is used to extract the first feature information of the point cloud block, and the feature sampling module is used to extract the The first feature information of the point cloud block is up-sampled to the second feature information, and the geometry generation module is used to map the second feature information of the point cloud block into a geometric space, so as to obtain the up-sampling of the point cloud block geometric information.
  2. 根据权利要求1所述的方法,其特征在于,所述特征提取模块包括密集连接的M个特征提取块;The method according to claim 1, wherein the feature extraction module comprises densely connected M feature extraction blocks;
    对于所述M个特征提取块中的第i+1个特征提取块,所述第i+1个特征提取块用于根据输入的第i个第四特征信息输出第i+1个第三特征信息,所述第i个第四特征信息是根据第i个特征提取块输出的第i个第三特征信息确定的,所述点云块的第一特征信息是根据所述M个特征提取块中第M个特征提取块所输出的第M个第三特征信息确定的,所述i为小于M的正整数。For the i+1th feature extraction block in the M feature extraction blocks, the i+1th feature extraction block is used to output the i+1th third feature according to the input i-th fourth feature information information, the i-th fourth feature information is determined according to the i-th third feature information output by the i-th feature extraction block, and the first feature information of the point cloud block is determined according to the M feature extraction blocks Determined by the Mth third feature information output by the Mth feature extraction block, the i is a positive integer smaller than M.
  3. 根据权利要求2所述的方法,其特征在于,The method according to claim 2, characterized in that,
    若i不等于1,则所述第i个第四特征信息为所述M个特征提取块中位于所述第i个特征提取块之前的各特征提取块所提取的第三特征信息、与所述第i个特征提取块所提取的第三特征信息进行级联后的特征信息;If i is not equal to 1, the i-th fourth feature information is the third feature information extracted by each feature extraction block before the i-th feature extraction block in the M feature extraction blocks, and the The feature information after cascading the third feature information extracted by the i-th feature extraction block;
    若i等于1,则所述第i个第四特征信息为所述M个特征提取块中第一个特征提取块所输出的第一个第三特征信息。If i is equal to 1, the ith fourth feature information is the first third feature information output by the first feature extraction block in the M feature extraction blocks.
  4. 根据权利要求2所述的方法,其特征在于,所述特征提取块包括:第一特征提取单元和串联连接的S个第二特征提取单元,所述S为正整数;The method according to claim 2, wherein the feature extraction block comprises: a first feature extraction unit and S second feature extraction units connected in series, wherein S is a positive integer;
    对于所述第i+1个特征提取块中的第一提取单元,所述第一提取单元用于针对所述点云块中的当前点,搜索所述当前点的K个邻近点,并基于所述点云块的第i个第四特征信息,将所述当前点的第四特征信息与所述邻近点的第四特征信息进行相减,得到K个残差特征信息,并将所述K个残差特征信息与所述当前点的第四特征信息进行级联,得到所述当前点的第i个级联特征信息,根据所述当前点的第i个级联特征信息,得到所述点云块的第i个级联特征信息,并将所述点云块的第i个级联特征信息输入所述S个第二特征提取单元中的第一个第二特征提取单元;For the first extraction unit in the i+1th feature extraction block, the first extraction unit is used to search for K neighboring points of the current point for the current point in the point cloud block, and based on For the ith fourth feature information of the point cloud block, the fourth feature information of the current point is subtracted from the fourth feature information of the adjacent point to obtain K residual feature information, and the The K residual feature information is concatenated with the fourth feature information of the current point to obtain the i-th concatenated feature information of the current point, and according to the i-th concatenated feature information of the current point, the The i-th concatenated feature information of the point cloud block, and input the i-th concatenated feature information of the point cloud block into the first second feature extraction unit in the S second feature extraction units;
    所述第一个第二特征提取单元用于根据所述点云块的第i个级联特征信息,输出第一个第五特征信息至第二个第二特征提取单元,其中所述点云块的第i+1个第三特征信息为所述S个第二特征提取单元中最后一个第二特征提取单元输出的第五特征信息。The first second feature extraction unit is used to output the first fifth feature information to the second second feature extraction unit according to the i-th cascaded feature information of the point cloud block, wherein the point cloud The i+1th third feature information of the block is the fifth feature information output by the last second feature extraction unit among the S second feature extraction units.
  5. 根据权利要求4所述的方法,其特征在于,所述第二特征提取单元包括P个残差块,所述P为正整数;The method according to claim 4, wherein the second feature extraction unit comprises P residual blocks, and the P is a positive integer;
    对于第s个第二特征提取单元中的第j+1个残差块,所述第j+1个残差块用于根据所述第s个第二特征提取单元中的第j个残差块所输出的第j个第一残差信息和输入所述第s个第二特征提取单元的第五特征信息,输出第j+1个第一残差信息,其中,所述j为小于P的正整数,所述s为小于或等于S的正整数;For the j+1th residual block in the sth second feature extraction unit, the j+1th residual block is used according to the jth residual in the sth second feature extraction unit The j-th first residual information output by the block and the fifth feature information input to the s-th second feature extraction unit, and the j+1-th first residual information is output, wherein the j is less than P A positive integer, the s is a positive integer less than or equal to S;
    所述第s个第二特征提取单元输出的第五特征信息是根据所述第s个第二特征提取单元中至少一个残差块输出的第一残差信息信息,以及输入所述第s个第二特征提取单元的第五特征信息确定的。The fifth feature information output by the sth second feature extraction unit is based on the first residual information output by at least one residual block in the sth second feature extraction unit, and input to the sth second feature extraction unit determined by the fifth feature information of the second feature extraction unit.
  6. 根据权利要求5所述的方法,其特征在于,所述方法还包括:将所述第s个第二特征提取单元中的第j个残差块所输出的第j个第一残差信息和输入所述第s个第二特征提取单元的第五特征信息进行相加后,输入所述第s个第二特征提取单元中的第j+1个残差块。The method according to claim 5, wherein the method further comprises: combining the jth first residual information output by the jth residual block in the sth second feature extraction unit and After the fifth feature information is input to the s th second feature extraction unit for addition, it is input to the j+1 th residual block in the s th second feature extraction unit.
  7. 根据权利要求5所述的方法,其特征在于,所述第s个第二特征提取单元输出的第五特征信息是根据所述第s个第二特征提取单元中最后一个残差块输出的第一残差信息信息、与P-1个残差块中至少一个残差块输出的第一残差信息信息进行级联后的特征信息、与输入所述第s个第二特征提取 单元的第五特征信息确定的,其中,所述P-1个残差块为所述第s个第二特征提取单元的P个残差块中除最后一个残差块之外的残差块。The method according to claim 5, wherein the fifth feature information output by the s second feature extraction unit is based on the output of the last residual block in the s second feature extraction unit. Residual information information, feature information concatenated with the first residual information information output by at least one residual block in the P-1 residual blocks, and input to the sth second feature extraction unit. determined by five feature information, wherein, the P-1 residual blocks are residual blocks except the last residual block among the P residual blocks of the s-th second feature extraction unit.
  8. 根据权利要求7所述的方法,其特征在于,所述第s个第二特征提取单元输出的第五特征信息是根据所述第s个第二特征提取单元中最后一个残差块输出的第一残差信息信息、与P-1个残差块中至少一个残差块输出的第一残差信息信息进行级联后特征信息、与输入所述第s个第二特征提取单元的第五特征信息进行相加后确定的。The method according to claim 7, wherein the fifth feature information output by the s second feature extraction unit is based on the output of the last residual block in the s second feature extraction unit. Residual information information, feature information concatenated with the first residual information information output by at least one residual block in the P-1 residual blocks, and the fifth input to the sth second feature extraction unit The feature information is determined after adding.
  9. 根据权利要求8所述的方法,其特征在于,所述第二特征提取单元还包括门控单元,The method according to claim 8, wherein the second feature extraction unit further comprises a gating unit,
    对于所述第s个第二特征提取单元中的门控单元,所述门控单元用于对所述第s个第二特征提取单元中的最后一个残差块输出的第一残差信息、与所述P-1个残差块中至少一个残差块输出的第一残差信息级联后的特征信息进行去冗余,输出去冗余后的特征信息;所述第s个第二特征提取单元输出的第五特征信息是根据所述去冗余后的特征信息和所述输入所述第s个第二特征提取单元的第五特征信息进行相加后确定的。For the gating unit in the s th second feature extraction unit, the gating unit is used for the first residual information output by the last residual block in the s th second feature extraction unit, De-redundancy is performed on the feature information concatenated with the first residual information output by at least one residual block in the P-1 residual blocks, and the de-redundant feature information is output; the s second The fifth feature information output by the feature extraction unit is determined after adding the de-redundant feature information and the fifth feature information input to the s-th second feature extraction unit.
  10. 根据权利要求1-9任一项所述的方法,其特征在于,所述特征上采样模块包括:特征上采样子模块和特征提取子模块;The method according to any one of claims 1-9, wherein the feature upsampling module comprises: a feature upsampling submodule and a feature extraction submodule;
    所述特征上采样子模块用于按照预设的上采样率r,将所述点云块的第一特征信息复制r份,并对复制后的第一特征信息在特征维度上增加一个n维向量,得到所述点云块的上采样特征信息,并将所述点云块的上采样特征信息输入特征提取子模块,其中不同第一特征信息对应的n维向量的值不相同;The feature upsampling submodule is used to copy r copies of the first feature information of the point cloud block according to the preset upsampling rate r, and add an n dimension to the feature dimension of the copied first feature information vector, obtain the upsampling feature information of the point cloud block, and input the upsampling feature information of the point cloud block into the feature extraction submodule, wherein the values of n-dimensional vectors corresponding to different first feature information are different;
    所述特征提取子模块用于根据所述点云块的上采样特征信息,输出所述点云块的第二特征信息。The feature extraction submodule is configured to output second feature information of the point cloud block according to the upsampled feature information of the point cloud block.
  11. 根据权利要求10所述的方法,其特征在于,所述特征提取子模块包括Q个第三特征提取单元,所述Q为正整数;The method according to claim 10, wherein the feature extraction submodule includes Q third feature extraction units, and the Q is a positive integer;
    针对所述Q个第三特征提取单元中的第k+1个第三特征提取单元,所述第k+1个第三特征提取单元用于根据第k个第三特征提取单元所提取的所述点云块的第k个增强上采样特征信息,输出所述点云块的第k+1个增强上采样特征信息,所述k为小于Q的正整数;For the k+1th third feature extraction unit among the Q third feature extraction units, the k+1th third feature extraction unit is used to extract all The kth enhanced upsampling feature information of the point cloud block, output the k+1th enhanced upsampling feature information of the point cloud block, and the k is a positive integer less than Q;
    所述点云块的第二特征信息为所述Q个第三特征提取单元中最后一个第三特征提取单元所提取的所述点云块的第Q个增强上采样特征信息。The second feature information of the point cloud block is the Qth enhanced upsampling feature information of the point cloud block extracted by the last third feature extraction unit among the Q third feature extraction units.
  12. 根据权利要求11所述的方法,其特征在于,所述第三特征提取单元包括L个残差块,所述L为正整数;The method according to claim 11, wherein the third feature extraction unit includes L residual blocks, and the L is a positive integer;
    对于所述第k+1个第三特征提取单元中的第l+1个残差块,所述第l+1个残差块用于根据所述第k+1个第三特征提取单元中的第l个残差块输出的第l个第二残差信息和输入所述第k+1个第三特征提取单元的第k个增强上采样特征信息,输出第l+1个第二残差信息,所述l为小于L的正整数;For the l+1th residual block in the k+1th third feature extraction unit, the l+1th residual block is used according to the k+1th third feature extraction unit The lth second residual information output by the lth residual block and the kth enhanced upsampling feature information input to the k+1th third feature extraction unit, output the l+1th second residual difference information, the l is a positive integer less than L;
    所述点云块的第k+1个增强上采样特征信息是根据所述第k+1个第三特征提取单元中至少一个残差块输出的第二残差信息,以及所述第k个增强上采样特征信息确定的。The k+1th enhanced upsampling feature information of the point cloud block is the second residual information output from at least one residual block in the k+1th third feature extraction unit, and the kth Enhanced upsampling feature information determined.
  13. 根据权利要求12所述的方法,其特征在于,所述方法还包括:对所述第l个残差块输出的第l个第二残差信息和所述第k个增强上采样特征信息进行相加后,输入所述第l+1个残差块。The method according to claim 12, characterized in that, the method further comprises: performing an operation on the lth second residual information output by the lth residual block and the kth enhanced upsampling feature information After the addition, input the l+1th residual block.
  14. 根据权利要求13所述的方法,其特征在于,所述点云块的第k+1个增强上采样特征信息根据所述L个残差块中最后一个残差块输出的第二残差信息、与L-1个残差块中至少一个残差块输出的第二残差信息进行级联后的特征信息和所述第k个增强上采样特征信息确定的,其中,所述L-1个残差块为所述第k+1个第三特征提取单元的L个残差块中除最后一个残差块之外的残差块。The method according to claim 13, characterized in that, the k+1th enhanced upsampled feature information of the point cloud block is based on the second residual information output by the last residual block in the L residual blocks , determined by the feature information concatenated with the second residual information output by at least one residual block in the L-1 residual blocks and the kth enhanced upsampling feature information, wherein the L-1 The residual blocks are the residual blocks except the last residual block among the L residual blocks of the k+1 third feature extraction unit.
  15. 根据权利要求14所述的方法,其特征在于,所述点云块的第k+1个增强上采样特征信息根据所述L个残差块中最后一个残差块输出的第二残差信息、与L-1个残差块中至少一个残差块输出的第二残差信息进行级联后的特征信息和所述第k个增强上采样特征信息进行相加后确定的。The method according to claim 14, wherein the k+1th enhanced upsampling feature information of the point cloud block is based on the second residual information output by the last residual block in the L residual blocks , and determined after adding the feature information concatenated with the second residual information output by at least one residual block in the L-1 residual blocks and the kth enhanced upsampling feature information.
  16. 根据权利要求15所述的方法,其特征在于,所述第三特征提取单元还包括门控单元;The method according to claim 15, wherein the third feature extraction unit further comprises a gating unit;
    对于所述第k+1个第三特征提取单元中的门控单元,所述门控单元用于对所述第k+1个第三特征提取单元中的最后一个残差块输出的第二残差信息、与所述L-1个残差块中至少一个残差块输出的第二特征信息进行级联后的特征信息进行去冗余,输出去冗余后的特征信息;For the gating unit in the k+1th third feature extraction unit, the gating unit is used for the second output of the last residual block in the k+1th third feature extraction unit performing de-redundancy on the residual information and the feature information concatenated with the second feature information output by at least one of the L-1 residual blocks, and outputting the de-redundant feature information;
    所述点云块的第k+1个增强上采样特征信息是根据去冗余后的特征信息与所述第k个增强上采 样特征信息进行相加后确定的。The k+1th enhanced upsampling feature information of the point cloud block is determined after adding the kth enhanced upsampling feature information according to the deredundant feature information.
  17. 根据权利要求10所述的方法,其特征在于,所述特征上采样模块还包括第一自相关注意力网络;The method according to claim 10, wherein the feature upsampling module further comprises a first autocorrelation attention network;
    所述第一自相关注意力网络用于对所述特征上采样子模块输出的所述点云块的上采样特征信息进行特征交互,输出特征交互后的所述点云块的上采样特征信息至所述特征提取子模块;The first autocorrelation attention network is used to perform feature interaction on the upsampling feature information of the point cloud block output by the feature upsampling submodule, and output the upsampling feature information of the point cloud block after feature interaction to the feature extraction submodule;
    所述特征提取子模块用于根据特征交互后的所述点云块的上采样特征信息,输出所述点云块的第二特征信息。The feature extraction submodule is configured to output second feature information of the point cloud block according to the upsampled feature information of the point cloud block after feature interaction.
  18. 根据权利要求17所述的方法,其特征在于,所述特征交互后的所述点云块的上采样特征信息的特征维度低于所述点云块的上采样特征信息的特征维度。The method according to claim 17, wherein the feature dimension of the upsampled feature information of the point cloud block after the feature interaction is lower than the feature dimension of the upsampled feature information of the point cloud block.
  19. 根据权利要求1所述的方法,其特征在于,所述几何生成模块包括多个全连接层;The method according to claim 1, wherein the geometry generation module comprises a plurality of fully connected layers;
    所述多个全连接层用于根据所述点云块的第二特征信息,输出所述点云块的上采样几何信息。The multiple fully connected layers are used to output upsampled geometric information of the point cloud block according to the second feature information of the point cloud block.
  20. 根据权利要求1所述的方法,其特征在于,所述几何生成模块包括:几何重建单元、滤波单元和下采样单元;The method according to claim 1, wherein the geometry generation module comprises: a geometry reconstruction unit, a filtering unit, and a downsampling unit;
    所述几何重建单元用于对所述将所述点云块的第二特征信息进行几何重建,输出所述点云块的初始上采样几何信息至所述滤波单元;The geometric reconstruction unit is used to geometrically reconstruct the second feature information of the point cloud block, and output the initial upsampling geometric information of the point cloud block to the filtering unit;
    所述滤波单元用于对所述点云块的初始上采样几何信息进行除噪,输出所述点云块滤除噪点的初始上采样几何信息至所述下采样单元;The filtering unit is used to denoise the initial upsampling geometric information of the point cloud block, and output the initial upsampling geometric information of the point cloud block to filter noise to the downsampling unit;
    所述下采样单元用于对所述点云块滤除噪点的初始上采样几何信息下采样至目标上采样率,输出所述点云块的上采样几何信息。The down-sampling unit is configured to down-sample the initial up-sampled geometric information of the point cloud block after filtering noise to a target up-sampling rate, and output the up-sampled geometric information of the point cloud block.
  21. 根据权利要求20所述的方法,其特征在于,所述目标上采样率小于或等于所述特征上采样模块的上采样率。The method according to claim 20, wherein the target upsampling rate is less than or equal to the upsampling rate of the feature upsampling module.
  22. 根据权利要求20所述的方法,其特征在于,所述方法还包括:The method according to claim 20, further comprising:
    解码所述点云码流,得到所述目标上采样率。Decoding the point cloud stream to obtain the target upsampling rate.
  23. 一种点云解码器,其特征在于,包括:A point cloud decoder, characterized in that it comprises:
    解码单元,用于解码点云码流,得到点云的几何信息;The decoding unit is used to decode the point cloud code stream to obtain the geometric information of the point cloud;
    划分单元,用于根据所述点云的几何信息,将所述点云划分成至少一个点云块;A division unit, configured to divide the point cloud into at least one point cloud block according to the geometric information of the point cloud;
    上采样单元,用于将所述点云块的几何信息输入生成器中进行上采样,得到所述点云块的上采样几何信息;An upsampling unit, configured to input the geometric information of the point cloud block into the generator for upsampling, to obtain the upsampled geometric information of the point cloud block;
    其中,所述生成器包括:特征提取模块、特征上采样模块和几何生成模块,所述特征提取模块用于提取所述点云块的第一特征信息,所述特征采样模块用于将所述点云块的第一特征信息上采样为第二特征信息,所述几何生成模块用于将所述点云块的第二特征信息映射至几何空间中,以得到所述点云块的上采样几何信息。Wherein, the generator includes: a feature extraction module, a feature upsampling module and a geometry generation module, the feature extraction module is used to extract the first feature information of the point cloud block, and the feature sampling module is used to extract the The first feature information of the point cloud block is up-sampled to the second feature information, and the geometry generation module is used to map the second feature information of the point cloud block into a geometric space, so as to obtain the up-sampling of the point cloud block geometric information.
  24. 一种点云解码器,其特征在于,包括:处理器和存储器;A point cloud decoder, characterized in that, comprising: a processor and a memory;
    所述存储器用于存储计算机程序;The memory is used to store computer programs;
    所述处理器用于调用并运行所述存储器中存储的计算机程序,以执行如权利要求1-22任一项所述的点云编码方法。The processor is used for invoking and running the computer program stored in the memory, so as to execute the point cloud encoding method according to any one of claims 1-22.
  25. 一种点云上采样方法,其特征在于,包括:A point cloud upsampling method is characterized in that, comprising:
    获取待上采样点云的几何信息;Obtain the geometric information of the point cloud to be upsampled;
    根据所述待上采样点云的几何信息,将所述待上采样点云划分成至少一个点云块;Divide the point cloud to be upsampled into at least one point cloud block according to the geometric information of the point cloud to be upsampled;
    将所述点云块的几何信息输入生成器中进行上采样,得到所述点云块的上采样几何信息;Input the geometric information of the point cloud block into the generator for upsampling, and obtain the upsampling geometric information of the point cloud block;
    其中,所述生成器包括:特征提取模块、特征上采样模块和几何生成模块,所述特征提取模块用于提取所述点云块的第一特征信息,所述特征采样模块用于将所述点云块的第一特征信息上采样为第二特征信息,所述几何生成模块用于将所述点云块的第二特征信息映射至几何空间中,以得到所述点云块的上采样几何信息。Wherein, the generator includes: a feature extraction module, a feature upsampling module and a geometry generation module, the feature extraction module is used to extract the first feature information of the point cloud block, and the feature sampling module is used to extract the The first feature information of the point cloud block is up-sampled to the second feature information, and the geometry generation module is used to map the second feature information of the point cloud block into a geometric space, so as to obtain the up-sampling of the point cloud block geometric information.
  26. 根据权利要求25所述的方法,其特征在于,所述特征提取模块包括密集连接的M个特征提 取块;The method according to claim 25, wherein the feature extraction module comprises densely connected M feature extraction blocks;
    对于所述M个特征提取块中的第i+1个特征提取块,所述第i+1个特征提取块用于根据输入的第i个第四特征信息输出第i+1个第三特征信息,所述第i个第四特征信息是根据第i个特征提取块输出的第i个第三特征信息确定的,所述点云块的第一特征信息是根据所述M个特征提取块中第M个特征提取块所输出的第M个第三特征信息确定的,所述i为小于M的正整数。For the i+1th feature extraction block in the M feature extraction blocks, the i+1th feature extraction block is used to output the i+1th third feature according to the input i-th fourth feature information information, the i-th fourth feature information is determined according to the i-th third feature information output by the i-th feature extraction block, and the first feature information of the point cloud block is determined according to the M feature extraction blocks Determined by the Mth third feature information output by the Mth feature extraction block, the i is a positive integer smaller than M.
  27. 根据权利要求26所述的方法,其特征在于,The method of claim 26, wherein
    若i不等于1,则所述第i个第四特征信息为所述M个特征提取块中位于所述第i个特征提取块之前的各特征提取块所提取的第三特征信息、与所述第i个特征提取块所提取的第三特征信息进行级联后的特征信息;If i is not equal to 1, the i-th fourth feature information is the third feature information extracted by each feature extraction block before the i-th feature extraction block in the M feature extraction blocks, and the The feature information after cascading the third feature information extracted by the i-th feature extraction block;
    若i等于1,则所述第i个第四特征信息为所述M个特征提取块中第一个特征提取块所输出的第一个第三特征信息。If i is equal to 1, the ith fourth feature information is the first third feature information output by the first feature extraction block in the M feature extraction blocks.
  28. 根据权利要求26所述的方法,其特征在于,所述特征提取块包括:第一特征提取单元和串联连接的S个第二特征提取单元,所述S为正整数;The method according to claim 26, wherein the feature extraction block comprises: a first feature extraction unit and S second feature extraction units connected in series, wherein S is a positive integer;
    对于所述第i+1个特征提取块中的第一提取单元,所述第一提取单元用于针对所述点云块中的当前点,搜索所述当前点的K个邻近点,并基于所述点云块的第i个第四特征信息,将所述当前点的第四特征信息与所述邻近点的第四特征信息进行相减,得到K个残差特征信息,并将所述K个残差特征信息与所述当前点的第四特征信息进行级联,得到所述当前点的第i个级联特征信息,根据所述当前点的第i个级联特征信息,得到所述点云块的第i个级联特征信息,并将所述点云块的第i个级联特征信息输入所述S个第二特征提取单元中的第一个第二特征提取单元;For the first extraction unit in the i+1th feature extraction block, the first extraction unit is used to search for K neighboring points of the current point for the current point in the point cloud block, and based on For the ith fourth feature information of the point cloud block, the fourth feature information of the current point is subtracted from the fourth feature information of the adjacent point to obtain K residual feature information, and the The K residual feature information is concatenated with the fourth feature information of the current point to obtain the i-th concatenated feature information of the current point, and according to the i-th concatenated feature information of the current point, the The i-th concatenated feature information of the point cloud block, and input the i-th concatenated feature information of the point cloud block into the first second feature extraction unit in the S second feature extraction units;
    所述第一个第二特征提取单元用于根据所述点云块的第i个级联特征信息,输出第一个第五特征信息至第二个第二特征提取单元,其中所述点云块的第i+1个第三特征信息为所述S个第二特征提取单元中最后一个第二特征提取单元输出的第五特征信息。The first second feature extraction unit is used to output the first fifth feature information to the second second feature extraction unit according to the i-th cascaded feature information of the point cloud block, wherein the point cloud The i+1th third feature information of the block is the fifth feature information output by the last second feature extraction unit among the S second feature extraction units.
  29. 根据权利要求28所述的方法,其特征在于,所述第二特征提取单元包括P个残差块,所述P为正整数;The method according to claim 28, wherein the second feature extraction unit comprises P residual blocks, and the P is a positive integer;
    对于第s个第二特征提取单元中的第j+1个残差块,所述第j+1个残差块用于根据所述第s个第二特征提取单元中的第j个残差块所输出的第j个第一残差信息和输入所述第s个第二特征提取单元的第五特征信息,输出第j+1个第一残差信息,其中,所述j为小于P的正整数,所述s为小于或等于S的正整数;For the j+1th residual block in the sth second feature extraction unit, the j+1th residual block is used according to the jth residual in the sth second feature extraction unit The j-th first residual information output by the block and the fifth feature information input to the s-th second feature extraction unit, and the j+1-th first residual information is output, wherein the j is less than P A positive integer, the s is a positive integer less than or equal to S;
    所述第s个第二特征提取单元输出的第五特征信息是根据所述第s个第二特征提取单元中至少一个残差块输出的第一残差信息信息,以及输入所述第s个第二特征提取单元的第五特征信息确定的。The fifth feature information output by the sth second feature extraction unit is based on the first residual information output by at least one residual block in the sth second feature extraction unit, and input to the sth second feature extraction unit determined by the fifth feature information of the second feature extraction unit.
  30. 根据权利要求29所述的方法,其特征在于,所述方法还包括:将所述第s个第二特征提取单元中的第j个残差块所输出的第j个第一残差信息和输入所述第s个第二特征提取单元的第五特征信息进行相加后,输入所述第s个第二特征提取单元中的第j+1个残差块。The method according to claim 29, characterized in that the method further comprises: combining the jth first residual information output by the jth residual block in the sth second feature extraction unit and After the fifth feature information is input to the s th second feature extraction unit for addition, it is input to the j+1 th residual block in the s th second feature extraction unit.
  31. 根据权利要求29所述的方法,其特征在于,所述第s个第二特征提取单元输出的第五特征信息是根据所述第s个第二特征提取单元中最后一个残差块输出的第一残差信息信息、与P-1个残差块中至少一个残差块输出的第一残差信息信息进行级联后的特征信息、与输入所述第s个第二特征提取单元的第五特征信息确定的,其中,所述P-1个残差块为所述第s个第二特征提取单元的P个残差块中除最后一个残差块之外的残差块。The method according to claim 29, wherein the fifth feature information output by the s th second feature extraction unit is based on the s th second feature information output by the last residual block in the s th second feature extraction unit Residual information information, feature information concatenated with the first residual information information output by at least one residual block in the P-1 residual blocks, and input to the sth second feature extraction unit. determined by five feature information, wherein, the P-1 residual blocks are residual blocks except the last residual block among the P residual blocks of the s-th second feature extraction unit.
  32. 根据权利要求31所述的方法,其特征在于,所述第s个第二特征提取单元输出的第五特征信息是根据所述第s个第二特征提取单元中最后一个残差块输出的第一残差信息信息、与P-1个残差块中至少一个残差块输出的第一残差信息信息进行级联后特征信息、与输入所述第s个第二特征提取单元的第五特征信息进行相加后确定的。The method according to claim 31, wherein the fifth feature information output by the s th second feature extraction unit is based on the s th second feature information output by the last residual block in the s th second feature extraction unit Residual information information, feature information concatenated with the first residual information information output by at least one residual block in the P-1 residual blocks, and the fifth input to the sth second feature extraction unit The feature information is determined after adding.
  33. 根据权利要求32所述的方法,其特征在于,所述第二特征提取单元还包括门控单元,The method according to claim 32, wherein the second feature extraction unit further comprises a gating unit,
    对于所述第s个第二特征提取单元中的门控单元,所述门控单元用于对所述第s个第二特征提取单元中的最后一个残差块输出的第一残差信息、与所述P-1个残差块中至少一个残差块输出的第一残差信息级联后的特征信息进行去冗余,输出去冗余后的特征信息;所述第s个第二特征提取单元输出的第五特征信息是根据所述去冗余后的特征信息和所述输入所述第s个第二特征提取单元的第五特征 信息进行相加后确定的。For the gating unit in the s th second feature extraction unit, the gating unit is used for the first residual information output by the last residual block in the s th second feature extraction unit, De-redundancy is performed on the feature information concatenated with the first residual information output by at least one residual block in the P-1 residual blocks, and the de-redundant feature information is output; the s second The fifth feature information output by the feature extraction unit is determined after adding the de-redundant feature information and the fifth feature information input to the s-th second feature extraction unit.
  34. 根据权利要求25-33任一项所述的方法,其特征在于,所述特征上采样模块包括:特征上采样子模块和特征提取子模块;The method according to any one of claims 25-33, wherein the feature upsampling module comprises: a feature upsampling submodule and a feature extraction submodule;
    所述特征上采样子模块用于按照预设的上采样率r,将所述点云块的第一特征信息复制r份,并对复制后的第一特征信息在特征维度上增加一个n维向量,得到所述点云块的上采样特征信息,并将所述点云块的上采样特征信息输入特征提取子模块,其中不同第一特征信息对应的n维向量的值不相同;The feature upsampling submodule is used to copy r copies of the first feature information of the point cloud block according to the preset upsampling rate r, and add an n dimension to the feature dimension of the copied first feature information vector, obtain the upsampling feature information of the point cloud block, and input the upsampling feature information of the point cloud block into the feature extraction submodule, wherein the values of n-dimensional vectors corresponding to different first feature information are different;
    所述特征提取子模块用于根据所述点云块的上采样特征信息,输出所述点云块的第二特征信息。The feature extraction submodule is configured to output second feature information of the point cloud block according to the upsampled feature information of the point cloud block.
  35. 根据权利要求34所述的方法,其特征在于,所述特征提取子模块包括Q个第三特征提取单元,所述Q为正整数;The method according to claim 34, wherein the feature extraction submodule includes Q third feature extraction units, and the Q is a positive integer;
    针对所述Q个第三特征提取单元中的第k+1个第三特征提取单元,所述第k+1个第三特征提取单元用于根据第k个第三特征提取单元所提取的所述点云块的第k个增强上采样特征信息,输出所述点云块的第k+1个增强上采样特征信息,所述k为小于Q的正整数;For the k+1th third feature extraction unit among the Q third feature extraction units, the k+1th third feature extraction unit is used to extract all The kth enhanced upsampling feature information of the point cloud block, output the k+1th enhanced upsampling feature information of the point cloud block, and the k is a positive integer less than Q;
    所述点云块的第二特征信息为所述Q个第三特征提取单元中最后一个第三特征提取单元所提取的所述点云块的第Q个增强上采样特征信息。The second feature information of the point cloud block is the Qth enhanced upsampling feature information of the point cloud block extracted by the last third feature extraction unit among the Q third feature extraction units.
  36. 根据权利要求35所述的方法,其特征在于,所述第三特征提取单元包括L个残差块,所述L为正整数;The method according to claim 35, wherein the third feature extraction unit comprises L residual blocks, and the L is a positive integer;
    对于所述第k+1个第三特征提取单元中的第l+1个残差块,所述第l+1个残差块用于根据所述第k+1个第三特征提取单元中的第l个残差块输出的第l个第二残差信息和输入所述第k+1个第三特征提取单元的第k个增强上采样特征信息,输出第l+1个第二残差信息,所述l为小于L的正整数;For the l+1th residual block in the k+1th third feature extraction unit, the l+1th residual block is used according to the k+1th third feature extraction unit The lth second residual information output by the lth residual block and the kth enhanced upsampling feature information input to the k+1th third feature extraction unit, output the l+1th second residual difference information, the l is a positive integer less than L;
    所述点云块的第k+1个增强上采样特征信息是根据所述第k+1个第三特征提取单元中至少一个残差块输出的第二残差信息,以及所述第k个增强上采样特征信息确定的。The k+1th enhanced upsampling feature information of the point cloud block is the second residual information output from at least one residual block in the k+1th third feature extraction unit, and the kth Enhanced upsampling feature information determined.
  37. 根据权利要求36所述的方法,其特征在于,所述方法还包括:对所述第l个残差块输出的第l个第二残差信息和所述第k个增强上采样特征信息进行相加后,输入所述第l+1个残差块。The method according to claim 36, characterized in that, the method further comprises: performing an operation on the lth second residual information output by the lth residual block and the kth enhanced upsampling feature information After the addition, input the l+1th residual block.
  38. 根据权利要求37所述的方法,其特征在于,所述点云块的第k+1个增强上采样特征信息根据所述L个残差块中最后一个残差块输出的第二残差信息、与L-1个残差块中至少一个残差块输出的第二残差信息进行级联后的特征信息和所述第k个增强上采样特征信息确定的,其中,所述L-1个残差块为所述第k+1个第三特征提取单元的L个残差块中除最后一个残差块之外的残差块。The method according to claim 37, characterized in that, the k+1th enhanced upsampling feature information of the point cloud block is based on the second residual information output by the last residual block in the L residual blocks , determined by the feature information concatenated with the second residual information output by at least one residual block in the L-1 residual blocks and the kth enhanced upsampling feature information, wherein the L-1 The residual blocks are the residual blocks except the last residual block among the L residual blocks of the k+1 third feature extraction unit.
  39. 根据权利要求38所述的方法,其特征在于,所述点云块的第k+1个增强上采样特征信息根据所述L个残差块中最后一个残差块输出的第二残差信息、与L-1个残差块中至少一个残差块输出的第二残差信息进行级联后的特征信息和所述第k个增强上采样特征信息进行相加后确定的。The method according to claim 38, characterized in that, the k+1th enhanced upsampled feature information of the point cloud block is based on the second residual information output by the last residual block in the L residual blocks , and determined after adding the feature information concatenated with the second residual information output by at least one residual block in the L-1 residual blocks and the kth enhanced upsampling feature information.
  40. 根据权利要求39所述的方法,其特征在于,所述第三特征提取单元还包括门控单元;The method according to claim 39, wherein the third feature extraction unit further comprises a gating unit;
    对于所述第k+1个第三特征提取单元中的门控单元,所述门控单元用于对所述第k+1个第三特征提取单元中的最后一个残差块输出的第二残差信息、与所述L-1个残差块中至少一个残差块输出的第二特征信息进行级联后的特征信息进行去冗余,输出去冗余后的特征信息;For the gating unit in the k+1th third feature extraction unit, the gating unit is used for the second output of the last residual block in the k+1th third feature extraction unit performing de-redundancy on the residual information and the feature information concatenated with the second feature information output by at least one of the L-1 residual blocks, and outputting the de-redundant feature information;
    所述点云块的第k+1个增强上采样特征信息是根据去冗余后的特征信息与所述第k个增强上采样特征信息进行相加后确定的。The k+1th enhanced upsampling feature information of the point cloud block is determined after adding the deredundant feature information to the kth enhanced upsampling feature information.
  41. 根据权利要求34所述的方法,其特征在于,所述特征上采样模块还包括第一自相关注意力网络;The method according to claim 34, wherein the feature upsampling module further comprises a first autocorrelation attention network;
    所述第一自相关注意力网络用于对所述特征上采样子模块输出的所述点云块的上采样特征信息进行特征交互,输出特征交互后的所述点云块的上采样特征信息至所述特征提取子模块;The first autocorrelation attention network is used to perform feature interaction on the upsampling feature information of the point cloud block output by the feature upsampling submodule, and output the upsampling feature information of the point cloud block after feature interaction to the feature extraction submodule;
    所述特征提取子模块用于根据特征交互后的所述点云块的上采样特征信息,输出所述点云块的第二特征信息。The feature extraction submodule is configured to output second feature information of the point cloud block according to the upsampled feature information of the point cloud block after feature interaction.
  42. 根据权利要求41所述的方法,其特征在于,所述特征交互后的所述点云块的上采样特征信息的特征维度低于所述点云块的上采样特征信息的特征维度。The method according to claim 41, wherein the feature dimension of the upsampled feature information of the point cloud block after the feature interaction is lower than the feature dimension of the upsampled feature information of the point cloud block.
  43. 根据权利要求25所述的方法,其特征在于,所述几何生成模块包括多个全连接层;The method according to claim 25, wherein the geometry generation module comprises a plurality of fully connected layers;
    所述多个全连接层用于根据所述点云块的第二特征信息,输出所述点云块的上采样几何信息。The multiple fully connected layers are used to output upsampled geometric information of the point cloud block according to the second feature information of the point cloud block.
  44. 根据权利要求25所述的方法,其特征在于,所述几何生成模块包括:几何重建单元、滤波单元和下采样单元;The method according to claim 25, wherein the geometry generation module comprises: a geometry reconstruction unit, a filtering unit, and a downsampling unit;
    所述几何重建单元用于对所述将所述点云块的第二特征信息进行几何重建,输出所述点云块的初始上采样几何信息至所述滤波单元;The geometric reconstruction unit is used to geometrically reconstruct the second feature information of the point cloud block, and output the initial upsampling geometric information of the point cloud block to the filtering unit;
    所述滤波单元用于对所述点云块的初始上采样几何信息进行除噪,输出所述点云块滤除噪点的初始上采样几何信息至所述下采样单元;The filtering unit is used to denoise the initial upsampling geometric information of the point cloud block, and output the initial upsampling geometric information of the point cloud block to filter noise to the downsampling unit;
    所述下采样单元用于对所述点云块滤除噪点的初始上采样几何信息下采样至目标上采样率,输出所述点云块的上采样几何信息。The down-sampling unit is configured to down-sample the initial up-sampled geometric information of the point cloud block after filtering noise to a target up-sampling rate, and output the up-sampled geometric information of the point cloud block.
  45. 根据权利要求44所述的方法,其特征在于,所述目标上采样率小于或等于所述特征上采样模块的上采样率。The method according to claim 44, wherein the target upsampling rate is less than or equal to the upsampling rate of the feature upsampling module.
  46. 一种点云上采样装置,其特征在于,包括:A point cloud upsampling device is characterized in that, comprising:
    获取单元,用于获取待上采样点云的几何信息;An acquisition unit, configured to acquire geometric information of the point cloud to be upsampled;
    划分单元,用于根据所述待上采样点云的几何信息,将所述待上采样点云划分成至少一个点云块;A division unit, configured to divide the point cloud to be upsampled into at least one point cloud block according to the geometric information of the point cloud to be upsampled;
    上采样单元,用于将所述点云块的几何信息输入生成器中进行上采样,得到所述点云块的上采样几何信息;An upsampling unit, configured to input the geometric information of the point cloud block into the generator for upsampling, to obtain the upsampled geometric information of the point cloud block;
    其中,所述生成器包括:特征提取模块、特征上采样模块和几何生成模块,所述特征提取模块用于提取所述点云块的第一特征信息,所述特征采样模块用于将所述点云块的第一特征信息上采样为第二特征信息,所述几何生成模块用于将所述点云块的第二特征信息映射至几何空间中,以得到所述点云块的上采样几何信息。Wherein, the generator includes: a feature extraction module, a feature upsampling module and a geometry generation module, the feature extraction module is used to extract the first feature information of the point cloud block, and the feature sampling module is used to extract the The first feature information of the point cloud block is up-sampled to the second feature information, and the geometry generation module is used to map the second feature information of the point cloud block into a geometric space, so as to obtain the up-sampling of the point cloud block geometric information.
  47. 一种点云上采样设备,其特征在于,包括:处理器和存储器;A point cloud upsampling device, characterized in that it includes: a processor and a memory;
    所述存储器用于存储计算机程序;The memory is used to store computer programs;
    所述处理器用于调用并运行所述存储器中存储的计算机程序,以执行如权利要求25-45任一项所述的方法。The processor is used for invoking and running the computer program stored in the memory, so as to execute the method according to any one of claims 25-45.
  48. 一种模型训练方法,其特征在于,包括:A model training method, characterized in that, comprising:
    获取训练点云的几何信息,并根据所述训练点云的几何信息,将所述训练点云划分成至少一个训练点云块;Obtaining geometric information of the training point cloud, and dividing the training point cloud into at least one training point cloud block according to the geometric information of the training point cloud;
    将所述训练点云块的几何信息输入生成器的特征提取模块进行特征提取,得到所述训练点云块的第一特征信息;The feature extraction module of the geometric information input generator of described training point cloud block is carried out feature extraction, obtains the first feature information of described training point cloud block;
    将所述训练点云块的第一特征信息输入所述生成器的特征上采样模块进行上采样,得到所述训练点云块的第二特征信息;Inputting the first feature information of the training point cloud block into the feature upsampling module of the generator for upsampling to obtain the second feature information of the training point cloud block;
    将所述训练点云块的第二特征信息输入所述生成器的几何生成模块进行几何重建,得到所述训练点云块的预测上采样几何信息;Inputting the second feature information of the training point cloud block into the geometric generation module of the generator for geometric reconstruction, and obtaining the predicted upsampling geometric information of the training point cloud block;
    根据所述训练点云块的预测上采样几何信息,对所述生成器中的特征提取模块、特征上采样模块和几何生成模块进行训练,得到训练后的生成器。According to the predicted upsampling geometric information of the training point cloud block, the feature extraction module, feature upsampling module and geometry generation module in the generator are trained to obtain the trained generator.
  49. 根据权利要求48所述的方法,其特征在于,所述根据所述训练点云块的预测上采样几何信息,对所述生成器中的特征提取模块、特征上采样模块和几何生成模块进行训练,得到训练后的生成器,包括:The method according to claim 48, wherein the feature extraction module, feature upsampling module and geometry generation module in the generator are trained according to the predicted upsampling geometric information of the training point cloud block , to get the trained generator, including:
    将所述训练点云块的预测上采样几何信息输入判别器,得到所述判别器的第一判别结果,所述判别器用于判断输入所述判别器的数据是否为所述训练点云块的上采样真值;The predicted upsampling geometric information of the training point cloud block is input into the discriminator to obtain the first discrimination result of the discriminator, and the discriminator is used to judge whether the data input to the discriminator is the data of the training point cloud block upsampled true value;
    根据所述判别器的第一判别结果,对所述生成器中的特征提取模块、特征上采样模块和几何生成模块进行训练,得到训练后的生成器。According to the first discrimination result of the discriminator, the feature extraction module, feature upsampling module and geometry generation module in the generator are trained to obtain a trained generator.
  50. 根据权利要求48或49所述的方法,其特征在于,所述特征提取模块包括M个密集连接的特征提取块,所述将所述训练点云块的几何信息输入所述生成器的特征提取模块进行特征提取,得到所述训练点云块的第一特征信息,包括:The method according to claim 48 or 49, wherein the feature extraction module comprises M densely connected feature extraction blocks, and the geometric information of the training point cloud block is input into the feature extraction of the generator Module carries out feature extraction, obtains the first characteristic information of described training point cloud block, comprises:
    将所述训练点云块的几何信息输入所述特征提取模块中,获取所述M个特征提取块中第i个特 征提取块所提取的所述训练点云块的第i个第三特征信息,所述i为小于M的正整数;Input the geometric information of the training point cloud block into the feature extraction module, and obtain the ith third feature information of the training point cloud block extracted by the ith feature extraction block in the M feature extraction blocks , the i is a positive integer less than M;
    根据所述训练点云块的第i个第三特征信息,得到所述训练点云块的第i个第四特征信息;According to the i-th third feature information of the training point cloud block, the i-th fourth feature information of the training point cloud block is obtained;
    将所述训练点云块的第i个第四特征信息输入第i+1个特征提取块中,得到所述训练点云块的第i+1个第三特征信息;Inputting the ith fourth feature information of the training point cloud block into the i+1 feature extraction block to obtain the i+1 third feature information of the training point cloud block;
    将所述训练点云块的第M个特征提取块所提取的第M个第三特征信息,作为所述训练点云块的第一特征信息。The Mth third feature information extracted by the Mth feature extraction block of the training point cloud block is used as the first feature information of the training point cloud block.
  51. 根据权利要求50所述的方法,其特征在于,所述根据所述训练点云块的第i个第三特征信息,得到所述训练点云块的第i个第四特征信息,包括:The method according to claim 50, wherein the i-th fourth feature information of the training point cloud block is obtained according to the i-th third feature information of the training point cloud block, comprising:
    若i不等于1,则获取所述M个特征提取块中位于所述第i个特征提取块之前的各特征提取块所提取的第三特征信息;并将位于所述第i个特征提取块之前的各特征提取块所提取的第三特征信息、与所述第i个特征提取块所提取的第三特征信息进行级联,作为所述训练点云块的第i个第四特征信息;If i is not equal to 1, then obtain the third feature information extracted by each feature extraction block before the ith feature extraction block in the M feature extraction blocks; and place it in the ith feature extraction block The third feature information extracted by each previous feature extraction block is concatenated with the third feature information extracted by the i-th feature extraction block, as the i-th fourth feature information of the training point cloud block;
    若i等于1,则所述M个特征提取单元中第一特征提取块所提取的第一个第三特征信息,作为所述训练点云块的第i个第四特征信息。If i is equal to 1, the first third feature information extracted by the first feature extraction block in the M feature extraction units is used as the ith fourth feature information of the training point cloud block.
  52. 根据权利要求50所述的方法,其特征在于,所述特征提取块包括:第一特征提取单元和串联连接的至少一个第二特征提取单元,所述将所述训练点云块的第i个第四特征信息输入第i+1个特征提取块中,得到所述训练点云块的第i+1个第三特征信息,包括:The method according to claim 50, wherein the feature extraction block comprises: a first feature extraction unit and at least one second feature extraction unit connected in series, and the i-th feature extraction unit of the training point cloud block is The fourth feature information is input in the i+1 feature extraction block, and the i+1 third feature information of the training point cloud block is obtained, including:
    将所述训练点云块的第i个第四特征信息输入所述第i+1个特征提取块中的第一特征提取单元,以使所述第一特征提取单元针对所述训练点云块中的当前点,搜索所述当前点的K个邻近点,并基于所述第i个第四特征信息,将所述当前点的第四特征信息与所述邻近点的第四特征信息进行相减,得到K个残差特征信息;将所述K个残差特征信息与所述当前点的第四特征信息进行级联,得到所述当前点的第i个级联特征信息,并根据所述当前点的第i个级联特征信息,得到所述训练点云块的第i个级联特征信息;Input the i-th fourth feature information of the training point cloud block into the first feature extraction unit in the i+1 feature extraction block, so that the first feature extraction unit is specific to the training point cloud block For the current point in , search K neighboring points of the current point, and based on the ith fourth characteristic information, compare the fourth characteristic information of the current point with the fourth characteristic information of the neighboring points subtraction to obtain K residual feature information; concatenate the K residual feature information with the fourth feature information of the current point to obtain the ith concatenated feature information of the current point, and according to the The i-th cascade feature information of the current point is obtained to obtain the i-th cascade feature information of the training point cloud block;
    将所述训练点云块的第i个级联特征信息输入所述第i+1个特征提取块中的第一个第二特征提取单元,得到第一个第一个第五特征信息,并将所述第一个第一个第五特征信息输入所述第i+1个特征提取块中的第二个第二特征提取单元中,得到第二个第五特征信息;Inputting the i-th cascaded feature information of the training point cloud block into the first second feature extraction unit in the i+1th feature extraction block to obtain the first, first, and fifth feature information, and Inputting the first and fifth feature information into the second second feature extraction unit in the i+1th feature extraction block to obtain the second fifth feature information;
    将所述第i+1个特征提取块中最后一个第二特征提取单元提取的第五特征信息,作为所述训练点云块的第i+1个第三特征信息。The fifth feature information extracted by the last second feature extraction unit in the i+1th feature extraction block is used as the i+1th third feature information of the training point cloud block.
  53. 根据权利要求52所述的方法,其特征在于,所述第二特征提取单元包括P个残差块,所述P为正整数,所述将所述训练点云块的第i个级联特征信息输入所述第i+1个特征提取块中的第一个第二特征提取单元,得到第一个第一个第五特征信息,包括:The method according to claim 52, wherein the second feature extraction unit includes P residual blocks, the P is a positive integer, and the i-th concatenated feature of the training point cloud block is The information is input into the first second feature extraction unit in the i+1th feature extraction block to obtain the first first fifth feature information, including:
    将所述第i个级联特征信息输入所述第i+1个特征提取块中的第一个第二特征提取单元,获得所述第一个第二特征提取单元中第j个残差块输出的第一残差信息,所述j为小于或等于P的正整数;Inputting the i-th cascaded feature information into the first second feature extraction unit in the i+1th feature extraction block to obtain the j-th residual block in the first second feature extraction unit The first residual information output, the j is a positive integer less than or equal to P;
    将所述第j个残差块输出的第一残差信息和所述第i个级联特征信息输入所述第一个第二特征提取单元中的第j+1个残差块中,得到所述第j+1个残差块输出的第一残差信息;inputting the first residual information output by the jth residual block and the ith concatenated feature information into the j+1th residual block in the first second feature extraction unit, to obtain The first residual information output by the j+1th residual block;
    根据所述第一个第二特征提取单元中的P个残差块中至少一个残差块输出的第一残差信息,以及所述第i个级联特征信息,确定所述第一个第二特征提取单元输出的第五特征信息。According to the first residual information output by at least one of the P residual blocks in the first second feature extraction unit, and the i-th concatenated feature information, determine the first and second The fifth feature information output by the feature extraction unit.
  54. 根据权利要求53所述的方法,其特征在于,所述将所述第j个残差块输出的第一残差信息和所述第i个级联特征信息输入所述第一个第二特征提取单元中的第j+1个残差块中,得到所述第j+1个残差块输出的第一残差信息,包括:The method according to claim 53, wherein the first residual information output by the jth residual block and the ith concatenated feature information are input into the first second feature In the j+1th residual block in the extraction unit, the first residual information output by the j+1th residual block is obtained, including:
    对所述第j个残差块输出的第一残差信息和所述第i个级联特征信息进行相加,并将相加后的特征信息输入所述第j+1个残差块中,得到所述第j+1个残差块输出的第一残差信息。Adding the first residual information output by the jth residual block and the ith concatenated feature information, and inputting the added feature information into the j+1th residual block , to obtain the first residual information output by the j+1th residual block.
  55. 根据权利要求53所述的方法,其特征在于,所述根据所述第一个第二特征提取单元中的P个残差块中至少一个残差块输出的第一残差信息,以及所述第i个级联特征信息,确定所述第一个第二特征提取单元输出的第五特征信息,包括:The method according to claim 53, characterized in that the first residual information output from at least one residual block in the P residual blocks in the first second feature extraction unit, and the The i-th cascaded feature information is used to determine the fifth feature information output by the first second feature extraction unit, including:
    将所述P个残差块中最后一个残差块输出的第一残差信息、与P-1个残差块中至少一个残差块输出的第一残差信息进行级联,其中所述P-1个残差块为所述P个残差块中除所述最后一个残差块之外 的残差块;Concatenating the first residual information output by the last residual block in the P residual blocks with the first residual information output by at least one residual block in the P-1 residual blocks, wherein the The P-1 residual blocks are residual blocks other than the last residual block among the P residual blocks;
    根据级联后的特征信息和所述第i个级联特征信息,确定所述第一个第二特征提取单元输出的第五特征信息。The fifth feature information output by the first second feature extraction unit is determined according to the concatenated feature information and the i-th concatenated feature information.
  56. 根据权利要求55所述的方法,其特征在于,所述根据级联后的特征信息和所述第i个级联特征信息,确定所述第一个第二特征提取单元输出的第五特征信息,包括:The method according to claim 55, characterized in that, according to the concatenated feature information and the ith concatenated feature information, the fifth feature information output by the first second feature extraction unit is determined ,include:
    将所述级联后的特征信息和所述第i个级联特征信息进行相加,作为所述第一个第二特征提取单元输出的第五特征信息。Adding the concatenated feature information to the i-th concatenated feature information is used as fifth feature information output by the first and second feature extraction units.
  57. 根据权利要求55所述的方法,其特征在于,所述第二特征提取单元还包括门控单元,所述根据级联后的特征信息和所述第i个级联特征信息,确定所述第一个第二特征提取单元输出的第五特征信息,包括:The method according to claim 55, wherein the second feature extraction unit further comprises a gating unit, and according to the concatenated feature information and the i-th concatenated feature information, the i-th concatenated feature information is used to determine the The fifth feature information output by a second feature extraction unit includes:
    将所述级联后的特征信息输入所述门控单元进行去冗余,得到去冗余后的特征信息;Inputting the cascaded feature information into the gating unit for de-redundancy, to obtain de-redundant feature information;
    将去冗余后的特征信息与所述第i个级联特征信息进行相加,作为所述第一个第二特征提取单元输出的第五特征信息。Adding the feature information after de-redundancy to the ith concatenated feature information is used as the fifth feature information output by the first second feature extraction unit.
  58. 根据权利要求48或49任一项所述的方法,其特征在于,所述特征上采样模块包括:特征上采样子模块和特征提取子模块,所述将所述训练点云块的第一特征信息输入所述生成器的特征上采样模块进行上采样,得到所述训练点云块的第二特征信息,包括:The method according to any one of claims 48 or 49, wherein the feature upsampling module comprises: a feature upsampling submodule and a feature extraction submodule, the first feature of the training point cloud block The feature upsampling module of information input described generator carries out upsampling, obtains the second feature information of described training point cloud block, comprises:
    将所述训练点云块的第一特征信息输入所述特征上采样子模块,以使所述特征上采样子模块按照预设的上采样率r,将所述训练点云块的第一特征信息复制r份,并对复制后的第一特征信息在特征维度上增加一个n维向量,得到所述训练点云块的上采样特征信息,其中不同第一特征信息对应的n维向量的值不相同;Input the first feature information of the training point cloud block into the feature upsampling submodule, so that the feature upsampling submodule uses the first feature of the training point cloud block according to the preset upsampling rate r Copy r copies of the information, and add an n-dimensional vector on the feature dimension to the first feature information after copying, so as to obtain the up-sampling feature information of the training point cloud block, wherein the values of the n-dimensional vectors corresponding to different first feature information Are not the same;
    将所述训练点云块的上采样特征信息输入所述特征提取子模块,得到所述特征提取子模块提取的所述训练点云块的第二特征信息。Inputting the upsampled feature information of the training point cloud block into the feature extraction submodule to obtain the second feature information of the training point cloud block extracted by the feature extraction submodule.
  59. 根据权利要求58所述的方法,其特征在于,所述特征提取子模块包括串联连接的Q个第三特征提取单元,所述Q为正整数,所述将所述训练点云块的上采样特征信息输入所述特征提取子模块,得到所述特征提取子模块提取的所述训练点云块的第二特征信息,包括:The method according to claim 58, wherein the feature extraction submodule comprises Q third feature extraction units connected in series, the Q is a positive integer, and the upsampling of the training point cloud block Feature information is input into the feature extraction submodule to obtain the second feature information of the training point cloud block extracted by the feature extraction submodule, including:
    将所述训练点云块的上采样特征信息输入所述特征提取子模块,获得第k个第三特征提取单元所提取的所述训练点云块的第k个增强上采样特征信息;Input the upsampling feature information of the training point cloud block into the feature extraction submodule, and obtain the kth enhanced upsampling feature information of the training point cloud block extracted by the kth third feature extraction unit;
    将所述训练点云块的第k个增强上采样特征信息输入第k+1个第三特征提取单元,得到所述第k+1个第三特征提取单元提取的所述训练点云块的第k+1个增强上采样特征信息;Input the kth enhanced upsampling feature information of the training point cloud block into the k+1 third feature extraction unit to obtain the training point cloud block extracted by the k+1 third feature extraction unit The k+1 enhanced upsampling feature information;
    将所述Q个第三特征提取单元中最后一个第三特征提取单元所提取的所述训练点云块的第Q个增强上采样特征信息,作为所述训练点云块的第二特征信息。The Qth enhanced upsampling feature information of the training point cloud block extracted by the last third feature extraction unit of the Q third feature extraction units is used as the second feature information of the training point cloud block.
  60. 根据权利要求59所述的方法,其特征在于,所述第三特征提取单元包括L个残差块,所述L为正整数,所述将所述训练点云块的第k个增强上采样特征信息输入第k+1个第三特征提取单元,得到所述第k+1个第三特征提取单元提取的所述训练点云块的第k+1个增强上采样特征信息,包括:The method according to claim 59, wherein the third feature extraction unit includes L residual blocks, the L is a positive integer, and the kth enhanced upsampling of the training point cloud block The feature information is input into the k+1th third feature extraction unit to obtain the k+1th enhanced upsampling feature information of the training point cloud block extracted by the k+1th third feature extraction unit, including:
    将所述训练点云块的第k个增强上采样特征信息输入所述第k+1个第三特征提取单元,获得所述第k+1个第三特征提取单元中第l个残差块输出的第二残差信息,所述l为小于或等于L的正整数;Inputting the kth enhanced upsampling feature information of the training point cloud block into the k+1th third feature extraction unit to obtain the lth residual block in the k+1th third feature extraction unit The second residual information output, the l is a positive integer less than or equal to L;
    将所述第l个残差块输出的第二残差信息和所述第k个增强上采样特征信息输入第l+1个残差块中,得到所述第l+1个残差块输出的第二残差信息;inputting the second residual information output by the lth residual block and the kth enhanced upsampling feature information into the l+1th residual block to obtain the output of the l+1th residual block The second residual information of ;
    根据所述L个残差块中至少一个残差块输出的第二残差信息,以及所述第k个增强上采样特征信息,得到所述训练点云块的第k+1个增强上采样特征信息。According to the second residual information output by at least one residual block in the L residual blocks, and the kth enhanced upsampling feature information, obtain the k+1th enhanced upsampling of the training point cloud block characteristic information.
  61. 根据权利要求60所述的方法,其特征在于,所述将所述第l个残差块输出的第二残差信息和所述第k个增强上采样特征信息输入第l+1个残差块中,得到所述第l+1个残差块输出的第二残差信息,包括:The method according to claim 60, wherein the second residual information output by the lth residual block and the kth enhanced upsampling feature information are input into the l+1th residual In the block, the second residual information output by the l+1 residual block is obtained, including:
    对所述第l个残差块输出的第二残差信息和所述第k个增强上采样特征信息进行相加,并将相加后的特征信息输入所述第l+1个残差块中,确定所述第l+1个残差块输出的第二残差信息。adding the second residual information output by the lth residual block to the kth enhanced upsampling feature information, and inputting the added feature information into the l+1th residual block , determine the second residual information output by the l+1th residual block.
  62. 根据权利要求60所述的方法,其特征在于,所述根据所述L个残差块中至少一个残差块输出的第二残差信息,以及所述第k个增强上采样特征信息,确定所述训练点云块的第k+1个增强上采 样特征信息,包括:The method according to claim 60, wherein, according to the second residual information output by at least one residual block in the L residual blocks, and the kth enhanced upsampling feature information, determine The k+1 enhanced upsampling feature information of the training point cloud block includes:
    将所述L个残差块中最后一个残差块输出的第二残差信息、与L-1个残差块中至少一个残差块输出的第二残差信息进行级联,其中所述L-1个残差块为所述L个残差块中除所述最后一个残差块之外的残差块;Concatenate the second residual information output by the last residual block in the L residual blocks with the second residual information output by at least one residual block in the L-1 residual blocks, wherein the The L-1 residual blocks are residual blocks other than the last residual block among the L residual blocks;
    根据级联后的特征信息和所述第k个增强上采样特征信息,确定所述训练点云块的第k+1个增强上采样特征信息。According to the concatenated feature information and the kth enhanced upsampled feature information, determine the k+1th enhanced upsampled feature information of the training point cloud block.
  63. 根据权利要求62所述的方法,其特征在于,所述根据级联后的特征信息和所述第k个增强上采样特征信息,确定所述训练点云块的第k+1个增强上采样特征信息,包括:The method according to claim 62, wherein the k+1th enhanced upsampling of the training point cloud block is determined according to the concatenated feature information and the kth enhanced upsampled feature information Characteristic information, including:
    将所述级联后的特征信息和所述第k个增强上采样特征信息进行相加,作为所述训练点云块的第k+1个增强上采样特征信息。Adding the concatenated feature information and the kth enhanced upsampled feature information as the k+1th enhanced upsampled feature information of the training point cloud block.
  64. 根据权利要求62所述的方法,其特征在于,所述第三特征提取单元还包括门控单元,所述根据级联后的特征信息和所述第k个增强上采样特征信息,确定所述训练点云块的第k+1个增强上采样特征信息,包括:The method according to claim 62, wherein the third feature extraction unit further includes a gating unit, and according to the cascaded feature information and the kth enhanced upsampling feature information, determine the The k+1th enhanced upsampling feature information of the training point cloud block, including:
    将所述级联后的特征信息输入所述门控单元进行去冗余,得到去冗余后的特征信息;Inputting the cascaded feature information into the gating unit for de-redundancy, to obtain de-redundant feature information;
    将去冗余后的特征信息与所述第k个增强上采样特征信息进行相加,作为所述训练点云块的第k+1个增强上采样特征信息。The feature information after de-redundancy is added to the kth enhanced upsampled feature information, and used as the k+1th enhanced upsampled feature information of the training point cloud block.
  65. 根据权利要求58所述的方法,其特征在于,所述特征上采样模块还包括第一自相关注意力网络,所述将所述训练点云块的上采样特征信息输入所述特征提取子模块,得到所述特征提取单元提取的所述训练点云块的第二特征信息,包括:The method according to claim 58, wherein the feature upsampling module further comprises a first autocorrelation attention network, and the upsampling feature information of the training point cloud block is input into the feature extraction submodule , obtaining the second feature information of the training point cloud block extracted by the feature extraction unit, including:
    将所述训练点云块的上采样特征信息输入所述第一自相关注意力网络进行特征交互,得到特征交互后的所述训练点云块的上采样特征信息;The upsampling feature information of the training point cloud block is input into the first autocorrelation attention network to perform feature interaction, and the upsampling feature information of the training point cloud block after feature interaction is obtained;
    将特征交互后的所述训练点云块的上采样特征信息输入所述特征提取子模块进行特征提取,得到所述训练点云块的第二特征信息。Inputting the upsampled feature information of the training point cloud block after the feature interaction into the feature extraction sub-module for feature extraction to obtain second feature information of the training point cloud block.
  66. 根据权利要求65所述的方法,其特征在于,所述特征交互后的所述训练点云块的上采样特征信息的特征维度低于所述训练点云块的上采样特征信息的特征维度。The method according to claim 65, wherein the feature dimension of the upsampled feature information of the training point cloud block after the feature interaction is lower than the feature dimension of the upsampled feature information of the training point cloud block.
  67. 根据权利要求48或49所述的方法,其特征在于,所述几何生成模块包括多个全连接层,所述将所述训练点云块的第二特征信息输入所述生成器的几何生成模块进行几何重建,得到所述训练点云块的预测上采样几何信息,包括:The method according to claim 48 or 49, wherein the geometry generation module comprises a plurality of fully connected layers, and the second feature information of the training point cloud block is input into the geometry generation module of the generator Perform geometric reconstruction to obtain the predicted upsampling geometric information of the training point cloud block, including:
    将所述训练点云块的第二特征信息输入所述多个全连接层,得到所述训练点云块的预测上采样几何信息。Inputting the second feature information of the training point cloud block into the plurality of fully connected layers to obtain the predicted upsampling geometric information of the training point cloud block.
  68. 根据权利要求48或49所述的方法,其特征在于,所述几何生成模块包括:几何重建单元、滤波单元和下采样单元,所述将所述训练点云块的第二特征信息输入所述生成器的几何生成模块,得到所述训练点云块的预测上采样几何信息,包括:The method according to claim 48 or 49, wherein the geometry generation module comprises: a geometry reconstruction unit, a filter unit and a downsampling unit, and the second feature information of the training point cloud block is input into the The geometric generation module of the generator obtains the predicted upsampling geometric information of the training point cloud block, including:
    所述将所述训练点云块的第二特征信息输入所述几何重建单元进行几何重建,得到所述训练点云块的初始上采样几何信息;The second feature information of the training point cloud block is input into the geometric reconstruction unit to perform geometric reconstruction, and the initial upsampling geometric information of the training point cloud block is obtained;
    将所述训练点云块的初始上采样几何信息输入滤波单元进行除噪,得到所述训练点云块滤除噪点的初始上采样几何信息;The initial upsampling geometric information of the training point cloud block is input into the filtering unit for denoising, and the initial upsampling geometric information of the training point cloud block for filtering noise is obtained;
    将所述训练点云块滤除噪点的初始上采样几何信息输入所述下采样单元中进行下采样,得到所述训练点云块的预测上采样几何信息。Inputting the initial upsampling geometric information of the training point cloud block to filter out noises into the downsampling unit for downsampling to obtain the predicted upsampling geometric information of the training point cloud block.
  69. 根据权利要求68所述的方法,其特征在于,所述训练点云块的上采样几何信息对应的上采样率小于或等于所述特征上采样模块的上采样率。The method according to claim 68, wherein the upsampling rate corresponding to the upsampling geometric information of the training point cloud block is less than or equal to the upsampling rate of the feature upsampling module.
  70. 根据权利要求49所述的方法,其特征在于,所述判别器为预先训练好的判别器。The method according to claim 49, wherein the discriminator is a pre-trained discriminator.
  71. 根据权利要求49所述的方法,其特征在于,所述方法还包括:The method according to claim 49, further comprising:
    使用所述训练点云块的几何信息对所述判别器进行训练。The discriminator is trained using the geometric information of the training point cloud blocks.
  72. 根据权利要求71所述的方法,其特征在于,所述使用所述训练点云块的几何信息对所述判别器进行训练,包括:The method according to claim 71, wherein said using the geometric information of said training point cloud block to train said discriminator comprises:
    将所述生成器生成的所述训练点云块的预测上采样几何信息输入所述判别器中,得到所述判别器 的第二判别结果;The predicted upsampling geometry information of the training point cloud blocks generated by the generator is input in the discriminator to obtain the second discriminant result of the discriminator;
    将所述训练点云块的几何信息的上采样真值输入所述判别器,得到所述判别器的第三判别结果;Inputting the upsampling true value of the geometric information of the training point cloud block into the discriminator to obtain the third discriminant result of the discriminator;
    根据所述第二判别结果和第三判别结果,确定所述判别器的损失;determining a loss of the discriminator according to the second discrimination result and the third discrimination result;
    根据所述判别器的损失,对所述判别器进行训练。Based on the loss of the discriminator, the discriminator is trained.
  73. 根据权利要求72所述的方法,其特征在于,所述根据所述第二判别结果和第三判别结果,确定所述判别器的损失,包括:The method according to claim 72, wherein the determining the loss of the discriminator according to the second discrimination result and the third discrimination result comprises:
    根据所述第二判别结果和第三判别结果,采用最小二乘损失函数,确定所述判别器的损失。The loss of the discriminator is determined by using a least squares loss function according to the second discrimination result and the third discrimination result.
  74. 根据权利要求72所述的方法,其特征在于,所述判别器包括:全局判别模块、边界判别模块和全连接模块,所述方法还包括:The method according to claim 72, wherein the discriminator comprises: a global discriminant module, a boundary discriminant module and a fully connected module, and the method further comprises:
    获取目标点云块的边界点的几何信息;Obtain the geometric information of the boundary points of the target point cloud block;
    将所述目标点云块的边界点的几何信息输入所述边界判别模块,得到所述目标点云块的边界特征信息;Input the geometric information of the boundary points of the target point cloud block into the boundary discrimination module to obtain the boundary feature information of the target point cloud block;
    将所述目标点云块的几何信息输入所述全局判别模块,得到所述目标点云块的全局特征信息;Input the geometric information of the target point cloud block into the global discrimination module to obtain the global feature information of the target point cloud block;
    将所述目标点云块的全局特征信息和边界特征信息输入所述全连接模块,得到所述判别器的目标判别结果;Inputting the global feature information and boundary feature information of the target point cloud block into the fully connected module to obtain the target discrimination result of the discriminator;
    其中,若所述目标点云块为所述生成器上采样后的训练点云块,且所述判断器未经过所述训练点云块训练,则所述目标判别结果为所述第二判别结果;若所述目标点云块为所述训练点云块的上采样真值,则所述目标判别结果为所述第三判别结果;若所述目标点云块为所述生成器上采样后的训练点云块,且所述判断器经过所述训练点云块训练,则所述目标判别结果为所述第一判别结果。Wherein, if the target point cloud block is a training point cloud block sampled by the generator, and the judger has not been trained by the training point cloud block, then the target discrimination result is the second discrimination Result; if the target point cloud block is the upsampling true value of the training point cloud block, then the target discrimination result is the third discrimination result; if the target point cloud block is the generator upsampling After the training point cloud block, and the judge is trained by the training point cloud block, the target discrimination result is the first discrimination result.
  75. 根据权利要求74所述的方法,其特征在于,所述获取目标点云块的边界点的几何信息,包括:The method according to claim 74, wherein said acquiring the geometric information of the boundary points of the target point cloud block comprises:
    使用高通图滤波器提取所述目标点云块的边界点的几何信息。Using a high-pass image filter to extract the geometric information of the boundary points of the target point cloud block.
  76. 根据权利要求74所述的方法,其特征在于,所述将所述目标点云块的全局特征信息和边界特征信息输入所述全连接模块,得到所述判别器的目标判别结果,包括:The method according to claim 74, wherein the inputting the global feature information and boundary feature information of the target point cloud block into the fully connected module to obtain the target discrimination result of the discriminator includes:
    将所述目标点云块的全局特征信息和边界特征信息进行级联;Cascading the global feature information and boundary feature information of the target point cloud block;
    将级联后的全局特征信息和边界特征信息输入所述全连接模块,得到所述判别器的目标判别结果。The cascaded global feature information and boundary feature information are input into the fully connected module to obtain the target discrimination result of the discriminator.
  77. 根据权利要求74所述的方法,其特征在于,所述全局判别模块沿着网络深度方向依次包括:第一数量个多层感知机、第一最大池化层、第二自相关注意力网络、第二数量个多层感知机和第二最大池化层;所述将所述目标点云块的几何信息输入所述全局判别模块,得到所述目标点云块的全局特征信息,包括:The method according to claim 74, wherein the global discrimination module comprises sequentially along the network depth direction: a first number of multi-layer perceptrons, a first maximum pooling layer, a second autocorrelation attention network, The second number of multi-layer perceptrons and the second maximum pooling layer; the geometric information of the target point cloud block is input into the global discrimination module to obtain the global feature information of the target point cloud block, including:
    将所述目标点云块的几何信息输入所述第一数量个多层感知机进行特征提取,得到所述目标点云块的第一全局特征信息;Inputting the geometric information of the target point cloud block into the first number of multi-layer perceptrons for feature extraction to obtain the first global feature information of the target point cloud block;
    将所述第一全局特征信息输入所述第一最大池化层进行降维处理,得到所述目标点云块的第二全局特征信息;Inputting the first global feature information into the first maximum pooling layer for dimensionality reduction processing to obtain second global feature information of the target point cloud block;
    将所述第一全局特征信息和所述第二全局特征信息输入所述第二自相关注意力网络进行特征交互,得到所述目标点云块的第三全局特征信息;Inputting the first global feature information and the second global feature information into the second autocorrelation attention network for feature interaction to obtain the third global feature information of the target point cloud block;
    将所述第三全局特征信息输入所述第二数量个多层感知机进而特征提取,得到所述目标点云块的第四全局特征信息;Inputting the third global feature information into the second number of multi-layer perceptrons for feature extraction to obtain the fourth global feature information of the target point cloud block;
    将所述第四全局特征信息输入所述第二最大池化层进行降维处理,得到所述目标点云块的全局特征信息。Inputting the fourth global feature information into the second maximum pooling layer to perform dimensionality reduction processing to obtain the global feature information of the target point cloud block.
  78. 根据权利要求77所述的方法,其特征在于,所述将所述第一全局特征信息和所述第二全局特征信息输入所述第二自相关注意力网络进行特征交互,得到所述目标点云块的第三全局特征信息,包括:The method according to claim 77, wherein the first global feature information and the second global feature information are input into the second autocorrelation attention network for feature interaction to obtain the target point The third global feature information of the cloud block, including:
    将所述第一全局特征信息和所述第二全局特征信息进行级联;cascading the first global feature information and the second global feature information;
    将级联后的所述第一全局特征信息和所述第二全局特征信息,输入所述第二自相关注意力网络进行特征交互,得到所述目标点云块的第三全局特征信息。The concatenated first global feature information and the second global feature information are input into the second autocorrelation attention network for feature interaction to obtain third global feature information of the target point cloud block.
  79. 根据权利要求77所述的方法,其特征在于,所述第一数量等于所述第二数量。The method of claim 77, wherein said first amount is equal to said second amount.
  80. 根据权利要求79所述的方法,其特征在于,所述第一数量与所述第二数量均等于2。The method of claim 79, wherein the first number and the second number are both equal to two.
  81. 根据权利要求80所述的方法,其特征在于,所述第一数量个多层感知机包括第一层多层感知机和第二层多层感知机,所述第二数量个多层感知机包括第三层多层感知机和第四层多层感知机,所述第一层多层感知机、所述第二层多层感知机、所述第三层多层感知机和所述第四层多层感知机的特征维度依次逐渐增加。The method of claim 80, wherein the first number of multilayer perceptrons includes a first layer of multilayer perceptrons and a second layer of multilayer perceptrons, and the second number of multilayer perceptrons Including the third layer multi-layer perceptron and the fourth layer multi-layer perceptron, the first layer multi-layer perceptron, the second layer multi-layer perceptron, the third layer multi-layer perceptron and the fourth layer multi-layer perceptron The feature dimensions of the four-layer multilayer perceptron gradually increase sequentially.
  82. 根据权利要求81所述的方法,其特征在于,所述第一层多层感知机的特征维度为32,所述第二层多层感知机的特征维度为64,所述第三层多层感知机的特征维度为128,所述第四层多层感知机的特征维度为256。The method according to claim 81, wherein the feature dimension of the first layer of multi-layer perceptron is 32, the feature dimension of the second layer of multi-layer perceptron is 64, and the feature dimension of the third layer of multi-layer perceptron is The feature dimension of the perceptron is 128, and the feature dimension of the fourth layer multi-layer perceptron is 256.
  83. 根据权利要求74所述的方法,其特征在于,所述边界判别模块沿着网络深度方向依次包括:第三数量个多层感知机、第三最大池化层、第三自相关注意力网络、第四数量个多层感知机和第四最大池化层;所述将所述目标点云块的边界点的几何信息输入所述边界判别模块,得到所述目标点云块的边界特征信息,包括:The method according to claim 74, wherein the boundary discrimination module comprises in sequence along the network depth direction: a third number of multi-layer perceptrons, a third maximum pooling layer, a third autocorrelation attention network, The fourth number of multilayer perceptrons and the fourth maximum pooling layer; the geometric information of the boundary points of the target point cloud block is input into the boundary discrimination module to obtain the boundary feature information of the target point cloud block, include:
    将所述目标点云块的边界点的几何信息输入所述第三数量个多层感知机中进行特征提取,得到所述目标点云块的第一边界特征信息;Input the geometric information of the boundary points of the target point cloud block into the third number of multi-layer perceptrons for feature extraction, and obtain the first boundary feature information of the target point cloud block;
    将所述第一边界特征信息输入所述第三最大池化层进行降维处理,得到所述目标点云块的第二边界特征信息;Inputting the first boundary feature information into the third maximum pooling layer for dimensionality reduction processing to obtain second boundary feature information of the target point cloud block;
    将所述第一边界特征信息和所述第二边界特征信息输入所述第三自相关注意力网络进行特征交互,得到所述目标点云块的第三边界特征信息;Inputting the first boundary feature information and the second boundary feature information into the third autocorrelation attention network for feature interaction to obtain the third boundary feature information of the target point cloud block;
    将所述第三边界特征信息输入所述第四数量个多层感知机进行特征提取,得到所述目标点云块的第四边界特征信息;Inputting the third boundary feature information into the fourth number of multi-layer perceptrons for feature extraction to obtain the fourth boundary feature information of the target point cloud block;
    将所述第四边界特征信息输入所述第四最大池化层进行降维处理,得到所述目标点云块的边界特征信息。Inputting the fourth boundary feature information into the fourth maximum pooling layer for dimensionality reduction processing to obtain boundary feature information of the target point cloud block.
  84. 根据权利要求83所述的方法,其特征在于,所述将所述第一边界特征信息和所述第二边界特征信息输入所述第三自相关注意力网络进行特征交互,得到所述目标点云块的第三边界特征信息,包括:The method according to claim 83, wherein the first boundary feature information and the second boundary feature information are input into the third autocorrelation attention network for feature interaction to obtain the target point The third boundary feature information of the cloud block, including:
    将所述第一边界特征信息和所述第二边界特征信息进行级联;cascading the first boundary feature information and the second boundary feature information;
    将级联后的所述第一边界特征信息和所述第二边界特征信息,输入所述第三自相关注意力网络进行特征交互,得到所述目标点云块的第三边界特征信息。The concatenated first boundary feature information and the second boundary feature information are input into the third autocorrelation attention network for feature interaction to obtain third boundary feature information of the target point cloud block.
  85. 根据权利要求83所述的方法,其特征在于,所述第三数量等于所述第四数量。The method of claim 83, wherein said third amount is equal to said fourth amount.
  86. 根据权利要求85所述的方法,其特征在于,所述第三数量与所述第四数量均等于2。The method of claim 85, wherein the third number and the fourth number are both equal to two.
  87. 根据权利要求86所述的方法,其特征在于,所述第三数量个多层感知机包括第五层多层感知机和第六层多层感知机,所述第四数量个多层感知机包括第七层多层感知机和第八层多层感知机,所述第五层多层感知机、所述第六层多层感知机、所述第七层多层感知机和所述第八层多层感知机的特征维度依次逐渐增加。The method of claim 86, wherein the third number of multilayer perceptrons includes a fifth layer of multilayer perceptrons and a sixth layer of multilayer perceptrons, and the fourth number of multilayer perceptrons Including the seventh layer multi-layer perceptron and the eighth layer multi-layer perceptron, the fifth layer multi-layer perceptron, the sixth layer multi-layer perceptron, the seventh layer multi-layer perceptron and the first layer The feature dimensions of the eight-layer multilayer perceptron gradually increase sequentially.
  88. 根据权利要求87所述的方法,其特征在于,所述第八层多层感知机的特征维度大于或等于所述第七层多层感知机的特征维度,且小于或等于第四层多层感知机的特征维度。The method according to claim 87, wherein the feature dimension of the eighth-layer multi-layer perceptron is greater than or equal to the feature dimension of the seventh-layer multi-layer perceptron, and is less than or equal to the fourth-layer multi-layer The feature dimension of the perceptron.
  89. 根据权利要求88所述的方法,其特征在于,所述第五层多层感知机的特征维度为32,所述第六层多层感知机的特征维度为64,所述第七层多层感知机的特征维度为128,所述第八层多层感知机的特征维度为192。The method according to claim 88, wherein the feature dimension of the fifth layer multi-layer perceptron is 32, the feature dimension of the sixth layer multi-layer perceptron is 64, and the seventh layer multi-layer perceptron The feature dimension of the perceptron is 128, and the feature dimension of the eighth layer multi-layer perceptron is 192.
  90. 根据权利要求49所述的方法,其特征在于,所述根据所述判别器的第一判别结果,对所述生成器中的特征提取模块、特征上采样模块和几何生成模块进行训练,得到训练后的生成器,包括:The method according to claim 49, characterized in that, according to the first discrimination result of the discriminator, the feature extraction module, feature upsampling module and geometry generation module in the generator are trained to obtain the training After the generator, including:
    根据所述第一判别结果,确定所述生成器的第一损失;determining a first loss of the generator according to the first discrimination result;
    根据所述第一损失,确定所述生成器中的特征提取模块、特征上采样模块和几何生成模块的参数矩阵。According to the first loss, the parameter matrix of the feature extraction module, feature upsampling module and geometry generation module in the generator is determined.
  91. 根据权利要求90所述的方法,其特征在于,所述根据所述第一判别结果,确定所述生成器的第一损失,包括:The method according to claim 90, wherein the determining the first loss of the generator according to the first discrimination result comprises:
    根据所述第一判别结果,采用最小二乘损失函数,确定所述生成器的第一损失。According to the first discrimination result, a least squares loss function is used to determine a first loss of the generator.
  92. 根据权利要求90所述的方法,其特征在于,所述根据所述第一损失,确定所述生成器中的特征提取模块、特征上采样模块和几何生成模块的参数矩阵,包括:The method according to claim 90, wherein, according to the first loss, determining the parameter matrix of the feature extraction module, feature upsampling module and geometry generation module in the generator comprises:
    确定所述生成器的至少一个第二损失;determining at least one second loss for the generator;
    根据所述生成器的第一损失和所述生成器的至少一个第二损失,确定所述生成器的目标损失;determining a target loss for the generator based on a first loss for the generator and at least one second loss for the generator;
    根据所述生成器的目标损失,确定所述生成器中的特征提取模块、特征上采样模块和几何生成模块的参数矩阵。According to the target loss of the generator, the parameter matrix of the feature extraction module, feature upsampling module and geometry generation module in the generator is determined.
  93. 根据权利要求92所述的方法,其特征在于,所述确定所述生成器的至少一个第二损失,包括:The method of claim 92, wherein said determining at least one second loss of said generator comprises:
    根据所述训练点云块的上采样几何信息和所述训练点云块的几何信息的上采样真值,采用地动距离方式,确定所述生成器的一个第二损失。According to the upsampled geometric information of the training point cloud block and the upsampled true value of the geometric information of the training point cloud block, a second loss of the generator is determined by using ground motion distance.
  94. 根据权利要求92所述的方法,其特征在于,所述确定所述生成器的至少一个第二损失,包括:The method of claim 92, wherein said determining at least one second loss of said generator comprises:
    将所述训练点云块的上采样几何信息进行下采样,得到与所述训练点云块相同点数的下采样训练点云块;Downsampling the upsampling geometric information of the training point cloud block to obtain a downsampling training point cloud block with the same number of points as the training point cloud block;
    根据所述下采样训练点云块的几何信息和所述训练点云块的几何信息,采样地动距离方式,确定所述生成器的一个第二损失。According to the geometric information of the downsampled training point cloud block and the geometric information of the training point cloud block, a ground motion distance is sampled to determine a second loss of the generator.
  95. 根据权利要求94所述的方法,其特征在于,所述根据所述下采样训练点云块的几何信息和所述训练点云块的几何信息,采样地动距离方式,确定所述生成器的一个第二损失,包括:The method according to claim 94, characterized in that, according to the geometric information of the down-sampled training point cloud block and the geometric information of the training point cloud block, the method of sampling ground motion distance is used to determine the generator's A second loss, including:
    根据如下公式,确定所述生成器的一个第二损失:A second loss of the generator is determined according to the following formula:
    Figure PCTCN2021096287-appb-100001
    Figure PCTCN2021096287-appb-100001
    其中,所述L id为所述生成器的第二损失,所述P ori为训练点云块,所述P low为下采样后的训练点云块,φ:P low→P ori表示由P low和P ori构成的双射,有且只有唯一的一种移动方式让P low与P ori移动到彼此点集的距离最小,所述
    Figure PCTCN2021096287-appb-100002
    为所述P low中的第k个点,所述
    Figure PCTCN2021096287-appb-100003
    为所述
    Figure PCTCN2021096287-appb-100004
    在所述P ori中对应的点。
    Wherein, the L id is the second loss of the generator, the P ori is a training point cloud block, and the P low is a downsampled training point cloud block, and φ:P low →P ori means that P For the bijection composed of low and P ori , there is only one and only way to move P low and P ori to the minimum distance between the point sets of each other.
    Figure PCTCN2021096287-appb-100002
    is the kth point in the P low , the
    Figure PCTCN2021096287-appb-100003
    for the said
    Figure PCTCN2021096287-appb-100004
    Corresponding point in the P ori .
  96. 根据权利要求92所述的方法,其特征在于,所述确定所述生成器的至少一个第二损失,包括:The method of claim 92, wherein said determining at least one second loss of said generator comprises:
    根据均匀损失函数,确定所述生成器的至少一个第二损失。Based on a uniform loss function, at least one second loss of the generator is determined.
  97. 根据权利要求92所述的方法,其特征在于,根据所述生成器的第一损失和至少一个第二损失,确定所述生成器的目标损失,包括:The method according to claim 92, wherein determining the target loss of the generator based on the first loss of the generator and at least one second loss comprises:
    将所述生成器的第一损失和所述至少一个第二损失的加权平均值,确定所述生成器的目标损失。A weighted average of the generator's first loss and the at least one second loss is used to determine the generator's target loss.
  98. 一种生成器的训练装置,其特征在于,包括:A generator training device, characterized in that it comprises:
    获取单元,用于获取训练点云的几何信息;An acquisition unit is used to acquire the geometric information of the training point cloud;
    划分单元,用于根据所述训练点云的几何信息,将所述训练点云划分成至少一个训练点云块;A division unit, configured to divide the training point cloud into at least one training point cloud block according to the geometric information of the training point cloud;
    训练单元,用于将所述训练点云块的几何信息输入生成器的特征提取模块进行特征提取,得到所述训练点云块的第一特征信息;将所述训练点云块的第一特征信息输入所述生成器的特征上采样模块进行上采样,得到所述训练点云块的第二特征信息;将所述训练点云块的第二特征信息输入所述生成器的几何生成模块进行几何重建,得到所述训练点云块的预测上采样几何信息;根据所述训练点云块的预测上采样几何信息,对所述生成器中的特征提取模块、特征上采样模块和几何生成模块进行训练,得到训练后的生成器。The training unit is used to input the geometric information of the training point cloud block into the feature extraction module of the generator for feature extraction to obtain the first feature information of the training point cloud block; the first feature information of the training point cloud block Information is input into the feature upsampling module of the generator for upsampling to obtain the second feature information of the training point cloud block; the second feature information of the training point cloud block is input into the geometry generation module of the generator for Geometric reconstruction to obtain the predicted upsampling geometric information of the training point cloud block; according to the predicted upsampling geometric information of the training point cloud block, the feature extraction module, feature upsampling module and geometry generation module in the generator Perform training to obtain the trained generator.
  99. 一种生成器训练设备,其特征在于,包括:处理器和存储器;A generator training device, characterized in that it includes: a processor and a memory;
    所述存储器用于存储计算机程序;The memory is used to store computer programs;
    所述处理器用于调用并运行所述存储器中存储的计算机程序,以执行如权利要求48-97任一项所述的方法。The processor is used for invoking and running the computer program stored in the memory, so as to execute the method according to any one of claims 48-97.
  100. 一种计算机可读存储介质,其特征在于,用于存储计算机程序,所述计算机程序使得计算机执行如权利要求1至23或25至45或48至97任一项所述的方法。A computer-readable storage medium, characterized by being used to store a computer program, the computer program causes a computer to execute the method according to any one of claims 1 to 23 or 25 to 45 or 48 to 97.
PCT/CN2021/096287 2021-05-27 2021-05-27 Point cloud decoding and upsampling and model training methods and apparatus WO2022246724A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN202180096083.7A CN117242493A (en) 2021-05-27 2021-05-27 Point cloud decoding, upsampling and model training method and device
PCT/CN2021/096287 WO2022246724A1 (en) 2021-05-27 2021-05-27 Point cloud decoding and upsampling and model training methods and apparatus

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2021/096287 WO2022246724A1 (en) 2021-05-27 2021-05-27 Point cloud decoding and upsampling and model training methods and apparatus

Publications (1)

Publication Number Publication Date
WO2022246724A1 true WO2022246724A1 (en) 2022-12-01

Family

ID=84229438

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/096287 WO2022246724A1 (en) 2021-05-27 2021-05-27 Point cloud decoding and upsampling and model training methods and apparatus

Country Status (2)

Country Link
CN (1) CN117242493A (en)
WO (1) WO2022246724A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116664645A (en) * 2023-07-28 2023-08-29 之江实验室 Point cloud matching method and device, electronic device and storage medium
CN117456078A (en) * 2023-12-19 2024-01-26 北京渲光科技有限公司 Neural radiation field rendering method, system and equipment based on various sampling strategies

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2020190093A1 (en) * 2019-03-20 2020-09-24 엘지전자 주식회사 Point cloud data transmission device, point cloud data transmission method, point cloud data reception device, and point cloud data reception method
CN111862289A (en) * 2020-08-04 2020-10-30 天津大学 Point cloud up-sampling method based on GAN network
CN112218077A (en) * 2019-07-11 2021-01-12 腾讯美国有限责任公司 Method and device for encoding point cloud attributes between channels and readable storage medium
CN112241997A (en) * 2020-09-14 2021-01-19 西北大学 Three-dimensional model establishing and repairing method and system based on multi-scale point cloud up-sampling
CN112565734A (en) * 2020-12-03 2021-03-26 西安电子科技大学 Point cloud attribute coding and decoding method and device based on hybrid coding
US20210104013A1 (en) * 2019-10-03 2021-04-08 Lg Electronics Inc. Point cloud data transmission device, point cloud data transmission method, point cloud data reception device, and point cloud data reception method

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2020190093A1 (en) * 2019-03-20 2020-09-24 엘지전자 주식회사 Point cloud data transmission device, point cloud data transmission method, point cloud data reception device, and point cloud data reception method
CN112218077A (en) * 2019-07-11 2021-01-12 腾讯美国有限责任公司 Method and device for encoding point cloud attributes between channels and readable storage medium
US20210104013A1 (en) * 2019-10-03 2021-04-08 Lg Electronics Inc. Point cloud data transmission device, point cloud data transmission method, point cloud data reception device, and point cloud data reception method
CN111862289A (en) * 2020-08-04 2020-10-30 天津大学 Point cloud up-sampling method based on GAN network
CN112241997A (en) * 2020-09-14 2021-01-19 西北大学 Three-dimensional model establishing and repairing method and system based on multi-scale point cloud up-sampling
CN112565734A (en) * 2020-12-03 2021-03-26 西安电子科技大学 Point cloud attribute coding and decoding method and device based on hybrid coding

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116664645A (en) * 2023-07-28 2023-08-29 之江实验室 Point cloud matching method and device, electronic device and storage medium
CN116664645B (en) * 2023-07-28 2023-10-20 之江实验室 Point cloud matching method and device, electronic device and storage medium
CN117456078A (en) * 2023-12-19 2024-01-26 北京渲光科技有限公司 Neural radiation field rendering method, system and equipment based on various sampling strategies
CN117456078B (en) * 2023-12-19 2024-03-26 北京渲光科技有限公司 Neural radiation field rendering method, system and equipment based on various sampling strategies

Also Published As

Publication number Publication date
CN117242493A (en) 2023-12-15

Similar Documents

Publication Publication Date Title
CN112771583B (en) Method, device and storage medium for processing point cloud data
KR20230074137A (en) Instance adaptive image and video compression using machine learning systems
WO2022246724A1 (en) Point cloud decoding and upsampling and model training methods and apparatus
US20230047400A1 (en) Method for predicting point cloud attribute, encoder, decoder, and storage medium
WO2023130333A1 (en) Encoding and decoding method, encoder, decoder, and storage medium
Balasubramani et al. Efficient image compression techniques for compressing multimodal medical images using neural network radial basis function approach
WO2022133753A1 (en) Point cloud encoding and decoding methods and systems, point cloud encoder, and point cloud decoder
WO2022100140A1 (en) Compression encoding method and apparatus, and decompression method and apparatus
WO2022067775A1 (en) Point cloud encoding and decoding method, encoder, decoder and codec system
Feng et al. Neural subspaces for light fields
WO2022257145A1 (en) Point cloud attribute prediction method and apparatus, and codec
CN115086716B (en) Method and device for selecting neighbor points in point cloud and coder-decoder
WO2022140937A1 (en) Point cloud encoding method and system, point cloud decoding method and system, point cloud encoder, and point cloud decoder
WO2024026712A1 (en) Point cloud coding method and apparatus, point cloud decoding method and apparatus, and device and storage medium
WO2023024842A1 (en) Point cloud encoding/decoding method, apparatus and device, and storage medium
Ungureanu et al. Image-Compression Techniques: Classical and “Region-of-Interest-Based” Approaches Presented in Recent Papers
US20230082456A1 (en) Point cloud attribute prediction method and apparatus, and related device
WO2023103565A1 (en) Point cloud attribute information encoding and decoding method and apparatus, device, and storage medium
US20230237704A1 (en) Point cloud decoding and encoding method, and decoder, encoder and encoding and decoding system
WO2022257150A1 (en) Point cloud encoding and decoding methods and apparatus, point cloud codec, and storage medium
WO2023050381A1 (en) Image and video coding using multi-sensor collaboration
WO2024011381A1 (en) Point cloud encoding method and apparatus, point cloud decoding method and apparatus, device and storage medium
CN115086658B (en) Point cloud data processing method and device, storage medium and encoding and decoding equipment
WO2024007144A1 (en) Encoding method, decoding method, code stream, encoders, decoders and storage medium
WO2023178662A1 (en) Image and video coding using multi-sensor collaboration and frequency adaptive processing

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21942304

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE