CN110278444B - Sparse representation three-dimensional point cloud compression method adopting geometric guidance - Google Patents

Sparse representation three-dimensional point cloud compression method adopting geometric guidance Download PDF

Info

Publication number
CN110278444B
CN110278444B CN201910645303.1A CN201910645303A CN110278444B CN 110278444 B CN110278444 B CN 110278444B CN 201910645303 A CN201910645303 A CN 201910645303A CN 110278444 B CN110278444 B CN 110278444B
Authority
CN
China
Prior art keywords
point cloud
dimensional point
block
mean value
adopting
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910645303.1A
Other languages
Chinese (zh)
Other versions
CN110278444A (en
Inventor
曾焕强
谷帅
侯军辉
陈婧
朱建清
张联昌
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Huaqiao University
Original Assignee
Huaqiao University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Huaqiao University filed Critical Huaqiao University
Priority to CN201910645303.1A priority Critical patent/CN110278444B/en
Publication of CN110278444A publication Critical patent/CN110278444A/en
Application granted granted Critical
Publication of CN110278444B publication Critical patent/CN110278444B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/124Quantisation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/176Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/597Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • H04N19/63Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding using sub-band based transform, e.g. wavelets
    • H04N19/64Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding using sub-band based transform, e.g. wavelets characterised by ordering of coefficients or of bits for transmission
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/90Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using coding techniques not provided for in groups H04N19/10-H04N19/85, e.g. fractals
    • H04N19/91Entropy coding, e.g. variable length coding [VLC] or arithmetic coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/90Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using coding techniques not provided for in groups H04N19/10-H04N19/85, e.g. fractals
    • H04N19/96Tree coding, e.g. quad-tree coding

Abstract

The invention discloses a sparse representation three-dimensional point cloud compression method adopting geometric guidance, which belongs to the field of video coding and comprises the following steps: partitioning the input three-dimensional point cloud by using an octree; obtaining an original redundant dictionary by adopting a graph transformation method; utilizing the geometric information of the point cloud in the block to perform downsampling on the original redundant dictionary; carrying out mean value removal on each unit block, and then carrying out sparse representation on the color information subjected to mean value removal on a down-sampling dictionary; carrying out predictive coding on the mean value of each coding unit block by using an octree-based block mean value prediction algorithm; coding the quantized sparse coefficient by adopting a Run-Level method; finally, all the encoded parameters are entropy encoded by an arithmetic encoder. The invention utilizes sparse representation, can efficiently compress huge three-dimensional point cloud data, and greatly improves the transmission and storage efficiency of the three-dimensional point cloud.

Description

Sparse representation three-dimensional point cloud compression method adopting geometric guidance
Technical Field
The invention relates to the field of video coding, in particular to a sparse representation three-dimensional point cloud compression method adopting geometric guidance.
Background
With the rapid development of three-dimensional acquisition and multimedia communication technologies, three-dimensional point clouds are widely applied to multiple fields of remote communication, augmented reality, immersive communication and the like as a new transmission medium. A three-dimensional point cloud is composed of a large set of points with specific three-dimensional position information, where each point has one or more features (e.g., color, normal, etc.). As a novel spatial data type, the three-dimensional point cloud can more effectively represent three-dimensional objects or scenes with more complex topological structures. Its enormous volume of data presents significant challenges to its storage and transmission. Therefore, how to construct an efficient and stable three-dimensional point cloud compression method is very important.
Unlike traditional natural image video, the three-dimensional point cloud is characterized by irregularities, which are mainly manifested in that not every voxel is occupied, which poses a great challenge to the compression of the three-dimensional point cloud. As research work progresses, sparse representation can be used to efficiently compress high-dimensional data, but how to build a redundant dictionary and sparsely represent color information of a point cloud on the redundant dictionary remains a problem to be solved.
Disclosure of Invention
The invention mainly aims to overcome the defects in the prior art and provides a sparse representation three-dimensional point cloud compression method adopting geometric guidance. The method fully considers the correlation of adjacent points in the three-dimensional point cloud data, converts color information into sparse coefficients by a sparse representation method, and realizes effective compression of the huge three-dimensional point cloud data by encoding nonzero coefficients with less quantity.
The invention adopts the following technical scheme:
a sparse representation three-dimensional point cloud compression method adopting geometric guidance is characterized by comprising the following steps:
1) Dividing the input three-dimensional point cloud to obtain a plurality of unit blocks;
2) Calculating an original redundant dictionary of the unit block;
3) Acquiring the original redundant dictionary by using a descending method based on the geometric information of the inner points of the current unit block to obtain a descending sampling redundant dictionary;
4) Calculating the color average value in each unit block, and performing the average value removal on the color information of each unit block;
5) Sparse representation is carried out on the color information subjected to mean value removal on a down-sampled redundant dictionary, and quantized sparse coefficients are obtained;
6) Carrying out predictive coding on the color mean value of each unit block by using an octree-based block mean value prediction algorithm;
7) Coding the quantized sparse coefficient by adopting a Run-Level method;
8) Entropy coding the coded parameters obtained in the steps 6) and 7) by adopting an arithmetic coder.
Preferably, in step 1), the input three-dimensional point cloud is segmented into unit blocks with uniform sizes by using octree.
Preferably, in step 2), the original redundant dictionary of the unit block is obtained by a graphic transformation.
As can be seen from the above description of the present invention, compared with the prior art, the present invention has the following advantages:
1. the method of the invention adopts the octree to segment the three-dimensional point cloud, thereby not only effectively keeping the correlation between points, but also ensuring that the size of each segmented unit block is the same.
2. The method of the invention utilizes the strong correlation between points, and obtains the original redundant dictionary by assuming that the voxels in each block are occupied and then adopting a graph transformation method; obtaining a down-sampling redundant dictionary by using the geometric information of the points in the block; the color information of the points in the block is sparsely represented on the downsampled redundant dictionary, so that the energy of color signals can be concentrated on a small number of non-zero coefficients, and the data volume of the three-dimensional point cloud is greatly reduced.
3. The invention designs an octree-based block mean prediction mode, which can eliminate color redundancy among each unit block; and for the transformed sparse coefficient, efficiently coding the quantized sparse coefficient by using a Run-level method.
Drawings
FIG. 1 is a main flow diagram of the process of the present invention;
FIG. 2 is an octree partitioning method of the present invention;
FIG. 3 is a schematic diagram of a block mean prediction method based on octree according to the present invention;
FIG. 4 is a flow chart of Run-Level encoding method of the present invention.
The invention is described in further detail below with reference to the figures and specific examples.
Detailed Description
The invention is further described below by means of specific embodiments.
Referring to fig. 1, a sparse representation three-dimensional point cloud compression method using geometric guidance performs octree segmentation on an input three-dimensional point cloud, and then assumes that voxels in each block are occupied, obtains an original redundant dictionary using graph transformation, and obtains a downsampled redundant dictionary through geometric information in each block. And then, sparsely representing the color information in each block on the downsampled redundant dictionary to obtain a sparse coefficient. And the mean value of each block is subjected to predictive coding and quantization by using the block mean value prediction method based on the octree. And coding the non-zero coefficient of the quantized sparse coefficient by adopting a Run-Level coding method. And finally, entropy coding is carried out on all coding parameters by adopting an algorithm coder. The specific implementation steps are as follows:
1) The octree is used to segment the original input point cloud to obtain unit blocks with uniform size, i.e. code blocks, as shown in fig. 2. Given a segmentation depth n, the original input three-dimensional point cloud is first divided into eight equal parts, and after segmentation, a block containing points is marked as 1, and a block without points is marked as 0. The block labeled 1 then continues to be octaved, the same block is labeled 0,1, based on the absence of a point within the block, and then the block labeled 1 continues to be partitioned. This process is repeated until the division is made to a given depth, and the number of divided blocks is Z.
2) Using the unit block as a processing unit, assuming that all the voxels in each block are occupied, calculating the weight value W between each voxeli,jThen, the original redundant dictionary D can be obtained by using a graph transformation, specifically, the following formula is used:
Figure BDA0002133241020000041
A=diag(d1,d2,...,dn)
Figure BDA0002133241020000042
L=A-W=DΛD-1
wherein dist represents the Euclidean distance between the ith voxel and the jth voxel, τ is a threshold, σ is a preset parameter of each voxel, L is a graph Laplace operator, and Λ is a diagonal matrix translation operator.
3) Downsampled redundant dictionary
Figure BDA0002133241020000043
The method can be obtained through geometric information of points in the block, and particularly, for the ith block, the redundant dictionary is downsampled
Figure BDA0002133241020000044
Can be obtained by the following formula:
Figure BDA0002133241020000045
wherein SiIs a down-sampling matrix obtained from an identity matrix by removing the rows of unoccupied voxels in the ith block.
4) The color mean value x may be obtained by averaging, and the color information x for sparse representation is obtained by averaging the original color information. Specifically, for the ith block, the following formula is used to obtain:
Figure BDA0002133241020000046
Figure BDA0002133241020000047
5) Sparse coefficient ciThe acquisition can be sparsely represented on the downsampled redundant dictionary by the de-averaged color information. Specifically, for the ith block, the following optimization problem can be solved and obtained by using an OMP algorithm:
Figure BDA0002133241020000051
where N is the number of occupied blocks and ε is the reconstruction error, used to control the sparse coefficient ciSparsity of ·0Is the 0 norm of the vector and is used for counting the number of nonzero elements of the input vector.
6) The mean value of each unit block is coded by adopting a prediction algorithm based on an octree. Fig. 3 shows an octree prediction method, which specifically includes the following steps:
first, assume that after octree segmentation, the original three-dimensional point cloud has been divided into m × m × m blocks, each having its own three-dimensional space coordinates (p) in three-dimensional spacex,py,pz) Wherein, 1 is less than or equal to px≤m,1≤py≤m,1≤pzM is less than or equal to m. Fig. 3 (a) shows 8 neighboring blocks of the current coding block, where the coordinates of the gray block number 1 is (p)x,py,pz) The coordinates of the other seven blocks with different colors numbered 2-8 are (p) for the current coding blockx-Δpx,py-Δpy,pz-Δpz) Wherein Δ px,Δpy,ΔpzE {0,1}, which is the already encoded reference block. For the current cell block (indicated by grey blocks in fig. 3 (b), with a block label of 1), 9 prediction modes will be used, where the 9 prediction modes include oneDC mode and 8 angle mode. Of the 8 angular modes, mode1 to mode7 are modes of seven reference neighboring blocks; mode8 is the mode of the reference macroblock, which is a block with 7 reference blocks as a whole. By comparing with the neighboring blocks and macroblocks, the best prediction mode can be obtained by the following formula:
modedecision=min|b-bi|
by comparing with 8 reference blocks, 8 prediction residuals can be obtained, and the minimum absolute value of the 8 residuals corresponds to the best prediction mode.
Where b is the mean of the current block of coding units, b is the mean of the current block of coding unitsiIs the mean of the reference block, i ∈ {1,2, ·,8}; | is an absolute value sign. Since the three-dimensional coordinates comprise one block, namely the block at the outermost layer of the whole three-dimensional point cloud, and the blocks have no reference block, the color mean value of each block is directly quantized, namely, the DC mode is adopted.
And 7) the quantized sparse coefficient can be coded by a Run-Level method.
Specifically, as shown in fig. 4, the quantized sparse coefficients are represented as 2 one-dimensional arrays, which are denoted as Run and Level, respectively. Since the number of non-zero coefficients in each block is different, numNonzeroCoef is introduced to record the number of non-zero coefficients in each block. For the ith block, only if NumNonorcoef [ i ]]Is not equal to 0 (run)i,leveli) And (5) coding, wherein i is more than or equal to 1 and less than or equal to Z. For quantized coefficient ciIf no non-zero coefficients are present, numNonorcoef [ i ]]A value of 0 indicates that no coefficients need to be encoded. If c isiIn which is NiA non-zero coefficient, N is first determinediStored in variable NumNonzeroCoef [ i ]]Then, run is stored separatelyiAnd leveliInto Run { i } and Level { i }. The above process is repeated until all the unit blocks are coded.
And step 8) entropy coding all the coded parameters in the previous step by adopting an arithmetic coder, wherein the coded parameters comprise quantized prediction residual errors, an optimal prediction mode, run, level and NumNonorcoef.
The above description is only an embodiment of the present invention, but the design concept of the present invention is not limited thereto, and any insubstantial modifications made by using the design concept should fall within the scope of infringing the present invention.

Claims (3)

1. A sparse representation three-dimensional point cloud compression method adopting geometric guidance is characterized by comprising the following steps:
1) Dividing the input three-dimensional point cloud to obtain a plurality of unit blocks;
2) Calculating an original redundant dictionary of the unit block;
3) Based on the geometric information of the inner points of the current cell block, downsampling the original redundant dictionary to obtain a downsampled redundant dictionary;
4) Calculating the color average value in each unit block, and performing the average value removal on the color information of each unit block;
5) Sparse representation is carried out on the color information subjected to mean value removal on a down-sampled redundant dictionary, and quantized sparse coefficients are obtained;
6) Carrying out predictive coding on the color mean value of each unit block by using an octree-based block mean value prediction algorithm;
7) Coding the quantized sparse coefficient by adopting a Run-Level method;
8) Entropy coding the coded parameters obtained in the steps 6) and 7) by adopting an arithmetic coder.
2. The method of claim 1, wherein in step 1), the input three-dimensional point cloud is segmented into uniform-sized unit blocks by using octree.
3. The method of compressing a three-dimensional point cloud with geometric guidance according to claim 1, wherein in step 2), the original redundant dictionary of unit blocks is obtained by graphic transformation.
CN201910645303.1A 2019-07-17 2019-07-17 Sparse representation three-dimensional point cloud compression method adopting geometric guidance Active CN110278444B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910645303.1A CN110278444B (en) 2019-07-17 2019-07-17 Sparse representation three-dimensional point cloud compression method adopting geometric guidance

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910645303.1A CN110278444B (en) 2019-07-17 2019-07-17 Sparse representation three-dimensional point cloud compression method adopting geometric guidance

Publications (2)

Publication Number Publication Date
CN110278444A CN110278444A (en) 2019-09-24
CN110278444B true CN110278444B (en) 2022-11-01

Family

ID=67964759

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910645303.1A Active CN110278444B (en) 2019-07-17 2019-07-17 Sparse representation three-dimensional point cloud compression method adopting geometric guidance

Country Status (1)

Country Link
CN (1) CN110278444B (en)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114009014A (en) * 2019-09-30 2022-02-01 Oppo广东移动通信有限公司 Color component prediction method, encoder, decoder, and computer storage medium
CN111866518B (en) * 2020-07-29 2022-05-27 西安邮电大学 Self-adaptive three-dimensional point cloud compression method based on feature extraction
CN112184840B (en) * 2020-09-22 2022-09-02 上海交通大学 3D point cloud compression system based on multi-scale structured dictionary learning
CN112256652B (en) * 2020-10-19 2022-09-16 济南大学 Three-dimensional point cloud attribute compression method, system and terminal

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103489203A (en) * 2013-01-31 2014-01-01 清华大学 Image coding method and system based on dictionary learning
CN104331913A (en) * 2014-11-19 2015-02-04 西安电子科技大学 Polarized SAR polarization method based on sparse K-SVD (Singular Value Decomposition)
CN109166160A (en) * 2018-09-17 2019-01-08 华侨大学 A kind of three-dimensional point cloud compression method predicted using figure
CN109345619A (en) * 2018-08-10 2019-02-15 华北电力大学(保定) Massive point cloud space management based on class octree encoding

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170214943A1 (en) * 2016-01-22 2017-07-27 Mitsubishi Electric Research Laboratories, Inc. Point Cloud Compression using Prediction and Shape-Adaptive Transforms
US10909725B2 (en) * 2017-09-18 2021-02-02 Apple Inc. Point cloud compression

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103489203A (en) * 2013-01-31 2014-01-01 清华大学 Image coding method and system based on dictionary learning
CN104331913A (en) * 2014-11-19 2015-02-04 西安电子科技大学 Polarized SAR polarization method based on sparse K-SVD (Singular Value Decomposition)
CN109345619A (en) * 2018-08-10 2019-02-15 华北电力大学(保定) Massive point cloud space management based on class octree encoding
CN109166160A (en) * 2018-09-17 2019-01-08 华侨大学 A kind of three-dimensional point cloud compression method predicted using figure

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
基于压缩感知的点云数据编码与重建;吴鑫;《中国优秀硕士学位论文全文数据库》;20150331;全文 *
采用改进图形变换的3D点云压缩;谷帅等;《信号处理》;20190131;全文 *

Also Published As

Publication number Publication date
CN110278444A (en) 2019-09-24

Similar Documents

Publication Publication Date Title
CN110278444B (en) Sparse representation three-dimensional point cloud compression method adopting geometric guidance
Sun et al. A novel point cloud compression algorithm based on clustering
CN110418135B (en) Point cloud intra-frame prediction method and device based on neighbor weight optimization
Davoine et al. Fractal image compression based on Delaunay triangulation and vector quantization
US11836954B2 (en) 3D point cloud compression system based on multi-scale structured dictionary learning
Golla et al. Real-time point cloud compression
Huang et al. Octree-Based Progressive Geometry Coding of Point Clouds.
CN109166160B (en) Three-dimensional point cloud compression method adopting graph prediction
CN103295197B (en) Based on the image super-resolution rebuilding method of dictionary learning and bilateral canonical
CN111953998A (en) Point cloud attribute coding and decoding method, device and system based on DCT (discrete cosine transformation)
KR20120036834A (en) Method for encoding/decoding a 3d mesh model that comprises one or more components
CN101383972A (en) Remote sensed image compression method based on space prediction and transformation
Sun et al. A novel coding architecture for lidar point cloud sequence
CN113518226A (en) G-PCC point cloud coding improvement method based on ground segmentation
WO2022166957A1 (en) Point cloud data preprocessing method, point cloud geometry coding method and device, and point cloud geometry decoding method and device
CN114332259A (en) Point cloud coding and decoding method based on vehicle-mounted laser radar
CN110349228B (en) Triangular mesh compression method for data-driven least square prediction
CN116665053B (en) High-resolution remote sensing image building identification method and system considering shadow information
CN114554175B (en) Classification rearrangement-based lossless compression method for two-dimensional point cloud distance images
CN114915793B (en) Point cloud coding and decoding method and device based on two-dimensional regularized plane projection
CN115278255B (en) Data storage system for safety management of strength instrument
CN114025146B (en) Dynamic point cloud geometric compression method based on scene flow network and time entropy model
CN114915794B (en) Point cloud coding and decoding method and device based on two-dimensional regularized plane projection
CN115118998A (en) Inter-frame coding method, decoding method and device for enhancing projection correlation
CN117974818A (en) Point cloud geometric compression method based on spatial channel mixed context model

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant