CN115499660A - Method and system for rapidly determining dynamic 3D point cloud compression interframe coding mode - Google Patents

Method and system for rapidly determining dynamic 3D point cloud compression interframe coding mode Download PDF

Info

Publication number
CN115499660A
CN115499660A CN202211185785.5A CN202211185785A CN115499660A CN 115499660 A CN115499660 A CN 115499660A CN 202211185785 A CN202211185785 A CN 202211185785A CN 115499660 A CN115499660 A CN 115499660A
Authority
CN
China
Prior art keywords
coding
mode
target coding
preset condition
frame
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202211185785.5A
Other languages
Chinese (zh)
Inventor
李跃
黄�俊
刘杰
万亚平
王超峰
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
University of South China
Original Assignee
University of South China
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by University of South China filed Critical University of South China
Priority to CN202211185785.5A priority Critical patent/CN115499660A/en
Publication of CN115499660A publication Critical patent/CN115499660A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/157Assigned coding mode, i.e. the coding mode being predefined or preselected to be further used for selection of another element or parameter
    • H04N19/159Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/119Adaptive subdivision aspects, e.g. subdivision of a picture into rectangular or non-rectangular coding blocks
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/146Data rate or code amount at the encoder output
    • H04N19/147Data rate or code amount at the encoder output according to rate distortion criteria

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

A method and a system for rapidly determining a dynamic 3D point cloud compression interframe coding mode relate to the technical field of point cloud compression. The invention predicts whether the optimal mode is the SKIP/Merge mode by judging whether the geometric picture frame/texture picture frame meets the first preset condition and the second preset condition, predicts whether the optimal mode is the Inter _2 Nx 2N mode by judging whether the geometric picture frame/texture picture frame meets the third preset condition and the fourth preset condition, and does not calculate the residual interframe prediction modes as long as the preset conditions are met, thereby achieving the purpose of saving time. Compared with the prior art, the method can ensure the coding quality of the target code, and can also obviously reduce the complexity of the calculation time in the process of making a coding mode decision on the target code.

Description

Method and system for rapidly determining dynamic 3D point cloud compression interframe coding mode
Technical Field
The invention relates to the technical field of point cloud compression, in particular to a method and a system for quickly determining a dynamic 3D point cloud compression interframe coding mode.
Background
With the recent development of visual capture technology, capture and digitization of three-dimensional world scenes has become more popular, where point clouds are one of the main 3D data representations and, in addition to providing spatial coordinates, provide attributes (such as color or reflectivity) associated with points in the 3D world. The point cloud in the original format requires a large amount of storage memory or transmission bandwidth. Classical dynamic point clouds as used for entertainment typically contain about 100 ten thousand points per frame, with a total bandwidth of 3.6Gbps at 30 frames per second, without compression. In addition, the advent of high resolution point cloud capture techniques has placed higher demands on the size of the point cloud in turn. In order to promote the point cloud technology and apply the point cloud technology to practical production, the compression technology is in urgent need of development.
In order to effectively compress dynamic 3D Point clouds, MPEG (Moving Picture Experts Group) is formulating a method (High Efficiency Video Coding, HEVC) for dynamic 3D Point Cloud Compression, i.e., video-based Point Cloud Compression V-PCC (Video-based Point Cloud Compression). In the HEVC compression Coding process, when Inter prediction mode selection is performed on each Coding Unit (CU), rate distortion cost (RDcost) calculations of modes such as SKIP/Merge mode, inter _2N × 2N mode, inter _ N × N mode, inter _2N × N mode, inter _ N × 2N mode, inter2N × nU mode, inter _2N × nD mode, inter _ nL × 2N mode, and Inter _ nR × 2N mode are performed in sequence, and finally, an optimal prediction mode is selected by comparing rdcosts. However, the selection of the optimal mode by calculating the RDcost also brings huge computational complexity to the compression coding, and seriously hinders the real-time transmission of the dynamic 3D point cloud. Furthermore, due to the very large difference between the geometry and color video generated by the dynamic 3D point cloud mapping and the natural video, this may result in that the proposed method for fast determination of HEVC inter-frame coding mode for the natural video cannot be well used in video-based dynamic 3D point cloud compression coding optimization. Therefore, a method for rapidly determining an inter-frame coding mode in dynamic 3D point cloud compression is urgently needed to be developed so as to solve the problem of network real-time transmission and improve user experience.
Therefore, how to reduce the computational complexity of inter-frame mode selection in the video-based dynamic 3D point cloud compression process is a technical problem that needs to be solved by those skilled in the art.
Disclosure of Invention
One objective of the present invention is to provide a method for quickly determining a dynamic 3D point cloud compression inter-frame coding mode to reduce the computational complexity in mode selection.
In order to solve the technical problems, the invention adopts the following technical scheme: a method for rapidly determining a dynamic 3D point cloud compression interframe coding mode comprises the following steps of after a target coding CU calculates the rate distortion cost of a Merge mode, judging the target coding CU as follows:
if the target coding CU is the geometric figure CU, executing first prediction distortion transformation on the target coding CU, judging whether a first preset condition is met, and if so, skipping the calculation of a subsequent inter-frame prediction mode; and when the target coding CU does not meet the first preset condition, after the calculation of the Inter2 Nx 2N mode is finished, executing third prediction distortion transformation on the target coding CU, judging whether the third preset condition is met, and if so, skipping the calculation of the subsequent Inter mode.
If the target coding CU is the texture map CU, performing second prediction distortion transformation on the target coding CU, judging whether a second preset condition is met, and if so, skipping the calculation of a subsequent inter-frame mode; and when the target coding CU does not meet the second preset condition, after the calculation of the Inter2 Nx 2N mode is finished, executing fourth prediction distortion transformation on the target coding CU, judging whether the fourth preset condition is met, and if so, skipping the calculation of the subsequent Inter mode.
In the method, specifically, in a dynamic 3D point cloud compressed intra-frame compression coding configuration, after a rate distortion cost is calculated in a Merge mode, it is determined whether a current dynamic 3D point cloud compressed coding frame does not belong to a placeholder image frame and is not an I frame;
if yes, sequentially extracting CU needing compression coding w×h Obtaining the target coding CU w×h Judging a geometric map and a texture map by the chrominance component;
obtaining a target coding CU w×h Pre-encoding of luminance componentsMeasuring distortion Err w×h (i,j);
If the current coding CU w×h Is a geometric frame, then the prediction distortion Err is corrected w×h (i, j) performing a first predictive distortion transformation to obtain Tra1 w×h (i,j);
Wherein the content of the first and second substances,
Figure BDA0003867661780000031
wherein the content of the first and second substances,
Figure BDA0003867661780000032
wherein w and h are the target coding CU respectively w×h Width and height of (ii), ori w×h (i, j) is the target coding CU w×h Of the original luminance value, pre w×h (i, j) encoding the CU for the target w×h The predicted brightness values i and j obtained after the inter-frame prediction mode calculation are respectively the target coding CU w×h The abscissa and the ordinate of the pixel point of (1), and QP is a quantization parameter of the current compressed coding frame;
for current CU w×h The region is divided into two parts horizontally and vertically to obtain two sub-matrixes horizontally and two sub-matrixes vertically to obtain the maximum Tra1 corresponding to the four sub-matrixes w×h (i, j) variance value to obtain maxVar B
For current CU w×h The region is divided into four parts horizontally and vertically to obtain four sub-matrixes horizontally and four sub-matrixes vertically to obtain the maximum Tra1 corresponding to the eight sub-matrixes w×h (i, j) variance value to obtain maxVar Q (ii) a Determining the target coding CU w×h Whether a first preset condition is met or not;
if not, acquiring the target coding CU after the Inter2 Nx 2N prediction mode w×h Predicted distortion Err of (1) w×h (i, j) on which a third predictive distortion transform is performed, resulting in Tra3 w×h (i,j);
Wherein the content of the first and second substances,
Figure BDA0003867661780000041
wherein the content of the first and second substances,
Figure BDA0003867661780000042
wherein w and h are the target coding CU w×h Width and height of (ii), ori w×h (i, j) original luminance value, pre, of the target encoded CU w×h (i, j) is the target coding CU w×h The predicted brightness values i and j obtained after the inter-frame prediction mode calculation are respectively the target coding CU w×h The abscissa and the ordinate of the pixel point of (1), and QP is a quantization parameter of the current compressed coding frame;
for the current CU w×h The region is divided into two parts horizontally and vertically to obtain two sub-matrixes horizontally and two sub-matrixes vertically to obtain the maximum Tra3 corresponding to the four sub-matrixes w×h (i, j) variance value to obtain maxVar B
For current CU w×h The area is divided into four parts horizontally and vertically, four sub-matrixes are obtained horizontally and four sub-matrixes are obtained vertically, and the maximum Tra3 corresponding to the eight sub-matrixes is obtained w×h (i, j) variance value to obtain maxVar Q (ii) a Determining the target coding CU w×h Whether a third preset condition is met;
if yes, skipping the calculation of the residual inter-frame prediction mode;
if the current coding CU w×h If it is a texture map frame, then a second prediction distortion transformation is performed on it to obtain Tra2 w×h (i,j);
Wherein, the first and the second end of the pipe are connected with each other,
Figure BDA0003867661780000043
wherein, the first and the second end of the pipe are connected with each other,
Figure BDA0003867661780000051
wherein w and h are the target coding CU respectively w×h Width and height of (ii), ori w×h (i, j) is the target coding CU w×h Of original brightnessValue of, pre w×h (i, j) encoding the CU for the target w×h The predicted brightness values i and j obtained after the inter-frame prediction mode calculation are respectively the target coding CU w×h The abscissa and the ordinate of the pixel point of (1), and QP is a quantization parameter of the current compressed coding frame;
for current CU w×h The region is divided into two parts horizontally and vertically to obtain two sub-matrixes horizontally and two sub-matrixes vertically to obtain the maximum Tra2 corresponding to the four sub-matrixes w×h (i, j) variance value to obtain maxVar B
For current CU w×h The region is divided into four parts horizontally and vertically to obtain four sub-matrixes horizontally and four sub-matrixes vertically to obtain the maximum Tra2 corresponding to the eight sub-matrixes w×h (i, j) variance value to obtain maxVar Q (ii) a Determining the target coding CU w×h Whether a second preset condition is met or not;
if not, the prediction distortion Err is corrected after the Inter2N × 2N prediction mode w×h (i, j) performing a third predictive distortion transform to obtain Tra4 w×h (i,j);
Wherein the content of the first and second substances,
Figure BDA0003867661780000052
wherein the content of the first and second substances,
Figure BDA0003867661780000053
wherein w and h are the target coding CU respectively w×h Width and height of (1), ori w×h (i, j) is the original luminance value, pre, of the target encoded CU w×h (i, j) is the target coding CU w×h The predicted brightness values i and j obtained after the inter-frame prediction mode calculation are respectively the target coding CU w×h The abscissa and the ordinate of the pixel point of (1), and QP is a quantization parameter of the current compressed coding frame;
for current CU w×h The region is divided into two parts horizontally and vertically to obtain two sub-matrixes horizontally and two sub-matrixes vertically, and the pairs in the four sub-blocks are obtainedMaximum of response Tra4 w×h (i, j) variance value to obtain maxVar B
For the current CU w×h The region is divided into four parts horizontally and vertically to obtain four sub-matrixes horizontally and four sub-matrixes vertically to obtain the maximum Tra4 corresponding to the eight sub-matrixes w×h (i, j) variance value to obtain maxVar Q (ii) a Determining the target coding CU w×h Whether a fourth preset condition is met or not;
if yes, skipping the RDcost calculation of the residual inter-frame mode;
wherein the expression of the first preset condition is as follows:
POC >4& & POC! =8& & POC! =16 and
Figure BDA0003867661780000061
and is provided with
Figure BDA0003867661780000062
The expression of the second preset condition is as follows:
Figure BDA0003867661780000063
and is
Figure BDA0003867661780000064
The expression of the third preset condition is as follows:
POC >4& & POC! =8& & POC! =16 and
Figure BDA0003867661780000065
and is
Figure BDA0003867661780000066
The expression of the fourth preset condition is as follows:
Figure BDA0003867661780000067
and is provided with
Figure BDA0003867661780000068
In the formula
Figure BDA0003867661780000069
For adaptive adjustment of the threshold value, an
Figure BDA00038676617800000610
Figure BDA00038676617800000611
In the formula
Figure BDA00038676617800000612
Respectively is to get Tra w×h (i, j) dividing the matrix into two submatrices horizontally and then calculating the variance;
Figure BDA00038676617800000613
respectively is to mix Tra w×h (i, j) longitudinally dividing the matrix into two sub-matrices and respectively calculating the variance;
Figure BDA00038676617800000614
respectively is to mix Tra w×h (i, j) transversely dividing the matrix into four sub-matrixes and respectively calculating the obtained variances;
Figure BDA00038676617800000615
respectively is to mix Tra w×h (i, j) the variance calculated after longitudinally dividing the matrix into four submatrices, wherein Tra w×h (i, j) is Tra1 w×h (i, j) or Tra2 w×h (i, j) or Tra3 w×h (i, j) or Tra4 w×h (i,j)。
Preferably, after the process of determining whether the current dynamic 3D point cloud compression encoding frame is not an I frame, the method further includes: if not, the inter-frame prediction mode is not rapidly determined for the coding frame.
More preferably, after the process of determining whether the target coding CU satisfies the first preset condition, the method further includes: if so, the target encoding skips the calculation of the residual inter prediction mode.
More preferably, after determining whether the target coding CU satisfies the second preset condition, the method further includes: if yes, the target code skips the calculation of the residual inter prediction mode.
More preferably, the determining whether the target coding CU satisfies the third preset condition further includes: if not, the target coding continues to execute the calculation of the residual inter-frame prediction mode.
More preferably, after the process of determining whether the target coding CU satisfies the fourth preset condition, the method further includes: if not, the target coding continues to execute the calculation of the residual inter-frame prediction mode.
In addition, the invention also provides a system for rapidly determining the dynamic 3D point cloud compression interframe coding mode, which comprises the following steps:
an encoding CU judging module: and the processing unit is used for judging whether the target coding CU is the geometric map CU or the texture map CU or not after calculating the Merge mode rate distortion cost of the target coding CU.
A first encoding mode determination module: the method comprises the steps of processing a target coding CU in a geometric figure CU form, wherein the processing comprises the steps of executing first prediction distortion transformation on the target coding CU, judging whether a first preset condition is met, and skipping the calculation of a subsequent inter-frame prediction mode if the first preset condition is met; and when the target coding CU does not meet the first preset condition, after the calculation of the Inter2N multiplied by 2N mode between the frames is finished, executing third prediction distortion transformation on the target coding CU, judging whether the third preset condition is met, and if so, skipping the calculation of the subsequent Inter mode.
A second encoding mode determination module: the method comprises the steps of processing a target coding CU in the form of a texture map CU, wherein the processing comprises executing second prediction distortion transformation on the target coding CU, judging whether a second preset condition is met, and skipping the calculation of a subsequent inter mode if the second preset condition is met; and when the target coding CU does not meet the second preset condition, after the calculation of the Inter2 Nx 2N mode is finished, executing fourth prediction distortion transformation on the target coding CU, judging whether the fourth preset condition is met, and if so, skipping the calculation of the subsequent Inter mode.
The system can be operated by the method for quickly determining the dynamic 3D point cloud compression interframe coding mode.
The method predicts the optimal mode by judging whether the geometric picture frame/texture picture frame meets a first preset condition or not; similarly, the best mode can be predicted by determining whether the geometric frame/texture frame satisfies the third and fourth predetermined conditions. The calculation of the residual inter prediction mode may not be performed as long as the preset condition is satisfied, thereby achieving the purpose of saving time. Therefore, the method provided by the invention not only can ensure the coding quality of the target code, but also can obviously reduce the complexity of the calculation time in the process of carrying out compression coding on the target code.
Drawings
FIG. 1 is a schematic flow chart of a method for rapidly determining a dynamic 3D point cloud compression interframe coding mode according to an embodiment of the invention;
FIG. 2 is a binary division example and depicts the variance values in the formula of the present invention
Figure BDA0003867661780000081
FIG. 3 is a quad-partitioned example and depicts variance values in the formulas of the present invention
Figure BDA0003867661780000082
Detailed Description
In order to facilitate understanding of those skilled in the art, the present invention will be further described with reference to the following examples and drawings, which are not intended to limit the present invention.
As shown in fig. 1, a method for rapidly determining a dynamic 3D point cloud compression inter-frame coding mode includes the following steps:
step S1: judging whether the current dynamic 3D point cloud compression coding frame is an I frame, if so, skipping any inter-frame mode, otherwise, executing S2;
step S2: sequentially extracting CUs needing compression coding, and acquiring chrominance component judgment geometric graphs and texture graphs of target coding CUs;
and step S3: judging whether the current frame is a geometric picture frame, if so, executing a step S4, otherwise, executing a step S7;
and step S4: computing the residual Err of the original and predicted images w×h (i, j) and performing a first predictive distortion transformation to obtain Tra1 w×h (i,j);
Step S5: for Tra1 w×h (i, j) carrying out horizontal and longitudinal halving and quartering in the corresponding CU region, and respectively calculating the maximum variance of the subblocks corresponding to the halving and the quartering to obtain maxVar B And maxVar Q Judging maxVar B And maxVar Q Whether a first preset condition is met, if yes, executing a step S6, and if not, executing a step S9;
step S6: skipping the residual inter prediction mode;
step S7: computing the residual Err of the original and predicted images w×h (i, j) and performing a second predictive distortion transformation to obtain Tra2 w×h (i,j);
Step S8: for Tra2 w×h (i, j) performing horizontal and longitudinal halving and quartering on the corresponding CU region, and respectively calculating the maximum variance of the subblocks corresponding to the halving and the quartering to obtain maxVar B And maxVar Q (ii) a Judging maxVar B And maxVar Q Whether a second preset condition is met, if so, executing a step S6, otherwise, executing a step S12;
step S9: performing a third predictive distortion transformation on the geometry to obtain Tra3 w×h (i,j);
Step S10: for Tra3 w×h (i, j) performing horizontal and longitudinal halving and quartering on the corresponding CU region, and respectively calculating the maximum variance of the subblocks corresponding to the halving and the quartering to obtain maxVar B And maxVar Q (ii) a Judging maxVar B And maxVar Q If the third preset condition is met, executing the step S11 if the third preset condition is met, otherwise, continuing to perform residual inter-frame predictionMeasuring a mode;
step S11: skipping the residual inter prediction mode;
step S12: performing a fourth predictive distortion transformation on the geometry to obtain Tra4 w×h (i,j);
Step S13: for Tra4 w×h (i, j) performing horizontal and longitudinal halving and quartering on the corresponding CU region, and respectively calculating the maximum variance of the subblocks corresponding to the halving and the quartering to obtain maxVar B And maxVar Q (ii) a Judging maxVar B And maxVar Q Whether a fourth preset condition is met, if yes, executing the step S11, and otherwise, continuing to perform the residual inter-frame prediction mode;
wherein, the first and the second end of the pipe are connected with each other,
Figure BDA0003867661780000101
wherein the content of the first and second substances,
Figure BDA0003867661780000102
wherein the content of the first and second substances,
Figure BDA0003867661780000103
wherein the content of the first and second substances,
Figure BDA0003867661780000104
wherein the content of the first and second substances,
Figure BDA0003867661780000105
where w and h are the target coding CU w×h Width and height of (1), ori w×h (i, j) is the original luminance value, pre, of the target coded CU w×h (i, j) is the target coding CU w×h The predicted brightness values i and j obtained after the inter-frame prediction mode calculation are respectively target coding CU w×h The abscissa and ordinate of the pixel point of (1), QP is the quantization parameter of the current compressed encoded frame, hereinafter Tra3 w×h (i, j) and Tra4 w×h (i, j) are collectively referred to as Tra w×h (i,j);
The first preset conditional expression is as follows:
POC >4& & POC! =8& & POC! =16 and
Figure BDA0003867661780000111
and is provided with
Figure BDA0003867661780000112
The second preset conditional expression is as follows:
Figure BDA0003867661780000113
and is
Figure BDA0003867661780000114
The third preset conditional expression is as follows:
POC >4& & POC! =8& & POC! =16 and
Figure BDA0003867661780000115
and is provided with
Figure BDA0003867661780000116
The fourth preset conditional expression is as follows:
Figure BDA0003867661780000117
and is provided with
Figure BDA0003867661780000118
In the formula
Figure BDA0003867661780000119
For adaptive adjustment of threshold value, and
Figure BDA00038676617800001110
Figure BDA00038676617800001111
in the formula
Figure BDA00038676617800001112
Respectively is to get Tra w×h (i, j) dividing the matrix into two submatrices horizontally and then calculating the variance;
Figure BDA00038676617800001113
respectively is to get Tra w×h (i, j) longitudinally dividing the matrix into two sub-matrices and respectively calculating the variance;
Figure BDA00038676617800001114
respectively is to mix Tra w×h (i, j) dividing the matrix into four sub-matrixes horizontally and then calculating the variance;
Figure BDA00038676617800001115
respectively is to mix Tra w×h (i, j) dividing the matrix into four submatrices longitudinally and then calculating the variance.
S4-S8 are utilized to predict whether the optimal mode is the SKIP/Merge mode by judging whether the geometric picture frame/texture picture frame meets the first preset condition and the second preset condition; similarly, the steps S9-S13 can predict whether the optimal mode is the Inter _2N × 2N mode by determining whether the geometric frame/texture frame satisfies the third and fourth predetermined conditions. The calculation of the residual inter prediction mode may not be performed as long as the preset condition is satisfied, thereby achieving the purpose of saving time. Therefore, the method provided by the embodiment can ensure that the target coding CU can not only be ensured w×h Can also significantly reduce the coding quality of the target CU w×h The computational time complexity in the coding mode decision process is performed.
It can be seen that in this embodiment, the CU is currently encoded w×h Is predicted based on the transformed value Tra1 of the prediction distortion w×h (i, j) can be reversedMapping a current coding CU w×h The best inter prediction mode should be selected. In coding a target CU w×h In the process of inter-frame prediction, four judgment standards are set to judge the target coding CU w×h Whether the residual inter prediction mode can be skipped, i.e., by determining the target coding CU w×h Whether the predicted distortion transformation value satisfies a first preset condition, a second preset condition, a third preset condition and a fourth preset condition to determine the target coding CU w×h Whether the calculation of the residual inter prediction mode can be skipped, and the target coding CU is coded w×h In the process of dividing, the threshold value can be adjusted in a self-adaptive way
Figure BDA0003867661780000121
To achieve a balance between savings in coding time and increased coding rate. Obviously, the target coding CU can be terminated early due to the method w×h Since various inter-frame prediction modes are performed, the method for quickly determining the inter-frame coding mode provided by the embodiment can significantly reduce the time spent on encoding the target CU w×h Temporal complexity in the inter mode selection process is performed.
On the basis of the above, the present embodiment further describes and optimizes the technical solution, and as a preferred implementation, in the step S1: and if the current dynamic 3D point cloud compressed coding frame is the I frame, the method does not perform the rapid mode determination on the current coding frame.
It should be appreciated by those skilled in the art that in the dynamic 3D point cloud compression interframe compression coding configuration, both the placeholder map frame and the geometric map frame and the texture map frame are included, and both the I frame, the P frame and the B frame are included, and the proposed method is a method for performing fast mode determination only for the B frame of the geometric map and the texture map.
By the method provided by the embodiment, the CU can be further ensured to be coded on the target CU w×h Integrity in the mode determination process is performed.
Based on the technical content disclosed by the above embodiment, dynamic 3D point cloud compression reference software TCM2-V7.0 is taken as a test platform, and the corresponding HEVC reference software is HM-16.20+ scm-8.8. The rapid determination method disclosed in the present embodiment is executed on a PC of Inter (R) Core (TM) i7_9700 cpu, 169g RAM, and thus the feasibility and effectiveness of the method are evaluated. The dynamic 3D point cloud sequence of the universal test includes "root", "readandblack", "soldier", "queen", "longaddress", and each dynamic 3D point cloud includes 32 frames. The encoding configuration is Random Access (RA) and employs five combinations of encoding Quantization Parameters (QP), (32, 42), (28, 37), (24, 32), (20, 27), (16, 22), respectively, for geometric video and texture video compression in dynamic 3D point clouds. For the evaluation of the geometric video compression performance of the dynamic 3D point cloud, the code rate change conditions (BD-rates) of point-to-point PSNR (D1) and point-to-plane PSNR (D2) are adopted, and for the evaluation of the color video compression performance of the dynamic 3D point cloud, the BD-rates of Luma, cb and Cr are adopted. The coding time savings are defined as:
Figure BDA0003867661780000131
wherein i represents 5 different QP values, T O For compression of the original test pattern, T P For compression coding time, TS, after the invention is applied to the original test model GP And TS TP The time for compressing and encoding the geometric video P frame of the dynamic 3D point cloud is saved, and the time for compressing and encoding the texture video P frame is saved.
Please refer to table 1, wherein table 1 shows the performance comparison results of the method of this embodiment on TCM2-V7.0 testing platform, wherein
Figure BDA0003867661780000132
Figure BDA0003867661780000133
TABLE 1 Performance comparison of the methods provided in this example on the TCM2-V7.0 test platform (unit:%)
Figure BDA0003867661780000141
Please refer to table 2, table 2 is a comparison result of the performance of the method provided in this embodiment on the TCM2-V7.0 testing platform, wherein
Figure BDA0003867661780000142
Figure BDA0003867661780000143
TABLE 2 Performance comparison of the methods provided in this example on the TCM2-V7.0 test platform (unit:%)
Figure BDA0003867661780000144
Please refer to table 3, table 3 shows the performance comparison results of the method of this embodiment on the TCM2-V7.0 testing platform, wherein
Figure BDA0003867661780000145
Figure BDA0003867661780000146
TABLE 3 Performance comparison of the methods provided in this example on the TCM2-V7.0 test platform (unit:%)
Figure BDA0003867661780000151
Please refer to table 4, table 4 shows the performance comparison results of the method of this embodiment on the TCM2-V7.0 testing platform, wherein
Figure BDA0003867661780000152
Figure BDA0003867661780000153
TABLE 4 Performance comparison of the methods provided in this example on the TCM2-V7.0 test platform (unit:%)
Figure BDA0003867661780000154
As can be seen from the comparison of BD-rates and encoding time savings for geometry video and texture video for dynamic 3D point clouds in tables 1, 2, 3 and 4, the adaptive adjustment threshold is used
Figure BDA0003867661780000161
For different values, the method provided by the embodiment can save the geometric video and color video coding time (39.91%, 47.48%), (45.47%, 50.95%), (51.41%, 58.54%) and (55.51%, 64.13%) respectively on average, while the BD-rates (D1 and D2) of the geometric video and the BD-rates (Luma) of the texture video increase on average (-0.18%, -0.25%, 0.09%), (-0.30%, -0.39%, 0.20%), (-0.13%, -0.24%, 0.58%) and (0.21%, 0.05%, 1.45%) respectively. With adaptive adjustment of the threshold
Figure BDA0003867661780000162
The more time savings are made in the compression encoding of the geometric and color videos of the dynamic 3D point cloud, and the BD-rates of D1, D2 and Luma increase.
In conclusion, the rapid determination method provided by the invention can greatly reduce the computational complexity of the inter-frame prediction mode under the condition of hardly reducing the compression quality of the dynamic 3D point cloud compression geometry and the texture video. Meanwhile, the method is simple in design, does not refer to CU time-space domain information, and can be integrated into a parallel compression coding framework to further reduce the dynamic 3D point cloud compression coding time.
The method for rapidly determining the dynamic 3D point cloud compression interframe coding mode provided by the invention is described in detail above, and specific examples are applied in the text to explain the principle and the implementation of the invention, and the description of the above embodiments is only used to help understanding the method and the core idea of the invention; meanwhile, for a person skilled in the art, according to the idea of the present invention, there may be variations in the specific embodiments and the application scope, and in summary, the content of the present specification should not be construed as a limitation to the present invention.

Claims (9)

1. The method for rapidly determining the dynamic 3D point cloud compression interframe coding mode is characterized by comprising the following steps of: after the target coding CU calculates the Merge mode rate-distortion cost, the target coding CU is judged as follows:
if the target coding CU is the geometric graph CU, performing first prediction distortion transformation on the target coding CU, judging whether a first preset condition is met, and if so, skipping the calculation of a subsequent inter-frame prediction mode; when the target coding CU does not meet the first preset condition, after the calculation of the Inter2 Nx 2N mode between the frames is finished, executing third prediction distortion transformation on the target coding CU, judging whether the third preset condition is met, and if so, skipping the calculation of the subsequent Inter mode;
if the target coding CU is the texture map CU, performing second prediction distortion transformation on the target coding CU, judging whether a second preset condition is met, and if so, skipping the calculation of a subsequent inter-frame mode; and when the target coding CU does not meet the second preset condition, after the calculation of the Inter2 Nx 2N mode is finished, executing fourth prediction distortion transformation on the target coding CU, judging whether the fourth preset condition is met, and if so, skipping the calculation of the subsequent Inter mode.
2. The method for fast determination of inter-frame coding mode for dynamic 3D point cloud compression according to claim 1, wherein: in the dynamic 3D point cloud compressed intra-frame compression coding configuration, after the rate distortion cost is calculated in the Merge mode, whether the current dynamic 3D point cloud compressed coding frame does not belong to a placeholder image frame or not and whether the current dynamic 3D point cloud compressed coding frame does not belong to an I frame or not is judged;
if both are, sequentially extracting CU requiring compression coding w×h Obtaining the target coding CU w×h Judging a geometric map and a texture map by the chrominance component;
obtaining a target coding CU w×h Prediction distortion Err of luminance component w×h (i,j);
If the current coding CU w×h Is a geometric picture frame, then the predicted distortion Err is corrected w×h (i, j) performing a first predictive distortion transformation to obtain Tra1 w×h (i,j);
Wherein the content of the first and second substances,
Figure FDA0003867661770000021
wherein the content of the first and second substances,
Figure FDA0003867661770000022
wherein w and h are the target coding CU respectively w×h Width and height of (ii), ori w×h (i, j) is the target coding CU w×h Of the original brightness value, pre w×h (i, j) encoding the CU for the target w×h The predicted brightness values i and j obtained after the inter-frame prediction mode calculation are respectively the target coding CU w×h The abscissa and the ordinate of the pixel point of (1), and QP is a quantization parameter of the current compressed coding frame;
for current CU w×h The region is divided into two parts horizontally and vertically to obtain two sub-matrixes horizontally and two sub-matrixes vertically to obtain the maximum Tra1 corresponding to the four sub-matrixes w×h (i, j) variance value to obtain maxVar B
For current CU w×h The region is divided into four parts horizontally and vertically to obtain four sub-matrixes horizontally and four sub-matrixes vertically to obtain the maximum Tra1 corresponding to the eight sub-matrixes w×h (i, j) variance value to obtain maxVar Q
Determining the target coding CU w×h Whether a first preset condition is met or not;
if not, acquiring the target coding CU after the Inter2N multiplied by 2N prediction mode w×h Predicted distortion Err of (1) w×h (i, j) on which a third predictive distortion transform is performed, resulting in Tra3 w×h (i,j);
Wherein, the first and the second end of the pipe are connected with each other,
Figure FDA0003867661770000023
wherein, the first and the second end of the pipe are connected with each other,
Figure FDA0003867661770000024
wherein w and h are the target coding CU respectively w×h Width and height of (ii), ori w×h (i, j) original luminance value, pre, of the target encoded CU w×h (i, j) is the target coding CU w×h The predicted brightness values i and j obtained after the inter-frame prediction mode calculation are respectively the target coding CU w×h The abscissa and the ordinate of the pixel point of (1), and QP is a quantization parameter of the current compressed coding frame;
for the current CU w×h The region is divided into two parts horizontally and vertically to obtain two sub-matrixes horizontally and two sub-matrixes vertically to obtain the maximum Tra3 corresponding to the four sub-matrixes w×h (i, j) variance value to obtain maxVar B
For current CU w×h The area is divided into four parts horizontally and vertically, four sub-matrixes are obtained horizontally and four sub-matrixes are obtained vertically, and the maximum Tra3 corresponding to the eight sub-matrixes is obtained w×h (i, j) variance value to obtain maxVar Q
Determining the target coding CU w×h Whether a third preset condition is met;
if yes, skipping the calculation of the residual inter-frame prediction mode;
if the current coding CU w×h If it is a texture map frame, then a second prediction distortion transformation is performed on it to obtain Tra2 w×h (i,j);
Wherein the content of the first and second substances,
Figure FDA0003867661770000031
wherein the content of the first and second substances,
Figure FDA0003867661770000032
wherein w and h are the target coding CU w×h Width and height of (1), ori w×h (i, j) is the target coding CU w×h Of the original brightness value, pre w×h (i, j) encoding the CU for the target w×h The predicted brightness values i and j obtained after the inter-frame prediction mode calculation are respectively the target coding CU w×h The abscissa and the ordinate of the pixel point of (1), and QP is a quantization parameter of the current compressed coding frame;
for current CU w×h The region is divided into two parts horizontally and vertically to obtain two sub-matrixes horizontally and two sub-matrixes vertically to obtain the maximum Tra2 corresponding to the four sub-matrixes w×h (i, j) variance value to obtain maxVar B
For the current CU w×h The area is divided into four parts horizontally and vertically to obtain four sub-matrixes horizontally and four sub-matrixes vertically, and the maximum Tra2 corresponding to the eight sub-matrixes is obtained w×h (i, j) variance value to obtain maxVar Q
Determining the target coding CU w×h Whether a second preset condition is met or not;
if not, the prediction distortion Err is corrected after the Inter2N × 2N prediction mode w×h (i, j) performing a third predictive distortion transformation to obtain Tra4 w×h (i,j);
Wherein, the first and the second end of the pipe are connected with each other,
Figure FDA0003867661770000041
wherein the content of the first and second substances,
Figure FDA0003867661770000042
wherein w and h are the target coding CU w×h Width and height of (ii), ori w×h (i, j) original luminance value, pre, of the target encoded CU w×h (i, j) is the target coding CU w×h The predicted brightness values i and j obtained after the inter-frame prediction mode calculation are respectively the target coding CU w×h The horizontal coordinate and the vertical coordinate of the pixel point, and QP is a quantization parameter of the current compressed coding frame;
for current CU w×h The areas are transversely and longitudinally divided into twoEqually dividing, horizontally obtaining two sub-matrixes, longitudinally obtaining two sub-matrixes, and solving the maximum Tra4 corresponding to the four sub-matrixes w×h (i, j) variance value to obtain maxVar B
For current CU w×h The region is divided into four parts horizontally and vertically to obtain four sub-matrixes horizontally and four sub-matrixes vertically to obtain the maximum Tra4 corresponding to the eight sub-matrixes w×h (i, j) variance value to obtain maxVar Q
Determining the target coding CU w×h Whether a fourth preset condition is met or not;
if yes, skipping the RDcost calculation of the rest inter-frame modes;
wherein the expression of the first preset condition is as follows:
POC >4& & POC! =8& & POC! =16 and
Figure FDA0003867661770000043
and is
Figure FDA0003867661770000044
The expression of the second preset condition is as follows:
Figure FDA0003867661770000045
and is provided with
Figure FDA0003867661770000046
The expression of the third preset condition is as follows:
POC >4& & POC! =8& & POC! =16 and
Figure FDA0003867661770000051
and is provided with
Figure FDA0003867661770000052
The expression of the fourth preset condition is as follows:
Figure FDA0003867661770000053
and is
Figure FDA0003867661770000054
In the formula
Figure FDA0003867661770000055
For adaptive adjustment of threshold value, and
Figure FDA0003867661770000056
Figure FDA0003867661770000057
in the formula
Figure FDA0003867661770000058
Respectively is to mix Tra w×h (i, j) dividing the matrix into two submatrices horizontally and then calculating the variance;
Figure FDA0003867661770000059
respectively is to get Tra w×h (i, j) longitudinally dividing the matrix into two sub-matrices and respectively calculating the variance;
Figure FDA00038676617700000510
respectively is to get Tra w×h (i, j) transversely dividing the matrix into four sub-matrixes and respectively calculating the obtained variances;
Figure FDA00038676617700000511
respectively is to mix Tra w×h (i, j) is longitudinally divided into four sub-matrixes and then respectively calculatedThe variance of (a) to (b), tra therein w×h (i, j) is Tra1 w×h (i, j) or Tra2 w×h (i, j) or Tra3 w×h (i, j) or Tra4 w×h (i,j)。
3. The method for fast determining the dynamic 3D point cloud compression interframe coding mode according to claim 2, wherein: after the process of judging whether the current dynamic 3D point cloud compression coding frame is not an I frame is carried out, the method further comprises the following steps: and if not, quickly determining the inter-frame prediction mode of the coded frame.
4. The method for fast determining the dynamic 3D point cloud compression interframe coding mode according to claim 2, wherein: after the process of determining whether the target coding CU satisfies the first preset condition, the method further includes: if yes, the target code skips the calculation of the residual inter prediction mode.
5. The method for fast determining the dynamic 3D point cloud compression interframe coding mode according to claim 2, wherein: after judging whether the target coding CU satisfies the second preset condition, the method further includes: if so, the target encoding skips the calculation of the residual inter prediction mode.
6. The method for fast determining the dynamic 3D point cloud compression interframe coding mode according to claim 2, wherein: after the process of determining whether the target CU satisfies the third preset condition, the method further includes: if not, the target coding continues to execute the calculation of the residual inter-frame prediction mode.
7. The method for fast determining the dynamic 3D point cloud compression interframe coding mode according to claim 2, wherein: after the process of determining whether the target coding CU satisfies the fourth preset condition, the method further includes: if not, the target coding continues to perform the calculation of the residual inter-frame prediction mode.
8. A dynamic 3D point cloud compression interframe coding mode fast determination system is characterized by comprising:
an encoding CU judging module: after the Merge mode rate-distortion cost of the target coding CU is calculated, judging whether the target coding CU is a geometric graph CU or a texture graph CU;
a first encoding mode determination module: the method comprises the steps of processing a target coding CU in a geometric graph CU form, wherein the processing comprises the steps of executing first prediction distortion transformation on the target coding CU, judging whether a first preset condition is met, and skipping the calculation of a subsequent inter-frame prediction mode if the first preset condition is met; when the target coding CU does not meet the first preset condition, after the calculation of the Inter2 Nx 2N mode of the frame is finished, executing third prediction distortion transformation on the target coding CU, judging whether the third preset condition is met, and if so, skipping the calculation of the subsequent frame mode;
a second encoding mode determination module: the method comprises the steps of processing a target coding CU in the form of a texture map CU, wherein the processing comprises executing second prediction distortion transformation on the target coding CU, judging whether a second preset condition is met, and skipping the calculation of a subsequent inter mode if the second preset condition is met; and when the target coding CU does not meet the second preset condition, after the calculation of the Inter2N multiplied by 2N mode between the frames is finished, executing fourth prediction distortion transformation on the target coding CU, judging whether the fourth preset condition is met, and if so, skipping the calculation of the subsequent Inter mode.
9. The system of claim 8, wherein the dynamic 3D point cloud compression interframe coding mode fast determination system comprises: the dynamic 3D point cloud compression interframe coding mode fast determination method of claims 2-7.
CN202211185785.5A 2022-09-27 2022-09-27 Method and system for rapidly determining dynamic 3D point cloud compression interframe coding mode Pending CN115499660A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202211185785.5A CN115499660A (en) 2022-09-27 2022-09-27 Method and system for rapidly determining dynamic 3D point cloud compression interframe coding mode

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202211185785.5A CN115499660A (en) 2022-09-27 2022-09-27 Method and system for rapidly determining dynamic 3D point cloud compression interframe coding mode

Publications (1)

Publication Number Publication Date
CN115499660A true CN115499660A (en) 2022-12-20

Family

ID=84471859

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202211185785.5A Pending CN115499660A (en) 2022-09-27 2022-09-27 Method and system for rapidly determining dynamic 3D point cloud compression interframe coding mode

Country Status (1)

Country Link
CN (1) CN115499660A (en)

Similar Documents

Publication Publication Date Title
CN102595140B (en) Intra-frame prediction video coding method based on image inpainting and vector prediction operators
CN109379594B (en) Video coding compression method, device, equipment and medium
CN102065298B (en) High-performance macroblock coding implementation method
CN106170093B (en) Intra-frame prediction performance improving coding method
WO2013143396A1 (en) Digital video quality control method and device thereof
TWI740801B (en) Picture encoding apparatus, dynamic picture decoding apparatus, and storage media
KR20190122615A (en) Method and Apparatus for image encoding
WO2021129007A1 (en) Method and device for determining video bitrate, computer apparatus, and storage medium
CN106303521B (en) A kind of HEVC Rate-distortion optimization method based on sensitivity of awareness
CN111741299B (en) Method, device and equipment for selecting intra-frame prediction mode and storage medium
CN109587491A (en) A kind of intra-frame prediction method, device and storage medium
CN110446052A (en) The quick CU depth selection method of depth map in a kind of 3D-HEVC frame
CN112188196A (en) Method for rapid intra-frame prediction of general video coding based on texture
CN101742301B (en) Block mode coding method and device
CN112153382B (en) Dynamic 3D point cloud compression rapid CU partitioning method and device and storage medium
CN108401185A (en) Reference frame selecting method, video transcoding method, electronic equipment and storage medium
CN112153381B (en) Method, device and medium for rapidly dividing CU in dynamic 3D point cloud compression frame
CN112001854A (en) Method for repairing coded image and related system and device
US8369423B2 (en) Method and device for coding
CN115499660A (en) Method and system for rapidly determining dynamic 3D point cloud compression interframe coding mode
CN114827606A (en) Quick decision-making method for coding unit division
CN108900842B (en) Y data compression processing method, device and equipment and WebP compression system
CN112509107A (en) Point cloud attribute recoloring method, device and encoder
CN111246218A (en) JND model-based CU partition prediction and mode decision texture coding method
CN113055670B (en) HEVC/H.265-based video coding method and system

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination