CN115499660A - Method and system for rapidly determining dynamic 3D point cloud compression interframe coding mode - Google Patents
Method and system for rapidly determining dynamic 3D point cloud compression interframe coding mode Download PDFInfo
- Publication number
- CN115499660A CN115499660A CN202211185785.5A CN202211185785A CN115499660A CN 115499660 A CN115499660 A CN 115499660A CN 202211185785 A CN202211185785 A CN 202211185785A CN 115499660 A CN115499660 A CN 115499660A
- Authority
- CN
- China
- Prior art keywords
- coding
- mode
- target coding
- preset condition
- frame
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/157—Assigned coding mode, i.e. the coding mode being predefined or preselected to be further used for selection of another element or parameter
- H04N19/159—Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/119—Adaptive subdivision aspects, e.g. subdivision of a picture into rectangular or non-rectangular coding blocks
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/146—Data rate or code amount at the encoder output
- H04N19/147—Data rate or code amount at the encoder output according to rate distortion criteria
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
Abstract
A method and a system for rapidly determining a dynamic 3D point cloud compression interframe coding mode relate to the technical field of point cloud compression. The invention predicts whether the optimal mode is the SKIP/Merge mode by judging whether the geometric picture frame/texture picture frame meets the first preset condition and the second preset condition, predicts whether the optimal mode is the Inter _2 Nx 2N mode by judging whether the geometric picture frame/texture picture frame meets the third preset condition and the fourth preset condition, and does not calculate the residual interframe prediction modes as long as the preset conditions are met, thereby achieving the purpose of saving time. Compared with the prior art, the method can ensure the coding quality of the target code, and can also obviously reduce the complexity of the calculation time in the process of making a coding mode decision on the target code.
Description
Technical Field
The invention relates to the technical field of point cloud compression, in particular to a method and a system for quickly determining a dynamic 3D point cloud compression interframe coding mode.
Background
With the recent development of visual capture technology, capture and digitization of three-dimensional world scenes has become more popular, where point clouds are one of the main 3D data representations and, in addition to providing spatial coordinates, provide attributes (such as color or reflectivity) associated with points in the 3D world. The point cloud in the original format requires a large amount of storage memory or transmission bandwidth. Classical dynamic point clouds as used for entertainment typically contain about 100 ten thousand points per frame, with a total bandwidth of 3.6Gbps at 30 frames per second, without compression. In addition, the advent of high resolution point cloud capture techniques has placed higher demands on the size of the point cloud in turn. In order to promote the point cloud technology and apply the point cloud technology to practical production, the compression technology is in urgent need of development.
In order to effectively compress dynamic 3D Point clouds, MPEG (Moving Picture Experts Group) is formulating a method (High Efficiency Video Coding, HEVC) for dynamic 3D Point Cloud Compression, i.e., video-based Point Cloud Compression V-PCC (Video-based Point Cloud Compression). In the HEVC compression Coding process, when Inter prediction mode selection is performed on each Coding Unit (CU), rate distortion cost (RDcost) calculations of modes such as SKIP/Merge mode, inter _2N × 2N mode, inter _ N × N mode, inter _2N × N mode, inter _ N × 2N mode, inter2N × nU mode, inter _2N × nD mode, inter _ nL × 2N mode, and Inter _ nR × 2N mode are performed in sequence, and finally, an optimal prediction mode is selected by comparing rdcosts. However, the selection of the optimal mode by calculating the RDcost also brings huge computational complexity to the compression coding, and seriously hinders the real-time transmission of the dynamic 3D point cloud. Furthermore, due to the very large difference between the geometry and color video generated by the dynamic 3D point cloud mapping and the natural video, this may result in that the proposed method for fast determination of HEVC inter-frame coding mode for the natural video cannot be well used in video-based dynamic 3D point cloud compression coding optimization. Therefore, a method for rapidly determining an inter-frame coding mode in dynamic 3D point cloud compression is urgently needed to be developed so as to solve the problem of network real-time transmission and improve user experience.
Therefore, how to reduce the computational complexity of inter-frame mode selection in the video-based dynamic 3D point cloud compression process is a technical problem that needs to be solved by those skilled in the art.
Disclosure of Invention
One objective of the present invention is to provide a method for quickly determining a dynamic 3D point cloud compression inter-frame coding mode to reduce the computational complexity in mode selection.
In order to solve the technical problems, the invention adopts the following technical scheme: a method for rapidly determining a dynamic 3D point cloud compression interframe coding mode comprises the following steps of after a target coding CU calculates the rate distortion cost of a Merge mode, judging the target coding CU as follows:
if the target coding CU is the geometric figure CU, executing first prediction distortion transformation on the target coding CU, judging whether a first preset condition is met, and if so, skipping the calculation of a subsequent inter-frame prediction mode; and when the target coding CU does not meet the first preset condition, after the calculation of the Inter2 Nx 2N mode is finished, executing third prediction distortion transformation on the target coding CU, judging whether the third preset condition is met, and if so, skipping the calculation of the subsequent Inter mode.
If the target coding CU is the texture map CU, performing second prediction distortion transformation on the target coding CU, judging whether a second preset condition is met, and if so, skipping the calculation of a subsequent inter-frame mode; and when the target coding CU does not meet the second preset condition, after the calculation of the Inter2 Nx 2N mode is finished, executing fourth prediction distortion transformation on the target coding CU, judging whether the fourth preset condition is met, and if so, skipping the calculation of the subsequent Inter mode.
In the method, specifically, in a dynamic 3D point cloud compressed intra-frame compression coding configuration, after a rate distortion cost is calculated in a Merge mode, it is determined whether a current dynamic 3D point cloud compressed coding frame does not belong to a placeholder image frame and is not an I frame;
if yes, sequentially extracting CU needing compression coding w×h Obtaining the target coding CU w×h Judging a geometric map and a texture map by the chrominance component;
obtaining a target coding CU w×h Pre-encoding of luminance componentsMeasuring distortion Err w×h (i,j);
If the current coding CU w×h Is a geometric frame, then the prediction distortion Err is corrected w×h (i, j) performing a first predictive distortion transformation to obtain Tra1 w×h (i,j);
wherein w and h are the target coding CU respectively w×h Width and height of (ii), ori w×h (i, j) is the target coding CU w×h Of the original luminance value, pre w×h (i, j) encoding the CU for the target w×h The predicted brightness values i and j obtained after the inter-frame prediction mode calculation are respectively the target coding CU w×h The abscissa and the ordinate of the pixel point of (1), and QP is a quantization parameter of the current compressed coding frame;
for current CU w×h The region is divided into two parts horizontally and vertically to obtain two sub-matrixes horizontally and two sub-matrixes vertically to obtain the maximum Tra1 corresponding to the four sub-matrixes w×h (i, j) variance value to obtain maxVar B ;
For current CU w×h The region is divided into four parts horizontally and vertically to obtain four sub-matrixes horizontally and four sub-matrixes vertically to obtain the maximum Tra1 corresponding to the eight sub-matrixes w×h (i, j) variance value to obtain maxVar Q (ii) a Determining the target coding CU w×h Whether a first preset condition is met or not;
if not, acquiring the target coding CU after the Inter2 Nx 2N prediction mode w×h Predicted distortion Err of (1) w×h (i, j) on which a third predictive distortion transform is performed, resulting in Tra3 w×h (i,j);
wherein w and h are the target coding CU w×h Width and height of (ii), ori w×h (i, j) original luminance value, pre, of the target encoded CU w×h (i, j) is the target coding CU w×h The predicted brightness values i and j obtained after the inter-frame prediction mode calculation are respectively the target coding CU w×h The abscissa and the ordinate of the pixel point of (1), and QP is a quantization parameter of the current compressed coding frame;
for the current CU w×h The region is divided into two parts horizontally and vertically to obtain two sub-matrixes horizontally and two sub-matrixes vertically to obtain the maximum Tra3 corresponding to the four sub-matrixes w×h (i, j) variance value to obtain maxVar B ;
For current CU w×h The area is divided into four parts horizontally and vertically, four sub-matrixes are obtained horizontally and four sub-matrixes are obtained vertically, and the maximum Tra3 corresponding to the eight sub-matrixes is obtained w×h (i, j) variance value to obtain maxVar Q (ii) a Determining the target coding CU w×h Whether a third preset condition is met;
if yes, skipping the calculation of the residual inter-frame prediction mode;
if the current coding CU w×h If it is a texture map frame, then a second prediction distortion transformation is performed on it to obtain Tra2 w×h (i,j);
wherein w and h are the target coding CU respectively w×h Width and height of (ii), ori w×h (i, j) is the target coding CU w×h Of original brightnessValue of, pre w×h (i, j) encoding the CU for the target w×h The predicted brightness values i and j obtained after the inter-frame prediction mode calculation are respectively the target coding CU w×h The abscissa and the ordinate of the pixel point of (1), and QP is a quantization parameter of the current compressed coding frame;
for current CU w×h The region is divided into two parts horizontally and vertically to obtain two sub-matrixes horizontally and two sub-matrixes vertically to obtain the maximum Tra2 corresponding to the four sub-matrixes w×h (i, j) variance value to obtain maxVar B ;
For current CU w×h The region is divided into four parts horizontally and vertically to obtain four sub-matrixes horizontally and four sub-matrixes vertically to obtain the maximum Tra2 corresponding to the eight sub-matrixes w×h (i, j) variance value to obtain maxVar Q (ii) a Determining the target coding CU w×h Whether a second preset condition is met or not;
if not, the prediction distortion Err is corrected after the Inter2N × 2N prediction mode w×h (i, j) performing a third predictive distortion transform to obtain Tra4 w×h (i,j);
wherein w and h are the target coding CU respectively w×h Width and height of (1), ori w×h (i, j) is the original luminance value, pre, of the target encoded CU w×h (i, j) is the target coding CU w×h The predicted brightness values i and j obtained after the inter-frame prediction mode calculation are respectively the target coding CU w×h The abscissa and the ordinate of the pixel point of (1), and QP is a quantization parameter of the current compressed coding frame;
for current CU w×h The region is divided into two parts horizontally and vertically to obtain two sub-matrixes horizontally and two sub-matrixes vertically, and the pairs in the four sub-blocks are obtainedMaximum of response Tra4 w×h (i, j) variance value to obtain maxVar B ;
For the current CU w×h The region is divided into four parts horizontally and vertically to obtain four sub-matrixes horizontally and four sub-matrixes vertically to obtain the maximum Tra4 corresponding to the eight sub-matrixes w×h (i, j) variance value to obtain maxVar Q (ii) a Determining the target coding CU w×h Whether a fourth preset condition is met or not;
if yes, skipping the RDcost calculation of the residual inter-frame mode;
wherein the expression of the first preset condition is as follows:
POC >4& & POC! =8& & POC! =16 and
The expression of the second preset condition is as follows:
The expression of the third preset condition is as follows:
POC >4& & POC! =8& & POC! =16 and
The expression of the fourth preset condition is as follows:
In the formulaRespectively is to get Tra w×h (i, j) dividing the matrix into two submatrices horizontally and then calculating the variance;respectively is to mix Tra w×h (i, j) longitudinally dividing the matrix into two sub-matrices and respectively calculating the variance;respectively is to mix Tra w×h (i, j) transversely dividing the matrix into four sub-matrixes and respectively calculating the obtained variances;respectively is to mix Tra w×h (i, j) the variance calculated after longitudinally dividing the matrix into four submatrices, wherein Tra w×h (i, j) is Tra1 w×h (i, j) or Tra2 w×h (i, j) or Tra3 w×h (i, j) or Tra4 w×h (i,j)。
Preferably, after the process of determining whether the current dynamic 3D point cloud compression encoding frame is not an I frame, the method further includes: if not, the inter-frame prediction mode is not rapidly determined for the coding frame.
More preferably, after the process of determining whether the target coding CU satisfies the first preset condition, the method further includes: if so, the target encoding skips the calculation of the residual inter prediction mode.
More preferably, after determining whether the target coding CU satisfies the second preset condition, the method further includes: if yes, the target code skips the calculation of the residual inter prediction mode.
More preferably, the determining whether the target coding CU satisfies the third preset condition further includes: if not, the target coding continues to execute the calculation of the residual inter-frame prediction mode.
More preferably, after the process of determining whether the target coding CU satisfies the fourth preset condition, the method further includes: if not, the target coding continues to execute the calculation of the residual inter-frame prediction mode.
In addition, the invention also provides a system for rapidly determining the dynamic 3D point cloud compression interframe coding mode, which comprises the following steps:
an encoding CU judging module: and the processing unit is used for judging whether the target coding CU is the geometric map CU or the texture map CU or not after calculating the Merge mode rate distortion cost of the target coding CU.
A first encoding mode determination module: the method comprises the steps of processing a target coding CU in a geometric figure CU form, wherein the processing comprises the steps of executing first prediction distortion transformation on the target coding CU, judging whether a first preset condition is met, and skipping the calculation of a subsequent inter-frame prediction mode if the first preset condition is met; and when the target coding CU does not meet the first preset condition, after the calculation of the Inter2N multiplied by 2N mode between the frames is finished, executing third prediction distortion transformation on the target coding CU, judging whether the third preset condition is met, and if so, skipping the calculation of the subsequent Inter mode.
A second encoding mode determination module: the method comprises the steps of processing a target coding CU in the form of a texture map CU, wherein the processing comprises executing second prediction distortion transformation on the target coding CU, judging whether a second preset condition is met, and skipping the calculation of a subsequent inter mode if the second preset condition is met; and when the target coding CU does not meet the second preset condition, after the calculation of the Inter2 Nx 2N mode is finished, executing fourth prediction distortion transformation on the target coding CU, judging whether the fourth preset condition is met, and if so, skipping the calculation of the subsequent Inter mode.
The system can be operated by the method for quickly determining the dynamic 3D point cloud compression interframe coding mode.
The method predicts the optimal mode by judging whether the geometric picture frame/texture picture frame meets a first preset condition or not; similarly, the best mode can be predicted by determining whether the geometric frame/texture frame satisfies the third and fourth predetermined conditions. The calculation of the residual inter prediction mode may not be performed as long as the preset condition is satisfied, thereby achieving the purpose of saving time. Therefore, the method provided by the invention not only can ensure the coding quality of the target code, but also can obviously reduce the complexity of the calculation time in the process of carrying out compression coding on the target code.
Drawings
FIG. 1 is a schematic flow chart of a method for rapidly determining a dynamic 3D point cloud compression interframe coding mode according to an embodiment of the invention;
FIG. 2 is a binary division example and depicts the variance values in the formula of the present invention
Detailed Description
In order to facilitate understanding of those skilled in the art, the present invention will be further described with reference to the following examples and drawings, which are not intended to limit the present invention.
As shown in fig. 1, a method for rapidly determining a dynamic 3D point cloud compression inter-frame coding mode includes the following steps:
step S1: judging whether the current dynamic 3D point cloud compression coding frame is an I frame, if so, skipping any inter-frame mode, otherwise, executing S2;
step S2: sequentially extracting CUs needing compression coding, and acquiring chrominance component judgment geometric graphs and texture graphs of target coding CUs;
and step S3: judging whether the current frame is a geometric picture frame, if so, executing a step S4, otherwise, executing a step S7;
and step S4: computing the residual Err of the original and predicted images w×h (i, j) and performing a first predictive distortion transformation to obtain Tra1 w×h (i,j);
Step S5: for Tra1 w×h (i, j) carrying out horizontal and longitudinal halving and quartering in the corresponding CU region, and respectively calculating the maximum variance of the subblocks corresponding to the halving and the quartering to obtain maxVar B And maxVar Q Judging maxVar B And maxVar Q Whether a first preset condition is met, if yes, executing a step S6, and if not, executing a step S9;
step S6: skipping the residual inter prediction mode;
step S7: computing the residual Err of the original and predicted images w×h (i, j) and performing a second predictive distortion transformation to obtain Tra2 w×h (i,j);
Step S8: for Tra2 w×h (i, j) performing horizontal and longitudinal halving and quartering on the corresponding CU region, and respectively calculating the maximum variance of the subblocks corresponding to the halving and the quartering to obtain maxVar B And maxVar Q (ii) a Judging maxVar B And maxVar Q Whether a second preset condition is met, if so, executing a step S6, otherwise, executing a step S12;
step S9: performing a third predictive distortion transformation on the geometry to obtain Tra3 w×h (i,j);
Step S10: for Tra3 w×h (i, j) performing horizontal and longitudinal halving and quartering on the corresponding CU region, and respectively calculating the maximum variance of the subblocks corresponding to the halving and the quartering to obtain maxVar B And maxVar Q (ii) a Judging maxVar B And maxVar Q If the third preset condition is met, executing the step S11 if the third preset condition is met, otherwise, continuing to perform residual inter-frame predictionMeasuring a mode;
step S11: skipping the residual inter prediction mode;
step S12: performing a fourth predictive distortion transformation on the geometry to obtain Tra4 w×h (i,j);
Step S13: for Tra4 w×h (i, j) performing horizontal and longitudinal halving and quartering on the corresponding CU region, and respectively calculating the maximum variance of the subblocks corresponding to the halving and the quartering to obtain maxVar B And maxVar Q (ii) a Judging maxVar B And maxVar Q Whether a fourth preset condition is met, if yes, executing the step S11, and otherwise, continuing to perform the residual inter-frame prediction mode;
where w and h are the target coding CU w×h Width and height of (1), ori w×h (i, j) is the original luminance value, pre, of the target coded CU w×h (i, j) is the target coding CU w×h The predicted brightness values i and j obtained after the inter-frame prediction mode calculation are respectively target coding CU w×h The abscissa and ordinate of the pixel point of (1), QP is the quantization parameter of the current compressed encoded frame, hereinafter Tra3 w×h (i, j) and Tra4 w×h (i, j) are collectively referred to as Tra w×h (i,j);
The first preset conditional expression is as follows:
POC >4& & POC! =8& & POC! =16 and
The second preset conditional expression is as follows:
The third preset conditional expression is as follows:
POC >4& & POC! =8& & POC! =16 and
The fourth preset conditional expression is as follows:
in the formulaRespectively is to get Tra w×h (i, j) dividing the matrix into two submatrices horizontally and then calculating the variance;respectively is to get Tra w×h (i, j) longitudinally dividing the matrix into two sub-matrices and respectively calculating the variance;respectively is to mix Tra w×h (i, j) dividing the matrix into four sub-matrixes horizontally and then calculating the variance;respectively is to mix Tra w×h (i, j) dividing the matrix into four submatrices longitudinally and then calculating the variance.
S4-S8 are utilized to predict whether the optimal mode is the SKIP/Merge mode by judging whether the geometric picture frame/texture picture frame meets the first preset condition and the second preset condition; similarly, the steps S9-S13 can predict whether the optimal mode is the Inter _2N × 2N mode by determining whether the geometric frame/texture frame satisfies the third and fourth predetermined conditions. The calculation of the residual inter prediction mode may not be performed as long as the preset condition is satisfied, thereby achieving the purpose of saving time. Therefore, the method provided by the embodiment can ensure that the target coding CU can not only be ensured w×h Can also significantly reduce the coding quality of the target CU w×h The computational time complexity in the coding mode decision process is performed.
It can be seen that in this embodiment, the CU is currently encoded w×h Is predicted based on the transformed value Tra1 of the prediction distortion w×h (i, j) can be reversedMapping a current coding CU w×h The best inter prediction mode should be selected. In coding a target CU w×h In the process of inter-frame prediction, four judgment standards are set to judge the target coding CU w×h Whether the residual inter prediction mode can be skipped, i.e., by determining the target coding CU w×h Whether the predicted distortion transformation value satisfies a first preset condition, a second preset condition, a third preset condition and a fourth preset condition to determine the target coding CU w×h Whether the calculation of the residual inter prediction mode can be skipped, and the target coding CU is coded w×h In the process of dividing, the threshold value can be adjusted in a self-adaptive wayTo achieve a balance between savings in coding time and increased coding rate. Obviously, the target coding CU can be terminated early due to the method w×h Since various inter-frame prediction modes are performed, the method for quickly determining the inter-frame coding mode provided by the embodiment can significantly reduce the time spent on encoding the target CU w×h Temporal complexity in the inter mode selection process is performed.
On the basis of the above, the present embodiment further describes and optimizes the technical solution, and as a preferred implementation, in the step S1: and if the current dynamic 3D point cloud compressed coding frame is the I frame, the method does not perform the rapid mode determination on the current coding frame.
It should be appreciated by those skilled in the art that in the dynamic 3D point cloud compression interframe compression coding configuration, both the placeholder map frame and the geometric map frame and the texture map frame are included, and both the I frame, the P frame and the B frame are included, and the proposed method is a method for performing fast mode determination only for the B frame of the geometric map and the texture map.
By the method provided by the embodiment, the CU can be further ensured to be coded on the target CU w×h Integrity in the mode determination process is performed.
Based on the technical content disclosed by the above embodiment, dynamic 3D point cloud compression reference software TCM2-V7.0 is taken as a test platform, and the corresponding HEVC reference software is HM-16.20+ scm-8.8. The rapid determination method disclosed in the present embodiment is executed on a PC of Inter (R) Core (TM) i7_9700 cpu, 169g RAM, and thus the feasibility and effectiveness of the method are evaluated. The dynamic 3D point cloud sequence of the universal test includes "root", "readandblack", "soldier", "queen", "longaddress", and each dynamic 3D point cloud includes 32 frames. The encoding configuration is Random Access (RA) and employs five combinations of encoding Quantization Parameters (QP), (32, 42), (28, 37), (24, 32), (20, 27), (16, 22), respectively, for geometric video and texture video compression in dynamic 3D point clouds. For the evaluation of the geometric video compression performance of the dynamic 3D point cloud, the code rate change conditions (BD-rates) of point-to-point PSNR (D1) and point-to-plane PSNR (D2) are adopted, and for the evaluation of the color video compression performance of the dynamic 3D point cloud, the BD-rates of Luma, cb and Cr are adopted. The coding time savings are defined as:
wherein i represents 5 different QP values, T O For compression of the original test pattern, T P For compression coding time, TS, after the invention is applied to the original test model GP And TS TP The time for compressing and encoding the geometric video P frame of the dynamic 3D point cloud is saved, and the time for compressing and encoding the texture video P frame is saved.
Please refer to table 1, wherein table 1 shows the performance comparison results of the method of this embodiment on TCM2-V7.0 testing platform, wherein
TABLE 1 Performance comparison of the methods provided in this example on the TCM2-V7.0 test platform (unit:%)
Please refer to table 2, table 2 is a comparison result of the performance of the method provided in this embodiment on the TCM2-V7.0 testing platform, wherein
TABLE 2 Performance comparison of the methods provided in this example on the TCM2-V7.0 test platform (unit:%)
Please refer to table 3, table 3 shows the performance comparison results of the method of this embodiment on the TCM2-V7.0 testing platform, wherein
TABLE 3 Performance comparison of the methods provided in this example on the TCM2-V7.0 test platform (unit:%)
Please refer to table 4, table 4 shows the performance comparison results of the method of this embodiment on the TCM2-V7.0 testing platform, wherein
TABLE 4 Performance comparison of the methods provided in this example on the TCM2-V7.0 test platform (unit:%)
As can be seen from the comparison of BD-rates and encoding time savings for geometry video and texture video for dynamic 3D point clouds in tables 1, 2, 3 and 4, the adaptive adjustment threshold is usedFor different values, the method provided by the embodiment can save the geometric video and color video coding time (39.91%, 47.48%), (45.47%, 50.95%), (51.41%, 58.54%) and (55.51%, 64.13%) respectively on average, while the BD-rates (D1 and D2) of the geometric video and the BD-rates (Luma) of the texture video increase on average (-0.18%, -0.25%, 0.09%), (-0.30%, -0.39%, 0.20%), (-0.13%, -0.24%, 0.58%) and (0.21%, 0.05%, 1.45%) respectively. With adaptive adjustment of the thresholdThe more time savings are made in the compression encoding of the geometric and color videos of the dynamic 3D point cloud, and the BD-rates of D1, D2 and Luma increase.
In conclusion, the rapid determination method provided by the invention can greatly reduce the computational complexity of the inter-frame prediction mode under the condition of hardly reducing the compression quality of the dynamic 3D point cloud compression geometry and the texture video. Meanwhile, the method is simple in design, does not refer to CU time-space domain information, and can be integrated into a parallel compression coding framework to further reduce the dynamic 3D point cloud compression coding time.
The method for rapidly determining the dynamic 3D point cloud compression interframe coding mode provided by the invention is described in detail above, and specific examples are applied in the text to explain the principle and the implementation of the invention, and the description of the above embodiments is only used to help understanding the method and the core idea of the invention; meanwhile, for a person skilled in the art, according to the idea of the present invention, there may be variations in the specific embodiments and the application scope, and in summary, the content of the present specification should not be construed as a limitation to the present invention.
Claims (9)
1. The method for rapidly determining the dynamic 3D point cloud compression interframe coding mode is characterized by comprising the following steps of: after the target coding CU calculates the Merge mode rate-distortion cost, the target coding CU is judged as follows:
if the target coding CU is the geometric graph CU, performing first prediction distortion transformation on the target coding CU, judging whether a first preset condition is met, and if so, skipping the calculation of a subsequent inter-frame prediction mode; when the target coding CU does not meet the first preset condition, after the calculation of the Inter2 Nx 2N mode between the frames is finished, executing third prediction distortion transformation on the target coding CU, judging whether the third preset condition is met, and if so, skipping the calculation of the subsequent Inter mode;
if the target coding CU is the texture map CU, performing second prediction distortion transformation on the target coding CU, judging whether a second preset condition is met, and if so, skipping the calculation of a subsequent inter-frame mode; and when the target coding CU does not meet the second preset condition, after the calculation of the Inter2 Nx 2N mode is finished, executing fourth prediction distortion transformation on the target coding CU, judging whether the fourth preset condition is met, and if so, skipping the calculation of the subsequent Inter mode.
2. The method for fast determination of inter-frame coding mode for dynamic 3D point cloud compression according to claim 1, wherein: in the dynamic 3D point cloud compressed intra-frame compression coding configuration, after the rate distortion cost is calculated in the Merge mode, whether the current dynamic 3D point cloud compressed coding frame does not belong to a placeholder image frame or not and whether the current dynamic 3D point cloud compressed coding frame does not belong to an I frame or not is judged;
if both are, sequentially extracting CU requiring compression coding w×h Obtaining the target coding CU w×h Judging a geometric map and a texture map by the chrominance component;
obtaining a target coding CU w×h Prediction distortion Err of luminance component w×h (i,j);
If the current coding CU w×h Is a geometric picture frame, then the predicted distortion Err is corrected w×h (i, j) performing a first predictive distortion transformation to obtain Tra1 w×h (i,j);
wherein w and h are the target coding CU respectively w×h Width and height of (ii), ori w×h (i, j) is the target coding CU w×h Of the original brightness value, pre w×h (i, j) encoding the CU for the target w×h The predicted brightness values i and j obtained after the inter-frame prediction mode calculation are respectively the target coding CU w×h The abscissa and the ordinate of the pixel point of (1), and QP is a quantization parameter of the current compressed coding frame;
for current CU w×h The region is divided into two parts horizontally and vertically to obtain two sub-matrixes horizontally and two sub-matrixes vertically to obtain the maximum Tra1 corresponding to the four sub-matrixes w×h (i, j) variance value to obtain maxVar B ;
For current CU w×h The region is divided into four parts horizontally and vertically to obtain four sub-matrixes horizontally and four sub-matrixes vertically to obtain the maximum Tra1 corresponding to the eight sub-matrixes w×h (i, j) variance value to obtain maxVar Q ;
Determining the target coding CU w×h Whether a first preset condition is met or not;
if not, acquiring the target coding CU after the Inter2N multiplied by 2N prediction mode w×h Predicted distortion Err of (1) w×h (i, j) on which a third predictive distortion transform is performed, resulting in Tra3 w×h (i,j);
wherein w and h are the target coding CU respectively w×h Width and height of (ii), ori w×h (i, j) original luminance value, pre, of the target encoded CU w×h (i, j) is the target coding CU w×h The predicted brightness values i and j obtained after the inter-frame prediction mode calculation are respectively the target coding CU w×h The abscissa and the ordinate of the pixel point of (1), and QP is a quantization parameter of the current compressed coding frame;
for the current CU w×h The region is divided into two parts horizontally and vertically to obtain two sub-matrixes horizontally and two sub-matrixes vertically to obtain the maximum Tra3 corresponding to the four sub-matrixes w×h (i, j) variance value to obtain maxVar B ;
For current CU w×h The area is divided into four parts horizontally and vertically, four sub-matrixes are obtained horizontally and four sub-matrixes are obtained vertically, and the maximum Tra3 corresponding to the eight sub-matrixes is obtained w×h (i, j) variance value to obtain maxVar Q ;
Determining the target coding CU w×h Whether a third preset condition is met;
if yes, skipping the calculation of the residual inter-frame prediction mode;
if the current coding CU w×h If it is a texture map frame, then a second prediction distortion transformation is performed on it to obtain Tra2 w×h (i,j);
wherein w and h are the target coding CU w×h Width and height of (1), ori w×h (i, j) is the target coding CU w×h Of the original brightness value, pre w×h (i, j) encoding the CU for the target w×h The predicted brightness values i and j obtained after the inter-frame prediction mode calculation are respectively the target coding CU w×h The abscissa and the ordinate of the pixel point of (1), and QP is a quantization parameter of the current compressed coding frame;
for current CU w×h The region is divided into two parts horizontally and vertically to obtain two sub-matrixes horizontally and two sub-matrixes vertically to obtain the maximum Tra2 corresponding to the four sub-matrixes w×h (i, j) variance value to obtain maxVar B ;
For the current CU w×h The area is divided into four parts horizontally and vertically to obtain four sub-matrixes horizontally and four sub-matrixes vertically, and the maximum Tra2 corresponding to the eight sub-matrixes is obtained w×h (i, j) variance value to obtain maxVar Q ;
Determining the target coding CU w×h Whether a second preset condition is met or not;
if not, the prediction distortion Err is corrected after the Inter2N × 2N prediction mode w×h (i, j) performing a third predictive distortion transformation to obtain Tra4 w×h (i,j);
wherein w and h are the target coding CU w×h Width and height of (ii), ori w×h (i, j) original luminance value, pre, of the target encoded CU w×h (i, j) is the target coding CU w×h The predicted brightness values i and j obtained after the inter-frame prediction mode calculation are respectively the target coding CU w×h The horizontal coordinate and the vertical coordinate of the pixel point, and QP is a quantization parameter of the current compressed coding frame;
for current CU w×h The areas are transversely and longitudinally divided into twoEqually dividing, horizontally obtaining two sub-matrixes, longitudinally obtaining two sub-matrixes, and solving the maximum Tra4 corresponding to the four sub-matrixes w×h (i, j) variance value to obtain maxVar B ;
For current CU w×h The region is divided into four parts horizontally and vertically to obtain four sub-matrixes horizontally and four sub-matrixes vertically to obtain the maximum Tra4 corresponding to the eight sub-matrixes w×h (i, j) variance value to obtain maxVar Q ;
Determining the target coding CU w×h Whether a fourth preset condition is met or not;
if yes, skipping the RDcost calculation of the rest inter-frame modes;
wherein the expression of the first preset condition is as follows:
POC >4& & POC! =8& & POC! =16 and
The expression of the second preset condition is as follows:
The expression of the third preset condition is as follows:
POC >4& & POC! =8& & POC! =16 and
The expression of the fourth preset condition is as follows:
in the formulaRespectively is to mix Tra w×h (i, j) dividing the matrix into two submatrices horizontally and then calculating the variance;respectively is to get Tra w×h (i, j) longitudinally dividing the matrix into two sub-matrices and respectively calculating the variance;respectively is to get Tra w×h (i, j) transversely dividing the matrix into four sub-matrixes and respectively calculating the obtained variances;respectively is to mix Tra w×h (i, j) is longitudinally divided into four sub-matrixes and then respectively calculatedThe variance of (a) to (b), tra therein w×h (i, j) is Tra1 w×h (i, j) or Tra2 w×h (i, j) or Tra3 w×h (i, j) or Tra4 w×h (i,j)。
3. The method for fast determining the dynamic 3D point cloud compression interframe coding mode according to claim 2, wherein: after the process of judging whether the current dynamic 3D point cloud compression coding frame is not an I frame is carried out, the method further comprises the following steps: and if not, quickly determining the inter-frame prediction mode of the coded frame.
4. The method for fast determining the dynamic 3D point cloud compression interframe coding mode according to claim 2, wherein: after the process of determining whether the target coding CU satisfies the first preset condition, the method further includes: if yes, the target code skips the calculation of the residual inter prediction mode.
5. The method for fast determining the dynamic 3D point cloud compression interframe coding mode according to claim 2, wherein: after judging whether the target coding CU satisfies the second preset condition, the method further includes: if so, the target encoding skips the calculation of the residual inter prediction mode.
6. The method for fast determining the dynamic 3D point cloud compression interframe coding mode according to claim 2, wherein: after the process of determining whether the target CU satisfies the third preset condition, the method further includes: if not, the target coding continues to execute the calculation of the residual inter-frame prediction mode.
7. The method for fast determining the dynamic 3D point cloud compression interframe coding mode according to claim 2, wherein: after the process of determining whether the target coding CU satisfies the fourth preset condition, the method further includes: if not, the target coding continues to perform the calculation of the residual inter-frame prediction mode.
8. A dynamic 3D point cloud compression interframe coding mode fast determination system is characterized by comprising:
an encoding CU judging module: after the Merge mode rate-distortion cost of the target coding CU is calculated, judging whether the target coding CU is a geometric graph CU or a texture graph CU;
a first encoding mode determination module: the method comprises the steps of processing a target coding CU in a geometric graph CU form, wherein the processing comprises the steps of executing first prediction distortion transformation on the target coding CU, judging whether a first preset condition is met, and skipping the calculation of a subsequent inter-frame prediction mode if the first preset condition is met; when the target coding CU does not meet the first preset condition, after the calculation of the Inter2 Nx 2N mode of the frame is finished, executing third prediction distortion transformation on the target coding CU, judging whether the third preset condition is met, and if so, skipping the calculation of the subsequent frame mode;
a second encoding mode determination module: the method comprises the steps of processing a target coding CU in the form of a texture map CU, wherein the processing comprises executing second prediction distortion transformation on the target coding CU, judging whether a second preset condition is met, and skipping the calculation of a subsequent inter mode if the second preset condition is met; and when the target coding CU does not meet the second preset condition, after the calculation of the Inter2N multiplied by 2N mode between the frames is finished, executing fourth prediction distortion transformation on the target coding CU, judging whether the fourth preset condition is met, and if so, skipping the calculation of the subsequent Inter mode.
9. The system of claim 8, wherein the dynamic 3D point cloud compression interframe coding mode fast determination system comprises: the dynamic 3D point cloud compression interframe coding mode fast determination method of claims 2-7.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202211185785.5A CN115499660A (en) | 2022-09-27 | 2022-09-27 | Method and system for rapidly determining dynamic 3D point cloud compression interframe coding mode |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202211185785.5A CN115499660A (en) | 2022-09-27 | 2022-09-27 | Method and system for rapidly determining dynamic 3D point cloud compression interframe coding mode |
Publications (1)
Publication Number | Publication Date |
---|---|
CN115499660A true CN115499660A (en) | 2022-12-20 |
Family
ID=84471859
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202211185785.5A Pending CN115499660A (en) | 2022-09-27 | 2022-09-27 | Method and system for rapidly determining dynamic 3D point cloud compression interframe coding mode |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN115499660A (en) |
-
2022
- 2022-09-27 CN CN202211185785.5A patent/CN115499660A/en active Pending
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN102595140B (en) | Intra-frame prediction video coding method based on image inpainting and vector prediction operators | |
CN109379594B (en) | Video coding compression method, device, equipment and medium | |
CN102065298B (en) | High-performance macroblock coding implementation method | |
CN106170093B (en) | Intra-frame prediction performance improving coding method | |
WO2013143396A1 (en) | Digital video quality control method and device thereof | |
TWI740801B (en) | Picture encoding apparatus, dynamic picture decoding apparatus, and storage media | |
KR20190122615A (en) | Method and Apparatus for image encoding | |
WO2021129007A1 (en) | Method and device for determining video bitrate, computer apparatus, and storage medium | |
CN106303521B (en) | A kind of HEVC Rate-distortion optimization method based on sensitivity of awareness | |
CN111741299B (en) | Method, device and equipment for selecting intra-frame prediction mode and storage medium | |
CN109587491A (en) | A kind of intra-frame prediction method, device and storage medium | |
CN110446052A (en) | The quick CU depth selection method of depth map in a kind of 3D-HEVC frame | |
CN112188196A (en) | Method for rapid intra-frame prediction of general video coding based on texture | |
CN101742301B (en) | Block mode coding method and device | |
CN112153382B (en) | Dynamic 3D point cloud compression rapid CU partitioning method and device and storage medium | |
CN108401185A (en) | Reference frame selecting method, video transcoding method, electronic equipment and storage medium | |
CN112153381B (en) | Method, device and medium for rapidly dividing CU in dynamic 3D point cloud compression frame | |
CN112001854A (en) | Method for repairing coded image and related system and device | |
US8369423B2 (en) | Method and device for coding | |
CN115499660A (en) | Method and system for rapidly determining dynamic 3D point cloud compression interframe coding mode | |
CN114827606A (en) | Quick decision-making method for coding unit division | |
CN108900842B (en) | Y data compression processing method, device and equipment and WebP compression system | |
CN112509107A (en) | Point cloud attribute recoloring method, device and encoder | |
CN111246218A (en) | JND model-based CU partition prediction and mode decision texture coding method | |
CN113055670B (en) | HEVC/H.265-based video coding method and system |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |