CN111464805B - Three-dimensional panoramic video rapid coding method based on panoramic saliency - Google Patents

Three-dimensional panoramic video rapid coding method based on panoramic saliency Download PDF

Info

Publication number
CN111464805B
CN111464805B CN202010175774.3A CN202010175774A CN111464805B CN 111464805 B CN111464805 B CN 111464805B CN 202010175774 A CN202010175774 A CN 202010175774A CN 111464805 B CN111464805 B CN 111464805B
Authority
CN
China
Prior art keywords
coding unit
current
frame
current frame
coded
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202010175774.3A
Other languages
Chinese (zh)
Other versions
CN111464805A (en
Inventor
杜宝祯
蒋刚毅
郁梅
郑雪蕾
晁涌
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Yiqilin Cultural Media Co.,Ltd.
Nantian Shujin (Beijing) Information Industry Development Co.,Ltd.
Shanghai Ruishenglian Information Technology Co ltd
Original Assignee
Ningbo University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Ningbo University filed Critical Ningbo University
Priority to CN202010175774.3A priority Critical patent/CN111464805B/en
Publication of CN111464805A publication Critical patent/CN111464805A/en
Application granted granted Critical
Publication of CN111464805B publication Critical patent/CN111464805B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/161Encoding, multiplexing or demultiplexing different image signal components

Abstract

The invention discloses a panoramic saliency-based stereoscopic panoramic video rapid coding method, which is used for calculating the saliency of a right viewpoint video frame in a stereoscopic panoramic video; dividing the rest maximum coding units except the maximum coding unit at the leftmost side and the topmost side in the right viewpoint video frame into a significant block and an insignificant block by combining the significance calculation result, respectively adopting two different coding fast termination modes aiming at the significant block and the insignificant block, and predicting and correcting the recursion depth interval of the current coding unit by using the optimal division depth of the adjacent block of the current coding unit in the time-space domain aiming at the insignificant block; for the significant block, judging whether the current coding unit reaches the optimal division depth or not by calculating and comparing the size relation between the root mean square error of the current coding unit and the division threshold value of the coding unit based on the panoramic perception distortion; the advantages are that the recursion complexity of the coding unit is effectively reduced and the coding time is saved.

Description

Three-dimensional panoramic video rapid coding method based on panoramic saliency
Technical Field
The invention relates to a video coding technology, in particular to a stereo panoramic video fast coding method based on panoramic saliency.
Background
In recent years, with the rapid development of multimedia networks, stereoscopic panoramic video systems are increasingly popular with people with brand new visual experiences. The stereoscopic panoramic video system can provide the information of the viewpoint of the left channel and the right channel for an observer, so that when the user watches the stereoscopic panoramic video, the depth perception information which is not contained in the single viewpoint panoramic video is increased, and the 'presence' is really realized. Therefore, the three-dimensional panoramic video system has wider application prospect in the fields of virtual reality, simulated driving and the like. However, a better visual experience comes at the cost of video data volume and processing. At present, too high complexity of the encoding time of the stereoscopic panoramic video also becomes one of the bottlenecks that restrict the massive application of the stereoscopic panoramic video. Therefore, how to effectively encode and compress the video is an urgent problem to be solved for the popularization and application of the stereoscopic panoramic video system.
Disclosure of Invention
The invention aims to solve the technical problem of providing a method for quickly encoding a three-dimensional panoramic video based on panoramic saliency, wherein the encoding time complexity is low.
The technical scheme adopted by the invention for solving the technical problems is as follows: a stereo panoramic video fast coding method based on panoramic saliency is characterized by comprising the following steps:
step 1: defining a current right viewpoint video frame to be processed except a 1 st frame in the stereoscopic panoramic video in the ERP projection format as a current frame; wherein the width of the current frame is W and the height is H;
step 2: carrying out significance calculation on the current frame to obtain a 3D-Sobel significance map of the current frame;
and step 3: defining a maximum coding unit to be processed currently in a current frame as a current maximum coding unit; wherein, the size of the current maximum coding unit is 64 × 64;
and 4, step 4: judging whether the current maximum coding unit is the uppermost or leftmost maximum coding unit in the current frame, if so, coding the current maximum coding unit by adopting a 3D-HEVC video coder, and then executing the step 11; otherwise, executing step 5;
and 5: calculating the significance strength of a region with the size of 64 multiplied by 64 corresponding to the current maximum coding unit in the 3D-Sobel significant map of the current frame, and marking the region as SILCU(ii) a And calculating a panorama significance threshold value of an area with the size of 64 multiplied by 64 corresponding to the current maximum coding unit in the 3D-Sobel significant image of the current frame, and recording the panorama significance threshold value as THS(ii) a Then judging SILCU≥THSIf yes, judging the current maximum coding unit as a significant block, redefining the current maximum coding unit as the current codingUnit, then execute step 9; if not, judging that the current maximum coding unit is an insignificant block, and then executing step 6;
step 6: let DLCU(View) represents the optimal recursive depth mean of the coded maximum coding unit corresponding to the current maximum coding unit in the left View video frame corresponding to the current frame, let DLCU(Col) represents the optimal recursive depth mean of the coded maximum coding unit corresponding to the current maximum coding unit in the right view video frame of the previous frame of the current frame, let DLCU(LT) represents the optimal recursive depth mean of the coded top-left maximum coding unit of the current maximum coding unit, let DLCU(L) represents the optimal recursive depth mean of the left-most coded unit of the current largest coded unit, let DLCU(T) represents an optimal recursive depth mean of the coded upper-side largest coding unit of the current largest coding unit; then, the recursion depth interval of the current maximum coding unit is predicted and is marked as [ D ]min,Dmax],
Figure BDA0002410766970000021
Figure BDA0002410766970000022
Wherein D isminMinimum partition depth, D, representing the current maximum coding unitmaxRepresents the maximum partition depth of the current maximum coding unit, min () is the function of taking the minimum value, max () is the function of taking the maximum value, the symbol
Figure BDA0002410766970000023
To round down the symbol, the symbol
Figure BDA0002410766970000024
Is a rounded up symbol;
and 7: jumping to a quadtree structure with the current maximum coding unit as a root node and dividing the depth into DminThe 3D-HEVC video encoder is adopted to encode all coding units in the CU layer in a depth-first traversal manner, and any coding unit in the CU layer is encodedUsing the current coding unit as the element, judging whether the maximum division depth of the current coding unit reaches D or not after the current coding unit is codedmaxOr whether the maximum division depth of the current coding unit reaches 3, if so, continuing to code the uncoded brother nodes of the current coding unit in a depth-first traversal mode until all the brother nodes of the current coding unit are coded, and then executing the step 11; if not, executing step 8;
and 8: calculating the significance strength of the region corresponding to the current coding unit in the 3D-Sobel significance map of the current frame, and recording the significance strength as SICUThen, SI is comparedCUAnd SILCUIf SI isCU>SILCUThen calculate the recursive depth interval of the current coding unit, which is marked as [ D ]min,D′max],
Figure BDA0002410766970000031
Then order Dmax=D′maxThen returning to the step 7 to continue the execution; if SI isCU≤SILCUThen calculate the recursive depth interval of the current coding unit, which is marked as [ D ]min,D″max],
Figure BDA0002410766970000032
Then order Dmax=D″maxThen returning to the step 7 to continue the execution; wherein D isCU(View) represents an optimal recursive depth mean of coded coding units corresponding to the current coding unit in the left View video frame corresponding to the current frame, DCU(Col) represents an optimal recursive depth mean, D, of the coded coding unit corresponding to the current coding unit in the right view video frame of the previous frame of the current frameCU(LT) represents the optimal recursive depth mean, D, of the coded upper left coding unit of the current coding unitCU(L) represents the optimal recursive depth mean, D, of the coded left coding unit of the current coding unitCU(T) represents an optimal recursive depth mean of the coded upper coding unit of the current coding unit;
and step 9: computing by using a BJND modelPerceptual distortion root mean square error, denoted MSE, of a preceding coding unitBjnd(ii) a And calculating the statistical root mean square error of the current coding unit, and recording as MSES
Figure BDA0002410766970000033
Then, the coding unit division threshold value based on the panoramic perception distortion is obtained through calculation and is recorded as THsplit,THsplit=η1MSES2MSEBjnd(ii) a Wherein e represents a natural base number, k is a slope, and the value is-2.3334, QstepRepresents a quantization step size of the current coding unit, QP represents a quantization parameter of the current coding unit,
Figure BDA0002410766970000034
MSEColrepresents the root mean square error of the coded coding unit corresponding to the current coding unit in the previous frame of the current frame,
Figure BDA0002410766970000035
representing the quantization step size, QP, of the coded coding unit corresponding to the current coding unit in the frame preceding the current frameColRepresenting the quantization parameter of the coded coding unit corresponding to the current coding unit in a frame preceding the current frame, b representing the intercept, with a value of 6.3751, nxn representing the size of the current coding unit, N having a value of 64 or 32 or 16 or 8, η1And η2Are all regulatory factors, η12=1;
Step 10: calculating the root mean square error of the current coding unit, and recording as MSECur(ii) a Then compares the MSECurAnd THsplitSize of (1), if MSECur≤THsplitIf the current coding unit reaches the optimal division depth, no further division is needed, a 3D-HEVC video encoder is adopted to encode the current coding unit, and then the step 11 is executed; if MSECur>THsplitJumping to a quadtree structure with the current maximum coding unit as a root node and dividing the depth into DminOf a 3D-HEVC video encoder, to depth all coding units in the CU layerCoding in a mode of degree-first traversal, regarding any coding unit in the CU layer as a current coding unit, then returning to the step 9 to continue execution until all sibling nodes of the current coding unit are coded completely, and then executing the step 11;
step 11: taking the next maximum coding unit to be processed in the current frame as the current maximum coding unit, then returning to the step 4 to continue executing until all the maximum coding units in the current frame are processed, and then executing the step 12;
step 12: and (3) taking the next frame of right viewpoint video frame to be processed in the three-dimensional panoramic video in the ERP projection format as a current frame, and then returning to the step (2) to continue executing until all video frames in the three-dimensional panoramic video in the ERP projection format are processed.
In the step 2, the significance of the current frame is calculated by adopting a 3D-Sobel model.
In the step 5, SI is calculatedLCUAnd calculating SI in said step 8CUThe process is the same, and the specific process is as follows: defining the area of which the significance strength is to be calculated as the area to be processed, recording the significance strength of the area to be processed as SI,
Figure BDA0002410766970000041
wherein the content of the first and second substances,
Figure BDA0002410766970000042
which indicates the size of the area to be treated,
Figure BDA0002410766970000043
has a value of 64 or 32 or 16 or 8,
Figure BDA0002410766970000044
the coordinate position of the pixel point at the upper left corner of the region to be processed in the 3D-Sobel saliency map of the current frame is represented,
Figure BDA0002410766970000045
the coordinate position in the 3D-Sobel saliency map representing the current frame is
Figure BDA0002410766970000046
The pixel value of the pixel point of (a),
Figure BDA0002410766970000047
in the step 5, THSThe calculation process of (2) is as follows:
step 5_ 1: calculating the ERP dimension weight of each pixel point in the current frame, and recording the ERP dimension weight of the pixel point with the coordinate position (x, y) in the current frame as wERP(x,y),
Figure BDA0002410766970000048
Wherein x is more than or equal to 0 and less than or equal to W-1, and y is more than or equal to 0 and less than or equal to H-1;
step 5_ 2: calculating the ERP dimension weight of the current maximum coding unit, and recording as wLCU
Figure BDA0002410766970000051
Wherein, N ' x N ' represents the size of the current maximum coding unit, i.e. the value of N ' is 64, (i ', j ') represents the coordinate position of the top left pixel point of the current maximum coding unit in the current frame, wERP(m ', n') represents ERP dimension weight of pixel point with coordinate position (m ', n') in the current frame, wherein m 'is more than or equal to 0 and less than or equal to W-1, and n' is more than or equal to 0 and less than or equal to H-1;
step 5_ 3: calculate THS,THS=THE+β×(1-wLCU) (ii) a Wherein TH isERepresents the significance threshold at half the height of the current frame, and β represents wLCUThe scaling factor of (2).
In the step 9, MSEBjndThe calculation process of (2) is as follows:
Figure BDA0002410766970000052
the N multiplied by N represents the size of the current coding unit, the value of N is 64 or 32 or 16 or 8, (i, j) represents the coordinate position of the upper left pixel point of the current coding unit in the current frame, BJND (m, N) represents the pixel value of the pixel point with the coordinate position (m, N) in the binocular just-noticeable distortion diagram of the current frame, m is more than or equal to 0 and less than or equal to W-1, and N is more than or equal to 0 and less than or equal to H-1.
In the step 9, MSEColThe calculation process of (2) is as follows:
Figure BDA0002410766970000053
the method comprises the steps of obtaining a current coding unit, obtaining a pixel value of a pixel point with a coordinate position (m, N) in a previous frame of a current frame, obtaining a pixel value of a pixel point with a coordinate position (m, N) in the previous frame of the current frame, obtaining a pixel value of a pixel point with a coordinate position (m, N) in a coded and reconstructed image of the previous frame of the current frame, and obtaining a pixel value of a pixel point with a coordinate position (m, N) in the previous frame of the current frame, wherein NxN represents the size of the current coding unit, N is 64 or 32 or 16 or 8, (I, j) represents the coordinate position of a pixel point in the upper left corner of the current coding unit, I (m, N) represents the pixel value of a pixel point with a coordinate position (m, N) in the previous frame of the current frame, I' (m, N) represents the pixel value of a pixel point with a pixel point (m, N) in the coded and reconstructed image in the previous frame of the current frame, m, N is greater than or equal to 0 and equal to or equal to W-1, and equal to or equal to 0.
In the step 10, MSECurThe calculation process of (2) is as follows:
Figure BDA0002410766970000054
wherein, NxN represents the size of the current coding unit, the value of N is 64 or 32 or 16 or 8, (i, j) represents the coordinate position of the upper left pixel point of the current coding unit in the current frame,
Figure BDA0002410766970000055
representing the pixel value of the pixel point with the coordinate position of (m, n) in the current frame,
Figure BDA0002410766970000056
and the pixel value of a pixel point with the coordinate position (m, n) in the coding predicted image of the current frame is represented, wherein m is more than or equal to 0 and less than or equal to W-1, and n is more than or equal to 0 and less than or equal to H-1.
Compared with the prior art, the invention has the advantages that:
the method analyzes the significance of a right viewpoint video frame in the stereoscopic panoramic video, provides a corresponding rapid termination mode respectively for a non-significant region and a significant region in the right viewpoint video frame, and predicts and corrects a recursion depth interval of a current coding unit by using the optimal division depth of adjacent blocks in a time-space domain of the current coding unit for the non-significant region; aiming at the significant region, whether the current coding unit reaches the optimal division depth is judged by calculating and comparing the size relation between the root mean square error of the current coding unit and the division threshold of the coding unit based on the panoramic perception distortion, and experimental tests show that the method can effectively reduce the recursion complexity of the coding unit and save the coding time.
Drawings
FIG. 1 is a block diagram of an overall implementation of the method of the present invention;
FIG. 2 is a flow diagram for fast termination of non-salient blocks;
fig. 3 is a flow chart for fast termination of a significant block.
Detailed Description
The invention is described in further detail below with reference to the accompanying examples.
The overall implementation block diagram of the method for rapidly encoding the stereoscopic panoramic video based on the panoramic saliency, which is provided by the invention, is shown in fig. 1, and the method comprises the following steps:
step 1: defining a right viewpoint video frame except a 1 st frame to be processed currently in a stereoscopic panoramic video in an ERP (Enterprise resource planning) projection format as a current frame; wherein the width of the current frame is W and the height is H. Here, all left view video frames and the 1 st frame right view video frame in the stereoscopic panoramic video are encoded by using the existing 3D-HEVC video encoder.
Step 2: and carrying out significance calculation on the current frame to obtain a 3D-Sobel significance map of the current frame.
In this embodiment, in step 2, a 3D-Sobel model is used to perform saliency calculation on the current frame.
And step 3: defining a maximum coding unit to be processed currently in a current frame as a current maximum coding unit; wherein the size of the current maximum coding unit is 64 × 64.
And 4, step 4: judging whether the current maximum coding unit is the uppermost or leftmost maximum coding unit in the current frame, if so, coding the current maximum coding unit by adopting a 3D-HEVC video coder, and then executing the step 11; otherwise, step 5 is executed. Here, the uppermost maximum coding unit in the current frame is the first row maximum coding unit in the current frame, and the leftmost maximum coding unit in the current frame is the first column maximum coding unit in the current frame.
And 5: calculating the significance strength of a region with the size of 64 multiplied by 64 corresponding to the current maximum coding unit in the 3D-Sobel significant map of the current frame, and marking the region as SILCU(ii) a And calculating a panorama significance threshold value of an area with the size of 64 multiplied by 64 corresponding to the current maximum coding unit in the 3D-Sobel significant image of the current frame, and recording the panorama significance threshold value as THS(ii) a Then judging SILCU≥THSIf yes, judging that the current maximum coding unit is a significant block, redefining the current maximum coding unit as the current coding unit, and executing the step 9; if not, the current maximum coding unit is determined to be an insignificant block, and then step 6 is performed.
In this embodiment, TH is performed in step 5SThe calculation process of (2) is as follows:
step 5_ 1: calculating the ERP dimension weight of each pixel point in the current frame, and recording the ERP dimension weight of the pixel point with the coordinate position (x, y) in the current frame as wERP(x,y),
Figure BDA0002410766970000071
Wherein x is more than or equal to 0 and less than or equal to W-1, and y is more than or equal to 0 and less than or equal to H-1.
Step 5_ 2: since the 3D-HEVC video encoder uses CU blocks in a video frame as basic coding units. For the convenience of application in coding, the ERP dimension weight of the current maximum coding unit is calculated and is marked as wLCU
Figure BDA0002410766970000072
Wherein, N ' x N ' represents the size of the current maximum coding unit, i.e. the value of N ' is 64, (i ', j ') represents the coordinate position of the top left pixel point of the current maximum coding unit in the current frame, wERPAnd (m ', n') represents the ERP dimension weight of the pixel point with the coordinate position (m ', n') in the current frame, wherein m 'is more than or equal to 0 and less than or equal to W-1, and n' is more than or equal to 0 and less than or equal to H-1.
Step 5_ 3: calculate THS,THS=THE+β×(1-wLCU) (ii) a Wherein TH isERepresenting the significance threshold at half height of the current frame, by a large number of realitiesTest to obtain THE0.3, beta represents wLCUβ is 0.2 through a number of experiments.
FIG. 2 presents a flow diagram for fast termination of non-salient blocks; fig. 3 presents a flow chart for fast termination of a significant block.
Step 6: let DLCU(View) represents the optimal recursive depth mean of the coded maximum coding unit corresponding to the current maximum coding unit in the left View video frame corresponding to the current frame, let DLCU(Col) represents the optimal recursive depth mean of the coded maximum coding unit corresponding to the current maximum coding unit in the right view video frame of the previous frame of the current frame, let DLCU(LT) represents the optimal recursive depth mean of the coded top-left maximum coding unit of the current maximum coding unit, let DLCU(L) represents the optimal recursive depth mean of the left-most coded unit of the current largest coded unit, let DLCU(T) represents an optimal recursive depth mean of the coded upper-side largest coding unit of the current largest coding unit; then, the recursion depth interval of the current maximum coding unit is predicted and is marked as [ D ]min,Dmax],
Figure BDA0002410766970000081
Figure BDA0002410766970000082
Wherein D isminMinimum partition depth, D, representing the current maximum coding unitmaxRepresents the maximum partition depth of the current maximum coding unit, min () is the function of taking the minimum value, max () is the function of taking the maximum value, the symbol
Figure BDA0002410766970000083
To round down the symbol, the symbol
Figure BDA0002410766970000084
Is a rounded up symbol; here, the coded upper left-most coding unit of the current maximum coding unit is the coded nearest neighbor maximum coding unit located at the upper left side of the current maximum coding unitAnd the encoded left-side maximum coding unit of the current maximum coding unit is the encoded nearest neighbor maximum coding unit positioned on the left side of the current maximum coding unit, and the encoded upper-side maximum coding unit of the current maximum coding unit is the encoded nearest neighbor maximum coding unit positioned on the upper side of the current maximum coding unit.
And 7: jumping to a quadtree structure with the current maximum coding unit as a root node and dividing the depth into DminThe coding unit layer(s) of (1) is (are) coded by a 3D-HEVC video coder in a depth-first traversal manner for all coding units in the CU layer, any coding unit in the CU layer is taken as a current coding unit, and after the current coding unit is coded, whether the maximum division depth of the current coding unit reaches D is judged firstmaxOr whether the maximum division depth of the current coding unit reaches 3, if so, continuing to code the uncoded brother nodes of the current coding unit in a depth-first traversal mode until all the brother nodes of the current coding unit are coded, and then executing the step 11; if not, step 8 is performed.
And 8: calculating the significance strength of the region corresponding to the current coding unit in the 3D-Sobel significance map of the current frame, and recording the significance strength as SICUThen, SI is comparedCUAnd SILCUIf SI isCU>SILCUAnd if the current coding unit is judged to be an insignificant block, the current maximum coding unit is located at a position where a significant region and an insignificant region in the current frame are bordered, and the current coding unit is located in a region with higher significance, so that to avoid quality degradation caused by too small coding depth, it is necessary to update the D according to the optimal division depth of the adjacent blocks in the time-space domain of the current coding unitmaxThus, the recursive depth interval of the current coding unit is calculated, denoted as [ D ]min,D′max],
Figure BDA0002410766970000091
Then order Dmax=D′maxThen returning to the step 7 to continue the execution; if SI isCU≤SILCUThen, it means that the significance strength of the current coding unit is consistent with or less than that of the current maximum coding unit, the current coding unit is inside the current maximum coding unit, the current coding unit is in a region where the significance is consistent with or less than the whole, and therefore the recursion depth interval of the current coding unit is calculated and is marked as [ D ]min,D″max],
Figure BDA0002410766970000092
Then order Dmax=D″maxThen returning to the step 7 to continue the execution; wherein D isCU(View) represents an optimal recursive depth mean of coded coding units corresponding to the current coding unit in the left View video frame corresponding to the current frame, DCU(Col) represents an optimal recursive depth mean, D, of the coded coding unit corresponding to the current coding unit in the right view video frame of the previous frame of the current frameCU(LT) represents the optimal recursive depth mean, D, of the coded upper left coding unit of the current coding unitCU(L) represents the optimal recursive depth mean, D, of the coded left coding unit of the current coding unitCU(T) represents an optimal recursive depth mean of the coded upper coding unit of the current coding unit; here, the coded upper-left coding unit of the current coding unit is a coded nearest neighbor coding unit located on the upper left side of the current coding unit, the coded left coding unit of the current coding unit is a coded nearest neighbor coding unit located on the left side of the current coding unit, and the coded upper-side coding unit of the current coding unit is a coded nearest neighbor coding unit located on the upper side of the current coding unit.
And step 9: the current maximum coding unit is a significant block, which indicates that the current maximum coding unit is in a texture dense area, and the perceptual Distortion root mean square error of the current coding unit is calculated by using the existing classical BJND (Binocular Just Noticeable Distortion) model and is recorded as MSEBjnd(ii) a And calculating the statistical root mean square error of the current coding unit, and recording as MSES
Figure BDA0002410766970000093
Then, the coding unit division threshold value based on the panoramic perception distortion is obtained through calculation and is recorded as THsplit,THsplit=η1MSES2MSEBjnd(ii) a Wherein e represents a natural base number, k is a slope, and the value is-2.3334, QstepRepresents a quantization step size of the current coding unit, QP represents a quantization parameter of the current coding unit,
Figure BDA0002410766970000094
MSEColrepresents the root mean square error of the coded coding unit corresponding to the current coding unit in the previous frame of the current frame,
Figure BDA0002410766970000095
representing the quantization step size, QP, of the coded coding unit corresponding to the current coding unit in the frame preceding the current frameColRepresenting the quantization parameter of the coded coding unit corresponding to the current coding unit in a frame preceding the current frame, b representing the intercept, with a value of 6.3751, nxn representing the size of the current coding unit, N having a value of 64 or 32 or 16 or 8, η1And η2Are all regulatory factors, η121 in this example, η1=η2=0.5。
In this embodiment, step 9, MSEBjndThe calculation process of (2) is as follows:
Figure BDA0002410766970000101
the N multiplied by N represents the size of the current coding unit, the value of N is 64 or 32 or 16 or 8, (i, j) represents the coordinate position of the upper left pixel point of the current coding unit in the current frame, BJND (m, N) represents the pixel value of the pixel point with the coordinate position of (m, N) in the binocular just-noticeable distortion diagram (namely BJND diagram) of the current frame, m is more than or equal to 0 and less than or equal to W-1, and N is more than or equal to 0 and less than or equal to H-1.
In this embodiment, step 9, MSEColThe calculation process of (2) is as follows:
Figure BDA0002410766970000102
the method comprises the steps of obtaining a current coding unit, obtaining a pixel value of a pixel point with a coordinate position (m, N) in a previous frame of a current frame, obtaining a pixel value of a pixel point with a coordinate position (m, N) in the previous frame of the current frame, obtaining a pixel value of a pixel point with a coordinate position (m, N) in a coded and reconstructed image of the previous frame of the current frame, and obtaining a pixel value of a pixel point with a coordinate position (m, N) in the previous frame of the current frame, wherein NxN represents the size of the current coding unit, N is 64 or 32 or 16 or 8, (I, j) represents the coordinate position of a pixel point in the upper left corner of the current coding unit, I (m, N) represents the pixel value of a pixel point with a coordinate position (m, N) in the previous frame of the current frame, I' (m, N) represents the pixel value of a pixel point with a pixel point (m, N) in the coded and reconstructed image in the previous frame of the current frame, m, N is greater than or equal to 0 and equal to or equal to W-1, and equal to or equal to 0.
Step 10: calculating the root mean square error of the current coding unit, and recording as MSECur(ii) a Then compares the MSECurAnd THsplitSize of (1), if MSECur≤THsplitIf the current coding unit reaches the optimal division depth, no further division is needed, a 3D-HEVC video encoder is adopted to encode the current coding unit, and then the step 11 is executed; if MSECur>THsplitJumping to a quadtree structure with the current maximum coding unit as a root node and dividing the depth into DminThe 3D-HEVC video encoder is adopted to encode all coding units in the CU layer in a depth-first traversal manner, and any coding unit in the CU layer is taken as a current coding unit, and then the process returns to step 9 to continue to be executed until all sibling nodes of the current coding unit are encoded, and then step 11 is executed.
In this embodiment, step 10, MSECurThe calculation process of (2) is as follows:
Figure BDA0002410766970000111
wherein, NxN represents the size of the current coding unit, the value of N is 64 or 32 or 16 or 8, (i, j) represents the coordinate position of the upper left pixel point of the current coding unit in the current frame,
Figure BDA0002410766970000112
representing the pixel value of the pixel point with the coordinate position of (m, n) in the current frame,
Figure BDA0002410766970000113
and the pixel value of a pixel point with the coordinate position (m, n) in the coding predicted image of the current frame is represented, wherein m is more than or equal to 0 and less than or equal to W-1, and n is more than or equal to 0 and less than or equal to H-1.
Step 11: and taking the next maximum coding unit to be processed in the current frame as the current maximum coding unit, then returning to the step 4 to continue executing until all the maximum coding units in the current frame are processed, and then executing the step 12.
Step 12: and (3) taking the next frame of right viewpoint video frame to be processed in the three-dimensional panoramic video in the ERP projection format as a current frame, and then returning to the step (2) to continue executing until all video frames in the three-dimensional panoramic video in the ERP projection format are processed.
In this embodiment, SI is calculated in step 5LCUAnd calculating SI in step 8CUThe process is the same, and the specific process is as follows: defining the area of which the significance strength is to be calculated as the area to be processed, recording the significance strength of the area to be processed as SI,
Figure BDA0002410766970000114
wherein the content of the first and second substances,
Figure BDA0002410766970000115
which indicates the size of the area to be treated,
Figure BDA0002410766970000116
has a value of 64 or 32 or 16 or 8,
Figure BDA0002410766970000117
the coordinate position of the pixel point at the upper left corner of the region to be processed in the 3D-Sobel saliency map of the current frame is represented,
Figure BDA0002410766970000118
the coordinate position in the 3D-Sobel saliency map representing the current frame is
Figure BDA0002410766970000119
The pixel value of the pixel point of (a),
Figure BDA00024107669700001110
to further illustrate the performance of the method of the present invention, the method of the present invention was tested.
In order to evaluate the effectiveness of the method, an HTM14.1 is selected as a test model, a 64-bit WIN7 operating system with a CPU Intel (R) core (TM) i3, a main frequency of 2.4GHz and a memory of 4G is configured in hardware, and a development tool selects VS 2013. Selecting stereo panoramic video sequences of 'chat', 'experience', 'photomraph', 'riverside', 'scientific _ spot', 'sign _ in', 'tourrist' and 'traffic' as standard test sequences, wherein the test frames are 100 frames, the coding structure is in an HBP random access mode, the GOP length of a group of pictures is 8, and the period of an I frame is 24. The initial quantization parameters QP of the independent viewpoints are 22, 27, 32 and 37 respectively, and the method is tested on the dependent viewpoints.
Table 1 lists specific information of "chat", "experience", "photograph", "riverside", "scientific _ spot", "sign _ in", "tourrist", and "traffic" stereoscopic panoramic video sequences.
TABLE 1 associated parameter information for stereoscopic panoramic video sequences
Figure BDA0002410766970000121
Table 2 shows the savings in coding time when the method of the present invention is used to code the stereoscopic panoramic video sequence listed in table 1, compared to the HTM original platform method. The time saving rate of coding by adopting the method of the invention compared with the coding by adopting the HTM original platform method is defined as delta TPRO-CU,ΔTPRO-CU=(TOrg-TPRO-CU)/TOrg×100[%]Wherein, TPRO-CURepresenting the coding time, T, of the coding by the method of the inventionOrgRepresenting the encoding time for encoding using the HTM native platform method.
TABLE 2 comparison of time savings for encoding using the method of the present invention versus HTM native platform method
Figure BDA0002410766970000131
Figure BDA0002410766970000141
As can be seen from table 2, the encoding using the method of the present invention can save 53.5% of encoding time on average. The method is adopted to carry out coding relative equalization aiming at 8 stereoscopic panoramic video sequences with different scenes and different motion conditions, the effect is better particularly for outdoor stereoscopic panoramic video sequences such as 'riverside' and 'scenic _ spot', and the coding time is saved by 57.8 percent and 56.5 percent respectively.
Table 3 lists the comparison of rate-distortion performance of the stereo panoramic video sequence listed in table 1 encoded by the method of the present invention under different quality evaluation methods. In the quality evaluation, PSNR (Peak Signal-to-Noise Rate), WS-PSNR (weighted to statistical uniform PSNR), and WS-SSIM (weighted-to-statistical-structural similarity, WS-SSIM) are used as quality evaluation indexes, and Rate distortion performance indexes under each quality evaluation method are respectively calculated and are correspondingly marked as BDBRPSNR(%)、BDBRWS-PSNR(%)、BDBRWS-SSIM(%) to evaluate the performance of the process of the invention.
TABLE 3 comparison of rate-distortion performance for different quality evaluation methods for encoding using the method of the present invention
Figure BDA0002410766970000151
As can be seen from Table 3, the code rate increase of the method of the present invention is below 1% under 3 quality indexes PSNR, WS-PSNR and WS-SSIM, and the average is 0.4%, 0.2% and 0.0%, respectively. WS-PSNR and BDBRWS-PSNR(%) is a quality evaluation index recommended by the panoramic video proposal, and compared with the traditional PSNR and BDBR, the method can be seenPSNR(%),The evaluation index can better embody the characteristics of the stereoscopic panoramic video. The WS-SSIM is proposed based on SSIM, and structural similarity and panoramic latitude factors are considered, so that the subjective quality performance of the coding method can be better reflected. Specifically, due to different scenes and different motion conditions of each stereoscopic panoramic video sequence, the BDBR (representing the percentage of code rate saved by a better coding method under the same objective quality) changes slightly differently, wherein the code rate of the stereoscopic panoramic video sequence (indoor) with complex texture and severe motion is slightly obviously increased, and the stereoscopic panoramic video sequence (indoor) with complex texture, severe motion is good in effect, the stereoscopic panoramic video is an outdoor sequence, north and south polar regions are mostly sky and ground, the occupation ratio is high, the texture is relatively simple, the severe motion degree is slow, and the BDBR is adopted for the stereoscopic panoramic video sequenceWS-PSNR(%)、BDBRWS-SSIMThe code rate is hardly increased by the evaluation index of (%).

Claims (7)

1. A stereo panoramic video fast coding method based on panoramic saliency is characterized by comprising the following steps:
step 1: defining a current right viewpoint video frame to be processed except a 1 st frame in the stereoscopic panoramic video in the ERP projection format as a current frame; wherein the width of the current frame is W and the height is H;
step 2: carrying out significance calculation on the current frame to obtain a 3D-Sobel significance map of the current frame;
and step 3: defining a maximum coding unit to be processed currently in a current frame as a current maximum coding unit; wherein, the size of the current maximum coding unit is 64 × 64;
and 4, step 4: judging whether the current maximum coding unit is the uppermost or leftmost maximum coding unit in the current frame, if so, coding the current maximum coding unit by adopting a 3D-HEVC video coder, and then executing the step 11; otherwise, executing step 5;
and 5: calculating the display of a region of size 64 x 64 corresponding to the current maximum coding unit in the 3D-Sobel saliency map of the current frameRemarkable strength, recorded as SILCU(ii) a And calculating a panorama significance threshold value of an area with the size of 64 multiplied by 64 corresponding to the current maximum coding unit in the 3D-Sobel significant image of the current frame, and recording the panorama significance threshold value as THS(ii) a Then judging SILCU≥THSIf yes, judging that the current maximum coding unit is a significant block, redefining the current maximum coding unit as the current coding unit, and executing the step 9; if not, judging that the current maximum coding unit is an insignificant block, and then executing step 6;
step 6: let DLCU(View) represents the optimal recursive depth mean of the coded maximum coding unit corresponding to the current maximum coding unit in the left View video frame corresponding to the current frame, let DLCU(Col) represents the optimal recursive depth mean of the coded maximum coding unit corresponding to the current maximum coding unit in the right view video frame of the previous frame of the current frame, let DLCU(LT) represents the optimal recursive depth mean of the coded top-left maximum coding unit of the current maximum coding unit, let DLCU(L) represents the optimal recursive depth mean of the left-most coded unit of the current largest coded unit, let DLCU(T) represents an optimal recursive depth mean of the coded upper-side largest coding unit of the current largest coding unit; then, the recursion depth interval of the current maximum coding unit is predicted and is marked as [ D ]min,Dmax],
Figure FDA0002410766960000011
Figure FDA0002410766960000012
Wherein D isminMinimum partition depth, D, representing the current maximum coding unitmaxRepresents the maximum partition depth of the current maximum coding unit, min () is the function of taking the minimum value, max () is the function of taking the maximum value, the symbol
Figure FDA0002410766960000021
To round down the symbol, the symbol
Figure FDA0002410766960000022
Is a rounded up symbol;
and 7: jumping to a quadtree structure with the current maximum coding unit as a root node and dividing the depth into DminThe method comprises the steps of coding all coding units in a CU layer in a depth-first traversal mode by adopting a 3D-HEVC video coder, regarding any coding unit in the CU layer as a current coding unit, and judging whether the maximum division depth of the current coding unit reaches D or not after the current coding unit is codedmaxOr whether the maximum division depth of the current coding unit reaches 3, if so, continuing to code the uncoded brother nodes of the current coding unit in a depth-first traversal mode until all the brother nodes of the current coding unit are coded, and then executing the step 11; if not, executing step 8;
and 8: calculating the significance strength of the region corresponding to the current coding unit in the 3D-Sobel significance map of the current frame, and recording the significance strength as SICUThen, SI is comparedCUAnd SILCUIf SI isCU>SILCUThen calculate the recursive depth interval of the current coding unit, which is marked as [ D ]min,D′max],
Figure FDA0002410766960000023
Then order Dmax=D′maxThen returning to the step 7 to continue the execution; if SI isCU≤SILCUThen calculate the recursive depth interval of the current coding unit, which is marked as [ D ]min,D″max],
Figure FDA0002410766960000024
Then order Dmax=D″maxThen returning to the step 7 to continue the execution; wherein D isCU(View) represents an optimal recursive depth mean of coded coding units corresponding to the current coding unit in the left View video frame corresponding to the current frame, DCU(Col) represents a unit coded in a right view video frame previous to the current frame and the current coding unitOptimal recursive depth mean, D, of the corresponding coded coding unitCU(LT) represents the optimal recursive depth mean, D, of the coded upper left coding unit of the current coding unitCU(L) represents the optimal recursive depth mean, D, of the coded left coding unit of the current coding unitCU(T) represents an optimal recursive depth mean of the coded upper coding unit of the current coding unit;
and step 9: calculating the root mean square error of the perceptual distortion of the current coding unit by using a BJND model, and recording the root mean square error as MSEBjnd(ii) a And calculating the statistical root mean square error of the current coding unit, and recording as MSES
Figure FDA0002410766960000025
Then, the coding unit division threshold value based on the panoramic perception distortion is obtained through calculation and is recorded as THsplit,THsplit=η1MSES2MSEBjnd(ii) a Wherein e represents a natural base number, k is a slope, and the value is-2.3334, QstepRepresents a quantization step size of the current coding unit, QP represents a quantization parameter of the current coding unit,
Figure FDA0002410766960000031
MSEColrepresents the root mean square error of the coded coding unit corresponding to the current coding unit in the previous frame of the current frame,
Figure FDA0002410766960000032
representing the quantization step size, QP, of the coded coding unit corresponding to the current coding unit in the frame preceding the current frameColRepresenting the quantization parameter of the coded coding unit corresponding to the current coding unit in a frame preceding the current frame, b representing the intercept, with a value of 6.3751, nxn representing the size of the current coding unit, N having a value of 64 or 32 or 16 or 8, η1And η2Are all regulatory factors, η12=1;
Step 10: calculating the root mean square error of the current coding unit, and recording as MSECur(ii) a Then compares the MSECurAnd THsplitSize of (1), if MSECur≤THsplitIf the current coding unit reaches the optimal division depth, no further division is needed, a 3D-HEVC video encoder is adopted to encode the current coding unit, and then the step 11 is executed; if MSECur>THsplitJumping to a quadtree structure with the current maximum coding unit as a root node and dividing the depth into DminThe 3D-HEVC video encoder is adopted to encode all coding units in the CU layer in a depth-first traversal mode, any coding unit in the CU layer is taken as a current coding unit, then the step 9 is returned to continue to be executed until all brother node encoding of the current coding unit is completed, and then the step 11 is executed;
step 11: taking the next maximum coding unit to be processed in the current frame as the current maximum coding unit, then returning to the step 4 to continue executing until all the maximum coding units in the current frame are processed, and then executing the step 12;
step 12: and (3) taking the next frame of right viewpoint video frame to be processed in the three-dimensional panoramic video in the ERP projection format as a current frame, and then returning to the step (2) to continue executing until all video frames in the three-dimensional panoramic video in the ERP projection format are processed.
2. The method according to claim 1, wherein in step 2, a 3D-Sobel model is used to perform saliency calculation on the current frame.
3. The method as claimed in claim 1 or 2, wherein the SI is calculated in step 5LCUAnd calculating SI in said step 8CUThe process is the same, and the specific process is as follows: defining the area of which the significance strength is to be calculated as the area to be processed, recording the significance strength of the area to be processed as SI,
Figure FDA0002410766960000033
wherein the content of the first and second substances,
Figure FDA0002410766960000034
which indicates the size of the area to be treated,
Figure FDA0002410766960000035
has a value of 64 or 32 or 16 or 8,
Figure FDA0002410766960000041
the coordinate position of the pixel point at the upper left corner of the region to be processed in the 3D-Sobel saliency map of the current frame is represented,
Figure FDA0002410766960000042
the coordinate position in the 3D-Sobel saliency map representing the current frame is
Figure FDA0002410766960000043
The pixel value of the pixel point of (a),
Figure FDA0002410766960000044
4. the method as claimed in claim 3, wherein TH in step 5 is THSThe calculation process of (2) is as follows:
step 5_ 1: calculating the ERP dimension weight of each pixel point in the current frame, and recording the ERP dimension weight of the pixel point with the coordinate position (x, y) in the current frame as wERP(x,y),
Figure FDA0002410766960000045
Wherein x is more than or equal to 0 and less than or equal to W-1, and y is more than or equal to 0 and less than or equal to H-1;
step 5_ 2: calculating the ERP dimension weight of the current maximum coding unit, and recording as wLCU
Figure FDA0002410766960000046
Wherein, N ' x N ' represents the size of the current maximum coding unit, i.e. the value of N ' is 64, (i ', j ') represents the coordinate position of the top left pixel point of the current maximum coding unit in the current frame, wERP(m ', n') represents ERP dimension weight of pixel point with coordinate position (m ', n') in the current frame, wherein m 'is more than or equal to 0 and less than or equal to W-1, and n' is more than or equal to 0 and less than or equal to H-1;
step 5_ 3: calculate THS,THS=THE+β×(1-wLCU) (ii) a Wherein TH isERepresents the significance threshold at half the height of the current frame, and β represents wLCUThe scaling factor of (2).
5. The method as claimed in claim 4, wherein the MSE in step 9 is implemented by using a stereo-panorama saliency-based stereo-panorama video fast coding methodBjndThe calculation process of (2) is as follows:
Figure FDA0002410766960000047
the N multiplied by N represents the size of the current coding unit, the value of N is 64 or 32 or 16 or 8, (i, j) represents the coordinate position of the upper left pixel point of the current coding unit in the current frame, BJND (m, N) represents the pixel value of the pixel point with the coordinate position (m, N) in the binocular just-noticeable distortion diagram of the current frame, m is more than or equal to 0 and less than or equal to W-1, and N is more than or equal to 0 and less than or equal to H-1.
6. The method as claimed in claim 5, wherein the MSE in step 9 is implemented by using a stereo-panorama saliency-based stereo-panorama video coding methodColThe calculation process of (2) is as follows:
Figure FDA0002410766960000051
the method comprises the steps of obtaining a current coding unit, obtaining a pixel value of a pixel point with a coordinate position (m, N) in a previous frame of a current frame, obtaining a pixel value of a pixel point with a coordinate position (m, N) in the previous frame of the current frame, obtaining a pixel value of a pixel point with a coordinate position (m, N) in a coded and reconstructed image of the previous frame of the current frame, and obtaining a pixel value of a pixel point with a coordinate position (m, N) in the previous frame of the current frame, wherein NxN represents the size of the current coding unit, N is 64 or 32 or 16 or 8, (I, j) represents the coordinate position of a pixel point in the upper left corner of the current coding unit, I (m, N) represents the pixel value of a pixel point with a coordinate position (m, N) in the previous frame of the current frame, I' (m, N) represents the pixel value of a pixel point with a pixel point (m, N) in the coded and reconstructed image in the previous frame of the current frame, m, N is greater than or equal to 0 and equal to or equal to W-1, and equal to or equal to 0.
7. The method as claimed in claim 6, wherein in step 10, MSE is used for fast coding of stereoscopic panoramic video based on panoramic saliencyCurThe calculation process of (2) is as follows:
Figure FDA0002410766960000052
wherein, NxN represents the size of the current coding unit, the value of N is 64 or 32 or 16 or 8, (i, j) represents the coordinate position of the upper left pixel point of the current coding unit in the current frame,
Figure FDA0002410766960000053
representing the pixel value of the pixel point with the coordinate position of (m, n) in the current frame,
Figure FDA0002410766960000054
and the pixel value of a pixel point with the coordinate position (m, n) in the coding predicted image of the current frame is represented, wherein m is more than or equal to 0 and less than or equal to W-1, and n is more than or equal to 0 and less than or equal to H-1.
CN202010175774.3A 2020-03-13 2020-03-13 Three-dimensional panoramic video rapid coding method based on panoramic saliency Active CN111464805B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010175774.3A CN111464805B (en) 2020-03-13 2020-03-13 Three-dimensional panoramic video rapid coding method based on panoramic saliency

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010175774.3A CN111464805B (en) 2020-03-13 2020-03-13 Three-dimensional panoramic video rapid coding method based on panoramic saliency

Publications (2)

Publication Number Publication Date
CN111464805A CN111464805A (en) 2020-07-28
CN111464805B true CN111464805B (en) 2021-08-10

Family

ID=71680747

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010175774.3A Active CN111464805B (en) 2020-03-13 2020-03-13 Three-dimensional panoramic video rapid coding method based on panoramic saliency

Country Status (1)

Country Link
CN (1) CN111464805B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113422956B (en) * 2021-06-17 2022-09-09 北京金山云网络技术有限公司 Image coding method and device, electronic equipment and storage medium

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108886598A (en) * 2016-01-12 2018-11-23 上海科技大学 The compression method and device of panoramic stereoscopic video system
CN110612553A (en) * 2017-05-09 2019-12-24 皇家Kpn公司 Encoding spherical video data

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108886598A (en) * 2016-01-12 2018-11-23 上海科技大学 The compression method and device of panoramic stereoscopic video system
CN110612553A (en) * 2017-05-09 2019-12-24 皇家Kpn公司 Encoding spherical video data

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
《A novel texture-based asymmetric visibility threshold model for stereoscopic video coding》;Du,B.,Yu,M.,Jiang,G.,Zhang,Y.,Zhu,T.;《Eurasip Journal on Image and Video Processing》;20181231;全文 *
《HEVC-compliant viewport-adaptive streaming of stereoscopic panoramic video》;Alireza Zare; Kashyap Kammachi Sreedhar;《2016 Picture Coding Symposium (PCS), 2016, pp. 1-5, doi: 10.1109/PCS.2016.7906401》;20161231;全文 *
《全景视频的显著性检测及其在编码传输中的应用》;苏群;《中国优秀硕士学位论文全文数据库-信息科技辑》;20190115(第1期);全文 *

Also Published As

Publication number Publication date
CN111464805A (en) 2020-07-28

Similar Documents

Publication Publication Date Title
CN110087087B (en) VVC inter-frame coding unit prediction mode early decision and block division early termination method
RU2518635C2 (en) Method and apparatus for encoding and decoding coding unit of picture boundary
EP3389276B1 (en) Hash-based encoder decisions for video coding
CN108322747B (en) Coding unit division optimization method for ultra-high definition video
CN108271024B (en) Image coding and decoding method and device
JP6664819B2 (en) System and method for mask-based processing of blocks of digital images
JP2018501698A (en) System and method for processing digital images
CN107371029B (en) Video packet priority distribution method based on content
WO2012098776A1 (en) Image coding device and image decoding device
CN110446052B (en) 3D-HEVC intra-frame depth map rapid CU depth selection method
CN102263951A (en) Quick fractal video compression and decompression method
JP6496821B2 (en) System and method for processing a block of a digital image
Chen et al. Sum-of-gradient based fast intra coding in 3D-HEVC for depth map sequence (SOG-FDIC)
JP2004242318A (en) Video block dividing method and its apparatus
CN111464805B (en) Three-dimensional panoramic video rapid coding method based on panoramic saliency
EP1690232A2 (en) Detection of local visual space-time details in a video signal
KR100865690B1 (en) Apparatus and method for fast mode selection
Zuo et al. Bi-layer texture discriminant fast depth intra coding for 3D-HEVC
CN111372079A (en) VVC inter-frame CU deep rapid dividing method
CN106658024A (en) Fast video coding method
Huang et al. Beyond VVC: Towards perceptual quality optimized video compression using multi-scale hybrid approaches
CN113132756B (en) Video coding and transcoding method
CN112104875B (en) Video encoding method, video encoding device and storage medium
Ma et al. A fast background model based surveillance video coding in HEVC
CN109274970B (en) Rapid scene switching detection method and system

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
EE01 Entry into force of recordation of patent licensing contract
EE01 Entry into force of recordation of patent licensing contract

Application publication date: 20200728

Assignee: Ningbo Science and Technology Innovation Association

Assignor: Ningbo University

Contract record no.: X2023980033633

Denomination of invention: A Fast Coding Method for Stereoscopic Panoramic Video Based on Panoramic Significance

Granted publication date: 20210810

License type: Common License

Record date: 20230317

TR01 Transfer of patent right
TR01 Transfer of patent right

Effective date of registration: 20230720

Address after: 100000 No. 159, Yard 1, Shuiding Road, Miaofengshan, Mentougou District, Beijing

Patentee after: Beijing Yiqilin Cultural Media Co.,Ltd.

Address before: Building C, No.888, Huanhu West 2nd Road, Lingang New District, China (Shanghai) pilot Free Trade Zone, Pudong New Area, Shanghai

Patentee before: Shanghai ruishenglian Information Technology Co.,Ltd.

Effective date of registration: 20230720

Address after: Building C, No.888, Huanhu West 2nd Road, Lingang New District, China (Shanghai) pilot Free Trade Zone, Pudong New Area, Shanghai

Patentee after: Shanghai ruishenglian Information Technology Co.,Ltd.

Address before: 315211, Fenghua Road, Jiangbei District, Zhejiang, Ningbo 818

Patentee before: Ningbo University

Effective date of registration: 20230720

Address after: Room 407, 4th Floor, No. 10 Shangdi Information Road, Haidian District, Beijing, 100000

Patentee after: Nantian Shujin (Beijing) Information Industry Development Co.,Ltd.

Address before: 100000 No. 159, Yard 1, Shuiding Road, Miaofengshan, Mentougou District, Beijing

Patentee before: Beijing Yiqilin Cultural Media Co.,Ltd.