WO2021253721A1 - 一种帧内预测方法、装置、终端及存储介质 - Google Patents
一种帧内预测方法、装置、终端及存储介质 Download PDFInfo
- Publication number
- WO2021253721A1 WO2021253721A1 PCT/CN2020/129553 CN2020129553W WO2021253721A1 WO 2021253721 A1 WO2021253721 A1 WO 2021253721A1 CN 2020129553 W CN2020129553 W CN 2020129553W WO 2021253721 A1 WO2021253721 A1 WO 2021253721A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- pixel
- image block
- pixel point
- intra
- circle
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 71
- 238000004590 computer program Methods 0.000 claims description 24
- 238000012545 processing Methods 0.000 abstract description 6
- 230000008569 process Effects 0.000 description 19
- 230000009466 transformation Effects 0.000 description 9
- 238000010586 diagram Methods 0.000 description 6
- 230000006870 function Effects 0.000 description 6
- 238000004422 calculation algorithm Methods 0.000 description 4
- 230000004044 response Effects 0.000 description 4
- 230000008878 coupling Effects 0.000 description 3
- 238000010168 coupling process Methods 0.000 description 3
- 238000005859 coupling reaction Methods 0.000 description 3
- 238000005516 engineering process Methods 0.000 description 3
- 238000004891 communication Methods 0.000 description 2
- 238000001514 detection method Methods 0.000 description 2
- 230000003190 augmentative effect Effects 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000005457 optimization Methods 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/103—Selection of coding mode or of prediction mode
- H04N19/11—Selection of coding mode or of prediction mode among a plurality of spatial predictive coding modes
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/134—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
- H04N19/146—Data rate or code amount at the encoder output
- H04N19/147—Data rate or code amount at the encoder output according to rate distortion criteria
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/17—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
- H04N19/176—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
Definitions
- This application belongs to the technical field of video processing, and in particular relates to an intra-frame prediction method, device, terminal, and storage medium.
- VR virtual reality
- AR augmented reality
- 360-degree video is one of the important forms of AR and VR.
- 360-degree video also provides a higher resolution and frame rate, and a better visual experience.
- the video usually needs to be encoded to achieve data storage or transmission.
- it is necessary to perform a projection transformation from a 3D sphere to a 2D plane before coding.
- the existing intra-frame prediction algorithms in video coding are designed for traditional 2D video.
- these algorithms do not consider the specific attributes and characteristics of the spherical domain and cannot be completely Eliminate spatial redundancy in 360-degree video, so that 360-degree video cannot achieve better coding performance.
- the embodiments of the application provide an intra-frame prediction method, device, terminal, and storage medium to solve the problem that the existing intra-frame prediction algorithm in video coding does not consider the specific attributes and characteristics of the spherical domain, and cannot completely eliminate the problem in the 360-degree video. Spatial redundancy is a problem that makes 360-degree video unable to achieve better coding performance.
- the first aspect of the embodiments of the present application provides an intra-frame prediction method, including:
- a reference pixel point corresponding to each pixel point in the image block is respectively determined from the neighborhood pixels, and each pixel point in the image block and its corresponding reference pixel point are located on the same preset arc ;
- the pixel value of the reference pixel corresponding to it is determined as the predicted value of the pixel.
- a second aspect of the embodiments of the present application provides an intra-frame prediction device, including:
- the first determining module is used to determine the neighboring pixels of the image block
- the second determining module is configured to determine the reference pixel corresponding to each pixel in the image block from the neighboring pixels, and each pixel in the image block and its corresponding reference pixel Located on the same preset arc;
- the third determining module is configured to determine the pixel value of the reference pixel corresponding to each pixel in the image block as the predicted value of the pixel.
- the third aspect of the embodiments of the present application provides a terminal, including a memory, a processor, and a computer program stored in the memory and capable of running on the processor.
- the processor executes the computer program when the computer program is executed. The steps of the method as described in the first aspect.
- the fourth aspect of the embodiments of the present application provides a computer-readable storage medium, the computer-readable storage medium stores a computer program, and the computer program implements the steps of the method described in the first aspect when the computer program is executed by a processor.
- the fifth aspect of the present application provides a computer program product, which when the computer program product runs on a terminal, causes the terminal to execute the steps of the method described in the first aspect.
- the reference pixels corresponding to each pixel in the image block are respectively determined from the neighboring pixels.
- Each pixel point of and its corresponding reference pixel point are located on the same preset arc; for each pixel point in the image block, the pixel value of the corresponding reference pixel point is determined as the predicted value of the pixel point.
- This process can meet the requirement of fully considering the specific attributes and characteristics of the spherical domain in the 360-degree video when performing intra-frame prediction based on the 2D planar video image obtained by the projection transformation, so that the execution of intra-frame prediction can be closer to the 360-degree video projection.
- the distribution of pixels in the 2D plane eliminates spatial redundancy in 360-degree videos, improves the accuracy of intra-frame prediction, and improves the final encoding performance of 360-degree videos.
- Fig. 1 is a first flowchart of an intra-frame prediction method provided by an embodiment of the present application
- Fig. 2 is a second flowchart of an intra-frame prediction method provided by an embodiment of the present application
- FIG. 3 is a schematic diagram 1 of an intra-frame prediction scene provided by an embodiment of the present application.
- FIG. 4 is a second schematic diagram of an intra-frame prediction scene provided by an embodiment of the present application.
- Fig. 5 is a structural diagram of an intra-frame prediction apparatus provided by an embodiment of the present application.
- Fig. 6 is a structural diagram of a terminal provided by an embodiment of the present application.
- the term “if” can be interpreted as “when” or “once” or “in response to determination” or “in response to detection” depending on the context .
- the phrase “if determined” or “if detected [described condition or event]” can be interpreted as meaning “once determined” or “in response to determination” or “once detected [described condition or event]” depending on the context ]” or “in response to detection of [condition or event described]”.
- the reference pixels corresponding to each pixel in the image block will be determined from the neighboring pixels.
- the pixel and its corresponding reference pixel are located on the same preset arc; for each pixel in the image block, the pixel value of the corresponding reference pixel is determined as the predicted value of the pixel.
- This process can meet the requirement of fully considering the specific attributes and characteristics of the spherical domain in the 360-degree video when performing intra-frame prediction based on the 2D planar video image obtained by the projection transformation, so that the execution of intra-frame prediction can be closer to the 360-degree video projection.
- the distribution of pixels in the 2D plane eliminates spatial redundancy in 360-degree videos, improves the accuracy of intra-frame prediction, and improves the final encoding performance of 360-degree videos.
- an intra-frame prediction method includes the following steps:
- Step 101 Determine the neighboring pixels of the image block.
- the image block is an image area in the image frame.
- An image block includes several pixels.
- the neighborhood pixel of the image block is the pixel in the area adjacent to the image block in the same image frame; more specifically, the neighborhood pixel is the edge pixel of the image block in the area adjacent to the image block Adjacent pixels.
- an image block is used as a pixel point prediction unit.
- the pixel value of each pixel included in the image block needs to be predicted separately until each pixel point is completed.
- intra-frame prediction the image block completes intra-frame prediction.
- the image block is a rectangular block.
- the neighborhood pixel may specifically be the left neighborhood pixel or the upper neighborhood pixel of the image block.
- Step 102 Determine a reference pixel point corresponding to each pixel point in the image block from the neighborhood pixels, and each pixel point in the image block and its corresponding reference pixel point are located in the same preset Arc.
- each pixel in the image block becomes a pixel to be predicted.
- intra-frame prediction on a pixel to be predicted in an image block, it is necessary to select a corresponding pixel from the neighboring pixels of the image block as a reference pixel. Specifically, a point on the same arc line as the selected pixel to be predicted among the neighboring pixels is used as the reference pixel corresponding to the pixel to be predicted.
- Step 103 For each pixel in the image block, the pixel value of the reference pixel corresponding to it is determined as the predicted value of the pixel.
- the pixel value of the reference pixel is used as the predicted value and directly assigned to the pixel value of the pixel to be predicted. . According to this operation, the intra-frame prediction of all pixels in the image block is completed.
- the reference pixels corresponding to each pixel in the image block are respectively determined from the neighboring pixels, and each pixel in the image block is The point and its corresponding reference pixel are located on the same preset arc; for each pixel in the image block, the pixel value of the corresponding reference pixel is determined as the predicted value of the pixel.
- This process can meet the requirement of fully considering the specific attributes and characteristics of the spherical domain in the 360-degree video when performing intra-frame prediction based on the 2D planar video image obtained by the projection transformation, so that the execution of intra-frame prediction can be closer to the 360-degree video projection.
- the distribution of pixels in the 2D plane eliminates spatial redundancy in 360-degree videos, improves the accuracy of intra-frame prediction, and improves the final encoding performance of 360-degree videos.
- the embodiments of the present application also provide different implementations of the intra prediction method.
- an intra-frame prediction method includes the following steps:
- Step 201 Determine the neighboring pixels of the image block.
- Step 202 Determine a pixel point located on a first circle in the neighborhood pixel point as a reference pixel point corresponding to a pixel point located on the first circle in the image block.
- the pixel point in the neighborhood pixel point that is located on the same circle as the pixel point in the image block is specifically determined as the reference pixel point.
- the first circle is the aforementioned preset arc.
- the determination of the pixel point located on the first circle in the neighborhood pixel point is a reference pixel point corresponding to the pixel point located on the first circle in the image block ,include:
- the pixel located on the first circle with the second pixel as the center is the reference pixel corresponding to the pixel located on the first circle in the image block .
- the distance between the center of the first circle and the center of the second circle (the first pixel in the image block to be predicted) is used to control the arc of the circle prediction to adapt to the distortion degree of the projection distortion.
- the first pixel point may be any pixel point in the image block, or a pixel point at the center position of the image block.
- the radius of the first circle can be specifically determined as required.
- one reference pixel corresponds to at least one pixel to be predicted.
- a circle is drawn with a radius of R with O(0,0) in the pixel block as the center.
- the pixel to be predicted is at least one; in the Figure 3 shows a plurality of pixels (including N(x, y)) located on the first circle in the pixel block, and pixels located on the circle (that is, the first circle) in the neighboring pixels of the pixel block Point P is used as the reference pixel of the pixel to be predicted.
- a circle is drawn with a radius R with O(0,0) in the pixel block as the center.
- the reference pixel when multiple concentric circles are drawn clockwise, the reference pixel is determined from the pixels in the left neighborhood among the neighborhood pixels of the image block. When multiple concentric circles are drawn counterclockwise, the reference pixel is determined from the upper neighboring pixel of the neighboring pixel of the image block. It can also be the opposite. This is only an exemplary description, and specific settings can be made according to actual needs in specific applications, and it is not limited to this.
- the method further includes:
- different M points can be selected to generate different prediction results respectively, and a better prediction result is selected from these prediction results as the final result.
- the value of the reference pixel is transferred to the pixel to be predicted in the image block until the pixel in the image block is filled.
- the intra-frame prediction method further includes:
- R ⁇ (Rmax-Rmin)+Rmin; where, ⁇ [0,1], ⁇ is the weight of the projection distortion degree, and the magnitude of the value of ⁇ is negative to the magnitude of the projection distortion degree Related; R is the specified length, Rmax is the maximum value of the specified length, and Rmin is the minimum value of the specified length.
- the content of the high-latitude area of the 360-degree video is over-sampled to a large extent and distorted and deformed to produce projection distortion. Therefore, in the 2D planar video image obtained by the transformation, for the high-latitude image area in the 360-degree video, the value of the above-mentioned specified length R should be smaller, and for the low-latitude image area in the 360-degree video, the value of the above-mentioned specified length R The value should be larger.
- the magnitude of the value of ⁇ is negatively related to the magnitude of the degree of projection distortion, that is, the smaller the value of ⁇ , the more serious the projection distortion, and the larger the value of ⁇ , the smaller the degree of projection distortion.
- the method further includes:
- the maximum value of the specified length and the minimum value of the specified length are set.
- the maximum value Rmax of the specified length can be set to 5l
- the minimum value Rmin of the specified length can be set to 0.8l
- l 2 w 2 +h 2
- w and h are the width and height of the image block.
- Step 203 For each pixel in the image block, the pixel value of the reference pixel corresponding to it is determined as the predicted value of the pixel.
- step 103 The implementation process of this step is the same as the implementation process of step 103 in the foregoing embodiment, and will not be repeated here.
- a specific circular prediction method is applied, which can adapt to the projection distortion from the 3D sphere to the 2D plane and realize the coding gain.
- intra-frame prediction methods involved in the aforementioned different embodiments can be applied to a video encoding method, which is applied to a video encoder, and the method includes:
- the at least two intra-frame prediction strategies include those described in the aforementioned different embodiments. Any one of the intra prediction methods involved; based on the rate-distortion cost value, the intra-frame prediction strategy with the least rate-distortion cost value is used as the target intra-frame prediction strategy; the target prediction strategy is calculated to frame the current image block The target prediction value of intra prediction; calculate the prediction residual according to the target prediction value and the real pixel value of the current image block; encode the video according to the target intra prediction strategy and the prediction residual, Generate video stream.
- the terminal device divides the video data into several frames of images, and each frame of image includes several image blocks.
- At least two intra prediction strategies include: traditional 2D image intra prediction methods (such as angular intra prediction (Angular Intra Prediction, AIP) methods) and any of the intra predictions involved in the embodiments of the present application method.
- traditional 2D image intra prediction methods such as angular intra prediction (Angular Intra Prediction, AIP) methods
- AIP Angular Intra Prediction
- the rate-distortion cost value is a variable that characterizes the prediction residual and the number of residual coding bits corresponding to the prediction residual.
- the prediction strategy with the least cost-effective rate-distortion is used as the target intra prediction strategy to reduce the bit rate of residual coding , Improve coding performance.
- encoding the video according to the target intra-frame prediction strategy and the prediction residual to generate a video code stream includes: setting a corresponding strategy for the current image block based on the target intra-frame prediction strategy Identification; encoding the video according to the strategy identification and the prediction residual to generate a video bitstream.
- the image block For different image blocks, determine the corresponding target intra-frame prediction strategy, and encode the strategy identifier corresponding to the target intra-frame prediction strategy into the video when encoding, and generate a video code stream to be able to target different
- the image block realizes the optimized coding operation and reduces the degree of video distortion.
- the above-mentioned intra-frame prediction method and video coding method can be used not only for the luminance component, but also for the chrominance component.
- different intra-frame prediction strategies are performed on both the luminance and chrominance components. Choose a strategy with a smaller rate-distortion cost as the target intra prediction strategy. Then, according to the selected target intra prediction strategy, the binary flag is set to 0 or 1, and encoding is performed. And send its code to the decoder.
- the strategy flag will be decoded before intra prediction. According to the strategy flag, the corresponding intra prediction will be performed. Then use the decoded residuals and intra-frame prediction results to reconstruct the 360-degree video.
- the intra-frame prediction methods involved in the aforementioned different embodiments can be applied to a video decoding method, which is applied to a video decoder, and the method includes:
- the at least two prediction strategies include any one of the intra-frame prediction methods involved in the aforementioned different embodiments; video reconstruction is performed according to the target prediction strategy and the prediction residual.
- decoding the video code stream to obtain a target intra prediction strategy corresponding to each image block includes: decoding the video code stream to obtain a strategy identifier corresponding to the image block; Strategy identification, determining the target intra-frame prediction strategy from the at least two prediction strategies.
- the reference pixels corresponding to each pixel in the image block are respectively determined from the neighboring pixels, and each pixel in the image block is The point and its corresponding reference pixel are located on the same preset arc; for each pixel in the image block, the pixel value of the corresponding reference pixel is determined as the predicted value of the pixel.
- This process can meet the requirement of fully considering the specific attributes and characteristics of the spherical domain in the 360-degree video when performing intra-frame prediction based on the 2D planar video image obtained by the projection transformation, so that the execution of intra-frame prediction can be closer to the 360-degree video projection.
- the distribution of pixels in the 2D plane eliminates spatial redundancy in 360-degree videos, improves the accuracy of intra-frame prediction, and improves the final encoding performance of 360-degree videos.
- FIG. 5 is a structural diagram of an intra-frame prediction apparatus provided by an embodiment of the present application. For ease of description, only parts related to the embodiment of the present application are shown.
- the intra prediction device 500 includes:
- the first determining module 501 is configured to determine the neighboring pixels of the image block
- the second determining module 502 is configured to respectively determine a reference pixel corresponding to each pixel in the image block from the neighboring pixels, and each pixel in the image block and its corresponding reference pixel The points are on the same preset arc;
- the third determining module 503 is configured to determine the pixel value of the reference pixel corresponding to each pixel in the image block as the predicted value of the pixel.
- the second determining module 502 includes:
- the determining sub-module is configured to determine a pixel point located on the first circle in the neighborhood pixel point, which is a reference pixel point corresponding to a pixel point located on the first circle in the image block.
- the determining sub-module is specifically used for:
- the pixel located on the first circle with the second pixel as the center is the reference pixel corresponding to the pixel located on the first circle in the image block .
- the intra-frame prediction apparatus 500 further includes:
- R ⁇ (Rmax-Rmin)+Rmin; where, ⁇ [0,1], ⁇ is the weight of the projection distortion degree, and the magnitude of the value of ⁇ is negative to the magnitude of the projection distortion degree Related; R is the specified length, Rmax is the maximum value of the specified length, and Rmin is the minimum value of the specified length.
- the intra-frame prediction apparatus 500 further includes:
- the setting module is configured to set the maximum value of the specified length and the minimum value of the specified length according to the width and height of the image block.
- the intra-frame prediction device provided in the embodiment of the present application can implement the various processes of the above-mentioned intra-frame prediction device method embodiment, and can achieve the same technical effect. In order to avoid repetition, details are not repeated here.
- Fig. 6 is a structural diagram of a terminal provided by an embodiment of the present application.
- the terminal 6 of this embodiment includes: at least one processor 60 (only one is shown in FIG. 6), a memory 61, and is stored in the memory 61 and can be stored on the at least one processor 60 A running computer program 62, when the processor 60 executes the computer program 62, the steps in any of the foregoing method embodiments are implemented.
- the terminal 6 may be a computing device such as a desktop computer, a notebook, a palmtop computer, and a cloud server.
- the terminal 6 may include, but is not limited to, a processor 60 and a memory 61.
- FIG. 6 is only an example of the terminal 6 and does not constitute a limitation on the terminal 6. It may include more or less components than shown in the figure, or a combination of certain components, or different components, such as
- the terminal may also include input and output devices, network access devices, buses, and so on.
- the so-called processor 60 may be a central processing unit (Central Processing Unit, CPU), other general-purpose processors, digital signal processors (Digital Signal Processor, DSP), application specific integrated circuits (Application Specific Integrated Circuit, ASIC), Ready-made programmable gate array (Field-Programmable Gate Array, FPGA) or other programmable logic devices, discrete gates or transistor logic devices, discrete hardware components, etc.
- the general-purpose processor may be a microprocessor or the processor may also be any conventional processor or the like.
- the memory 61 may be an internal storage unit of the terminal 6, such as a hard disk or memory of the terminal 6.
- the memory 61 may also be an external storage device of the terminal 6, such as a plug-in hard disk equipped on the terminal 6, a smart memory card (Smart Media Card, SMC), or a Secure Digital (SD) card, Flash Card, etc. Further, the memory 61 may also include both an internal storage unit of the terminal 6 and an external storage device.
- the memory 61 is used to store the computer program and other programs and data required by the terminal.
- the memory 61 can also be used to temporarily store data that has been output or will be output.
- the disclosed device/terminal and method may be implemented in other ways.
- the device/terminal embodiments described above are only illustrative.
- the division of the modules or units is only a logical function division, and there may be other divisions in actual implementation, such as multiple units or Components can be combined or integrated into another system, or some features can be omitted or not implemented.
- the displayed or discussed mutual coupling or direct coupling or communication connection may be indirect coupling or communication connection through some interfaces, devices or units, and may be in electrical, mechanical or other forms.
- the units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, they may be located in one place, or they may be distributed on multiple network units. Some or all of the units may be selected according to actual needs to achieve the objectives of the solutions of the embodiments.
- the functional units in the various embodiments of the present application may be integrated into one processing unit, or each unit may exist alone physically, or two or more units may be integrated into one unit.
- the above-mentioned integrated unit can be implemented in the form of hardware or software functional unit.
- the integrated module/unit is implemented in the form of a software functional unit and sold or used as an independent product, it can be stored in a computer readable storage medium.
- this application implements all or part of the processes in the above-mentioned embodiments and methods, and can also be completed by instructing relevant hardware through a computer program.
- the computer program can be stored in a computer-readable storage medium. When the program is executed by the processor, it can implement the steps of the foregoing method embodiments.
- the computer program includes computer program code, and the computer program code may be in the form of source code, object code, executable file, or some intermediate form.
- the computer-readable medium may include: any entity or device capable of carrying the computer program code, recording medium, U disk, mobile hard disk, magnetic disk, optical disk, computer memory, read-only memory (ROM, Read-Only Memory) , Random Access Memory (RAM, Random Access Memory), electrical carrier signal, telecommunications signal, and software distribution media, etc.
- ROM Read-Only Memory
- RAM Random Access Memory
- electrical carrier signal telecommunications signal
- software distribution media etc.
- the content contained in the computer-readable medium can be appropriately added or deleted according to the requirements of the legislation and patent practice in the jurisdiction.
- the computer-readable medium Does not include electrical carrier signals and telecommunication signals.
- This application implements all or part of the processes in the above-mentioned embodiment methods, and can also be implemented by a computer program product.
- the computer program product runs on a terminal, the steps in each of the above-mentioned method embodiments can be realized when the terminal is executed. .
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
Abstract
Description
Claims (10)
- 一种帧内预测方法,其特征在于,包括:确定图像块的邻域像素点;从所述邻域像素点中分别确定与所述图像块中的每个像素点对应的参考像素点,所述图像块中的每个像素点与其对应的参考像素点位于同一预设弧线上;对于所述图像块中的每个像素点,将与其对应的所述参考像素点的像素值确定为所述像素点的预测值。
- 根据权利要求1所述的帧内预测方法,其特征在于,所述从所述邻域像素点中分别确定与所述图像块中的每个像素点对应的参考像素点,包括:确定所述邻域像素点中位于第一圆周上的像素点,为与所述图像块中位于所述第一圆周上的像素点对应的参考像素点。
- 根据权利要求2所述的帧内预测方法,其特征在于,所述确定所述邻域像素点中位于第一圆周上的像素点,为与所述图像块中位于所述第一圆周上的像素点对应的参考像素点,包括:从以所述图像块中的第一像素点为圆心,以指定长度为半径的第二圆周上选取第二像素点;确定所述邻域像素点中,位于以所述第二像素点为圆心的第一圆周上的像素点,为与所述图像块中位于所述第一圆周上的像素点对应的参考像素点。
- 根据权利要求3所述的帧内预测方法,其特征在于,还包括:基于投影失真度,计算所述指定长度R;其中,R=ρ×(Rmax–Rmin)+Rmin;其中,ρ∈[0,1],ρ为所述投影失真度的权重,所述ρ的值的大小与所述投影失真度的大小负相关;R为所述指定长度,Rmax为所述指定长度的最大值,Rmin为所述指定长度的最小值。
- 根据权利要求4所述的帧内预测方法,其特征在于,还包括:根据所述图像块的宽度和高度,设置所述指定长度的最大值与所述指定长度的最小值。
- 一种帧内预测装置,其特征在于,包括:第一确定模块,用于确定图像块的邻域像素点;第二确定模块,用于从所述邻域像素点中分别确定与所述图像块中的每个像素点对应的参考像素点,所述图像块中的每个像素点与其对应的参考像素点位于同一预设弧线上;第三确定模块,用于对于所述图像块中的每个像素点,将与其对应的所述参考像素点的像素值确定为所述像素点的预测值。
- 根据权利要求6所述的帧内预测装置,其特征在于,所述第二确定模块,包括:确定子模块,用于确定所述邻域像素点中位于第一圆周上的像素点,为与所述图像块中位于所述第一圆周上的像素点对应的参考像素点。
- 根据权利要求7所述的帧内预测装置,其特征在于,所述确定子模块,具体用于:从以所述图像块中的第一像素点为圆心,以指定长度为半径的第二圆周上选取第二像素点;确定所述邻域像素点中,位于以所述第二像素点为圆心的第一圆周上的像素点,为与所述图像块中位于所述第一圆周上的像素点对应的参考像素点。
- 一种终端,包括存储器、处理器以及存储在所述存储器中并可在所述处理器上运行的计算机程序,其特征在于,所述处理器执行所述计算机程序时实现如权利要求1至5中任一项所述方法的步骤。
- 一种计算机可读存储介质,所述计算机可读存储介质存储有计算机程序,其特征在于,所述计算机程序被处理器执行时实现如权利要求1至5中任一项所述方法的步骤。
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010550240.4 | 2020-06-16 | ||
CN202010550240.4A CN111818333B (zh) | 2020-06-16 | 2020-06-16 | 一种帧内预测方法、装置、终端及存储介质 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2021253721A1 true WO2021253721A1 (zh) | 2021-12-23 |
Family
ID=72846572
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2020/129553 WO2021253721A1 (zh) | 2020-06-16 | 2020-11-17 | 一种帧内预测方法、装置、终端及存储介质 |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN111818333B (zh) |
WO (1) | WO2021253721A1 (zh) |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111818333B (zh) * | 2020-06-16 | 2022-04-29 | 中国科学院深圳先进技术研究院 | 一种帧内预测方法、装置、终端及存储介质 |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP1162830A2 (en) * | 2000-06-07 | 2001-12-12 | Be Here Corporation | Method and apparatus for electronically distributing motion panoramic images |
CN101006730A (zh) * | 2004-08-13 | 2007-07-25 | 庆熙大学校产学协力团 | 用于对环形图像中的基准像素进行插值以及对环形图像进行编码/解码的方法和设备 |
WO2018199792A1 (en) * | 2017-04-26 | 2018-11-01 | Huawei Technologies Co., Ltd | Apparatuses and methods for encoding and decoding a panoramic video signal |
WO2018199793A1 (en) * | 2017-04-27 | 2018-11-01 | Huawei Technologies Co., Ltd | Geodesic intra-prediction for panoramic video coding |
CN109804633A (zh) * | 2016-09-30 | 2019-05-24 | 交互数字Vc控股公司 | 利用自适应帧内预测的全向视频编码和解码的方法和装置 |
CN111066322A (zh) * | 2017-06-14 | 2020-04-24 | 华为技术有限公司 | 通过透视信息进行视频编码的帧内预测 |
CN111818333A (zh) * | 2020-06-16 | 2020-10-23 | 中国科学院深圳先进技术研究院 | 一种帧内预测方法、装置、终端及存储介质 |
Family Cites Families (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7817717B2 (en) * | 2002-06-18 | 2010-10-19 | Qualcomm Incorporated | Motion estimation techniques for video encoding |
EP2129135B1 (en) * | 2007-02-23 | 2012-08-15 | Nippon Telegraph and Telephone Corporation | Video coding intra prediction method and decoding method, their device, their program and program-recorded medium |
CN102685504B (zh) * | 2011-03-10 | 2015-08-19 | 华为技术有限公司 | 视频图像的编解码方法、编码装置、解码装置及其系统 |
CN103533374B (zh) * | 2012-07-06 | 2018-02-16 | 乐金电子(中国)研究开发中心有限公司 | 一种视频编码、解码的方法及装置 |
JP6005572B2 (ja) * | 2013-03-28 | 2016-10-12 | Kddi株式会社 | 動画像符号化装置、動画像復号装置、動画像符号化方法、動画像復号方法、およびプログラム |
EP3556096B1 (en) * | 2016-12-22 | 2024-05-22 | HFI Innovation Inc. | Method and apparatus of motion refinement for video coding |
CN110677644B (zh) * | 2018-07-03 | 2021-11-16 | 北京大学 | 一种视频编码、解码方法及视频编码帧内预测器 |
-
2020
- 2020-06-16 CN CN202010550240.4A patent/CN111818333B/zh active Active
- 2020-11-17 WO PCT/CN2020/129553 patent/WO2021253721A1/zh active Application Filing
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP1162830A2 (en) * | 2000-06-07 | 2001-12-12 | Be Here Corporation | Method and apparatus for electronically distributing motion panoramic images |
CN101006730A (zh) * | 2004-08-13 | 2007-07-25 | 庆熙大学校产学协力团 | 用于对环形图像中的基准像素进行插值以及对环形图像进行编码/解码的方法和设备 |
CN109804633A (zh) * | 2016-09-30 | 2019-05-24 | 交互数字Vc控股公司 | 利用自适应帧内预测的全向视频编码和解码的方法和装置 |
WO2018199792A1 (en) * | 2017-04-26 | 2018-11-01 | Huawei Technologies Co., Ltd | Apparatuses and methods for encoding and decoding a panoramic video signal |
WO2018199793A1 (en) * | 2017-04-27 | 2018-11-01 | Huawei Technologies Co., Ltd | Geodesic intra-prediction for panoramic video coding |
CN111066322A (zh) * | 2017-06-14 | 2020-04-24 | 华为技术有限公司 | 通过透视信息进行视频编码的帧内预测 |
CN111818333A (zh) * | 2020-06-16 | 2020-10-23 | 中国科学院深圳先进技术研究院 | 一种帧内预测方法、装置、终端及存储介质 |
Also Published As
Publication number | Publication date |
---|---|
CN111818333B (zh) | 2022-04-29 |
CN111818333A (zh) | 2020-10-23 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20190273929A1 (en) | De-Blocking Filtering Method and Terminal | |
RU2716156C1 (ru) | Способ и устройство для составления всенаправленного изображения в выходной формат макета | |
US10560678B2 (en) | Method and apparatus having video encoding function with syntax element signaling of rotation information of content-oriented rotation applied to 360-degree image content or 360-degree video content represented in projection format and associated method and apparatus having video decoding function | |
WO2020001565A1 (zh) | 点云编解码方法和编解码器 | |
WO2021036795A1 (zh) | 视频超分辨率处理方法及装置 | |
CN110913218A (zh) | 一种视频帧预测方法、装置及终端设备 | |
CN111971969A (zh) | 用于对点云的几何形状进行编码的方法和装置 | |
WO2023029858A1 (zh) | 点云媒体文件的封装与解封装方法、装置及存储介质 | |
WO2024037137A1 (zh) | 一种沉浸媒体的数据处理方法、装置、设备、介质和产品 | |
WO2021253721A1 (zh) | 一种帧内预测方法、装置、终端及存储介质 | |
WO2022121650A1 (zh) | 点云属性的预测方法、编码器、解码器及存储介质 | |
US20200068205A1 (en) | Geodesic intra-prediction for panoramic video coding | |
WO2019037558A1 (zh) | 图像处理方法和装置 | |
US11240512B2 (en) | Intra-prediction for video coding using perspective information | |
CA3110477C (en) | Picture partitioning method and apparatus | |
TWI696979B (zh) | 圖像處理方法和裝置及全景視頻圖像的處理方法和系統 | |
WO2021114100A1 (zh) | 帧内预测方法、视频编码、解码方法及相关设备 | |
CN104737539A (zh) | 减小含有相互类似的信息的多个图像的数据量的方法、程序及装置 | |
JP2010528516A (ja) | 画像を処理する方法およびシステム | |
CN111246208A (zh) | 视频处理方法、装置及电子设备 | |
Cai et al. | Towards 6DoF live video streaming system for immersive media | |
CN111885417B (zh) | Vr视频播放方法、装置、设备以及存储介质 | |
CN110830806A (zh) | 一种视频帧预测方法、装置及终端设备 | |
CN110115026B (zh) | 在电子装置中在矩形投影上产生360度内容的方法和系统 | |
WO2024051299A1 (zh) | 一种编解码方法及装置 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 20941219 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20941219 Country of ref document: EP Kind code of ref document: A1 |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20941219 Country of ref document: EP Kind code of ref document: A1 |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20941219 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 09/08/2023) |