WO2021253721A1 - 一种帧内预测方法、装置、终端及存储介质 - Google Patents

一种帧内预测方法、装置、终端及存储介质 Download PDF

Info

Publication number
WO2021253721A1
WO2021253721A1 PCT/CN2020/129553 CN2020129553W WO2021253721A1 WO 2021253721 A1 WO2021253721 A1 WO 2021253721A1 CN 2020129553 W CN2020129553 W CN 2020129553W WO 2021253721 A1 WO2021253721 A1 WO 2021253721A1
Authority
WO
WIPO (PCT)
Prior art keywords
pixel
image block
pixel point
intra
circle
Prior art date
Application number
PCT/CN2020/129553
Other languages
English (en)
French (fr)
Inventor
张云
朱林卫
李娜
皮金勇
Original Assignee
中国科学院深圳先进技术研究院
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 中国科学院深圳先进技术研究院 filed Critical 中国科学院深圳先进技术研究院
Publication of WO2021253721A1 publication Critical patent/WO2021253721A1/zh

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • H04N19/11Selection of coding mode or of prediction mode among a plurality of spatial predictive coding modes
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/146Data rate or code amount at the encoder output
    • H04N19/147Data rate or code amount at the encoder output according to rate distortion criteria
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/176Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock

Definitions

  • This application belongs to the technical field of video processing, and in particular relates to an intra-frame prediction method, device, terminal, and storage medium.
  • VR virtual reality
  • AR augmented reality
  • 360-degree video is one of the important forms of AR and VR.
  • 360-degree video also provides a higher resolution and frame rate, and a better visual experience.
  • the video usually needs to be encoded to achieve data storage or transmission.
  • it is necessary to perform a projection transformation from a 3D sphere to a 2D plane before coding.
  • the existing intra-frame prediction algorithms in video coding are designed for traditional 2D video.
  • these algorithms do not consider the specific attributes and characteristics of the spherical domain and cannot be completely Eliminate spatial redundancy in 360-degree video, so that 360-degree video cannot achieve better coding performance.
  • the embodiments of the application provide an intra-frame prediction method, device, terminal, and storage medium to solve the problem that the existing intra-frame prediction algorithm in video coding does not consider the specific attributes and characteristics of the spherical domain, and cannot completely eliminate the problem in the 360-degree video. Spatial redundancy is a problem that makes 360-degree video unable to achieve better coding performance.
  • the first aspect of the embodiments of the present application provides an intra-frame prediction method, including:
  • a reference pixel point corresponding to each pixel point in the image block is respectively determined from the neighborhood pixels, and each pixel point in the image block and its corresponding reference pixel point are located on the same preset arc ;
  • the pixel value of the reference pixel corresponding to it is determined as the predicted value of the pixel.
  • a second aspect of the embodiments of the present application provides an intra-frame prediction device, including:
  • the first determining module is used to determine the neighboring pixels of the image block
  • the second determining module is configured to determine the reference pixel corresponding to each pixel in the image block from the neighboring pixels, and each pixel in the image block and its corresponding reference pixel Located on the same preset arc;
  • the third determining module is configured to determine the pixel value of the reference pixel corresponding to each pixel in the image block as the predicted value of the pixel.
  • the third aspect of the embodiments of the present application provides a terminal, including a memory, a processor, and a computer program stored in the memory and capable of running on the processor.
  • the processor executes the computer program when the computer program is executed. The steps of the method as described in the first aspect.
  • the fourth aspect of the embodiments of the present application provides a computer-readable storage medium, the computer-readable storage medium stores a computer program, and the computer program implements the steps of the method described in the first aspect when the computer program is executed by a processor.
  • the fifth aspect of the present application provides a computer program product, which when the computer program product runs on a terminal, causes the terminal to execute the steps of the method described in the first aspect.
  • the reference pixels corresponding to each pixel in the image block are respectively determined from the neighboring pixels.
  • Each pixel point of and its corresponding reference pixel point are located on the same preset arc; for each pixel point in the image block, the pixel value of the corresponding reference pixel point is determined as the predicted value of the pixel point.
  • This process can meet the requirement of fully considering the specific attributes and characteristics of the spherical domain in the 360-degree video when performing intra-frame prediction based on the 2D planar video image obtained by the projection transformation, so that the execution of intra-frame prediction can be closer to the 360-degree video projection.
  • the distribution of pixels in the 2D plane eliminates spatial redundancy in 360-degree videos, improves the accuracy of intra-frame prediction, and improves the final encoding performance of 360-degree videos.
  • Fig. 1 is a first flowchart of an intra-frame prediction method provided by an embodiment of the present application
  • Fig. 2 is a second flowchart of an intra-frame prediction method provided by an embodiment of the present application
  • FIG. 3 is a schematic diagram 1 of an intra-frame prediction scene provided by an embodiment of the present application.
  • FIG. 4 is a second schematic diagram of an intra-frame prediction scene provided by an embodiment of the present application.
  • Fig. 5 is a structural diagram of an intra-frame prediction apparatus provided by an embodiment of the present application.
  • Fig. 6 is a structural diagram of a terminal provided by an embodiment of the present application.
  • the term “if” can be interpreted as “when” or “once” or “in response to determination” or “in response to detection” depending on the context .
  • the phrase “if determined” or “if detected [described condition or event]” can be interpreted as meaning “once determined” or “in response to determination” or “once detected [described condition or event]” depending on the context ]” or “in response to detection of [condition or event described]”.
  • the reference pixels corresponding to each pixel in the image block will be determined from the neighboring pixels.
  • the pixel and its corresponding reference pixel are located on the same preset arc; for each pixel in the image block, the pixel value of the corresponding reference pixel is determined as the predicted value of the pixel.
  • This process can meet the requirement of fully considering the specific attributes and characteristics of the spherical domain in the 360-degree video when performing intra-frame prediction based on the 2D planar video image obtained by the projection transformation, so that the execution of intra-frame prediction can be closer to the 360-degree video projection.
  • the distribution of pixels in the 2D plane eliminates spatial redundancy in 360-degree videos, improves the accuracy of intra-frame prediction, and improves the final encoding performance of 360-degree videos.
  • an intra-frame prediction method includes the following steps:
  • Step 101 Determine the neighboring pixels of the image block.
  • the image block is an image area in the image frame.
  • An image block includes several pixels.
  • the neighborhood pixel of the image block is the pixel in the area adjacent to the image block in the same image frame; more specifically, the neighborhood pixel is the edge pixel of the image block in the area adjacent to the image block Adjacent pixels.
  • an image block is used as a pixel point prediction unit.
  • the pixel value of each pixel included in the image block needs to be predicted separately until each pixel point is completed.
  • intra-frame prediction the image block completes intra-frame prediction.
  • the image block is a rectangular block.
  • the neighborhood pixel may specifically be the left neighborhood pixel or the upper neighborhood pixel of the image block.
  • Step 102 Determine a reference pixel point corresponding to each pixel point in the image block from the neighborhood pixels, and each pixel point in the image block and its corresponding reference pixel point are located in the same preset Arc.
  • each pixel in the image block becomes a pixel to be predicted.
  • intra-frame prediction on a pixel to be predicted in an image block, it is necessary to select a corresponding pixel from the neighboring pixels of the image block as a reference pixel. Specifically, a point on the same arc line as the selected pixel to be predicted among the neighboring pixels is used as the reference pixel corresponding to the pixel to be predicted.
  • Step 103 For each pixel in the image block, the pixel value of the reference pixel corresponding to it is determined as the predicted value of the pixel.
  • the pixel value of the reference pixel is used as the predicted value and directly assigned to the pixel value of the pixel to be predicted. . According to this operation, the intra-frame prediction of all pixels in the image block is completed.
  • the reference pixels corresponding to each pixel in the image block are respectively determined from the neighboring pixels, and each pixel in the image block is The point and its corresponding reference pixel are located on the same preset arc; for each pixel in the image block, the pixel value of the corresponding reference pixel is determined as the predicted value of the pixel.
  • This process can meet the requirement of fully considering the specific attributes and characteristics of the spherical domain in the 360-degree video when performing intra-frame prediction based on the 2D planar video image obtained by the projection transformation, so that the execution of intra-frame prediction can be closer to the 360-degree video projection.
  • the distribution of pixels in the 2D plane eliminates spatial redundancy in 360-degree videos, improves the accuracy of intra-frame prediction, and improves the final encoding performance of 360-degree videos.
  • the embodiments of the present application also provide different implementations of the intra prediction method.
  • an intra-frame prediction method includes the following steps:
  • Step 201 Determine the neighboring pixels of the image block.
  • Step 202 Determine a pixel point located on a first circle in the neighborhood pixel point as a reference pixel point corresponding to a pixel point located on the first circle in the image block.
  • the pixel point in the neighborhood pixel point that is located on the same circle as the pixel point in the image block is specifically determined as the reference pixel point.
  • the first circle is the aforementioned preset arc.
  • the determination of the pixel point located on the first circle in the neighborhood pixel point is a reference pixel point corresponding to the pixel point located on the first circle in the image block ,include:
  • the pixel located on the first circle with the second pixel as the center is the reference pixel corresponding to the pixel located on the first circle in the image block .
  • the distance between the center of the first circle and the center of the second circle (the first pixel in the image block to be predicted) is used to control the arc of the circle prediction to adapt to the distortion degree of the projection distortion.
  • the first pixel point may be any pixel point in the image block, or a pixel point at the center position of the image block.
  • the radius of the first circle can be specifically determined as required.
  • one reference pixel corresponds to at least one pixel to be predicted.
  • a circle is drawn with a radius of R with O(0,0) in the pixel block as the center.
  • the pixel to be predicted is at least one; in the Figure 3 shows a plurality of pixels (including N(x, y)) located on the first circle in the pixel block, and pixels located on the circle (that is, the first circle) in the neighboring pixels of the pixel block Point P is used as the reference pixel of the pixel to be predicted.
  • a circle is drawn with a radius R with O(0,0) in the pixel block as the center.
  • the reference pixel when multiple concentric circles are drawn clockwise, the reference pixel is determined from the pixels in the left neighborhood among the neighborhood pixels of the image block. When multiple concentric circles are drawn counterclockwise, the reference pixel is determined from the upper neighboring pixel of the neighboring pixel of the image block. It can also be the opposite. This is only an exemplary description, and specific settings can be made according to actual needs in specific applications, and it is not limited to this.
  • the method further includes:
  • different M points can be selected to generate different prediction results respectively, and a better prediction result is selected from these prediction results as the final result.
  • the value of the reference pixel is transferred to the pixel to be predicted in the image block until the pixel in the image block is filled.
  • the intra-frame prediction method further includes:
  • R ⁇ (Rmax-Rmin)+Rmin; where, ⁇ [0,1], ⁇ is the weight of the projection distortion degree, and the magnitude of the value of ⁇ is negative to the magnitude of the projection distortion degree Related; R is the specified length, Rmax is the maximum value of the specified length, and Rmin is the minimum value of the specified length.
  • the content of the high-latitude area of the 360-degree video is over-sampled to a large extent and distorted and deformed to produce projection distortion. Therefore, in the 2D planar video image obtained by the transformation, for the high-latitude image area in the 360-degree video, the value of the above-mentioned specified length R should be smaller, and for the low-latitude image area in the 360-degree video, the value of the above-mentioned specified length R The value should be larger.
  • the magnitude of the value of ⁇ is negatively related to the magnitude of the degree of projection distortion, that is, the smaller the value of ⁇ , the more serious the projection distortion, and the larger the value of ⁇ , the smaller the degree of projection distortion.
  • the method further includes:
  • the maximum value of the specified length and the minimum value of the specified length are set.
  • the maximum value Rmax of the specified length can be set to 5l
  • the minimum value Rmin of the specified length can be set to 0.8l
  • l 2 w 2 +h 2
  • w and h are the width and height of the image block.
  • Step 203 For each pixel in the image block, the pixel value of the reference pixel corresponding to it is determined as the predicted value of the pixel.
  • step 103 The implementation process of this step is the same as the implementation process of step 103 in the foregoing embodiment, and will not be repeated here.
  • a specific circular prediction method is applied, which can adapt to the projection distortion from the 3D sphere to the 2D plane and realize the coding gain.
  • intra-frame prediction methods involved in the aforementioned different embodiments can be applied to a video encoding method, which is applied to a video encoder, and the method includes:
  • the at least two intra-frame prediction strategies include those described in the aforementioned different embodiments. Any one of the intra prediction methods involved; based on the rate-distortion cost value, the intra-frame prediction strategy with the least rate-distortion cost value is used as the target intra-frame prediction strategy; the target prediction strategy is calculated to frame the current image block The target prediction value of intra prediction; calculate the prediction residual according to the target prediction value and the real pixel value of the current image block; encode the video according to the target intra prediction strategy and the prediction residual, Generate video stream.
  • the terminal device divides the video data into several frames of images, and each frame of image includes several image blocks.
  • At least two intra prediction strategies include: traditional 2D image intra prediction methods (such as angular intra prediction (Angular Intra Prediction, AIP) methods) and any of the intra predictions involved in the embodiments of the present application method.
  • traditional 2D image intra prediction methods such as angular intra prediction (Angular Intra Prediction, AIP) methods
  • AIP Angular Intra Prediction
  • the rate-distortion cost value is a variable that characterizes the prediction residual and the number of residual coding bits corresponding to the prediction residual.
  • the prediction strategy with the least cost-effective rate-distortion is used as the target intra prediction strategy to reduce the bit rate of residual coding , Improve coding performance.
  • encoding the video according to the target intra-frame prediction strategy and the prediction residual to generate a video code stream includes: setting a corresponding strategy for the current image block based on the target intra-frame prediction strategy Identification; encoding the video according to the strategy identification and the prediction residual to generate a video bitstream.
  • the image block For different image blocks, determine the corresponding target intra-frame prediction strategy, and encode the strategy identifier corresponding to the target intra-frame prediction strategy into the video when encoding, and generate a video code stream to be able to target different
  • the image block realizes the optimized coding operation and reduces the degree of video distortion.
  • the above-mentioned intra-frame prediction method and video coding method can be used not only for the luminance component, but also for the chrominance component.
  • different intra-frame prediction strategies are performed on both the luminance and chrominance components. Choose a strategy with a smaller rate-distortion cost as the target intra prediction strategy. Then, according to the selected target intra prediction strategy, the binary flag is set to 0 or 1, and encoding is performed. And send its code to the decoder.
  • the strategy flag will be decoded before intra prediction. According to the strategy flag, the corresponding intra prediction will be performed. Then use the decoded residuals and intra-frame prediction results to reconstruct the 360-degree video.
  • the intra-frame prediction methods involved in the aforementioned different embodiments can be applied to a video decoding method, which is applied to a video decoder, and the method includes:
  • the at least two prediction strategies include any one of the intra-frame prediction methods involved in the aforementioned different embodiments; video reconstruction is performed according to the target prediction strategy and the prediction residual.
  • decoding the video code stream to obtain a target intra prediction strategy corresponding to each image block includes: decoding the video code stream to obtain a strategy identifier corresponding to the image block; Strategy identification, determining the target intra-frame prediction strategy from the at least two prediction strategies.
  • the reference pixels corresponding to each pixel in the image block are respectively determined from the neighboring pixels, and each pixel in the image block is The point and its corresponding reference pixel are located on the same preset arc; for each pixel in the image block, the pixel value of the corresponding reference pixel is determined as the predicted value of the pixel.
  • This process can meet the requirement of fully considering the specific attributes and characteristics of the spherical domain in the 360-degree video when performing intra-frame prediction based on the 2D planar video image obtained by the projection transformation, so that the execution of intra-frame prediction can be closer to the 360-degree video projection.
  • the distribution of pixels in the 2D plane eliminates spatial redundancy in 360-degree videos, improves the accuracy of intra-frame prediction, and improves the final encoding performance of 360-degree videos.
  • FIG. 5 is a structural diagram of an intra-frame prediction apparatus provided by an embodiment of the present application. For ease of description, only parts related to the embodiment of the present application are shown.
  • the intra prediction device 500 includes:
  • the first determining module 501 is configured to determine the neighboring pixels of the image block
  • the second determining module 502 is configured to respectively determine a reference pixel corresponding to each pixel in the image block from the neighboring pixels, and each pixel in the image block and its corresponding reference pixel The points are on the same preset arc;
  • the third determining module 503 is configured to determine the pixel value of the reference pixel corresponding to each pixel in the image block as the predicted value of the pixel.
  • the second determining module 502 includes:
  • the determining sub-module is configured to determine a pixel point located on the first circle in the neighborhood pixel point, which is a reference pixel point corresponding to a pixel point located on the first circle in the image block.
  • the determining sub-module is specifically used for:
  • the pixel located on the first circle with the second pixel as the center is the reference pixel corresponding to the pixel located on the first circle in the image block .
  • the intra-frame prediction apparatus 500 further includes:
  • R ⁇ (Rmax-Rmin)+Rmin; where, ⁇ [0,1], ⁇ is the weight of the projection distortion degree, and the magnitude of the value of ⁇ is negative to the magnitude of the projection distortion degree Related; R is the specified length, Rmax is the maximum value of the specified length, and Rmin is the minimum value of the specified length.
  • the intra-frame prediction apparatus 500 further includes:
  • the setting module is configured to set the maximum value of the specified length and the minimum value of the specified length according to the width and height of the image block.
  • the intra-frame prediction device provided in the embodiment of the present application can implement the various processes of the above-mentioned intra-frame prediction device method embodiment, and can achieve the same technical effect. In order to avoid repetition, details are not repeated here.
  • Fig. 6 is a structural diagram of a terminal provided by an embodiment of the present application.
  • the terminal 6 of this embodiment includes: at least one processor 60 (only one is shown in FIG. 6), a memory 61, and is stored in the memory 61 and can be stored on the at least one processor 60 A running computer program 62, when the processor 60 executes the computer program 62, the steps in any of the foregoing method embodiments are implemented.
  • the terminal 6 may be a computing device such as a desktop computer, a notebook, a palmtop computer, and a cloud server.
  • the terminal 6 may include, but is not limited to, a processor 60 and a memory 61.
  • FIG. 6 is only an example of the terminal 6 and does not constitute a limitation on the terminal 6. It may include more or less components than shown in the figure, or a combination of certain components, or different components, such as
  • the terminal may also include input and output devices, network access devices, buses, and so on.
  • the so-called processor 60 may be a central processing unit (Central Processing Unit, CPU), other general-purpose processors, digital signal processors (Digital Signal Processor, DSP), application specific integrated circuits (Application Specific Integrated Circuit, ASIC), Ready-made programmable gate array (Field-Programmable Gate Array, FPGA) or other programmable logic devices, discrete gates or transistor logic devices, discrete hardware components, etc.
  • the general-purpose processor may be a microprocessor or the processor may also be any conventional processor or the like.
  • the memory 61 may be an internal storage unit of the terminal 6, such as a hard disk or memory of the terminal 6.
  • the memory 61 may also be an external storage device of the terminal 6, such as a plug-in hard disk equipped on the terminal 6, a smart memory card (Smart Media Card, SMC), or a Secure Digital (SD) card, Flash Card, etc. Further, the memory 61 may also include both an internal storage unit of the terminal 6 and an external storage device.
  • the memory 61 is used to store the computer program and other programs and data required by the terminal.
  • the memory 61 can also be used to temporarily store data that has been output or will be output.
  • the disclosed device/terminal and method may be implemented in other ways.
  • the device/terminal embodiments described above are only illustrative.
  • the division of the modules or units is only a logical function division, and there may be other divisions in actual implementation, such as multiple units or Components can be combined or integrated into another system, or some features can be omitted or not implemented.
  • the displayed or discussed mutual coupling or direct coupling or communication connection may be indirect coupling or communication connection through some interfaces, devices or units, and may be in electrical, mechanical or other forms.
  • the units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, they may be located in one place, or they may be distributed on multiple network units. Some or all of the units may be selected according to actual needs to achieve the objectives of the solutions of the embodiments.
  • the functional units in the various embodiments of the present application may be integrated into one processing unit, or each unit may exist alone physically, or two or more units may be integrated into one unit.
  • the above-mentioned integrated unit can be implemented in the form of hardware or software functional unit.
  • the integrated module/unit is implemented in the form of a software functional unit and sold or used as an independent product, it can be stored in a computer readable storage medium.
  • this application implements all or part of the processes in the above-mentioned embodiments and methods, and can also be completed by instructing relevant hardware through a computer program.
  • the computer program can be stored in a computer-readable storage medium. When the program is executed by the processor, it can implement the steps of the foregoing method embodiments.
  • the computer program includes computer program code, and the computer program code may be in the form of source code, object code, executable file, or some intermediate form.
  • the computer-readable medium may include: any entity or device capable of carrying the computer program code, recording medium, U disk, mobile hard disk, magnetic disk, optical disk, computer memory, read-only memory (ROM, Read-Only Memory) , Random Access Memory (RAM, Random Access Memory), electrical carrier signal, telecommunications signal, and software distribution media, etc.
  • ROM Read-Only Memory
  • RAM Random Access Memory
  • electrical carrier signal telecommunications signal
  • software distribution media etc.
  • the content contained in the computer-readable medium can be appropriately added or deleted according to the requirements of the legislation and patent practice in the jurisdiction.
  • the computer-readable medium Does not include electrical carrier signals and telecommunication signals.
  • This application implements all or part of the processes in the above-mentioned embodiment methods, and can also be implemented by a computer program product.
  • the computer program product runs on a terminal, the steps in each of the above-mentioned method embodiments can be realized when the terminal is executed. .

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

本申请适用于视频处理技术领域,提供一种帧内预测方法、装置、终端及存储介质,其中方法包括:确定图像块的邻域像素点;从所述邻域像素点中分别确定与所述图像块中的每个像素点对应的参考像素点,所述图像块中的每个像素点与其对应的参考像素点位于同一预设弧线上;对于所述图像块中的每个像素点,将与其对应的所述参考像素点的像素值确定为所述像素点的预测值。该方案使帧内预测的执行能够更贴近360度视频投影后在2D平面中的像素点分布情况,消除360度视频中的空间冗余,提升帧内预测的准确度,提升360度视频的最终编码性能。

Description

一种帧内预测方法、装置、终端及存储介质 技术领域
本申请属于视频处理技术领域,尤其涉及一种帧内预测方法、装置、终端及存储介质。
背景技术
在多媒体技术发展过程中,由于虚拟现实(Virtual Reality,VR)和增强现实(Augmented Reality,AR)技术能够向消费者提供身临其境的体验,因此它们已被逐渐应用于多个领域。
其中,360度视频是AR和VR的重要形式之一。它除了能够提供更广阔的视角之外,360度视频还提供了更高的分辨率和帧率,更好的视觉体验。
而视频通常需要通过编码来实现数据的存储或传输。现有技术中,为了使360度视频适应当前基于块的视频编码标准,在编码之前需要进行从3D球体到2D平面的投影变换。
现有的视频编码中的帧内预测算法是为传统的2D视频设计的,在基于投影变换得到的2D平面视频图像进行帧内预测时,这些算法没有考虑球形域特定的属性及特征,不能完全消除360度视频中的空间冗余,使得360度视频不能取得较好的编码性能。
发明内容
本申请实施例提供了一种帧内预测方法、装置、终端及存储介质,以解决现有视频编码中的帧内预测算法没有考虑球形域特定的属性及特征,不能完全消除360度视频中的空间冗余,使得360度视频不能取得较好的编码性能的问题。
本申请实施例的第一方面提供了一种帧内预测方法,包括:
确定图像块的邻域像素点;
从所述邻域像素点中分别确定与所述图像块中的每个像素点对应的参考像素点,所述图像块中的每个像素点与其对应的参考像素点位于同一预设弧线上;
对于所述图像块中的每个像素点,将与其对应的所述参考像素点的像素值确定为所述像素点的预测值。
本申请实施例的第二方面提供了一种帧内预测装置,包括:
第一确定模块,用于确定图像块的邻域像素点;
第二确定模块,用于从所述邻域像素点中分别确定与所述图像块中的每个像素点对应的参考像素点,所述图像块中的每个像素点与其对应的参考像素点位于同一预设弧线上;
第三确定模块,用于对于所述图像块中的每个像素点,将与其对应的所述参考像素点的像素值确定为所述像素点的预测值。
本申请实施例的第三方面提供了一种终端,包括存储器、处理器以及存储在所述存储器中并可在所述处理器上运行的计算机程序,所述处理器执行所述计算机程序时实现如第一方面所述方法的步骤。
本申请实施例的第四方面提供了一种计算机可读存储介质,所述计算机可读存储介质存储有计算机程序,所述计算机程序被处理器执行时实现如第一方面所述方法的步骤。
本申请的第五方面提供了一种计算机程序产品,当所述计算机程序产品在终端上运行时,使得所述终端执行上述第一方面所述方法的步骤。
由上可见,本申请实施例中,通过确定图像块的邻域像素点,从该邻域像素点中分别确定与所述图像块中的每个像素点对应的参考像素点,该图像块中的每个像素点与其对应的参考像素点位于同一预设弧线上;对于该图像块中的每个像素点,将与其对应的参考像素点的像素值确定为该像素点的预测值。该过程,能够满足在基于投影变换得到的2D平面视频图像进行帧内预测时,充 分考虑360度视频中球形域特定的属性及特征,使帧内预测的执行能够更贴近360度视频投影后在2D平面中的像素点分布情况,消除360度视频中的空间冗余,提升帧内预测的准确度,提升360度视频的最终编码性能。
附图说明
为了更清楚地说明本申请实施例中的技术方案,下面将对实施例或现有技术描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本申请的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其他的附图。
图1是本申请实施例提供的一种帧内预测方法的流程图一;
图2是本申请实施例提供的一种帧内预测方法的流程图二;
图3是本申请实施例提供的一种帧内预测场景示意图一;
图4是本申请实施例提供的一种帧内预测场景示意图二;
图5是本申请实施例提供的一种帧内预测装置的结构图;
图6是本申请实施例提供的一种终端的结构图。
具体实施方式
以下描述中,为了说明而不是为了限定,提出了诸如特定系统结构、技术之类的具体细节,以便透彻理解本申请实施例。然而,本领域的技术人员应当清楚,在没有这些具体细节的其它实施例中也可以实现本申请。在其它情况中,省略对众所周知的系统、装置、电路以及方法的详细说明,以免不必要的细节妨碍本申请的描述。
应当理解,当在本说明书和所附权利要求书中使用时,术语“包括”指示所描述特征、整体、步骤、操作、元素和/或组件的存在,但并不排除一个或多个其它特征、整体、步骤、操作、元素、组件和/或其集合的存在或添加。
还应当理解,在此本申请说明书中所使用的术语仅仅是出于描述特定实施 例的目的而并不意在限制本申请。如在本申请说明书和所附权利要求书中所使用的那样,除非上下文清楚地指明其它情况,否则单数形式的“一”、“一个”及“该”意在包括复数形式。
还应当进一步理解,在本申请说明书和所附权利要求书中使用的术语“和/或”是指相关联列出的项中的一个或多个的任何组合以及所有可能组合,并且包括这些组合。
如在本说明书和所附权利要求书中所使用的那样,术语“如果”可以依据上下文被解释为“当...时”或“一旦”或“响应于确定”或“响应于检测到”。类似地,短语“如果确定”或“如果检测到[所描述条件或事件]”可以依据上下文被解释为意指“一旦确定”或“响应于确定”或“一旦检测到[所描述条件或事件]”或“响应于检测到[所描述条件或事件]”。
应理解,本实施例中各步骤的序号的大小并不意味着执行顺序的先后,各过程的执行顺序应以其功能和内在逻辑确定,而不应对本申请实施例的实施过程构成任何限定。
本申请实施例中,将通过确定图像块的邻域像素点,从该邻域像素点中分别确定与所述图像块中的每个像素点对应的参考像素点,该图像块中的每个像素点与其对应的参考像素点位于同一预设弧线上;对于该图像块中的每个像素点,将与其对应的参考像素点的像素值确定为该像素点的预测值。该过程,能够满足在基于投影变换得到的2D平面视频图像进行帧内预测时,充分考虑360度视频中球形域特定的属性及特征,使帧内预测的执行能够更贴近360度视频投影后在2D平面中的像素点分布情况,消除360度视频中的空间冗余,提升帧内预测的准确度,提升360度视频的最终编码性能。
为了说明本申请所述的技术方案,下面通过具体实施例来进行说明。
参见图1,图1是本申请实施例提供的一种帧内预测方法的流程图一。如图1所示,一种帧内预测方法,该方法包括以下步骤:
步骤101,确定图像块的邻域像素点。
该图像块为图像帧中的一个图像区域。一个图像块中包括若干个像素点。
图像块的邻域像素点为在同一图像帧中与该图像块相邻区域中的像素点;更具体地,该邻域像素点为与该图像块相邻区域中与图像块的边缘像素点相邻的像素点。
具体地,将一个图像块作为一个像素点预测单元,在对一个图像块进行帧内预测时,需要分别对该图像块中所包含的各像素点进行像素值预测,直至每个像素点均完成帧内预测时,则该图像块完成帧内预测。
可选地,该图像块为一个矩形块。该邻域像素点具体可以是图像块的左侧邻域像素点或者上侧邻域像素点。
步骤102,从所述邻域像素点中分别确定与所述图像块中的每个像素点对应的参考像素点,所述图像块中的每个像素点与其对应的参考像素点位于同一预设弧线上。
这里,图像块中的每个像素点成为待预测像素点。在对图像块中的待预测像素点进行帧内预测时,需要从图像块的邻域像素点中选取对应的像素点作为参考像素点。具体是,将邻域像素点中与选取的待预测像素点处于同一弧形线上的点作为该待预测像素点所对应的参考像素点。以能够满足在基于投影变换得到的2D平面视频图像进行帧内预测时,充分考虑360度视频中球形域特定的属性及特征,使帧内预测的执行能够更贴近360度视频投影后在2D平面中的像素点分布情况,消除360度视频中的空间冗余,提升帧内预测的准确度,提升360度视频的最终编码性能。
步骤103,对于所述图像块中的每个像素点,将与其对应的所述参考像素点的像素值确定为所述像素点的预测值。
具体地,在实施过程中,在确定好与图像块中待预测像素点所对应的参考像素点后,将参考像素点的像素值作为预测值,直接赋值给待预测像素点的像素值即可。依照该操作处理,直至图像块中所有的像素点均帧内预测完毕。
本申请实施例中,通过确定图像块的邻域像素点,从该邻域像素点中分别 确定与所述图像块中的每个像素点对应的参考像素点,该图像块中的每个像素点与其对应的参考像素点位于同一预设弧线上;对于该图像块中的每个像素点,将与其对应的参考像素点的像素值确定为该像素点的预测值。该过程,能够满足在基于投影变换得到的2D平面视频图像进行帧内预测时,充分考虑360度视频中球形域特定的属性及特征,使帧内预测的执行能够更贴近360度视频投影后在2D平面中的像素点分布情况,消除360度视频中的空间冗余,提升帧内预测的准确度,提升360度视频的最终编码性能。
本申请实施例中还提供了帧内预测方法的不同实施方式。
参见图2,图2是本申请实施例提供的一种帧内预测方法的流程图二。如图2所示,一种帧内预测方法,该方法包括以下步骤:
步骤201,确定图像块的邻域像素点。
该步骤的实现过程与前述实施方式中步骤101的实现过程相同,此处不再赘述。
步骤202,确定所述邻域像素点中位于第一圆周上的像素点,为与所述图像块中位于所述第一圆周上的像素点对应的参考像素点。
这里,在确定图像块中像素点所对应的参考像素点时,具体将邻域像素点中与图像块中像素点位于同一圆周上的像素点确定为参考像素点。
该第一圆周即为前述的预设弧线。
其中,作为一可选的实施方式,该确定所述邻域像素点中位于第一圆周上的像素点,为与所述图像块中位于所述第一圆周上的像素点对应的参考像素点,包括:
从以所述图像块中的第一像素点为圆心,以指定长度为半径的第二圆周上选取第二像素点;
确定所述邻域像素点中,位于以所述第二像素点为圆心的第一圆周上的像素点,为与所述图像块中位于所述第一圆周上的像素点对应的参考像素点。
利用第一圆周的圆心与第二圆周的圆心(待预测的图像块中第一像素点) 之间的距离控制圆形预测的弧度,以适应投影失真的失真程度。
具体地,该第一像素点可以是图像块中任意一像素点,或者是该图像块中的中心位置处的像素点。该第一圆周的半径可以根据需要进行具体确定。
其中,一个参照像素点对应于至少一个待预测像素点。
具体地,结合图3所示,以像素块中O(0,0)为圆心,绘制一个圆,其半径为R。在以O(0,0)为圆心,半径为R的圆周(即第二圆周)上选择一个点M(x0,y0)。以该M(x0,y0)为圆心,绘制一个圆,像素块中位于该圆的圆周(即第一圆周)上的像素点作为待预测像素点(该待预测像素点为至少一个;在该图3中为像素块中位于第一圆周上的包括N(x,y)在内的多个像素点),像素块的邻域像素点中位于该圆周(即该第一圆周)上的像素点P则作为该待预测像素点的参考像素点。
结合图4所示,亦可以以该第二像素点为圆心按照不同半径长度画若干个同心圆,该若干个同心圆的圆周形成若干个第一圆周,该若干个第一圆周均与像素块相交,若干个第一圆周均与像素块之间具有像素点交集,以能够使得在第一圆周上同时存在参考像素点及与之对应的图像块中的待预测像素点。
具体地,在图4中,以像素块中O(0,0)为圆心,绘制一个圆,其半径为R。在以O(0,0)为圆心,半径为R的圆周(即第二圆周)上选择一个点M(x0,y0)。以该M(x0,y0)为圆心,在顺时针或逆时针方向上绘制多个同心圆,这些同心圆的圆周将与该像素块相交,最终该些同心圆的圆周将覆盖该像素块中的像素点。
其中,当顺时针绘制多个同心圆时,则从图像块的邻域像素点中的左侧邻域像素点中确定参照像素点。当逆时针绘制多个同心圆时,则从图像块的邻域像素点的上侧邻域像素点中确定参照像素点。亦可以相反。这里仅为示例性说明,在具体应用中可以根据实际需要进行具体设置,不以此为限。
进一步地,在从以O(0,0)为圆心,半径为R的圆周上选取点M时,不同的点M(x0,y0)将产生不同的预测结果。
对应地,在所述对于所述图像块中的每个像素点,将与其对应的所述参考像素点的像素值确定为所述像素点的预测值之后,还包括:
重新执行所述从所述邻域像素点中分别确定与所述图像块中的每个像素点对应的参考像素点的步骤;其中,所述第二像素点更新为从所述第二圆周上重新选取的新的像素点。
该过程中,可以选取不同的M点,分别对应产生不同的预测结果,从这些预测结果中选取一个较佳的预测结果作为最终结果。将参考像素点的值传递到图像块中的待预测像素点,直至图像块中像素点被填充完毕。
进一步地,作为一可选的实施方式,该帧内预测方法还包括:
基于投影失真度,计算所述指定长度R;
其中,R=ρ×(Rmax–Rmin)+Rmin;其中,ρ∈[0,1],ρ为所述投影失真度的权重,所述ρ的值的大小与所述投影失真度的大小负相关;R为所述指定长度,Rmax为所述指定长度的最大值,Rmin为所述指定长度的最小值。
在360度视频投影变换得到2D平面视频图像过程中,360度视频的高纬度区域的内容在很大程度上被过度采样而扭曲变形产生投影失真。因此,在变换得到的2D平面视频图像中,对于360度视频中的高纬度图像区域,上述指定长度R的值应较小,而对于360度视频中的低纬度图像区域,上述指定长度R的值应较大。
其中,ρ的值的大小与所述投影失真度的大小负相关,即当ρ的值越小表示投影失真越严重,ρ的值越大表示投影失真度越小。
具体地,权重ρ可计算为ρ(m,n)=cos(π(0.5–(n+0.5)/H)),其中,H是利用ERP格式投影后360度视频的高度,m和n是利用ERP格式投影后的图像块中目标点在图像帧中的坐标值。
进一步地,作为一可选的实施方式,该方法还包括:
根据所述图像块的宽度和高度,设置所述指定长度的最大值与所述指定长度的最小值。
具体地,该指定长度的最大值Rmax可以设置为5l,该指定长度的最小值Rmin可以设置为0.8l,l 2=w 2+h 2,其中w和h是图像块的宽度和高度。
步骤203,对于所述图像块中的每个像素点,将与其对应的所述参考像素点的像素值确定为所述像素点的预测值。
该步骤的实现过程与前述实施方式中步骤103的实现过程相同,此处不再赘述。
该过程中,应用了特定的圆形预测方式,该方式能适应从3D球体到2D平面的投影失真,实现编码增益。
进一步地,前述不同实施方式中所涉及的帧内预测方法可以应用至一种视频编码方法中,该视频编码方法应用于视频编码器,该方法包括:
确定视频中的当前图像块;计算采用至少两种帧内预测策略分别对所述当前图像块进行帧内预测的率失真代价值,该至少两种帧内预测策略包括如前述不同实施方式中所涉及的任一项帧内预测方法;基于所述率失真代价值,将率失真代价值最小的帧内预测策略作为目标帧内预测策略;计算所述目标预测策略对所述当前图像块进行帧内预测的目标预测值;根据所述目标预测值及所述当前图像块的真实像素值,计算预测残差;根据所述目标帧内预测策略和所述预测残差对所述视频进行编码,生成视频码流。
其中,终端设备将视频数据分成若干帧图像,每帧图像包括若干图像块。
具体地,至少两种帧内预测策略包括:传统的2D图像帧内预测方法(例如角度帧内预测(Angular Intra Prediction,AIP)方法)及本申请实施例中所涉及的任一项帧内预测方法。
其中,率失真代价值为表征预测残差及与预测残差对应的残差编码的比特数的变量,将率失真代价值最小的预测策略作为目标帧内预测策略,减少残差编码的比特率,提高编码性能。
其中,根据所述目标帧内预测策略和所述预测残差对所述视频进行编码,生成视频码流,包括:基于所述目标帧内预测策略,将所述当前图像块设置一 对应的策略标识;根据所述策略标识和所述预测残差对所述视频进行编码,生成视频码流。
实现针对不同的图像块,确定与之对应的目标帧内预测策略,并在编码时将与该目标帧内预测策略对应的策略标识一同编码进视频中,生成视频码流,以能够针对不同的图像块实现最佳化编码操作,减少视频失真度。
在具体应用时,上述帧内预测方法及视频编码方法不仅可以用于亮度分量,而且还可以用于色度分量。使用二进制标志分别激活亮度分量和色度分量的优化。在编码器上,对亮度和色度分量均执行不同的帧内预测策略。选择具有较小率失真代价的策略作为目标帧内预测策略。然后,根据所选目标帧内预测策略将二进制标志设置为0或1,执行编码。并将其编码发送给解码器。在解码器端,策略标志将在帧内预测之前被解码。根据策略标志,将执行相应的帧内预测。之后利用解码的残差和帧内预测结果,重建360度视频。
对应地,前述不同实施方式中所涉及的帧内预测方法可以应用至一种视频解码方法,应用于视频解码器,该方法包括:
获取视频码流;对所述视频码流进行解码,得到与视频帧中图像块对应的目标帧内预测策略和预测残差;所述目标帧内预测策略为从至少两种预测方法中确定得到;所述至少两种预测策略中包括如前述不同实施方式中所涉及的任一项帧内预测方法;根据所述目标预测策略和所述预测残差进行视频重建。
其中,对所述视频码流进行解码,得到与每一图像块对应的目标帧内预测策略,包括:对所述视频码流进行解码,得到与所述图像块对应的策略标识;根据所述策略标识,从所述至少两种预测策略中确定所述目标帧内预测策略。
本申请实施例中,通过确定图像块的邻域像素点,从该邻域像素点中分别确定与所述图像块中的每个像素点对应的参考像素点,该图像块中的每个像素点与其对应的参考像素点位于同一预设弧线上;对于该图像块中的每个像素点,将与其对应的参考像素点的像素值确定为该像素点的预测值。该过程,能够满足在基于投影变换得到的2D平面视频图像进行帧内预测时,充分考虑360度 视频中球形域特定的属性及特征,使帧内预测的执行能够更贴近360度视频投影后在2D平面中的像素点分布情况,消除360度视频中的空间冗余,提升帧内预测的准确度,提升360度视频的最终编码性能。
参见图5,图5是本申请实施例提供的一种帧内预测装置的结构图,为了便于说明,仅示出了与本申请实施例相关的部分。
所述帧内预测装置500包括:
第一确定模块501,用于确定图像块的邻域像素点;
第二确定模块502,用于从所述邻域像素点中分别确定与所述图像块中的每个像素点对应的参考像素点,所述图像块中的每个像素点与其对应的参考像素点位于同一预设弧线上;
第三确定模块503,用于对于所述图像块中的每个像素点,将与其对应的所述参考像素点的像素值确定为所述像素点的预测值。
其中,所述第二确定模块502,包括:
确定子模块,用于确定所述邻域像素点中位于第一圆周上的像素点,为与所述图像块中位于所述第一圆周上的像素点对应的参考像素点。
其中,所述确定子模块,具体用于:
从以所述图像块中的第一像素点为圆心,以指定长度为半径的第二圆周上选取第二像素点;
确定所述邻域像素点中,位于以所述第二像素点为圆心的第一圆周上的像素点,为与所述图像块中位于所述第一圆周上的像素点对应的参考像素点。
其中,该帧内预测装置500还包括:
计算模块,用于基于投影失真度,计算所述指定长度R;
其中,R=ρ×(Rmax–Rmin)+Rmin;其中,ρ∈[0,1],ρ为所述投影失真度的权重,所述ρ的值的大小与所述投影失真度的大小负相关;R为所述指定长度,Rmax为所述指定长度的最大值,Rmin为所述指定长度的最小值。
其中,该帧内预测装置500还包括:
设置模块,用于根据所述图像块的宽度和高度,设置所述指定长度的最大值与所述指定长度的最小值。
本申请实施例提供的帧内预测装置能够实现上述帧内预测装置方法的实施例的各个过程,且能达到相同的技术效果,为避免重复,这里不再赘述。
图6是本申请实施例提供的一种终端的结构图。如该图所示,该实施例的终端6包括:至少一个处理器60(图6中仅示出一个)、存储器61以及存储在所述存储器61中并可在所述至少一个处理器60上运行的计算机程序62,所述处理器60执行所述计算机程序62时实现上述任意各个方法实施例中的步骤。
所述终端6可以是桌上型计算机、笔记本、掌上电脑及云端服务器等计算设备。所述终端6可包括,但不仅限于,处理器60、存储器61。本领域技术人员可以理解,图6仅仅是终端6的示例,并不构成对终端6的限定,可以包括比图示更多或更少的部件,或者组合某些部件,或者不同的部件,例如所述终端还可以包括输入输出设备、网络接入设备、总线等。
所称处理器60可以是中央处理单元(Central Processing Unit,CPU),还可以是其他通用处理器、数字信号处理器(Digital Signal Processor,DSP)、专用集成电路(Application Specific Integrated Circuit,ASIC)、现成可编程门阵列(Field-Programmable Gate Array,FPGA)或者其他可编程逻辑器件、分立门或者晶体管逻辑器件、分立硬件组件等。通用处理器可以是微处理器或者该处理器也可以是任何常规的处理器等。
所述存储器61可以是所述终端6的内部存储单元,例如终端6的硬盘或内存。所述存储器61也可以是所述终端6的外部存储设备,例如所述终端6上配备的插接式硬盘,智能存储卡(Smart Media Card,SMC),安全数字(Secure Digital,SD)卡,闪存卡(Flash Card)等。进一步地,所述存储器61还可以既包括所述终端6的内部存储单元也包括外部存储设备。所述存储器61用于存储所述计算机程序以及所述终端所需的其他程序和数据。所述存储器61还可以用于暂时地存储已经输出或者将要输出的数据。
所属领域的技术人员可以清楚地了解到,为了描述的方便和简洁,仅以上述各功能单元、模块的划分进行举例说明,实际应用中,可以根据需要而将上述功能分配由不同的功能单元、模块完成,即将所述装置的内部结构划分成不同的功能单元或模块,以完成以上描述的全部或者部分功能。实施例中的各功能单元、模块可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中,上述集成的单元既可以采用硬件的形式实现,也可以采用软件功能单元的形式实现。另外,各功能单元、模块的具体名称也只是为了便于相互区分,并不用于限制本申请的保护范围。上述系统中单元、模块的具体工作过程,可以参考前述方法实施例中的对应过程,在此不再赘述。
在上述实施例中,对各个实施例的描述都各有侧重,某个实施例中没有详述或记载的部分,可以参见其它实施例的相关描述。
本领域普通技术人员可以意识到,结合本文中所公开的实施例描述的各示例的单元及算法步骤,能够以电子硬件、或者计算机软件和电子硬件的结合来实现。这些功能究竟以硬件还是软件方式来执行,取决于技术方案的特定应用和设计约束条件。专业技术人员可以对每个特定的应用来使用不同方法来实现所描述的功能,但是这种实现不应认为超出本申请的范围。
在本申请所提供的实施例中,应该理解到,所揭露的装置/终端和方法,可以通过其它的方式实现。例如,以上所描述的装置/终端实施例仅仅是示意性的,例如,所述模块或单元的划分,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分方式,例如多个单元或组件可以结合或者可以集成到另一个系统,或一些特征可以忽略,或不执行。另一点,所显示或讨论的相互之间的耦合或直接耦合或通讯连接可以是通过一些接口,装置或单元的间接耦合或通讯连接,可以是电性,机械或其它的形式。
所述作为分离部件说明的单元可以是或者也可以不是物理上分开的,作为单元显示的部件可以是或者也可以不是物理单元,即可以位于一个地方,或者 也可以分布到多个网络单元上。可以根据实际的需要选择其中的部分或者全部单元来实现本实施例方案的目的。
另外,在本申请各个实施例中的各功能单元可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中。上述集成的单元既可以采用硬件的形式实现,也可以采用软件功能单元的形式实现。
所述集成的模块/单元如果以软件功能单元的形式实现并作为独立的产品销售或使用时,可以存储在一个计算机可读取存储介质中。基于这样的理解,本申请实现上述实施例方法中的全部或部分流程,也可以通过计算机程序来指令相关的硬件来完成,所述的计算机程序可存储于一计算机可读存储介质中,该计算机程序在被处理器执行时,可实现上述各个方法实施例的步骤。其中,所述计算机程序包括计算机程序代码,所述计算机程序代码可以为源代码形式、对象代码形式、可执行文件或某些中间形式等。所述计算机可读介质可以包括:能够携带所述计算机程序代码的任何实体或装置、记录介质、U盘、移动硬盘、磁碟、光盘、计算机存储器、只读存储器(ROM,Read-Only Memory)、随机存取存储器(RAM,Random Access Memory)、电载波信号、电信信号以及软件分发介质等。需要说明的是,所述计算机可读介质包含的内容可以根据司法管辖区内立法和专利实践的要求进行适当的增减,例如在某些司法管辖区,根据立法和专利实践,计算机可读介质不包括电载波信号和电信信号。
本申请实现上述实施例方法中的全部或部分流程,也可以通过计算机程序产品来实现,当计算机程序产品在终端上运行时,使得所述终端执行时实现可实现上述各个方法实施例中的步骤。
以上所述实施例仅用以说明本申请的技术方案,而非对其限制;尽管参照前述实施例对本申请进行了详细的说明,本领域的普通技术人员应当理解:其依然可以对前述各实施例所记载的技术方案进行修改,或者对其中部分技术特征进行等同替换;而这些修改或者替换,并不使相应技术方案的本质脱离本申 请各实施例技术方案的精神和范围,均应包含在本申请的保护范围之内。

Claims (10)

  1. 一种帧内预测方法,其特征在于,包括:
    确定图像块的邻域像素点;
    从所述邻域像素点中分别确定与所述图像块中的每个像素点对应的参考像素点,所述图像块中的每个像素点与其对应的参考像素点位于同一预设弧线上;
    对于所述图像块中的每个像素点,将与其对应的所述参考像素点的像素值确定为所述像素点的预测值。
  2. 根据权利要求1所述的帧内预测方法,其特征在于,所述从所述邻域像素点中分别确定与所述图像块中的每个像素点对应的参考像素点,包括:
    确定所述邻域像素点中位于第一圆周上的像素点,为与所述图像块中位于所述第一圆周上的像素点对应的参考像素点。
  3. 根据权利要求2所述的帧内预测方法,其特征在于,所述确定所述邻域像素点中位于第一圆周上的像素点,为与所述图像块中位于所述第一圆周上的像素点对应的参考像素点,包括:
    从以所述图像块中的第一像素点为圆心,以指定长度为半径的第二圆周上选取第二像素点;
    确定所述邻域像素点中,位于以所述第二像素点为圆心的第一圆周上的像素点,为与所述图像块中位于所述第一圆周上的像素点对应的参考像素点。
  4. 根据权利要求3所述的帧内预测方法,其特征在于,还包括:
    基于投影失真度,计算所述指定长度R;
    其中,R=ρ×(Rmax–Rmin)+Rmin;其中,ρ∈[0,1],ρ为所述投影失真度的权重,所述ρ的值的大小与所述投影失真度的大小负相关;R为所述指定长度,Rmax为所述指定长度的最大值,Rmin为所述指定长度的最小值。
  5. 根据权利要求4所述的帧内预测方法,其特征在于,还包括:
    根据所述图像块的宽度和高度,设置所述指定长度的最大值与所述指定长度的最小值。
  6. 一种帧内预测装置,其特征在于,包括:
    第一确定模块,用于确定图像块的邻域像素点;
    第二确定模块,用于从所述邻域像素点中分别确定与所述图像块中的每个像素点对应的参考像素点,所述图像块中的每个像素点与其对应的参考像素点位于同一预设弧线上;
    第三确定模块,用于对于所述图像块中的每个像素点,将与其对应的所述参考像素点的像素值确定为所述像素点的预测值。
  7. 根据权利要求6所述的帧内预测装置,其特征在于,所述第二确定模块,包括:
    确定子模块,用于确定所述邻域像素点中位于第一圆周上的像素点,为与所述图像块中位于所述第一圆周上的像素点对应的参考像素点。
  8. 根据权利要求7所述的帧内预测装置,其特征在于,所述确定子模块,具体用于:
    从以所述图像块中的第一像素点为圆心,以指定长度为半径的第二圆周上选取第二像素点;
    确定所述邻域像素点中,位于以所述第二像素点为圆心的第一圆周上的像素点,为与所述图像块中位于所述第一圆周上的像素点对应的参考像素点。
  9. 一种终端,包括存储器、处理器以及存储在所述存储器中并可在所述处理器上运行的计算机程序,其特征在于,所述处理器执行所述计算机程序时实现如权利要求1至5中任一项所述方法的步骤。
  10. 一种计算机可读存储介质,所述计算机可读存储介质存储有计算机程序,其特征在于,所述计算机程序被处理器执行时实现如权利要求1至5中任一项所述方法的步骤。
PCT/CN2020/129553 2020-06-16 2020-11-17 一种帧内预测方法、装置、终端及存储介质 WO2021253721A1 (zh)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202010550240.4 2020-06-16
CN202010550240.4A CN111818333B (zh) 2020-06-16 2020-06-16 一种帧内预测方法、装置、终端及存储介质

Publications (1)

Publication Number Publication Date
WO2021253721A1 true WO2021253721A1 (zh) 2021-12-23

Family

ID=72846572

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/129553 WO2021253721A1 (zh) 2020-06-16 2020-11-17 一种帧内预测方法、装置、终端及存储介质

Country Status (2)

Country Link
CN (1) CN111818333B (zh)
WO (1) WO2021253721A1 (zh)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111818333B (zh) * 2020-06-16 2022-04-29 中国科学院深圳先进技术研究院 一种帧内预测方法、装置、终端及存储介质

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1162830A2 (en) * 2000-06-07 2001-12-12 Be Here Corporation Method and apparatus for electronically distributing motion panoramic images
CN101006730A (zh) * 2004-08-13 2007-07-25 庆熙大学校产学协力团 用于对环形图像中的基准像素进行插值以及对环形图像进行编码/解码的方法和设备
WO2018199792A1 (en) * 2017-04-26 2018-11-01 Huawei Technologies Co., Ltd Apparatuses and methods for encoding and decoding a panoramic video signal
WO2018199793A1 (en) * 2017-04-27 2018-11-01 Huawei Technologies Co., Ltd Geodesic intra-prediction for panoramic video coding
CN109804633A (zh) * 2016-09-30 2019-05-24 交互数字Vc控股公司 利用自适应帧内预测的全向视频编码和解码的方法和装置
CN111066322A (zh) * 2017-06-14 2020-04-24 华为技术有限公司 通过透视信息进行视频编码的帧内预测
CN111818333A (zh) * 2020-06-16 2020-10-23 中国科学院深圳先进技术研究院 一种帧内预测方法、装置、终端及存储介质

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7817717B2 (en) * 2002-06-18 2010-10-19 Qualcomm Incorporated Motion estimation techniques for video encoding
EP2129135B1 (en) * 2007-02-23 2012-08-15 Nippon Telegraph and Telephone Corporation Video coding intra prediction method and decoding method, their device, their program and program-recorded medium
CN102685504B (zh) * 2011-03-10 2015-08-19 华为技术有限公司 视频图像的编解码方法、编码装置、解码装置及其系统
CN103533374B (zh) * 2012-07-06 2018-02-16 乐金电子(中国)研究开发中心有限公司 一种视频编码、解码的方法及装置
JP6005572B2 (ja) * 2013-03-28 2016-10-12 Kddi株式会社 動画像符号化装置、動画像復号装置、動画像符号化方法、動画像復号方法、およびプログラム
EP3556096B1 (en) * 2016-12-22 2024-05-22 HFI Innovation Inc. Method and apparatus of motion refinement for video coding
CN110677644B (zh) * 2018-07-03 2021-11-16 北京大学 一种视频编码、解码方法及视频编码帧内预测器

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1162830A2 (en) * 2000-06-07 2001-12-12 Be Here Corporation Method and apparatus for electronically distributing motion panoramic images
CN101006730A (zh) * 2004-08-13 2007-07-25 庆熙大学校产学协力团 用于对环形图像中的基准像素进行插值以及对环形图像进行编码/解码的方法和设备
CN109804633A (zh) * 2016-09-30 2019-05-24 交互数字Vc控股公司 利用自适应帧内预测的全向视频编码和解码的方法和装置
WO2018199792A1 (en) * 2017-04-26 2018-11-01 Huawei Technologies Co., Ltd Apparatuses and methods for encoding and decoding a panoramic video signal
WO2018199793A1 (en) * 2017-04-27 2018-11-01 Huawei Technologies Co., Ltd Geodesic intra-prediction for panoramic video coding
CN111066322A (zh) * 2017-06-14 2020-04-24 华为技术有限公司 通过透视信息进行视频编码的帧内预测
CN111818333A (zh) * 2020-06-16 2020-10-23 中国科学院深圳先进技术研究院 一种帧内预测方法、装置、终端及存储介质

Also Published As

Publication number Publication date
CN111818333B (zh) 2022-04-29
CN111818333A (zh) 2020-10-23

Similar Documents

Publication Publication Date Title
US20190273929A1 (en) De-Blocking Filtering Method and Terminal
RU2716156C1 (ru) Способ и устройство для составления всенаправленного изображения в выходной формат макета
US10560678B2 (en) Method and apparatus having video encoding function with syntax element signaling of rotation information of content-oriented rotation applied to 360-degree image content or 360-degree video content represented in projection format and associated method and apparatus having video decoding function
WO2020001565A1 (zh) 点云编解码方法和编解码器
WO2021036795A1 (zh) 视频超分辨率处理方法及装置
CN110913218A (zh) 一种视频帧预测方法、装置及终端设备
CN111971969A (zh) 用于对点云的几何形状进行编码的方法和装置
WO2023029858A1 (zh) 点云媒体文件的封装与解封装方法、装置及存储介质
WO2024037137A1 (zh) 一种沉浸媒体的数据处理方法、装置、设备、介质和产品
WO2021253721A1 (zh) 一种帧内预测方法、装置、终端及存储介质
WO2022121650A1 (zh) 点云属性的预测方法、编码器、解码器及存储介质
US20200068205A1 (en) Geodesic intra-prediction for panoramic video coding
WO2019037558A1 (zh) 图像处理方法和装置
US11240512B2 (en) Intra-prediction for video coding using perspective information
CA3110477C (en) Picture partitioning method and apparatus
TWI696979B (zh) 圖像處理方法和裝置及全景視頻圖像的處理方法和系統
WO2021114100A1 (zh) 帧内预测方法、视频编码、解码方法及相关设备
CN104737539A (zh) 减小含有相互类似的信息的多个图像的数据量的方法、程序及装置
JP2010528516A (ja) 画像を処理する方法およびシステム
CN111246208A (zh) 视频处理方法、装置及电子设备
Cai et al. Towards 6DoF live video streaming system for immersive media
CN111885417B (zh) Vr视频播放方法、装置、设备以及存储介质
CN110830806A (zh) 一种视频帧预测方法、装置及终端设备
CN110115026B (zh) 在电子装置中在矩形投影上产生360度内容的方法和系统
WO2024051299A1 (zh) 一种编解码方法及装置

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20941219

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20941219

Country of ref document: EP

Kind code of ref document: A1

122 Ep: pct application non-entry in european phase

Ref document number: 20941219

Country of ref document: EP

Kind code of ref document: A1

122 Ep: pct application non-entry in european phase

Ref document number: 20941219

Country of ref document: EP

Kind code of ref document: A1

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 09/08/2023)