CN117596392A - Coding information determining method of coding block and related product - Google Patents

Coding information determining method of coding block and related product Download PDF

Info

Publication number
CN117596392A
CN117596392A CN202311277999.XA CN202311277999A CN117596392A CN 117596392 A CN117596392 A CN 117596392A CN 202311277999 A CN202311277999 A CN 202311277999A CN 117596392 A CN117596392 A CN 117596392A
Authority
CN
China
Prior art keywords
coding block
coding
video frame
frame
prediction mode
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202311277999.XA
Other languages
Chinese (zh)
Inventor
段晨辉
陈靖
高敏
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shuhang Technology Beijing Co ltd
Original Assignee
Shuhang Technology Beijing Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shuhang Technology Beijing Co ltd filed Critical Shuhang Technology Beijing Co ltd
Priority to CN202311277999.XA priority Critical patent/CN117596392A/en
Publication of CN117596392A publication Critical patent/CN117596392A/en
Pending legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/119Adaptive subdivision aspects, e.g. subdivision of a picture into rectangular or non-rectangular coding blocks
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/146Data rate or code amount at the encoder output
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/157Assigned coding mode, i.e. the coding mode being predefined or preselected to be further used for selection of another element or parameter
    • H04N19/159Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/172Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/176Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/58Motion compensation with long-term prediction, i.e. the reference frame for a current frame not being the temporally closest one

Abstract

The embodiment of the application discloses a coding information determining method of a coding block and a related product. The method comprises the following steps: dividing a j-th video frame into a plurality of coding blocks, wherein the j-th video frame is any video frame in a video to be coded; and when the jth video frame has a corresponding reference frame, determining the coding information of the nth coding block based on the coding information of the corresponding coding block of the nth coding block in the reference frame and the coding information of the first neighborhood coding block of the corresponding coding block, wherein the nth coding block is any one of the coding blocks.

Description

Coding information determining method of coding block and related product
Technical Field
The present disclosure relates to the field of video coding technologies, and in particular, to a method for determining coding information of a coding block and a related product.
Background
As network-based video applications become increasingly popular, modern devices and network infrastructure together drive the rapid growth of high resolution and low latency video content, such as video on demand, real-time streaming media, screen sharing, and the like. The application form of these video contents makes the demand for efficient video compression technology very strong. Currently, in the existing video compression technology, when determining the coding information of each coding block of each video frame in a video, all coding information is traversed to determine, and the method for determining the coding information reduces the video coding speed.
Disclosure of Invention
The embodiment of the application provides a coding information determining method of a coding block and a related product, wherein the coding information of a corresponding coding block of a reference frame of a current frame and the coding information of a neighborhood coding block of the corresponding coding block are utilized to skip the traversal of some coding information so as to improve the coding speed.
In a first aspect, an embodiment of the present application provides a method for determining coding information of a coding block, where the method includes:
dividing a j-th video frame into a plurality of coding blocks, wherein the j-th video frame is any video frame in a video to be coded; and when the jth video frame has a corresponding reference frame, determining the coding information of the nth coding block based on the coding information of the corresponding coding block of the nth coding block in the reference frame and the coding information of the first neighborhood coding block of the corresponding coding block, wherein the nth coding block is any one of the coding blocks.
In a second aspect, an embodiment of the present application provides an encoding information determining apparatus for an encoding block, including: a transceiver unit and a processing unit;
the receiving and transmitting unit is used for acquiring a video to be encoded;
the processing unit is configured to divide a jth video frame into a plurality of encoding blocks, where the jth video frame is any video frame in the video to be encoded; the processing unit is configured to determine, when the jth video frame has a corresponding reference frame, coding information of an nth coding block based on coding information of a corresponding coding block of the nth coding block in the reference frame and coding information of a first neighboring coding block of the corresponding coding block, where the nth coding block is any one of the plurality of coding blocks.
In a third aspect, an embodiment of the present application provides an electronic device, including: a processor and a memory, the processor being connected to the memory, the memory being for storing a computer program, the processor being for executing the computer program stored in the memory to cause the electronic device to perform the method as in the first aspect.
In a fourth aspect, embodiments of the present application provide a computer-readable storage medium storing a computer program, the computer program causing a computer to perform the method as in the first aspect.
In a fifth aspect, embodiments of the present application provide a computer program product comprising a non-transitory computer readable storage medium storing a computer program, the computer being operable to cause a computer to perform a method as in the first aspect.
The implementation of the embodiment of the application has the following beneficial effects:
it can be seen that in the embodiment of the present application, the jth video frame is divided into a plurality of encoding blocks, where the jth video frame is any video frame in the video to be encoded; the nth code block is any one of the plurality of code blocks. When determining the coding information of the nth coding block, not all the coding information is directly traversed to obtain the coding information of the nth coding block, but the coding information of the nth coding block is determined based on the correlation between adjacent video frames. Specifically, when the jth video frame has a corresponding reference frame, determining a prediction mode of the nth coding block based on a prediction mode of the corresponding coding block of the nth coding block in the reference frame and a prediction mode of a first neighboring coding block of the corresponding coding block, where it is noted that the nth coding block is any one of the plurality of coding blocks. For example, when the coding information of the first neighboring coding block and the coding information of the corresponding coding block are both inter-prediction modes, it is indicated that the j-th video frame has a relatively large correlation with the reference frame, so that when the prediction mode of the n-th coding block is traversed, the intra-prediction mode can be skipped to obtain the coding information of the n-th coding block, and the coding speed can be effectively improved without reducing the video quality.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present application, the drawings that are needed in the description of the embodiments will be briefly introduced below, and it is obvious that the drawings in the following description are some embodiments of the present application, and other drawings may be obtained according to these drawings without inventive effort for a person skilled in the art.
Fig. 1 is a schematic diagram of a coding information determining system of a coding block according to an embodiment of the present application;
fig. 2 is a flow chart of a method for determining coding information of a coding block according to an embodiment of the present application;
fig. 3 is a flowchart of a method for determining a prediction mode of a coding block according to an embodiment of the present application;
fig. 4 is a functional unit block diagram of an encoding information determining apparatus of an encoding block according to an embodiment of the present application;
fig. 5 is a schematic structural diagram of an electronic device according to an embodiment of the present application.
Detailed Description
The following description of the embodiments of the present application will be made clearly and fully with reference to the accompanying drawings, in which it is evident that the embodiments described are some, but not all, of the embodiments of the present application. All other embodiments, which can be made by one of ordinary skill in the art based on the embodiments herein without making any inventive effort, are intended to be within the scope of the present application.
The terms "first," "second," "third," and "fourth" and the like in the description and in the claims and drawings are used for distinguishing between different objects and not for describing a particular sequential order. Furthermore, the terms "comprise" and "have," as well as any variations thereof, are intended to cover a non-exclusive inclusion. For example, a process, method, system, article, or apparatus that comprises a list of steps or elements is not limited to only those listed steps or elements but may include other steps or elements not listed or inherent to such process, method, article, or apparatus.
Reference herein to "an embodiment" means that a particular feature, result, or characteristic described in connection with the embodiment may be included in at least one embodiment of the application. The appearances of such phrases in various places in the specification are not necessarily all referring to the same embodiment, nor are separate or alternative embodiments mutually exclusive of other embodiments. Those of skill in the art will explicitly and implicitly appreciate that the embodiments described herein may be combined with other embodiments.
In order to facilitate understanding of the technical solutions of the present application, first, related technical knowledge related to the present application is explained and explained.
Frame type: in the current video encoder, the pictures are organized in units of groups of pictures (Group of Pictures, GOP), one group of pictures is a data stream after a section of pictures is encoded, starting with an I frame and ending with the next I frame, and one group is a series of data streams after a section of pictures with high content similarity is encoded. A complete group of pictures includes three different frame types: i-frames, P-frames, and B-frames, can also be understood as I-type, P-type, and B-type. The I frame is the first frame of the picture group, only one I frame exists in one group, the I frame is an intra-frame coding frame, and the I frame is an independent frame with all information and can be independently coded without referring to other images. P frames are inter-coded frames that need to be referenced to either the previous I frame or P frame to be coded, and have no complete picture data and only picture difference data from the previous reference frame. P frames typically occupy fewer data bits than I frames. The B frame is a bi-directional predictive encoded frame, that is, the B frame records the difference between the current frame and the previous and subsequent reference frames, that is, the B frame is to be decoded, not only the previous buffered picture, but also the picture after decoding is to be obtained, and the final picture is obtained by overlapping the previous and subsequent pictures with the current frame data.
Prediction mode: the present application mainly describes a prediction mode in an open media alliance video One (AV 1) video codec mode as an example. Depending on the type of prediction mode, the prediction modes can be divided into two groups: intra prediction mode and inter prediction mode. The intra-frame prediction mode and the inter-frame prediction mode respectively correspond to a plurality of sub-intra-frame prediction modes and sub-inter-frame prediction modes. In the AV1 video codec mode, the intra prediction modes include: directional intra-prediction mode, non-directional intra-prediction mode, intra-filtering mode, and chroma intra-prediction mode; the inter prediction mode includes: motion vector prediction mode, spatial motion vector prediction mode, and temporal motion vector prediction mode. The prediction mode determination method under other video codecs is similar to the present application, and is not described here too much.
Intra prediction: intra prediction refers to predicting a current pixel by using the similarity or correlation between adjacent pixels in a video and using the adjacent pixels coded by the current image, thereby achieving the purpose of removing spatial redundancy, and the obtained prediction residual is further processed by subsequent modules such as transformation, quantization, entropy coding and the like to generate a final code stream.
Inter prediction: inter-frame prediction refers to a process of predicting a block to be encoded in a current image from an adjacent encoded image to obtain a reference block, and aims to remove temporal redundancy of a video signal, and the block to be encoded in the current frame is searched in the reference frame to obtain a best matching block according to a matching criterion of the encoding block.
Referring to fig. 1, fig. 1 is a schematic diagram of a coding information determining system of a coding block according to an embodiment of the present application. The prediction mode determination system of the encoded block includes a terminal device 101 and a server 102.
When the terminal apparatus 101 needs to upload a video (this video is referred to as a video to be encoded in this application) to the server 102, the terminal apparatus 101 needs to encode the video before uploading to the server 102. Specifically, the terminal device 101 firstly frames the video to be encoded to obtain a plurality of video frames. Then, the terminal device 101 divides a j-th video frame into a plurality of encoding blocks, wherein the j-th video frame is any one of the plurality of video frames of the video to be encoded; when the jth video frame has a corresponding reference frame, the terminal device 101 determines coding information of an nth coding block, which is any one of a plurality of coding blocks, based on coding information of a first neighboring coding block of a sum of corresponding coding blocks of the nth coding block in the reference frame. Further, the terminal device 101 may obtain the coding information of each coding block in the jth video frame based on the coding information of the nth coding block, and code the jth video frame based on the coding information of each coding block, thereby completing coding of a plurality of video frames and further implementing coding of the video to be coded. Finally, the terminal device 101 uploads the encoded video to be encoded to the server 102.
It should be noted that, if the server 102 needs to issue a video to the terminal device 101, it may also encode according to the above encoding policy to obtain an encoded video, and issue the encoded video to the terminal device 101, which will not be described in detail. Therefore, the method for determining the coding block in the present application may be performed on the terminal device side or may be performed on the server side, which is not limited in the present application.
It can be seen that, in the embodiment of the present application, the terminal device 101 determines the coding information of the nth coding block based on the coding information of the corresponding coding block of the nth coding block in the reference frame and the coding information of the first neighboring coding block of the corresponding coding block. The coding information of the block to be coded is determined by utilizing the time and space correlation between the nth coding block and the reference frame, specifically, the coding information of the nth coding block is determined by utilizing the time correlation between the nth coding block of the jth video frame and the corresponding coding block of the reference frame, the space correlation between the nth coding block of the jth video frame and the first neighborhood coding block, and the time and space correlation between the nth coding block and the second neighborhood coding block of the corresponding coding block, so that the coding speed can be effectively improved under the condition of ensuring the video quality.
Referring to fig. 2, fig. 2 is a flowchart of a method for determining coding information of a coding block according to an embodiment of the present application, where the method includes, but is not limited to, steps 201-203:
201: dividing the j-th video frame into a plurality of coding blocks, wherein the j-th video frame is any video frame in the video to be coded.
When the video to be encoded is encoded, firstly framing the video to be encoded to obtain a plurality of video frames, then dividing each video into encoding blocks to obtain a plurality of encoding blocks corresponding to each video frame, and then determining the prediction mode of each encoding block in each video frame to encode each video frame. It should be understood that a video frame is an image framed from a video, and the arrangement order of the video frames corresponds to the display time of the image one by one. Thus, the jth video frame may also be referred to as the video frame at the j instant.
And 202, when the jth video frame has a corresponding reference frame, determining the coding information of the nth coding block based on the coding information of the corresponding coding block of the nth coding block in the reference frame and the coding information of the first neighborhood coding block of the corresponding coding block, wherein the nth coding block is any one of a plurality of coding blocks.
It should be noted that the coding information of the coding block is the prediction mode of the coding block, and the position of the corresponding coding block in the reference frame is the same as the position of the nth coding block in the jth video frame.
Specifically, when traversing the prediction modes of the nth encoding block, some prediction modes are skipped using the temporal correlation and the spatial correlation between video frames to increase the encoding speed. When the prediction mode of the nth coding block is determined, the prediction mode of the second neighborhood coding block of the nth coding block, the prediction mode of the corresponding coding block of the nth coding block in the reference frame and the prediction mode of the first neighborhood coding block of the corresponding coding block are firstly obtained to obtain the prediction mode of the nth coding block. It should be noted that, the first neighboring coding block may be an upper neighboring coding block and a left neighboring coding block of the corresponding coding block; the second neighboring coding block may be an upper neighboring coding block and a left neighboring coding block of the nth coding block, which is not limited in this application.
For example, when the prediction mode of the first neighboring coding block, the prediction mode of the second neighboring coding block, and the prediction mode of the corresponding coding block are all inter-prediction modes, if the prediction mode of the nth coding block is traversed, the intra-prediction mode is skipped to obtain the prediction mode of the nth coding block. When the intra prediction mode is skipped, all the intra prediction modes of the intra prediction mode are skipped, and the prediction modes except the intra prediction mode are traversed to obtain the prediction mode of the nth encoded block.
For example, when the prediction mode of the first neighboring coding block, the prediction mode of the second neighboring coding block, and the prediction mode of the corresponding coding block are intra-frame prediction modes, if the prediction mode of the nth coding block is traversed, skipping part or all of the inter-subframe prediction modes to obtain the prediction mode of the nth coding block, wherein the inter-frame prediction modes include a plurality of inter-subframe prediction modes. When the inter prediction mode is skipped, not all the inter prediction modes are skipped directly, but the inter prediction mode corresponding to the rate distortion cost is determined by obtaining the rate distortion cost of the corresponding coding block, so as to determine whether to skip the inter prediction mode corresponding to the rate distortion cost when the prediction mode of the nth coding block is encoded.
Specifically, skipping part of the inter-subframe modes includes: and obtaining the rate distortion cost when the corresponding coding block is coded under each sub-inter prediction mode, and obtaining a plurality of rate distortion costs, wherein each rate distortion cost corresponds to each sub-inter prediction mode one by one. For example, when there is only one reference frame, that is, when the frame type of the jth video frame is P-type, only the rate-distortion cost of the corresponding coding speed in each sub-inter prediction mode needs to be obtained; when two reference frames exist, namely when the frame type of the jth video frame is B type, the rate distortion cost of the corresponding coding block of each reference frame in each sub-inter-frame prediction mode is required to be acquired respectively, the average value of the rate distortion costs of the two corresponding coding blocks is determined, and the average value is taken as the rate distortion cost of the corresponding coding speed. And when the h rate-distortion cost is greater than a preset threshold, skipping a sub-frame prediction mode corresponding to the h rate-distortion cost, wherein the h rate-distortion cost is any one of a plurality of rate-distortion costs. The optimal prediction mode is selected by comparing the rate-distortion costs when each prediction mode is used for encoding the encoded block, and selecting the prediction mode corresponding to the smallest rate-distortion cost as the optimal prediction mode.
It can be seen that, in the embodiment of the present application, when the prediction mode of the first neighboring coding block, the prediction mode of the second neighboring coding block, and the prediction mode of the corresponding coding block are intra-frame prediction modes, if the prediction mode of the nth coding block is traversed, the rate distortion cost when the corresponding coding block is encoded in each sub-inter-frame prediction mode is calculated, and the skipped part or all of the inter-subframe prediction modes are determined by comparing the rate distortion cost with the size of the preset threshold, so as to obtain the prediction mode of the nth coding block. The prediction mode of the nth coding block is traversed, possible inter-subframe prediction modes are not omitted, the accuracy of selection of the prediction mode is guaranteed, and the coding speed can be effectively improved.
In one embodiment of the present application, before determining the coding information of the nth coding block based on the coding information of the corresponding coding block of the nth coding block in the reference frame and the coding information of the first neighboring coding block of the corresponding coding block, the method further includes:
a reference frame corresponding to the jth video frame is determined based on the frame type of the jth video frame and the similarity of the jth video frame to neighboring frames. Specifically, in the embodiment of the present application, when selecting the adjacent frame of the jth video frame, instead of directly selecting the previous video frame or the next video frame of the jth video frame as the adjacent frame of the jth video frame, the selecting manner of the adjacent frame and the number of reference frames are determined based on the frame type of the jth video frame.
For example, when the frame type of the jth video frame is P-type, the jth video frame needs to refer to the video frame before it for encoding, and thus, the previous k video frames of the jth video frame are acquired as the adjacent frames of the jth video frame, so as to select the reference frame of the jth video frame. Specifically, the similarity between each video frame in the previous k video frames and the jth video frame is determined, a plurality of first similarities are obtained, and the video frame corresponding to the maximum value in the plurality of first similarities is used as a reference frame corresponding to the jth video frame.
For example, when the frame type of the jth video frame is B type, when encoding the jth video frame, not only the adjacent video frame before the jth video frame but also the adjacent video frame after the jth video frame are acquired to select at least two reference frames so as to encode each encoding block of the jth video frame. Specifically, the first m video frames and the last t video frames of the jth video frame are obtained as adjacent frames of the jth video frame; a reference frame corresponding to the jth video frame is determined based on the similarity of the first m video frames and the last t video frames to the jth video frame. Determining the similarity of each video frame in the first m video frames and the jth video frame to obtain a plurality of second similarities; determining the similarity of each video frame in the t video frames and the j video frame to obtain a plurality of third similarity; and taking the video frame corresponding to the maximum value in the second similarity and the video frame corresponding to the maximum value in the third similarity as reference frames corresponding to the j-th video frame.
It should be noted that, when the frame type of the jth video frame is I type, that is, the jth video frame is the first video frame, it is unnecessary to calculate the similarity between the adjacent frame and the jth video frame to determine the reference frame, because the I type is an intra-frame encoded frame, it is an independent frame with all information, and it is unnecessary to refer to other images to perform encoding independently.
It can be seen that, in the embodiment of the present application, by determining the frame type of the jth video frame of the current encoded frame to obtain the reference frame determining method under each frame type, different frame types correspond to different numbers of reference frames, and different reference frame selecting methods, for example, when the frame type of the jth video frame is B type, not only the reference frame needs to be selected from the adjacent frames before the jth video frame, but also the reference frame needs to be selected from the adjacent frames after the jth video frame, so that the time correlation between the jth video frame and the adjacent frames of the video frame to be encoded is fully utilized, so that the determination of the reference frame is more accurate, and the video encoding quality is ensured. Further, when determining the prediction mode of the nth coding block, determining the prediction mode of the block to be coded by using the time and space correlation between the nth coding block and the reference frame, specifically, using the time correlation between the nth coding block of the jth video frame and the corresponding coding block of the reference frame, the space correlation between the nth coding block of the jth video frame and the first neighborhood coding block, and the time and space correlation between the nth coding block and the second neighborhood coding block of the corresponding coding block, and determining the prediction mode of the nth coding block, thereby effectively improving the coding speed under the condition of ensuring the video quality.
In one embodiment of the present application, when the jth video frame does not have a corresponding reference frame, that is, when the frame type of the jth video frame is I-type, the prediction mode of the nth encoded block is determined to be an intra prediction mode. Because the frame type of the jth video frame is I type, the jth video frame does not have a reference frame, the jth video frame does not need to encode the encoding block of the jth video frame by utilizing the time correlation between the jth video frame and the adjacent frames, and only the correlation between adjacent pixels in the video frame, namely the spatial correlation, is utilized to encode the encoding block of the jth video frame, thereby achieving the purpose of removing the spatial redundancy.
Referring to fig. 3, fig. 3 is a flowchart of a method for determining a prediction mode of a coding block according to an embodiment of the present application. In this embodiment, the contents repeated with steps 201 to 202 are not repeated. The method includes, but is not limited to, the following steps:
301: the j-1 th video frame is encoded, and the j-1 th video frame is divided into a plurality of encoded blocks.
Specifically, the j-1 th video frame is a video frame at a time preceding the j-th video frame.
302: traversing each coding block of the j-1 video frame to obtain a prediction mode of each coding block.
It should be noted that, before encoding the jth video frame, not only each coding block of the jth-1 video frame is traversed to obtain the prediction mode of each coding block, but also the video frame before the jth video frame is traversed, where the video frame before the jth video frame includes the first I frame to the jth video frame before the jth video frame. To obtain a prediction mode corresponding to each coding block of each of the first I frame to the j video frame before the j video frame.
303: the prediction mode of each coding block of the j-1 video frame is stored into a global variable after loop filtering.
Specifically, the prediction mode of each of the encoded blocks of the j-1 th video frame is stored into a global variable so as to acquire the prediction mode of each of the encoded blocks of the j-1 th video frame. Further, the video coding loop filtering can reduce errors and distortions after coding and remove blocking effects by filtering pixel values in the coding process. In particular, loop filtering adjusts the predicted pixel values in the encoder to match the original pixel values. When the prediction mode is stored in the global variable, not only the prediction mode of each coding block of the j-1 th video frame, but also the prediction mode corresponding to each coding block of each of the first I frame to the j-th video frame before the j-th video frame are stored.
304: the j-th video frame is encoded, and the j-th video frame is divided into a plurality of encoding blocks.
305: each coded block in the jth video frame is traversed.
306: it is determined whether a reference frame exists for the jth video frame.
Specifically, a reference frame corresponding to the jth video frame is determined based on the frame type of the jth video frame and the similarity of the jth video frame to the neighboring frames. The implementation process of the above-mentioned embodiment 306 refers to the implementation process of the above-mentioned embodiment, and will not be described here too much.
307a: if yes, the prediction mode of each coding block of the reference frame in the global variable is obtained.
For example, when the reference frame is the j-1 th video frame, the prediction mode of each of the encoded blocks of the j-1 th video frame stored in the global variable is acquired. It should be noted that, when two reference frames exist, the prediction mode of each coding block of each reference frame needs to be acquired separately.
307b: if not, determining the prediction mode of the jth video frame as the intra-frame prediction mode.
It should be noted that, after determining that the prediction mode is the intra prediction mode, the method further needs to continue traversing a plurality of intra sub-frame prediction modes in the intra prediction mode to obtain the prediction mode of the jth video frame.
308: when an nth encoded block in a jth video frame is encoded, a prediction mode of the encoded block in the jth video frame is obtained, wherein the nth encoded block is any one of a plurality of encoded blocks of the jth video frame.
Specifically, when encoding to the nth encoding block in the jth video frame, the prediction modes of all encoding blocks encoded before the nth encoding block in the jth video frame are acquired. Wherein the nth coding block is any one of a plurality of coding blocks.
309: determining whether the prediction mode of the first neighborhood coding block of the nth coding block, the prediction mode of the corresponding coding block of the nth coding block in the reference frame and the prediction mode of the second neighborhood coding block of the corresponding coding block are all inter prediction modes.
It should be noted that, the position of the corresponding coding block in the reference frame is the same as the position of the nth coding block in the jth video frame, and the first neighboring coding block may be an upper neighboring coding block and a left neighboring coding block of the nth coding block; the second neighboring coding block may be an upper neighboring coding block and a left neighboring coding block of the corresponding coding block, which is not limited in this application.
310: if yes, skipping the intra-frame prediction mode when traversing the prediction mode of the nth coding block to obtain the prediction mode of the nth coding block.
For example, when the prediction mode of the first neighboring coding block, the prediction mode of the second neighboring coding block, and the prediction mode of the corresponding coding block are all inter-prediction modes, if the prediction mode of the nth coding block is traversed, the intra-prediction mode is skipped to obtain the prediction mode of the nth coding block. When the intra prediction mode is skipped, all the intra prediction modes of the intra prediction mode are skipped, and the prediction modes except the intra prediction mode are traversed to obtain the prediction mode of the nth encoded block.
Referring to fig. 4, fig. 4 is a functional unit block diagram of an encoding information determining apparatus for an encoding block according to an embodiment of the present application. The encoding information determining apparatus 400 of the encoding block includes: a transceiver unit 401 and a processing unit 402;
the transceiver unit 401 is configured to obtain a video to be encoded;
the processing unit 402 is configured to divide a jth video frame into a plurality of encoding blocks, where the jth video frame is any video frame in the video to be encoded; the processing unit 402 is configured to determine, when the jth video frame has a corresponding reference frame, coding information of an nth coding block based on coding information of a corresponding coding block of the jth video frame in the reference frame and coding information of a first neighboring coding block of the corresponding coding block, where the nth coding block is any one of the plurality of coding blocks.
In one embodiment of the present application, the processing unit 402 is specifically configured to, in determining the coding information of the nth coding block based on the coding information of the corresponding coding block of the nth coding block in the reference frame and the coding information of the first neighboring coding block of the corresponding coding block:
acquiring the coding information of a second neighborhood coding block of the nth coding block;
and determining the coding information of the nth coding block based on the coding information of the corresponding coding block, the coding information of the first neighborhood coding block and the coding information of the second neighborhood coding block.
In one embodiment of the present application, the coding information of the coding block is a prediction mode of the coding block; the processing unit 402 is specifically configured to determine the coding information aspect of the nth coding block based on the coding information of the corresponding coding block, the coding information of the first neighboring coding block, and the coding information of the second neighboring coding block:
when the prediction mode of the first neighborhood coding block, the prediction mode of the second neighborhood coding block and the prediction mode of the corresponding coding block are all inter-frame prediction modes, if the prediction mode of the nth coding block is traversed, skipping the intra-frame prediction mode to obtain the prediction mode of the nth coding block;
And when the prediction mode of the first neighborhood coding block, the prediction mode of the second neighborhood coding block and the prediction mode of the corresponding coding block are all the intra-frame prediction modes, if the prediction mode of the nth coding block is traversed, skipping part or all of the inter-subframe prediction modes to obtain the prediction mode of the nth coding block, wherein the inter-frame prediction modes comprise a plurality of inter-subframe prediction modes.
In one embodiment of the present application, the processing unit 402 is specifically configured to skip part or all of the inter prediction modes:
obtaining rate distortion costs when each sub-inter prediction mode is used for encoding the corresponding encoding block, and obtaining a plurality of rate distortion costs, wherein each rate distortion cost corresponds to each sub-inter prediction mode one by one;
and skipping a sub-frame prediction mode corresponding to the h rate-distortion cost when the h rate-distortion cost is greater than a preset threshold, wherein the h rate-distortion cost is any one of the plurality of rate-distortion costs.
In one embodiment of the present application, the position of the corresponding coding block in the reference frame is the same as the position of the nth coding block in the jth video frame.
In one embodiment of the present application, the first neighboring coding block includes an upper neighboring coding block and a left neighboring coding block of the corresponding coding block;
the second neighbor coding block includes an upper side neighboring coding block and a left side neighboring coding block of the nth coding block.
In one embodiment of the present application, before determining the coding information of the nth coding block based on the coding information of the first neighboring coding block of the nth coding block, the coding information of the corresponding coding block of the nth coding block in the reference frame, and the coding information of the second neighboring coding block of the corresponding coding block, the processing unit 402 is specifically configured to:
when the frame type of the jth video frame is P type, acquiring the first k video frames of the jth video frame as adjacent frames of the jth video frame;
determining the similarity of each video frame in the previous k video frames and the jth video frame to obtain a plurality of first similarities;
taking the video frame corresponding to the maximum value in the first similarity as a reference frame corresponding to the j-th video frame;
when the frame type of the jth video frame is B type, acquiring the first m video frames and the last t video frames of the jth video frame as adjacent frames of the jth video frame;
And determining a reference frame corresponding to the jth video frame based on the similarity of the first m video frames and the last t video frames with the jth video frame.
In one embodiment of the present application, the processing unit 402 is specifically configured to, in determining, based on the similarity between the first m video frames and the last t video frames and the jth video frame, a reference frame corresponding to the jth video frame:
determining the similarity of each video frame in the first m video frames and the jth video frame to obtain a plurality of second similarities;
determining the similarity of each video frame in the last t video frames and the jth video frame to obtain a plurality of third similarity;
and taking the video frame corresponding to the maximum value in the second similarity and the video frame corresponding to the maximum value in the third similarity as reference frames corresponding to the j-th video frame.
In one embodiment of the present application, when the jth video frame does not have a corresponding reference frame, the processing unit 402 is specifically configured to:
and determining the prediction mode of the nth coding block as the intra prediction mode.
Referring to fig. 5, fig. 5 is a schematic structural diagram of an electronic device according to an embodiment of the present application. As shown in fig. 5, the electronic device 500 includes a transceiver 501, a processor 502, and a memory 503. Which are connected by a bus 504. The memory 503 is used to store computer programs and data, and the data stored in the memory 503 may be transferred to the processor 502.
The processor 502 is configured to read a computer program in the memory 503 to perform the following operations:
dividing a j-th video frame into a plurality of coding blocks, wherein the j-th video frame is any video frame in a video to be coded;
and when the jth video frame has a corresponding reference frame, determining the coding information of the nth coding block based on the coding information of the corresponding coding block of the nth coding block in the reference frame and the coding information of the first neighborhood coding block of the corresponding coding block, wherein the nth coding block is any one of the coding blocks.
In one embodiment of the present application, the processing unit 402 is specifically configured to, in determining the coding information of the nth coding block based on the coding information of the corresponding coding block of the nth coding block in the reference frame and the coding information of the first neighboring coding block of the corresponding coding block:
acquiring the coding information of a second neighborhood coding block of the nth coding block;
and determining the coding information of the nth coding block based on the coding information of the corresponding coding block, the coding information of the first neighborhood coding block and the coding information of the second neighborhood coding block.
In one embodiment of the present application, the coding information of the coding block is a prediction mode of the coding block; the processor 502 is specifically configured to perform the following steps of:
when the prediction mode of the first neighborhood coding block, the prediction mode of the second neighborhood coding block and the prediction mode of the corresponding coding block are all inter-frame prediction modes, if the prediction mode of the nth coding block is traversed, skipping the intra-frame prediction mode to obtain the prediction mode of the nth coding block;
and when the prediction mode of the first neighborhood coding block, the prediction mode of the second neighborhood coding block and the prediction mode of the corresponding coding block are all the intra-frame prediction modes, if the prediction mode of the nth coding block is traversed, skipping part or all of the inter-subframe prediction modes to obtain the prediction mode of the nth coding block, wherein the inter-frame prediction modes comprise a plurality of inter-subframe prediction modes.
In one embodiment of the present application, the processor 502 is specifically configured to perform the following steps in terms of skipping part or all of the inter prediction modes:
Obtaining rate distortion costs when each sub-inter prediction mode is used for encoding the corresponding encoding block, and obtaining a plurality of rate distortion costs, wherein each rate distortion cost corresponds to each sub-inter prediction mode one by one;
and skipping a sub-frame prediction mode corresponding to the h rate-distortion cost when the h rate-distortion cost is greater than a preset threshold, wherein the h rate-distortion cost is any one of the plurality of rate-distortion costs.
In one embodiment of the present application, the position of the corresponding coding block in the reference frame is the same as the position of the nth coding block in the jth video frame.
In one embodiment of the present application, the first neighboring coding block includes an upper neighboring coding block and a left neighboring coding block of the corresponding coding block;
the second neighbor coding block includes an upper side neighboring coding block and a left side neighboring coding block of the nth coding block.
In one embodiment of the present application, before determining the coding information of the nth coding block based on the coding information of the first neighboring coding block of the nth coding block, the coding information of the corresponding coding block of the nth coding block in the reference frame, and the coding information of the second neighboring coding block of the corresponding coding block, the processor 502 is specifically configured to perform the following steps:
When the frame type of the jth video frame is P type, acquiring the first k video frames of the jth video frame as adjacent frames of the jth video frame;
determining the similarity of each video frame in the previous k video frames and the jth video frame to obtain a plurality of first similarities,
taking the video frame corresponding to the maximum value in the first similarity as a reference frame corresponding to the j-th video frame;
when the frame type of the jth video frame is B type, acquiring the first m video frames and the last t video frames of the jth video frame as adjacent frames of the jth video frame;
and determining a reference frame corresponding to the jth video frame based on the similarity of the first m video frames and the last t video frames with the jth video frame.
In one embodiment of the present application, the processor 502 is specifically configured to perform the following steps in determining a reference frame corresponding to the jth video frame based on the similarity between the first m video frames and the last t video frames and the jth video frame:
determining the similarity of each video frame in the first m video frames and the jth video frame to obtain a plurality of second similarities;
Determining the similarity of each video frame in the last t video frames and the jth video frame to obtain a plurality of third similarity;
and taking the video frame corresponding to the maximum value in the second similarity and the video frame corresponding to the maximum value in the third similarity as reference frames corresponding to the j-th video frame.
In one embodiment of the present application, when the jth video frame does not have a corresponding reference frame, the processing unit 402 is specifically configured to:
and determining the prediction mode of the nth coding block as the intra prediction mode.
Specifically, the transceiver 501 may be the transceiver unit 401 of the coding information determining apparatus 400 of one coding block in the embodiment of fig. 4, and the processor 502 may be the processing unit 402 of the coding information determining apparatus 400 of one coding block in the embodiment of fig. 4.
It should be understood that the electronic device in the present application may include a smart Phone (such as an Android mobile Phone, an iOS mobile Phone, a Windows Phone mobile Phone, etc.), a tablet computer, a palm computer, a notebook computer, a mobile internet device MID (Mobile Internet Devices, abbreviated as MID) or a wearable device, etc. The above-described electronic devices are merely examples and are not intended to be exhaustive and include, but are not limited to, the above-described electronic devices. In practical applications, the electronic device may further include: intelligent vehicle terminals, computer devices, etc.
The present application also provides a computer-readable storage medium storing a computer program that is executed by a processor to implement some or all of the steps of the coding information determining method of any one of the coding blocks described in the above method embodiments.
The present application also provides a computer program product comprising a non-transitory computer-readable storage medium storing a computer program operable to cause a computer to perform part or all of the steps of a method for determining coding information of any one of the coding blocks as described in the above method embodiments.
It should be noted that, for simplicity of description, the foregoing method embodiments are all expressed as a series of action combinations, but it should be understood by those skilled in the art that the present application is not limited by the order of actions described, as some steps may be performed in other order or simultaneously in accordance with the present application. Further, those skilled in the art will also appreciate that the embodiments described in the specification are all alternative embodiments, and that the acts and modules referred to are not necessarily required in the present application.
In the foregoing embodiments, the descriptions of the embodiments are emphasized, and for parts of one embodiment that are not described in detail, reference may be made to related descriptions of other embodiments.
In the several embodiments provided in this application, it should be understood that the disclosed apparatus may be implemented in other ways. For example, the apparatus embodiments described above are merely illustrative, such as a division of units, merely a division of logic functions, and there may be additional divisions in actual implementation, such as multiple units or components may be combined or integrated into another system, or some features may be omitted, or not performed. Alternatively, the coupling or direct coupling or communication connection shown or discussed with each other may be an indirect coupling or communication connection via some interfaces, devices or units, or may be in electrical or other forms.
The units described as separate units may or may not be physically separate, and units shown as units may or may not be physical units, may be located in one place, or may be distributed over a plurality of network units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution of this embodiment.
In addition, each functional unit in each embodiment of the present application may be integrated in one processing unit, or each unit may exist alone physically, or two or more units may be integrated in one unit. The integrated units described above may be implemented either in hardware or in software program modules.
The integrated units, if implemented in the form of software program modules and sold or used as a stand-alone product, may be stored in a computer readable memory. Based on such understanding, the technical solution of the present application may be embodied in essence or a part contributing to the prior art or all or part of the technical solution in the form of a software product stored in a memory, including several instructions for causing a computer device (which may be a personal computer, a server or a network device, etc.) to perform all or part of the steps of the methods of the embodiments of the present application. And the aforementioned memory includes: a U-disk, a Read-Only Memory (ROM), a random access Memory (RAM, random Access Memory), a removable hard disk, a magnetic disk, or an optical disk, or other various media capable of storing program codes.
Those of ordinary skill in the art will appreciate that all or a portion of the steps in the various methods of the above embodiments may be implemented by a program that instructs associated hardware, and the program may be stored in a computer readable memory, which may include: flash disk, read-Only Memory (ROM), random access Memory (Random Access Memory, RAM), magnetic disk or optical disk.
The foregoing has outlined rather broadly the more detailed description of embodiments of the present application, wherein specific examples are provided herein to illustrate the principles and embodiments of the present application, the above examples being provided solely to assist in the understanding of the methods of the present application and the core ideas thereof; meanwhile, as those skilled in the art will vary in the specific embodiments and application scope according to the ideas of the present application, the contents of the present specification should not be construed as limiting the present application in summary.

Claims (12)

1. A coding information determining method of a coding block, comprising:
dividing a j-th video frame into a plurality of coding blocks, wherein the j-th video frame is any video frame in a video to be coded;
and when the jth video frame has a corresponding reference frame, determining the coding information of the nth coding block based on the coding information of the corresponding coding block of the nth coding block in the reference frame and the coding information of the first neighborhood coding block of the corresponding coding block, wherein the nth coding block is any one of the coding blocks.
2. The method according to claim 1, wherein the determining the coding information of the nth coding block based on the coding information of the corresponding coding block of the nth coding block in the reference frame and the coding information of the first neighboring coding block of the corresponding coding block includes:
acquiring the coding information of a second neighborhood coding block of the nth coding block;
and determining the coding information of the nth coding block based on the coding information of the corresponding coding block, the coding information of the first neighborhood coding block and the coding information of the second neighborhood coding block.
3. The method according to claim 2, wherein the coding information of the coding block is a prediction mode of the coding block; the determining the coding information of the nth coding block based on the coding information of the corresponding coding block, the coding information of the first neighboring coding block and the coding information of the second neighboring coding block includes:
when the prediction mode of the first neighborhood coding block, the prediction mode of the second neighborhood coding block and the prediction mode of the corresponding coding block are all inter-frame prediction modes, if the prediction mode of the nth coding block is traversed, skipping the intra-frame prediction mode to obtain the prediction mode of the nth coding block;
And when the prediction mode of the first neighborhood coding block, the prediction mode of the second neighborhood coding block and the prediction mode of the corresponding coding block are all the intra-frame prediction modes, skipping part or all of the inter-subframe prediction modes when traversing the prediction mode of the nth coding block so as to obtain the prediction mode of the nth coding block, wherein the inter-frame prediction modes comprise a plurality of inter-subframe prediction modes.
4. The method of claim 3, wherein skipping some or all of the inter-subframe prediction modes comprises:
obtaining rate distortion costs when each sub-inter prediction mode is used for encoding the corresponding encoding block, and obtaining a plurality of rate distortion costs, wherein each rate distortion cost corresponds to each sub-inter prediction mode one by one;
and skipping a sub-frame prediction mode corresponding to the h rate-distortion cost when the h rate-distortion cost is greater than a preset threshold, wherein the h rate-distortion cost is any one of the plurality of rate-distortion costs.
5. The method according to any one of claim 1 to 4, wherein,
the position of the corresponding coding block in the reference frame is the same as the position of the nth coding block in the jth video frame.
6. The method according to any one of claims 1 to 5, wherein,
the first neighborhood coding block comprises an upper side adjacent coding block and a left side adjacent coding block of the corresponding coding block;
the second neighbor coding block includes an upper side neighboring coding block and a left side neighboring coding block of the nth coding block.
7. The method according to any one of claims 1-6, wherein before determining the coding information of the nth coding block based on the coding information of the corresponding coding block of the nth coding block in the reference frame and the coding information of the first neighboring coding block of the corresponding coding block, the method further comprises:
when the frame type of the jth video frame is P type, acquiring the first k video frames of the jth video frame as adjacent frames of the jth video frame;
determining the similarity of each video frame in the previous k video frames and the jth video frame to obtain a plurality of first similarities;
taking the video frame corresponding to the maximum value in the first similarity as a reference frame corresponding to the j-th video frame;
when the frame type of the jth video frame is B type, acquiring the first m video frames and the last t video frames of the jth video frame as adjacent frames of the jth video frame;
And determining a reference frame corresponding to the jth video frame based on the similarity of the first m video frames and the last t video frames with the jth video frame.
8. The method of claim 7, wherein the determining a reference frame corresponding to the jth video frame based on the similarity of the first m video frames and the last t video frames to the jth video frame comprises:
determining the similarity of each video frame in the first m video frames and the jth video frame to obtain a plurality of second similarities;
determining the similarity of each video frame in the last t video frames and the jth video frame to obtain a plurality of third similarity;
and taking the video frame corresponding to the maximum value in the second similarity and the video frame corresponding to the maximum value in the third similarity as reference frames corresponding to the j-th video frame.
9. The method of any of claims 1-8, wherein when the j-th video frame does not have a corresponding reference frame, the method further comprises:
and determining the prediction mode of the nth coding block as the intra prediction mode.
10. An encoding information determination apparatus for an encoding block, comprising: a transceiver unit and a processing unit;
The receiving and transmitting unit is used for acquiring a video to be encoded;
the processing unit is configured to divide a jth video frame into a plurality of encoding blocks, where the jth video frame is any video frame in the video to be encoded;
the processing unit is configured to determine, when the jth video frame has a corresponding reference frame, coding information of an nth coding block based on coding information of a corresponding coding block of the nth coding block in the reference frame and coding information of a first neighboring coding block of the corresponding coding block, where the nth coding block is any one of the plurality of coding blocks.
11. An electronic device, comprising: a processor and a memory, the processor being connected to the memory, the memory being for storing a computer program, the processor being for executing the computer program stored in the memory to cause the electronic device to perform the method of any one of claims 1-9.
12. A computer readable storage medium, characterized in that the computer readable storage medium stores a computer program, which is executed by a processor to implement the method of any one of claims 1-9.
CN202311277999.XA 2023-09-28 2023-09-28 Coding information determining method of coding block and related product Pending CN117596392A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202311277999.XA CN117596392A (en) 2023-09-28 2023-09-28 Coding information determining method of coding block and related product

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202311277999.XA CN117596392A (en) 2023-09-28 2023-09-28 Coding information determining method of coding block and related product

Publications (1)

Publication Number Publication Date
CN117596392A true CN117596392A (en) 2024-02-23

Family

ID=89912199

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202311277999.XA Pending CN117596392A (en) 2023-09-28 2023-09-28 Coding information determining method of coding block and related product

Country Status (1)

Country Link
CN (1) CN117596392A (en)

Similar Documents

Publication Publication Date Title
KR101859155B1 (en) Tuning video compression for high frame rate and variable frame rate capture
US10142654B2 (en) Method for encoding/decoding video by oblong intra prediction
CN109688407B (en) Reference block selection method and device for coding unit, electronic equipment and storage medium
WO2019114721A1 (en) Interframe prediction method and device for video data
CN113056910A (en) Motion vector predictor index coding for video coding
US20160050431A1 (en) Method and system for organizing pixel information in memory
US9565404B2 (en) Encoding techniques for banding reduction
AU2020261145B2 (en) Picture prediction method and apparatus, and computer-readable storage medium
CN113596475A (en) Image/video encoding method, apparatus, system, and computer-readable storage medium
AU2024201357A1 (en) Picture prediction method and apparatus, and computer-readable storage medium
US20050089232A1 (en) Method of video compression that accommodates scene changes
CN113382249B (en) Image/video encoding method, apparatus, system, and computer-readable storage medium
CN108401185B (en) Reference frame selection method, video transcoding method, electronic device and storage medium
WO2020253681A1 (en) Method and device for constructing merge candidate motion information list, and codec
US10142633B2 (en) Flexible coding unit ordering and block sizing
CN114257818A (en) Video encoding and decoding method, device, equipment and storage medium
WO2023011420A1 (en) Encoding method and apparatus, and decoding method and apparatus
WO2023020320A1 (en) Entropy encoding and decoding method and device
WO2023092256A1 (en) Video encoding method and related apparatus therefor
CN117596392A (en) Coding information determining method of coding block and related product
CN112055970B (en) Construction method of candidate motion information list, inter-frame prediction method and device
CN114793282A (en) Neural network based video compression with bit allocation
US20180109791A1 (en) A method and a module for self-adaptive motion estimation
CN117616751A (en) Video encoding and decoding of moving image group
RU2789454C2 (en) Video or image encoding based on in-block coding

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination