WO2019084792A1 - Procédé et dispositif de codage - Google Patents

Procédé et dispositif de codage Download PDF

Info

Publication number
WO2019084792A1
WO2019084792A1 PCT/CN2017/108651 CN2017108651W WO2019084792A1 WO 2019084792 A1 WO2019084792 A1 WO 2019084792A1 CN 2017108651 W CN2017108651 W CN 2017108651W WO 2019084792 A1 WO2019084792 A1 WO 2019084792A1
Authority
WO
WIPO (PCT)
Prior art keywords
image block
current image
coding
motion vector
intra
Prior art date
Application number
PCT/CN2017/108651
Other languages
English (en)
Chinese (zh)
Inventor
苏文艺
赵亮
朱磊
Original Assignee
深圳市大疆创新科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 深圳市大疆创新科技有限公司 filed Critical 深圳市大疆创新科技有限公司
Priority to CN201780008551.4A priority Critical patent/CN108702510A/zh
Priority to PCT/CN2017/108651 priority patent/WO2019084792A1/fr
Publication of WO2019084792A1 publication Critical patent/WO2019084792A1/fr

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/157Assigned coding mode, i.e. the coding mode being predefined or preselected to be further used for selection of another element or parameter
    • H04N19/159Prediction type, e.g. intra-frame, inter-frame or bidirectional frame prediction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/176Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding

Definitions

  • the present application relates to the field of video coding, and more particularly to an encoding method and apparatus.
  • An I frame sometimes referred to as an intra-coded frame, is a key frame in a video image.
  • An I frame is an independent frame with all the information, and can be independently coded without reference to other image frames.
  • a P frame (or a B frame) may also be referred to as an inter-coded frame, and an inter-frame coding technique is used, that is, information needs to be coded and referenced with reference to other image frames.
  • Inter-frame coding is equivalent to temporal low-pass filtering of the image, so that consecutive P-frames have similar similarity. Therefore, during video playback, if the played video content is a video content carried in a P frame, the human eye will feel that the playback process of the video content is smooth.
  • I frames are usually encoded periodically during video encoding.
  • the periodically occurring I frame blocks the smooth propagation of the video content, causing the video content viewed by the human eye to flicker.
  • the present application provides an encoding method and apparatus, which can reduce the degree of flicker of a video to some extent.
  • an encoding method including: acquiring a current image block to be encoded, the current image block being any one of an I frame or an I strip; inter-coding the current image block, Obtaining inter-frame coding information of the current image block, where the inter-frame coding information includes a reconstructed image block of the current image block; determining, according to the inter-frame coding information, whether the current image block generates flicker; The current image block is generated to generate flicker, and the reconstructed image block is intra-coded to obtain an intra-frame coding result of the current image block; according to the intra-frame coding result, the generation The code stream information corresponding to the current image block is described.
  • an encoder comprising: a memory for storing a program; a processor, configured to execute a program stored in the memory, when the program is executed, the processor is configured to execute as the first The encoding method described in the aspect.
  • a third aspect provides an image processing system, comprising: an image acquisition system, configured to acquire an image; and the encoder according to the second aspect, configured to encode an image collected by the image acquisition system to obtain a code stream. Information; an image transmission system for transmitting the code stream information.
  • a drone comprising the image processing system of the third aspect.
  • a computer storage medium comprising computer instructions that, when executed on a computer, cause the computer to perform the method of the first aspect.
  • a computer program product that, when executed on a computer, causes the computer to perform the method of the first aspect.
  • the technical solution provided by the present application changes the encoding mode of the image block in the traditional I frame or the I strip, first inter-codes the current image block in the I frame or the I strip, and determines the current image based on the inter-frame encoding information. Whether the intraframe coding of the block causes a flicker effect. If the intra-coding of the current image block causes a flickering effect, the reconstructed image block obtained by inter-coding is used instead of the current image block for intra-coding.
  • the interframe coding technique has the characteristics of time low-pass filtering.
  • the video content of the corresponding image block in the first few frames may be more similar, and intra-coding of the reconstructed image block may avoid a large jump of the encoded video content to a certain extent, thereby reducing the video content to a certain extent.
  • the degree of flicker during playback may be more similar, and intra-coding of the reconstructed image block may avoid a large jump of the encoded video content to a certain extent, thereby reducing the video content to a certain extent.
  • the degree of flicker during playback may be more similar, and intra-coding of the reconstructed image block may avoid a large jump of the encoded video content to a certain extent, thereby reducing the video content to a certain extent.
  • the degree of flicker during playback may be more similar, and intra-coding of the reconstructed image block may avoid a large jump of the encoded video content to a certain extent, thereby reducing the video content to a certain
  • FIG. 1 is an exemplary diagram of an arrangement of I frames.
  • FIG. 2 is an exemplary diagram of an arrangement of I strips.
  • FIG. 3 is a schematic structural diagram of an encoding system to which an embodiment of the present invention is applicable.
  • FIG. 4 is a schematic flowchart of an encoding method according to an embodiment of the present invention.
  • FIG. 5 is a schematic flow diagram of one implementation of step 430 of FIG.
  • FIG. 6 is a schematic flow diagram of another implementation of step 430 of FIG.
  • FIG. 7 is a diagram showing an example of surrounding image blocks of a current image block according to an embodiment of the present invention.
  • FIG. 8 is a schematic structural diagram of an encoder according to an embodiment of the present invention.
  • FIG. 9 is a schematic structural diagram of an image processing system according to an embodiment of the present invention.
  • FIG. 10 is a schematic structural diagram of a drone according to an embodiment of the present invention.
  • the technical solution provided by the present application can be applied to various systems related to video coding, such as a video codec system, a video storage system, a video transmission system, a drone image transmission system, and the like.
  • the technical solution provided by the present application can be applied to various video coding standards, such as H.263, H.264, H.265/AVC (advanced video coding), moving picture experts group (moving picture experts group, MPEG) and so on.
  • video coding standards such as H.263, H.264, H.265/AVC (advanced video coding), moving picture experts group (moving picture experts group, MPEG) and so on.
  • the video encoding process produces I and P frames (or B frames).
  • the appearance of an I frame blocks the smooth propagation of the video content, causing the video content viewed by the human eye to flicker (this phenomenon can be referred to as the flicker effect of the video).
  • flicker effect of the video this phenomenon can be referred to as the flicker effect of the video.
  • the causes of the flickering effect of the video will be described in detail below.
  • I is usually periodically encoded. frame.
  • GDR gradual refresh
  • the flickering effect of the video is generally generated.
  • the P frame is encoded by the inter-frame coding mode, and the inter-frame coding is equivalent to performing low-pass filtering on the consecutive P frames, which makes the P-frames of consecutive occurrences have higher similarity, and the human eye is watching P.
  • the video content corresponding to the frame the video content is perceived to be smooth.
  • this type of image frame blocks the smooth propagation of the video content, causing periodic flicker of the video content.
  • the human eye since the position of the I strip changes periodically, the human eye not only feels the periodic flicker of the video content, but also feels that the flickering portion of the video is moving, forming a roller effect.
  • Embodiments of the present invention provide an encoding method, which can reduce the degree of flicker of a video.
  • an encoding system 300 to which an embodiment of the present invention is applicable will be described in detail in conjunction with FIG. 3.
  • the authoring system 300 can include an initialization module 310, a motion estimation module 320, Inter-coding module 330, intra-coding module 340 and entropy encoding module 350.
  • the initialization module 310 can be used to initialize the video content to be encoded, such as block division of the image frame to be encoded, to form an image block to be encoded (sometimes referred to as a macro bloack (MB)).
  • MB macro bloack
  • the different video coding and decoding standards have their respective corresponding image block division manners, which are not specifically limited in the embodiment of the present invention.
  • the motion estimation module 320 can be used for motion estimation of the current image block to be encoded, such as calculating a motion vector (MV) of the current image block, a reference image block, and the like.
  • the information output by the motion estimation module 320 can be input to the inter-frame coding module 330 as inter-frame coding information.
  • the inter coding module 330 and the intra coding module 340 can be used for inter coding and intra coding of the current image block, respectively.
  • Encoding system 300 typically encodes the current image block by selecting one of the inter-coding modes and/or intra-coding modes of the current image block.
  • the coding mode may be, for example, an coding mode in which the rate distortion cost is the smallest in the inter coding mode or the intra coding mode, that is, the optimal coding mode. It should be noted that, for a P frame or a P strip (which may also be a B frame or a B strip), the encoding system 300 usually selects an inter coding mode from the inter coding mode to encode the current image block.
  • the encoding system 300 typically encodes the current image block from one of the intra-coded modes of the intra-coded mode.
  • all frames may be first inter-coded by the inter-frame coding module 330, and then whether intra-coding is required, and intra-frame coding is required, and whether to use the current image block for intra-frame coding is confirmed. Or reconstruct the picture block for intra-frame coding; it can also determine whether the frame requires intra-frame coding, and if it is determined that intra-frame coding is required, it is determined whether intra-frame coding is required, which is not limited herein.
  • the entropy encoding module 350 may perform an entropy encoding operation on the data encoded by the inter encoding module 330 or the intra encoding module 340 to obtain a code stream to be stored or to be transmitted.
  • FIG. 4 is a schematic flowchart of an encoding method according to an embodiment of the present invention.
  • the encoding method of Figure 4 can be performed, for example, by an encoder, such as a video encoder.
  • the encoding method can include steps 410-450. The steps in Figure 4 are described in detail below.
  • a current image block to be encoded is obtained.
  • the current image block may be any one of an I frame or an I slice.
  • step 420 the current image block is inter-coded to obtain inter-frame coding information of the current image block.
  • the inter coding mode may also be referred to as an inter mode or an inter mode.
  • the current image block may be inter-coded using one or more inter-coding modules.
  • One or more The inter-frame coding mode may be selected according to actual needs, or may be determined according to the codec standard actually used, which is not limited by the embodiment of the present invention. Taking H.264/AVC as an example, the at least one inter-coding mode may include, for example, one or more of the following inter-coding modes: inter-frame 16 ⁇ 16 mode, inter-frame 16 ⁇ 8 mode, inter-frame 8 ⁇ 16 mode, inter-frame 8 ⁇ 8 mode (inter-frame 8 ⁇ 8 mode can be further subdivided into inter-frame 4 ⁇ 4 mode) and the like.
  • step 420 can include inter-coding the current image block using an optimal inter coding mode to obtain inter-coded information of the current image block.
  • the interframe coding information obtained in step 420 can be understood as some kind of information related to interframe coding.
  • the information may be, for example, the information generated by the inter-frame coding process, or the information used in the inter-frame coding process, or the information included in the inter-frame coding result, which is not limited by the embodiment of the present invention.
  • the inter-coded information may include reconstructed image blocks of the current image block. Assuming that the current image block is encoded in an inter-coding mode, the reconstructed image block may be a reconstructed image block corresponding to the current image block in the inter-coding mode. Assuming that the current image block is encoded in a plurality of inter-coding modes, the reconstructed image block may be any one of the plurality of inter-coding modes (hereinafter referred to as the target inter-coding mode). Reconstructed image block under ).
  • the rate-distortion cost of the plurality of inter-coding modes is estimated, and the target inter-coding mode is selected from the plurality of inter-coding modes according to the rate-distortion cost of the plurality of inter-coding modes.
  • the target inter-coding mode may be an optimal inter-coding mode among the plurality of inter-coding modes described above.
  • the optimal inter coding mode referred to in this application may, for example, refer to an inter coding mode in which the rate distortion cost is the smallest among a plurality of inter coding modes.
  • the inter-frame coding information may further include at least one of the following: a reference image block corresponding to the current image block, a motion vector corresponding to the current image block, and the like.
  • the reference image block corresponding to the current image block may be an image block in a reference image frame corresponding to the current image frame where the current image block is located, and the position of the reference image frame and/or the reference image block at the reference image frame may be an inter-frame coding mode determine. Assuming that the current image block is encoded in an inter-coding mode, the reference image block corresponding to the current image block may be a reference image block corresponding to the current image block in the inter-coding mode.
  • the reference image block corresponding to the current image block may be a reference image block corresponding to the current image block in the inter-coding mode.
  • the reference image corresponding to the current image block The block may be a reference image block corresponding to the current image block in any one of the plurality of inter-coding modes.
  • the reference image block corresponding to the current image block may be a reference image block corresponding to the current image block in the optimal inter coding mode among the plurality of inter coding modes described above.
  • the motion vector corresponding to the image block may correspond to the image block in the inter-coding mode.
  • Sport vector Assuming that the image block is encoded in a plurality of inter-coding modes, the motion vector corresponding to the image block may be a motion vector corresponding to the image block in any one of the plurality of inter-coding modes. . As an example, the motion vector corresponding to the image block may be a motion vector corresponding to the optimal inter coding mode of the image block in the foregoing multiple inter coding modes.
  • step 430 it is determined whether the current image block will produce flicker based on the interframe coding information.
  • the current image block produces flickering, which means that when the video is played to the video content contained in the current image block, the continuity of the video playback is interrupted, and the human eye perceives that the video content has a jump.
  • step 430 there are a plurality of implementations of the step 430, which are not limited by the embodiment of the present invention.
  • whether the current image block will generate flicker may be determined according to the similarity between the reconstructed image block of the current image block and the current image block.
  • whether the current image block may generate flicker may be determined according to the similarity of the reconstructed image block of the current image block and the reference image block corresponding to the current image block.
  • whether the current image block may generate flicker may be determined according to the consistency of the motion vector corresponding to the current image block and the motion vector corresponding to the surrounding image block of the current image block.
  • two or more of the above examples may also be combined to determine if the current image block will produce a flicker.
  • the specific implementation of step 430 is detailed below, and will not be described in detail herein.
  • step 440 if the current image block is to be flickering, the reconstructed image block is intra-coded to obtain an intra-coded result of the current image block.
  • an encoding object (or a coded source image block) is a current image block.
  • the reconstructed image block may be used as the encoding target instead of the current image block for intra coding.
  • the reconstructed image block is an image block obtained based on inter-frame coding, and the inter-frame coding technique has the characteristics of time low-pass filtering. Therefore, compared with the current image block, the reconstructed image block and the corresponding image in the previous or previous frames are reconstructed.
  • the video content of the block may be closer, and intra-coding of the reconstructed image block will be To a certain extent, it avoids a large jump in the encoded video content, thereby reducing the degree of flicker of the video content during playback.
  • the quantization parameter (QP) corresponding to the intra-frame coding may be the same as or different from the quantization parameter corresponding to the inter-frame coding.
  • the quantization parameter corresponding to the intra-frame coding may be set to be smaller than the inter-frame coding corresponding. Quantitative parameters.
  • the encoded data may not be quantized, so that distortion of the video content generated by the quantization process may be avoided.
  • Intra_qp may represent a quantization parameter corresponding to intra coding
  • delta_qp may indicate a difference between a quantization parameter corresponding to the inter coding and a quantization parameter corresponding to the intra coding.
  • the value of delta_qp may be a positive integer, for example, may be a positive integer greater than 3.
  • the method of FIG. 4 may further include: if the current image block does not generate flicker, the current image block may be intra-coded to obtain an intra-coded result of the current image block.
  • step 450 code stream information corresponding to the current image block is generated according to the intra-frame coding result.
  • the intra-coded result may be subsequently processed based on a conventional manner to generate code stream information.
  • an operation such as entropy coding can be performed on the intraframe coding result.
  • data can be deblocked and filtered.
  • the embodiment of the present invention changes the encoding mode of the image block in the traditional I frame or the I strip, first performs interframe coding on the current image block in the I frame or the I strip, and determines the current image block based on the interframe encoding information. Whether intraframe coding will cause a flicker effect. If the intra-coding of the current image block causes a flickering effect, the reconstructed image block obtained by inter-coding is used instead of the current image block for intra-coding.
  • the interframe coding technique has the characteristics of time low-pass filtering. Therefore, compared with the current image block, the reconstructed image block may be more similar to the video content of the corresponding image block in the previous frame or the previous few frames.
  • the intra-coding of the image block will avoid the large jump of the encoded video content to a certain extent, thus reducing the degree of flicker of the video content during playback.
  • step 430 is illustrated by way of example with reference to specific embodiments.
  • the inter-frame coding information may further include a reference image block corresponding to the current image block.
  • step 430 can include steps 432-434.
  • step 432 the similarity of the reconstructed image block and the reference image block is determined.
  • step 434 it is determined whether the current image block will produce flicker based on the similarity between the reconstructed image block and the reference image block.
  • the degree of similarity between the reconstructed image block and the reference image block can be measured in a variety of ways. For example, the sum of absolute error (SAE) and/or sum of squares for error (SSE) of the reconstructed image block and the reference image block can be compared.
  • SAE sum of absolute error
  • SSE sum of squares for error
  • step 434 is related to the definition of the value of the similarity. It is assumed that the higher the similarity between the reconstructed image block and the reference image block, the larger the similarity of the reconstructed image block and the reference image block, the step 434 may include: if the similarity is less than or less than or equal to the first preset threshold, determining The current image block will produce a flicker. Further, in some embodiments, step 434 may further include determining that the current image block does not generate flicker if the similarity is greater than or equal to the first predetermined threshold.
  • the step 434 may include: if the similarity is greater than or equal to the first predetermined threshold, determining The current image block will produce a flicker. Further, in some embodiments, step 434 may further include determining that the current image block does not produce flicker if the similarity is less than or less than or equal to the first predetermined threshold.
  • inter-frame coding has the characteristics of temporal low-pass filtering, unless the video content itself is abrupt, the similarity between the reconstructed image block and the reference image block is generally higher (or less). Therefore, if the reconstructed image block and the reference image block have low similarity (or a large difference), it can be understood that the video content itself is abrupt, for example, the scene depicted by the video suddenly switches from one scene to another completely different scene, which A natural change in video content, not the flickering effect of the video due to encoding.
  • the embodiment of the present invention determines that the current image block does not generate flicker, and may still generate an intra-frame coding result of the current image block by directly performing intra-frame coding on the current image block.
  • the embodiment of the invention can distinguish the normal mutation of the video content from the flicker of the video content, and adopt different coding strategies, thereby improving the coding quality of the video.
  • the current image block may be directly used as the reconstructed image block corresponding to the current image block.
  • step 434 in the embodiment of FIG. 5 may be replaced with determining whether the current image block will produce a flicker based on the similarity of the current image block and the reference image block.
  • the interframe coding information may also include a motion vector corresponding to the current image block.
  • step 430 can include steps 436-438.
  • step 436 a difference between the motion vector corresponding to the current image block and the motion vector corresponding to the surrounding image block of the current image block is determined.
  • the surrounding image block of the current image block may refer to the image block that has been encoded around the current image block.
  • the surrounding image blocks of the current image block can be defined in various ways.
  • a surrounding image block of the current image block may include adjacent image blocks of the current image block.
  • a surrounding image block of a current image block can include adjacent image blocks of adjacent image blocks of the current image block.
  • the surrounding image block of the current image block may include one image block, and may also include two or more image blocks.
  • the surrounding image blocks of the image block X may include the image block A, the image block B, the image block C, and the image block D.
  • the surrounding image blocks of the image block X may be the image block A and the image block B.
  • step 438 it is determined whether the current image block will generate flicker according to the difference between the motion vector corresponding to the current image block and the motion vector corresponding to the surrounding image block.
  • the embodiment of the present invention determines that the current image block does not generate flicker, and still generates an intra-frame coding result of the current image block by directly performing intra-frame coding on the current image block.
  • the embodiment of the invention can distinguish the normal mutation of the video content from the flicker of the video content, and adopt different coding strategies, thereby improving the coding quality of the video.
  • the difference between the motion vector corresponding to the current image block and the motion vector corresponding to the surrounding image block may be defined in various ways.
  • the difference between the motion vector corresponding to the current image block and the motion vector corresponding to the surrounding image block may be defined as a motion vector corresponding to the current image block and each surrounding image block pair.
  • the minimum value of the difference in the motion vector may be defined as an average value of the difference between the motion vector corresponding to the current image block and the motion vector corresponding to each surrounding image block.
  • Step 438 may include determining that the current image block produces a flicker if the difference between the motion vector corresponding to the current image block and the motion vector corresponding to the surrounding image block is less than or less than a second predetermined threshold. Further, in some embodiments, step 438 may further include determining that the current image block does not generate flicker if the difference between the motion vector corresponding to the current image block and the motion vector corresponding to the surrounding image block is greater than a second predetermined threshold.
  • the current image block is the image block X shown in FIG. 7
  • the surrounding image block of the current image block is taken as an example of the image block A-D as shown in FIG. 7 .
  • the motion vector corresponding to the image block X is mv x .
  • the motion vectors corresponding to the image block AD are mv x , mv a , mv b , mv c and mv d , respectively .
  • the components of mv x along the x and y directions are denoted by mv xx and mv yy , respectively.
  • the components of mv a along the x and y directions are denoted by mv ax and mv ay , respectively.
  • the components of mv b along the x and y directions are denoted by mv bx and mv by , respectively.
  • the components of mv c along the x and y directions are denoted by mv cx and mv cy , respectively.
  • the components of mv d along the x and y directions are denoted by mv dx and mv dy , respectively.
  • the difference mv_diff between the motion vector corresponding to the current image block and the motion vector corresponding to the surrounding image block can be calculated by the following formula:
  • Mv_diff (mv_diff a +mv_diff b +mv_diff c +mv_diff d )/8, where:
  • Mv_diff a ABS(mv xx -mv ax )+ABS(mv xy -mv ay );
  • Mv_diff b ABS(mv xx -mv bx )+ABS(mv xy -mv by );
  • Mv_diff c ABS(mv xx -mv cx )+ABS(mv xy -mv cy );
  • Mv_diff d ABS(mv xx -mv dx )+ABS(mv xy -mv dy ).
  • mv_diff can be compared to a second predetermined threshold mv_thres. If mv_diff is greater than mv_thres, it can be considered that the motion vector corresponding to the current image block and the motion vector corresponding to the surrounding image block have a large difference (or poor consistency). If mv_diff is smaller than mv_thres, it can be considered that the difference between the motion vector corresponding to the current image block and the motion vector corresponding to the surrounding image block is small (or the consistency is good). Mv_thres can be determined empirically or experimentally, and mv_thres can be set to 6, for example, or set to 8.
  • the embodiment of FIG. 6 may be performed separately or in combination with the embodiment of FIG. 5. For example, after determining, by the embodiment of FIG. 5, that the similarity between the reconstructed image block and the reference image block is higher (greater than the first preset threshold), the embodiment of FIG. 6 may be further performed to determine when The difference between the motion vector corresponding to the previous image block and the motion vector corresponding to the surrounding image block of the current image block (eg, whether the difference is greater than a second predetermined threshold), and the result obtained in the embodiment of FIG. 6 is taken as the final result.
  • the embodiment of FIG. 5 may be further performed to determine the similarity between the reconstructed image block and the reference image block (eg, determining whether the similarity is Greater than the first predetermined threshold), and the results obtained in the embodiment of FIG. 5 are taken as the final result.
  • the encoder 800 can include a memory 810 and a processor 820.
  • Memory 810 can be used to store programs.
  • the processor 820 can be used to execute programs stored in the memory 810.
  • the processor 820 may be configured to acquire a current image block to be encoded, where the current image block is any one of an I frame or an I slice; and interframe coding the current image block, Obtaining inter-frame coding information of the current image block, where the inter-frame coding information includes a reconstructed image block of the current image block; determining, according to the inter-frame coding information, whether the current image block generates flicker; The current image block is generated to generate flicker, and the reconstructed image block is intra-coded to obtain an intra-frame coding result of the current image block; and the code stream information corresponding to the current image block is generated according to the intra-frame coding result.
  • the inter-frame coding information further includes a reference image block corresponding to the current image block
  • the processor 820 is configured to determine a similarity between the reconstructed image block and the reference image block. And determining, according to the similarity between the reconstructed image block and the reference image block, whether the current image block generates a flicker.
  • the processor 820 is configured to determine that the current image block generates a flicker if the similarity is less than or equal to a first preset threshold.
  • the processor 820 is configured to determine that the current image block does not generate flicker if the similarity is greater than or equal to a first preset threshold.
  • the reference image block is a reference image block corresponding to the current image block in an optimal inter coding mode.
  • the reference image block is an image block in a reference image frame corresponding to a current image frame where the current image block is located, and the reference image block is in the reference image frame.
  • the position is determined by the optimal inter coding mode.
  • the inter-frame coding information further includes the current image block pair a motion vector
  • the processor 820 is configured to determine a difference between a motion vector corresponding to the current image block and a motion vector corresponding to a surrounding image block of the current image block; and a motion vector sum corresponding to the current image block A difference in a motion vector corresponding to the surrounding image block determines whether the current image block generates a flicker.
  • the processor 820 is configured to determine, if a difference between a motion vector corresponding to the current image block and a motion vector corresponding to the surrounding image block is less than or equal to a second preset threshold. The current image block will produce a flicker.
  • the processor 820 is configured to determine, if a difference between a motion vector corresponding to the current image block and a motion vector corresponding to the surrounding image block is greater than the second preset threshold, The current image block does not produce flicker.
  • the surrounding image block is an adjacent image block of the current image block.
  • the motion vector corresponding to the current image block is a motion vector corresponding to an optimal inter coding mode of the current image block.
  • the motion vector corresponding to the surrounding image block is a motion vector corresponding to an optimal inter coding mode of the surrounding image block.
  • the processor 820 is configured to perform intra coding on the reconstructed image block according to a quantization parameter corresponding to intra coding, where the quantization parameter corresponding to the intra coding is smaller than The interframe encodes a corresponding quantization parameter.
  • the processor 820 is further configured to: according to the quantization parameter corresponding to the inter-frame coding, and the preset quantization parameter corresponding to the inter-frame coding, corresponding to the intra-frame coding The difference between the quantization parameters determines a quantization parameter corresponding to the intra coding mode.
  • the processor 820 is further configured to perform intra coding on the current image block if the current image block does not generate flicker, to obtain intra coding of the current image block. result.
  • the processor 820 is configured to entropy encode the intra-coded result to obtain code stream information corresponding to the current image block.
  • An embodiment of the present invention also provides an image processing system.
  • the image processing system 900 can include an image acquisition system 910, an encoder 800, and an image transmission system 920.
  • the encoder 800 can be used to encode the image acquired by the image acquisition system 910 to obtain code stream information.
  • Image transmission system 920 can be used to transmit code stream information.
  • the embodiment of the invention further provides a drone.
  • the drone 1000 can include an image processing system 900.
  • the drone 1000 can also include a housing, and the image processing system 900 can be located inside the housing.
  • the computer program product includes one or more computer instructions.
  • the computer can be a general purpose computer, a special purpose computer, a computer network, or other programmable device.
  • the computer instructions can be stored in a computer readable storage medium or transferred from one computer readable storage medium to another computer readable storage medium, for example, the computer instructions can be from a website site, computer, server or data center Transmission to another website site, computer, server or data center via wired (eg coaxial cable, fiber optic, digital subscriber line (DSL)) or wireless (eg infrared, wireless, microwave, etc.).
  • the computer readable storage medium can be any available media that can be accessed by a computer or a data storage device such as a server, data center, or the like that includes one or more available media.
  • the usable medium may be a magnetic medium (for example, a floppy disk, a hard disk, a magnetic tape), an optical medium (such as a digital video disc (DVD)), or a semiconductor medium (such as a solid state disk (SSD)).
  • a magnetic medium for example, a floppy disk, a hard disk, a magnetic tape
  • an optical medium such as a digital video disc (DVD)
  • a semiconductor medium such as a solid state disk (SSD)
  • the disclosed systems, devices, and methods may be implemented in other manners.
  • the device embodiments described above are merely illustrative.
  • the division of the unit is only a logical function division.
  • there may be another division manner for example, multiple units or components may be combined or Can be integrated into another system, or some features can be ignored or not executed.
  • Another point that is shown or discussed between each other The coupling or direct coupling or communication connection may be an indirect coupling or communication connection through some interface, device or unit, and may be in an electrical, mechanical or other form.
  • the units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, may be located in one place, or may be distributed to multiple network units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution of the embodiment.
  • each functional unit in each embodiment of the present application may be integrated into one processing unit, or each unit may exist physically separately, or two or more units may be integrated into one unit.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

L'invention concerne un procédé et un dispositif de codage. Le procédé de codage consiste à : acquérir un bloc d'image actuel devant être codé, le bloc d'image actuel étant un bloc d'image quelconque dans une trame I ou une bande I ; exécuter un codage inter-trame sur le bloc d'image actuel ; déterminer si le bloc d'image actuel clignote ou non d'après des informations de codage inter-trame ; dans l'affirmative, exécuter un codage intra-trame sur un bloc d'image reconstruit ; et générer des informations de flux de code correspondant au bloc d'image actuel d'après un résultat de codage intra-trame. La solution technique fournie par la présente invention permet de déterminer, d'après des informations de codage inter-trame, si le codage intra-trame sur le bloc d'image actuel peut provoquer un effet de clignotement, ou non ; et, si le codage intra-trame sur le bloc d'image actuel peut provoquer un effet de clignotement, un bloc d'image reconstruit obtenu par un codage inter-trame est utilisé pour remplacer le codage intra-trame sur le bloc d'image actuel, ce qui peut réduire dans une certaine mesure le degré de clignotement d'un contenu vidéo durant un processus de lecture.
PCT/CN2017/108651 2017-10-31 2017-10-31 Procédé et dispositif de codage WO2019084792A1 (fr)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN201780008551.4A CN108702510A (zh) 2017-10-31 2017-10-31 一种编码方法及装置
PCT/CN2017/108651 WO2019084792A1 (fr) 2017-10-31 2017-10-31 Procédé et dispositif de codage

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2017/108651 WO2019084792A1 (fr) 2017-10-31 2017-10-31 Procédé et dispositif de codage

Publications (1)

Publication Number Publication Date
WO2019084792A1 true WO2019084792A1 (fr) 2019-05-09

Family

ID=63844095

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2017/108651 WO2019084792A1 (fr) 2017-10-31 2017-10-31 Procédé et dispositif de codage

Country Status (2)

Country Link
CN (1) CN108702510A (fr)
WO (1) WO2019084792A1 (fr)

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2020107268A1 (fr) * 2018-11-28 2020-06-04 深圳市大疆创新科技有限公司 Procédé de codage de flux de code gdr, dispositif terminal, et support de stockage lisible par machine
CN111247800B (zh) * 2019-01-25 2022-07-19 深圳市大疆创新科技有限公司 确定图像信息量的方法与装置
WO2020258256A1 (fr) * 2019-06-28 2020-12-30 深圳市大疆创新科技有限公司 Procédé et appareil de traitement d'image, et véhicule aérien sans pilote
CN110351560A (zh) * 2019-07-17 2019-10-18 深圳市网心科技有限公司 一种编码方法、系统及电子设备和存储介质
CN110996102B (zh) * 2020-03-03 2020-05-22 眸芯科技(上海)有限公司 抑制p/b帧中帧内块呼吸效应的视频编码方法及装置

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101091393A (zh) * 2004-12-28 2007-12-19 日本电气株式会社 运动图像编码方法以及利用其的装置、计算机程序
US20100315557A1 (en) * 2009-06-11 2010-12-16 Texas Instruments Incorporated Reducing flicker in the display of video streams
CN102833532A (zh) * 2011-06-16 2012-12-19 安讯士有限公司 用于对数字视频数据进行编码的方法和数字视频编码器系统
CN103179394A (zh) * 2013-01-21 2013-06-26 北京航空航天大学 一种基于区域视频质量平稳的i帧码率控制方法

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4449915B2 (ja) * 2006-02-08 2010-04-14 ソニー株式会社 符号化装置、符号化方法およびプログラム、並びに、記録媒体

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101091393A (zh) * 2004-12-28 2007-12-19 日本电气株式会社 运动图像编码方法以及利用其的装置、计算机程序
US20100315557A1 (en) * 2009-06-11 2010-12-16 Texas Instruments Incorporated Reducing flicker in the display of video streams
CN102833532A (zh) * 2011-06-16 2012-12-19 安讯士有限公司 用于对数字视频数据进行编码的方法和数字视频编码器系统
CN103179394A (zh) * 2013-01-21 2013-06-26 北京航空航天大学 一种基于区域视频质量平稳的i帧码率控制方法

Also Published As

Publication number Publication date
CN108702510A (zh) 2018-10-23

Similar Documents

Publication Publication Date Title
WO2019084792A1 (fr) Procédé et dispositif de codage
CN105409216B (zh) 丢失视频数据的条件隐藏
US9414086B2 (en) Partial frame utilization in video codecs
US20150312575A1 (en) Advanced video coding method, system, apparatus, and storage medium
JP5555741B2 (ja) デジタル映像データを符号化するための方法およびデジタル映像符号化システム
US11743475B2 (en) Advanced video coding method, system, apparatus, and storage medium
WO2010135609A1 (fr) Décision de type à image adaptative pour codage vidéo
JP2006519565A (ja) ビデオ符号化
JP2013524669A (ja) 高効率な動画像符号化のためのスーパーブロック
US8781004B1 (en) System and method for encoding video using variable loop filter
US20050129125A1 (en) Method and apparatus for pitcure compression using variable block of arbitrary size
JP2015023579A (ja) ストリーミング配信システム
US9565404B2 (en) Encoding techniques for banding reduction
US9432694B2 (en) Signal shaping techniques for video data that is susceptible to banding artifacts
CN113259671B (zh) 视频编解码中的环路滤波方法、装置、设备及存储介质
CN115118976A (zh) 一种图像编码方法、可读介质及其电子设备
US20150189306A1 (en) Video encoder with block merging and methods for use therewith
WO2014160569A1 (fr) Appareils et procédés pour rafraîchissement intra de trame échelonnée
WO2017015958A1 (fr) Procédé et dispositif de codage et de décodage vidéo
JP2012015603A (ja) 画像処理装置及び画像映像処理方法
CN114222127A (zh) 一种视频编码方法、视频解码方法及装置
CN117616751A (zh) 动态图像组的视频编解码
US20160360219A1 (en) Preventing i-frame popping in video encoding and decoding
US9542611B1 (en) Logo detection for macroblock-based video processing
JP4936557B2 (ja) 符号化装置

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17931071

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 17931071

Country of ref document: EP

Kind code of ref document: A1