WO2019135447A1 - Procédé et dipositif de codage vidéo et procédé et dispositif de décodage vidéo faisant appel à une technique basée sur une prédiction de mouvement - Google Patents

Procédé et dipositif de codage vidéo et procédé et dispositif de décodage vidéo faisant appel à une technique basée sur une prédiction de mouvement Download PDF

Info

Publication number
WO2019135447A1
WO2019135447A1 PCT/KR2018/003803 KR2018003803W WO2019135447A1 WO 2019135447 A1 WO2019135447 A1 WO 2019135447A1 KR 2018003803 W KR2018003803 W KR 2018003803W WO 2019135447 A1 WO2019135447 A1 WO 2019135447A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
encoding unit
outline
block
samples
Prior art date
Application number
PCT/KR2018/003803
Other languages
English (en)
Korean (ko)
Inventor
박민수
최웅일
김찬열
Original Assignee
삼성전자 주식회사
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 삼성전자 주식회사 filed Critical 삼성전자 주식회사
Priority to KR1020207017845A priority Critical patent/KR20200096551A/ko
Priority to PCT/KR2018/007812 priority patent/WO2019135457A1/fr
Publication of WO2019135447A1 publication Critical patent/WO2019135447A1/fr

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/563Motion estimation with padding, i.e. with filling of non-object values in an arbitrarily shaped picture block or region for estimation purposes
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • H04N19/105Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/176Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a block, e.g. a macroblock
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/577Motion compensation with bidirectional frame interpolation, i.e. using B-pictures

Definitions

  • the image can be encoded or decoded using peripheral information of the reference image.
  • an inter prediction method for predicting a current image using a previous image or a next image is used during a video encoding / decoding process on the assumption that the content of the current image is similar to the previous or next image of the current image.
  • Inter prediction can also be performed for each block.
  • the distance between the current block and the closest block among the blocks of other images can be expressed mainly as a motion vector. It is possible to predict the sample values of the current block by using the motion vector determined by the inter prediction.
  • sample values of a current block can be predicted using a sample value of a block (reference block) that is separated from a current block by a motion vector.
  • a video encoding method comprising: obtaining prediction information of blocks adjacent to an inside of an outline of a reference image when a current block refers to an out-of-bounds region of the reference image in motion reference; Determining reference samples for predicting samples located outside the outline of the reference image using prediction information of a block adjacent to the outline; And determining the samples located outside the outline of the reference image using the reference samples.
  • the reference samples may be located in an image.
  • the prediction information is inter-prediction information of a block adjacent to the outline, and the inter-prediction information may include at least one of a reference picture list, a reference index, and a motion vector.
  • the reference picture list of the inter prediction information indicates the L0 reference list and the L1 reference list
  • the motion vector of the L0 reference list and the areas of the areas indicated by the motion vectors of the L1 reference list the predictive information corresponding to the motion vector of the L0 reference list and the predictive information corresponding to the motion vector of the L1 reference list are used for a region of a smaller size among the neighboring regions, the reference samples can be determined in a linear manner.
  • the motion vector of the L0 reference list and the motion vector of the L1 reference list are shifted in the outward direction of the area indicated by the first motion vector
  • the reference samples may be determined using the prediction information excluding the prediction information corresponding to the first motion vector when the neighboring region is outside the image.
  • the reference picture list of the inter prediction information indicates the L0 reference list and the L1 reference list
  • the motion vector of the L0 reference list and the motion vector of the L1 reference list are shifted in the outward direction of the area indicated by the first motion vector
  • the reference samples can be generated using the samples adjacent to the inside of the outline of the reference image.
  • the prediction information may be intra mode information of a block adjacent to the inside of the outline, and the reference samples may be determined from samples in the reference image according to the intra mode information.
  • the prediction information is information on a characteristic of a block adjacent to the outline, and the reference samples may be determined according to information on the characteristic.
  • the prediction information may include affine mode information, and the reference samples may be determined according to the affine mode information.
  • the characteristics of the block adjacent to the inside of the outline may be determined using one of a gradient and an edge detection.
  • the prediction information may include a characteristic of a block adjacent to the inside of the outline Information.
  • a block related to a block adjacent to the contour line may be determined by a template matching method, and the prediction information may include prediction information of the related block.
  • the reference image is one of a plurality of images combined according to a 360-degree image format, if prediction information of a block adjacent to an inside of an outline of the reference image exists, It is possible to generate sample values of an area outside the outline.
  • the 360 degree image may be generated by one of Rotated Sphere Projection (RSP), Equirectangular Projection (ERP), Icosahedral Projection (ISP), Segmented Sphere Projection (SSP), CubeMap Projection (CMP), Equal Area Cube Map .
  • RSP Rotated Sphere Projection
  • ERP Equirectangular Projection
  • ISP Icosahedral Projection
  • SSP Segmented Sphere Projection
  • CMP CubeMap Projection
  • a video decoding method comprising: obtaining prediction information of blocks adjacent to an inside of an outline of a reference image when a current block refers to an out-of-bounds region of the reference image in motion reference; Determining reference samples for predicting samples located outside the outline of the reference image using prediction information of a block adjacent to the outline; And determining the samples located outside the outline of the reference image using the reference samples.
  • a video encoding apparatus including: a prediction information obtaining unit that obtains prediction information of blocks adjacent to an inside of an outline of a reference image, when a current block refers to an out-of- A reference sample determining unit for determining reference samples for predicting samples located outside the outline of the reference image using prediction information of a block adjacent to the inside of the outline; And an outer sample determination unit for determining a sample located outside the outline of the reference image using the reference samples.
  • a video decoding apparatus comprising: a prediction information obtaining unit that obtains prediction information of blocks adjacent to an inside of an outline of a reference image, when a current block refers to an out-of- A reference sample determining unit for determining reference samples for predicting samples located outside the outline of the reference image using prediction information of a block adjacent to the inside of the outline; And an outer sample determination unit for determining a sample located outside the outline of the reference image using the reference samples.
  • FIG. 1A shows a flowchart of a padding method of an image outside region using prediction information according to an exemplary embodiment.
  • FIG. 1B is a flowchart illustrating a method of predicting a motion of an outer region of a reference image using prediction information according to another embodiment.
  • FIG. 2A illustrates a block diagram of an image padding apparatus using prediction information according to an embodiment.
  • FIG. 2B is a block diagram of a motion prediction apparatus to which a padding method of a reference image outer region using prediction information is applied according to another embodiment.
  • FIG. 3 illustrates a method of padding an outer region of a reference image according to a conventional padding method.
  • 4A illustrates a method of padding an outer region of a reference image using inter information according to an embodiment.
  • 4B illustrates a method of padding an outer region of a reference image using intra information according to an exemplary embodiment of the present invention.
  • FIG. 5 shows an example of a reference image to which a conventional padding method is applied.
  • FIG. 6 illustrates an example of a reference image to which the padding method according to an embodiment is applied.
  • FIG. 7 shows an example of a 360-degree image to which the conventional padding method is applied.
  • FIG. 8 illustrates an example of a reference image of a 360-degree image to which a padding method of a reference image outside region according to an embodiment is applied.
  • FIG. 9 illustrates an example of a reference image of a 360-degree image to which a padding method of a reference image outside area is applied according to another embodiment.
  • 10A is a block diagram of a video encoding apparatus to which a padding method of a reference image outside region using prediction information according to an embodiment is applied.
  • FIG. 10B is a block diagram of a video decoding apparatus to which a padding method of a reference image outside region using prediction information according to an embodiment is applied.
  • FIG. 11 illustrates a process in which an image decoding apparatus determines at least one encoding unit by dividing a current encoding unit according to an embodiment.
  • FIG. 12 illustrates a process in which an image decoding apparatus determines at least one encoding unit by dividing a non-square encoding unit according to an embodiment.
  • FIG. 13 illustrates a process in which an image decoding apparatus divides an encoding unit based on at least one of information on a block format information and a format mode according to an embodiment.
  • FIG. 14 illustrates a method for an image decoding apparatus to determine a predetermined encoding unit among odd number of encoding units according to an embodiment.
  • FIG. 15 illustrates a sequence in which a plurality of coding units are processed when an image decoding apparatus determines a plurality of coding units by dividing a current coding unit according to an embodiment.
  • 16 illustrates a process of determining that the current encoding unit is divided into odd number of encoding units when the image decoding apparatus can not process the encoding units in a predetermined order according to an embodiment.
  • FIG. 17 illustrates a process in which an image decoding apparatus determines at least one encoding unit by dividing a first encoding unit according to an embodiment.
  • FIG. 18 is a diagram illustrating an example in which when the non-square second encoding unit determined by dividing the first encoding unit by the image decoding apparatus satisfies a predetermined condition, a form in which the second encoding unit can be divided is limited Lt; / RTI >
  • FIG. 19 illustrates a process in which an image decoding apparatus divides a square-shaped encoding unit when the information on the split mode mode can not be divided into four square-shaped encoding units according to an exemplary embodiment.
  • FIG. 20 illustrates that a processing order among a plurality of coding units may be changed according to a division process of coding units according to an embodiment.
  • FIG. 21 illustrates a process in which the depth of an encoding unit is determined according to a change in type and size of an encoding unit when a plurality of encoding units are determined by recursively dividing an encoding unit according to an exemplary embodiment.
  • FIG. 22 illustrates a depth index (hereinafter referred to as a PID) for classifying a depth and a coding unit that can be determined according to the type and size of coding units according to an exemplary embodiment.
  • a PID depth index
  • FIG. 23 illustrates that a plurality of coding units are determined according to a plurality of predetermined data units included in a picture according to an embodiment.
  • FIG. 24 shows a processing block serving as a reference for determining a determination order of reference encoding units included in a picture according to an embodiment.
  • part used in the specification means software or hardware component, and "part " However, “part” is not meant to be limited to software or hardware. “Part” may be configured to reside on an addressable storage medium and may be configured to play back one or more processors.
  • part (s) refers to components such as software components, object oriented software components, class components and task components, and processes, Subroutines, segments of program code, drivers, firmware, microcode, circuitry, data, databases, data structures, tables, arrays and variables.
  • the functions provided in the components and “parts " may be combined into a smaller number of components and” parts " or further separated into additional components and “parts ".
  • processor may be embodied in a processor and memory.
  • the term “processor” should be broadly interpreted to include a general purpose processor, a central processing unit (CPU), a microprocessor, a digital signal processor (DSP), a controller, a microcontroller, a state machine, In some circumstances, a “processor” may refer to an application specific integrated circuit (ASIC), a programmable logic device (PLD), a field programmable gate array (FPGA)
  • ASIC application specific integrated circuit
  • PLD programmable logic device
  • FPGA field programmable gate array
  • processor refers to a combination of processing devices, such as, for example, a combination of a DSP and a microprocessor, a combination of a plurality of microprocessors, a combination of one or more microprocessors in conjunction with a DSP core, It can also be called.
  • memory should be broadly interpreted to include any electronic component capable of storing electronic information.
  • the terminology memory may be any suitable memory such as random access memory (RAM), read-only memory (ROM), non-volatile random access memory (NVRAM), programmable read-only memory (PROM), erase- May refer to various types of processor-readable media such as erasable programmable read-only memory (PROM), flash memory, magnetic or optical data storage devices, registers, and the like.
  • RAM random access memory
  • ROM read-only memory
  • NVRAM non-volatile random access memory
  • PROM programmable read-only memory
  • erase- May to various types of processor-readable media such as erasable programmable read-only memory (PROM), flash memory, magnetic or optical data storage devices, registers, and the like.
  • a memory is said to be in electronic communication with a processor if the processor can read information from and / or write information to the memory.
  • the memory integrated in the processor is in electronic communication with the processor.
  • the "image” may be a static image such as a still image of a video or a dynamic image such as a moving image, i.e., the video itself.
  • sample means data to be processed as data assigned to a sampling position of an image.
  • pixel values in the image of the spatial domain, and transform coefficients on the transform domain may be samples.
  • a unit including at least one of these samples may be defined as a block.
  • the "padding” operation refers to an operation of copying a sample value of another pixel to generate a sample value of the current pixel.
  • FIG. 1A shows a flowchart of a padding method of an image outside region using prediction information according to an exemplary embodiment.
  • the current block means the block of the current image
  • the reference block means the block indicated by the motion vector of the current image in the reference image of the current image.
  • step s10 prediction information of a current block adjacent to an outline of a current image is obtained.
  • An image padding method attempts to predict a sample value of an outer region located outside an outline of a current image using prediction information of a current block when prediction information of a current block exists.
  • the reference samples for predicting the samples located outside the outline of the current block can be determined using the prediction information of the current block in step s12.
  • Reference samples may be located within the current or reference image.
  • the samples located outside the outline of the current image, which are adjacent to the current block using the reference samples determined in step s12, may be determined. As a result, the samples located in the outer region of the current image can be padded.
  • the prediction information of the motion prediction method to which the padding method of the reference region is applied may be intra prediction information of a neighboring block or intra mode information of a neighboring block or may be affine motion information of neighboring blocks.
  • the inter prediction information of a current block used for padding of an image outside area may include at least one of a reference picture list, a reference index, and a motion vector. More specifically, the reference image of the current block can be determined using the inter prediction information. The reference image may be determined from among the images included in at least one of the reference list of the current block, i.e., the L0 list and the L1 list. The position indicated by the motion vector of the current block among the blocks of the reference image may be the position of the reference block.
  • the reference picture list of the inter prediction information may indicate at least one of the L0 reference list and the L1 reference list.
  • one reference image among the images included in the L0 reference list can be determined using the L0 reference index.
  • one reference image among the images included in the L1 reference list can be determined using the L1 reference index.
  • the image padding method may use the neighboring blocks of the reference block determined according to the inter prediction information of the current block to padd the samples outside the outline of the current image. That is, the samples of the pixels included in the neighboring block of the reference block indicated by the motion vector of the current block may be determined as reference samples, and the sample values of the image outside region adjacent to the current block may be determined using the reference samples.
  • the outer samples of the current block can be determined using reference samples of neighboring blocks adjacent to the left side of the reference block of the current block.
  • the outer samples of the current block can be determined using reference samples of neighboring blocks adjacent to the upper side of the reference block of the current block.
  • the outer samples of the current block can be determined using reference samples of neighboring blocks adjacent to the right side of the reference block of the current block.
  • the outer samples of the current block can be determined using the reference samples of the neighboring block below the reference block of the current block.
  • the reference samples can be determined using the remaining prediction information excluding the prediction information corresponding to the motion vector.
  • samples adjacent to the outline among the samples of the current block are determined as reference samples, The image outside area can be padded.
  • reference samples for padding the outer region can be determined according to the intra mode information of the current block.
  • the intra mode information of the current block used for padding of the out-of-picture area may be an intra prediction mode indicating a direction of neighboring samples referred to predict a sample of the current block.
  • the intra-prediction mode may include 33 directional prediction modes, a DC mode, and a planar mode.
  • the present invention is not limited thereto, and the direction prediction mode of the intra prediction mode can be subdivided into 62 direction prediction modes.
  • the intra-prediction mode may include at least one of a DC mode, a vertical mode, a horizontal mode, a planar mode, a diagonal mode in an upper left and lower right direction, and a diagonal mode in an upper right and lower left direction .
  • the samples in the current image which are located according to the direction indicated by the intra prediction mode of the current block, can be determined as reference samples for padding the outer sample.
  • a method of padding an image outside area according to the intra prediction mode of the current block will be described later with reference to FIG. 4B.
  • the characteristics of a block adjacent to an outline may be determined using a method such as gradient, edge detect, etc.
  • the prediction information may include information on the characteristics of the block adjacent to the outline . ≪ / RTI >
  • a method of padding an outer region of an image may be applied. More specifically, a method of padding a reference block when a reference block for motion prediction is located outside the reference image will be described below with reference to FIGS. 1B and 2B.
  • a conventional padding method can be used to generate reference samples values. That is, at the time of in-loop filtering, a reference sample can be generated by copying samples inside an image adjacent to a reference sample around a video outline to padding a reference sample located outside the image .
  • a padding method for an image outside region can find a block related to a current block through template matching, and can perform padding using a related block.
  • Template matching is a method of determining a template area of a certain type surrounding a block and determining that a block surrounded by the detected area is a similar block of the block when the closest area to the template area is detected in the reference image. Accordingly, if template matching is applied to the padding method according to the present disclosure, the following procedure can be followed. If the current block is adjacent to the outline of the current image, a template of the current block in the current image may be determined, and a block surrounded by the area detected by template matching may be determined as a similar block of the current block. The sample values of the block adjacent to the pseudo block of the current block can be determined as the sample values for padding the image outside region adjacent to the current block.
  • FIG. 1B shows a flowchart of an applied motion prediction method of a padding method of a reference image outside region according to an embodiment.
  • a reference block for inter prediction of a current block is divided into sub- Prediction information of at least one block adjacent to the inside of the outline of the reference image is obtained.
  • the motion vector of the current block can be determined by inter prediction of the current block.
  • the sample value of the current block can be compensated using the sample value of the reference block, that is, the area indicated by the motion vector of the current block.
  • the motion vector is 'unavailable' Is defined.
  • the motion prediction method includes the steps of acquiring prediction information of another block adjacent to an inside of an outline of a reference image when a motion vector of the current block indicates an area outside the outline of the reference image, To estimate the sample values of the reference block located outside the outline.
  • the reference image may be determined from among the images included in at least one of the reference list of the current block, i.e., the L0 list and the L1 list.
  • the position indicated by the motion vector of the current block among the blocks of the reference image may be the position of the reference block.
  • the positions of all the pixels included in the reference block may be located outside the outline of the reference image. Or the position of some pixels included in the reference block may be located outside the outline of the reference image.
  • the reference samples used for predicting the samples located outside the outline of the reference image are determined using the prediction information of the block adjacent to the inside of the outline of the reference image in step s13.
  • the samples located outside the outline of the reference image are determined using the determined reference samples.
  • a reference picture selected from a reference list of a current block is referred to as a 'current reference picture', and a reference block or a region indicated by a current block motion vector in a 'current reference picture' Reference block " and " current reference region ".
  • an image for predicting a sample located outside the outline of the current reference image (hereinafter referred to as an "external sample”) is referred to as a reference image by prediction information of a block adjacent to the inside of the outline of the current reference image it means.
  • the prediction information of the motion prediction method to which the padding method of the reference image outside region is applied may be determined based on the inter prediction information of the block adjacent to the inside of the outline of the current reference image or the intra prediction mode of the block adjacent to the inside of the outline of the current reference image, Information.
  • inter prediction information among prediction information of adjacent blocks in the outline may be used.
  • the 'image for predicting the external samples' of the current reference image can be determined from the images included in at least one of the reference list of the 'block adjacent to the inside of the outline' of the current reference image, that is, the L0 list and the L1 list.
  • the reference block of the 'block adjacent to the outline' can be determined based on the position indicated by the motion vector of the 'block adjacent to the outline' within the 'reference image for predicting the external samples'.
  • Outer samples of the current reference image can be predicted using samples adjacent in the outline direction of the 'reference image of a block adjacent to the outline' among the samples adjacent to the reference block of the 'block adjacent to the inside of the outline'
  • Reference samples for padding outer samples may be located within an image for predicting outer samples.
  • the outer samples can be padded using the samples adjacent to the inside of the outline of the current reference image.
  • the outer samples may be samples of the pixels included in the reference block of the current block.
  • some pixels included in the reference block of the current block may be padded with samples of some pixels, i.e., the outer samples.
  • the prediction information of the block adjacent to the inside of the outline may include the affine mode information, and the affine mode information , It is possible to determine reference samples for predicting outer samples of the current reference image.
  • the inter prediction information of a block adjacent to the inside of the outline of the current reference image may include at least one of a reference picture list, a reference index, and a motion vector .
  • a reference picture list an image indicated by the reference index in the reference picture list of a block adjacent to the inside of the outline of the current reference image is determined as a reference image for predicting the external samples.
  • the position indicated by the motion vector of the block adjacent to the inside of the outline of the reference image among the reference images for predicting the external samples can be determined as a block to be a reference object.
  • Neighboring samples of a block to be referred to can be determined as samples for predicting external samples, i.e., samples located outside the outline of the current reference image.
  • the inter prediction information of the block adjacent to the inside of the left outline is determined.
  • Blocks or samples located outside the left outline of the current reference image may be predicted using neighboring blocks or samples adjacent to the left side of the reference block in the determined image.
  • the inter prediction information of the block adjacent to the inside of the upper outline The reference block in the image and the image is determined. Blocks or samples located outside the upper outline of the current reference image may be predicted using neighboring blocks or samples adjacent to the upper side of the reference block in the determined image.
  • Blocks or samples located outside the right outline of the current reference image may be predicted using neighboring blocks or samples adjacent to the right side of the reference block within the determined image.
  • a reference block in a video and an image to be a reference of a block adjacent to the inside of the outline is determined.
  • a block or samples located outside the lower outline of the current reference image may be predicted using neighboring blocks or samples that are adjacent to the lower side of the reference block in the determined image.
  • a reference picture list of inter prediction information of a block adjacent to an inside of an outline of a current reference picture includes at least one of an L0 reference list and an L1 reference list Can point to.
  • an L0 reference list of a block adjacent to an inside of an outline of a current reference image an image indicated by the L0 reference index of a block adjacent to the inside of the outline of the current reference image among the images included in the L0 reference list is referred to
  • the L0 image for predicting the samples located outside the outline of the image can be determined.
  • an L1 reference list of a block adjacent to an inside of an outline of a current reference image an image indicated by an L1 reference index of a block adjacent to the inside of the outline of the current reference image among the images included in the L1 reference list, It is possible to determine the L1 image for predicting the samples located outside the outline of the reference image. If the position of the block neighboring the outline of the block indicated by the L1 motion vector of the block adjacent to the inside of the current reference image is located outside the L1 image for predicting the samples located outside the outline, Samples located outside the outline may be padded with the sample values within the outline of the current reference image.
  • Prediction mode according to the Bi-prediction mode available for the L0 reference list and the L1 reference list of the block adjacent to the inside of the outline of the current reference image the reference image according to the L0 reference list of the block adjacent to the inside of the outline
  • the samples of the reference image according to the L1 reference list the outer samples of the current reference image can be predicted.
  • the L0 reference image in the L0 reference list of the block adjacent to the inside of the current reference image is determined to be an image for predicting a sample outside the outline, and among the L0 reference images, the L0 motion vector The L0 reference block can be determined.
  • the L1 reference picture of the L1 reference list of the block adjacent to the inside of the current reference picture is an image for predicting the sample outside the outline, and the L1 motion vector Lt; / RTI > reference block may be determined.
  • the neighboring samples of the L0 reference block in the L0 reference image of the block adjacent to the inside of the outline and the neighboring samples of the L1 reference block in the L1 reference image can be used to predict the outer samples of the current reference image. For example, using the weight based on the distance between the current reference image and the L0 reference image and the distance based on the distance between the current reference image and the L1 reference image, weights of neighboring samples of the L0 reference block and neighboring samples of the L1 reference block May be determined to be the value of the outer samples of the current reference image.
  • the inter prediction according to the Bi-prediction mode available for the L0 reference list and the L1 reference list of the block adjacent to the inside of the current reference image is performed and the L0 motion vector of the block adjacent in the outline of the current reference image and the L1
  • a motion indicating the outside of the image for predicting the samples located outside the outline It is possible to determine reference samples for padding samples located outside the outline of the current reference image using the prediction information other than the prediction information corresponding to the vector.
  • Another example is a case where an area neighboring the outline of the area indicated by the L0 motion vector and the L1 motion vector of the block adjacent to the inside of the current reference image is located outside the image for predicting the samples located outside the outline It is possible to padd samples located outside the outline of the reference image using the samples touched inside the outline of the current reference image.
  • inter prediction according to the Bi-prediction mode available for the L0 reference list and the L1 reference list of the block adjacent to the current reference image is performed, and L0
  • the reference image of the motion vector indicating the smaller one of the two neighboring areas is determined as an image for predicting samples located outside the outline of the reference image, and neighboring samples of the L0 reference block in the determined image and neighboring samples of the L1 reference block in the L1 reference image are used, Samples can be predicted.
  • a reference sample for padding an outer area of an outline of a reference image may be determined from a block in a reference image according to intra mode information of a motion estimation method to which a padding method of a reference image outer area according to an embodiment is applied.
  • the intra mode information of a block adjacent to an inside of an outline of a current reference image used for padding of a reference image outside region may be predicted by predicting a sample of a block adjacent to an inside of an outline of a current reference image, Prediction mode indicating the direction of neighboring samples to be referred to in order to determine the direction of neighboring samples.
  • the characteristics of the block adjacent to the inside of the outline of the current reference image may be determined using a method such as gradient, edge detection, etc.
  • Prediction information of the blocks adjacent to each other may include information on the characteristics of blocks adjacent to the outline.
  • a method of padding a reference block according to intra information of a block adjacent to an inside of an outline of a current reference image will be described with reference to FIG. 4B.
  • a motion estimation method to which a padding method of an outer region of a reference image is applied is to find a block related to a block adjacent to an inside of an outline of a current reference image through a template matching method, Padding can be performed by using.
  • template matching is applied to the padding method of the reference image outside region according to the present disclosure, the following procedure can be followed.
  • the reference block of the current block is out of the reference image, the template of the boundary block within the reference image adjacent to the reference block is determined, and the block surrounded by the area detected by the template matching is determined as the similar block of the boundary block .
  • the reference block of the current block can be padded using the sample values of the block adjacent to the pseudo block of the boundary block.
  • FIG. 2A illustrates a block diagram of an image padding apparatus using prediction information according to an embodiment.
  • an image padding apparatus 20 using prediction information includes a prediction information obtaining unit 22, a reference sample determining unit 24, and an external sample determining unit 26.
  • the image padding apparatus 20 using prediction information includes a central processor (not shown) for controlling the prediction information obtaining unit 22, the reference sample determining unit 24, and the external sample determining unit 26 can do.
  • the prediction information obtaining section 22, the reference sample determining section 24 and the external sample determining section 26 are operated by respective ones of the processors (not shown), and the processors (not shown)
  • the motion prediction apparatus to which the padding method of the reference region is applied can be operated as a whole.
  • the prediction information obtaining unit 22 of the image padding apparatus 20 using the prediction information can obtain the prediction information of the current block adjacent to the outline of the current image.
  • the reference sample determining unit 24 can use the prediction information of the current block to determine reference samples for predicting samples located outside the outline of the current block.
  • the outer sample determination unit 26 may determine samples located outside the outline of the current image adjacent to the current block using the reference samples.
  • the prediction information obtaining section 22 determines that the prediction information of the current block exists, Whether the pixel position is located in one of the upper left, upper right, lower left, and right lower sides of the current picture. If the prediction information of the current block exists and the pixel position of the outer region is not a position adjacent to one of the upper left side corner, the upper right side corner, the lower left side corner, and the lower right side corner of the current picture, the reference sample determination unit 24 It can be confirmed whether there is a neighboring sample of the reference block indicated by the prediction information of the current block.
  • the inter prediction information of the current block it can be confirmed whether neighbor samples neighboring the reference block among the reference images of the current block exist. If intraprediction information of the current block is used, it can be confirmed whether there is a neighboring sample located in a direction according to the intra prediction mode of the current block from the pixel position of the outer region, among the samples adjacent to the inside of the outline of the current image.
  • the sample value of the current area can be determined using the neighboring sample.
  • the sample value of the current image may be determined using the sample in the outline of the current image adjacent to the outer region in the horizontal or vertical direction .
  • the sample value of the current image can be determined using the sample in the outline of the current image adjacent to the outside region in the horizontal or vertical direction.
  • the pixel positions of the outer region are adjacent to the left upper corner of the current picture in the upper left diagonal direction, the upper right corner is adjacent to the upper right corner in the right upper diagonal direction, the lower left corner is adjacent in the lower left diagonal direction,
  • the sample value of the current image can be determined using the sample in the outline of the current image which is adjacent to the outside region in the horizontal or vertical direction.
  • the method of padding the outer region of the image using the prediction information of the current block corresponds to the above description with reference to FIG. 1A, and thus a duplicated description will be omitted.
  • FIG. 2B is a block diagram of a motion prediction apparatus 21 to which a padding method of a reference image outside region using prediction information is applied according to an embodiment.
  • the motion prediction apparatus 21 to which the reference region padding method according to an embodiment is applied includes a prediction information obtaining unit 23, a reference sample determining unit 25, and an external sample determining unit 27 .
  • the motion prediction apparatus 21 to which the padding method of the reference region according to the embodiment is applied includes a central processor (not shown) for controlling the prediction information obtaining unit 23, the reference sample determining unit 25 and the outer sample determining unit 27 Time).
  • the prediction information obtaining section 23, the reference sample determining section 25, and the external sample determining section 27 are operated by respective ones of the processors (not shown), and the processors (not shown) Accordingly, the motion prediction apparatus to which the padding method of the outer region of the reference image is applied can be operated as a whole.
  • the prediction information obtaining unit 23 of the motion prediction apparatus to which the padding method of the reference region according to an embodiment applies the current reference is referred to as a current reference when the motion vector determined by the inter prediction of the current block indicates an area outside the current reference image
  • Prediction information of neighboring blocks within the outline of the image can be obtained.
  • the reference sample determining unit 25 can determine reference samples for predicting samples located outside the outline by using prediction information of a block adjacent to the inside of the outline of the current reference image.
  • the outer sample determination unit 27 may determine a sample located outside the outline of the current reference image using the determined reference samples for predicting the samples located outside the current reference image.
  • a method of determining reference samples for padding an outer region of a current reference image using prediction information of a block adjacent to an inside of an outline of a current reference image is as follows. Referring to FIG. 1B, Method, and thus redundant description will be omitted.
  • a reconstructed image generated by decoding a current image can be used as a reference image for motion prediction of another image.
  • the video coding system and the video decoding system may pre-padd the outer area of the restored image of the current image to generate a padding larger than the current image, A padded picture can be generated.
  • the image padding apparatus 20 and the motion prediction apparatus 21 may be configured to classify the outer region of the decoded reconstructed image into reference samples determined using intra prediction information or intra prediction information in the reconstructed image By padding, an image larger than the original size of the restored image can be stored in the memory.
  • the image padding device 20 and the motion prediction device 21 may store the decoded original size reconstructed image first in the memory, and determine that the reference block for motion prediction of the other image exists in the outer region of the reconstructed image It is possible to generate the sample values of the reference block by padding the outer region with the reference samples determined using the intra prediction information or the intra prediction information in the reconstructed image.
  • FIG. 3 illustrates a method of padding an outer region of a reference image according to a conventional padding method.
  • the conventional padding method is a method of padding samples 33 located outside the reference image contour 31, by using an outline line contiguous to the image contour line 31 among the pixels of the contour block 35 And copies the sample values of the pixels to generate sample values of the samples 33 located outside the reference image contour 31.
  • the mirroring method of using outline pixels arranged in the padding direction 37 among the pixels in the outline pixels and copying in the direction symmetrical about the contour line 31 And also generates the sample values of the samples 33.
  • 4A illustrates a method of padding an outer region of a reference image using inter information according to an embodiment.
  • the predicted motion may be used to pad the sample values of the outer pixels 43 that are located outside the outline of the reference image 40.
  • the sample values of the external pixels 43 can be predicted using the inter-information of the outline block 45 adjacent to the inside of the outline of the reference image 40 adjacent to the external pixels 43.
  • the block 41 to be a reference of the outline block 45 can be determined using the reference list, the reference index, and the motion vector of the outline block 45. More specifically, in the reference list of the outline block 45, the image indicated by the reference index becomes the reference image 42 of the outline block 45, and the motion vector of the outline block 45 of the reference image 42 The block 41 to be referred to can be determined.
  • the samples of the neighboring block 49 adjacent to the reference block 41 may be determined as reference samples for padding the outer pixels 43 of the reference image 40. [ Therefore, the sample values of the outer pixels 43 of the reference image 40 can be replaced by the sample values of the neighboring block 49 adjacent to the block 41 to be referred to.
  • 4B illustrates a method of padding an outer region of a reference image using intra information according to an exemplary embodiment of the present invention.
  • the intra prediction mode 44 of the block 45 in order to padding the outer pixels 43 located outside the outline of the reference image 40, the intra prediction mode 44 of the block 45 to be used.
  • the intra prediction mode 44 may indicate that neighboring sample values located in the lower-right direction 48 may be used to predict a sample of a block 45 that is contiguous within the outline of the reference image. (45, 48) adjacent to the inside of the outline of the reference image (40) according to the direction indicated by the intra prediction mode (44) of the block (45) adjacent to the inside of the outline of the reference image.
  • the external pixels 43 of the image 40 may be padded.
  • the sample values of the pixels positioned in the lower right direction of the outer pixels 43 of the reference image 40 among the pixels adjacent to the reference image outline in the outline blocks 45 and 48 are detected by the reference image 40,
  • the sample values of the external pixels 43 of the pixel P may be determined.
  • FIG. 5 shows an example of a reference image to which a conventional padding method is applied.
  • the conventional padding method is a method in which sample values of pixels of a reference image outside area 55 outside the outline 53 of a reference image 51 are sampled by the inner pixels 53 adjacent to the outline 53 of the reference image,
  • the reference image outside area 55 is copied. That is, the sample values of the reference image outside region 55 adjacent to the vertical outline of the outline 53 copy the sample values of the internal pixels of the reference image 51 adjacent to the vertical outline in the horizontal direction in the horizontal direction.
  • the sample values of the reference image outside region 55 adjacent to the horizontal outline of the outline 53 copy the sample values of the internal pixels of the reference image 51 adjacent to the horizontal outline in the vertical direction in the vertical direction.
  • FIG. 6 illustrates an example of a reference image to which the padding method according to an embodiment is applied.
  • a method of padding a reference image outside area 65 using inter-prediction information of a block adjacent to an inside of an outline 63 of a reference image 61 is applied.
  • the reference image outside area 65 located outside the outline line 63 of the reference image 61 is included in the other image to be referenced indicated by the inter prediction information of the adjacent block inside the outline line 63 Can be padded using reference samples. Therefore, a portion 67 of the reference image outside area 65 may be generated so as to include an area not included in the reference image 61.
  • the reference image outside area 65 outside the outline 63 is predicted using the prediction information of the block inside the reference image 61, so that the reference image outside area 65 May also include information that can not be obtained from within the reference image 61. Therefore, the accuracy of motion prediction using the reference image including the reference image outside area 65 is improved, and the coding efficiency of the inter prediction can be improved.
  • a motion prediction method to which a padding method of a reference image using prediction information is applied can be applied to a method of decoding a 360 degree image.
  • a 360 degree image consists of a number of images taken in various directions.
  • the existing video encoding method and video decoding method assume that the image has a rectangular shape.
  • a plurality of images constituting a 360 degree image are arranged in a predetermined order to constitute one image, and the existing video coding method and video decoding method are applied to the image.
  • FIG. 7 shows an example of a reference image of a 360-degree image to which a conventional padding method is applied.
  • a reference image 75 of a 360-degree image is composed of images 75-1, 75-2, 75-3, 75-4, 75-5, and 75-6 photographed in six directions . Since there is no external sample value of the image outline 71 of the reference image 75 of the 360-degree image, the same value as the samples adjacent to the inside of the outline 71 of the 360- By padding, the outer region 73 of the reference image 75 of the 360 degree image can be generated.
  • FIG. 8 illustrates an example of a reference image of a 360-degree image to which a padding method of a reference image outside region according to an embodiment is applied.
  • a motion prediction method to which a padding method of a reference image outside area according to an embodiment is applied is applied to a reference image 75 of a 360-degree image of FIG.
  • the reference image 75 of the 360 degree image is divided into the respective images 75-1, 75-2, 75-3, 75-4, 75-5, and 75-6, Can be applied to each of the images 75-1, 75-2, 75-3, 75-4, 75-5, and 75-6.
  • the reference image outside area 83-1 can be generated.
  • the reference image is the image 75-2, 75-3, 75-4, 75-5, 75-6
  • the image 75-2, 75-3, 75-4, 75-5 81-2 and 81-6 based on the inter prediction information of the contour blocks adjacent to the inside of the contour lines 81-2, 81-3, 81-4, 81-5, 83-3, 83-4, 83-5, and 83-6 outside the reference image region outside the reference image regions 83-2, 83-4, 81-5, and 81-6.
  • FIG. 9 illustrates an example of a 360-degree image to which a padding method of a reference image outside area is applied according to another embodiment.
  • a single image is composed of a Rotated Sphere Projection (RSP), an Equirectangular Projection (ERP), an Icosahedral Projection (ISP), a Segmented Sphere Projection (SSP), a CubeMap Projection (EAC) and so on.
  • RSP Rotated Sphere Projection
  • ERP Equirectangular Projection
  • ISP Icosahedral Projection
  • SSP Segmented Sphere Projection
  • EAC CubeMap Projection
  • the actual region photographed may be different from the region to be encoded.
  • the area to be encoded is larger than the actual area, it is a conventional method to fill the empty area other than the actual area through a method such as interpolation.
  • a padding method of a reference image outside area may be applied to a free area of the current image.
  • the reference image 90 of the 360-degree image of FIG. 9 is an image generated by Rotated Sphere Projection (RSP).
  • the reference image 90 of the 360-degree image may include an image 95 and a free area (or padding area) 93.
  • the reference image may be a 360 degree image 90 and the empty area 93 may exist outside the reference image outline 91.
  • the sample value of the free area 93 can be predicted using the reference sample values determined using the inter prediction information or the intra prediction information of the block located inside the outline 91.
  • the accuracy of the motion prediction is increased by filling the outline region outside the actual value by using the prediction information of the outline block inside the outline of the reference image.
  • FIG. 10A is a block diagram of a video encoding apparatus 2000 to which an image padding method using prediction information according to an embodiment is applied.
  • the video encoding apparatus 2000 includes a predictive encoding unit 2015, a transform and quantization unit 2020, an entropy encoding unit 2025, an inverse quantization and inverse transformation unit 2030, a deblocking filtering unit 2035, (2040).
  • the predictive coding unit 2015 can determine prediction information for indicating a predictive sample by inter prediction or intra prediction on a current block of the input image.
  • inter prediction the prediction samples of the current block are determined from the reference image that has been coded first, and in the case of intra prediction, the prediction samples of the current block can be determined from the samples that are coded first in the current image.
  • the transform and quantization unit 2020 transforms and quantizes the residual blocks, which are difference components between the predicted samples of the predicted block and the samples of the current block, and outputs the quantized transform coefficients.
  • the entropy encoding unit 2025 can perform entropy encoding on the quantized transform coefficients to output a bitstream.
  • the inverse quantization and inverse transform unit 2030 performs inverse quantization and inverse transform on the quantized transform coefficients to reconstruct the residual block in the spatial domain.
  • the reconstructed block in which the reconstructed residual block is combined with the prediction block is input to the deblocking filtering unit 2035 and the loop filtering unit 2040, and the reconstructed block is filtered and output to the reconstructed block.
  • the reconstructed image composed of reconstructed blocks modified through loop filtering can be used as a reference image for inter prediction of the next block in the predictive coding unit 2015.
  • a method of padding the outer region of the current image described above with reference to FIGS. 1A to 9 may be applied when the current block is adjacent to the outline of the current image and inter prediction information or intra prediction information for the current block exists. That is, the operation of the image padding apparatus 20 described above with reference to FIG. 2A may be performed in the predictive encoding unit 2015. FIG. The operation of the image padding device 20 performed in the predictive encoding unit 2015 is the same as the operation described above with reference to FIGS. 1A to 9, and thus a duplicated description will be omitted.
  • the padding method of the reference picture outside region described above with reference to Figs. 1B to 9 is applied . That is, the operation of the motion prediction apparatus 21 described above with reference to FIG. 2B may be performed by the predictive encoding unit 2015. FIG.
  • the operation of the motion prediction unit 21 performed in the predictive encoding unit 2015 is the same as the operation described above with reference to FIGS. 1B and 2B, and thus a duplicated description will be omitted.
  • FIG. 10B shows a block diagram of a video decoding apparatus 2080 to which an image padding method using prediction information according to an embodiment is applied.
  • the video decoding apparatus 2080 includes an entropy decoding unit 2055, an inverse quantization and inverse transform unit 2060, a predictive decoding unit 2075, a deblocking filtering unit 2065 and a loop filtering unit 2070.
  • the entropy decoding unit 2055 performs entropy decoding on the received bitstream to obtain quantized transform coefficients of the image.
  • the inverse quantization and inverse transform unit 2060 performs inverse quantization and inverse transform on the quantized transform coefficients of the transform block to reconstruct the residual block in the spatial domain.
  • the prediction decoding unit 2075 can determine reference blocks of the reference block using the prediction information for inter prediction or intra prediction for the current block and determine prediction blocks using the reference samples.
  • inter prediction the prediction samples of the current block are determined first among reconstructed reference images, and in the case of intra prediction, the prediction samples of the current block can be determined among the samples reconstructed in the current image.
  • a reconstruction block in the spatial domain is determined by combining the prediction block and the residual block, the sample values of the reconstruction block are modified through the deblocking filtering unit 2065 and the loop filtering unit 2070, A restored image of the current image can be output.
  • the reconstructed image can be used as a reference image for predicting the next image in the predictive decoding unit 2075.
  • the padding method of the reference region described above with reference to FIGS. Can be applied. That is, the operation of the image padding apparatus 20 described above with reference to FIG. 2A may be performed in the predictive decoding unit 2075.
  • the operation of the image padding device 20 performed in the predictive encoding unit 2075 is the same as the operation described above with reference to FIGS. 1A to 9, and thus a duplicated description will be omitted.
  • FIG. 1B to 9 is applied . That is, the operation of the motion prediction apparatus 21 described above with reference to FIG. 2B may be performed in the predictive decoding unit 2075.
  • FIG. The operation of the motion prediction apparatus 21 performed in the predictive encoding unit 2075 is the same as the operation described above with reference to FIGS. 1B and 2B, and thus a duplicated description will be omitted.
  • FIG. 11 illustrates a process in which the video decoding apparatus 2080 divides a current encoding unit to determine at least one encoding unit according to an embodiment.
  • the block shape may include 4Nx4N, 4Nx2N, 2Nx4N, 4NxN, or Nx4N. Where N may be a positive integer.
  • the block type information is information indicating at least one of a ratio, or a size, of a shape, direction, width, and height of an encoding unit.
  • the shape of the encoding unit may include a square and a non-square. If the width and height of the encoding unit are the same (4Nx4N), the image decoding apparatus 2080 can determine the block type information of the encoding unit as a square. The image decoding apparatus 2080 can determine the shape of the encoding unit to be non-square.
  • the image decoding apparatus 2080 can determine the block type information of the encoding unit as a non-square. If the shape of the coding unit is a non-square shape, the image decoding apparatus 2080 divides the width and height ratio of the block type information of the coding unit by 1: 2, 2: 1, 1: 4, 4: Or 8: 1. Further, based on the length of the width of the coding unit and the length of the height, the video decoding apparatus 2080 can determine whether the coding unit is the horizontal direction or the vertical direction. Further, the image decoding apparatus 2080 can determine the size of the encoding unit based on at least one of the width of the encoding unit, the length of the height, and the width.
  • the image decoding apparatus 2080 can determine a type of an encoding unit using block type information and determine a type of an encoding unit to be divided using information on the division type mode. That is, the division method of the coding unit indicated by the information on the division mode can be determined according to which block type the block type information used by the video decoding apparatus 2080 represents.
  • the image decoding apparatus 2080 can obtain information on the split mode mode from the bit stream. However, the present invention is not limited thereto, and the image decoding apparatus 2080 and the image encoding apparatus 2200 can acquire information on the promised split mode mode on the basis of the block type information.
  • the image decoding apparatus 2080 can acquire information on the promised divided mode mode for the maximum encoding unit or the minimum encoding unit. For example, the image decoding apparatus 2080 can determine the size of the maximum encoding unit to be 256x256.
  • the image decoding apparatus 2080 can determine the information about the promised division mode in advance by quad split. Quad partitioning is a split mode mode that bisects both the width and the height of the encoding unit.
  • the image decoding apparatus 2080 can obtain a 128x128 encoding unit from the maximum encoding unit of 256x256 size based on the information on the split mode mode. Also, the image decoding apparatus 2080 can determine the size of the minimum encoding unit to be 4x4. The image decoding apparatus 2080 can acquire information on the division mode mode indicating "not divided" for the minimum encoding unit.
  • the image decoding apparatus 2080 may use block type information indicating that the current encoding unit is a square format. For example, the image decoding apparatus 2080 can determine whether to divide a square encoding unit into vertically, horizontally, or four encoding units according to the information on the split mode mode.
  • the video decoding apparatus 2080 decodes the current encoding unit 300 according to information on the split mode mode indicating that the current block is not divided, It is possible to determine the divided coding units 310b, 310c, and 310d based on the information on the division mode mode indicating the predetermined division method, without dividing the coding unit 310a having the same size as the coding unit 310a.
  • the image decoding apparatus 2080 includes two encoding units 310b (FIG. 11) that divide a current encoding unit 300 in the vertical direction based on information on a split mode mode indicating that the image is divided vertically according to an embodiment. Can be determined.
  • the image decoding apparatus 2080 can determine two encoding units 310c in which the current encoding unit 300 is divided in the horizontal direction based on the information on the split mode mode indicating that the image is divided in the horizontal direction.
  • the image decoding apparatus 2080 can determine four encoding units 310d obtained by dividing the current encoding unit 300 in the vertical direction and the horizontal direction based on the information on the split mode mode indicating that the image is divided vertically and horizontally have.
  • the division type in which the square encoding unit can be divided should not be limited to the above-mentioned form, but may include various forms in which the information on the division type mode can be represented.
  • the predetermined divisional form in which the square encoding unit is divided will be described in detail by way of various embodiments below.
  • FIG. 12 illustrates a process in which the image decoding apparatus 2080 determines at least one encoding unit by dividing a non-square encoding unit according to an embodiment.
  • the image decoding apparatus 2080 may use block type information indicating that the current encoding unit is a non-square format.
  • the image decoding apparatus 2080 can determine whether to divide the non-square current encoding unit according to the information on the split mode mode or not in a predetermined method.
  • the image decoding apparatus 2080 decodes the current encoding
  • the encoding unit 410 or 460 having the same size as the unit 400 or 450 is determined or the encoding unit 420a, 420b, 430a, or 430b divided based on the information on the division mode mode indicating the predetermined division method , 430c, 470a, 470b, 480a, 480b, 480c.
  • the predetermined division method in which the non-square coding unit is divided will be described in detail through various embodiments.
  • the image decoding apparatus 2080 may determine a type in which an encoding unit is divided using information on a split mode mode.
  • information on the split mode mode may include at least one Lt; / RTI > can be represented by the number of encoding units.
  • the image decoding apparatus 2080 decodes the current encoding
  • the unit 400 or 450 may be divided to determine two encoding units 420a, 420b, or 470a and 470b included in the current encoding unit.
  • the image decoding apparatus 2080 decodes the non- The current encoding unit can be divided in consideration of the position of the long side of the current encoding unit (400 or 450). For example, the image decoding apparatus 2080 divides the current encoding unit 400 or 450 in the direction of dividing the longer side of the current encoding unit 400 or 450 in consideration of the type of the current encoding unit 400 or 450 So that a plurality of encoding units can be determined.
  • the image decoding apparatus 2080 when the information on the split mode mode indicates that an encoding unit is divided into an odd number of blocks (tri-split), the image decoding apparatus 2080 includes the current encoding unit 400 or 450 An odd number of encoding units can be determined. For example, when the information on the split mode mode indicates that the current encoding unit 400 or 450 is divided into three encoding units, the image decoding device 2080 encodes the current encoding unit 400 or 450 into three encoding Can be divided into units 430a, 430b, 430c, 480a, 480b, and 480c.
  • the ratio of the width and height of the current encoding unit 400 or 450 may be 4: 1 or 1: 4. If the ratio of width to height is 4: 1, the length of the width is longer than the length of the height, so the block type information may be horizontal. If the ratio of width to height is 1: 4, the block type information may be vertical because the length of the width is shorter than the length of the height.
  • the image decoding apparatus 2080 can determine to divide the current encoding unit into odd number blocks based on the information on the split mode mode. The image decoding apparatus 2080 can determine the division direction of the current encoding unit 400 or 450 based on the block type information of the current encoding unit 400 or 450.
  • the image decoding apparatus 2080 can determine the encoding units 430a, 430b, and 430c by dividing the current encoding unit 400 in the horizontal direction. If the current encoding unit 450 is in the horizontal direction, the image decoding apparatus 2080 can determine the encoding units 480a, 480b, and 480c by dividing the current encoding unit 450 in the vertical direction.
  • the image decoding apparatus 2080 may determine an odd number of encoding units included in the current encoding unit 400 or 450, and the sizes of the determined encoding units may not be the same. For example, the size of a predetermined encoding unit 430b or 480b among the determined odd number of encoding units 430a, 430b, 430c, 480a, 480b, and 480c is different from the size of the other encoding units 430a, 430c, 480a, and 480c .
  • an encoding unit that can be determined by dividing the current encoding unit (400 or 450) may have a plurality of types of sizes, and an odd number of encoding units (430a, 430b, 430c, 480a, 480b, 480c) May have different sizes.
  • the image decoding apparatus 2080 can determine an odd number of encoding units included in the current encoding unit 400 or 450 Further, the image decoding apparatus 2080 may set a predetermined restriction on at least one encoding unit among odd number of encoding units generated by division. Referring to FIG.
  • the image decoding apparatus 2080 decodes an encoding unit (not shown) located at the center among the three encoding units 430a, 430b, 430c, 480a, 480b, and 480c generated by dividing the current encoding unit 400 or 450,
  • the decoding process for the coding units 430b and 480b may be different from the coding units 430a, 430c, 480a, and 480c.
  • the image decoding apparatus 2080 may restrict the coding units 430b and 480b positioned at the center to be not further divided unlike the other coding units 430a, 430c, 480a, and 480c, It can be limited to be divided.
  • FIG. 13 illustrates a process in which an image decoding apparatus 2080 divides an encoding unit based on at least one of information on a block format information and a format mode according to an embodiment.
  • the image decoding apparatus 2080 decides not to divide or divide the first encoding unit 500 of the square shape into encoding units based on at least one of the information on the block type information and the information on the division mode mode . If the information on the split mode mode indicates that the first encoding unit 500 is divided in the horizontal direction according to the embodiment, the image decoding apparatus 2080 divides the first encoding unit 500 in the horizontal direction, 2 encoding unit 510, as shown in FIG.
  • the first encoding unit, the second encoding unit, and the third encoding unit used according to an embodiment are terms used to understand the relation before and after the division between encoding units.
  • the second encoding unit can be determined, and if the second encoding unit is divided, the third encoding unit can be determined.
  • the relationship between the first coding unit, the second coding unit and the third coding unit used can be understood to be in accordance with the above-mentioned characteristic.
  • the image decoding apparatus 2080 may determine that the determined second encoding unit 510 is not divided or divided into encoding units based on at least one of the block type information and the information on the split mode mode . 13, the image decoding apparatus 2080 decodes the first encoding unit 500 based on at least one of information on the block type information and the information on the split mode, 510) may be divided into at least one third encoding unit 520a, 520b, 520c, 520d, or the second encoding unit 510 may not be divided.
  • the image decoding apparatus 2080 may acquire at least one of information on the block type information and the division mode mode and the image decoding apparatus 2080 may acquire the block type information on the basis of at least one of the acquired block type information and the information on the division mode mode
  • the second encoding unit 510 may divide the first encoding unit 500 into a plurality of second encoding units of various types (for example, 510), and the second encoding unit 510 may divide the block type information and the information
  • the first encoding unit 500 may be divided according to a manner in which the first encoding unit 500 is divided.
  • the first encoding unit 500 is divided into the second encoding units 510 based on at least one of the block type information for the first encoding unit 500 and the information about the split mode mode 520b, 520c, and 520d (e.g., 520a, 520b, 520c, and 520d) based on at least one of the block type information on the second encoding unit 510 and the information on the split mode mode, Etc.). That is, the encoding unit may be recursively divided based on at least one of the information on the split mode mode and the block type information associated with each of the encoding units. Therefore, a square encoding unit may be determined in a non-square encoding unit, and a non-square encoding unit may be determined by dividing the square encoding unit recursively.
  • predetermined encoding units for example, An encoding unit or a square-shaped encoding unit
  • the square-shaped third coding unit 520b which is one of the odd-numbered third coding units 520b, 520c, and 520d, may be divided in the horizontal direction and divided into a plurality of fourth coding units.
  • the non-square fourth encoding unit 530b or 530d which is one of the plurality of fourth encoding units 530a, 530b, 530c, and 530d, may be further divided into a plurality of encoding units.
  • the fourth encoding unit 530b or 530d in the non-square form may be divided again into odd number of encoding units.
  • a method which can be used for recursive division of an encoding unit will be described later in various embodiments.
  • the image decoding apparatus 2080 divides each of the third encoding units 520a, 520b, 520c, and 520d into encoding units based on at least one of the block type information and the information on the split mode mode . Also, the image decoding apparatus 2080 may determine that the second encoding unit 510 is not divided based on at least one of the block type information and the information on the split mode mode. The image decoding apparatus 2080 may divide the non-square-shaped second coding unit 510 into odd-numbered third coding units 520b, 520c and 520d according to an embodiment.
  • the image decoding apparatus 2080 may set a predetermined limit on a predetermined third encoding unit among odd numbered third encoding units 520b, 520c, and 520d. For example, the image decoding apparatus 2080 may limit the encoding unit 520c located in the middle among the odd numbered third encoding units 520b, 520c, and 520d to no longer be divided, or be divided into a set number of times .
  • the image decoding apparatus 2080 includes an encoding unit (not shown) located in the middle among the odd third encoding units 520b, 520c, and 520d included in the second encoding unit 510 in the non- 520c may not be further divided or may be limited to being divided into a predetermined division form (for example, divided into four coding units only or divided into a form corresponding to a form in which the second coding units 510 are divided) (For example, dividing only n times, n > 0).
  • a predetermined division form for example, divided into four coding units only or divided into a form corresponding to a form in which the second coding units 510 are divided
  • the above restriction on the coding unit 520c positioned at the center is merely an example and should not be construed to be limited to the above embodiments and the coding unit 520c positioned at the center is not limited to the coding units 520b and 520d Quot;), < / RTI > which can be decoded differently.
  • the image decoding apparatus 2080 can acquire at least one of the block type information used for dividing the current encoding unit and the information about the split mode using a predetermined position in the current encoding unit.
  • FIG. 14 illustrates a method for an image decoding apparatus 2080 to determine a predetermined encoding unit among odd number of encoding units according to an embodiment.
  • At least one of the block type information of the current encoding units 600 and 650 and the information of the split mode mode is a sample of a predetermined position among a plurality of samples included in the current encoding units 600 and 650 For example, samples 640 and 690 positioned in the middle).
  • the predetermined position in the current coding unit 600 in which at least one of the block type information and the division mode information can be obtained should not be limited to the middle position shown in Fig. 14, It should be understood that various positions (e.g., top, bottom, left, right, top left, bottom left, top right, or bottom right, etc.) that may be included in unit 600 may be included.
  • the video decoding apparatus 2080 may determine that the current encoding unit is divided or not divided into encoding units of various types and sizes by acquiring at least one of the block type information and the division type mode information obtained from the predetermined position .
  • the image decoding apparatus 2080 can select one of the encoding units.
  • the method for selecting one of the plurality of encoding units may be various, and description of these methods will be described later in various embodiments.
  • the image decoding apparatus 2080 may divide the current encoding unit into a plurality of encoding units and determine a predetermined encoding unit.
  • the image decoding apparatus 2080 may use information indicating the positions of odd-numbered coding units in order to determine a coding unit located in the middle among odd-numbered coding units. 14, the image decoding apparatus 2080 divides the current encoding unit 600 or the current encoding unit 650 into odd number of encoding units 620a, 620b, and 620c or odd number of encoding units 660a, 660b, and 660c.
  • the image decoding apparatus 2080 decodes the encoded data in the middle encoding unit 620b or the middle encoding unit 620b using information on the positions of odd number of encoding units 620a, 620b, and 620c or odd number of encoding units 660a, 660b, and 660c. (660b). For example, the image decoding apparatus 2080 determines the positions of the encoding units 620a, 620b, and 620c based on information indicating the positions of predetermined samples included in the encoding units 620a, 620b, and 620c, The encoding unit 620b located in the encoding unit 620b can be determined.
  • the image decoding apparatus 2080 decodes the encoding units 620a, 620b, and 620c based on information indicating the positions of the upper left samples 630a, 630b, and 630c of the encoding units 620a, 620b, and 620c,
  • the encoding unit 620b located in the center can be determined.
  • Information indicating the positions of the upper left samples 630a, 630b, and 630c included in the coding units 620a, 620b, and 620c according to one embodiment is stored in the pictures of the coding units 620a, 620b, and 620c Or information about the position or coordinates of the object.
  • Information indicating the positions of the upper left samples 630a, 630b, and 630c included in the coding units 620a, 620b, and 620c according to one embodiment is stored in the coding units 620a , 620b, and 620c, and the width or height may correspond to information indicating the difference between the coordinates of the encoding units 620a, 620b, and 620c in the picture.
  • the image decoding apparatus 2080 directly uses the information on the position or the coordinates in the pictures of the coding units 620a, 620b, and 620c or the information on the width or height of the coding units corresponding to the difference between the coordinates
  • the encoding unit 620b located in the center can be determined.
  • the information indicating the position of the upper left sample 630a of the upper coding unit 620a may indicate the coordinates (xa, ya) and the upper left sample 530b of the middle coding unit 620b May indicate the coordinates (xb, yb), and the information indicating the position of the upper left sample 630c of the lower coding unit 620c may indicate the coordinates (xc, yc).
  • the image decoding apparatus 2080 can determine the center encoding unit 620b using the coordinates of the upper left samples 630a, 630b, and 630c included in the encoding units 620a, 620b, and 620c.
  • the coding unit 620b including (xb, yb) coordinates of the sample 630b located at the center, Can be determined as a coding unit located in the middle of the coding units 620a, 620b, and 620c determined by dividing the current coding unit 600.
  • the coordinates indicating the positions of the samples 630a, 630b and 630c in the upper left corner may indicate the coordinates indicating the absolute position in the picture
  • the position of the upper left sample 630a of the upper coding unit 620a may be (Dxb, dyb), which is information indicating the relative position of the sample 630b at the upper left of the middle encoding unit 620b, and the relative position of the sample 630c at the upper left of the lower encoding unit 620c
  • Information dyn (dxc, dyc) coordinates may also be used.
  • the method of determining the coding unit at a predetermined position by using the coordinates of the sample as information indicating the position of the sample included in the coding unit should not be limited to the above-described method, and various arithmetic Should be interpreted as a method.
  • the image decoding apparatus 2080 may divide the current encoding unit 600 into a plurality of encoding units 620a, 620b, and 620c, and may encode a predetermined one of the encoding units 620a, 620b, and 620c
  • the encoding unit can be selected according to the criterion. For example, the image decoding apparatus 2080 can select an encoding unit 620b having a different size from among the encoding units 620a, 620b, and 620c.
  • the image decoding apparatus 2080 includes a (xa, ya) coordinate which is information indicating the position of the upper left sample 630a of the upper encoding unit 620a, (Xc, yc) coordinates, which is information indicating the position of the lower-stage coding unit 630b and the position of the upper-left sample 630c of the lower-stage coding unit 620c, , 620b, and 620c, respectively.
  • the image decoding apparatus 2080 decodes the encoding units 620a and 620b using the coordinates (xa, ya), (xb, yb), (xc, yc) indicating the positions of the encoding units 620a, 620b and 620c , And 620c, respectively.
  • the image decoding apparatus 2080 can determine the width of the upper encoding unit 620a as the width of the current encoding unit 600.
  • the image decoding apparatus 2080 can determine the height of the upper encoding unit 620a as yb-ya.
  • the image decoding apparatus 2080 may determine the width of the middle encoding unit 620b as the width of the current encoding unit 600 according to an embodiment.
  • the image decoding apparatus 2080 can determine the height of the center encoding unit 620b as yc-yb.
  • the image decoding apparatus 2080 can determine the width or height of the lower coding unit by using the width or height of the current coding unit and the width and height of the upper coding unit 620a and the middle coding unit 620b .
  • the image decoding apparatus 2080 can determine an encoding unit having a different size from the other encoding units based on the widths and heights of the determined encoding units 620a, 620b, and 620c. Referring to FIG.
  • the image decoding apparatus 2080 can determine a coding unit 620b as a coding unit at a predetermined position while having a size different from that of the upper coding unit 620a and the lower coding unit 620c.
  • the process of determining the encoding unit having a size different from that of the other encoding units by the image decoding apparatus 2080 may be the same as that of the first embodiment in which the encoding unit of a predetermined position is determined using the size of the encoding unit determined based on the sample coordinates , Various processes may be used for determining the encoding unit at a predetermined position by comparing the sizes of the encoding units determined according to predetermined sample coordinates.
  • the image decoding apparatus 2080 decodes the position of the left upper sample 670a of the left coding unit 660a and the position (xd, yd) of the left upper sample of the left coding unit 660b of the middle coding unit 660b, 660b and 660c using the (xf, yf) coordinates, which is information indicating the (xe, ye) coordinate which is the information indicating the position of the right encoding unit 660c and the position of the sample 670c at the upper left of the right encoding unit 660c, Each width or height can be determined.
  • the image decoding apparatus 2080 decodes the encoding units 660a and 660b using the coordinates (xd, yd), (xe, ye), (xf, yf) indicating the positions of the encoding units 660a, 660b and 660c And 660c, respectively.
  • the image decoding apparatus 2080 can determine the width of the left encoding unit 660a as xe-xd.
  • the image decoding apparatus 2080 can determine the height of the left encoding unit 660a as the height of the current encoding unit 650.
  • the image decoding apparatus 2080 can determine the width of the middle encoding unit 660b as xf-xe.
  • the image decoding apparatus 2080 can determine the height of the middle encoding unit 660b as the height of the current encoding unit 600.
  • the width or height of the right encoding unit 660c is determined by the width or height of the current encoding unit 650 and the width and height of the left encoding unit 660a and the middle encoding unit 660b . ≪ / RTI >
  • the image decoding apparatus 2080 can determine an encoding unit having a different size from the other encoding units based on the widths and heights of the determined encoding units 660a, 660b, and 660c. Referring to FIG. 14, the image decoding apparatus 2080 can determine a coding unit 660b as a coding unit at a predetermined position while having a size different from that of the left coding unit 660a and the right coding unit 660c.
  • the process of determining the encoding unit having a size different from that of the other encoding units by the image decoding apparatus 2080 may be the same as that of the first embodiment in which the encoding unit of a predetermined position is determined using the size of the encoding unit determined based on the sample coordinates , Various processes may be used for determining the encoding unit at a predetermined position by comparing the sizes of the encoding units determined according to predetermined sample coordinates.
  • the position of the sample to be considered for determining the position of the coding unit should not be interpreted as being limited to the left upper end, and information about the position of any sample included in the coding unit can be interpreted as being available.
  • the image decoding apparatus 2080 can select a coding unit at a predetermined position among the odd number of coding units determined by dividing the current coding unit considering the type of the current coding unit. For example, if the current coding unit is a non-square shape having a width greater than the height, the image decoding apparatus 2080 can determine a coding unit at a predetermined position along the horizontal direction. That is, the image decoding apparatus 2080 may determine one of the encoding units which are located in the horizontal direction and limit the encoding unit. If the current coding unit is a non-square shape having a height greater than the width, the image decoding apparatus 2080 can determine a coding unit at a predetermined position in the vertical direction. That is, the image decoding apparatus 2080 may determine one of the encoding units which are located in the vertical direction and limit the encoding unit.
  • the image decoding apparatus 2080 may use information indicating positions of even-numbered encoding units in order to determine an encoding unit of a predetermined position among the even-numbered encoding units.
  • the image decoding apparatus 2080 can determine an even number of coding units by binary coding the current coding unit and determine a coding unit at a predetermined position using information on the positions of even number of coding units. A concrete procedure for this is omitted because it may be a process corresponding to a process of determining a coding unit of a predetermined position (for example, the middle position) among the above-mentioned odd number of coding units in FIG.
  • a non-square current encoding unit is divided into a plurality of encoding units
  • in order to determine an encoding unit at a predetermined position among a plurality of encoding units Can be used.
  • the image decoding apparatus 2080 decodes the block type information stored in the sample included in the middle encoding unit, Mode may be used.
  • the image decoding apparatus 2080 divides the current encoding unit 600 into a plurality of encoding units 620a, 620b, and 620c based on at least one of information on the block type information and the information on the split mode mode And the encoding unit 620b located in the middle of the plurality of encoding units 620a, 620b, and 620c can be determined. Furthermore, the image decoding apparatus 2080 can determine a coding unit 620b positioned at the center in consideration of the position at which at least one of the block type information and the division type mode information is obtained.
  • At least one of the block type information of the current encoding unit 600 and the information of the division mode mode can be obtained in the sample 640 located in the middle of the current encoding unit 600, If the current encoding unit 600 is divided into a plurality of encoding units 620a, 620b, and 620c based on at least one of the information on the division mode mode and the encoding unit 620b including the sample 640, As shown in FIG.
  • the information used for determining the encoding unit located in the middle should not be limited to at least one of the block type information and the information about the division mode mode, and a process of determining an encoding unit in which various types of information are located in the middle ≪ / RTI >
  • predetermined information for identifying a coding unit at a predetermined position may be obtained from a predetermined sample included in a coding unit to be determined.
  • the image decoding apparatus 2080 encodes a predetermined number of coding units (for example, a plurality of divided coding units) among a plurality of coding units 620a, 620b, and 620c determined by dividing the current coding unit 600 (For example, a sample located in the middle of the current encoding unit 600) at a predetermined position in the current encoding unit 600 in order to determine an encoding unit located in the middle of the encoding unit, And at least one of information on the split mode mode may be used.
  • the image decoding apparatus 2080 can determine the sample of the predetermined position in consideration of the block form of the current encoding unit 600, and the image decoding apparatus 2080 decodes the plural A coding unit 620b including samples from which predetermined information (for example, at least one of information on the block type information and the division mode information) can be obtained from the plurality of coding units 620a, 620b, and 620c So that a predetermined limit can be set.
  • predetermined information for example, at least one of information on the block type information and the division mode information
  • the image decoding apparatus 2080 can determine a sample 640 positioned in the center of the current encoding unit 600 as a sample from which predetermined information can be obtained,
  • the coding unit 2080 may limit the coding unit 620b including the sample 640 to a predetermined limit in the decoding process.
  • the position of the sample from which the predetermined information can be obtained should not be construed to be limited to the above-mentioned position, but may be interpreted as samples at arbitrary positions included in the encoding unit 620b to be determined for limiting.
  • the position of a sample from which predetermined information can be obtained may be determined according to the type of the current encoding unit 600.
  • the block type information can determine whether the current encoding unit is a square or a non-square, and determine the position of a sample from which predetermined information can be obtained according to the shape.
  • the video decoding apparatus 2080 may use at least one of the information on the width of the current coding unit and the information on the height to position at least one of the width and the height of the current coding unit in half The sample can be determined as a sample from which predetermined information can be obtained.
  • the image decoding apparatus 2080 decodes one of the samples adjacent to the boundary dividing the longer side of the current encoding unit into half, Can be determined as a sample from which the information of < / RTI >
  • the image decoding apparatus 2080 decodes the block type information and the split mode mode to determine a coding unit of a predetermined position among the plurality of encoding units.
  • Information can be used.
  • the image decoding apparatus 2080 can acquire at least one of the block type information and the information on the split mode mode from a sample of a predetermined position included in the encoding unit, and the image decoding apparatus 2080
  • the plurality of coding units generated by dividing the unit may be divided using at least one of the information on the division mode and the block type information obtained from the sample at the predetermined position included in each of the plurality of the coding units.
  • the coding unit can be recursively divided using at least one of the block type information obtained in the sample at the predetermined position included in each of the coding units and the information about the division mode.
  • the recursive division process of the encoding unit is described in detail with reference to FIG. 13, and a detailed description thereof will be omitted.
  • the image decoding apparatus 2080 can determine at least one encoding unit by dividing the current encoding unit, and the order in which the at least one encoding unit is decoded is determined as a predetermined block (for example, ). ≪ / RTI >
  • FIG. 15 shows a sequence in which a plurality of coding units are processed when the image decoding apparatus 2080 determines a plurality of coding units by dividing a current coding unit according to an embodiment.
  • the image decoding apparatus 2080 divides the first encoding unit 700 in the vertical direction according to the information on the block type information and the division mode, thereby determining the second encoding units 710a and 710b,
  • the second encoding units 730a and 730b may be determined by dividing the first encoding units 700a to 750c by dividing the first encoding units 700a and 750b in the horizontal direction to divide the first encoding units 700 in the vertical direction and the horizontal direction, , 750d can be determined.
  • the image decoding apparatus 2080 can determine the order in which the second encoding units 710a and 710b determined by dividing the first encoding unit 700 in the vertical direction are processed in the horizontal direction 710c .
  • the image decoding apparatus 2080 can determine the processing order of the second encoding units 730a and 730b determined by dividing the first encoding unit 700 in the horizontal direction as the vertical direction 730c.
  • the image decoding apparatus 2080 processes the encoding units located in one row of the second encoding units 750a, 750b, 750c, and 750d determined by dividing the first encoding unit 700 in the vertical direction and the horizontal direction, (For example, a raster scan order or a z scan order 750e) in which the encoding units located in the next row are processed.
  • the image decoding apparatus 2080 can recursively divide encoding units. 15, the image decoding apparatus 2080 can determine a plurality of encoding units 710a, 710b, 730a, 730b, 750a, 750b, 750c and 750d by dividing the first encoding unit 700, The determined plurality of encoding units 710a, 710b, 730a, 730b, 750a, 750b, 750c, and 750d can be recursively divided.
  • the method of dividing the plurality of encoding units 710a, 710b, 730a, 730b, 750a, 750b, 750c, and 750d may be a method corresponding to the method of dividing the first encoding unit 700.
  • the plurality of encoding units 710a, 710b, 730a, 730b, 750a, 750b, 750c, and 750d may be independently divided into a plurality of encoding units.
  • the image decoding apparatus 2080 can determine the second encoding units 710a and 710b by dividing the first encoding unit 700 in the vertical direction, and can further determine the second encoding units 710a and 710b Can be determined not to divide or separate independently.
  • the image decoding apparatus 2080 may divide the second encoding unit 710a on the left side in the horizontal direction into the third encoding units 720a and 720b and the second encoding units 710b ) May not be divided.
  • the processing order of the encoding units may be determined based on the division process of the encoding units.
  • the processing order of the divided coding units can be determined based on the processing order of the coding units immediately before being divided.
  • the image decoding apparatus 2080 can independently determine the order in which the third encoding units 720a and 720b determined by dividing the second encoding unit 710a on the left side are processed to the second encoding unit 710b on the right side.
  • the third encoding units 720a and 720b may be processed in the vertical direction 720c because the second encoding units 710a on the left side are divided in the horizontal direction and the third encoding units 720a and 720b are determined.
  • the order in which the left second encoding unit 710a and the right second encoding unit 710b are processed corresponds to the horizontal direction 710c
  • the right encoding unit 710b can be processed after the blocks 720a and 720b are processed in the vertical direction 720c.
  • the above description is intended to explain the process sequence in which encoding units are determined according to the encoding units before division. Therefore, it should not be construed to be limited to the above-described embodiments, It should be construed as being used in various ways that can be handled independently in sequence.
  • 16 illustrates a process of determining that the current encoding unit is divided into odd number of encoding units when the image decoding apparatus 2080 can not process the encoding units in a predetermined order according to an embodiment.
  • the image decoding apparatus 2080 can determine that the current encoding unit is divided into odd number of encoding units based on the obtained block type information and information on the split mode mode.
  • the first encoding unit 800 in the form of a square may be divided into second non-square encoding units 810a and 810b, and the second encoding units 810a and 810b may be independently 3 encoding units 820a, 820b, 820c, 820d, and 820e.
  • the image decoding apparatus 2080 can determine the plurality of third encoding units 820a and 820b by dividing the left encoding unit 810a of the second encoding unit in the horizontal direction, and the right encoding unit 810b Can be divided into an odd number of third encoding units 820c, 820d, and 820e.
  • the image decoding apparatus 2080 determines whether or not the third encoding units 820a, 820b, 820c, 820d, and 820e can be processed in a predetermined order and determines whether there are odd-numbered encoding units You can decide. Referring to FIG. 16, the image decoding apparatus 2080 can recursively divide the first encoding unit 800 to determine the third encoding units 820a, 820b, 820c, 820d, and 820e.
  • the image decoding apparatus 2080 may encode the first encoding unit 800, the second encoding units 810a and 810b or the third encoding units 820a and 820b based on at least one of the block type information and the information on the split mode mode , 820c, 820d, and 820e are divided into odd number of coding units among the divided types. For example, an encoding unit located on the right of the second encoding units 810a and 810b may be divided into odd third encoding units 820c, 820d, and 820e.
  • the order in which the plurality of coding units included in the first coding unit 800 are processed may be a predetermined order (for example, a z-scan order 830) 2080 can determine whether the third encoding units 820c, 820d, and 820e determined by dividing the right second encoding unit 810b into odd numbers satisfy the condition that the third encoding units 820c, 820d, and 820e can be processed according to the predetermined order.
  • a predetermined order for example, a z-scan order 830
  • the image decoding apparatus 2080 satisfies the condition that the third encoding units 820a, 820b, 820c, 820d, and 820e included in the first encoding unit 800 can be processed in a predetermined order And it is determined whether or not at least one of the widths and heights of the second encoding units 810a and 810b is divided in half according to the boundaries of the third encoding units 820a, 820b, 820c, 820d, and 820e, .
  • the third encoding units 820a and 820b which are determined by dividing the height of the left second encoding unit 810a in the non-square shape by half, can satisfy the condition.
  • the boundaries of the third encoding units 820c, 820d, and 820e determined by dividing the right second encoding unit 810b into three encoding units do not divide the width or height of the right second encoding unit 810b in half ,
  • the third encoding units 820c, 820d, and 820e may be determined as not satisfying the condition.
  • the image decoding apparatus 2080 may determine that the scan sequence is disconnection in the case of such unsatisfactory condition and determine that the right second encoding unit 810b is divided into odd number of encoding units based on the determination result.
  • the image decoding apparatus 2080 may limit a coding unit at a predetermined position among the divided coding units when the coding unit is divided into odd number of coding units. Since the embodiment has been described above, a detailed description thereof will be omitted.
  • FIG. 17 illustrates a process in which the image decoding apparatus 2080 determines at least one encoding unit by dividing a first encoding unit 900 according to an embodiment.
  • the image decoding apparatus 2080 can divide the first encoding unit 900 based on at least one of the block type information obtained through the entropy decoding unit 2055 and the information on the split mode mode .
  • the first coding unit 900 in the form of a square may be divided into four coding units having a square form, or may be divided into a plurality of non-square coding units. For example, referring to FIG. 17, if the block type information indicates that the first encoding unit 900 is square, and the information on the split mode mode is divided into non-square encoding units, the image decoding apparatus 2080 1 encoding unit 900 into a plurality of non-square encoding units.
  • the image decoding apparatus 2080 decodes the first encoding
  • the unit 900 can be divided into the second coding units 910a, 910b and 910c divided in the vertical direction as the odd number of coding units or the second coding units 920a, 920b and 920c divided in the horizontal direction .
  • the image decoding apparatus 2080 may be configured such that the second encoding units 910a, 910b, 910c, 920a, 920b, and 920c included in the first encoding unit 900 are processed in a predetermined order And the condition is that at least one of the width and the height of the first encoding unit 900 is divided in half according to the boundaries of the second encoding units 910a, 910b, 910c, 920a, 920b, and 920c .
  • the boundaries of the second encoding units 910a, 910b, and 910c which are determined by dividing the first encoding unit 900 in the vertical direction into a square shape, are divided in half by the width of the first encoding unit 900
  • the first encoding unit 900 can be determined as not satisfying a condition that can be processed in a predetermined order.
  • the boundaries of the second encoding units 920a, 920b, and 920c which are determined by dividing the first encoding unit 900 in the horizontal direction into the horizontal direction, can not divide the width of the first encoding unit 900 in half, 1 encoding unit 900 may be determined as not satisfying a condition that can be processed in a predetermined order.
  • the image decoding apparatus 2080 may determine that the scan sequence is disconnection in the case of such unsatisfactory condition and determine that the first encoding unit 900 is divided into odd number of encoding units based on the determination result. According to an exemplary embodiment, the image decoding apparatus 2080 may limit a coding unit at a predetermined position among the divided coding units when the coding unit is divided into odd number of coding units. Since the embodiment has been described above, a detailed description thereof will be omitted.
  • the image decoding apparatus 2080 may divide the first encoding unit to determine various types of encoding units.
  • the image decoding apparatus 2080 may divide a first coding unit 900 in a square form and a first coding unit 930 or 950 in a non-square form into various types of coding units .
  • the image decoding apparatus 2080 decodes the first encoding unit 1000 in the form of a square based on at least one of the block type information and the division type mode information acquired through the entropy decoding unit 2055 - the second encoding units 1010a, 1010b, 1020a, and 1020b in the form of a square.
  • the second encoding units 1010a, 1010b, 1020a, and 1020b may be independently divided.
  • the image decoding apparatus 2080 divides or divides the image data into a plurality of encoding units based on at least one of the block type information and the information on the split mode mode associated with each of the second encoding units 1010a, 1010b, 1020a, and 1020b You can decide not to. According to one embodiment, the image decoding apparatus 2080 divides the non-square left second encoding unit 1010a determined by dividing the first encoding unit 1000 in the vertical direction into a horizontal direction, 1012a, and 1012b.
  • the right second encoding unit 1010b is horizontally aligned with the left second encoding unit 1010a in a direction in which the left second encoding unit 1010a is divided, As shown in Fig. If the right second encoding unit 1010b is divided in the same direction and the third encoding units 1014a and 1014b are determined, the left second encoding unit 1010a and the right second encoding unit 1010b are arranged in the horizontal direction
  • the third encoding units 1012a, 1012b, 1014a, and 1014b can be determined by being independently divided.
  • the image decoding apparatus 2080 may convert the first encoding unit 1000 into four square-shaped second encoding units 1030a, 1030b, 1030c, and 1030d based on at least one of the block type information and the information on the split mode mode ), which may be inefficient in terms of image decoding.
  • the image decoding apparatus 2080 divides the non-square second encoding unit 1020a or 1020b determined by dividing the first encoding unit 1000 in the horizontal direction into a vertical direction, (1022a, 1022b, 1024a, 1024b).
  • the video decoding apparatus 2080 decodes the second coding unit (for example, The encoding unit 1020b) can be restricted such that the upper second encoding unit 1020a can not be divided vertically in the same direction as the divided direction.
  • FIG. 19 illustrates a process in which the image decoding apparatus 2080 divides a square-shaped encoding unit when the information on the split mode mode can not indicate division into four square-shaped encoding units according to an embodiment .
  • the image decoding apparatus 2080 divides the first encoding unit 1100 based on at least one of information on the block type information and the information on the split mode mode to generate second encoding units 1110a, 1110b, 1120a, and 1120b Etc.) can be determined.
  • the information on the division type mode may include information on various types in which the coding unit can be divided, but information on various types may not include information for dividing into four square units of coding units.
  • the image decoding apparatus 2080 can not divide the first encoding unit 1100 in the square form into the second encoding units 1130a, 1130b, 1130c, and 1130d in the form of four squares .
  • the image decoding apparatus 2080 can determine the second encoding units 1110a, 1110b, 1120a, and 1120b in the non-square form based on the information on the split mode mode.
  • the image decoding apparatus 2080 can independently divide the non-square second encoding units 1110a, 1110b, 1120a, and 1120b, respectively.
  • Each of the second encoding units 1110a, 1110b, 1120a, and 1120b may be divided in a predetermined order through a recursive method, and the first encoding units 1110a, 1110b, 1120a, and 1120b may be divided according to at least one of the block type information, May be a division method corresponding to how the unit 1100 is divided.
  • the image decoding apparatus 2080 can determine the third encoding units 1112a and 1112b in the form of a square by dividing the left second encoding unit 1110a in the horizontal direction, and the second right encoding unit 1110b It is possible to determine the third encoding units 1114a and 1114b in the form of a square by being divided in the horizontal direction. Furthermore, the image decoding apparatus 2080 may divide the left second encoding unit 1110a and the right second encoding unit 1110b in the horizontal direction to determine the third encoding units 1116a, 1116b, 1116c, and 1116d in the form of a square have. In this case, the encoding unit can be determined in the same manner as the first encoding unit 1100 is divided into the four second square encoding units 1130a, 1130b, 1130c, and 1130d.
  • the image decoding apparatus 2080 can determine the third encoding units 1122a and 1122b in the shape of a square by dividing the upper second encoding unit 1120a in the vertical direction, and the lower second encoding units 1120b May be divided in the vertical direction to determine the third encoding units 1124a and 1124b in the form of a square. Further, the image decoding apparatus 2080 may divide the upper second encoding unit 1120a and the lower second encoding unit 1120b in the vertical direction to determine the square-shaped third encoding units 1126a, 1126b, 1126a, and 1126b have. In this case, the encoding unit can be determined in the same manner as the first encoding unit 1100 is divided into the four second square encoding units 1130a, 1130b, 1130c, and 1130d.
  • FIG. 20 illustrates that a processing order among a plurality of coding units may be changed according to a division process of coding units according to an embodiment.
  • the image decoding apparatus 2080 may divide the first encoding unit 1200 based on information on the block type information and the split mode mode.
  • the image decoding apparatus 2080 decodes the first encoding
  • the unit 1200 can be divided to determine the second encoding unit (e.g., 1210a, 1210b, 1220a, 1220b, etc.). Referring to FIG.
  • the non-square second encoding units 1210a, 1210b, 1220a, and 1220b which are determined by dividing the first encoding unit 1200 only in the horizontal direction or the vertical direction, Can be independently divided based on information on the < / RTI >
  • the image decoding apparatus 2080 divides the second encoding units 1210a and 1210b generated by dividing the first encoding unit 1200 in the vertical direction into the horizontal direction, and outputs the third encoding units 1216a, 1216b, 1216c and 1216d can be determined and the second encoding units 1220a and 1220b generated by dividing the first encoding unit 1200 in the horizontal direction are divided in the horizontal direction and the third encoding units 1226a, , 1226d. Since the process of dividing the second encoding units 1210a, 1210b, 1220a, and 1220b has been described above with reference to FIG. 19, a detailed description thereof will be omitted.
  • the image decoding apparatus 2080 can process a coding unit in a predetermined order. Features of processing of a coding unit in a predetermined order have been described above with reference to FIG. 15, and a detailed description thereof will be omitted. Referring to FIG. 20, the image decoding apparatus 2080 divides a first encoding unit 1200 of a square shape into 4 pieces of fourth encoding units 1216a, 1216b, 1216c, 1216d, 1226a, 1226b, 1226c, 1226d Can be determined.
  • the image decoding apparatus 2080 decodes the processing order of the third encoding units 1216a, 1216b, 1216c, 1216d, 1226a, 1226b, 1226c, and 1226d according to the form in which the first encoding unit 1200 is divided You can decide.
  • the image decoding apparatus 2080 divides the second encoding units 1210a and 1210b generated in the vertical direction into horizontal units to determine the third encoding units 1216a, 1216b, 1216c, and 1216d And the image decoding apparatus 2080 processes the third encoding units 1216a and 1216c included in the left second encoding unit 1210a first in the vertical direction and then processes the third encoding units 1216a and 1216c included in the right second encoding unit 1210b The third encoding units 1216a, 1216b, 1216c, and 1216d can be processed according to the order 1217 of processing the third encoding units 1216b and 1216d in the vertical direction.
  • the image decoding apparatus 2080 divides the second encoding units 1220a and 1220b generated in the horizontal direction into vertical directions to determine the third encoding units 1226a, 1226b, 1226c and 1226d And the image decoding apparatus 2080 first processes the third encoding units 1226a and 1226b included in the upper second encoding unit 1220a in the horizontal direction and then processes the third encoding units 1226a and 1226b included in the lower second encoding unit 1220b The third encoding units 1226a, 1226b, 1226c, and 1226d may be processed in accordance with an order 1227 for processing the third encoding units 1226c and 1226d in the horizontal direction.
  • the second encoding units 1210a, 1210b, 1220a, and 1220b are divided to determine the third encoding units 1216a, 1216b, 1216c, 1216d, 1226a, 1226b, 1226c, and 1226d, have.
  • the second encoding units 1210a and 1210b determined to be divided in the vertical direction and the second encoding units 1220a and 1220b determined to be divided in the horizontal direction are divided into different formats, but the third encoding units 1216a , 1216b, 1216c, 1216d, 1226a, 1226b, 1226c, and 1226d, the result is that the first encoding unit 1200 is divided into the same type of encoding units.
  • the image decoding apparatus 2080 divides the encoding unit recursively through different processes based on at least one of the block type information and the information about the division mode, thereby determining the same type of encoding units as a result, A plurality of encoding units may be processed in different orders.
  • FIG. 21 illustrates a process in which the depth of an encoding unit is determined according to a change in type and size of an encoding unit when a plurality of encoding units are determined by recursively dividing an encoding unit according to an exemplary embodiment.
  • the image decoding apparatus 2080 can determine the depth of a coding unit according to a predetermined criterion.
  • a predetermined criterion may be a length of a long side of a coding unit.
  • the depth of the current coding unit is smaller than the depth of the coding unit before being divided it can be determined that the depth is increased by n.
  • an encoding unit with an increased depth is expressed as a lower-depth encoding unit.
  • the image decoding apparatus 2080 may include a square- 1 encoding unit 1300 can be divided to determine the second encoding unit 1302, the third encoding unit 1304, etc. of the lower depth. If the size of the first encoding unit 1300 in the form of a square is 2Nx2N, the second encoding unit 1302 determined by dividing the width and height of the first encoding unit 1300 by 1/2 may have a size of NxN have.
  • the third encoding unit 1304 determined by dividing the width and height of the second encoding unit 1302 by a half size may have a size of N / 2xN / 2.
  • the width and height of the third encoding unit 1304 correspond to 1/4 of the first encoding unit 1300. If the depth of the first encoding unit 1300 is D, the depth of the second encoding unit 1302, which is half the width and height of the first encoding unit 1300, may be D + 1, The depth of the third encoding unit 1304, which is one fourth of the width and height of the third encoding unit 1300, may be D + 2.
  • block type information indicating a non-square shape for example, block type information is' 1: NS_VER 'indicating that the height is a non-square having a width greater than the width or' 2 >: NS_HOR '
  • the image decoding apparatus 2080 divides the non-square first coding unit 1310 or 1320 into a second lower coding unit 1312 or 1322, The third encoding unit 1314 or 1324, or the like.
  • the image decoding apparatus 2080 can determine a second coding unit (e.g., 1302, 1312, 1322, etc.) by dividing at least one of the width and the height of the first coding unit 1310 of Nx2N size. That is, the image decoding apparatus 2080 can determine the second encoding unit 1302 of NxN size or the second encoding unit 1322 of NxN / 2 size by dividing the first encoding unit 1310 in the horizontal direction, It is also possible to determine the second encoding unit 1312 of N / 2xN size by dividing it in the horizontal direction and the vertical direction.
  • a second coding unit e.g., 1302, 1312, 1322, etc.
  • the image decoding apparatus 2080 divides at least one of the width and the height of the 2NxN first encoding unit 1320 to determine a second encoding unit (for example, 1302, 1312, 1322, etc.) It is possible. That is, the image decoding apparatus 2080 can determine the second encoding unit 1302 of NxN size or the second encoding unit 1312 of N / 2xN size by dividing the first encoding unit 1320 in the vertical direction, The second encoding unit 1322 of the NxN / 2 size may be determined by dividing the image data in the horizontal direction and the vertical direction.
  • a second encoding unit for example, 1302, 1312, 1322, etc.
  • the image decoding apparatus 2080 divides at least one of the width and the height of the second encoding unit 1302 of NxN size to determine a third encoding unit (for example, 1304, 1314, 1324, etc.) It is possible. That is, the image decoding apparatus 2080 divides the second encoding unit 1302 in the vertical direction and the horizontal direction to determine a third encoding unit 1304 of N / 2xN / 2 size or a N / 4xN / 3 encoding unit 1314 or a third encoding unit 1324 of N / 2xN / 4 size.
  • a third encoding unit for example, 1304, 1314, 1324, etc.
  • the image decoding apparatus 2080 divides at least one of the width and the height of the second encoding unit 1312 of N / 2xN size into a third encoding unit (e.g., 1304, 1314, 1324, etc.) . That is, the image decoding apparatus 2080 divides the second encoding unit 1312 in the horizontal direction to generate a third encoding unit 1304 of N / 2xN / 2 or a third encoding unit 1324 of N / 2xN / 4 size ) Or may be divided in the vertical and horizontal directions to determine the third encoding unit 1314 of N / 4xN / 2 size.
  • a third encoding unit e.g. 1304, 1314, 1324, etc.
  • the image decoding apparatus 2080 may divide at least one of the width and the height of the second encoding unit 1322 of NxN / 2 size into a third encoding unit (e.g., 1304, 1314, 1324, etc.) . That is, the image decoding apparatus 2080 divides the second encoding unit 1322 in the vertical direction to obtain a third encoding unit 1304 of N / 2xN / 2 or a third encoding unit 1314 of N / 4xN / 2 size ) Or may be divided in the vertical and horizontal directions to determine the third encoding unit 1324 of N / 2xN / 4 size.
  • a third encoding unit e.g. 1304, 1314, 1324, etc.
  • the image decoding apparatus 2080 may divide a square-shaped encoding unit (for example, 1300, 1302, and 1304) into a horizontal direction or a vertical direction.
  • a square-shaped encoding unit for example, 1300, 1302, and 1304
  • the first encoding unit 1300 having a size of 2Nx2N is divided in the vertical direction to determine a first encoding unit 1310 having a size of Nx2N or the first encoding unit 1310 having a size of 2NxN to determine a first encoding unit 1320 having a size of 2NxN .
  • the depth of the encoding unit when the depth is determined based on the length of the longest side of the encoding unit, the depth of the encoding unit, which is determined by dividing the first encoding unit 1300 of 2Nx2N size in the horizontal direction or the vertical direction, May be the same as the depth of the unit (1300).
  • the width and height of the third encoding unit 1314 or 1324 may correspond to one fourth of the first encoding unit 1310 or 1320.
  • the depth of the first coding unit 1310 or 1320 is D
  • the depth of the second coding unit 1312 or 1322 which is half the width and height of the first coding unit 1310 or 1320 is D +
  • the depth of the third encoding unit 1314 or 1324, which is one fourth of the width and height of the first encoding unit 1310 or 1320 may be D + 2.
  • FIG. 22 illustrates a depth index (hereinafter referred to as a PID) for classifying a depth and a coding unit that can be determined according to the type and size of coding units according to an exemplary embodiment.
  • a PID depth index
  • the image decoding apparatus 2080 may divide the first encoding unit 1400 in a square form to determine various types of second encoding units. 22, the image decoding apparatus 2080 divides the first encoding unit 1400 into at least one of the vertical direction and the horizontal direction according to the information on the division mode mode, and outputs the second encoding units 1402a and 1402b , 1404a, 1404b, 1406a, 1406b, 1406c, 1406d. That is, the image decoding apparatus 2080 determines the second encoding units 1402a, 1402b, 1404a, 1404b, 1406a, 1406b, 1406c, and 1406d based on the information on the split mode mode for the first encoding unit 1400 .
  • the second encoding units 1402a, 1402b, 1404a, 1404b, 1406a, 1406b, 1406c, and 1406d which are determined according to the information on the split mode mode for the first encoding unit 1400 in the square form.
  • the depth can be determined based on the length of the sides. For example, since the length of one side of the first encoding unit 1400 in the square form is the same as the length of long sides of the second encoding units 1402a, 1402b, 1404a, and 1404b in the non-square form, 1400) and the non-square type second encoding units 1402a, 1402b, 1404a, 1404b are denoted by D in the same manner.
  • the length of one side of the second encoding units 1406a, 1406b, 1406c and 1406d in the form of the second encoding units 1406a, 1406b, 1406c and 1406d is 1/2 of the length of one side of the first encoding unit 1400, May be a depth of D + 1 that is one depth lower than D, which is the depth of the first encoding unit 1400.
  • the image decoding apparatus 2080 divides the first coding unit 1410 having a height greater than the width in the horizontal direction according to the information about the division mode, and generates a plurality of second coding units 1412a and 1412b , 1414a, 1414b, and 1414c. According to an embodiment, the image decoding apparatus 2080 divides the first encoding unit 1420 having a shape whose length is greater than the height, in the vertical direction according to the information on the division mode, and generates a plurality of second encoding units 1422a and 1422b , 1424a, 1424b, 1424c.
  • 1412b, 1414a, 1414b, 1414c. 1422a, 1422b, 1414c, 1414b, 1414c, 1414b, 1414c, 1414b, 1414c, 1414b, 1414c, 1424a, 1424b, 1424c can be determined in depth based on the length of the long side. For example, since the length of one side of the square-shaped second encoding units 1412a and 1412b is 1/2 times the length of one side of the non-square first encoding unit 1410 whose height is longer than the width, The depth of the second encoding units 1412a and 1412b of the form is D + 1 which is one depth lower than the depth D of the first encoding unit 1410 of the non-square form.
  • the image decoding apparatus 2080 can divide the non-square first encoding unit 1410 into odd number of second encoding units 1414a, 1414b and 1414c based on the information on the split mode mode.
  • the odd number of second encoding units 1414a, 1414b and 1414c may include non-square second encoding units 1414a and 1414c and a square second encoding unit 1414b.
  • the length of the long sides of the non-square type second encoding units 1414a and 1414c and the length of one side of the second encoding unit 1414b in the square form are set to 1/10 of the length of one side of the first encoding unit 1410
  • the depth of the second encoding units 1414a, 1414b, and 1414c may be a depth of D + 1 which is one depth lower than D, which is the depth of the first encoding unit 1410.
  • the image decoding apparatus 2080 is connected to the first encoding unit 1420 in the form of a non-square shape whose width is longer than the height in a manner corresponding to the scheme for determining the depths of the encoding units associated with the first encoding unit 1410 The depth of the encoding units can be determined.
  • the image decoding apparatus 2080 may calculate the size ratio between the coding units The index can be determined based on the index. Referring to FIG. 22, the coding unit 1414b positioned at the center among the odd-numbered coding units 1414a, 1414b, and 1414c has the same width as the other coding units 1414a and 1414c, Lt; / RTI > 1414a and 1414c. That is, in this case, the encoding unit 1414b positioned in the middle may include two of the other encoding units 1414a and 1414c.
  • the coding unit 1414c positioned next to the coding unit 1414c may be three days in which the index is increased by two. That is, there may be a discontinuity in the value of the index.
  • the image decoding apparatus 2080 may determine whether odd-numbered encoding units are not the same size based on the presence or absence of an index discontinuity for distinguishing between the divided encoding units.
  • the image decoding apparatus 2080 can determine whether the image is divided into a specific division form based on an index value for identifying a plurality of coding units divided and determined from the current coding unit. 22, the image decoding apparatus 2080 divides a first coding unit 1410 of a rectangular shape whose height is longer than the width to determine an even number of coding units 1412a and 1412b or an odd number of coding units 1414a and 1414b , And 1414c.
  • the image decoding apparatus 2080 may use an index (PID) indicating each coding unit in order to distinguish each of the plurality of coding units.
  • the PID may be obtained at a sample of a predetermined position of each coding unit (e.g., the upper left sample).
  • the image decoding apparatus 2080 may determine an encoding unit at a predetermined position among the encoding units determined by using the index for classifying the encoding unit. According to an exemplary embodiment, when the information on the split mode format for the rectangular first-type encoding unit 1410 whose height is longer than the width is divided into three encoding units, the image decoding apparatus 2080 decodes the first encoding unit 1410 can be divided into three coding units 1414a, 1414b, 1414c. The image decoding apparatus 2080 can assign an index to each of the three encoding units 1414a, 1414b, and 1414c.
  • the image decoding apparatus 2080 can compare the indexes of the respective coding units in order to determine the middle coding unit among the coding units divided into odd numbers.
  • the image decoding apparatus 2080 encodes the encoding unit 1414b having an index corresponding to the middle value among the indices based on the indexes of the encoding units by encoding the middle position among the encoding units determined by dividing the first encoding unit 1410 Can be determined as a unit.
  • the image decoding apparatus 2080 may determine an index based on a size ratio between coding units when the coding units are not the same size in determining the index for dividing the divided coding units .
  • the coding unit 1414b generated by dividing the first coding unit 1410 is divided into coding units 1414a and 1414c having the same width as the other coding units 1414a and 1414c but different in height Can be double the height.
  • the index (PID) of the coding unit 1414b positioned at the center is 1, the coding unit 1414c located next to the coding unit 1414c may be three days in which the index is increased by two.
  • the image decoding apparatus 2080 may determine that the image is divided into a plurality of encoding units including encoding units having different sizes from other encoding units.
  • the image decoding apparatus 2080 decodes the encoding unit (for example, the middle encoding unit) at a predetermined position among the odd number of encoding units
  • the current encoding unit can be divided into different sizes.
  • the image decoding apparatus 2080 can determine an encoding unit having a different size by using an index (PID) for the encoding unit.
  • PID index
  • the index and the size or position of the encoding unit at a predetermined position to be determined are specific for explaining an embodiment, and thus should not be construed to be limited thereto, and various indexes, positions and sizes of encoding units can be used Should be interpreted.
  • the image decoding apparatus 2080 may use a predetermined data unit in which recursive division of an encoding unit starts.
  • FIG. 23 illustrates that a plurality of coding units are determined according to a plurality of predetermined data units included in a picture according to an embodiment.
  • a predetermined data unit may be defined as a unit of data in which an encoding unit starts to be recursively segmented using at least one of block type information and information on a division mode mode. That is, it may correspond to a coding unit of the highest depth used in the process of determining a plurality of coding units for dividing the current picture.
  • a predetermined data unit is referred to as a reference data unit for convenience of explanation.
  • the reference data unit may represent a predetermined size and shape.
  • the reference encoding unit may comprise samples of MxN.
  • M and N may be equal to each other, or may be an integer represented by a multiplier of 2. That is, the reference data unit may represent a square or a non-square shape, and may be divided into an integer number of encoding units.
  • the image decoding apparatus 2080 may divide the current picture into a plurality of reference data units. According to an exemplary embodiment, the image decoding apparatus 2080 may divide a plurality of reference data units for dividing a current picture into pieces using information on a division mode mode for each reference data unit. The segmentation process of the reference data unit may correspond to the segmentation process using a quad-tree structure.
  • the image decoding apparatus 2080 can determine in advance the minimum size that the reference data unit included in the current picture can have. Accordingly, the image decoding apparatus 2080 can determine reference data units of various sizes having a size of a minimum size or more, and use at least one of the block type information and the division mode mode information based on the determined reference data unit The encoding unit can be determined.
  • the image decoding apparatus 2080 can use a square-shaped reference encoding unit 1500 or a non-square-shaped reference encoding unit 1502.
  • the type and size of the reference encoding unit may include various data units (e.g., a sequence, a picture, a slice, a slice segment a slice segment, a maximum encoding unit, and the like).
  • the entropy decoding unit 2055 of the image decoding apparatus 2080 may obtain at least one of the information on the type of the reference encoding unit and the size of the reference encoding unit from the bitstream for each of the various data units .
  • the process of determining at least one encoding unit included in the reference type encoding unit 1500 is described in detail in the process of dividing the current encoding unit 300 of FIG. 11, and the process of dividing the non- ) Is determined through the process of dividing the current encoding unit 400 or 450 of FIG. 12, so that a detailed description thereof will be omitted.
  • the image decoding apparatus 2080 may include an index for identifying the size and type of the reference encoding unit Can be used. That is, the entropy decoding unit 2055 decodes a predetermined condition (for example, data having a size smaller than a slice) among the various data units (e.g., a sequence, a picture, a slice, a slice segment, Unit), it is possible to obtain only an index for identification of the size and type of the reference encoding unit for each slice, slice segment, maximum encoding unit, and the like.
  • a predetermined condition for example, data having a size smaller than a slice
  • the various data units e.g., a sequence, a picture, a slice, a slice segment, Unit
  • the image decoding apparatus 2080 can determine the size and type of the reference data unit for each data unit that satisfies the predetermined condition by using the index.
  • the information on the type of the reference encoding unit and the information on the size of the reference encoding unit are obtained from the bitstream for each relatively small data unit and used, the use efficiency of the bitstream may not be good. Therefore, Information on the size of the reference encoding unit and information on the size of the reference encoding unit can be acquired and used. In this case, at least one of the size and the type of the reference encoding unit corresponding to the index indicating the size and type of the reference encoding unit may be predetermined.
  • the image decoding apparatus 2080 selects at least one of a size and a type of a predetermined reference encoding unit according to an index, thereby obtaining at least one of a size and a type of a reference encoding unit included in a data unit serving as a reference for obtaining an index You can decide.
  • the image decoding apparatus 2080 may use at least one reference encoding unit included in one maximum encoding unit. That is, the maximum encoding unit for dividing an image may include at least one reference encoding unit, and the encoding unit may be determined through a recursive division process of each reference encoding unit. According to an exemplary embodiment, at least one of the width and the height of the maximum encoding unit may correspond to at least one integer multiple of the width and height of the reference encoding unit. According to an exemplary embodiment, the size of the reference encoding unit may be a size obtained by dividing the maximum encoding unit n times according to a quadtree structure.
  • the image decoding apparatus 2080 can determine the reference encoding unit by dividing the maximum encoding unit n times according to the quad tree structure, and can calculate the reference encoding unit as block type information and information on the split mode Based on at least one of them.
  • FIG. 24 shows a processing block serving as a reference for determining a determination order of reference encoding units included in the picture 1600 according to an embodiment.
  • the image decoding apparatus 2080 can determine at least one processing block that divides a picture.
  • the processing block is a data unit including at least one reference encoding unit for dividing an image, and at least one reference encoding unit included in the processing block may be determined in a specific order. That is, the order of determination of at least one reference encoding unit determined in each processing block may correspond to one of various kinds of order in which the reference encoding unit can be determined, and the reference encoding unit determination order determined in each processing block May be different for each processing block.
  • the order of determination of the reference encoding unit determined for each processing block is a raster scan, a Z scan, an N scan, an up-right diagonal scan, a horizontal scan a horizontal scan, and a vertical scan. However, the order that can be determined should not be limited to the scan orders.
  • the image decoding apparatus 2080 may obtain information on the size of the processing block to determine the size of the at least one processing block included in the image.
  • the image decoding apparatus 2080 may obtain information on the size of the processing block from the bitstream to determine the size of the at least one processing block included in the image.
  • the size of such a processing block may be a predetermined size of a data unit represented by information on the size of the processing block.
  • the entropy decoding unit 2055 of the image decoding apparatus 2080 can obtain information on the size of the processing block from the bitstream for each specific data unit.
  • information on the size of a processing block can be obtained from a bitstream in units of data such as an image, a sequence, a picture, a slice, a slice segment, and the like. That is, the entropy decoding unit 2055 can obtain information on the size of the processing block from the bitstream for each of the plurality of data units, and the image decoding apparatus 2080 divides the picture using the information on the size of the obtained processing block
  • the size of the processing block may be an integer multiple of the reference encoding unit.
  • the image decoding apparatus 2080 can determine the sizes of the processing blocks 1602 and 1612 included in the picture 1600. [ For example, the image decoding apparatus 2080 can determine the size of the processing block based on information on the size of the processing block obtained from the bitstream. 24, the image decoding apparatus 2080 decodes the horizontal size of the processing blocks 1602 and 1612 by four times the horizontal size of the reference encoding unit, four times the vertical size of the reference encoding unit, You can decide. The image decoding apparatus 2080 can determine an order in which at least one reference encoding unit is determined in at least one processing block.
  • the video decoding apparatus 2080 can determine each processing block 1602, 1612 included in the picture 1600 based on the size of the processing block, and can include in the processing blocks 1602, 1612 The determination order of at least one reference encoding unit is determined.
  • the determination of the reference encoding unit may include determining the size of the reference encoding unit according to an embodiment.
  • the image decoding apparatus 2080 may obtain information on a determination order of at least one reference encoding unit included in at least one processing block from a bitstream, So that the order in which at least one reference encoding unit is determined can be determined.
  • the information on the decision order can be defined in the order or direction in which the reference encoding units are determined in the processing block. That is, the order in which the reference encoding units are determined may be independently determined for each processing block.
  • the image decoding apparatus 2080 may obtain information on a determination order of a reference encoding unit from a bit stream for each specific data unit.
  • the entropy decoding unit 2055 can obtain information on the determination order of the reference encoding unit from a bitstream for each data unit such as an image, a sequence, a picture, a slice, a slice segment, and a processing block. Since the information on the determination order of the reference encoding unit indicates the reference encoding unit determination order in the processing block, the information on the determination order can be obtained for each specific data unit including an integer number of processing blocks.
  • the image decoding apparatus 2080 may determine at least one reference encoding unit based on the determined order according to an embodiment.
  • the entropy decoding unit 2055 can obtain information on the reference encoding unit determination order from the bitstream as information related to the processing blocks 1602 and 1612, and the video decoding apparatus 2080 can obtain the information It is possible to determine the order of determining at least one reference encoding unit included in the blocks 1602 and 1612 and determine at least one reference encoding unit included in the picture 1600 according to the determination order of the encoding units. Referring to FIG. 24, the video decoding apparatus 2080 can determine the determination order 1604 and 1614 of at least one reference encoding unit associated with each of the processing blocks 1602 and 1612.
  • the reference encoding unit determination order associated with each processing block 1602, 1612 may be different for each processing block. If the reference encoding unit determination order 1604 related to the processing block 1602 is a raster scan order, the reference encoding unit included in the processing block 1602 can be determined according to the raster scan order. On the other hand, when the reference encoding unit determination order 1614 related to the other processing block 1612 is a reverse order of the raster scan order, the reference encoding unit included in the processing block 1612 can be determined according to the reverse order of the raster scan order.
  • the image decoding apparatus 2080 may decode the determined at least one reference encoding unit according to an embodiment.
  • the image decoding apparatus 2080 can decode an image based on the reference encoding unit determined through the above-described embodiment.
  • the method of decoding the reference encoding unit may include various methods of decoding the image.
  • the image decoding apparatus 2080 may obtain block type information indicating a type of a current encoding unit or information on a split mode mode indicating a method of dividing a current encoding unit from a bitstream.
  • Information about the block type information or the split mode mode may be included in a bitstream related to various data units.
  • the video decoding apparatus 2080 includes a sequence parameter set, a picture parameter set, a video parameter set, a slice header, a slice segment header slice block type information included in the segment header or information on the split mode mode can be used.
  • the image decoding apparatus 2080 can acquire and use a syntax element corresponding to information on the maximum encoding unit, the reference encoding unit, the block type information from the bit stream, or the split mode mode for each processing block from the bit stream.
  • the above-described embodiments of the present disclosure can be embodied in a general-purpose digital computer that can be created as a program that can be executed by a computer and operates the program using a computer-readable recording medium.
  • the computer readable recording medium includes a storage medium such as a magnetic storage medium (e.g., ROM, floppy disk, hard disk, etc.), optical reading medium (e.g., CD ROM, DVD, etc.).

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

Le procédé de décodage vidéo selon un mode de réalisation de la présente invention comprend les étapes consistant à : obtenir des informations de prédiction de blocs adjacents à l'intérieur d'un contour d'une image de référence, si un bloc courant se réfère à une région à l'extérieur d'une limite externe de l'image de référence pendant une référence de mouvement ; déterminer des échantillons de référence pour prédire des échantillons positionnés à l'extérieur du contour, en faisant intervenir des informations de prédiction d'un bloc adjacent à l'intérieur du contour ; et déterminer des échantillons positionnés à l'extérieur du contour de l'image de référence à l'aide des échantillons de référence.
PCT/KR2018/003803 2018-01-02 2018-03-30 Procédé et dipositif de codage vidéo et procédé et dispositif de décodage vidéo faisant appel à une technique basée sur une prédiction de mouvement WO2019135447A1 (fr)

Priority Applications (2)

Application Number Priority Date Filing Date Title
KR1020207017845A KR20200096551A (ko) 2018-01-02 2018-07-10 움직임 예측에 의한 패딩 기법을 이용한 비디오 부호화 방법 및 장치, 비디오 복호화 방법 및 장치
PCT/KR2018/007812 WO2019135457A1 (fr) 2018-01-02 2018-07-10 Procédé et appareil de codage vidéo et procédé et appareil de décodage vidéo, utilisant une technique de remplissage basée sur la prédiction de mouvement

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US201862612765P 2018-01-02 2018-01-02
US62/612,765 2018-01-02

Publications (1)

Publication Number Publication Date
WO2019135447A1 true WO2019135447A1 (fr) 2019-07-11

Family

ID=67144281

Family Applications (2)

Application Number Title Priority Date Filing Date
PCT/KR2018/003803 WO2019135447A1 (fr) 2018-01-02 2018-03-30 Procédé et dipositif de codage vidéo et procédé et dispositif de décodage vidéo faisant appel à une technique basée sur une prédiction de mouvement
PCT/KR2018/007812 WO2019135457A1 (fr) 2018-01-02 2018-07-10 Procédé et appareil de codage vidéo et procédé et appareil de décodage vidéo, utilisant une technique de remplissage basée sur la prédiction de mouvement

Family Applications After (1)

Application Number Title Priority Date Filing Date
PCT/KR2018/007812 WO2019135457A1 (fr) 2018-01-02 2018-07-10 Procédé et appareil de codage vidéo et procédé et appareil de décodage vidéo, utilisant une technique de remplissage basée sur la prédiction de mouvement

Country Status (2)

Country Link
KR (1) KR20200096551A (fr)
WO (2) WO2019135447A1 (fr)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20120118463A (ko) * 2010-01-18 2012-10-26 소니 주식회사 화상 처리 장치 및 방법, 및 프로그램
WO2013147557A1 (fr) * 2012-03-29 2013-10-03 엘지전자 주식회사 Procédé de prédiction entre couches, et dispositif de codage et dispositif de décodage l'utilisant
KR20140074377A (ko) * 2011-10-04 2014-06-17 퀄컴 인코포레이티드 비디오 코딩을 위한 모션 벡터 예측자 후보 클리핑 제거
WO2014093079A1 (fr) * 2012-12-12 2014-06-19 Qualcomm Incorporated Dispositif et procédé pour le codage échelonnable d'informations vidéo en fonction de codage vidéo hautement efficace
US20170353737A1 (en) * 2016-06-07 2017-12-07 Mediatek Inc. Method and Apparatus of Boundary Padding for VR Video Processing

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1582063B1 (fr) * 2003-01-07 2018-03-07 Thomson Licensing DTV Intercodage/intracodage vidéo mélange de partitions de macroblocs
US7440500B2 (en) * 2003-07-15 2008-10-21 Lsi Logic Corporation Supporting motion vectors outside picture boundaries in motion estimation process
DE602006020556D1 (de) * 2005-04-01 2011-04-21 Panasonic Corp Bilddecodierungsvorrichtung und bilddecodierungsverfahren
KR20140122189A (ko) * 2013-04-05 2014-10-17 한국전자통신연구원 계층 간 결합된 화면 내 예측을 이용한 영상 부호화/복호화 방법 및 그 장치
WO2015194896A1 (fr) * 2014-06-18 2015-12-23 삼성전자 주식회사 Procédé et dispositif d'encodage vidéo inter-couche pour compenser une différence de luminance, et procédé et dispositif de décodage vidéo

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20120118463A (ko) * 2010-01-18 2012-10-26 소니 주식회사 화상 처리 장치 및 방법, 및 프로그램
KR20140074377A (ko) * 2011-10-04 2014-06-17 퀄컴 인코포레이티드 비디오 코딩을 위한 모션 벡터 예측자 후보 클리핑 제거
WO2013147557A1 (fr) * 2012-03-29 2013-10-03 엘지전자 주식회사 Procédé de prédiction entre couches, et dispositif de codage et dispositif de décodage l'utilisant
WO2014093079A1 (fr) * 2012-12-12 2014-06-19 Qualcomm Incorporated Dispositif et procédé pour le codage échelonnable d'informations vidéo en fonction de codage vidéo hautement efficace
US20170353737A1 (en) * 2016-06-07 2017-12-07 Mediatek Inc. Method and Apparatus of Boundary Padding for VR Video Processing

Also Published As

Publication number Publication date
KR20200096551A (ko) 2020-08-12
WO2019135457A1 (fr) 2019-07-11

Similar Documents

Publication Publication Date Title
WO2021006692A1 (fr) Procédé et appareil de décodage vidéo, et procédé et appareil de codage vidéo
WO2019143093A1 (fr) Procédé et dispositif de décodage vidéo et procédé et dispositif de codage vidéo
WO2019066384A1 (fr) Procédé et dispositif de décodage vidéo utilisant une prédiction inter-composante, et procédé et dispositif de codage de vidéo utilisant une prédiction inter-composante
WO2019172676A1 (fr) Procédé et dispositif de décodage vidéo, et procédé et dispositif de codage vidéo
WO2020040619A1 (fr) Procédé et appareil de décodage vidéo, et procédé et appareil de codage vidéo
WO2020027551A1 (fr) Procédé et appareil de codage d'image, et procédé et appareil de décodage d'image
WO2019009502A1 (fr) Procédé et dispositif de codage vidéo, procédé et dispositif de décodage vidéo
WO2019135558A1 (fr) Procédé et appareil de décodage vidéo et procédé et appareil de codage vidéo
WO2020076097A1 (fr) Procédé de codage et de décodage vidéo à l'aide d'une valeur différentielle de vecteur de mouvement, et appareil de codage et de décodage d'informations de mouvement
WO2019168347A1 (fr) Procédé et appareil de décodage vidéo, et procédé et appareil de codage vidéo
WO2020013627A1 (fr) Procédé et dispositif de décodage vidéo, et procédé et dispositif de codage vidéo
WO2019066472A1 (fr) Procédé et appareil de codage d'image, et procédé et appareil de décodage d'image
WO2019066514A1 (fr) Procédé de codage et appareil associé, et procédé de décodage et appareil associé
WO2020076130A1 (fr) Procédé et dispositif de codage et de décodage vidéo utilisant des tuiles et des groupes de tuiles
WO2019009503A1 (fr) Procédé et dispositif d'encodage vidéo, procédé et dispositif de décodage vidéo
WO2020235951A1 (fr) Procédé et dispositif de codage d'image, et procédé et dispositif de décodage d'image
WO2017090968A1 (fr) Procédé pour coder/décoder une image et dispositif associé
WO2019009666A1 (fr) Procédé et appareil d'encodage d'image, et procédé et appareil de décodage d'image
WO2019209028A1 (fr) Procédé et dispositif de codage vidéo, et procédé et dispositif de décodage vidéo
WO2020040623A1 (fr) Procédé et appareil de codage d'image, et procédé et appareil de décodage d'image
WO2020117010A1 (fr) Procédé et dispositif de décodage de vidéo et procédé et dispositif d'encodage de vidéo
WO2019078630A1 (fr) Procédé et appareil de décodage vidéo, et procédé et appareil de codage vidéo
WO2019216710A1 (fr) Procédé de segmentation d'image et appareil de codage et de décodage d'image
WO2019066574A1 (fr) Procédé et dispositif de codage, et procédé et dispositif de décodage
WO2019216712A1 (fr) Procédé et appareil de codage de vidéo et procédé et appareil de décodage de vidéo

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 18898944

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 18898944

Country of ref document: EP

Kind code of ref document: A1